diff --git a/llava-next-video-7b_lora-True_qlora-False/README.md b/llava-next-video-7b_lora-True_qlora-False/README.md new file mode 100644 index 0000000000000000000000000000000000000000..972a036203d877262d3c6673f4d81814e7409dc5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/README.md @@ -0,0 +1,202 @@ +--- +base_model: llava-hf/LLaVA-NeXT-Video-7B-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/adapter_config.json b/llava-next-video-7b_lora-True_qlora-False/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..719f6e47b29561ff93d519680ccd115b480d825e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/adapter_config.json @@ -0,0 +1,132 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "llava-hf/LLaVA-NeXT-Video-7B-hf", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 8, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [], + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "model.layers.17.self_attn.q_proj", + "model.layers.17.self_attn.k_proj", + "model.layers.13.self_attn.k_proj", + "model.layers.20.self_attn.q_proj", + "model.layers.18.self_attn.k_proj", + "29.self_attn.k_proj", + "model.layers.16.self_attn.v_proj", + "model.layers.14.self_attn.q_proj", + "29.self_attn.v_proj", + "24.self_attn.k_proj", + "model.layers.18.self_attn.v_proj", + "model.layers.5.self_attn.q_proj", + "model.layers.11.self_attn.k_proj", + "27.self_attn.v_proj", + "model.layers.16.self_attn.q_proj", + "model.layers.10.self_attn.q_proj", + "model.layers.12.self_attn.v_proj", + "model.layers.20.self_attn.v_proj", + "model.layers.0.self_attn.v_proj", + "model.layers.7.self_attn.q_proj", + "model.layers.23.self_attn.q_proj", + "model.layers.1.self_attn.q_proj", + "25.self_attn.v_proj", + "model.layers.4.self_attn.v_proj", + "model.layers.19.self_attn.q_proj", + "model.layers.10.self_attn.k_proj", + "model.layers.22.self_attn.k_proj", + "model.layers.14.self_attn.v_proj", + "model.layers.13.self_attn.v_proj", + "25.self_attn.k_proj", + "24.self_attn.v_proj", + "model.layers.21.self_attn.q_proj", + "model.layers.8.self_attn.q_proj", + "model.layers.6.self_attn.q_proj", + "model.layers.6.self_attn.k_proj", + "model.layers.1.self_attn.v_proj", + "model.layers.9.self_attn.k_proj", + "model.layers.13.self_attn.q_proj", + "o_proj", + "model.layers.16.self_attn.k_proj", + "model.layers.6.self_attn.v_proj", + "model.layers.21.self_attn.k_proj", + "model.layers.5.self_attn.k_proj", + "27.self_attn.q_proj", + "model.layers.20.self_attn.k_proj", + "model.layers.11.self_attn.v_proj", + "model.layers.22.self_attn.q_proj", + "model.layers.14.self_attn.k_proj", + "28.self_attn.v_proj", + "model.layers.22.self_attn.v_proj", + "gate_proj", + "model.layers.0.self_attn.k_proj", + "model.layers.5.self_attn.v_proj", + "30.self_attn.q_proj", + "model.layers.3.self_attn.k_proj", + "29.self_attn.q_proj", + "model.layers.19.self_attn.k_proj", + "model.layers.10.self_attn.v_proj", + "model.layers.15.self_attn.k_proj", + "model.layers.7.self_attn.v_proj", + "model.layers.9.self_attn.v_proj", + "up_proj", + "model.layers.4.self_attn.q_proj", + "28.self_attn.q_proj", + "model.layers.3.self_attn.q_proj", + "24.self_attn.q_proj", + "30.self_attn.v_proj", + "model.layers.12.self_attn.k_proj", + "model.layers.23.self_attn.k_proj", + "26.self_attn.q_proj", + "28.self_attn.k_proj", + "31.self_attn.k_proj", + "down_proj", + "model.layers.2.self_attn.q_proj", + "31.self_attn.q_proj", + "model.layers.15.self_attn.q_proj", + "model.layers.12.self_attn.q_proj", + "model.layers.0.self_attn.q_proj", + "model.layers.2.self_attn.k_proj", + "model.layers.3.self_attn.v_proj", + "model.layers.11.self_attn.q_proj", + "25.self_attn.q_proj", + "model.layers.23.self_attn.v_proj", + "30.self_attn.k_proj", + "31.self_attn.v_proj", + "model.layers.15.self_attn.v_proj", + "model.layers.8.self_attn.v_proj", + "26.self_attn.k_proj", + "model.layers.19.self_attn.v_proj", + "model.layers.2.self_attn.v_proj", + "26.self_attn.v_proj", + "model.layers.21.self_attn.v_proj", + "model.layers.17.self_attn.v_proj", + "model.layers.8.self_attn.k_proj", + "model.layers.18.self_attn.q_proj", + "model.layers.7.self_attn.k_proj", + "model.layers.4.self_attn.k_proj", + "model.layers.9.self_attn.q_proj", + "27.self_attn.k_proj", + "model.layers.1.self_attn.k_proj" + ], + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/adapter_model.safetensors b/llava-next-video-7b_lora-True_qlora-False/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fca48c618e79d6c9d59165c4271e1468a35bf25e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcffdf981d90fd323630fa03bc9b94cc6c0d2e47a9bf7387c891fbcc8c4b54bc +size 40043208 diff --git a/llava-next-video-7b_lora-True_qlora-False/arguments/data.yaml b/llava-next-video-7b_lora-True_qlora-False/arguments/data.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ff26eea17400697c386f9b461e3cba07ef7fe58b --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/arguments/data.yaml @@ -0,0 +1,7 @@ +assistant_key: gpt +data_path: /home/nartay_aikyn/side_repo/CapERA_Llava-compatible_train.json +eval_data_path: null +image_folder: /home/nartay_aikyn/side_repo/Dataset +num_frames: 8 +user_key: human +video_folder: /home/nartay_aikyn/side_repo/Dataset diff --git a/llava-next-video-7b_lora-True_qlora-False/arguments/lora.yaml b/llava-next-video-7b_lora-True_qlora-False/arguments/lora.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3d32d18d50573b306bdb8441a811e4b4d8e151ab --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/arguments/lora.yaml @@ -0,0 +1,8 @@ +lora_alpha: 8 +lora_bias: none +lora_dropout: 0.05 +lora_r: 8 +lora_weight_path: '' +q_lora: false +use_lora: true +use_vision_lora: false diff --git a/llava-next-video-7b_lora-True_qlora-False/arguments/model.yaml b/llava-next-video-7b_lora-True_qlora-False/arguments/model.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4d2cd7f62c84c1861d2f668e7b6841010b6ec895 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/arguments/model.yaml @@ -0,0 +1,2 @@ +model_id: llava-next-video-7b +model_local_path: llava-hf/LLaVA-NeXT-Video-7B-hf diff --git a/llava-next-video-7b_lora-True_qlora-False/arguments/training.yaml b/llava-next-video-7b_lora-True_qlora-False/arguments/training.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6ad6f58a13c1aef3ef243e5f8df29bd6f12c2ef0 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/arguments/training.yaml @@ -0,0 +1,153 @@ +_n_gpu: 1 +accelerator_config: + dispatch_batches: null + even_batches: true + gradient_accumulation_kwargs: null + non_blocking: false + split_batches: false + use_configured_state: false + use_seedable_sampler: true +adafactor: false +adam_beta1: 0.9 +adam_beta2: 0.999 +adam_epsilon: 1.0e-08 +auto_find_batch_size: false +batch_eval_metrics: false +bf16: true +bf16_full_eval: false +data_seed: null +dataloader_drop_last: false +dataloader_num_workers: 4 +dataloader_persistent_workers: false +dataloader_pin_memory: true +dataloader_prefetch_factor: null +ddp_backend: null +ddp_broadcast_buffers: null +ddp_bucket_cap_mb: null +ddp_find_unused_parameters: null +ddp_timeout: 1800 +debug: [] +deepspeed: ./ds_configs/zero3.json +disable_tqdm: false +dispatch_batches: null +do_eval: true +do_predict: false +do_train: false +eval_accumulation_steps: null +eval_delay: 0 +eval_do_concat_batches: true +eval_on_start: false +eval_steps: null +eval_strategy: !!python/object/apply:transformers.trainer_utils.IntervalStrategy +- epoch +eval_use_gather_object: false +evaluation_strategy: null +fp16: false +fp16_backend: auto +fp16_full_eval: false +fp16_opt_level: O1 +fsdp: [] +fsdp_config: + min_num_params: 0 + xla: false + xla_fsdp_grad_ckpt: false + xla_fsdp_v2: false +fsdp_min_num_params: 0 +fsdp_transformer_layer_cls_to_wrap: null +full_determinism: false +gradient_accumulation_steps: 1 +gradient_checkpointing: true +gradient_checkpointing_kwargs: null +greater_is_better: null +group_by_length: false +half_precision_backend: auto +hub_always_push: false +hub_model_id: null +hub_private_repo: false +hub_strategy: !!python/object/apply:transformers.trainer_utils.HubStrategy +- every_save +hub_token: null +ignore_data_skip: false +include_inputs_for_metrics: false +include_num_input_tokens_seen: false +include_tokens_per_second: false +jit_mode_eval: false +label_names: null +label_smoothing_factor: 0.0 +learning_rate: 2.0e-05 +length_column_name: length +load_best_model_at_end: false +local_rank: 1 +log_level: passive +log_level_replica: warning +log_on_each_node: true +logging_dir: ./checkpoints/llava-next-video-7b_lora-True_qlora-False/runs/Jun19_11-26-10_node007 +logging_first_step: false +logging_nan_inf_filter: true +logging_steps: 1.0 +logging_strategy: !!python/object/apply:transformers.trainer_utils.IntervalStrategy +- steps +lr_scheduler_kwargs: {} +lr_scheduler_type: !!python/object/apply:transformers.trainer_utils.SchedulerType +- cosine +mask_question_tokens: true +max_grad_norm: 1.0 +max_steps: -1 +metric_for_best_model: null +model_max_length: 4096 +mp_parameters: '' +neftune_noise_alpha: null +no_cuda: false +num_train_epochs: 5.0 +optim: !!python/object/apply:transformers.training_args.OptimizerNames +- adamw_torch +optim_args: null +optim_target_modules: null +output_dir: ./checkpoints/llava-next-video-7b_lora-True_qlora-False +overwrite_output_dir: false +past_index: -1 +per_device_eval_batch_size: 2 +per_device_train_batch_size: 2 +per_gpu_eval_batch_size: null +per_gpu_train_batch_size: null +prediction_loss_only: false +push_to_hub: false +push_to_hub_model_id: null +push_to_hub_organization: null +push_to_hub_token: null +ray_scope: last +remove_unused_columns: false +report_to: +- wandb +restore_callback_states_from_checkpoint: false +resume_from_checkpoint: null +run_name: llava-next-video-7b_lora-True_qlora-False +save_on_each_node: false +save_only_model: false +save_safetensors: true +save_steps: 500 +save_strategy: !!python/object/apply:transformers.trainer_utils.IntervalStrategy +- epoch +save_total_limit: 5 +seed: 42 +skip_memory_metrics: true +split_batches: null +tf32: true +torch_compile: false +torch_compile_backend: null +torch_compile_mode: null +torch_empty_cache_steps: null +torchdynamo: null +tpu_metrics_debug: false +tpu_num_cores: null +train_vision_encoder: false +train_vision_projector: false +use_cpu: false +use_flash_attn: false +use_ipex: false +use_legacy_prediction_loop: false +use_liger_kernel: false +use_mps_device: false +warmup_ratio: 0.03 +warmup_steps: 0 +weight_decay: 0.0 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/README.md b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/README.md new file mode 100644 index 0000000000000000000000000000000000000000..972a036203d877262d3c6673f4d81814e7409dc5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/README.md @@ -0,0 +1,202 @@ +--- +base_model: llava-hf/LLaVA-NeXT-Video-7B-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/adapter_config.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..719f6e47b29561ff93d519680ccd115b480d825e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/adapter_config.json @@ -0,0 +1,132 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "llava-hf/LLaVA-NeXT-Video-7B-hf", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 8, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [], + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "model.layers.17.self_attn.q_proj", + "model.layers.17.self_attn.k_proj", + "model.layers.13.self_attn.k_proj", + "model.layers.20.self_attn.q_proj", + "model.layers.18.self_attn.k_proj", + "29.self_attn.k_proj", + "model.layers.16.self_attn.v_proj", + "model.layers.14.self_attn.q_proj", + "29.self_attn.v_proj", + "24.self_attn.k_proj", + "model.layers.18.self_attn.v_proj", + "model.layers.5.self_attn.q_proj", + "model.layers.11.self_attn.k_proj", + "27.self_attn.v_proj", + "model.layers.16.self_attn.q_proj", + "model.layers.10.self_attn.q_proj", + "model.layers.12.self_attn.v_proj", + "model.layers.20.self_attn.v_proj", + "model.layers.0.self_attn.v_proj", + "model.layers.7.self_attn.q_proj", + "model.layers.23.self_attn.q_proj", + "model.layers.1.self_attn.q_proj", + "25.self_attn.v_proj", + "model.layers.4.self_attn.v_proj", + "model.layers.19.self_attn.q_proj", + "model.layers.10.self_attn.k_proj", + "model.layers.22.self_attn.k_proj", + "model.layers.14.self_attn.v_proj", + "model.layers.13.self_attn.v_proj", + "25.self_attn.k_proj", + "24.self_attn.v_proj", + "model.layers.21.self_attn.q_proj", + "model.layers.8.self_attn.q_proj", + "model.layers.6.self_attn.q_proj", + "model.layers.6.self_attn.k_proj", + "model.layers.1.self_attn.v_proj", + "model.layers.9.self_attn.k_proj", + "model.layers.13.self_attn.q_proj", + "o_proj", + "model.layers.16.self_attn.k_proj", + "model.layers.6.self_attn.v_proj", + "model.layers.21.self_attn.k_proj", + "model.layers.5.self_attn.k_proj", + "27.self_attn.q_proj", + "model.layers.20.self_attn.k_proj", + "model.layers.11.self_attn.v_proj", + "model.layers.22.self_attn.q_proj", + "model.layers.14.self_attn.k_proj", + "28.self_attn.v_proj", + "model.layers.22.self_attn.v_proj", + "gate_proj", + "model.layers.0.self_attn.k_proj", + "model.layers.5.self_attn.v_proj", + "30.self_attn.q_proj", + "model.layers.3.self_attn.k_proj", + "29.self_attn.q_proj", + "model.layers.19.self_attn.k_proj", + "model.layers.10.self_attn.v_proj", + "model.layers.15.self_attn.k_proj", + "model.layers.7.self_attn.v_proj", + "model.layers.9.self_attn.v_proj", + "up_proj", + "model.layers.4.self_attn.q_proj", + "28.self_attn.q_proj", + "model.layers.3.self_attn.q_proj", + "24.self_attn.q_proj", + "30.self_attn.v_proj", + "model.layers.12.self_attn.k_proj", + "model.layers.23.self_attn.k_proj", + "26.self_attn.q_proj", + "28.self_attn.k_proj", + "31.self_attn.k_proj", + "down_proj", + "model.layers.2.self_attn.q_proj", + "31.self_attn.q_proj", + "model.layers.15.self_attn.q_proj", + "model.layers.12.self_attn.q_proj", + "model.layers.0.self_attn.q_proj", + "model.layers.2.self_attn.k_proj", + "model.layers.3.self_attn.v_proj", + "model.layers.11.self_attn.q_proj", + "25.self_attn.q_proj", + "model.layers.23.self_attn.v_proj", + "30.self_attn.k_proj", + "31.self_attn.v_proj", + "model.layers.15.self_attn.v_proj", + "model.layers.8.self_attn.v_proj", + "26.self_attn.k_proj", + "model.layers.19.self_attn.v_proj", + "model.layers.2.self_attn.v_proj", + "26.self_attn.v_proj", + "model.layers.21.self_attn.v_proj", + "model.layers.17.self_attn.v_proj", + "model.layers.8.self_attn.k_proj", + "model.layers.18.self_attn.q_proj", + "model.layers.7.self_attn.k_proj", + "model.layers.4.self_attn.k_proj", + "model.layers.9.self_attn.q_proj", + "27.self_attn.k_proj", + "model.layers.1.self_attn.k_proj" + ], + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/adapter_model.safetensors b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7c86e619cb6e86541b91e5bb1743fa44634d5ef4 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a0784065d2ea5f7b1eb0216ac23c972c97203881c9652334b00e87118e58a96 +size 40043208 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..1de52637298a6fb5e70bc15a2028631e66f49a3b --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80e7b9fa18944de352a72ce8239ba84b00a1dd107251c760da2e9634db681dfb +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..3b18a87eca054441b581d8029ecdccf57cd55669 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4ae56ab40ed2b5aafa2b4c421c95f7beef77ba1c34641daaf05b1e834758cc4 +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/zero_pp_rank_0_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/zero_pp_rank_0_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..36b8b8213ba46df043b51d42dca9872ec42c5145 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/zero_pp_rank_0_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63e12daf98ca3feba8a8c02d54ef3250248b154e3cf81c723745a86c9cfa3965 +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/zero_pp_rank_1_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/zero_pp_rank_1_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..f00d5cea9986bcf6a5abd1811f65ee6eedf6cc3c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/global_step1842/zero_pp_rank_1_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2a22ce0f42f50675e10f4517cd1d1d4f3ca52b09fef0ca3ae3f7ff9f7ca28c2 +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/latest b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/latest new file mode 100644 index 0000000000000000000000000000000000000000..c01fa95c617fda050f0a01110a8327a70aa12d7a --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/latest @@ -0,0 +1 @@ +global_step1842 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/rng_state_0.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..39c7c1e2eb461b13184e31da4c0400624eb327d0 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2acff294a85a25208b8e58089c5fcb48583b1dece9ba42e71b67d1bdbd55e4d5 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/rng_state_1.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..b9a8a1cf194e321b5230dbfb4a25011bd8d50968 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75e7fb492cb7725c64b09d28f65f074fb81ca826fce818d005a70169f3b1f697 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/scheduler.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..201d1d35518ca504205cee68e9c5c8de202719f2 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88b3137e149efb3dd118254475f25d13e3e2f97096b045874d201baa08a6ead0 +size 1064 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/trainer_state.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a69d42c3fa519337b77bc65bd7d933c6ee9e11e5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/trainer_state.json @@ -0,0 +1,12927 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 1842, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005428881650380022, + "grad_norm": 2.1320506281889067, + "learning_rate": 7.220216606498195e-08, + "loss": 5.1936, + "step": 1 + }, + { + "epoch": 0.0010857763300760044, + "grad_norm": 1.8633488766761366, + "learning_rate": 1.444043321299639e-07, + "loss": 4.6995, + "step": 2 + }, + { + "epoch": 0.0016286644951140066, + "grad_norm": 2.063286098879776, + "learning_rate": 2.1660649819494586e-07, + "loss": 4.4902, + "step": 3 + }, + { + "epoch": 0.002171552660152009, + "grad_norm": 1.7123050230907495, + "learning_rate": 2.888086642599278e-07, + "loss": 4.9125, + "step": 4 + }, + { + "epoch": 0.0027144408251900108, + "grad_norm": 1.7499657276799891, + "learning_rate": 3.610108303249098e-07, + "loss": 3.7648, + "step": 5 + }, + { + "epoch": 0.003257328990228013, + "grad_norm": 2.0419045636066886, + "learning_rate": 4.332129963898917e-07, + "loss": 4.7842, + "step": 6 + }, + { + "epoch": 0.003800217155266015, + "grad_norm": 1.4960282529105982, + "learning_rate": 5.054151624548736e-07, + "loss": 4.9333, + "step": 7 + }, + { + "epoch": 0.004343105320304018, + "grad_norm": 1.640954967476452, + "learning_rate": 5.776173285198556e-07, + "loss": 3.9576, + "step": 8 + }, + { + "epoch": 0.004885993485342019, + "grad_norm": 2.0590500459610244, + "learning_rate": 6.498194945848375e-07, + "loss": 4.3833, + "step": 9 + }, + { + "epoch": 0.0054288816503800215, + "grad_norm": 1.07709353257816, + "learning_rate": 7.220216606498196e-07, + "loss": 3.7669, + "step": 10 + }, + { + "epoch": 0.005971769815418024, + "grad_norm": 1.9506235655782926, + "learning_rate": 7.942238267148016e-07, + "loss": 4.5975, + "step": 11 + }, + { + "epoch": 0.006514657980456026, + "grad_norm": 1.411983096954528, + "learning_rate": 8.664259927797834e-07, + "loss": 3.9706, + "step": 12 + }, + { + "epoch": 0.007057546145494028, + "grad_norm": 1.7110053131090028, + "learning_rate": 9.386281588447654e-07, + "loss": 4.9081, + "step": 13 + }, + { + "epoch": 0.00760043431053203, + "grad_norm": 1.3689315757199692, + "learning_rate": 1.0108303249097473e-06, + "loss": 4.2426, + "step": 14 + }, + { + "epoch": 0.008143322475570033, + "grad_norm": 1.5612091671030481, + "learning_rate": 1.0830324909747293e-06, + "loss": 4.5144, + "step": 15 + }, + { + "epoch": 0.008686210640608035, + "grad_norm": 1.3725168953865754, + "learning_rate": 1.1552346570397112e-06, + "loss": 4.9378, + "step": 16 + }, + { + "epoch": 0.009229098805646038, + "grad_norm": 2.027359913808926, + "learning_rate": 1.2274368231046932e-06, + "loss": 3.7477, + "step": 17 + }, + { + "epoch": 0.009771986970684038, + "grad_norm": 2.1695456962160224, + "learning_rate": 1.299638989169675e-06, + "loss": 5.7636, + "step": 18 + }, + { + "epoch": 0.01031487513572204, + "grad_norm": 3.2637002287680428, + "learning_rate": 1.3718411552346572e-06, + "loss": 6.1883, + "step": 19 + }, + { + "epoch": 0.010857763300760043, + "grad_norm": 2.5723616120643857, + "learning_rate": 1.4440433212996392e-06, + "loss": 4.0267, + "step": 20 + }, + { + "epoch": 0.011400651465798045, + "grad_norm": 1.4473754577877487, + "learning_rate": 1.516245487364621e-06, + "loss": 4.526, + "step": 21 + }, + { + "epoch": 0.011943539630836048, + "grad_norm": 1.6098208731761, + "learning_rate": 1.5884476534296031e-06, + "loss": 4.5163, + "step": 22 + }, + { + "epoch": 0.01248642779587405, + "grad_norm": 2.899624682573202, + "learning_rate": 1.6606498194945849e-06, + "loss": 4.8188, + "step": 23 + }, + { + "epoch": 0.013029315960912053, + "grad_norm": 2.273571202553131, + "learning_rate": 1.7328519855595669e-06, + "loss": 5.0074, + "step": 24 + }, + { + "epoch": 0.013572204125950055, + "grad_norm": 1.1877323869171315, + "learning_rate": 1.805054151624549e-06, + "loss": 4.4331, + "step": 25 + }, + { + "epoch": 0.014115092290988056, + "grad_norm": 1.9365501019385494, + "learning_rate": 1.8772563176895308e-06, + "loss": 5.0525, + "step": 26 + }, + { + "epoch": 0.014657980456026058, + "grad_norm": 1.7710485953105986, + "learning_rate": 1.949458483754513e-06, + "loss": 4.1346, + "step": 27 + }, + { + "epoch": 0.01520086862106406, + "grad_norm": 1.231848881292397, + "learning_rate": 2.0216606498194946e-06, + "loss": 4.2431, + "step": 28 + }, + { + "epoch": 0.015743756786102063, + "grad_norm": 1.6372838462475754, + "learning_rate": 2.0938628158844768e-06, + "loss": 5.4317, + "step": 29 + }, + { + "epoch": 0.016286644951140065, + "grad_norm": 1.7326568199523718, + "learning_rate": 2.1660649819494585e-06, + "loss": 4.4774, + "step": 30 + }, + { + "epoch": 0.016829533116178068, + "grad_norm": 2.1315641946923387, + "learning_rate": 2.2382671480144407e-06, + "loss": 4.1418, + "step": 31 + }, + { + "epoch": 0.01737242128121607, + "grad_norm": 1.916538366799816, + "learning_rate": 2.3104693140794225e-06, + "loss": 5.7926, + "step": 32 + }, + { + "epoch": 0.017915309446254073, + "grad_norm": 2.057531274363787, + "learning_rate": 2.3826714801444047e-06, + "loss": 4.2862, + "step": 33 + }, + { + "epoch": 0.018458197611292075, + "grad_norm": 2.6727877745045023, + "learning_rate": 2.4548736462093864e-06, + "loss": 5.1323, + "step": 34 + }, + { + "epoch": 0.019001085776330078, + "grad_norm": 1.5043525736330898, + "learning_rate": 2.527075812274368e-06, + "loss": 4.0634, + "step": 35 + }, + { + "epoch": 0.019543973941368076, + "grad_norm": 2.6324248758076862, + "learning_rate": 2.59927797833935e-06, + "loss": 5.4211, + "step": 36 + }, + { + "epoch": 0.02008686210640608, + "grad_norm": 1.6581905646911415, + "learning_rate": 2.6714801444043326e-06, + "loss": 4.9825, + "step": 37 + }, + { + "epoch": 0.02062975027144408, + "grad_norm": 1.6667970880898455, + "learning_rate": 2.7436823104693144e-06, + "loss": 4.3219, + "step": 38 + }, + { + "epoch": 0.021172638436482084, + "grad_norm": 1.7448150724822071, + "learning_rate": 2.815884476534296e-06, + "loss": 4.3912, + "step": 39 + }, + { + "epoch": 0.021715526601520086, + "grad_norm": 3.720503777726443, + "learning_rate": 2.8880866425992783e-06, + "loss": 6.0592, + "step": 40 + }, + { + "epoch": 0.02225841476655809, + "grad_norm": 1.6853001074010774, + "learning_rate": 2.96028880866426e-06, + "loss": 3.8395, + "step": 41 + }, + { + "epoch": 0.02280130293159609, + "grad_norm": 1.797029067943667, + "learning_rate": 3.032490974729242e-06, + "loss": 4.4651, + "step": 42 + }, + { + "epoch": 0.023344191096634093, + "grad_norm": 1.6298824221707349, + "learning_rate": 3.1046931407942245e-06, + "loss": 4.4507, + "step": 43 + }, + { + "epoch": 0.023887079261672096, + "grad_norm": 1.4793189209059368, + "learning_rate": 3.1768953068592062e-06, + "loss": 4.7849, + "step": 44 + }, + { + "epoch": 0.024429967426710098, + "grad_norm": 1.1741346570472608, + "learning_rate": 3.249097472924188e-06, + "loss": 3.5897, + "step": 45 + }, + { + "epoch": 0.0249728555917481, + "grad_norm": 2.1468867306169064, + "learning_rate": 3.3212996389891698e-06, + "loss": 3.8969, + "step": 46 + }, + { + "epoch": 0.025515743756786103, + "grad_norm": 1.657494000697021, + "learning_rate": 3.393501805054152e-06, + "loss": 4.3412, + "step": 47 + }, + { + "epoch": 0.026058631921824105, + "grad_norm": 1.7125864055710684, + "learning_rate": 3.4657039711191337e-06, + "loss": 3.8059, + "step": 48 + }, + { + "epoch": 0.026601520086862108, + "grad_norm": 2.9515347000925622, + "learning_rate": 3.5379061371841155e-06, + "loss": 4.1069, + "step": 49 + }, + { + "epoch": 0.02714440825190011, + "grad_norm": 3.0185188918098262, + "learning_rate": 3.610108303249098e-06, + "loss": 5.2646, + "step": 50 + }, + { + "epoch": 0.02768729641693811, + "grad_norm": 1.5546972252207274, + "learning_rate": 3.68231046931408e-06, + "loss": 3.7212, + "step": 51 + }, + { + "epoch": 0.02823018458197611, + "grad_norm": 2.6376587426282474, + "learning_rate": 3.7545126353790616e-06, + "loss": 4.2862, + "step": 52 + }, + { + "epoch": 0.028773072747014114, + "grad_norm": 2.053905428064986, + "learning_rate": 3.826714801444043e-06, + "loss": 3.8304, + "step": 53 + }, + { + "epoch": 0.029315960912052116, + "grad_norm": 2.9083183656849725, + "learning_rate": 3.898916967509026e-06, + "loss": 4.8223, + "step": 54 + }, + { + "epoch": 0.02985884907709012, + "grad_norm": 1.9919239162425852, + "learning_rate": 3.971119133574007e-06, + "loss": 3.9054, + "step": 55 + }, + { + "epoch": 0.03040173724212812, + "grad_norm": 2.4076122411223135, + "learning_rate": 4.043321299638989e-06, + "loss": 4.8568, + "step": 56 + }, + { + "epoch": 0.030944625407166124, + "grad_norm": 3.854434602830178, + "learning_rate": 4.115523465703971e-06, + "loss": 5.7771, + "step": 57 + }, + { + "epoch": 0.031487513572204126, + "grad_norm": 2.510303564095041, + "learning_rate": 4.1877256317689535e-06, + "loss": 5.5851, + "step": 58 + }, + { + "epoch": 0.03203040173724213, + "grad_norm": 2.8884448366576305, + "learning_rate": 4.259927797833936e-06, + "loss": 4.1774, + "step": 59 + }, + { + "epoch": 0.03257328990228013, + "grad_norm": 2.0893778739721, + "learning_rate": 4.332129963898917e-06, + "loss": 4.4679, + "step": 60 + }, + { + "epoch": 0.03311617806731813, + "grad_norm": 1.5816439039579229, + "learning_rate": 4.404332129963899e-06, + "loss": 4.1999, + "step": 61 + }, + { + "epoch": 0.033659066232356136, + "grad_norm": 1.9591574137365761, + "learning_rate": 4.4765342960288814e-06, + "loss": 4.1539, + "step": 62 + }, + { + "epoch": 0.03420195439739414, + "grad_norm": 3.1533909413995924, + "learning_rate": 4.548736462093864e-06, + "loss": 4.5842, + "step": 63 + }, + { + "epoch": 0.03474484256243214, + "grad_norm": 2.543715437360663, + "learning_rate": 4.620938628158845e-06, + "loss": 4.1473, + "step": 64 + }, + { + "epoch": 0.03528773072747014, + "grad_norm": 2.7608093188531955, + "learning_rate": 4.693140794223827e-06, + "loss": 4.9354, + "step": 65 + }, + { + "epoch": 0.035830618892508145, + "grad_norm": 4.192163162864289, + "learning_rate": 4.765342960288809e-06, + "loss": 5.4263, + "step": 66 + }, + { + "epoch": 0.03637350705754615, + "grad_norm": 2.809206875027472, + "learning_rate": 4.837545126353791e-06, + "loss": 4.4259, + "step": 67 + }, + { + "epoch": 0.03691639522258415, + "grad_norm": 3.9318073942863534, + "learning_rate": 4.909747292418773e-06, + "loss": 4.1483, + "step": 68 + }, + { + "epoch": 0.03745928338762215, + "grad_norm": 2.935065385226121, + "learning_rate": 4.981949458483755e-06, + "loss": 4.6854, + "step": 69 + }, + { + "epoch": 0.038002171552660155, + "grad_norm": 2.3687140876301305, + "learning_rate": 5.054151624548736e-06, + "loss": 4.5875, + "step": 70 + }, + { + "epoch": 0.03854505971769816, + "grad_norm": 2.987709377109457, + "learning_rate": 5.126353790613719e-06, + "loss": 5.5784, + "step": 71 + }, + { + "epoch": 0.03908794788273615, + "grad_norm": 2.618514747422557, + "learning_rate": 5.1985559566787e-06, + "loss": 3.9901, + "step": 72 + }, + { + "epoch": 0.039630836047774155, + "grad_norm": 3.729239515680902, + "learning_rate": 5.270758122743683e-06, + "loss": 4.6192, + "step": 73 + }, + { + "epoch": 0.04017372421281216, + "grad_norm": 2.488658936956188, + "learning_rate": 5.342960288808665e-06, + "loss": 3.9889, + "step": 74 + }, + { + "epoch": 0.04071661237785016, + "grad_norm": 2.6055640979236223, + "learning_rate": 5.415162454873647e-06, + "loss": 4.3166, + "step": 75 + }, + { + "epoch": 0.04125950054288816, + "grad_norm": 2.091905540809132, + "learning_rate": 5.487364620938629e-06, + "loss": 3.5848, + "step": 76 + }, + { + "epoch": 0.041802388707926165, + "grad_norm": 4.153140315460812, + "learning_rate": 5.559566787003611e-06, + "loss": 4.4889, + "step": 77 + }, + { + "epoch": 0.04234527687296417, + "grad_norm": 4.828033123535034, + "learning_rate": 5.631768953068592e-06, + "loss": 4.886, + "step": 78 + }, + { + "epoch": 0.04288816503800217, + "grad_norm": 3.106146715385438, + "learning_rate": 5.7039711191335744e-06, + "loss": 4.3698, + "step": 79 + }, + { + "epoch": 0.04343105320304017, + "grad_norm": 4.458698539182903, + "learning_rate": 5.776173285198557e-06, + "loss": 4.7018, + "step": 80 + }, + { + "epoch": 0.043973941368078175, + "grad_norm": 4.619902873416999, + "learning_rate": 5.848375451263538e-06, + "loss": 4.3449, + "step": 81 + }, + { + "epoch": 0.04451682953311618, + "grad_norm": 2.411841615517987, + "learning_rate": 5.92057761732852e-06, + "loss": 3.9952, + "step": 82 + }, + { + "epoch": 0.04505971769815418, + "grad_norm": 3.6594837742651705, + "learning_rate": 5.992779783393502e-06, + "loss": 4.1308, + "step": 83 + }, + { + "epoch": 0.04560260586319218, + "grad_norm": 2.3760888273966074, + "learning_rate": 6.064981949458484e-06, + "loss": 4.0315, + "step": 84 + }, + { + "epoch": 0.046145494028230184, + "grad_norm": 4.579655616004927, + "learning_rate": 6.137184115523466e-06, + "loss": 4.1051, + "step": 85 + }, + { + "epoch": 0.04668838219326819, + "grad_norm": 3.5909339610827846, + "learning_rate": 6.209386281588449e-06, + "loss": 4.3126, + "step": 86 + }, + { + "epoch": 0.04723127035830619, + "grad_norm": 3.1030422711991625, + "learning_rate": 6.28158844765343e-06, + "loss": 4.3038, + "step": 87 + }, + { + "epoch": 0.04777415852334419, + "grad_norm": 4.134211324790428, + "learning_rate": 6.3537906137184125e-06, + "loss": 4.906, + "step": 88 + }, + { + "epoch": 0.048317046688382194, + "grad_norm": 3.538946913288816, + "learning_rate": 6.425992779783395e-06, + "loss": 4.2182, + "step": 89 + }, + { + "epoch": 0.048859934853420196, + "grad_norm": 1.9828152543749473, + "learning_rate": 6.498194945848376e-06, + "loss": 3.8487, + "step": 90 + }, + { + "epoch": 0.0494028230184582, + "grad_norm": 2.127969936938475, + "learning_rate": 6.570397111913358e-06, + "loss": 3.9221, + "step": 91 + }, + { + "epoch": 0.0499457111834962, + "grad_norm": 2.740653026382279, + "learning_rate": 6.6425992779783395e-06, + "loss": 4.4661, + "step": 92 + }, + { + "epoch": 0.050488599348534204, + "grad_norm": 3.278496788216742, + "learning_rate": 6.714801444043322e-06, + "loss": 4.1995, + "step": 93 + }, + { + "epoch": 0.051031487513572206, + "grad_norm": 1.5455749978381657, + "learning_rate": 6.787003610108304e-06, + "loss": 3.3941, + "step": 94 + }, + { + "epoch": 0.05157437567861021, + "grad_norm": 2.6938438745410975, + "learning_rate": 6.859205776173285e-06, + "loss": 3.9863, + "step": 95 + }, + { + "epoch": 0.05211726384364821, + "grad_norm": 2.754193515919841, + "learning_rate": 6.9314079422382674e-06, + "loss": 4.1621, + "step": 96 + }, + { + "epoch": 0.05266015200868621, + "grad_norm": 1.787757508585929, + "learning_rate": 7.00361010830325e-06, + "loss": 3.728, + "step": 97 + }, + { + "epoch": 0.053203040173724216, + "grad_norm": 1.6756343581979376, + "learning_rate": 7.075812274368231e-06, + "loss": 3.4483, + "step": 98 + }, + { + "epoch": 0.05374592833876222, + "grad_norm": 3.0418610323064317, + "learning_rate": 7.148014440433214e-06, + "loss": 4.1131, + "step": 99 + }, + { + "epoch": 0.05428881650380022, + "grad_norm": 2.710846119482301, + "learning_rate": 7.220216606498196e-06, + "loss": 5.0449, + "step": 100 + }, + { + "epoch": 0.054831704668838216, + "grad_norm": 1.906382949957388, + "learning_rate": 7.2924187725631776e-06, + "loss": 3.3181, + "step": 101 + }, + { + "epoch": 0.05537459283387622, + "grad_norm": 2.4669878697837, + "learning_rate": 7.36462093862816e-06, + "loss": 3.7996, + "step": 102 + }, + { + "epoch": 0.05591748099891422, + "grad_norm": 2.3755842712374586, + "learning_rate": 7.436823104693142e-06, + "loss": 4.0321, + "step": 103 + }, + { + "epoch": 0.05646036916395222, + "grad_norm": 2.032793864149989, + "learning_rate": 7.509025270758123e-06, + "loss": 3.4121, + "step": 104 + }, + { + "epoch": 0.057003257328990226, + "grad_norm": 2.1561554352810686, + "learning_rate": 7.5812274368231055e-06, + "loss": 4.2585, + "step": 105 + }, + { + "epoch": 0.05754614549402823, + "grad_norm": 1.9627108654626784, + "learning_rate": 7.653429602888087e-06, + "loss": 4.3085, + "step": 106 + }, + { + "epoch": 0.05808903365906623, + "grad_norm": 1.8622753095278954, + "learning_rate": 7.72563176895307e-06, + "loss": 3.8459, + "step": 107 + }, + { + "epoch": 0.05863192182410423, + "grad_norm": 1.4438765470100654, + "learning_rate": 7.797833935018051e-06, + "loss": 3.4914, + "step": 108 + }, + { + "epoch": 0.059174809989142235, + "grad_norm": 1.6164085905580508, + "learning_rate": 7.870036101083033e-06, + "loss": 3.4958, + "step": 109 + }, + { + "epoch": 0.05971769815418024, + "grad_norm": 1.6075014519578008, + "learning_rate": 7.942238267148014e-06, + "loss": 3.202, + "step": 110 + }, + { + "epoch": 0.06026058631921824, + "grad_norm": 1.7630994896021268, + "learning_rate": 8.014440433212997e-06, + "loss": 3.0916, + "step": 111 + }, + { + "epoch": 0.06080347448425624, + "grad_norm": 1.7600666006428431, + "learning_rate": 8.086642599277978e-06, + "loss": 3.5791, + "step": 112 + }, + { + "epoch": 0.061346362649294245, + "grad_norm": 2.324243905399748, + "learning_rate": 8.158844765342961e-06, + "loss": 3.6247, + "step": 113 + }, + { + "epoch": 0.06188925081433225, + "grad_norm": 1.2715449173786617, + "learning_rate": 8.231046931407943e-06, + "loss": 3.2381, + "step": 114 + }, + { + "epoch": 0.06243213897937025, + "grad_norm": 1.9029014476619814, + "learning_rate": 8.303249097472926e-06, + "loss": 3.5269, + "step": 115 + }, + { + "epoch": 0.06297502714440825, + "grad_norm": 1.8782574049467182, + "learning_rate": 8.375451263537907e-06, + "loss": 2.8698, + "step": 116 + }, + { + "epoch": 0.06351791530944625, + "grad_norm": 1.8522383705853351, + "learning_rate": 8.447653429602888e-06, + "loss": 3.5003, + "step": 117 + }, + { + "epoch": 0.06406080347448426, + "grad_norm": 1.6480852413427443, + "learning_rate": 8.519855595667871e-06, + "loss": 3.5665, + "step": 118 + }, + { + "epoch": 0.06460369163952226, + "grad_norm": 2.1628448626122685, + "learning_rate": 8.592057761732853e-06, + "loss": 3.343, + "step": 119 + }, + { + "epoch": 0.06514657980456026, + "grad_norm": 1.568456113423694, + "learning_rate": 8.664259927797834e-06, + "loss": 2.474, + "step": 120 + }, + { + "epoch": 0.06568946796959826, + "grad_norm": 1.8054357065473743, + "learning_rate": 8.736462093862817e-06, + "loss": 2.9228, + "step": 121 + }, + { + "epoch": 0.06623235613463627, + "grad_norm": 1.9418450135728782, + "learning_rate": 8.808664259927798e-06, + "loss": 3.1995, + "step": 122 + }, + { + "epoch": 0.06677524429967427, + "grad_norm": 1.321851521267606, + "learning_rate": 8.88086642599278e-06, + "loss": 3.5515, + "step": 123 + }, + { + "epoch": 0.06731813246471227, + "grad_norm": 1.390198419578882, + "learning_rate": 8.953068592057763e-06, + "loss": 2.7741, + "step": 124 + }, + { + "epoch": 0.06786102062975027, + "grad_norm": 2.119370139523294, + "learning_rate": 9.025270758122744e-06, + "loss": 3.3427, + "step": 125 + }, + { + "epoch": 0.06840390879478828, + "grad_norm": 2.042189734142807, + "learning_rate": 9.097472924187727e-06, + "loss": 3.1414, + "step": 126 + }, + { + "epoch": 0.06894679695982628, + "grad_norm": 2.0399787373405123, + "learning_rate": 9.169675090252709e-06, + "loss": 3.0862, + "step": 127 + }, + { + "epoch": 0.06948968512486428, + "grad_norm": 2.0523727295223915, + "learning_rate": 9.24187725631769e-06, + "loss": 2.6461, + "step": 128 + }, + { + "epoch": 0.07003257328990228, + "grad_norm": 1.459190908283032, + "learning_rate": 9.314079422382673e-06, + "loss": 2.9907, + "step": 129 + }, + { + "epoch": 0.07057546145494029, + "grad_norm": 1.7403521896723462, + "learning_rate": 9.386281588447654e-06, + "loss": 3.0681, + "step": 130 + }, + { + "epoch": 0.07111834961997829, + "grad_norm": 2.2283604183714383, + "learning_rate": 9.458483754512636e-06, + "loss": 3.4619, + "step": 131 + }, + { + "epoch": 0.07166123778501629, + "grad_norm": 1.3291461832292721, + "learning_rate": 9.530685920577619e-06, + "loss": 3.096, + "step": 132 + }, + { + "epoch": 0.0722041259500543, + "grad_norm": 1.7839357457835638, + "learning_rate": 9.6028880866426e-06, + "loss": 2.7512, + "step": 133 + }, + { + "epoch": 0.0727470141150923, + "grad_norm": 1.7836465130823984, + "learning_rate": 9.675090252707581e-06, + "loss": 3.3118, + "step": 134 + }, + { + "epoch": 0.0732899022801303, + "grad_norm": 2.0407172747925135, + "learning_rate": 9.747292418772564e-06, + "loss": 2.1032, + "step": 135 + }, + { + "epoch": 0.0738327904451683, + "grad_norm": 2.335226908691354, + "learning_rate": 9.819494584837546e-06, + "loss": 2.9039, + "step": 136 + }, + { + "epoch": 0.0743756786102063, + "grad_norm": 2.567081845339778, + "learning_rate": 9.891696750902527e-06, + "loss": 3.1071, + "step": 137 + }, + { + "epoch": 0.0749185667752443, + "grad_norm": 1.9549184920601244, + "learning_rate": 9.96389891696751e-06, + "loss": 2.8253, + "step": 138 + }, + { + "epoch": 0.07546145494028231, + "grad_norm": 1.6389815727698025, + "learning_rate": 1.0036101083032491e-05, + "loss": 2.5701, + "step": 139 + }, + { + "epoch": 0.07600434310532031, + "grad_norm": 1.6558924195158018, + "learning_rate": 1.0108303249097473e-05, + "loss": 2.6428, + "step": 140 + }, + { + "epoch": 0.07654723127035831, + "grad_norm": 2.0257343427634087, + "learning_rate": 1.0180505415162456e-05, + "loss": 2.2828, + "step": 141 + }, + { + "epoch": 0.07709011943539631, + "grad_norm": 1.879991343961764, + "learning_rate": 1.0252707581227437e-05, + "loss": 3.1436, + "step": 142 + }, + { + "epoch": 0.07763300760043432, + "grad_norm": 1.8718236498893788, + "learning_rate": 1.032490974729242e-05, + "loss": 3.2434, + "step": 143 + }, + { + "epoch": 0.0781758957654723, + "grad_norm": 1.8252190287923398, + "learning_rate": 1.03971119133574e-05, + "loss": 2.4851, + "step": 144 + }, + { + "epoch": 0.07871878393051031, + "grad_norm": 2.6564413093838213, + "learning_rate": 1.0469314079422383e-05, + "loss": 2.7315, + "step": 145 + }, + { + "epoch": 0.07926167209554831, + "grad_norm": 2.054556112475997, + "learning_rate": 1.0541516245487366e-05, + "loss": 2.647, + "step": 146 + }, + { + "epoch": 0.07980456026058631, + "grad_norm": 1.072343898976271, + "learning_rate": 1.0613718411552347e-05, + "loss": 2.5868, + "step": 147 + }, + { + "epoch": 0.08034744842562432, + "grad_norm": 1.1443585178571194, + "learning_rate": 1.068592057761733e-05, + "loss": 2.3247, + "step": 148 + }, + { + "epoch": 0.08089033659066232, + "grad_norm": 2.0424670761496646, + "learning_rate": 1.0758122743682312e-05, + "loss": 1.989, + "step": 149 + }, + { + "epoch": 0.08143322475570032, + "grad_norm": 1.3715615162918033, + "learning_rate": 1.0830324909747295e-05, + "loss": 2.3989, + "step": 150 + }, + { + "epoch": 0.08197611292073832, + "grad_norm": 1.3740116291532023, + "learning_rate": 1.0902527075812274e-05, + "loss": 1.8726, + "step": 151 + }, + { + "epoch": 0.08251900108577633, + "grad_norm": 2.5824759155435655, + "learning_rate": 1.0974729241877257e-05, + "loss": 2.3019, + "step": 152 + }, + { + "epoch": 0.08306188925081433, + "grad_norm": 1.5584514490971877, + "learning_rate": 1.1046931407942239e-05, + "loss": 2.4209, + "step": 153 + }, + { + "epoch": 0.08360477741585233, + "grad_norm": 1.6184456176727229, + "learning_rate": 1.1119133574007222e-05, + "loss": 2.2377, + "step": 154 + }, + { + "epoch": 0.08414766558089033, + "grad_norm": 2.6665151568462586, + "learning_rate": 1.1191335740072201e-05, + "loss": 2.6381, + "step": 155 + }, + { + "epoch": 0.08469055374592833, + "grad_norm": 1.442929928712006, + "learning_rate": 1.1263537906137184e-05, + "loss": 2.3405, + "step": 156 + }, + { + "epoch": 0.08523344191096634, + "grad_norm": 1.906547769868643, + "learning_rate": 1.1335740072202166e-05, + "loss": 2.4391, + "step": 157 + }, + { + "epoch": 0.08577633007600434, + "grad_norm": 1.2153727696829317, + "learning_rate": 1.1407942238267149e-05, + "loss": 1.8401, + "step": 158 + }, + { + "epoch": 0.08631921824104234, + "grad_norm": 1.4810729195972283, + "learning_rate": 1.1480144404332132e-05, + "loss": 2.7248, + "step": 159 + }, + { + "epoch": 0.08686210640608034, + "grad_norm": 2.9401118999625058, + "learning_rate": 1.1552346570397113e-05, + "loss": 2.6582, + "step": 160 + }, + { + "epoch": 0.08740499457111835, + "grad_norm": 1.145106458573816, + "learning_rate": 1.1624548736462096e-05, + "loss": 2.6648, + "step": 161 + }, + { + "epoch": 0.08794788273615635, + "grad_norm": 1.9228128025198137, + "learning_rate": 1.1696750902527076e-05, + "loss": 2.3702, + "step": 162 + }, + { + "epoch": 0.08849077090119435, + "grad_norm": 1.318978856783607, + "learning_rate": 1.1768953068592059e-05, + "loss": 1.9388, + "step": 163 + }, + { + "epoch": 0.08903365906623235, + "grad_norm": 1.58104775083856, + "learning_rate": 1.184115523465704e-05, + "loss": 2.0774, + "step": 164 + }, + { + "epoch": 0.08957654723127036, + "grad_norm": 1.8206646323260607, + "learning_rate": 1.1913357400722023e-05, + "loss": 2.5221, + "step": 165 + }, + { + "epoch": 0.09011943539630836, + "grad_norm": 1.0815960351491627, + "learning_rate": 1.1985559566787005e-05, + "loss": 2.128, + "step": 166 + }, + { + "epoch": 0.09066232356134636, + "grad_norm": 1.3341486432977188, + "learning_rate": 1.2057761732851988e-05, + "loss": 1.9418, + "step": 167 + }, + { + "epoch": 0.09120521172638436, + "grad_norm": 1.3772486036170464, + "learning_rate": 1.2129963898916967e-05, + "loss": 1.84, + "step": 168 + }, + { + "epoch": 0.09174809989142237, + "grad_norm": 1.42874152479176, + "learning_rate": 1.220216606498195e-05, + "loss": 2.2172, + "step": 169 + }, + { + "epoch": 0.09229098805646037, + "grad_norm": 1.9691394708584433, + "learning_rate": 1.2274368231046932e-05, + "loss": 2.9247, + "step": 170 + }, + { + "epoch": 0.09283387622149837, + "grad_norm": 1.4773974635910447, + "learning_rate": 1.2346570397111915e-05, + "loss": 2.1028, + "step": 171 + }, + { + "epoch": 0.09337676438653637, + "grad_norm": 1.5937344901476087, + "learning_rate": 1.2418772563176898e-05, + "loss": 2.4073, + "step": 172 + }, + { + "epoch": 0.09391965255157438, + "grad_norm": 1.630783959399718, + "learning_rate": 1.2490974729241878e-05, + "loss": 2.6978, + "step": 173 + }, + { + "epoch": 0.09446254071661238, + "grad_norm": 1.44736447594164, + "learning_rate": 1.256317689530686e-05, + "loss": 1.8851, + "step": 174 + }, + { + "epoch": 0.09500542888165038, + "grad_norm": 1.3909459856510267, + "learning_rate": 1.2635379061371842e-05, + "loss": 1.576, + "step": 175 + }, + { + "epoch": 0.09554831704668838, + "grad_norm": 1.7838393677575515, + "learning_rate": 1.2707581227436825e-05, + "loss": 2.3895, + "step": 176 + }, + { + "epoch": 0.09609120521172639, + "grad_norm": 1.1294127385799255, + "learning_rate": 1.2779783393501806e-05, + "loss": 2.4855, + "step": 177 + }, + { + "epoch": 0.09663409337676439, + "grad_norm": 1.259662421149648, + "learning_rate": 1.285198555956679e-05, + "loss": 2.5469, + "step": 178 + }, + { + "epoch": 0.09717698154180239, + "grad_norm": 1.4621367635995284, + "learning_rate": 1.2924187725631769e-05, + "loss": 1.9015, + "step": 179 + }, + { + "epoch": 0.09771986970684039, + "grad_norm": 1.6960126954903645, + "learning_rate": 1.2996389891696752e-05, + "loss": 2.9545, + "step": 180 + }, + { + "epoch": 0.0982627578718784, + "grad_norm": 1.5419022503429414, + "learning_rate": 1.3068592057761733e-05, + "loss": 1.9912, + "step": 181 + }, + { + "epoch": 0.0988056460369164, + "grad_norm": 1.4073422496120207, + "learning_rate": 1.3140794223826716e-05, + "loss": 2.3353, + "step": 182 + }, + { + "epoch": 0.0993485342019544, + "grad_norm": 1.2881659254471862, + "learning_rate": 1.3212996389891696e-05, + "loss": 2.7784, + "step": 183 + }, + { + "epoch": 0.0998914223669924, + "grad_norm": 1.361969663712106, + "learning_rate": 1.3285198555956679e-05, + "loss": 2.523, + "step": 184 + }, + { + "epoch": 0.1004343105320304, + "grad_norm": 2.5841731119141316, + "learning_rate": 1.3357400722021662e-05, + "loss": 2.43, + "step": 185 + }, + { + "epoch": 0.10097719869706841, + "grad_norm": 1.3203537847943392, + "learning_rate": 1.3429602888086643e-05, + "loss": 1.9404, + "step": 186 + }, + { + "epoch": 0.10152008686210641, + "grad_norm": 1.393097901493087, + "learning_rate": 1.3501805054151626e-05, + "loss": 2.6202, + "step": 187 + }, + { + "epoch": 0.10206297502714441, + "grad_norm": 0.9414377380652101, + "learning_rate": 1.3574007220216608e-05, + "loss": 2.0879, + "step": 188 + }, + { + "epoch": 0.10260586319218241, + "grad_norm": 1.6547581806701488, + "learning_rate": 1.3646209386281591e-05, + "loss": 2.8912, + "step": 189 + }, + { + "epoch": 0.10314875135722042, + "grad_norm": 1.1236261864015666, + "learning_rate": 1.371841155234657e-05, + "loss": 2.1802, + "step": 190 + }, + { + "epoch": 0.10369163952225842, + "grad_norm": 1.1766384878462062, + "learning_rate": 1.3790613718411554e-05, + "loss": 1.9127, + "step": 191 + }, + { + "epoch": 0.10423452768729642, + "grad_norm": 2.0453724866488656, + "learning_rate": 1.3862815884476535e-05, + "loss": 1.8522, + "step": 192 + }, + { + "epoch": 0.10477741585233442, + "grad_norm": 1.550964370734908, + "learning_rate": 1.3935018050541518e-05, + "loss": 2.5852, + "step": 193 + }, + { + "epoch": 0.10532030401737243, + "grad_norm": 1.303227180334529, + "learning_rate": 1.40072202166065e-05, + "loss": 2.407, + "step": 194 + }, + { + "epoch": 0.10586319218241043, + "grad_norm": 1.2009193115283474, + "learning_rate": 1.4079422382671482e-05, + "loss": 1.8978, + "step": 195 + }, + { + "epoch": 0.10640608034744843, + "grad_norm": 1.353648963417667, + "learning_rate": 1.4151624548736462e-05, + "loss": 2.283, + "step": 196 + }, + { + "epoch": 0.10694896851248643, + "grad_norm": 1.266154745219424, + "learning_rate": 1.4223826714801445e-05, + "loss": 2.0679, + "step": 197 + }, + { + "epoch": 0.10749185667752444, + "grad_norm": 4.186447043886807, + "learning_rate": 1.4296028880866428e-05, + "loss": 1.9175, + "step": 198 + }, + { + "epoch": 0.10803474484256244, + "grad_norm": 1.147065238667786, + "learning_rate": 1.436823104693141e-05, + "loss": 2.2748, + "step": 199 + }, + { + "epoch": 0.10857763300760044, + "grad_norm": 1.2760699256137895, + "learning_rate": 1.4440433212996392e-05, + "loss": 2.3031, + "step": 200 + }, + { + "epoch": 0.10912052117263844, + "grad_norm": 0.9452152631580967, + "learning_rate": 1.4512635379061372e-05, + "loss": 1.4833, + "step": 201 + }, + { + "epoch": 0.10966340933767643, + "grad_norm": 1.3980490034032074, + "learning_rate": 1.4584837545126355e-05, + "loss": 1.5865, + "step": 202 + }, + { + "epoch": 0.11020629750271443, + "grad_norm": 1.6440838011007093, + "learning_rate": 1.4657039711191336e-05, + "loss": 1.632, + "step": 203 + }, + { + "epoch": 0.11074918566775244, + "grad_norm": 1.3969764794458068, + "learning_rate": 1.472924187725632e-05, + "loss": 1.4301, + "step": 204 + }, + { + "epoch": 0.11129207383279044, + "grad_norm": 1.223663090997016, + "learning_rate": 1.48014440433213e-05, + "loss": 2.3993, + "step": 205 + }, + { + "epoch": 0.11183496199782844, + "grad_norm": 1.183881027275754, + "learning_rate": 1.4873646209386284e-05, + "loss": 1.4987, + "step": 206 + }, + { + "epoch": 0.11237785016286644, + "grad_norm": 1.3788483357578956, + "learning_rate": 1.4945848375451264e-05, + "loss": 2.7742, + "step": 207 + }, + { + "epoch": 0.11292073832790445, + "grad_norm": 1.1260622559595288, + "learning_rate": 1.5018050541516247e-05, + "loss": 1.755, + "step": 208 + }, + { + "epoch": 0.11346362649294245, + "grad_norm": 1.8439396062381286, + "learning_rate": 1.5090252707581228e-05, + "loss": 2.9875, + "step": 209 + }, + { + "epoch": 0.11400651465798045, + "grad_norm": 1.2789617727573501, + "learning_rate": 1.5162454873646211e-05, + "loss": 2.0119, + "step": 210 + }, + { + "epoch": 0.11454940282301845, + "grad_norm": 1.3385913561767113, + "learning_rate": 1.5234657039711192e-05, + "loss": 1.7693, + "step": 211 + }, + { + "epoch": 0.11509229098805646, + "grad_norm": 1.049162702021749, + "learning_rate": 1.5306859205776174e-05, + "loss": 1.7541, + "step": 212 + }, + { + "epoch": 0.11563517915309446, + "grad_norm": 1.0797407129441892, + "learning_rate": 1.537906137184116e-05, + "loss": 2.0427, + "step": 213 + }, + { + "epoch": 0.11617806731813246, + "grad_norm": 1.3661735952693994, + "learning_rate": 1.545126353790614e-05, + "loss": 2.722, + "step": 214 + }, + { + "epoch": 0.11672095548317046, + "grad_norm": 1.7084933140661187, + "learning_rate": 1.552346570397112e-05, + "loss": 2.2125, + "step": 215 + }, + { + "epoch": 0.11726384364820847, + "grad_norm": 1.3814428141700028, + "learning_rate": 1.5595667870036102e-05, + "loss": 2.448, + "step": 216 + }, + { + "epoch": 0.11780673181324647, + "grad_norm": 1.298309196615407, + "learning_rate": 1.5667870036101084e-05, + "loss": 2.2706, + "step": 217 + }, + { + "epoch": 0.11834961997828447, + "grad_norm": 1.4553148411944858, + "learning_rate": 1.5740072202166065e-05, + "loss": 2.0087, + "step": 218 + }, + { + "epoch": 0.11889250814332247, + "grad_norm": 1.5083115470870359, + "learning_rate": 1.581227436823105e-05, + "loss": 1.7179, + "step": 219 + }, + { + "epoch": 0.11943539630836048, + "grad_norm": 1.2335179878839475, + "learning_rate": 1.5884476534296028e-05, + "loss": 2.5337, + "step": 220 + }, + { + "epoch": 0.11997828447339848, + "grad_norm": 1.1729282902419653, + "learning_rate": 1.5956678700361013e-05, + "loss": 1.9604, + "step": 221 + }, + { + "epoch": 0.12052117263843648, + "grad_norm": 1.5700347334422509, + "learning_rate": 1.6028880866425994e-05, + "loss": 2.3655, + "step": 222 + }, + { + "epoch": 0.12106406080347448, + "grad_norm": 1.435764746128967, + "learning_rate": 1.6101083032490975e-05, + "loss": 2.109, + "step": 223 + }, + { + "epoch": 0.12160694896851248, + "grad_norm": 1.0866216936301165, + "learning_rate": 1.6173285198555957e-05, + "loss": 2.3334, + "step": 224 + }, + { + "epoch": 0.12214983713355049, + "grad_norm": 1.8060789361277199, + "learning_rate": 1.624548736462094e-05, + "loss": 1.6605, + "step": 225 + }, + { + "epoch": 0.12269272529858849, + "grad_norm": 1.539101200174871, + "learning_rate": 1.6317689530685923e-05, + "loss": 2.2597, + "step": 226 + }, + { + "epoch": 0.12323561346362649, + "grad_norm": 1.7392984350243024, + "learning_rate": 1.6389891696750904e-05, + "loss": 2.5979, + "step": 227 + }, + { + "epoch": 0.1237785016286645, + "grad_norm": 1.6792982939949113, + "learning_rate": 1.6462093862815885e-05, + "loss": 2.0626, + "step": 228 + }, + { + "epoch": 0.1243213897937025, + "grad_norm": 2.067856070276011, + "learning_rate": 1.6534296028880867e-05, + "loss": 2.056, + "step": 229 + }, + { + "epoch": 0.1248642779587405, + "grad_norm": 1.2773224600362139, + "learning_rate": 1.660649819494585e-05, + "loss": 1.5852, + "step": 230 + }, + { + "epoch": 0.1254071661237785, + "grad_norm": 1.381245750855955, + "learning_rate": 1.6678700361010833e-05, + "loss": 2.0675, + "step": 231 + }, + { + "epoch": 0.1259500542888165, + "grad_norm": 1.4215755121774605, + "learning_rate": 1.6750902527075814e-05, + "loss": 2.1002, + "step": 232 + }, + { + "epoch": 0.1264929424538545, + "grad_norm": 1.4581673093175722, + "learning_rate": 1.6823104693140795e-05, + "loss": 1.9279, + "step": 233 + }, + { + "epoch": 0.1270358306188925, + "grad_norm": 1.2656102558648035, + "learning_rate": 1.6895306859205777e-05, + "loss": 1.4913, + "step": 234 + }, + { + "epoch": 0.1275787187839305, + "grad_norm": 1.3985246778048182, + "learning_rate": 1.6967509025270758e-05, + "loss": 2.1629, + "step": 235 + }, + { + "epoch": 0.1281216069489685, + "grad_norm": 1.757194392177075, + "learning_rate": 1.7039711191335743e-05, + "loss": 2.5688, + "step": 236 + }, + { + "epoch": 0.12866449511400652, + "grad_norm": 1.3766112824147358, + "learning_rate": 1.711191335740072e-05, + "loss": 2.3344, + "step": 237 + }, + { + "epoch": 0.12920738327904452, + "grad_norm": 1.2985043105700416, + "learning_rate": 1.7184115523465706e-05, + "loss": 2.0626, + "step": 238 + }, + { + "epoch": 0.12975027144408252, + "grad_norm": 1.715796674825951, + "learning_rate": 1.7256317689530687e-05, + "loss": 2.1559, + "step": 239 + }, + { + "epoch": 0.13029315960912052, + "grad_norm": 1.6374209306607368, + "learning_rate": 1.7328519855595668e-05, + "loss": 2.0801, + "step": 240 + }, + { + "epoch": 0.13083604777415853, + "grad_norm": 1.244682506367769, + "learning_rate": 1.7400722021660653e-05, + "loss": 1.6125, + "step": 241 + }, + { + "epoch": 0.13137893593919653, + "grad_norm": 1.3001838781366861, + "learning_rate": 1.7472924187725634e-05, + "loss": 2.0934, + "step": 242 + }, + { + "epoch": 0.13192182410423453, + "grad_norm": 1.1600879910417528, + "learning_rate": 1.7545126353790616e-05, + "loss": 1.7862, + "step": 243 + }, + { + "epoch": 0.13246471226927253, + "grad_norm": 1.3503312968411385, + "learning_rate": 1.7617328519855597e-05, + "loss": 2.1636, + "step": 244 + }, + { + "epoch": 0.13300760043431054, + "grad_norm": 2.1299761628528167, + "learning_rate": 1.768953068592058e-05, + "loss": 2.4826, + "step": 245 + }, + { + "epoch": 0.13355048859934854, + "grad_norm": 1.3232052544377066, + "learning_rate": 1.776173285198556e-05, + "loss": 2.0497, + "step": 246 + }, + { + "epoch": 0.13409337676438654, + "grad_norm": 1.5079892467642568, + "learning_rate": 1.7833935018050544e-05, + "loss": 2.4552, + "step": 247 + }, + { + "epoch": 0.13463626492942454, + "grad_norm": 1.716355318914646, + "learning_rate": 1.7906137184115526e-05, + "loss": 1.5309, + "step": 248 + }, + { + "epoch": 0.13517915309446255, + "grad_norm": 2.1523878955969193, + "learning_rate": 1.7978339350180507e-05, + "loss": 2.7403, + "step": 249 + }, + { + "epoch": 0.13572204125950055, + "grad_norm": 1.8169301927997485, + "learning_rate": 1.805054151624549e-05, + "loss": 2.3324, + "step": 250 + }, + { + "epoch": 0.13626492942453855, + "grad_norm": 1.4464924096148426, + "learning_rate": 1.812274368231047e-05, + "loss": 1.8399, + "step": 251 + }, + { + "epoch": 0.13680781758957655, + "grad_norm": 1.9818298103877654, + "learning_rate": 1.8194945848375454e-05, + "loss": 2.4237, + "step": 252 + }, + { + "epoch": 0.13735070575461455, + "grad_norm": 1.6371003739671923, + "learning_rate": 1.8267148014440436e-05, + "loss": 2.0477, + "step": 253 + }, + { + "epoch": 0.13789359391965256, + "grad_norm": 1.2944832148952543, + "learning_rate": 1.8339350180505417e-05, + "loss": 2.2269, + "step": 254 + }, + { + "epoch": 0.13843648208469056, + "grad_norm": 1.4284289887933213, + "learning_rate": 1.84115523465704e-05, + "loss": 2.2465, + "step": 255 + }, + { + "epoch": 0.13897937024972856, + "grad_norm": 1.4633226693151502, + "learning_rate": 1.848375451263538e-05, + "loss": 2.1582, + "step": 256 + }, + { + "epoch": 0.13952225841476656, + "grad_norm": 1.6169546473154692, + "learning_rate": 1.855595667870036e-05, + "loss": 1.9526, + "step": 257 + }, + { + "epoch": 0.14006514657980457, + "grad_norm": 1.47082651679944, + "learning_rate": 1.8628158844765346e-05, + "loss": 2.0314, + "step": 258 + }, + { + "epoch": 0.14060803474484257, + "grad_norm": 1.2279964838302115, + "learning_rate": 1.8700361010830327e-05, + "loss": 1.4345, + "step": 259 + }, + { + "epoch": 0.14115092290988057, + "grad_norm": 1.5727648286246085, + "learning_rate": 1.877256317689531e-05, + "loss": 2.7371, + "step": 260 + }, + { + "epoch": 0.14169381107491857, + "grad_norm": 1.5536143039045338, + "learning_rate": 1.884476534296029e-05, + "loss": 1.6691, + "step": 261 + }, + { + "epoch": 0.14223669923995658, + "grad_norm": 1.504953118413033, + "learning_rate": 1.891696750902527e-05, + "loss": 1.8422, + "step": 262 + }, + { + "epoch": 0.14277958740499458, + "grad_norm": 1.6092868309857633, + "learning_rate": 1.8989169675090253e-05, + "loss": 1.6446, + "step": 263 + }, + { + "epoch": 0.14332247557003258, + "grad_norm": 1.434484546738067, + "learning_rate": 1.9061371841155237e-05, + "loss": 1.5945, + "step": 264 + }, + { + "epoch": 0.14386536373507058, + "grad_norm": 1.3472204852366787, + "learning_rate": 1.913357400722022e-05, + "loss": 1.4511, + "step": 265 + }, + { + "epoch": 0.1444082519001086, + "grad_norm": 2.493842366649246, + "learning_rate": 1.92057761732852e-05, + "loss": 2.1144, + "step": 266 + }, + { + "epoch": 0.1449511400651466, + "grad_norm": 1.3138428233046642, + "learning_rate": 1.927797833935018e-05, + "loss": 1.6592, + "step": 267 + }, + { + "epoch": 0.1454940282301846, + "grad_norm": 1.7201458829291465, + "learning_rate": 1.9350180505415163e-05, + "loss": 1.6895, + "step": 268 + }, + { + "epoch": 0.1460369163952226, + "grad_norm": 1.9777013455620547, + "learning_rate": 1.9422382671480147e-05, + "loss": 2.2628, + "step": 269 + }, + { + "epoch": 0.1465798045602606, + "grad_norm": 1.8559862870685497, + "learning_rate": 1.949458483754513e-05, + "loss": 1.3374, + "step": 270 + }, + { + "epoch": 0.1471226927252986, + "grad_norm": 1.919404449298866, + "learning_rate": 1.956678700361011e-05, + "loss": 2.3805, + "step": 271 + }, + { + "epoch": 0.1476655808903366, + "grad_norm": 1.4760010543412088, + "learning_rate": 1.963898916967509e-05, + "loss": 1.9791, + "step": 272 + }, + { + "epoch": 0.1482084690553746, + "grad_norm": 1.4491842831658432, + "learning_rate": 1.9711191335740073e-05, + "loss": 1.8119, + "step": 273 + }, + { + "epoch": 0.1487513572204126, + "grad_norm": 1.4098796806726177, + "learning_rate": 1.9783393501805054e-05, + "loss": 1.7398, + "step": 274 + }, + { + "epoch": 0.1492942453854506, + "grad_norm": 1.5009369942229867, + "learning_rate": 1.985559566787004e-05, + "loss": 1.5194, + "step": 275 + }, + { + "epoch": 0.1498371335504886, + "grad_norm": 1.5073462671342261, + "learning_rate": 1.992779783393502e-05, + "loss": 1.8125, + "step": 276 + }, + { + "epoch": 0.1503800217155266, + "grad_norm": 1.9315604890805302, + "learning_rate": 2e-05, + "loss": 1.5429, + "step": 277 + }, + { + "epoch": 0.15092290988056462, + "grad_norm": 1.5059106777628657, + "learning_rate": 1.999999938159203e-05, + "loss": 1.6402, + "step": 278 + }, + { + "epoch": 0.15146579804560262, + "grad_norm": 1.7255630320192266, + "learning_rate": 1.9999997526368205e-05, + "loss": 2.2806, + "step": 279 + }, + { + "epoch": 0.15200868621064062, + "grad_norm": 1.7985407961594548, + "learning_rate": 1.999999443432874e-05, + "loss": 2.5213, + "step": 280 + }, + { + "epoch": 0.15255157437567862, + "grad_norm": 1.801471230172198, + "learning_rate": 1.999999010547403e-05, + "loss": 1.8362, + "step": 281 + }, + { + "epoch": 0.15309446254071662, + "grad_norm": 3.236881896883975, + "learning_rate": 1.999998453980461e-05, + "loss": 2.7148, + "step": 282 + }, + { + "epoch": 0.15363735070575463, + "grad_norm": 1.7343786362856708, + "learning_rate": 1.9999977737321156e-05, + "loss": 1.8195, + "step": 283 + }, + { + "epoch": 0.15418023887079263, + "grad_norm": 1.5006942215232695, + "learning_rate": 1.999996969802452e-05, + "loss": 1.8744, + "step": 284 + }, + { + "epoch": 0.15472312703583063, + "grad_norm": 4.171967924853298, + "learning_rate": 1.99999604219157e-05, + "loss": 2.4329, + "step": 285 + }, + { + "epoch": 0.15526601520086863, + "grad_norm": 1.7546486556973808, + "learning_rate": 1.9999949908995832e-05, + "loss": 1.3931, + "step": 286 + }, + { + "epoch": 0.15580890336590664, + "grad_norm": 2.0223702721102375, + "learning_rate": 1.999993815926622e-05, + "loss": 1.6704, + "step": 287 + }, + { + "epoch": 0.1563517915309446, + "grad_norm": 1.8736762651082781, + "learning_rate": 1.9999925172728324e-05, + "loss": 1.9895, + "step": 288 + }, + { + "epoch": 0.15689467969598261, + "grad_norm": 1.8878827291463698, + "learning_rate": 1.9999910949383742e-05, + "loss": 2.4939, + "step": 289 + }, + { + "epoch": 0.15743756786102062, + "grad_norm": 1.7277080078786893, + "learning_rate": 1.9999895489234234e-05, + "loss": 1.9265, + "step": 290 + }, + { + "epoch": 0.15798045602605862, + "grad_norm": 1.7402172969699101, + "learning_rate": 1.999987879228172e-05, + "loss": 1.8101, + "step": 291 + }, + { + "epoch": 0.15852334419109662, + "grad_norm": 2.077278713661093, + "learning_rate": 1.9999860858528257e-05, + "loss": 2.1763, + "step": 292 + }, + { + "epoch": 0.15906623235613462, + "grad_norm": 2.247624069618552, + "learning_rate": 1.9999841687976067e-05, + "loss": 1.9727, + "step": 293 + }, + { + "epoch": 0.15960912052117263, + "grad_norm": 1.6712153667666279, + "learning_rate": 1.999982128062752e-05, + "loss": 1.9921, + "step": 294 + }, + { + "epoch": 0.16015200868621063, + "grad_norm": 1.7126188149003763, + "learning_rate": 1.999979963648514e-05, + "loss": 1.9375, + "step": 295 + }, + { + "epoch": 0.16069489685124863, + "grad_norm": 1.949228534239786, + "learning_rate": 1.99997767555516e-05, + "loss": 2.1628, + "step": 296 + }, + { + "epoch": 0.16123778501628663, + "grad_norm": 1.8253305155157105, + "learning_rate": 1.9999752637829734e-05, + "loss": 1.8293, + "step": 297 + }, + { + "epoch": 0.16178067318132464, + "grad_norm": 1.5881557560095252, + "learning_rate": 1.9999727283322524e-05, + "loss": 1.3688, + "step": 298 + }, + { + "epoch": 0.16232356134636264, + "grad_norm": 2.4008202914116503, + "learning_rate": 1.9999700692033112e-05, + "loss": 1.5919, + "step": 299 + }, + { + "epoch": 0.16286644951140064, + "grad_norm": 1.610839420604411, + "learning_rate": 1.9999672863964778e-05, + "loss": 1.8724, + "step": 300 + }, + { + "epoch": 0.16340933767643864, + "grad_norm": 1.4663439746230646, + "learning_rate": 1.9999643799120964e-05, + "loss": 1.4628, + "step": 301 + }, + { + "epoch": 0.16395222584147665, + "grad_norm": 1.7051101810518976, + "learning_rate": 1.9999613497505272e-05, + "loss": 1.2916, + "step": 302 + }, + { + "epoch": 0.16449511400651465, + "grad_norm": 1.8967399031374104, + "learning_rate": 1.9999581959121443e-05, + "loss": 1.7559, + "step": 303 + }, + { + "epoch": 0.16503800217155265, + "grad_norm": 1.6206520761555416, + "learning_rate": 1.9999549183973382e-05, + "loss": 1.9196, + "step": 304 + }, + { + "epoch": 0.16558089033659065, + "grad_norm": 2.3545652344963544, + "learning_rate": 1.999951517206514e-05, + "loss": 1.2731, + "step": 305 + }, + { + "epoch": 0.16612377850162866, + "grad_norm": 1.7952736257095967, + "learning_rate": 1.9999479923400926e-05, + "loss": 2.0502, + "step": 306 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 1.7199420678903319, + "learning_rate": 1.99994434379851e-05, + "loss": 1.7059, + "step": 307 + }, + { + "epoch": 0.16720955483170466, + "grad_norm": 2.3318553047346, + "learning_rate": 1.9999405715822167e-05, + "loss": 1.4667, + "step": 308 + }, + { + "epoch": 0.16775244299674266, + "grad_norm": 1.618961362672766, + "learning_rate": 1.9999366756916804e-05, + "loss": 1.1866, + "step": 309 + }, + { + "epoch": 0.16829533116178066, + "grad_norm": 2.0805239071290114, + "learning_rate": 1.999932656127382e-05, + "loss": 1.7601, + "step": 310 + }, + { + "epoch": 0.16883821932681867, + "grad_norm": 1.782022796323131, + "learning_rate": 1.9999285128898193e-05, + "loss": 1.6207, + "step": 311 + }, + { + "epoch": 0.16938110749185667, + "grad_norm": 2.2438870289045028, + "learning_rate": 1.9999242459795045e-05, + "loss": 2.1329, + "step": 312 + }, + { + "epoch": 0.16992399565689467, + "grad_norm": 2.2550840823235587, + "learning_rate": 1.9999198553969652e-05, + "loss": 1.6055, + "step": 313 + }, + { + "epoch": 0.17046688382193267, + "grad_norm": 1.8012459587153988, + "learning_rate": 1.9999153411427445e-05, + "loss": 2.018, + "step": 314 + }, + { + "epoch": 0.17100977198697068, + "grad_norm": 1.8498840013582993, + "learning_rate": 1.9999107032174007e-05, + "loss": 1.8113, + "step": 315 + }, + { + "epoch": 0.17155266015200868, + "grad_norm": 2.2799557163953965, + "learning_rate": 1.9999059416215078e-05, + "loss": 1.9494, + "step": 316 + }, + { + "epoch": 0.17209554831704668, + "grad_norm": 2.383891228288174, + "learning_rate": 1.999901056355654e-05, + "loss": 1.8323, + "step": 317 + }, + { + "epoch": 0.17263843648208468, + "grad_norm": 1.9081077971876759, + "learning_rate": 1.9998960474204443e-05, + "loss": 1.5476, + "step": 318 + }, + { + "epoch": 0.1731813246471227, + "grad_norm": 1.9010783437994137, + "learning_rate": 1.999890914816498e-05, + "loss": 1.4602, + "step": 319 + }, + { + "epoch": 0.1737242128121607, + "grad_norm": 2.295573334252978, + "learning_rate": 1.9998856585444493e-05, + "loss": 1.0965, + "step": 320 + }, + { + "epoch": 0.1742671009771987, + "grad_norm": 2.1014332124066954, + "learning_rate": 1.999880278604949e-05, + "loss": 1.3726, + "step": 321 + }, + { + "epoch": 0.1748099891422367, + "grad_norm": 2.0418722317428206, + "learning_rate": 1.9998747749986625e-05, + "loss": 1.8875, + "step": 322 + }, + { + "epoch": 0.1753528773072747, + "grad_norm": 1.768897707524193, + "learning_rate": 1.99986914772627e-05, + "loss": 1.8841, + "step": 323 + }, + { + "epoch": 0.1758957654723127, + "grad_norm": 2.603335158781435, + "learning_rate": 1.9998633967884676e-05, + "loss": 1.4262, + "step": 324 + }, + { + "epoch": 0.1764386536373507, + "grad_norm": 2.1485447383197607, + "learning_rate": 1.999857522185967e-05, + "loss": 1.9283, + "step": 325 + }, + { + "epoch": 0.1769815418023887, + "grad_norm": 1.8398472624384716, + "learning_rate": 1.9998515239194945e-05, + "loss": 1.9089, + "step": 326 + }, + { + "epoch": 0.1775244299674267, + "grad_norm": 2.6988014220525374, + "learning_rate": 1.9998454019897918e-05, + "loss": 1.3636, + "step": 327 + }, + { + "epoch": 0.1780673181324647, + "grad_norm": 2.182794188412918, + "learning_rate": 1.9998391563976166e-05, + "loss": 1.7041, + "step": 328 + }, + { + "epoch": 0.1786102062975027, + "grad_norm": 1.970858895313084, + "learning_rate": 1.9998327871437405e-05, + "loss": 1.4917, + "step": 329 + }, + { + "epoch": 0.1791530944625407, + "grad_norm": 1.829655959511745, + "learning_rate": 1.9998262942289524e-05, + "loss": 1.9694, + "step": 330 + }, + { + "epoch": 0.17969598262757872, + "grad_norm": 1.9990017706312653, + "learning_rate": 1.9998196776540545e-05, + "loss": 2.067, + "step": 331 + }, + { + "epoch": 0.18023887079261672, + "grad_norm": 2.2829395740777234, + "learning_rate": 1.9998129374198655e-05, + "loss": 1.6707, + "step": 332 + }, + { + "epoch": 0.18078175895765472, + "grad_norm": 2.174263401724357, + "learning_rate": 1.9998060735272186e-05, + "loss": 1.6897, + "step": 333 + }, + { + "epoch": 0.18132464712269272, + "grad_norm": 1.8438064757899841, + "learning_rate": 1.9997990859769633e-05, + "loss": 1.4446, + "step": 334 + }, + { + "epoch": 0.18186753528773072, + "grad_norm": 2.0988396037795507, + "learning_rate": 1.9997919747699638e-05, + "loss": 1.8837, + "step": 335 + }, + { + "epoch": 0.18241042345276873, + "grad_norm": 2.582616969632433, + "learning_rate": 1.999784739907099e-05, + "loss": 2.459, + "step": 336 + }, + { + "epoch": 0.18295331161780673, + "grad_norm": 2.5198677585207707, + "learning_rate": 1.9997773813892644e-05, + "loss": 1.5117, + "step": 337 + }, + { + "epoch": 0.18349619978284473, + "grad_norm": 2.1361317951692635, + "learning_rate": 1.9997698992173697e-05, + "loss": 1.5892, + "step": 338 + }, + { + "epoch": 0.18403908794788273, + "grad_norm": 1.7284990063627486, + "learning_rate": 1.9997622933923406e-05, + "loss": 1.5151, + "step": 339 + }, + { + "epoch": 0.18458197611292074, + "grad_norm": 2.076752898704788, + "learning_rate": 1.9997545639151176e-05, + "loss": 2.3424, + "step": 340 + }, + { + "epoch": 0.18512486427795874, + "grad_norm": 2.479369276757794, + "learning_rate": 1.999746710786657e-05, + "loss": 1.5923, + "step": 341 + }, + { + "epoch": 0.18566775244299674, + "grad_norm": 1.9193986648366674, + "learning_rate": 1.9997387340079294e-05, + "loss": 1.6222, + "step": 342 + }, + { + "epoch": 0.18621064060803474, + "grad_norm": 1.8622986153150223, + "learning_rate": 1.999730633579922e-05, + "loss": 1.8402, + "step": 343 + }, + { + "epoch": 0.18675352877307275, + "grad_norm": 2.4012194427686238, + "learning_rate": 1.999722409503637e-05, + "loss": 1.9406, + "step": 344 + }, + { + "epoch": 0.18729641693811075, + "grad_norm": 2.3641518463127538, + "learning_rate": 1.9997140617800907e-05, + "loss": 2.0108, + "step": 345 + }, + { + "epoch": 0.18783930510314875, + "grad_norm": 2.406749148558192, + "learning_rate": 1.9997055904103156e-05, + "loss": 1.1323, + "step": 346 + }, + { + "epoch": 0.18838219326818675, + "grad_norm": 2.002785642517938, + "learning_rate": 1.99969699539536e-05, + "loss": 1.8745, + "step": 347 + }, + { + "epoch": 0.18892508143322476, + "grad_norm": 2.0002882431712816, + "learning_rate": 1.9996882767362874e-05, + "loss": 1.6311, + "step": 348 + }, + { + "epoch": 0.18946796959826276, + "grad_norm": 2.14418043520414, + "learning_rate": 1.9996794344341744e-05, + "loss": 1.7789, + "step": 349 + }, + { + "epoch": 0.19001085776330076, + "grad_norm": 2.0976983604919144, + "learning_rate": 1.9996704684901163e-05, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.19055374592833876, + "grad_norm": 2.017704706993719, + "learning_rate": 1.9996613789052214e-05, + "loss": 1.7974, + "step": 351 + }, + { + "epoch": 0.19109663409337677, + "grad_norm": 1.9907769296520095, + "learning_rate": 1.999652165680614e-05, + "loss": 1.9079, + "step": 352 + }, + { + "epoch": 0.19163952225841477, + "grad_norm": 2.3031451756435746, + "learning_rate": 1.999642828817433e-05, + "loss": 1.7022, + "step": 353 + }, + { + "epoch": 0.19218241042345277, + "grad_norm": 1.6722822353766713, + "learning_rate": 1.9996333683168342e-05, + "loss": 1.6655, + "step": 354 + }, + { + "epoch": 0.19272529858849077, + "grad_norm": 2.2974087830498027, + "learning_rate": 1.9996237841799874e-05, + "loss": 1.761, + "step": 355 + }, + { + "epoch": 0.19326818675352878, + "grad_norm": 1.9181254747298617, + "learning_rate": 1.9996140764080777e-05, + "loss": 1.8259, + "step": 356 + }, + { + "epoch": 0.19381107491856678, + "grad_norm": 2.07933834321742, + "learning_rate": 1.9996042450023053e-05, + "loss": 1.3783, + "step": 357 + }, + { + "epoch": 0.19435396308360478, + "grad_norm": 2.8332787295887436, + "learning_rate": 1.9995942899638875e-05, + "loss": 1.9796, + "step": 358 + }, + { + "epoch": 0.19489685124864278, + "grad_norm": 1.7924418183793758, + "learning_rate": 1.9995842112940545e-05, + "loss": 1.5063, + "step": 359 + }, + { + "epoch": 0.19543973941368079, + "grad_norm": 2.78627522830663, + "learning_rate": 1.9995740089940532e-05, + "loss": 1.891, + "step": 360 + }, + { + "epoch": 0.1959826275787188, + "grad_norm": 2.067311381839208, + "learning_rate": 1.9995636830651453e-05, + "loss": 1.9235, + "step": 361 + }, + { + "epoch": 0.1965255157437568, + "grad_norm": 2.539112412358468, + "learning_rate": 1.9995532335086078e-05, + "loss": 2.3062, + "step": 362 + }, + { + "epoch": 0.1970684039087948, + "grad_norm": 1.9040915545393904, + "learning_rate": 1.999542660325734e-05, + "loss": 1.9263, + "step": 363 + }, + { + "epoch": 0.1976112920738328, + "grad_norm": 1.7664296640025938, + "learning_rate": 1.9995319635178305e-05, + "loss": 1.5562, + "step": 364 + }, + { + "epoch": 0.1981541802388708, + "grad_norm": 1.687746217581272, + "learning_rate": 1.9995211430862206e-05, + "loss": 1.738, + "step": 365 + }, + { + "epoch": 0.1986970684039088, + "grad_norm": 1.9292930460598687, + "learning_rate": 1.9995101990322428e-05, + "loss": 1.5446, + "step": 366 + }, + { + "epoch": 0.1992399565689468, + "grad_norm": 1.9890390069401638, + "learning_rate": 1.9994991313572508e-05, + "loss": 1.9086, + "step": 367 + }, + { + "epoch": 0.1997828447339848, + "grad_norm": 1.8632430045841502, + "learning_rate": 1.999487940062613e-05, + "loss": 1.9747, + "step": 368 + }, + { + "epoch": 0.2003257328990228, + "grad_norm": 2.0677818303742237, + "learning_rate": 1.999476625149714e-05, + "loss": 1.4377, + "step": 369 + }, + { + "epoch": 0.2008686210640608, + "grad_norm": 1.7412205305358945, + "learning_rate": 1.9994651866199527e-05, + "loss": 1.9352, + "step": 370 + }, + { + "epoch": 0.2014115092290988, + "grad_norm": 1.8562175126383418, + "learning_rate": 1.9994536244747448e-05, + "loss": 1.5469, + "step": 371 + }, + { + "epoch": 0.20195439739413681, + "grad_norm": 1.8688345857586197, + "learning_rate": 1.9994419387155194e-05, + "loss": 1.8327, + "step": 372 + }, + { + "epoch": 0.20249728555917482, + "grad_norm": 2.1538407513899647, + "learning_rate": 1.9994301293437223e-05, + "loss": 1.9621, + "step": 373 + }, + { + "epoch": 0.20304017372421282, + "grad_norm": 2.4849664096044752, + "learning_rate": 1.999418196360814e-05, + "loss": 1.8179, + "step": 374 + }, + { + "epoch": 0.20358306188925082, + "grad_norm": 2.098949652063271, + "learning_rate": 1.99940613976827e-05, + "loss": 1.8719, + "step": 375 + }, + { + "epoch": 0.20412595005428882, + "grad_norm": 1.8802547433747465, + "learning_rate": 1.999393959567582e-05, + "loss": 1.5697, + "step": 376 + }, + { + "epoch": 0.20466883821932683, + "grad_norm": 2.090861017905227, + "learning_rate": 1.9993816557602567e-05, + "loss": 1.8249, + "step": 377 + }, + { + "epoch": 0.20521172638436483, + "grad_norm": 1.775481861127165, + "learning_rate": 1.999369228347815e-05, + "loss": 1.3095, + "step": 378 + }, + { + "epoch": 0.20575461454940283, + "grad_norm": 2.260330600420479, + "learning_rate": 1.999356677331794e-05, + "loss": 1.9153, + "step": 379 + }, + { + "epoch": 0.20629750271444083, + "grad_norm": 2.0769016143551484, + "learning_rate": 1.999344002713747e-05, + "loss": 1.958, + "step": 380 + }, + { + "epoch": 0.20684039087947884, + "grad_norm": 2.063584775140761, + "learning_rate": 1.9993312044952408e-05, + "loss": 1.7887, + "step": 381 + }, + { + "epoch": 0.20738327904451684, + "grad_norm": 1.9638736053910157, + "learning_rate": 1.9993182826778588e-05, + "loss": 1.3178, + "step": 382 + }, + { + "epoch": 0.20792616720955484, + "grad_norm": 1.7752329715548703, + "learning_rate": 1.9993052372631988e-05, + "loss": 1.4473, + "step": 383 + }, + { + "epoch": 0.20846905537459284, + "grad_norm": 2.1002641405528952, + "learning_rate": 1.999292068252874e-05, + "loss": 1.9698, + "step": 384 + }, + { + "epoch": 0.20901194353963085, + "grad_norm": 2.757803188261632, + "learning_rate": 1.999278775648514e-05, + "loss": 1.1517, + "step": 385 + }, + { + "epoch": 0.20955483170466885, + "grad_norm": 3.0341284449704378, + "learning_rate": 1.9992653594517624e-05, + "loss": 1.987, + "step": 386 + }, + { + "epoch": 0.21009771986970685, + "grad_norm": 2.445830333821562, + "learning_rate": 1.9992518196642786e-05, + "loss": 1.7656, + "step": 387 + }, + { + "epoch": 0.21064060803474485, + "grad_norm": 2.737783962568244, + "learning_rate": 1.9992381562877368e-05, + "loss": 2.3012, + "step": 388 + }, + { + "epoch": 0.21118349619978286, + "grad_norm": 2.0108432149096793, + "learning_rate": 1.9992243693238275e-05, + "loss": 2.1096, + "step": 389 + }, + { + "epoch": 0.21172638436482086, + "grad_norm": 2.1448465445226397, + "learning_rate": 1.9992104587742558e-05, + "loss": 1.6912, + "step": 390 + }, + { + "epoch": 0.21226927252985886, + "grad_norm": 2.842739011518123, + "learning_rate": 1.999196424640742e-05, + "loss": 1.9987, + "step": 391 + }, + { + "epoch": 0.21281216069489686, + "grad_norm": 2.026900017823922, + "learning_rate": 1.9991822669250216e-05, + "loss": 1.6251, + "step": 392 + }, + { + "epoch": 0.21335504885993486, + "grad_norm": 2.139818571931489, + "learning_rate": 1.9991679856288462e-05, + "loss": 1.4181, + "step": 393 + }, + { + "epoch": 0.21389793702497287, + "grad_norm": 2.061327407258535, + "learning_rate": 1.999153580753982e-05, + "loss": 2.0976, + "step": 394 + }, + { + "epoch": 0.21444082519001087, + "grad_norm": 1.9510646261617408, + "learning_rate": 1.9991390523022105e-05, + "loss": 1.3091, + "step": 395 + }, + { + "epoch": 0.21498371335504887, + "grad_norm": 2.0892427860245664, + "learning_rate": 1.9991244002753287e-05, + "loss": 1.3693, + "step": 396 + }, + { + "epoch": 0.21552660152008687, + "grad_norm": 1.9528708238712815, + "learning_rate": 1.9991096246751483e-05, + "loss": 1.2807, + "step": 397 + }, + { + "epoch": 0.21606948968512488, + "grad_norm": 2.0084632909809983, + "learning_rate": 1.9990947255034977e-05, + "loss": 1.7429, + "step": 398 + }, + { + "epoch": 0.21661237785016288, + "grad_norm": 2.19414527352725, + "learning_rate": 1.999079702762219e-05, + "loss": 1.5962, + "step": 399 + }, + { + "epoch": 0.21715526601520088, + "grad_norm": 2.184484203373308, + "learning_rate": 1.9990645564531702e-05, + "loss": 1.6887, + "step": 400 + }, + { + "epoch": 0.21769815418023888, + "grad_norm": 2.214473402714986, + "learning_rate": 1.9990492865782248e-05, + "loss": 1.5226, + "step": 401 + }, + { + "epoch": 0.2182410423452769, + "grad_norm": 2.1467748369867414, + "learning_rate": 1.9990338931392714e-05, + "loss": 1.5634, + "step": 402 + }, + { + "epoch": 0.21878393051031486, + "grad_norm": 2.1969353910959684, + "learning_rate": 1.999018376138214e-05, + "loss": 1.8652, + "step": 403 + }, + { + "epoch": 0.21932681867535286, + "grad_norm": 2.245031544698939, + "learning_rate": 1.9990027355769715e-05, + "loss": 2.1811, + "step": 404 + }, + { + "epoch": 0.21986970684039087, + "grad_norm": 2.381275606702572, + "learning_rate": 1.9989869714574784e-05, + "loss": 1.6356, + "step": 405 + }, + { + "epoch": 0.22041259500542887, + "grad_norm": 2.8847531540936053, + "learning_rate": 1.9989710837816846e-05, + "loss": 1.9953, + "step": 406 + }, + { + "epoch": 0.22095548317046687, + "grad_norm": 2.383074182943421, + "learning_rate": 1.9989550725515553e-05, + "loss": 1.5589, + "step": 407 + }, + { + "epoch": 0.22149837133550487, + "grad_norm": 2.3867744628533987, + "learning_rate": 1.99893893776907e-05, + "loss": 1.5351, + "step": 408 + }, + { + "epoch": 0.22204125950054288, + "grad_norm": 2.4139767232435463, + "learning_rate": 1.998922679436225e-05, + "loss": 1.5949, + "step": 409 + }, + { + "epoch": 0.22258414766558088, + "grad_norm": 3.034840719531937, + "learning_rate": 1.9989062975550313e-05, + "loss": 1.7994, + "step": 410 + }, + { + "epoch": 0.22312703583061888, + "grad_norm": 2.6245673280102544, + "learning_rate": 1.9988897921275144e-05, + "loss": 2.2593, + "step": 411 + }, + { + "epoch": 0.22366992399565688, + "grad_norm": 2.403934048820897, + "learning_rate": 1.998873163155716e-05, + "loss": 1.3179, + "step": 412 + }, + { + "epoch": 0.22421281216069489, + "grad_norm": 2.0174357446052524, + "learning_rate": 1.998856410641693e-05, + "loss": 1.8698, + "step": 413 + }, + { + "epoch": 0.2247557003257329, + "grad_norm": 2.25869172419699, + "learning_rate": 1.998839534587517e-05, + "loss": 1.596, + "step": 414 + }, + { + "epoch": 0.2252985884907709, + "grad_norm": 3.384596068816443, + "learning_rate": 1.9988225349952758e-05, + "loss": 1.5232, + "step": 415 + }, + { + "epoch": 0.2258414766558089, + "grad_norm": 2.2109577343653246, + "learning_rate": 1.9988054118670712e-05, + "loss": 1.2403, + "step": 416 + }, + { + "epoch": 0.2263843648208469, + "grad_norm": 2.1825505717325053, + "learning_rate": 1.9987881652050215e-05, + "loss": 1.7205, + "step": 417 + }, + { + "epoch": 0.2269272529858849, + "grad_norm": 2.7992033856118583, + "learning_rate": 1.99877079501126e-05, + "loss": 1.7486, + "step": 418 + }, + { + "epoch": 0.2274701411509229, + "grad_norm": 2.6351781282432793, + "learning_rate": 1.9987533012879344e-05, + "loss": 1.6693, + "step": 419 + }, + { + "epoch": 0.2280130293159609, + "grad_norm": 2.2385207734696384, + "learning_rate": 1.9987356840372088e-05, + "loss": 1.1296, + "step": 420 + }, + { + "epoch": 0.2285559174809989, + "grad_norm": 2.4921881643709107, + "learning_rate": 1.998717943261262e-05, + "loss": 1.1379, + "step": 421 + }, + { + "epoch": 0.2290988056460369, + "grad_norm": 2.842685080646979, + "learning_rate": 1.9987000789622884e-05, + "loss": 1.7963, + "step": 422 + }, + { + "epoch": 0.2296416938110749, + "grad_norm": 2.4776027353414145, + "learning_rate": 1.9986820911424972e-05, + "loss": 1.6939, + "step": 423 + }, + { + "epoch": 0.2301845819761129, + "grad_norm": 2.688785422525701, + "learning_rate": 1.9986639798041134e-05, + "loss": 1.2282, + "step": 424 + }, + { + "epoch": 0.23072747014115091, + "grad_norm": 3.4758865457335495, + "learning_rate": 1.998645744949377e-05, + "loss": 1.7664, + "step": 425 + }, + { + "epoch": 0.23127035830618892, + "grad_norm": 2.5586740557561605, + "learning_rate": 1.9986273865805432e-05, + "loss": 1.5637, + "step": 426 + }, + { + "epoch": 0.23181324647122692, + "grad_norm": 2.645306116903878, + "learning_rate": 1.9986089046998827e-05, + "loss": 1.5331, + "step": 427 + }, + { + "epoch": 0.23235613463626492, + "grad_norm": 2.748145883999443, + "learning_rate": 1.998590299309681e-05, + "loss": 1.7316, + "step": 428 + }, + { + "epoch": 0.23289902280130292, + "grad_norm": 2.4357923833746438, + "learning_rate": 1.99857157041224e-05, + "loss": 1.1479, + "step": 429 + }, + { + "epoch": 0.23344191096634093, + "grad_norm": 2.7322633978331377, + "learning_rate": 1.9985527180098755e-05, + "loss": 1.6089, + "step": 430 + }, + { + "epoch": 0.23398479913137893, + "grad_norm": 2.625531215380387, + "learning_rate": 1.9985337421049193e-05, + "loss": 1.8973, + "step": 431 + }, + { + "epoch": 0.23452768729641693, + "grad_norm": 2.1586487666504754, + "learning_rate": 1.9985146426997185e-05, + "loss": 1.664, + "step": 432 + }, + { + "epoch": 0.23507057546145493, + "grad_norm": 2.748704307086674, + "learning_rate": 1.9984954197966355e-05, + "loss": 1.677, + "step": 433 + }, + { + "epoch": 0.23561346362649294, + "grad_norm": 2.345532526959197, + "learning_rate": 1.9984760733980476e-05, + "loss": 1.5133, + "step": 434 + }, + { + "epoch": 0.23615635179153094, + "grad_norm": 2.1655585346308848, + "learning_rate": 1.9984566035063473e-05, + "loss": 1.4206, + "step": 435 + }, + { + "epoch": 0.23669923995656894, + "grad_norm": 2.697584333049271, + "learning_rate": 1.9984370101239434e-05, + "loss": 1.8131, + "step": 436 + }, + { + "epoch": 0.23724212812160694, + "grad_norm": 2.235069158786981, + "learning_rate": 1.9984172932532583e-05, + "loss": 1.7839, + "step": 437 + }, + { + "epoch": 0.23778501628664495, + "grad_norm": 2.548562657856099, + "learning_rate": 1.998397452896731e-05, + "loss": 1.4876, + "step": 438 + }, + { + "epoch": 0.23832790445168295, + "grad_norm": 2.1698874138883673, + "learning_rate": 1.9983774890568163e-05, + "loss": 1.2916, + "step": 439 + }, + { + "epoch": 0.23887079261672095, + "grad_norm": 2.3237954527681084, + "learning_rate": 1.998357401735982e-05, + "loss": 1.398, + "step": 440 + }, + { + "epoch": 0.23941368078175895, + "grad_norm": 2.448364938314344, + "learning_rate": 1.9983371909367135e-05, + "loss": 1.2663, + "step": 441 + }, + { + "epoch": 0.23995656894679696, + "grad_norm": 1.935899726785714, + "learning_rate": 1.99831685666151e-05, + "loss": 1.2448, + "step": 442 + }, + { + "epoch": 0.24049945711183496, + "grad_norm": 2.14357016947003, + "learning_rate": 1.9982963989128864e-05, + "loss": 1.4786, + "step": 443 + }, + { + "epoch": 0.24104234527687296, + "grad_norm": 2.5830135196717148, + "learning_rate": 1.998275817693373e-05, + "loss": 1.8324, + "step": 444 + }, + { + "epoch": 0.24158523344191096, + "grad_norm": 3.159289299734819, + "learning_rate": 1.9982551130055157e-05, + "loss": 1.9091, + "step": 445 + }, + { + "epoch": 0.24212812160694897, + "grad_norm": 3.2712239674501755, + "learning_rate": 1.9982342848518753e-05, + "loss": 1.7869, + "step": 446 + }, + { + "epoch": 0.24267100977198697, + "grad_norm": 3.35834841128443, + "learning_rate": 1.998213333235027e-05, + "loss": 1.2772, + "step": 447 + }, + { + "epoch": 0.24321389793702497, + "grad_norm": 2.9379666136057354, + "learning_rate": 1.998192258157563e-05, + "loss": 1.3497, + "step": 448 + }, + { + "epoch": 0.24375678610206297, + "grad_norm": 2.334119075845609, + "learning_rate": 1.9981710596220897e-05, + "loss": 1.5541, + "step": 449 + }, + { + "epoch": 0.24429967426710097, + "grad_norm": 2.781469856763384, + "learning_rate": 1.998149737631229e-05, + "loss": 1.9413, + "step": 450 + }, + { + "epoch": 0.24484256243213898, + "grad_norm": 3.0196781534130452, + "learning_rate": 1.9981282921876177e-05, + "loss": 1.3238, + "step": 451 + }, + { + "epoch": 0.24538545059717698, + "grad_norm": 3.547169321727429, + "learning_rate": 1.9981067232939086e-05, + "loss": 1.9952, + "step": 452 + }, + { + "epoch": 0.24592833876221498, + "grad_norm": 3.2411499530913535, + "learning_rate": 1.9980850309527693e-05, + "loss": 1.8244, + "step": 453 + }, + { + "epoch": 0.24647122692725298, + "grad_norm": 2.670313260104859, + "learning_rate": 1.9980632151668822e-05, + "loss": 1.607, + "step": 454 + }, + { + "epoch": 0.247014115092291, + "grad_norm": 2.8822100638306143, + "learning_rate": 1.9980412759389468e-05, + "loss": 1.8868, + "step": 455 + }, + { + "epoch": 0.247557003257329, + "grad_norm": 2.4776152417583317, + "learning_rate": 1.9980192132716748e-05, + "loss": 1.778, + "step": 456 + }, + { + "epoch": 0.248099891422367, + "grad_norm": 2.2001923672712076, + "learning_rate": 1.9979970271677967e-05, + "loss": 1.3544, + "step": 457 + }, + { + "epoch": 0.248642779587405, + "grad_norm": 2.7694932683911837, + "learning_rate": 1.9979747176300553e-05, + "loss": 1.6521, + "step": 458 + }, + { + "epoch": 0.249185667752443, + "grad_norm": 2.8464573075472845, + "learning_rate": 1.99795228466121e-05, + "loss": 1.8803, + "step": 459 + }, + { + "epoch": 0.249728555917481, + "grad_norm": 2.577989994947286, + "learning_rate": 1.9979297282640365e-05, + "loss": 1.8838, + "step": 460 + }, + { + "epoch": 0.250271444082519, + "grad_norm": 2.9859834231033164, + "learning_rate": 1.997907048441323e-05, + "loss": 1.5122, + "step": 461 + }, + { + "epoch": 0.250814332247557, + "grad_norm": 3.0857580735568098, + "learning_rate": 1.9978842451958757e-05, + "loss": 1.7789, + "step": 462 + }, + { + "epoch": 0.251357220412595, + "grad_norm": 2.523783848348806, + "learning_rate": 1.9978613185305145e-05, + "loss": 1.5815, + "step": 463 + }, + { + "epoch": 0.251900108577633, + "grad_norm": 2.699094362013039, + "learning_rate": 1.9978382684480747e-05, + "loss": 1.7448, + "step": 464 + }, + { + "epoch": 0.252442996742671, + "grad_norm": 2.736268550628698, + "learning_rate": 1.997815094951408e-05, + "loss": 1.7052, + "step": 465 + }, + { + "epoch": 0.252985884907709, + "grad_norm": 3.1150828428090014, + "learning_rate": 1.99779179804338e-05, + "loss": 1.7743, + "step": 466 + }, + { + "epoch": 0.253528773072747, + "grad_norm": 3.3513164619888482, + "learning_rate": 1.997768377726872e-05, + "loss": 2.3905, + "step": 467 + }, + { + "epoch": 0.254071661237785, + "grad_norm": 2.400886982379507, + "learning_rate": 1.9977448340047808e-05, + "loss": 1.6096, + "step": 468 + }, + { + "epoch": 0.254614549402823, + "grad_norm": 2.633533044966171, + "learning_rate": 1.9977211668800186e-05, + "loss": 1.7796, + "step": 469 + }, + { + "epoch": 0.255157437567861, + "grad_norm": 3.9297139007235042, + "learning_rate": 1.997697376355512e-05, + "loss": 1.4602, + "step": 470 + }, + { + "epoch": 0.255700325732899, + "grad_norm": 5.588260619709643, + "learning_rate": 1.9976734624342044e-05, + "loss": 2.0389, + "step": 471 + }, + { + "epoch": 0.256243213897937, + "grad_norm": 2.600723171476426, + "learning_rate": 1.9976494251190522e-05, + "loss": 1.3676, + "step": 472 + }, + { + "epoch": 0.25678610206297503, + "grad_norm": 2.61945002649116, + "learning_rate": 1.9976252644130297e-05, + "loss": 1.7902, + "step": 473 + }, + { + "epoch": 0.25732899022801303, + "grad_norm": 3.7513561790803838, + "learning_rate": 1.997600980319124e-05, + "loss": 1.5997, + "step": 474 + }, + { + "epoch": 0.25787187839305103, + "grad_norm": 2.8832359552778737, + "learning_rate": 1.9975765728403395e-05, + "loss": 1.9636, + "step": 475 + }, + { + "epoch": 0.25841476655808904, + "grad_norm": 3.0975018752600243, + "learning_rate": 1.9975520419796942e-05, + "loss": 1.1165, + "step": 476 + }, + { + "epoch": 0.25895765472312704, + "grad_norm": 2.121708502818221, + "learning_rate": 1.9975273877402227e-05, + "loss": 1.5108, + "step": 477 + }, + { + "epoch": 0.25950054288816504, + "grad_norm": 2.4073592870530116, + "learning_rate": 1.997502610124974e-05, + "loss": 1.7828, + "step": 478 + }, + { + "epoch": 0.26004343105320304, + "grad_norm": 2.6858679986632974, + "learning_rate": 1.997477709137013e-05, + "loss": 1.8483, + "step": 479 + }, + { + "epoch": 0.26058631921824105, + "grad_norm": 3.283752190131325, + "learning_rate": 1.997452684779419e-05, + "loss": 1.6105, + "step": 480 + }, + { + "epoch": 0.26112920738327905, + "grad_norm": 2.500181185675909, + "learning_rate": 1.997427537055287e-05, + "loss": 1.5475, + "step": 481 + }, + { + "epoch": 0.26167209554831705, + "grad_norm": 2.3992324550953885, + "learning_rate": 1.9974022659677278e-05, + "loss": 1.6062, + "step": 482 + }, + { + "epoch": 0.26221498371335505, + "grad_norm": 2.7692293822867837, + "learning_rate": 1.9973768715198667e-05, + "loss": 1.4995, + "step": 483 + }, + { + "epoch": 0.26275787187839306, + "grad_norm": 2.971423364277874, + "learning_rate": 1.9973513537148447e-05, + "loss": 1.7904, + "step": 484 + }, + { + "epoch": 0.26330076004343106, + "grad_norm": 2.2769736321644105, + "learning_rate": 1.9973257125558177e-05, + "loss": 1.3121, + "step": 485 + }, + { + "epoch": 0.26384364820846906, + "grad_norm": 2.5949046051899254, + "learning_rate": 1.997299948045957e-05, + "loss": 1.4555, + "step": 486 + }, + { + "epoch": 0.26438653637350706, + "grad_norm": 3.0210593683445204, + "learning_rate": 1.997274060188449e-05, + "loss": 1.8211, + "step": 487 + }, + { + "epoch": 0.26492942453854507, + "grad_norm": 3.076626224616319, + "learning_rate": 1.9972480489864962e-05, + "loss": 1.6366, + "step": 488 + }, + { + "epoch": 0.26547231270358307, + "grad_norm": 3.1369728365663536, + "learning_rate": 1.9972219144433148e-05, + "loss": 1.5027, + "step": 489 + }, + { + "epoch": 0.26601520086862107, + "grad_norm": 2.2870450349164635, + "learning_rate": 1.9971956565621383e-05, + "loss": 1.2784, + "step": 490 + }, + { + "epoch": 0.2665580890336591, + "grad_norm": 3.0130036065633776, + "learning_rate": 1.9971692753462134e-05, + "loss": 1.2083, + "step": 491 + }, + { + "epoch": 0.2671009771986971, + "grad_norm": 2.6063157323029733, + "learning_rate": 1.9971427707988034e-05, + "loss": 1.0083, + "step": 492 + }, + { + "epoch": 0.2676438653637351, + "grad_norm": 2.579371053895234, + "learning_rate": 1.997116142923186e-05, + "loss": 1.1937, + "step": 493 + }, + { + "epoch": 0.2681867535287731, + "grad_norm": 2.532537971800688, + "learning_rate": 1.9970893917226554e-05, + "loss": 1.4735, + "step": 494 + }, + { + "epoch": 0.2687296416938111, + "grad_norm": 2.3483150144294105, + "learning_rate": 1.997062517200519e-05, + "loss": 1.7269, + "step": 495 + }, + { + "epoch": 0.2692725298588491, + "grad_norm": 2.594809867192747, + "learning_rate": 1.997035519360102e-05, + "loss": 1.8283, + "step": 496 + }, + { + "epoch": 0.2698154180238871, + "grad_norm": 2.613813750609998, + "learning_rate": 1.9970083982047428e-05, + "loss": 1.2302, + "step": 497 + }, + { + "epoch": 0.2703583061889251, + "grad_norm": 2.648279162964909, + "learning_rate": 1.9969811537377956e-05, + "loss": 1.6225, + "step": 498 + }, + { + "epoch": 0.2709011943539631, + "grad_norm": 2.3790190706794325, + "learning_rate": 1.9969537859626308e-05, + "loss": 1.5172, + "step": 499 + }, + { + "epoch": 0.2714440825190011, + "grad_norm": 2.7054998578606364, + "learning_rate": 1.9969262948826326e-05, + "loss": 1.4525, + "step": 500 + }, + { + "epoch": 0.2719869706840391, + "grad_norm": 2.412151508264948, + "learning_rate": 1.9968986805012012e-05, + "loss": 1.3299, + "step": 501 + }, + { + "epoch": 0.2725298588490771, + "grad_norm": 2.4836460319285414, + "learning_rate": 1.9968709428217525e-05, + "loss": 1.5217, + "step": 502 + }, + { + "epoch": 0.2730727470141151, + "grad_norm": 3.3112944949184606, + "learning_rate": 1.9968430818477168e-05, + "loss": 2.0643, + "step": 503 + }, + { + "epoch": 0.2736156351791531, + "grad_norm": 3.266395629954733, + "learning_rate": 1.9968150975825397e-05, + "loss": 1.1719, + "step": 504 + }, + { + "epoch": 0.2741585233441911, + "grad_norm": 2.3487351103507073, + "learning_rate": 1.996786990029683e-05, + "loss": 1.6876, + "step": 505 + }, + { + "epoch": 0.2747014115092291, + "grad_norm": 3.0672442719402673, + "learning_rate": 1.9967587591926227e-05, + "loss": 1.3946, + "step": 506 + }, + { + "epoch": 0.2752442996742671, + "grad_norm": 2.84320621483769, + "learning_rate": 1.99673040507485e-05, + "loss": 1.3259, + "step": 507 + }, + { + "epoch": 0.2757871878393051, + "grad_norm": 2.5078449617010707, + "learning_rate": 1.9967019276798728e-05, + "loss": 0.8478, + "step": 508 + }, + { + "epoch": 0.2763300760043431, + "grad_norm": 3.137784699454796, + "learning_rate": 1.9966733270112126e-05, + "loss": 1.1688, + "step": 509 + }, + { + "epoch": 0.2768729641693811, + "grad_norm": 3.7277826491955017, + "learning_rate": 1.996644603072407e-05, + "loss": 1.1091, + "step": 510 + }, + { + "epoch": 0.2774158523344191, + "grad_norm": 2.5276781604415635, + "learning_rate": 1.996615755867008e-05, + "loss": 1.1299, + "step": 511 + }, + { + "epoch": 0.2779587404994571, + "grad_norm": 3.1012700661738744, + "learning_rate": 1.996586785398584e-05, + "loss": 1.7218, + "step": 512 + }, + { + "epoch": 0.2785016286644951, + "grad_norm": 2.285166712515903, + "learning_rate": 1.9965576916707182e-05, + "loss": 1.2868, + "step": 513 + }, + { + "epoch": 0.27904451682953313, + "grad_norm": 2.33097906349044, + "learning_rate": 1.9965284746870088e-05, + "loss": 0.9887, + "step": 514 + }, + { + "epoch": 0.27958740499457113, + "grad_norm": 2.6473787082237927, + "learning_rate": 1.9964991344510697e-05, + "loss": 1.8543, + "step": 515 + }, + { + "epoch": 0.28013029315960913, + "grad_norm": 2.4628160599533366, + "learning_rate": 1.996469670966529e-05, + "loss": 1.2263, + "step": 516 + }, + { + "epoch": 0.28067318132464714, + "grad_norm": 3.2897847068350905, + "learning_rate": 1.9964400842370314e-05, + "loss": 1.6338, + "step": 517 + }, + { + "epoch": 0.28121606948968514, + "grad_norm": 2.4439319341540324, + "learning_rate": 1.9964103742662363e-05, + "loss": 1.0836, + "step": 518 + }, + { + "epoch": 0.28175895765472314, + "grad_norm": 2.3221991020412003, + "learning_rate": 1.996380541057818e-05, + "loss": 1.2331, + "step": 519 + }, + { + "epoch": 0.28230184581976114, + "grad_norm": 2.9571040634251564, + "learning_rate": 1.9963505846154662e-05, + "loss": 1.3066, + "step": 520 + }, + { + "epoch": 0.28284473398479915, + "grad_norm": 3.7512706020225624, + "learning_rate": 1.996320504942886e-05, + "loss": 1.7482, + "step": 521 + }, + { + "epoch": 0.28338762214983715, + "grad_norm": 2.4620109793388267, + "learning_rate": 1.9962903020437983e-05, + "loss": 1.5334, + "step": 522 + }, + { + "epoch": 0.28393051031487515, + "grad_norm": 3.030374272795485, + "learning_rate": 1.9962599759219383e-05, + "loss": 1.8957, + "step": 523 + }, + { + "epoch": 0.28447339847991315, + "grad_norm": 2.452389821491403, + "learning_rate": 1.9962295265810563e-05, + "loss": 1.5438, + "step": 524 + }, + { + "epoch": 0.28501628664495116, + "grad_norm": 2.713028369466205, + "learning_rate": 1.996198954024919e-05, + "loss": 1.4272, + "step": 525 + }, + { + "epoch": 0.28555917480998916, + "grad_norm": 3.810321275175567, + "learning_rate": 1.996168258257307e-05, + "loss": 1.9028, + "step": 526 + }, + { + "epoch": 0.28610206297502716, + "grad_norm": 2.7774100977441236, + "learning_rate": 1.9961374392820173e-05, + "loss": 1.5644, + "step": 527 + }, + { + "epoch": 0.28664495114006516, + "grad_norm": 2.6798712089104186, + "learning_rate": 1.9961064971028616e-05, + "loss": 1.188, + "step": 528 + }, + { + "epoch": 0.28718783930510317, + "grad_norm": 2.889465990486677, + "learning_rate": 1.9960754317236666e-05, + "loss": 1.5393, + "step": 529 + }, + { + "epoch": 0.28773072747014117, + "grad_norm": 3.465251366831076, + "learning_rate": 1.996044243148275e-05, + "loss": 2.1899, + "step": 530 + }, + { + "epoch": 0.28827361563517917, + "grad_norm": 2.645941940974219, + "learning_rate": 1.9960129313805437e-05, + "loss": 1.3691, + "step": 531 + }, + { + "epoch": 0.2888165038002172, + "grad_norm": 2.3914199977194293, + "learning_rate": 1.9959814964243455e-05, + "loss": 1.5219, + "step": 532 + }, + { + "epoch": 0.2893593919652552, + "grad_norm": 2.3023810529281343, + "learning_rate": 1.995949938283569e-05, + "loss": 1.5147, + "step": 533 + }, + { + "epoch": 0.2899022801302932, + "grad_norm": 2.7362205671791155, + "learning_rate": 1.9959182569621164e-05, + "loss": 1.7571, + "step": 534 + }, + { + "epoch": 0.2904451682953312, + "grad_norm": 3.971162331076012, + "learning_rate": 1.9958864524639066e-05, + "loss": 1.3425, + "step": 535 + }, + { + "epoch": 0.2909880564603692, + "grad_norm": 3.4144928239616514, + "learning_rate": 1.9958545247928727e-05, + "loss": 1.6962, + "step": 536 + }, + { + "epoch": 0.2915309446254072, + "grad_norm": 3.5063126675319043, + "learning_rate": 1.9958224739529647e-05, + "loss": 1.6406, + "step": 537 + }, + { + "epoch": 0.2920738327904452, + "grad_norm": 2.9013783116047547, + "learning_rate": 1.995790299948146e-05, + "loss": 1.6376, + "step": 538 + }, + { + "epoch": 0.2926167209554832, + "grad_norm": 2.4827691033904693, + "learning_rate": 1.9957580027823957e-05, + "loss": 1.8672, + "step": 539 + }, + { + "epoch": 0.2931596091205212, + "grad_norm": 3.1245563354940242, + "learning_rate": 1.9957255824597087e-05, + "loss": 1.885, + "step": 540 + }, + { + "epoch": 0.2937024972855592, + "grad_norm": 2.7941860466759896, + "learning_rate": 1.9956930389840945e-05, + "loss": 1.0903, + "step": 541 + }, + { + "epoch": 0.2942453854505972, + "grad_norm": 3.1348104146717772, + "learning_rate": 1.9956603723595784e-05, + "loss": 2.1446, + "step": 542 + }, + { + "epoch": 0.2947882736156352, + "grad_norm": 2.4248222663066747, + "learning_rate": 1.995627582590201e-05, + "loss": 1.3503, + "step": 543 + }, + { + "epoch": 0.2953311617806732, + "grad_norm": 3.1538245864476337, + "learning_rate": 1.995594669680017e-05, + "loss": 1.3275, + "step": 544 + }, + { + "epoch": 0.2958740499457112, + "grad_norm": 3.436905240480997, + "learning_rate": 1.9955616336330976e-05, + "loss": 1.7249, + "step": 545 + }, + { + "epoch": 0.2964169381107492, + "grad_norm": 3.9145352507266393, + "learning_rate": 1.9955284744535287e-05, + "loss": 2.0089, + "step": 546 + }, + { + "epoch": 0.2969598262757872, + "grad_norm": 2.8600509325831185, + "learning_rate": 1.9954951921454113e-05, + "loss": 1.4527, + "step": 547 + }, + { + "epoch": 0.2975027144408252, + "grad_norm": 2.865760632888347, + "learning_rate": 1.995461786712862e-05, + "loss": 1.4553, + "step": 548 + }, + { + "epoch": 0.2980456026058632, + "grad_norm": 3.3559882259900706, + "learning_rate": 1.9954282581600127e-05, + "loss": 2.0456, + "step": 549 + }, + { + "epoch": 0.2985884907709012, + "grad_norm": 3.3424656149019008, + "learning_rate": 1.9953946064910098e-05, + "loss": 1.8253, + "step": 550 + }, + { + "epoch": 0.2991313789359392, + "grad_norm": 2.569119318410074, + "learning_rate": 1.9953608317100153e-05, + "loss": 1.3623, + "step": 551 + }, + { + "epoch": 0.2996742671009772, + "grad_norm": 2.8918953822102424, + "learning_rate": 1.995326933821207e-05, + "loss": 1.7521, + "step": 552 + }, + { + "epoch": 0.3002171552660152, + "grad_norm": 2.7627900710192246, + "learning_rate": 1.995292912828777e-05, + "loss": 1.775, + "step": 553 + }, + { + "epoch": 0.3007600434310532, + "grad_norm": 2.8053609300694804, + "learning_rate": 1.9952587687369334e-05, + "loss": 1.7536, + "step": 554 + }, + { + "epoch": 0.30130293159609123, + "grad_norm": 2.9775715301146803, + "learning_rate": 1.995224501549899e-05, + "loss": 1.6715, + "step": 555 + }, + { + "epoch": 0.30184581976112923, + "grad_norm": 3.146995410263436, + "learning_rate": 1.9951901112719123e-05, + "loss": 1.1032, + "step": 556 + }, + { + "epoch": 0.30238870792616723, + "grad_norm": 3.219197817112143, + "learning_rate": 1.9951555979072266e-05, + "loss": 1.6326, + "step": 557 + }, + { + "epoch": 0.30293159609120524, + "grad_norm": 2.676508071644292, + "learning_rate": 1.99512096146011e-05, + "loss": 1.4836, + "step": 558 + }, + { + "epoch": 0.30347448425624324, + "grad_norm": 3.8806900857620374, + "learning_rate": 1.9950862019348474e-05, + "loss": 1.7794, + "step": 559 + }, + { + "epoch": 0.30401737242128124, + "grad_norm": 2.547409032322543, + "learning_rate": 1.995051319335737e-05, + "loss": 1.3263, + "step": 560 + }, + { + "epoch": 0.30456026058631924, + "grad_norm": 3.025189850713409, + "learning_rate": 1.995016313667094e-05, + "loss": 1.2409, + "step": 561 + }, + { + "epoch": 0.30510314875135724, + "grad_norm": 3.3644665856402614, + "learning_rate": 1.9949811849332476e-05, + "loss": 1.2988, + "step": 562 + }, + { + "epoch": 0.30564603691639525, + "grad_norm": 3.4126932100522755, + "learning_rate": 1.9949459331385422e-05, + "loss": 1.6126, + "step": 563 + }, + { + "epoch": 0.30618892508143325, + "grad_norm": 3.3751971677759416, + "learning_rate": 1.994910558287338e-05, + "loss": 1.6243, + "step": 564 + }, + { + "epoch": 0.30673181324647125, + "grad_norm": 3.0948952637866105, + "learning_rate": 1.9948750603840102e-05, + "loss": 1.6553, + "step": 565 + }, + { + "epoch": 0.30727470141150925, + "grad_norm": 3.1116929004314224, + "learning_rate": 1.9948394394329494e-05, + "loss": 1.0466, + "step": 566 + }, + { + "epoch": 0.30781758957654726, + "grad_norm": 2.649244034262683, + "learning_rate": 1.9948036954385613e-05, + "loss": 1.2914, + "step": 567 + }, + { + "epoch": 0.30836047774158526, + "grad_norm": 3.9338023936000965, + "learning_rate": 1.9947678284052667e-05, + "loss": 1.7532, + "step": 568 + }, + { + "epoch": 0.30890336590662326, + "grad_norm": 2.887133065759567, + "learning_rate": 1.9947318383375017e-05, + "loss": 1.8001, + "step": 569 + }, + { + "epoch": 0.30944625407166126, + "grad_norm": 3.6996119456889915, + "learning_rate": 1.9946957252397173e-05, + "loss": 2.4852, + "step": 570 + }, + { + "epoch": 0.30998914223669927, + "grad_norm": 2.829855772438557, + "learning_rate": 1.9946594891163808e-05, + "loss": 1.6048, + "step": 571 + }, + { + "epoch": 0.31053203040173727, + "grad_norm": 4.288772356209683, + "learning_rate": 1.9946231299719732e-05, + "loss": 1.4841, + "step": 572 + }, + { + "epoch": 0.31107491856677527, + "grad_norm": 2.37958986026152, + "learning_rate": 1.9945866478109914e-05, + "loss": 1.1797, + "step": 573 + }, + { + "epoch": 0.3116178067318133, + "grad_norm": 2.9977970906442932, + "learning_rate": 1.9945500426379483e-05, + "loss": 1.224, + "step": 574 + }, + { + "epoch": 0.3121606948968513, + "grad_norm": 3.406596416686285, + "learning_rate": 1.9945133144573705e-05, + "loss": 1.4793, + "step": 575 + }, + { + "epoch": 0.3127035830618892, + "grad_norm": 3.3772239188722244, + "learning_rate": 1.994476463273801e-05, + "loss": 1.3696, + "step": 576 + }, + { + "epoch": 0.3132464712269272, + "grad_norm": 2.729299768057245, + "learning_rate": 1.9944394890917977e-05, + "loss": 1.448, + "step": 577 + }, + { + "epoch": 0.31378935939196523, + "grad_norm": 2.6526406591248297, + "learning_rate": 1.9944023919159335e-05, + "loss": 1.6905, + "step": 578 + }, + { + "epoch": 0.31433224755700323, + "grad_norm": 2.9512501195242944, + "learning_rate": 1.9943651717507965e-05, + "loss": 1.6277, + "step": 579 + }, + { + "epoch": 0.31487513572204123, + "grad_norm": 2.8292191327941723, + "learning_rate": 1.9943278286009903e-05, + "loss": 1.0532, + "step": 580 + }, + { + "epoch": 0.31541802388707924, + "grad_norm": 3.2105870294745436, + "learning_rate": 1.9942903624711335e-05, + "loss": 1.1823, + "step": 581 + }, + { + "epoch": 0.31596091205211724, + "grad_norm": 3.224137115744835, + "learning_rate": 1.9942527733658602e-05, + "loss": 1.5409, + "step": 582 + }, + { + "epoch": 0.31650380021715524, + "grad_norm": 3.126134123334164, + "learning_rate": 1.9942150612898194e-05, + "loss": 1.2423, + "step": 583 + }, + { + "epoch": 0.31704668838219324, + "grad_norm": 2.709954974374804, + "learning_rate": 1.994177226247675e-05, + "loss": 1.3419, + "step": 584 + }, + { + "epoch": 0.31758957654723124, + "grad_norm": 2.9030367366777927, + "learning_rate": 1.9941392682441066e-05, + "loss": 1.427, + "step": 585 + }, + { + "epoch": 0.31813246471226925, + "grad_norm": 3.272875981776567, + "learning_rate": 1.9941011872838092e-05, + "loss": 2.0196, + "step": 586 + }, + { + "epoch": 0.31867535287730725, + "grad_norm": 3.3480185179867, + "learning_rate": 1.994062983371493e-05, + "loss": 1.6038, + "step": 587 + }, + { + "epoch": 0.31921824104234525, + "grad_norm": 2.801738772846361, + "learning_rate": 1.9940246565118822e-05, + "loss": 1.7505, + "step": 588 + }, + { + "epoch": 0.31976112920738325, + "grad_norm": 2.743550967049156, + "learning_rate": 1.993986206709718e-05, + "loss": 1.3914, + "step": 589 + }, + { + "epoch": 0.32030401737242126, + "grad_norm": 2.481845489278486, + "learning_rate": 1.9939476339697555e-05, + "loss": 1.6927, + "step": 590 + }, + { + "epoch": 0.32084690553745926, + "grad_norm": 3.145019330878407, + "learning_rate": 1.993908938296765e-05, + "loss": 1.42, + "step": 591 + }, + { + "epoch": 0.32138979370249726, + "grad_norm": 3.326427662456394, + "learning_rate": 1.9938701196955335e-05, + "loss": 1.2516, + "step": 592 + }, + { + "epoch": 0.32193268186753526, + "grad_norm": 3.195815443331326, + "learning_rate": 1.9938311781708616e-05, + "loss": 2.1428, + "step": 593 + }, + { + "epoch": 0.32247557003257327, + "grad_norm": 2.461395361556941, + "learning_rate": 1.9937921137275657e-05, + "loss": 1.4448, + "step": 594 + }, + { + "epoch": 0.32301845819761127, + "grad_norm": 3.1793351214549794, + "learning_rate": 1.993752926370477e-05, + "loss": 1.4609, + "step": 595 + }, + { + "epoch": 0.32356134636264927, + "grad_norm": 2.783909288864463, + "learning_rate": 1.9937136161044427e-05, + "loss": 1.3355, + "step": 596 + }, + { + "epoch": 0.3241042345276873, + "grad_norm": 2.8156929696256734, + "learning_rate": 1.9936741829343247e-05, + "loss": 2.101, + "step": 597 + }, + { + "epoch": 0.3246471226927253, + "grad_norm": 3.793352093788154, + "learning_rate": 1.993634626865e-05, + "loss": 2.19, + "step": 598 + }, + { + "epoch": 0.3251900108577633, + "grad_norm": 2.6680863350639545, + "learning_rate": 1.993594947901361e-05, + "loss": 1.5199, + "step": 599 + }, + { + "epoch": 0.3257328990228013, + "grad_norm": 2.6498586813134297, + "learning_rate": 1.9935551460483155e-05, + "loss": 0.9282, + "step": 600 + }, + { + "epoch": 0.3262757871878393, + "grad_norm": 3.443063689484479, + "learning_rate": 1.993515221310786e-05, + "loss": 1.9107, + "step": 601 + }, + { + "epoch": 0.3268186753528773, + "grad_norm": 2.45686855619251, + "learning_rate": 1.9934751736937103e-05, + "loss": 1.2929, + "step": 602 + }, + { + "epoch": 0.3273615635179153, + "grad_norm": 3.3974155688994077, + "learning_rate": 1.9934350032020417e-05, + "loss": 1.546, + "step": 603 + }, + { + "epoch": 0.3279044516829533, + "grad_norm": 2.3692191456624783, + "learning_rate": 1.993394709840749e-05, + "loss": 1.2239, + "step": 604 + }, + { + "epoch": 0.3284473398479913, + "grad_norm": 4.044352925058249, + "learning_rate": 1.993354293614815e-05, + "loss": 1.801, + "step": 605 + }, + { + "epoch": 0.3289902280130293, + "grad_norm": 4.051192057617293, + "learning_rate": 1.993313754529239e-05, + "loss": 1.8043, + "step": 606 + }, + { + "epoch": 0.3295331161780673, + "grad_norm": 3.3680281079594634, + "learning_rate": 1.9932730925890344e-05, + "loss": 1.4915, + "step": 607 + }, + { + "epoch": 0.3300760043431053, + "grad_norm": 2.993677842102555, + "learning_rate": 1.9932323077992312e-05, + "loss": 1.3457, + "step": 608 + }, + { + "epoch": 0.3306188925081433, + "grad_norm": 3.325666927669253, + "learning_rate": 1.9931914001648726e-05, + "loss": 1.4221, + "step": 609 + }, + { + "epoch": 0.3311617806731813, + "grad_norm": 2.459382097232459, + "learning_rate": 1.993150369691019e-05, + "loss": 1.2551, + "step": 610 + }, + { + "epoch": 0.3317046688382193, + "grad_norm": 5.281638549560053, + "learning_rate": 1.993109216382745e-05, + "loss": 1.142, + "step": 611 + }, + { + "epoch": 0.3322475570032573, + "grad_norm": 2.6931033883715374, + "learning_rate": 1.99306794024514e-05, + "loss": 1.2573, + "step": 612 + }, + { + "epoch": 0.3327904451682953, + "grad_norm": 2.8066393604642714, + "learning_rate": 1.9930265412833097e-05, + "loss": 0.9847, + "step": 613 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 3.682172765988149, + "learning_rate": 1.992985019502374e-05, + "loss": 1.4482, + "step": 614 + }, + { + "epoch": 0.3338762214983713, + "grad_norm": 3.7190248655227873, + "learning_rate": 1.9929433749074684e-05, + "loss": 1.3682, + "step": 615 + }, + { + "epoch": 0.3344191096634093, + "grad_norm": 2.8594798198555527, + "learning_rate": 1.9929016075037438e-05, + "loss": 1.2396, + "step": 616 + }, + { + "epoch": 0.3349619978284473, + "grad_norm": 3.5439348644341035, + "learning_rate": 1.992859717296366e-05, + "loss": 1.5843, + "step": 617 + }, + { + "epoch": 0.3355048859934853, + "grad_norm": 2.3447575030783283, + "learning_rate": 1.992817704290516e-05, + "loss": 1.2025, + "step": 618 + }, + { + "epoch": 0.3360477741585233, + "grad_norm": 3.027039186876229, + "learning_rate": 1.99277556849139e-05, + "loss": 1.7133, + "step": 619 + }, + { + "epoch": 0.33659066232356133, + "grad_norm": 2.9868315770124343, + "learning_rate": 1.9927333099041992e-05, + "loss": 1.2309, + "step": 620 + }, + { + "epoch": 0.33713355048859933, + "grad_norm": 3.215889472723888, + "learning_rate": 1.9926909285341705e-05, + "loss": 1.2217, + "step": 621 + }, + { + "epoch": 0.33767643865363733, + "grad_norm": 3.293119071439456, + "learning_rate": 1.992648424386546e-05, + "loss": 1.7828, + "step": 622 + }, + { + "epoch": 0.33821932681867534, + "grad_norm": 2.7023629491691623, + "learning_rate": 1.992605797466582e-05, + "loss": 1.3745, + "step": 623 + }, + { + "epoch": 0.33876221498371334, + "grad_norm": 4.234534944070885, + "learning_rate": 1.9925630477795514e-05, + "loss": 1.3415, + "step": 624 + }, + { + "epoch": 0.33930510314875134, + "grad_norm": 3.6994706259321815, + "learning_rate": 1.9925201753307406e-05, + "loss": 1.4234, + "step": 625 + }, + { + "epoch": 0.33984799131378934, + "grad_norm": 3.156992028368943, + "learning_rate": 1.992477180125453e-05, + "loss": 1.4845, + "step": 626 + }, + { + "epoch": 0.34039087947882735, + "grad_norm": 3.1286393814611397, + "learning_rate": 1.9924340621690058e-05, + "loss": 0.9685, + "step": 627 + }, + { + "epoch": 0.34093376764386535, + "grad_norm": 3.336970766050665, + "learning_rate": 1.9923908214667323e-05, + "loss": 1.3268, + "step": 628 + }, + { + "epoch": 0.34147665580890335, + "grad_norm": 4.290496076646803, + "learning_rate": 1.99234745802398e-05, + "loss": 1.2785, + "step": 629 + }, + { + "epoch": 0.34201954397394135, + "grad_norm": 2.7491777959300023, + "learning_rate": 1.9923039718461127e-05, + "loss": 1.3798, + "step": 630 + }, + { + "epoch": 0.34256243213897936, + "grad_norm": 5.312529540414799, + "learning_rate": 1.992260362938509e-05, + "loss": 2.2512, + "step": 631 + }, + { + "epoch": 0.34310532030401736, + "grad_norm": 3.9557409230221676, + "learning_rate": 1.9922166313065618e-05, + "loss": 1.849, + "step": 632 + }, + { + "epoch": 0.34364820846905536, + "grad_norm": 3.7327574664165786, + "learning_rate": 1.9921727769556805e-05, + "loss": 1.4078, + "step": 633 + }, + { + "epoch": 0.34419109663409336, + "grad_norm": 3.7644055867642594, + "learning_rate": 1.992128799891289e-05, + "loss": 1.714, + "step": 634 + }, + { + "epoch": 0.34473398479913137, + "grad_norm": 3.923952524278675, + "learning_rate": 1.9920847001188258e-05, + "loss": 1.7613, + "step": 635 + }, + { + "epoch": 0.34527687296416937, + "grad_norm": 3.1523646374054155, + "learning_rate": 1.992040477643746e-05, + "loss": 1.2563, + "step": 636 + }, + { + "epoch": 0.34581976112920737, + "grad_norm": 3.6157151039083444, + "learning_rate": 1.991996132471519e-05, + "loss": 1.1795, + "step": 637 + }, + { + "epoch": 0.3463626492942454, + "grad_norm": 4.222007576307266, + "learning_rate": 1.991951664607629e-05, + "loss": 1.6605, + "step": 638 + }, + { + "epoch": 0.3469055374592834, + "grad_norm": 3.1555785318795544, + "learning_rate": 1.9919070740575764e-05, + "loss": 1.526, + "step": 639 + }, + { + "epoch": 0.3474484256243214, + "grad_norm": 3.2645269107818677, + "learning_rate": 1.991862360826876e-05, + "loss": 1.3182, + "step": 640 + }, + { + "epoch": 0.3479913137893594, + "grad_norm": 3.53044955128402, + "learning_rate": 1.991817524921058e-05, + "loss": 1.5583, + "step": 641 + }, + { + "epoch": 0.3485342019543974, + "grad_norm": 3.24710260787145, + "learning_rate": 1.9917725663456678e-05, + "loss": 0.9878, + "step": 642 + }, + { + "epoch": 0.3490770901194354, + "grad_norm": 3.9154742975791654, + "learning_rate": 1.991727485106266e-05, + "loss": 1.6034, + "step": 643 + }, + { + "epoch": 0.3496199782844734, + "grad_norm": 3.8199965805178318, + "learning_rate": 1.9916822812084282e-05, + "loss": 1.3768, + "step": 644 + }, + { + "epoch": 0.3501628664495114, + "grad_norm": 3.198793670499035, + "learning_rate": 1.9916369546577455e-05, + "loss": 1.4308, + "step": 645 + }, + { + "epoch": 0.3507057546145494, + "grad_norm": 2.908177540585374, + "learning_rate": 1.9915915054598237e-05, + "loss": 1.3964, + "step": 646 + }, + { + "epoch": 0.3512486427795874, + "grad_norm": 2.920889003436479, + "learning_rate": 1.9915459336202844e-05, + "loss": 1.056, + "step": 647 + }, + { + "epoch": 0.3517915309446254, + "grad_norm": 3.7046400158340864, + "learning_rate": 1.991500239144763e-05, + "loss": 1.9052, + "step": 648 + }, + { + "epoch": 0.3523344191096634, + "grad_norm": 4.412988121206581, + "learning_rate": 1.9914544220389124e-05, + "loss": 1.944, + "step": 649 + }, + { + "epoch": 0.3528773072747014, + "grad_norm": 3.87124697251994, + "learning_rate": 1.9914084823083988e-05, + "loss": 1.4951, + "step": 650 + }, + { + "epoch": 0.3534201954397394, + "grad_norm": 3.8734663848108584, + "learning_rate": 1.9913624199589037e-05, + "loss": 1.5462, + "step": 651 + }, + { + "epoch": 0.3539630836047774, + "grad_norm": 2.7283359997443126, + "learning_rate": 1.9913162349961248e-05, + "loss": 0.9188, + "step": 652 + }, + { + "epoch": 0.3545059717698154, + "grad_norm": 2.885333878853398, + "learning_rate": 1.991269927425774e-05, + "loss": 1.2292, + "step": 653 + }, + { + "epoch": 0.3550488599348534, + "grad_norm": 3.2425772460053257, + "learning_rate": 1.9912234972535788e-05, + "loss": 1.2863, + "step": 654 + }, + { + "epoch": 0.3555917480998914, + "grad_norm": 3.060966260943164, + "learning_rate": 1.991176944485281e-05, + "loss": 1.5056, + "step": 655 + }, + { + "epoch": 0.3561346362649294, + "grad_norm": 4.533553635387312, + "learning_rate": 1.99113026912664e-05, + "loss": 1.8229, + "step": 656 + }, + { + "epoch": 0.3566775244299674, + "grad_norm": 2.8582150809054045, + "learning_rate": 1.9910834711834267e-05, + "loss": 1.3233, + "step": 657 + }, + { + "epoch": 0.3572204125950054, + "grad_norm": 4.930680768753159, + "learning_rate": 1.9910365506614308e-05, + "loss": 1.5997, + "step": 658 + }, + { + "epoch": 0.3577633007600434, + "grad_norm": 3.0733335483268083, + "learning_rate": 1.9909895075664545e-05, + "loss": 0.9206, + "step": 659 + }, + { + "epoch": 0.3583061889250814, + "grad_norm": 3.4659704922712686, + "learning_rate": 1.990942341904317e-05, + "loss": 1.2752, + "step": 660 + }, + { + "epoch": 0.35884907709011943, + "grad_norm": 3.493850558129149, + "learning_rate": 1.9908950536808508e-05, + "loss": 1.588, + "step": 661 + }, + { + "epoch": 0.35939196525515743, + "grad_norm": 3.61256229272583, + "learning_rate": 1.9908476429019056e-05, + "loss": 1.469, + "step": 662 + }, + { + "epoch": 0.35993485342019543, + "grad_norm": 3.3430037514089803, + "learning_rate": 1.9908001095733445e-05, + "loss": 1.3038, + "step": 663 + }, + { + "epoch": 0.36047774158523344, + "grad_norm": 3.8918915484374557, + "learning_rate": 1.9907524537010467e-05, + "loss": 1.3683, + "step": 664 + }, + { + "epoch": 0.36102062975027144, + "grad_norm": 3.6032272045245053, + "learning_rate": 1.9907046752909064e-05, + "loss": 1.8694, + "step": 665 + }, + { + "epoch": 0.36156351791530944, + "grad_norm": 4.546268377661869, + "learning_rate": 1.9906567743488326e-05, + "loss": 1.2871, + "step": 666 + }, + { + "epoch": 0.36210640608034744, + "grad_norm": 3.5283169402860777, + "learning_rate": 1.9906087508807504e-05, + "loss": 1.5334, + "step": 667 + }, + { + "epoch": 0.36264929424538545, + "grad_norm": 2.7984191262784504, + "learning_rate": 1.9905606048925993e-05, + "loss": 1.1924, + "step": 668 + }, + { + "epoch": 0.36319218241042345, + "grad_norm": 3.54703835950394, + "learning_rate": 1.9905123363903335e-05, + "loss": 1.5972, + "step": 669 + }, + { + "epoch": 0.36373507057546145, + "grad_norm": 3.0126938064772855, + "learning_rate": 1.9904639453799236e-05, + "loss": 1.2294, + "step": 670 + }, + { + "epoch": 0.36427795874049945, + "grad_norm": 3.16358553989926, + "learning_rate": 1.990415431867354e-05, + "loss": 1.3564, + "step": 671 + }, + { + "epoch": 0.36482084690553745, + "grad_norm": 4.118910539945259, + "learning_rate": 1.990366795858626e-05, + "loss": 1.6838, + "step": 672 + }, + { + "epoch": 0.36536373507057546, + "grad_norm": 3.8466533902335502, + "learning_rate": 1.9903180373597534e-05, + "loss": 1.7986, + "step": 673 + }, + { + "epoch": 0.36590662323561346, + "grad_norm": 3.6195384409794684, + "learning_rate": 1.990269156376768e-05, + "loss": 1.6113, + "step": 674 + }, + { + "epoch": 0.36644951140065146, + "grad_norm": 3.509724379422402, + "learning_rate": 1.9902201529157152e-05, + "loss": 1.5496, + "step": 675 + }, + { + "epoch": 0.36699239956568946, + "grad_norm": 3.579737463592409, + "learning_rate": 1.9901710269826554e-05, + "loss": 1.4856, + "step": 676 + }, + { + "epoch": 0.36753528773072747, + "grad_norm": 3.780431786449365, + "learning_rate": 1.9901217785836655e-05, + "loss": 1.8519, + "step": 677 + }, + { + "epoch": 0.36807817589576547, + "grad_norm": 4.0659352627131735, + "learning_rate": 1.9900724077248354e-05, + "loss": 1.4382, + "step": 678 + }, + { + "epoch": 0.36862106406080347, + "grad_norm": 2.9524656376446274, + "learning_rate": 1.9900229144122723e-05, + "loss": 1.4541, + "step": 679 + }, + { + "epoch": 0.3691639522258415, + "grad_norm": 2.388269642575342, + "learning_rate": 1.989973298652097e-05, + "loss": 0.9567, + "step": 680 + }, + { + "epoch": 0.3697068403908795, + "grad_norm": 2.8322419431529453, + "learning_rate": 1.9899235604504467e-05, + "loss": 1.1187, + "step": 681 + }, + { + "epoch": 0.3702497285559175, + "grad_norm": 3.0469435898757613, + "learning_rate": 1.9898736998134726e-05, + "loss": 1.5571, + "step": 682 + }, + { + "epoch": 0.3707926167209555, + "grad_norm": 3.623982705749655, + "learning_rate": 1.9898237167473416e-05, + "loss": 1.2047, + "step": 683 + }, + { + "epoch": 0.3713355048859935, + "grad_norm": 3.669362267695381, + "learning_rate": 1.9897736112582357e-05, + "loss": 1.7747, + "step": 684 + }, + { + "epoch": 0.3718783930510315, + "grad_norm": 2.9109325710576353, + "learning_rate": 1.989723383352352e-05, + "loss": 1.0044, + "step": 685 + }, + { + "epoch": 0.3724212812160695, + "grad_norm": 3.4478434843760146, + "learning_rate": 1.9896730330359032e-05, + "loss": 1.1245, + "step": 686 + }, + { + "epoch": 0.3729641693811075, + "grad_norm": 3.2957514959337275, + "learning_rate": 1.989622560315116e-05, + "loss": 1.3243, + "step": 687 + }, + { + "epoch": 0.3735070575461455, + "grad_norm": 3.5500423086355988, + "learning_rate": 1.989571965196234e-05, + "loss": 2.0478, + "step": 688 + }, + { + "epoch": 0.3740499457111835, + "grad_norm": 2.621987663839103, + "learning_rate": 1.9895212476855136e-05, + "loss": 1.3135, + "step": 689 + }, + { + "epoch": 0.3745928338762215, + "grad_norm": 4.055390555970487, + "learning_rate": 1.989470407789228e-05, + "loss": 1.5165, + "step": 690 + }, + { + "epoch": 0.3751357220412595, + "grad_norm": 2.94204110816024, + "learning_rate": 1.989419445513666e-05, + "loss": 1.3458, + "step": 691 + }, + { + "epoch": 0.3756786102062975, + "grad_norm": 3.5714132672245977, + "learning_rate": 1.98936836086513e-05, + "loss": 1.3034, + "step": 692 + }, + { + "epoch": 0.3762214983713355, + "grad_norm": 3.659137782783938, + "learning_rate": 1.9893171538499382e-05, + "loss": 1.6203, + "step": 693 + }, + { + "epoch": 0.3767643865363735, + "grad_norm": 5.116097300755018, + "learning_rate": 1.9892658244744236e-05, + "loss": 2.1071, + "step": 694 + }, + { + "epoch": 0.3773072747014115, + "grad_norm": 2.8449662018307005, + "learning_rate": 1.9892143727449357e-05, + "loss": 1.0477, + "step": 695 + }, + { + "epoch": 0.3778501628664495, + "grad_norm": 4.021030963638319, + "learning_rate": 1.989162798667838e-05, + "loss": 1.9528, + "step": 696 + }, + { + "epoch": 0.3783930510314875, + "grad_norm": 3.377368666687089, + "learning_rate": 1.989111102249508e-05, + "loss": 1.2481, + "step": 697 + }, + { + "epoch": 0.3789359391965255, + "grad_norm": 3.403268816169458, + "learning_rate": 1.9890592834963406e-05, + "loss": 1.1864, + "step": 698 + }, + { + "epoch": 0.3794788273615635, + "grad_norm": 3.082879601892987, + "learning_rate": 1.9890073424147453e-05, + "loss": 1.7365, + "step": 699 + }, + { + "epoch": 0.3800217155266015, + "grad_norm": 4.122610250215441, + "learning_rate": 1.988955279011145e-05, + "loss": 1.4139, + "step": 700 + }, + { + "epoch": 0.3805646036916395, + "grad_norm": 3.298087120576001, + "learning_rate": 1.98890309329198e-05, + "loss": 1.5904, + "step": 701 + }, + { + "epoch": 0.3811074918566775, + "grad_norm": 3.1183850173888703, + "learning_rate": 1.9888507852637043e-05, + "loss": 1.2146, + "step": 702 + }, + { + "epoch": 0.38165038002171553, + "grad_norm": 3.128042971411966, + "learning_rate": 1.9887983549327873e-05, + "loss": 1.2751, + "step": 703 + }, + { + "epoch": 0.38219326818675353, + "grad_norm": 3.362555156234201, + "learning_rate": 1.988745802305714e-05, + "loss": 1.1363, + "step": 704 + }, + { + "epoch": 0.38273615635179153, + "grad_norm": 3.29574344438116, + "learning_rate": 1.988693127388984e-05, + "loss": 1.227, + "step": 705 + }, + { + "epoch": 0.38327904451682954, + "grad_norm": 4.2418853227437205, + "learning_rate": 1.9886403301891123e-05, + "loss": 1.7091, + "step": 706 + }, + { + "epoch": 0.38382193268186754, + "grad_norm": 4.165071133964158, + "learning_rate": 1.9885874107126287e-05, + "loss": 1.9403, + "step": 707 + }, + { + "epoch": 0.38436482084690554, + "grad_norm": 4.639609386348944, + "learning_rate": 1.9885343689660787e-05, + "loss": 1.221, + "step": 708 + }, + { + "epoch": 0.38490770901194354, + "grad_norm": 3.0746775894794034, + "learning_rate": 1.9884812049560226e-05, + "loss": 1.4958, + "step": 709 + }, + { + "epoch": 0.38545059717698155, + "grad_norm": 4.822804859325608, + "learning_rate": 1.9884279186890357e-05, + "loss": 1.6928, + "step": 710 + }, + { + "epoch": 0.38599348534201955, + "grad_norm": 4.299811491105524, + "learning_rate": 1.9883745101717084e-05, + "loss": 1.3075, + "step": 711 + }, + { + "epoch": 0.38653637350705755, + "grad_norm": 3.882170994492157, + "learning_rate": 1.9883209794106464e-05, + "loss": 1.9712, + "step": 712 + }, + { + "epoch": 0.38707926167209555, + "grad_norm": 4.164019029255246, + "learning_rate": 1.9882673264124705e-05, + "loss": 1.6109, + "step": 713 + }, + { + "epoch": 0.38762214983713356, + "grad_norm": 3.691226317297577, + "learning_rate": 1.9882135511838167e-05, + "loss": 1.3225, + "step": 714 + }, + { + "epoch": 0.38816503800217156, + "grad_norm": 5.777282505586183, + "learning_rate": 1.988159653731336e-05, + "loss": 1.7945, + "step": 715 + }, + { + "epoch": 0.38870792616720956, + "grad_norm": 3.3926512133724165, + "learning_rate": 1.9881056340616944e-05, + "loss": 1.5797, + "step": 716 + }, + { + "epoch": 0.38925081433224756, + "grad_norm": 3.7756965718668467, + "learning_rate": 1.988051492181573e-05, + "loss": 1.5515, + "step": 717 + }, + { + "epoch": 0.38979370249728557, + "grad_norm": 3.5509658648264613, + "learning_rate": 1.987997228097668e-05, + "loss": 1.2004, + "step": 718 + }, + { + "epoch": 0.39033659066232357, + "grad_norm": 4.682493032945398, + "learning_rate": 1.987942841816692e-05, + "loss": 1.2957, + "step": 719 + }, + { + "epoch": 0.39087947882736157, + "grad_norm": 2.923319971884201, + "learning_rate": 1.9878883333453704e-05, + "loss": 0.753, + "step": 720 + }, + { + "epoch": 0.3914223669923996, + "grad_norm": 2.8706465146507845, + "learning_rate": 1.987833702690445e-05, + "loss": 1.1057, + "step": 721 + }, + { + "epoch": 0.3919652551574376, + "grad_norm": 3.4516078574834057, + "learning_rate": 1.987778949858673e-05, + "loss": 1.1275, + "step": 722 + }, + { + "epoch": 0.3925081433224756, + "grad_norm": 3.3400553957950567, + "learning_rate": 1.9877240748568263e-05, + "loss": 1.1538, + "step": 723 + }, + { + "epoch": 0.3930510314875136, + "grad_norm": 3.155540544638446, + "learning_rate": 1.987669077691692e-05, + "loss": 1.2486, + "step": 724 + }, + { + "epoch": 0.3935939196525516, + "grad_norm": 3.1450276486107054, + "learning_rate": 1.987613958370072e-05, + "loss": 1.265, + "step": 725 + }, + { + "epoch": 0.3941368078175896, + "grad_norm": 3.0493327058864406, + "learning_rate": 1.9875587168987834e-05, + "loss": 1.2097, + "step": 726 + }, + { + "epoch": 0.3946796959826276, + "grad_norm": 3.716259137065701, + "learning_rate": 1.987503353284659e-05, + "loss": 1.5386, + "step": 727 + }, + { + "epoch": 0.3952225841476656, + "grad_norm": 3.441725300410296, + "learning_rate": 1.9874478675345458e-05, + "loss": 1.3936, + "step": 728 + }, + { + "epoch": 0.3957654723127036, + "grad_norm": 3.773353353497139, + "learning_rate": 1.9873922596553067e-05, + "loss": 1.731, + "step": 729 + }, + { + "epoch": 0.3963083604777416, + "grad_norm": 3.527645597828566, + "learning_rate": 1.987336529653819e-05, + "loss": 1.708, + "step": 730 + }, + { + "epoch": 0.3968512486427796, + "grad_norm": 3.8440953147603643, + "learning_rate": 1.9872806775369762e-05, + "loss": 1.6102, + "step": 731 + }, + { + "epoch": 0.3973941368078176, + "grad_norm": 3.4036622029265966, + "learning_rate": 1.9872247033116855e-05, + "loss": 0.9298, + "step": 732 + }, + { + "epoch": 0.3979370249728556, + "grad_norm": 4.034027632823603, + "learning_rate": 1.98716860698487e-05, + "loss": 1.78, + "step": 733 + }, + { + "epoch": 0.3984799131378936, + "grad_norm": 4.10524690369694, + "learning_rate": 1.987112388563468e-05, + "loss": 1.2354, + "step": 734 + }, + { + "epoch": 0.3990228013029316, + "grad_norm": 4.059289261823474, + "learning_rate": 1.9870560480544325e-05, + "loss": 1.804, + "step": 735 + }, + { + "epoch": 0.3995656894679696, + "grad_norm": 3.2544839104292667, + "learning_rate": 1.986999585464732e-05, + "loss": 1.1228, + "step": 736 + }, + { + "epoch": 0.4001085776330076, + "grad_norm": 3.741117309135591, + "learning_rate": 1.9869430008013496e-05, + "loss": 1.329, + "step": 737 + }, + { + "epoch": 0.4006514657980456, + "grad_norm": 4.446330579981585, + "learning_rate": 1.9868862940712838e-05, + "loss": 1.6506, + "step": 738 + }, + { + "epoch": 0.4011943539630836, + "grad_norm": 3.139296317304318, + "learning_rate": 1.9868294652815483e-05, + "loss": 1.144, + "step": 739 + }, + { + "epoch": 0.4017372421281216, + "grad_norm": 4.025270331554587, + "learning_rate": 1.986772514439172e-05, + "loss": 1.336, + "step": 740 + }, + { + "epoch": 0.4022801302931596, + "grad_norm": 3.4545518889756557, + "learning_rate": 1.986715441551198e-05, + "loss": 1.6006, + "step": 741 + }, + { + "epoch": 0.4028230184581976, + "grad_norm": 3.748598610621174, + "learning_rate": 1.986658246624686e-05, + "loss": 1.4812, + "step": 742 + }, + { + "epoch": 0.4033659066232356, + "grad_norm": 5.126058311631635, + "learning_rate": 1.9866009296667093e-05, + "loss": 2.0129, + "step": 743 + }, + { + "epoch": 0.40390879478827363, + "grad_norm": 8.145692299678405, + "learning_rate": 1.9865434906843574e-05, + "loss": 1.8321, + "step": 744 + }, + { + "epoch": 0.40445168295331163, + "grad_norm": 4.23918362592579, + "learning_rate": 1.9864859296847343e-05, + "loss": 1.5765, + "step": 745 + }, + { + "epoch": 0.40499457111834963, + "grad_norm": 4.85145039296263, + "learning_rate": 1.986428246674959e-05, + "loss": 1.6486, + "step": 746 + }, + { + "epoch": 0.40553745928338764, + "grad_norm": 3.7681072947070415, + "learning_rate": 1.986370441662166e-05, + "loss": 1.4466, + "step": 747 + }, + { + "epoch": 0.40608034744842564, + "grad_norm": 4.206136090938125, + "learning_rate": 1.986312514653505e-05, + "loss": 1.6717, + "step": 748 + }, + { + "epoch": 0.40662323561346364, + "grad_norm": 3.7336223866799965, + "learning_rate": 1.9862544656561403e-05, + "loss": 0.9599, + "step": 749 + }, + { + "epoch": 0.40716612377850164, + "grad_norm": 4.037759008136602, + "learning_rate": 1.986196294677251e-05, + "loss": 1.9387, + "step": 750 + }, + { + "epoch": 0.40770901194353965, + "grad_norm": 3.8832525632606156, + "learning_rate": 1.9861380017240324e-05, + "loss": 2.0184, + "step": 751 + }, + { + "epoch": 0.40825190010857765, + "grad_norm": 3.9521376069923546, + "learning_rate": 1.986079586803694e-05, + "loss": 1.0926, + "step": 752 + }, + { + "epoch": 0.40879478827361565, + "grad_norm": 5.128973903987411, + "learning_rate": 1.986021049923461e-05, + "loss": 1.5943, + "step": 753 + }, + { + "epoch": 0.40933767643865365, + "grad_norm": 4.5372892013581865, + "learning_rate": 1.9859623910905728e-05, + "loss": 1.643, + "step": 754 + }, + { + "epoch": 0.40988056460369166, + "grad_norm": 4.2137978897196415, + "learning_rate": 1.985903610312285e-05, + "loss": 1.7237, + "step": 755 + }, + { + "epoch": 0.41042345276872966, + "grad_norm": 2.713265124864733, + "learning_rate": 1.985844707595867e-05, + "loss": 1.0521, + "step": 756 + }, + { + "epoch": 0.41096634093376766, + "grad_norm": 3.362872283003281, + "learning_rate": 1.9857856829486045e-05, + "loss": 1.2879, + "step": 757 + }, + { + "epoch": 0.41150922909880566, + "grad_norm": 4.022235417044952, + "learning_rate": 1.9857265363777975e-05, + "loss": 1.398, + "step": 758 + }, + { + "epoch": 0.41205211726384366, + "grad_norm": 4.083280210123792, + "learning_rate": 1.9856672678907616e-05, + "loss": 1.6652, + "step": 759 + }, + { + "epoch": 0.41259500542888167, + "grad_norm": 4.469771898345589, + "learning_rate": 1.985607877494827e-05, + "loss": 1.6242, + "step": 760 + }, + { + "epoch": 0.41313789359391967, + "grad_norm": 3.0490808076449016, + "learning_rate": 1.9855483651973396e-05, + "loss": 1.0783, + "step": 761 + }, + { + "epoch": 0.41368078175895767, + "grad_norm": 3.488230005104396, + "learning_rate": 1.9854887310056593e-05, + "loss": 1.3953, + "step": 762 + }, + { + "epoch": 0.4142236699239957, + "grad_norm": 3.5610333355465653, + "learning_rate": 1.9854289749271624e-05, + "loss": 1.3692, + "step": 763 + }, + { + "epoch": 0.4147665580890337, + "grad_norm": 3.5572639096784533, + "learning_rate": 1.9853690969692393e-05, + "loss": 1.646, + "step": 764 + }, + { + "epoch": 0.4153094462540717, + "grad_norm": 3.811741925703646, + "learning_rate": 1.9853090971392953e-05, + "loss": 1.6637, + "step": 765 + }, + { + "epoch": 0.4158523344191097, + "grad_norm": 2.912609699248709, + "learning_rate": 1.9852489754447526e-05, + "loss": 0.9086, + "step": 766 + }, + { + "epoch": 0.4163952225841477, + "grad_norm": 3.7608762982374646, + "learning_rate": 1.985188731893046e-05, + "loss": 1.5131, + "step": 767 + }, + { + "epoch": 0.4169381107491857, + "grad_norm": 3.776000457813264, + "learning_rate": 1.985128366491627e-05, + "loss": 1.5929, + "step": 768 + }, + { + "epoch": 0.4174809989142237, + "grad_norm": 4.129873624081247, + "learning_rate": 1.9850678792479613e-05, + "loss": 1.5461, + "step": 769 + }, + { + "epoch": 0.4180238870792617, + "grad_norm": 3.2233425858963645, + "learning_rate": 1.9850072701695306e-05, + "loss": 0.9614, + "step": 770 + }, + { + "epoch": 0.4185667752442997, + "grad_norm": 3.8215876727885054, + "learning_rate": 1.984946539263831e-05, + "loss": 1.2549, + "step": 771 + }, + { + "epoch": 0.4191096634093377, + "grad_norm": 3.0062072101032125, + "learning_rate": 1.9848856865383732e-05, + "loss": 0.9796, + "step": 772 + }, + { + "epoch": 0.4196525515743757, + "grad_norm": 2.937513473620578, + "learning_rate": 1.984824712000684e-05, + "loss": 1.0204, + "step": 773 + }, + { + "epoch": 0.4201954397394137, + "grad_norm": 2.949279428268976, + "learning_rate": 1.984763615658305e-05, + "loss": 1.332, + "step": 774 + }, + { + "epoch": 0.4207383279044517, + "grad_norm": 3.455520053658832, + "learning_rate": 1.9847023975187925e-05, + "loss": 1.3786, + "step": 775 + }, + { + "epoch": 0.4212812160694897, + "grad_norm": 3.8352233275650733, + "learning_rate": 1.9846410575897183e-05, + "loss": 1.4897, + "step": 776 + }, + { + "epoch": 0.4218241042345277, + "grad_norm": 3.262791090623995, + "learning_rate": 1.984579595878669e-05, + "loss": 1.5097, + "step": 777 + }, + { + "epoch": 0.4223669923995657, + "grad_norm": 3.603809745585289, + "learning_rate": 1.9845180123932456e-05, + "loss": 1.8419, + "step": 778 + }, + { + "epoch": 0.4229098805646037, + "grad_norm": 3.9648489260279196, + "learning_rate": 1.9844563071410656e-05, + "loss": 1.5159, + "step": 779 + }, + { + "epoch": 0.4234527687296417, + "grad_norm": 4.576987136894564, + "learning_rate": 1.9843944801297605e-05, + "loss": 1.6866, + "step": 780 + }, + { + "epoch": 0.4239956568946797, + "grad_norm": 3.4164331578089406, + "learning_rate": 1.9843325313669774e-05, + "loss": 1.5533, + "step": 781 + }, + { + "epoch": 0.4245385450597177, + "grad_norm": 3.2121441901059367, + "learning_rate": 1.9842704608603774e-05, + "loss": 0.8879, + "step": 782 + }, + { + "epoch": 0.4250814332247557, + "grad_norm": 3.6213877563860537, + "learning_rate": 1.9842082686176388e-05, + "loss": 1.3098, + "step": 783 + }, + { + "epoch": 0.4256243213897937, + "grad_norm": 3.719993086277978, + "learning_rate": 1.9841459546464527e-05, + "loss": 1.331, + "step": 784 + }, + { + "epoch": 0.4261672095548317, + "grad_norm": 2.7952357248876516, + "learning_rate": 1.9840835189545266e-05, + "loss": 1.1298, + "step": 785 + }, + { + "epoch": 0.42671009771986973, + "grad_norm": 3.450597586563723, + "learning_rate": 1.9840209615495822e-05, + "loss": 1.0023, + "step": 786 + }, + { + "epoch": 0.42725298588490773, + "grad_norm": 4.338480682869939, + "learning_rate": 1.983958282439357e-05, + "loss": 1.7044, + "step": 787 + }, + { + "epoch": 0.42779587404994573, + "grad_norm": 3.4831324579282765, + "learning_rate": 1.983895481631603e-05, + "loss": 1.4948, + "step": 788 + }, + { + "epoch": 0.42833876221498374, + "grad_norm": 6.382810905837672, + "learning_rate": 1.9838325591340885e-05, + "loss": 2.3008, + "step": 789 + }, + { + "epoch": 0.42888165038002174, + "grad_norm": 3.063848408601439, + "learning_rate": 1.9837695149545945e-05, + "loss": 1.2045, + "step": 790 + }, + { + "epoch": 0.42942453854505974, + "grad_norm": 4.80361465821882, + "learning_rate": 1.9837063491009193e-05, + "loss": 1.2552, + "step": 791 + }, + { + "epoch": 0.42996742671009774, + "grad_norm": 4.16884813133452, + "learning_rate": 1.9836430615808745e-05, + "loss": 1.6618, + "step": 792 + }, + { + "epoch": 0.43051031487513575, + "grad_norm": 4.215747683493222, + "learning_rate": 1.9835796524022886e-05, + "loss": 1.2571, + "step": 793 + }, + { + "epoch": 0.43105320304017375, + "grad_norm": 4.894853560761769, + "learning_rate": 1.9835161215730038e-05, + "loss": 0.793, + "step": 794 + }, + { + "epoch": 0.43159609120521175, + "grad_norm": 3.264854272979086, + "learning_rate": 1.983452469100877e-05, + "loss": 1.0721, + "step": 795 + }, + { + "epoch": 0.43213897937024975, + "grad_norm": 3.2346237132359335, + "learning_rate": 1.9833886949937823e-05, + "loss": 1.2396, + "step": 796 + }, + { + "epoch": 0.43268186753528776, + "grad_norm": 3.7265266370894023, + "learning_rate": 1.9833247992596058e-05, + "loss": 1.8845, + "step": 797 + }, + { + "epoch": 0.43322475570032576, + "grad_norm": 3.0655010730082126, + "learning_rate": 1.9832607819062513e-05, + "loss": 1.556, + "step": 798 + }, + { + "epoch": 0.43376764386536376, + "grad_norm": 3.209417077142042, + "learning_rate": 1.983196642941636e-05, + "loss": 1.2183, + "step": 799 + }, + { + "epoch": 0.43431053203040176, + "grad_norm": 4.384697511143271, + "learning_rate": 1.9831323823736933e-05, + "loss": 2.0889, + "step": 800 + }, + { + "epoch": 0.43485342019543977, + "grad_norm": 3.2773759507081084, + "learning_rate": 1.9830680002103703e-05, + "loss": 0.919, + "step": 801 + }, + { + "epoch": 0.43539630836047777, + "grad_norm": 4.1436406822638805, + "learning_rate": 1.9830034964596304e-05, + "loss": 1.7759, + "step": 802 + }, + { + "epoch": 0.43593919652551577, + "grad_norm": 5.715436881791243, + "learning_rate": 1.9829388711294512e-05, + "loss": 1.4277, + "step": 803 + }, + { + "epoch": 0.4364820846905538, + "grad_norm": 3.166985124043999, + "learning_rate": 1.982874124227826e-05, + "loss": 1.56, + "step": 804 + }, + { + "epoch": 0.4370249728555918, + "grad_norm": 4.237249076660815, + "learning_rate": 1.9828092557627626e-05, + "loss": 1.2366, + "step": 805 + }, + { + "epoch": 0.4375678610206297, + "grad_norm": 3.073398749866342, + "learning_rate": 1.982744265742284e-05, + "loss": 1.1475, + "step": 806 + }, + { + "epoch": 0.4381107491856677, + "grad_norm": 3.356777818996544, + "learning_rate": 1.9826791541744285e-05, + "loss": 1.2666, + "step": 807 + }, + { + "epoch": 0.4386536373507057, + "grad_norm": 3.171383453112648, + "learning_rate": 1.982613921067249e-05, + "loss": 1.1407, + "step": 808 + }, + { + "epoch": 0.43919652551574373, + "grad_norm": 3.334493176451811, + "learning_rate": 1.9825485664288138e-05, + "loss": 1.3309, + "step": 809 + }, + { + "epoch": 0.43973941368078173, + "grad_norm": 3.199045603289457, + "learning_rate": 1.982483090267206e-05, + "loss": 1.1634, + "step": 810 + }, + { + "epoch": 0.44028230184581973, + "grad_norm": 2.9970299057571856, + "learning_rate": 1.9824174925905235e-05, + "loss": 0.9642, + "step": 811 + }, + { + "epoch": 0.44082519001085774, + "grad_norm": 3.9851247078132994, + "learning_rate": 1.98235177340688e-05, + "loss": 1.4926, + "step": 812 + }, + { + "epoch": 0.44136807817589574, + "grad_norm": 3.586604601363014, + "learning_rate": 1.9822859327244034e-05, + "loss": 1.3105, + "step": 813 + }, + { + "epoch": 0.44191096634093374, + "grad_norm": 4.198634753759357, + "learning_rate": 1.9822199705512372e-05, + "loss": 1.4437, + "step": 814 + }, + { + "epoch": 0.44245385450597174, + "grad_norm": 3.6499375531563096, + "learning_rate": 1.9821538868955394e-05, + "loss": 1.2063, + "step": 815 + }, + { + "epoch": 0.44299674267100975, + "grad_norm": 4.3973905860052716, + "learning_rate": 1.9820876817654836e-05, + "loss": 1.5002, + "step": 816 + }, + { + "epoch": 0.44353963083604775, + "grad_norm": 2.7871960857121287, + "learning_rate": 1.9820213551692585e-05, + "loss": 0.6589, + "step": 817 + }, + { + "epoch": 0.44408251900108575, + "grad_norm": 3.4126665543063184, + "learning_rate": 1.981954907115067e-05, + "loss": 1.1651, + "step": 818 + }, + { + "epoch": 0.44462540716612375, + "grad_norm": 4.229651941034759, + "learning_rate": 1.981888337611127e-05, + "loss": 1.6444, + "step": 819 + }, + { + "epoch": 0.44516829533116176, + "grad_norm": 3.4135993440364056, + "learning_rate": 1.981821646665673e-05, + "loss": 1.2614, + "step": 820 + }, + { + "epoch": 0.44571118349619976, + "grad_norm": 4.274078711060124, + "learning_rate": 1.9817548342869527e-05, + "loss": 1.66, + "step": 821 + }, + { + "epoch": 0.44625407166123776, + "grad_norm": 2.9406863812447903, + "learning_rate": 1.98168790048323e-05, + "loss": 1.0412, + "step": 822 + }, + { + "epoch": 0.44679695982627576, + "grad_norm": 4.300279758003301, + "learning_rate": 1.981620845262783e-05, + "loss": 1.3689, + "step": 823 + }, + { + "epoch": 0.44733984799131377, + "grad_norm": 3.130525530503507, + "learning_rate": 1.9815536686339056e-05, + "loss": 1.1771, + "step": 824 + }, + { + "epoch": 0.44788273615635177, + "grad_norm": 3.6606030046920597, + "learning_rate": 1.981486370604906e-05, + "loss": 1.0323, + "step": 825 + }, + { + "epoch": 0.44842562432138977, + "grad_norm": 3.495865966877518, + "learning_rate": 1.981418951184108e-05, + "loss": 1.4493, + "step": 826 + }, + { + "epoch": 0.4489685124864278, + "grad_norm": 3.621229661878418, + "learning_rate": 1.9813514103798498e-05, + "loss": 1.5392, + "step": 827 + }, + { + "epoch": 0.4495114006514658, + "grad_norm": 3.6798995423063485, + "learning_rate": 1.9812837482004853e-05, + "loss": 1.0593, + "step": 828 + }, + { + "epoch": 0.4500542888165038, + "grad_norm": 3.4514040085362954, + "learning_rate": 1.9812159646543824e-05, + "loss": 0.9262, + "step": 829 + }, + { + "epoch": 0.4505971769815418, + "grad_norm": 3.195454430712678, + "learning_rate": 1.9811480597499257e-05, + "loss": 1.1469, + "step": 830 + }, + { + "epoch": 0.4511400651465798, + "grad_norm": 4.590874916307442, + "learning_rate": 1.981080033495513e-05, + "loss": 1.6534, + "step": 831 + }, + { + "epoch": 0.4516829533116178, + "grad_norm": 3.668702667937113, + "learning_rate": 1.981011885899558e-05, + "loss": 1.0597, + "step": 832 + }, + { + "epoch": 0.4522258414766558, + "grad_norm": 2.7669639542025046, + "learning_rate": 1.98094361697049e-05, + "loss": 1.0387, + "step": 833 + }, + { + "epoch": 0.4527687296416938, + "grad_norm": 4.734510249253564, + "learning_rate": 1.9808752267167515e-05, + "loss": 1.6757, + "step": 834 + }, + { + "epoch": 0.4533116178067318, + "grad_norm": 9.006078348758516, + "learning_rate": 1.9808067151468018e-05, + "loss": 2.1559, + "step": 835 + }, + { + "epoch": 0.4538545059717698, + "grad_norm": 4.169764802563779, + "learning_rate": 1.9807380822691146e-05, + "loss": 1.2557, + "step": 836 + }, + { + "epoch": 0.4543973941368078, + "grad_norm": 3.970541817152741, + "learning_rate": 1.9806693280921783e-05, + "loss": 1.5586, + "step": 837 + }, + { + "epoch": 0.4549402823018458, + "grad_norm": 5.042448582457387, + "learning_rate": 1.9806004526244966e-05, + "loss": 1.3191, + "step": 838 + }, + { + "epoch": 0.4554831704668838, + "grad_norm": 3.7467475303782236, + "learning_rate": 1.9805314558745876e-05, + "loss": 1.3318, + "step": 839 + }, + { + "epoch": 0.4560260586319218, + "grad_norm": 4.214820131429537, + "learning_rate": 1.980462337850986e-05, + "loss": 1.3958, + "step": 840 + }, + { + "epoch": 0.4565689467969598, + "grad_norm": 3.9655409992603965, + "learning_rate": 1.980393098562239e-05, + "loss": 0.9978, + "step": 841 + }, + { + "epoch": 0.4571118349619978, + "grad_norm": 5.522638140748038, + "learning_rate": 1.980323738016912e-05, + "loss": 1.4151, + "step": 842 + }, + { + "epoch": 0.4576547231270358, + "grad_norm": 3.84359618797871, + "learning_rate": 1.9802542562235822e-05, + "loss": 1.3567, + "step": 843 + }, + { + "epoch": 0.4581976112920738, + "grad_norm": 5.115138943098397, + "learning_rate": 1.9801846531908437e-05, + "loss": 1.6437, + "step": 844 + }, + { + "epoch": 0.4587404994571118, + "grad_norm": 3.811847477591368, + "learning_rate": 1.9801149289273054e-05, + "loss": 1.2335, + "step": 845 + }, + { + "epoch": 0.4592833876221498, + "grad_norm": 4.4009848681180115, + "learning_rate": 1.98004508344159e-05, + "loss": 1.2147, + "step": 846 + }, + { + "epoch": 0.4598262757871878, + "grad_norm": 3.850991850744113, + "learning_rate": 1.9799751167423376e-05, + "loss": 1.4865, + "step": 847 + }, + { + "epoch": 0.4603691639522258, + "grad_norm": 3.1303042065802917, + "learning_rate": 1.9799050288382004e-05, + "loss": 1.4926, + "step": 848 + }, + { + "epoch": 0.4609120521172638, + "grad_norm": 4.225513214699549, + "learning_rate": 1.9798348197378475e-05, + "loss": 1.2044, + "step": 849 + }, + { + "epoch": 0.46145494028230183, + "grad_norm": 4.139979880818486, + "learning_rate": 1.9797644894499625e-05, + "loss": 1.5682, + "step": 850 + }, + { + "epoch": 0.46199782844733983, + "grad_norm": 4.414209735978709, + "learning_rate": 1.979694037983244e-05, + "loss": 1.9033, + "step": 851 + }, + { + "epoch": 0.46254071661237783, + "grad_norm": 5.079031167256292, + "learning_rate": 1.9796234653464057e-05, + "loss": 1.8745, + "step": 852 + }, + { + "epoch": 0.46308360477741584, + "grad_norm": 6.59971413476838, + "learning_rate": 1.9795527715481755e-05, + "loss": 1.2449, + "step": 853 + }, + { + "epoch": 0.46362649294245384, + "grad_norm": 3.863740127637264, + "learning_rate": 1.9794819565972973e-05, + "loss": 1.3912, + "step": 854 + }, + { + "epoch": 0.46416938110749184, + "grad_norm": 3.8578186204961327, + "learning_rate": 1.9794110205025302e-05, + "loss": 1.3487, + "step": 855 + }, + { + "epoch": 0.46471226927252984, + "grad_norm": 4.19751015099396, + "learning_rate": 1.9793399632726466e-05, + "loss": 1.129, + "step": 856 + }, + { + "epoch": 0.46525515743756785, + "grad_norm": 4.261483190159906, + "learning_rate": 1.979268784916436e-05, + "loss": 1.7298, + "step": 857 + }, + { + "epoch": 0.46579804560260585, + "grad_norm": 4.1774094914047435, + "learning_rate": 1.9791974854427008e-05, + "loss": 0.8716, + "step": 858 + }, + { + "epoch": 0.46634093376764385, + "grad_norm": 3.845723101011648, + "learning_rate": 1.9791260648602603e-05, + "loss": 0.9086, + "step": 859 + }, + { + "epoch": 0.46688382193268185, + "grad_norm": 3.4922208632687393, + "learning_rate": 1.9790545231779476e-05, + "loss": 1.3688, + "step": 860 + }, + { + "epoch": 0.46742671009771986, + "grad_norm": 2.976761133296452, + "learning_rate": 1.978982860404611e-05, + "loss": 0.7865, + "step": 861 + }, + { + "epoch": 0.46796959826275786, + "grad_norm": 3.934765402497862, + "learning_rate": 1.978911076549114e-05, + "loss": 1.2993, + "step": 862 + }, + { + "epoch": 0.46851248642779586, + "grad_norm": 3.029198772099951, + "learning_rate": 1.978839171620335e-05, + "loss": 1.1737, + "step": 863 + }, + { + "epoch": 0.46905537459283386, + "grad_norm": 3.5637545602445946, + "learning_rate": 1.978767145627167e-05, + "loss": 1.1932, + "step": 864 + }, + { + "epoch": 0.46959826275787186, + "grad_norm": 3.131289093388952, + "learning_rate": 1.9786949985785187e-05, + "loss": 0.69, + "step": 865 + }, + { + "epoch": 0.47014115092290987, + "grad_norm": 3.522468470152659, + "learning_rate": 1.978622730483313e-05, + "loss": 1.0277, + "step": 866 + }, + { + "epoch": 0.47068403908794787, + "grad_norm": 4.046439598363532, + "learning_rate": 1.9785503413504883e-05, + "loss": 1.2917, + "step": 867 + }, + { + "epoch": 0.47122692725298587, + "grad_norm": 5.55259287782346, + "learning_rate": 1.978477831188998e-05, + "loss": 1.866, + "step": 868 + }, + { + "epoch": 0.4717698154180239, + "grad_norm": 5.874766516228883, + "learning_rate": 1.97840520000781e-05, + "loss": 1.6221, + "step": 869 + }, + { + "epoch": 0.4723127035830619, + "grad_norm": 5.530080877073305, + "learning_rate": 1.9783324478159074e-05, + "loss": 1.7617, + "step": 870 + }, + { + "epoch": 0.4728555917480999, + "grad_norm": 4.996268624326209, + "learning_rate": 1.9782595746222886e-05, + "loss": 1.5794, + "step": 871 + }, + { + "epoch": 0.4733984799131379, + "grad_norm": 3.934049315300115, + "learning_rate": 1.9781865804359663e-05, + "loss": 1.2603, + "step": 872 + }, + { + "epoch": 0.4739413680781759, + "grad_norm": 4.534039775579426, + "learning_rate": 1.978113465265969e-05, + "loss": 1.2766, + "step": 873 + }, + { + "epoch": 0.4744842562432139, + "grad_norm": 4.281791787646871, + "learning_rate": 1.9780402291213393e-05, + "loss": 1.6424, + "step": 874 + }, + { + "epoch": 0.4750271444082519, + "grad_norm": 4.443571747878234, + "learning_rate": 1.977966872011135e-05, + "loss": 1.514, + "step": 875 + }, + { + "epoch": 0.4755700325732899, + "grad_norm": 4.432653652742011, + "learning_rate": 1.9778933939444298e-05, + "loss": 1.2405, + "step": 876 + }, + { + "epoch": 0.4761129207383279, + "grad_norm": 3.408851551186198, + "learning_rate": 1.9778197949303107e-05, + "loss": 0.9597, + "step": 877 + }, + { + "epoch": 0.4766558089033659, + "grad_norm": 3.13258933410851, + "learning_rate": 1.9777460749778812e-05, + "loss": 0.8015, + "step": 878 + }, + { + "epoch": 0.4771986970684039, + "grad_norm": 3.535707436018844, + "learning_rate": 1.977672234096259e-05, + "loss": 1.3375, + "step": 879 + }, + { + "epoch": 0.4777415852334419, + "grad_norm": 5.210333639552661, + "learning_rate": 1.9775982722945764e-05, + "loss": 1.5039, + "step": 880 + }, + { + "epoch": 0.4782844733984799, + "grad_norm": 4.14115837125392, + "learning_rate": 1.9775241895819818e-05, + "loss": 1.0267, + "step": 881 + }, + { + "epoch": 0.4788273615635179, + "grad_norm": 4.390843049257208, + "learning_rate": 1.977449985967637e-05, + "loss": 1.6639, + "step": 882 + }, + { + "epoch": 0.4793702497285559, + "grad_norm": 3.5906251649150405, + "learning_rate": 1.9773756614607205e-05, + "loss": 1.2382, + "step": 883 + }, + { + "epoch": 0.4799131378935939, + "grad_norm": 4.78880370244169, + "learning_rate": 1.9773012160704242e-05, + "loss": 1.9436, + "step": 884 + }, + { + "epoch": 0.4804560260586319, + "grad_norm": 4.024679917575935, + "learning_rate": 1.977226649805956e-05, + "loss": 1.7126, + "step": 885 + }, + { + "epoch": 0.4809989142236699, + "grad_norm": 4.789782875875197, + "learning_rate": 1.9771519626765384e-05, + "loss": 1.4639, + "step": 886 + }, + { + "epoch": 0.4815418023887079, + "grad_norm": 4.188325588019175, + "learning_rate": 1.9770771546914088e-05, + "loss": 0.9698, + "step": 887 + }, + { + "epoch": 0.4820846905537459, + "grad_norm": 4.491436069698424, + "learning_rate": 1.9770022258598192e-05, + "loss": 1.5504, + "step": 888 + }, + { + "epoch": 0.4826275787187839, + "grad_norm": 4.889270563723842, + "learning_rate": 1.9769271761910373e-05, + "loss": 1.6052, + "step": 889 + }, + { + "epoch": 0.4831704668838219, + "grad_norm": 3.7615089990618493, + "learning_rate": 1.9768520056943454e-05, + "loss": 1.2445, + "step": 890 + }, + { + "epoch": 0.4837133550488599, + "grad_norm": 4.153654659347469, + "learning_rate": 1.9767767143790403e-05, + "loss": 1.4304, + "step": 891 + }, + { + "epoch": 0.48425624321389793, + "grad_norm": 4.040032964648755, + "learning_rate": 1.9767013022544346e-05, + "loss": 1.2906, + "step": 892 + }, + { + "epoch": 0.48479913137893593, + "grad_norm": 4.6559235863058275, + "learning_rate": 1.976625769329855e-05, + "loss": 1.5214, + "step": 893 + }, + { + "epoch": 0.48534201954397393, + "grad_norm": 4.224494142526306, + "learning_rate": 1.976550115614644e-05, + "loss": 1.2299, + "step": 894 + }, + { + "epoch": 0.48588490770901194, + "grad_norm": 5.225621473809677, + "learning_rate": 1.9764743411181585e-05, + "loss": 1.3489, + "step": 895 + }, + { + "epoch": 0.48642779587404994, + "grad_norm": 3.298440710545404, + "learning_rate": 1.97639844584977e-05, + "loss": 1.2107, + "step": 896 + }, + { + "epoch": 0.48697068403908794, + "grad_norm": 4.8373671778516485, + "learning_rate": 1.9763224298188652e-05, + "loss": 1.7748, + "step": 897 + }, + { + "epoch": 0.48751357220412594, + "grad_norm": 3.9332967347674224, + "learning_rate": 1.9762462930348465e-05, + "loss": 1.1036, + "step": 898 + }, + { + "epoch": 0.48805646036916395, + "grad_norm": 3.555513906222738, + "learning_rate": 1.9761700355071306e-05, + "loss": 1.5988, + "step": 899 + }, + { + "epoch": 0.48859934853420195, + "grad_norm": 4.115001448738943, + "learning_rate": 1.9760936572451487e-05, + "loss": 0.9789, + "step": 900 + }, + { + "epoch": 0.48914223669923995, + "grad_norm": 3.955269044937872, + "learning_rate": 1.9760171582583476e-05, + "loss": 1.3833, + "step": 901 + }, + { + "epoch": 0.48968512486427795, + "grad_norm": 2.718909991804504, + "learning_rate": 1.975940538556189e-05, + "loss": 1.1708, + "step": 902 + }, + { + "epoch": 0.49022801302931596, + "grad_norm": 4.234679090051759, + "learning_rate": 1.975863798148149e-05, + "loss": 1.3002, + "step": 903 + }, + { + "epoch": 0.49077090119435396, + "grad_norm": 4.162638017444828, + "learning_rate": 1.975786937043719e-05, + "loss": 1.246, + "step": 904 + }, + { + "epoch": 0.49131378935939196, + "grad_norm": 4.466027820294199, + "learning_rate": 1.975709955252406e-05, + "loss": 1.7345, + "step": 905 + }, + { + "epoch": 0.49185667752442996, + "grad_norm": 4.120456111509664, + "learning_rate": 1.97563285278373e-05, + "loss": 1.1339, + "step": 906 + }, + { + "epoch": 0.49239956568946797, + "grad_norm": 3.7932999801551346, + "learning_rate": 1.9755556296472286e-05, + "loss": 1.241, + "step": 907 + }, + { + "epoch": 0.49294245385450597, + "grad_norm": 3.3311865780622107, + "learning_rate": 1.9754782858524515e-05, + "loss": 1.2329, + "step": 908 + }, + { + "epoch": 0.49348534201954397, + "grad_norm": 4.21189840914184, + "learning_rate": 1.9754008214089654e-05, + "loss": 1.089, + "step": 909 + }, + { + "epoch": 0.494028230184582, + "grad_norm": 4.882048591403942, + "learning_rate": 1.9753232363263513e-05, + "loss": 1.3579, + "step": 910 + }, + { + "epoch": 0.49457111834962, + "grad_norm": 3.8495412115497833, + "learning_rate": 1.9752455306142052e-05, + "loss": 0.876, + "step": 911 + }, + { + "epoch": 0.495114006514658, + "grad_norm": 5.35378911928701, + "learning_rate": 1.975167704282137e-05, + "loss": 1.5025, + "step": 912 + }, + { + "epoch": 0.495656894679696, + "grad_norm": 3.4776107222093176, + "learning_rate": 1.9750897573397733e-05, + "loss": 1.1576, + "step": 913 + }, + { + "epoch": 0.496199782844734, + "grad_norm": 4.3680755226003525, + "learning_rate": 1.9750116897967545e-05, + "loss": 1.2519, + "step": 914 + }, + { + "epoch": 0.496742671009772, + "grad_norm": 4.154395835162648, + "learning_rate": 1.9749335016627355e-05, + "loss": 1.3397, + "step": 915 + }, + { + "epoch": 0.49728555917481, + "grad_norm": 5.465210053229008, + "learning_rate": 1.9748551929473873e-05, + "loss": 1.7555, + "step": 916 + }, + { + "epoch": 0.497828447339848, + "grad_norm": 4.391089608848125, + "learning_rate": 1.9747767636603953e-05, + "loss": 0.9685, + "step": 917 + }, + { + "epoch": 0.498371335504886, + "grad_norm": 4.215127655197256, + "learning_rate": 1.9746982138114597e-05, + "loss": 1.3889, + "step": 918 + }, + { + "epoch": 0.498914223669924, + "grad_norm": 4.559291789845839, + "learning_rate": 1.9746195434102956e-05, + "loss": 1.2411, + "step": 919 + }, + { + "epoch": 0.499457111834962, + "grad_norm": 4.226267506705079, + "learning_rate": 1.9745407524666327e-05, + "loss": 0.9517, + "step": 920 + }, + { + "epoch": 0.5, + "grad_norm": 4.289151336000736, + "learning_rate": 1.9744618409902164e-05, + "loss": 0.9123, + "step": 921 + }, + { + "epoch": 0.500542888165038, + "grad_norm": 5.356043714827742, + "learning_rate": 1.9743828089908067e-05, + "loss": 1.4039, + "step": 922 + }, + { + "epoch": 0.501085776330076, + "grad_norm": 4.867388642606108, + "learning_rate": 1.9743036564781785e-05, + "loss": 1.877, + "step": 923 + }, + { + "epoch": 0.501628664495114, + "grad_norm": 4.034273796102648, + "learning_rate": 1.9742243834621207e-05, + "loss": 1.6151, + "step": 924 + }, + { + "epoch": 0.502171552660152, + "grad_norm": 4.67131234305618, + "learning_rate": 1.9741449899524393e-05, + "loss": 1.6415, + "step": 925 + }, + { + "epoch": 0.50271444082519, + "grad_norm": 4.609211834325983, + "learning_rate": 1.9740654759589524e-05, + "loss": 1.2702, + "step": 926 + }, + { + "epoch": 0.503257328990228, + "grad_norm": 3.8693712873725064, + "learning_rate": 1.973985841491495e-05, + "loss": 1.0518, + "step": 927 + }, + { + "epoch": 0.503800217155266, + "grad_norm": 4.224951197105749, + "learning_rate": 1.9739060865599163e-05, + "loss": 1.1367, + "step": 928 + }, + { + "epoch": 0.504343105320304, + "grad_norm": 4.851656117293343, + "learning_rate": 1.973826211174081e-05, + "loss": 1.2753, + "step": 929 + }, + { + "epoch": 0.504885993485342, + "grad_norm": 4.255261482914816, + "learning_rate": 1.9737462153438675e-05, + "loss": 1.1723, + "step": 930 + }, + { + "epoch": 0.50542888165038, + "grad_norm": 3.8038614228711616, + "learning_rate": 1.9736660990791705e-05, + "loss": 1.1919, + "step": 931 + }, + { + "epoch": 0.505971769815418, + "grad_norm": 3.934227455395002, + "learning_rate": 1.9735858623898984e-05, + "loss": 1.1311, + "step": 932 + }, + { + "epoch": 0.506514657980456, + "grad_norm": 6.087081162565673, + "learning_rate": 1.9735055052859752e-05, + "loss": 1.4226, + "step": 933 + }, + { + "epoch": 0.507057546145494, + "grad_norm": 3.901680463400142, + "learning_rate": 1.9734250277773398e-05, + "loss": 0.8336, + "step": 934 + }, + { + "epoch": 0.507600434310532, + "grad_norm": 4.29804205692996, + "learning_rate": 1.9733444298739454e-05, + "loss": 1.5767, + "step": 935 + }, + { + "epoch": 0.50814332247557, + "grad_norm": 4.114842645062852, + "learning_rate": 1.9732637115857606e-05, + "loss": 1.3353, + "step": 936 + }, + { + "epoch": 0.508686210640608, + "grad_norm": 3.8525228530533706, + "learning_rate": 1.9731828729227686e-05, + "loss": 0.8208, + "step": 937 + }, + { + "epoch": 0.509229098805646, + "grad_norm": 4.592070280997817, + "learning_rate": 1.973101913894968e-05, + "loss": 1.3908, + "step": 938 + }, + { + "epoch": 0.509771986970684, + "grad_norm": 3.5842386277465628, + "learning_rate": 1.9730208345123718e-05, + "loss": 1.1996, + "step": 939 + }, + { + "epoch": 0.510314875135722, + "grad_norm": 6.053647281932304, + "learning_rate": 1.9729396347850082e-05, + "loss": 1.565, + "step": 940 + }, + { + "epoch": 0.51085776330076, + "grad_norm": 4.515664489508703, + "learning_rate": 1.9728583147229196e-05, + "loss": 1.4823, + "step": 941 + }, + { + "epoch": 0.511400651465798, + "grad_norm": 4.7752750595730475, + "learning_rate": 1.9727768743361644e-05, + "loss": 1.3867, + "step": 942 + }, + { + "epoch": 0.511943539630836, + "grad_norm": 3.983262894008015, + "learning_rate": 1.972695313634815e-05, + "loss": 1.1327, + "step": 943 + }, + { + "epoch": 0.512486427795874, + "grad_norm": 3.8510372454489326, + "learning_rate": 1.9726136326289586e-05, + "loss": 1.0233, + "step": 944 + }, + { + "epoch": 0.5130293159609121, + "grad_norm": 3.7881258264140634, + "learning_rate": 1.9725318313286985e-05, + "loss": 1.0239, + "step": 945 + }, + { + "epoch": 0.5135722041259501, + "grad_norm": 5.532602386318144, + "learning_rate": 1.9724499097441513e-05, + "loss": 1.8325, + "step": 946 + }, + { + "epoch": 0.5141150922909881, + "grad_norm": 5.978249809716352, + "learning_rate": 1.9723678678854493e-05, + "loss": 1.4899, + "step": 947 + }, + { + "epoch": 0.5146579804560261, + "grad_norm": 4.164217442125778, + "learning_rate": 1.9722857057627398e-05, + "loss": 1.5428, + "step": 948 + }, + { + "epoch": 0.5152008686210641, + "grad_norm": 3.315839086509451, + "learning_rate": 1.9722034233861848e-05, + "loss": 1.161, + "step": 949 + }, + { + "epoch": 0.5157437567861021, + "grad_norm": 4.362448119896709, + "learning_rate": 1.9721210207659608e-05, + "loss": 1.3402, + "step": 950 + }, + { + "epoch": 0.5162866449511401, + "grad_norm": 4.196652705893848, + "learning_rate": 1.9720384979122594e-05, + "loss": 1.4782, + "step": 951 + }, + { + "epoch": 0.5168295331161781, + "grad_norm": 4.093486911730782, + "learning_rate": 1.9719558548352876e-05, + "loss": 0.8508, + "step": 952 + }, + { + "epoch": 0.5173724212812161, + "grad_norm": 3.623412868803193, + "learning_rate": 1.9718730915452664e-05, + "loss": 1.4979, + "step": 953 + }, + { + "epoch": 0.5179153094462541, + "grad_norm": 3.4964911024452845, + "learning_rate": 1.9717902080524324e-05, + "loss": 1.0495, + "step": 954 + }, + { + "epoch": 0.5184581976112921, + "grad_norm": 4.021867363378223, + "learning_rate": 1.9717072043670367e-05, + "loss": 1.5284, + "step": 955 + }, + { + "epoch": 0.5190010857763301, + "grad_norm": 4.383918034859983, + "learning_rate": 1.9716240804993454e-05, + "loss": 1.3092, + "step": 956 + }, + { + "epoch": 0.5195439739413681, + "grad_norm": 4.61200677033352, + "learning_rate": 1.971540836459639e-05, + "loss": 1.8809, + "step": 957 + }, + { + "epoch": 0.5200868621064061, + "grad_norm": 4.490685725956694, + "learning_rate": 1.9714574722582142e-05, + "loss": 1.4004, + "step": 958 + }, + { + "epoch": 0.5206297502714441, + "grad_norm": 3.4814367857792736, + "learning_rate": 1.9713739879053802e-05, + "loss": 0.7918, + "step": 959 + }, + { + "epoch": 0.5211726384364821, + "grad_norm": 4.751597566480749, + "learning_rate": 1.9712903834114635e-05, + "loss": 1.6595, + "step": 960 + }, + { + "epoch": 0.5217155266015201, + "grad_norm": 4.535333205317335, + "learning_rate": 1.9712066587868042e-05, + "loss": 1.5676, + "step": 961 + }, + { + "epoch": 0.5222584147665581, + "grad_norm": 4.285001993552699, + "learning_rate": 1.9711228140417577e-05, + "loss": 1.0156, + "step": 962 + }, + { + "epoch": 0.5228013029315961, + "grad_norm": 4.61361421766516, + "learning_rate": 1.9710388491866934e-05, + "loss": 0.9552, + "step": 963 + }, + { + "epoch": 0.5233441910966341, + "grad_norm": 5.760482951437034, + "learning_rate": 1.9709547642319968e-05, + "loss": 2.2573, + "step": 964 + }, + { + "epoch": 0.5238870792616721, + "grad_norm": 4.38974180509626, + "learning_rate": 1.9708705591880674e-05, + "loss": 0.9939, + "step": 965 + }, + { + "epoch": 0.5244299674267101, + "grad_norm": 4.856793718321527, + "learning_rate": 1.97078623406532e-05, + "loss": 1.6161, + "step": 966 + }, + { + "epoch": 0.5249728555917481, + "grad_norm": 4.164767885626303, + "learning_rate": 1.9707017888741838e-05, + "loss": 1.4327, + "step": 967 + }, + { + "epoch": 0.5255157437567861, + "grad_norm": 5.163022804287371, + "learning_rate": 1.970617223625104e-05, + "loss": 1.2063, + "step": 968 + }, + { + "epoch": 0.5260586319218241, + "grad_norm": 4.826390195584734, + "learning_rate": 1.9705325383285384e-05, + "loss": 1.609, + "step": 969 + }, + { + "epoch": 0.5266015200868621, + "grad_norm": 6.3570394668292405, + "learning_rate": 1.9704477329949617e-05, + "loss": 1.5845, + "step": 970 + }, + { + "epoch": 0.5271444082519001, + "grad_norm": 5.176208926963624, + "learning_rate": 1.9703628076348628e-05, + "loss": 1.5792, + "step": 971 + }, + { + "epoch": 0.5276872964169381, + "grad_norm": 4.968725075311641, + "learning_rate": 1.9702777622587452e-05, + "loss": 1.774, + "step": 972 + }, + { + "epoch": 0.5282301845819761, + "grad_norm": 3.733580423042607, + "learning_rate": 1.9701925968771277e-05, + "loss": 0.8424, + "step": 973 + }, + { + "epoch": 0.5287730727470141, + "grad_norm": 4.330993358163065, + "learning_rate": 1.9701073115005437e-05, + "loss": 1.6198, + "step": 974 + }, + { + "epoch": 0.5293159609120521, + "grad_norm": 4.64032700965104, + "learning_rate": 1.9700219061395408e-05, + "loss": 1.2958, + "step": 975 + }, + { + "epoch": 0.5298588490770901, + "grad_norm": 4.08538215893539, + "learning_rate": 1.969936380804683e-05, + "loss": 1.5371, + "step": 976 + }, + { + "epoch": 0.5304017372421281, + "grad_norm": 3.7672102191405012, + "learning_rate": 1.9698507355065478e-05, + "loss": 1.4831, + "step": 977 + }, + { + "epoch": 0.5309446254071661, + "grad_norm": 4.237599031471779, + "learning_rate": 1.969764970255728e-05, + "loss": 1.4561, + "step": 978 + }, + { + "epoch": 0.5314875135722041, + "grad_norm": 4.296489590459447, + "learning_rate": 1.9696790850628308e-05, + "loss": 1.3813, + "step": 979 + }, + { + "epoch": 0.5320304017372421, + "grad_norm": 4.037479773042692, + "learning_rate": 1.9695930799384787e-05, + "loss": 1.103, + "step": 980 + }, + { + "epoch": 0.5325732899022801, + "grad_norm": 3.388491089813824, + "learning_rate": 1.9695069548933097e-05, + "loss": 1.0416, + "step": 981 + }, + { + "epoch": 0.5331161780673181, + "grad_norm": 3.6808593992848873, + "learning_rate": 1.969420709937975e-05, + "loss": 1.3158, + "step": 982 + }, + { + "epoch": 0.5336590662323561, + "grad_norm": 3.574434479685402, + "learning_rate": 1.969334345083142e-05, + "loss": 0.9542, + "step": 983 + }, + { + "epoch": 0.5342019543973942, + "grad_norm": 4.205886917864195, + "learning_rate": 1.9692478603394926e-05, + "loss": 1.1542, + "step": 984 + }, + { + "epoch": 0.5347448425624322, + "grad_norm": 3.89660120144171, + "learning_rate": 1.9691612557177225e-05, + "loss": 1.2374, + "step": 985 + }, + { + "epoch": 0.5352877307274702, + "grad_norm": 5.508260629607227, + "learning_rate": 1.9690745312285443e-05, + "loss": 2.0836, + "step": 986 + }, + { + "epoch": 0.5358306188925082, + "grad_norm": 5.482641917586033, + "learning_rate": 1.968987686882683e-05, + "loss": 1.5608, + "step": 987 + }, + { + "epoch": 0.5363735070575462, + "grad_norm": 5.4640893579957845, + "learning_rate": 1.9689007226908807e-05, + "loss": 1.3632, + "step": 988 + }, + { + "epoch": 0.5369163952225842, + "grad_norm": 4.163793092253212, + "learning_rate": 1.9688136386638926e-05, + "loss": 1.0509, + "step": 989 + }, + { + "epoch": 0.5374592833876222, + "grad_norm": 7.431185891169344, + "learning_rate": 1.96872643481249e-05, + "loss": 2.2147, + "step": 990 + }, + { + "epoch": 0.5380021715526602, + "grad_norm": 4.243102861037138, + "learning_rate": 1.9686391111474574e-05, + "loss": 0.9978, + "step": 991 + }, + { + "epoch": 0.5385450597176982, + "grad_norm": 4.250454886057412, + "learning_rate": 1.968551667679596e-05, + "loss": 1.5265, + "step": 992 + }, + { + "epoch": 0.5390879478827362, + "grad_norm": 5.989466882905416, + "learning_rate": 1.9684641044197207e-05, + "loss": 1.3344, + "step": 993 + }, + { + "epoch": 0.5396308360477742, + "grad_norm": 6.003524822087355, + "learning_rate": 1.9683764213786617e-05, + "loss": 2.0785, + "step": 994 + }, + { + "epoch": 0.5401737242128122, + "grad_norm": 4.535858561115482, + "learning_rate": 1.9682886185672633e-05, + "loss": 1.7654, + "step": 995 + }, + { + "epoch": 0.5407166123778502, + "grad_norm": 3.304287633610362, + "learning_rate": 1.9682006959963854e-05, + "loss": 0.9151, + "step": 996 + }, + { + "epoch": 0.5412595005428882, + "grad_norm": 3.6154194334954797, + "learning_rate": 1.9681126536769022e-05, + "loss": 1.1871, + "step": 997 + }, + { + "epoch": 0.5418023887079262, + "grad_norm": 5.596808777173772, + "learning_rate": 1.968024491619703e-05, + "loss": 1.1773, + "step": 998 + }, + { + "epoch": 0.5423452768729642, + "grad_norm": 4.486728043355676, + "learning_rate": 1.9679362098356923e-05, + "loss": 1.6495, + "step": 999 + }, + { + "epoch": 0.5428881650380022, + "grad_norm": 5.254525660355539, + "learning_rate": 1.9678478083357882e-05, + "loss": 1.7632, + "step": 1000 + }, + { + "epoch": 0.5434310532030402, + "grad_norm": 4.084222439189464, + "learning_rate": 1.9677592871309248e-05, + "loss": 1.4108, + "step": 1001 + }, + { + "epoch": 0.5439739413680782, + "grad_norm": 3.7647794378542248, + "learning_rate": 1.9676706462320504e-05, + "loss": 1.0893, + "step": 1002 + }, + { + "epoch": 0.5445168295331162, + "grad_norm": 3.658128786211824, + "learning_rate": 1.967581885650128e-05, + "loss": 1.0626, + "step": 1003 + }, + { + "epoch": 0.5450597176981542, + "grad_norm": 4.111525087755858, + "learning_rate": 1.967493005396136e-05, + "loss": 1.1915, + "step": 1004 + }, + { + "epoch": 0.5456026058631922, + "grad_norm": 4.130761699023899, + "learning_rate": 1.967404005481067e-05, + "loss": 1.1007, + "step": 1005 + }, + { + "epoch": 0.5461454940282302, + "grad_norm": 4.610164776763781, + "learning_rate": 1.9673148859159292e-05, + "loss": 1.3626, + "step": 1006 + }, + { + "epoch": 0.5466883821932682, + "grad_norm": 3.3516729554956686, + "learning_rate": 1.9672256467117445e-05, + "loss": 1.1138, + "step": 1007 + }, + { + "epoch": 0.5472312703583062, + "grad_norm": 5.185808250719794, + "learning_rate": 1.9671362878795502e-05, + "loss": 1.5639, + "step": 1008 + }, + { + "epoch": 0.5477741585233442, + "grad_norm": 3.9549787550673323, + "learning_rate": 1.9670468094303983e-05, + "loss": 1.0094, + "step": 1009 + }, + { + "epoch": 0.5483170466883822, + "grad_norm": 5.025417088271204, + "learning_rate": 1.966957211375356e-05, + "loss": 1.9741, + "step": 1010 + }, + { + "epoch": 0.5488599348534202, + "grad_norm": 5.1560673391175, + "learning_rate": 1.9668674937255044e-05, + "loss": 0.9653, + "step": 1011 + }, + { + "epoch": 0.5494028230184582, + "grad_norm": 4.843540796107876, + "learning_rate": 1.9667776564919404e-05, + "loss": 1.4745, + "step": 1012 + }, + { + "epoch": 0.5499457111834962, + "grad_norm": 3.8379573247832193, + "learning_rate": 1.966687699685775e-05, + "loss": 1.537, + "step": 1013 + }, + { + "epoch": 0.5504885993485342, + "grad_norm": 4.291804816024517, + "learning_rate": 1.9665976233181342e-05, + "loss": 1.6859, + "step": 1014 + }, + { + "epoch": 0.5510314875135722, + "grad_norm": 4.827630960232388, + "learning_rate": 1.966507427400159e-05, + "loss": 1.3205, + "step": 1015 + }, + { + "epoch": 0.5515743756786102, + "grad_norm": 4.251540690047392, + "learning_rate": 1.9664171119430044e-05, + "loss": 0.8213, + "step": 1016 + }, + { + "epoch": 0.5521172638436482, + "grad_norm": 3.7693342275615476, + "learning_rate": 1.9663266769578414e-05, + "loss": 1.0349, + "step": 1017 + }, + { + "epoch": 0.5526601520086862, + "grad_norm": 4.075472936638236, + "learning_rate": 1.966236122455855e-05, + "loss": 1.0853, + "step": 1018 + }, + { + "epoch": 0.5532030401737242, + "grad_norm": 4.771167333111926, + "learning_rate": 1.9661454484482448e-05, + "loss": 1.2135, + "step": 1019 + }, + { + "epoch": 0.5537459283387622, + "grad_norm": 3.834854303591261, + "learning_rate": 1.9660546549462262e-05, + "loss": 1.0972, + "step": 1020 + }, + { + "epoch": 0.5542888165038002, + "grad_norm": 4.712560042125338, + "learning_rate": 1.9659637419610278e-05, + "loss": 0.8779, + "step": 1021 + }, + { + "epoch": 0.5548317046688382, + "grad_norm": 3.8174143787104566, + "learning_rate": 1.9658727095038942e-05, + "loss": 0.7395, + "step": 1022 + }, + { + "epoch": 0.5553745928338762, + "grad_norm": 5.970646625648977, + "learning_rate": 1.965781557586085e-05, + "loss": 1.4233, + "step": 1023 + }, + { + "epoch": 0.5559174809989142, + "grad_norm": 3.872798187838392, + "learning_rate": 1.9656902862188732e-05, + "loss": 0.8147, + "step": 1024 + }, + { + "epoch": 0.5564603691639523, + "grad_norm": 5.632609234869673, + "learning_rate": 1.9655988954135473e-05, + "loss": 1.1581, + "step": 1025 + }, + { + "epoch": 0.5570032573289903, + "grad_norm": 4.141188191911578, + "learning_rate": 1.9655073851814117e-05, + "loss": 1.1707, + "step": 1026 + }, + { + "epoch": 0.5575461454940283, + "grad_norm": 5.7835323867879564, + "learning_rate": 1.9654157555337837e-05, + "loss": 1.2383, + "step": 1027 + }, + { + "epoch": 0.5580890336590663, + "grad_norm": 6.410603262158535, + "learning_rate": 1.9653240064819965e-05, + "loss": 2.3461, + "step": 1028 + }, + { + "epoch": 0.5586319218241043, + "grad_norm": 4.661944777700264, + "learning_rate": 1.9652321380373974e-05, + "loss": 1.1718, + "step": 1029 + }, + { + "epoch": 0.5591748099891423, + "grad_norm": 4.224334614804915, + "learning_rate": 1.9651401502113497e-05, + "loss": 1.0937, + "step": 1030 + }, + { + "epoch": 0.5597176981541803, + "grad_norm": 4.292561821167698, + "learning_rate": 1.9650480430152295e-05, + "loss": 0.8306, + "step": 1031 + }, + { + "epoch": 0.5602605863192183, + "grad_norm": 4.515070808084738, + "learning_rate": 1.9649558164604293e-05, + "loss": 1.4042, + "step": 1032 + }, + { + "epoch": 0.5608034744842563, + "grad_norm": 4.906478484057087, + "learning_rate": 1.964863470558356e-05, + "loss": 1.3396, + "step": 1033 + }, + { + "epoch": 0.5613463626492943, + "grad_norm": 3.9530548146664306, + "learning_rate": 1.9647710053204307e-05, + "loss": 0.917, + "step": 1034 + }, + { + "epoch": 0.5618892508143323, + "grad_norm": 4.640160859146004, + "learning_rate": 1.96467842075809e-05, + "loss": 1.6625, + "step": 1035 + }, + { + "epoch": 0.5624321389793703, + "grad_norm": 4.921887921875984, + "learning_rate": 1.964585716882785e-05, + "loss": 1.5133, + "step": 1036 + }, + { + "epoch": 0.5629750271444083, + "grad_norm": 4.667679101301949, + "learning_rate": 1.964492893705981e-05, + "loss": 0.9713, + "step": 1037 + }, + { + "epoch": 0.5635179153094463, + "grad_norm": 6.0457527064758825, + "learning_rate": 1.9643999512391586e-05, + "loss": 1.1101, + "step": 1038 + }, + { + "epoch": 0.5640608034744843, + "grad_norm": 5.188403972218186, + "learning_rate": 1.964306889493813e-05, + "loss": 1.8276, + "step": 1039 + }, + { + "epoch": 0.5646036916395223, + "grad_norm": 4.449370010668479, + "learning_rate": 1.9642137084814548e-05, + "loss": 1.0797, + "step": 1040 + }, + { + "epoch": 0.5651465798045603, + "grad_norm": 5.97660491495462, + "learning_rate": 1.9641204082136085e-05, + "loss": 1.913, + "step": 1041 + }, + { + "epoch": 0.5656894679695983, + "grad_norm": 5.611282387277076, + "learning_rate": 1.9640269887018135e-05, + "loss": 2.4593, + "step": 1042 + }, + { + "epoch": 0.5662323561346363, + "grad_norm": 3.6226559029227996, + "learning_rate": 1.9639334499576237e-05, + "loss": 0.7939, + "step": 1043 + }, + { + "epoch": 0.5667752442996743, + "grad_norm": 4.701569628633213, + "learning_rate": 1.963839791992609e-05, + "loss": 1.4092, + "step": 1044 + }, + { + "epoch": 0.5673181324647123, + "grad_norm": 5.185222096624812, + "learning_rate": 1.9637460148183525e-05, + "loss": 1.3851, + "step": 1045 + }, + { + "epoch": 0.5678610206297503, + "grad_norm": 4.094772946058271, + "learning_rate": 1.963652118446453e-05, + "loss": 1.1872, + "step": 1046 + }, + { + "epoch": 0.5684039087947883, + "grad_norm": 4.653748340189311, + "learning_rate": 1.9635581028885233e-05, + "loss": 1.1546, + "step": 1047 + }, + { + "epoch": 0.5689467969598263, + "grad_norm": 3.9426831069951844, + "learning_rate": 1.9634639681561924e-05, + "loss": 0.9324, + "step": 1048 + }, + { + "epoch": 0.5694896851248643, + "grad_norm": 4.949304407204856, + "learning_rate": 1.9633697142611017e-05, + "loss": 1.2372, + "step": 1049 + }, + { + "epoch": 0.5700325732899023, + "grad_norm": 5.364689744345368, + "learning_rate": 1.9632753412149096e-05, + "loss": 1.5104, + "step": 1050 + }, + { + "epoch": 0.5705754614549403, + "grad_norm": 5.307969223857254, + "learning_rate": 1.9631808490292884e-05, + "loss": 1.212, + "step": 1051 + }, + { + "epoch": 0.5711183496199783, + "grad_norm": 5.197310781163775, + "learning_rate": 1.963086237715924e-05, + "loss": 1.0175, + "step": 1052 + }, + { + "epoch": 0.5716612377850163, + "grad_norm": 4.095170540464123, + "learning_rate": 1.9629915072865194e-05, + "loss": 1.2711, + "step": 1053 + }, + { + "epoch": 0.5722041259500543, + "grad_norm": 4.816474093418903, + "learning_rate": 1.9628966577527902e-05, + "loss": 0.7274, + "step": 1054 + }, + { + "epoch": 0.5727470141150923, + "grad_norm": 4.000604469041167, + "learning_rate": 1.962801689126468e-05, + "loss": 1.2913, + "step": 1055 + }, + { + "epoch": 0.5732899022801303, + "grad_norm": 4.31668064977018, + "learning_rate": 1.962706601419298e-05, + "loss": 1.2598, + "step": 1056 + }, + { + "epoch": 0.5738327904451683, + "grad_norm": 5.033983318277113, + "learning_rate": 1.9626113946430414e-05, + "loss": 1.5149, + "step": 1057 + }, + { + "epoch": 0.5743756786102063, + "grad_norm": 4.51738649586986, + "learning_rate": 1.9625160688094733e-05, + "loss": 1.2389, + "step": 1058 + }, + { + "epoch": 0.5749185667752443, + "grad_norm": 3.6899120907028853, + "learning_rate": 1.9624206239303837e-05, + "loss": 1.1632, + "step": 1059 + }, + { + "epoch": 0.5754614549402823, + "grad_norm": 4.423539955787714, + "learning_rate": 1.9623250600175775e-05, + "loss": 0.9526, + "step": 1060 + }, + { + "epoch": 0.5760043431053203, + "grad_norm": 5.062986942342047, + "learning_rate": 1.962229377082874e-05, + "loss": 1.6768, + "step": 1061 + }, + { + "epoch": 0.5765472312703583, + "grad_norm": 5.044415439938055, + "learning_rate": 1.962133575138108e-05, + "loss": 1.3808, + "step": 1062 + }, + { + "epoch": 0.5770901194353963, + "grad_norm": 3.1405283495786587, + "learning_rate": 1.962037654195128e-05, + "loss": 0.7588, + "step": 1063 + }, + { + "epoch": 0.5776330076004343, + "grad_norm": 5.449873922495929, + "learning_rate": 1.9619416142657974e-05, + "loss": 1.1544, + "step": 1064 + }, + { + "epoch": 0.5781758957654723, + "grad_norm": 4.268527750000251, + "learning_rate": 1.961845455361995e-05, + "loss": 1.1941, + "step": 1065 + }, + { + "epoch": 0.5787187839305103, + "grad_norm": 4.6641878464575095, + "learning_rate": 1.9617491774956137e-05, + "loss": 1.7162, + "step": 1066 + }, + { + "epoch": 0.5792616720955484, + "grad_norm": 4.583539478633983, + "learning_rate": 1.9616527806785612e-05, + "loss": 1.1415, + "step": 1067 + }, + { + "epoch": 0.5798045602605864, + "grad_norm": 4.325093393868848, + "learning_rate": 1.96155626492276e-05, + "loss": 1.3284, + "step": 1068 + }, + { + "epoch": 0.5803474484256244, + "grad_norm": 4.180109104666535, + "learning_rate": 1.9614596302401478e-05, + "loss": 1.136, + "step": 1069 + }, + { + "epoch": 0.5808903365906624, + "grad_norm": 4.108700840520527, + "learning_rate": 1.9613628766426762e-05, + "loss": 1.6142, + "step": 1070 + }, + { + "epoch": 0.5814332247557004, + "grad_norm": 5.1931558422083555, + "learning_rate": 1.9612660041423116e-05, + "loss": 0.8479, + "step": 1071 + }, + { + "epoch": 0.5819761129207384, + "grad_norm": 6.008601497593343, + "learning_rate": 1.9611690127510358e-05, + "loss": 1.4743, + "step": 1072 + }, + { + "epoch": 0.5825190010857764, + "grad_norm": 4.531622764164067, + "learning_rate": 1.9610719024808444e-05, + "loss": 1.5755, + "step": 1073 + }, + { + "epoch": 0.5830618892508144, + "grad_norm": 6.18647874774261, + "learning_rate": 1.9609746733437486e-05, + "loss": 1.7369, + "step": 1074 + }, + { + "epoch": 0.5836047774158524, + "grad_norm": 5.5880249590216025, + "learning_rate": 1.9608773253517738e-05, + "loss": 1.4402, + "step": 1075 + }, + { + "epoch": 0.5841476655808904, + "grad_norm": 4.798692544278573, + "learning_rate": 1.9607798585169595e-05, + "loss": 1.2257, + "step": 1076 + }, + { + "epoch": 0.5846905537459284, + "grad_norm": 5.635786862796798, + "learning_rate": 1.960682272851361e-05, + "loss": 1.183, + "step": 1077 + }, + { + "epoch": 0.5852334419109664, + "grad_norm": 3.3979344506068037, + "learning_rate": 1.9605845683670484e-05, + "loss": 0.6463, + "step": 1078 + }, + { + "epoch": 0.5857763300760044, + "grad_norm": 4.5152680576801405, + "learning_rate": 1.960486745076105e-05, + "loss": 0.8476, + "step": 1079 + }, + { + "epoch": 0.5863192182410424, + "grad_norm": 4.4606140059612835, + "learning_rate": 1.9603888029906305e-05, + "loss": 1.185, + "step": 1080 + }, + { + "epoch": 0.5868621064060804, + "grad_norm": 4.229219175605004, + "learning_rate": 1.960290742122738e-05, + "loss": 1.3672, + "step": 1081 + }, + { + "epoch": 0.5874049945711184, + "grad_norm": 4.964172517040214, + "learning_rate": 1.960192562484556e-05, + "loss": 1.0561, + "step": 1082 + }, + { + "epoch": 0.5879478827361564, + "grad_norm": 4.8348336861653785, + "learning_rate": 1.960094264088228e-05, + "loss": 1.1005, + "step": 1083 + }, + { + "epoch": 0.5884907709011944, + "grad_norm": 4.3153540955607514, + "learning_rate": 1.959995846945911e-05, + "loss": 1.6513, + "step": 1084 + }, + { + "epoch": 0.5890336590662324, + "grad_norm": 4.733352525701168, + "learning_rate": 1.9598973110697773e-05, + "loss": 1.6162, + "step": 1085 + }, + { + "epoch": 0.5895765472312704, + "grad_norm": 5.402815147148315, + "learning_rate": 1.959798656472015e-05, + "loss": 1.4178, + "step": 1086 + }, + { + "epoch": 0.5901194353963084, + "grad_norm": 3.8276911935691147, + "learning_rate": 1.9596998831648247e-05, + "loss": 1.1554, + "step": 1087 + }, + { + "epoch": 0.5906623235613464, + "grad_norm": 4.499779496087119, + "learning_rate": 1.9596009911604232e-05, + "loss": 0.9786, + "step": 1088 + }, + { + "epoch": 0.5912052117263844, + "grad_norm": 4.3705352636049595, + "learning_rate": 1.959501980471042e-05, + "loss": 1.0827, + "step": 1089 + }, + { + "epoch": 0.5917480998914224, + "grad_norm": 4.9770508472387025, + "learning_rate": 1.9594028511089264e-05, + "loss": 1.2512, + "step": 1090 + }, + { + "epoch": 0.5922909880564604, + "grad_norm": 4.23068807802663, + "learning_rate": 1.9593036030863376e-05, + "loss": 0.7528, + "step": 1091 + }, + { + "epoch": 0.5928338762214984, + "grad_norm": 4.321323716192747, + "learning_rate": 1.9592042364155496e-05, + "loss": 1.0773, + "step": 1092 + }, + { + "epoch": 0.5933767643865364, + "grad_norm": 3.380036112638087, + "learning_rate": 1.9591047511088535e-05, + "loss": 0.7238, + "step": 1093 + }, + { + "epoch": 0.5939196525515744, + "grad_norm": 3.9722637658815447, + "learning_rate": 1.959005147178553e-05, + "loss": 0.8132, + "step": 1094 + }, + { + "epoch": 0.5944625407166124, + "grad_norm": 4.354488865716506, + "learning_rate": 1.9589054246369673e-05, + "loss": 1.1214, + "step": 1095 + }, + { + "epoch": 0.5950054288816504, + "grad_norm": 4.337355429292092, + "learning_rate": 1.9588055834964307e-05, + "loss": 0.9039, + "step": 1096 + }, + { + "epoch": 0.5955483170466884, + "grad_norm": 5.864047590425705, + "learning_rate": 1.9587056237692912e-05, + "loss": 1.6648, + "step": 1097 + }, + { + "epoch": 0.5960912052117264, + "grad_norm": 4.904173131706714, + "learning_rate": 1.9586055454679123e-05, + "loss": 1.1465, + "step": 1098 + }, + { + "epoch": 0.5966340933767644, + "grad_norm": 4.411877930208604, + "learning_rate": 1.9585053486046717e-05, + "loss": 1.4579, + "step": 1099 + }, + { + "epoch": 0.5971769815418024, + "grad_norm": 4.985171448881133, + "learning_rate": 1.9584050331919616e-05, + "loss": 1.5497, + "step": 1100 + }, + { + "epoch": 0.5977198697068404, + "grad_norm": 4.950532301400581, + "learning_rate": 1.9583045992421902e-05, + "loss": 1.3811, + "step": 1101 + }, + { + "epoch": 0.5982627578718784, + "grad_norm": 4.154119544907876, + "learning_rate": 1.9582040467677782e-05, + "loss": 1.186, + "step": 1102 + }, + { + "epoch": 0.5988056460369164, + "grad_norm": 4.069018604381756, + "learning_rate": 1.9581033757811628e-05, + "loss": 0.9312, + "step": 1103 + }, + { + "epoch": 0.5993485342019544, + "grad_norm": 5.33070083129983, + "learning_rate": 1.9580025862947948e-05, + "loss": 1.4274, + "step": 1104 + }, + { + "epoch": 0.5998914223669924, + "grad_norm": 4.855606632567906, + "learning_rate": 1.95790167832114e-05, + "loss": 1.0875, + "step": 1105 + }, + { + "epoch": 0.6004343105320304, + "grad_norm": 5.30182267546013, + "learning_rate": 1.957800651872679e-05, + "loss": 1.6501, + "step": 1106 + }, + { + "epoch": 0.6009771986970684, + "grad_norm": 4.433903391255917, + "learning_rate": 1.957699506961907e-05, + "loss": 1.3897, + "step": 1107 + }, + { + "epoch": 0.6015200868621065, + "grad_norm": 4.687653782723915, + "learning_rate": 1.9575982436013335e-05, + "loss": 0.8484, + "step": 1108 + }, + { + "epoch": 0.6020629750271445, + "grad_norm": 5.650426140105992, + "learning_rate": 1.957496861803483e-05, + "loss": 1.923, + "step": 1109 + }, + { + "epoch": 0.6026058631921825, + "grad_norm": 4.024650772512973, + "learning_rate": 1.957395361580895e-05, + "loss": 1.2713, + "step": 1110 + }, + { + "epoch": 0.6031487513572205, + "grad_norm": 4.343921133786845, + "learning_rate": 1.9572937429461223e-05, + "loss": 1.3152, + "step": 1111 + }, + { + "epoch": 0.6036916395222585, + "grad_norm": 4.1467176317930265, + "learning_rate": 1.957192005911734e-05, + "loss": 1.3325, + "step": 1112 + }, + { + "epoch": 0.6042345276872965, + "grad_norm": 4.305465221711973, + "learning_rate": 1.9570901504903128e-05, + "loss": 0.9237, + "step": 1113 + }, + { + "epoch": 0.6047774158523345, + "grad_norm": 4.101045495900469, + "learning_rate": 1.9569881766944564e-05, + "loss": 1.3932, + "step": 1114 + }, + { + "epoch": 0.6053203040173725, + "grad_norm": 4.0920748344339835, + "learning_rate": 1.956886084536777e-05, + "loss": 1.0067, + "step": 1115 + }, + { + "epoch": 0.6058631921824105, + "grad_norm": 4.844431281019285, + "learning_rate": 1.956783874029902e-05, + "loss": 1.2726, + "step": 1116 + }, + { + "epoch": 0.6064060803474485, + "grad_norm": 4.329792546128024, + "learning_rate": 1.9566815451864723e-05, + "loss": 1.1144, + "step": 1117 + }, + { + "epoch": 0.6069489685124865, + "grad_norm": 5.430322578642178, + "learning_rate": 1.9565790980191447e-05, + "loss": 1.3307, + "step": 1118 + }, + { + "epoch": 0.6074918566775245, + "grad_norm": 3.672369600344093, + "learning_rate": 1.9564765325405895e-05, + "loss": 0.8514, + "step": 1119 + }, + { + "epoch": 0.6080347448425625, + "grad_norm": 3.6433282223345334, + "learning_rate": 1.9563738487634924e-05, + "loss": 1.1972, + "step": 1120 + }, + { + "epoch": 0.6085776330076005, + "grad_norm": 4.23249445053715, + "learning_rate": 1.956271046700553e-05, + "loss": 1.0638, + "step": 1121 + }, + { + "epoch": 0.6091205211726385, + "grad_norm": 3.9162807435580134, + "learning_rate": 1.956168126364487e-05, + "loss": 1.121, + "step": 1122 + }, + { + "epoch": 0.6096634093376765, + "grad_norm": 3.3651337591577812, + "learning_rate": 1.956065087768023e-05, + "loss": 0.9604, + "step": 1123 + }, + { + "epoch": 0.6102062975027145, + "grad_norm": 5.5905307606778685, + "learning_rate": 1.955961930923905e-05, + "loss": 1.4509, + "step": 1124 + }, + { + "epoch": 0.6107491856677525, + "grad_norm": 3.6120233107886404, + "learning_rate": 1.955858655844892e-05, + "loss": 1.0106, + "step": 1125 + }, + { + "epoch": 0.6112920738327905, + "grad_norm": 3.6242967727445468, + "learning_rate": 1.9557552625437574e-05, + "loss": 1.0768, + "step": 1126 + }, + { + "epoch": 0.6118349619978285, + "grad_norm": 4.856763858492221, + "learning_rate": 1.9556517510332883e-05, + "loss": 1.6372, + "step": 1127 + }, + { + "epoch": 0.6123778501628665, + "grad_norm": 4.648427809898363, + "learning_rate": 1.9555481213262873e-05, + "loss": 1.1622, + "step": 1128 + }, + { + "epoch": 0.6129207383279045, + "grad_norm": 4.228089396781667, + "learning_rate": 1.9554443734355723e-05, + "loss": 1.067, + "step": 1129 + }, + { + "epoch": 0.6134636264929425, + "grad_norm": 4.069786144676624, + "learning_rate": 1.9553405073739743e-05, + "loss": 1.1427, + "step": 1130 + }, + { + "epoch": 0.6140065146579805, + "grad_norm": 3.112182391585594, + "learning_rate": 1.9552365231543395e-05, + "loss": 0.7862, + "step": 1131 + }, + { + "epoch": 0.6145494028230185, + "grad_norm": 4.783518734299315, + "learning_rate": 1.955132420789529e-05, + "loss": 1.0366, + "step": 1132 + }, + { + "epoch": 0.6150922909880565, + "grad_norm": 4.905128048667836, + "learning_rate": 1.9550282002924187e-05, + "loss": 1.2566, + "step": 1133 + }, + { + "epoch": 0.6156351791530945, + "grad_norm": 4.643678450407587, + "learning_rate": 1.9549238616758987e-05, + "loss": 1.2749, + "step": 1134 + }, + { + "epoch": 0.6161780673181325, + "grad_norm": 4.18363664236292, + "learning_rate": 1.954819404952873e-05, + "loss": 1.1523, + "step": 1135 + }, + { + "epoch": 0.6167209554831705, + "grad_norm": 3.9776065520060335, + "learning_rate": 1.9547148301362623e-05, + "loss": 0.8672, + "step": 1136 + }, + { + "epoch": 0.6172638436482085, + "grad_norm": 5.417170242762362, + "learning_rate": 1.9546101372389994e-05, + "loss": 1.6434, + "step": 1137 + }, + { + "epoch": 0.6178067318132465, + "grad_norm": 3.3325133642660068, + "learning_rate": 1.9545053262740335e-05, + "loss": 1.1447, + "step": 1138 + }, + { + "epoch": 0.6183496199782845, + "grad_norm": 5.4727757064949545, + "learning_rate": 1.9544003972543273e-05, + "loss": 1.5013, + "step": 1139 + }, + { + "epoch": 0.6188925081433225, + "grad_norm": 4.619714236951128, + "learning_rate": 1.954295350192859e-05, + "loss": 0.9758, + "step": 1140 + }, + { + "epoch": 0.6194353963083605, + "grad_norm": 5.189787937143953, + "learning_rate": 1.954190185102621e-05, + "loss": 1.2771, + "step": 1141 + }, + { + "epoch": 0.6199782844733985, + "grad_norm": 6.571114312185869, + "learning_rate": 1.9540849019966198e-05, + "loss": 1.2805, + "step": 1142 + }, + { + "epoch": 0.6205211726384365, + "grad_norm": 5.6343515388196455, + "learning_rate": 1.9539795008878774e-05, + "loss": 1.4061, + "step": 1143 + }, + { + "epoch": 0.6210640608034745, + "grad_norm": 5.582702517092888, + "learning_rate": 1.9538739817894302e-05, + "loss": 1.613, + "step": 1144 + }, + { + "epoch": 0.6216069489685125, + "grad_norm": 6.194616996988272, + "learning_rate": 1.9537683447143287e-05, + "loss": 1.114, + "step": 1145 + }, + { + "epoch": 0.6221498371335505, + "grad_norm": 6.0395907363668515, + "learning_rate": 1.9536625896756377e-05, + "loss": 1.5233, + "step": 1146 + }, + { + "epoch": 0.6226927252985885, + "grad_norm": 4.672511820840566, + "learning_rate": 1.9535567166864382e-05, + "loss": 1.1491, + "step": 1147 + }, + { + "epoch": 0.6232356134636265, + "grad_norm": 4.954133576744497, + "learning_rate": 1.9534507257598244e-05, + "loss": 1.1974, + "step": 1148 + }, + { + "epoch": 0.6237785016286646, + "grad_norm": 4.447205012745297, + "learning_rate": 1.953344616908905e-05, + "loss": 1.019, + "step": 1149 + }, + { + "epoch": 0.6243213897937026, + "grad_norm": 4.284087771325503, + "learning_rate": 1.9532383901468038e-05, + "loss": 1.0814, + "step": 1150 + }, + { + "epoch": 0.6248642779587406, + "grad_norm": 3.576067836847051, + "learning_rate": 1.9531320454866595e-05, + "loss": 1.1025, + "step": 1151 + }, + { + "epoch": 0.6254071661237784, + "grad_norm": 3.6831292979487777, + "learning_rate": 1.9530255829416246e-05, + "loss": 0.967, + "step": 1152 + }, + { + "epoch": 0.6259500542888164, + "grad_norm": 3.842496728015195, + "learning_rate": 1.952919002524867e-05, + "loss": 1.0903, + "step": 1153 + }, + { + "epoch": 0.6264929424538545, + "grad_norm": 4.077633480381933, + "learning_rate": 1.952812304249568e-05, + "loss": 0.9227, + "step": 1154 + }, + { + "epoch": 0.6270358306188925, + "grad_norm": 4.157004955121598, + "learning_rate": 1.952705488128925e-05, + "loss": 0.9427, + "step": 1155 + }, + { + "epoch": 0.6275787187839305, + "grad_norm": 5.267614104282471, + "learning_rate": 1.952598554176149e-05, + "loss": 1.9875, + "step": 1156 + }, + { + "epoch": 0.6281216069489685, + "grad_norm": 3.8938350314376735, + "learning_rate": 1.952491502404465e-05, + "loss": 1.1662, + "step": 1157 + }, + { + "epoch": 0.6286644951140065, + "grad_norm": 4.851409026807682, + "learning_rate": 1.9523843328271144e-05, + "loss": 1.9559, + "step": 1158 + }, + { + "epoch": 0.6292073832790445, + "grad_norm": 4.4104613077358845, + "learning_rate": 1.9522770454573513e-05, + "loss": 0.8011, + "step": 1159 + }, + { + "epoch": 0.6297502714440825, + "grad_norm": 3.6883335641947856, + "learning_rate": 1.952169640308446e-05, + "loss": 1.1838, + "step": 1160 + }, + { + "epoch": 0.6302931596091205, + "grad_norm": 3.728203039459734, + "learning_rate": 1.9520621173936818e-05, + "loss": 1.1423, + "step": 1161 + }, + { + "epoch": 0.6308360477741585, + "grad_norm": 3.6272612276564296, + "learning_rate": 1.9519544767263574e-05, + "loss": 0.8376, + "step": 1162 + }, + { + "epoch": 0.6313789359391965, + "grad_norm": 3.928276333701112, + "learning_rate": 1.951846718319786e-05, + "loss": 1.0749, + "step": 1163 + }, + { + "epoch": 0.6319218241042345, + "grad_norm": 5.941949883397345, + "learning_rate": 1.951738842187296e-05, + "loss": 1.8535, + "step": 1164 + }, + { + "epoch": 0.6324647122692725, + "grad_norm": 3.4741200080360644, + "learning_rate": 1.951630848342229e-05, + "loss": 0.7537, + "step": 1165 + }, + { + "epoch": 0.6330076004343105, + "grad_norm": 5.392480994365815, + "learning_rate": 1.9515227367979416e-05, + "loss": 0.8776, + "step": 1166 + }, + { + "epoch": 0.6335504885993485, + "grad_norm": 5.203169892005258, + "learning_rate": 1.951414507567806e-05, + "loss": 1.2678, + "step": 1167 + }, + { + "epoch": 0.6340933767643865, + "grad_norm": 7.499226723318992, + "learning_rate": 1.9513061606652076e-05, + "loss": 1.3295, + "step": 1168 + }, + { + "epoch": 0.6346362649294245, + "grad_norm": 4.355172848230307, + "learning_rate": 1.9511976961035474e-05, + "loss": 0.9188, + "step": 1169 + }, + { + "epoch": 0.6351791530944625, + "grad_norm": 4.983220943925928, + "learning_rate": 1.9510891138962398e-05, + "loss": 1.1951, + "step": 1170 + }, + { + "epoch": 0.6357220412595005, + "grad_norm": 5.3098934312394395, + "learning_rate": 1.950980414056715e-05, + "loss": 0.855, + "step": 1171 + }, + { + "epoch": 0.6362649294245385, + "grad_norm": 5.125683116898323, + "learning_rate": 1.950871596598417e-05, + "loss": 1.1371, + "step": 1172 + }, + { + "epoch": 0.6368078175895765, + "grad_norm": 4.3010814264850215, + "learning_rate": 1.950762661534804e-05, + "loss": 1.7401, + "step": 1173 + }, + { + "epoch": 0.6373507057546145, + "grad_norm": 5.860625964437982, + "learning_rate": 1.95065360887935e-05, + "loss": 1.3557, + "step": 1174 + }, + { + "epoch": 0.6378935939196525, + "grad_norm": 6.3356570754666155, + "learning_rate": 1.9505444386455426e-05, + "loss": 1.2281, + "step": 1175 + }, + { + "epoch": 0.6384364820846905, + "grad_norm": 4.788366086551355, + "learning_rate": 1.9504351508468842e-05, + "loss": 1.3458, + "step": 1176 + }, + { + "epoch": 0.6389793702497285, + "grad_norm": 4.5553889359557065, + "learning_rate": 1.9503257454968914e-05, + "loss": 0.803, + "step": 1177 + }, + { + "epoch": 0.6395222584147665, + "grad_norm": 4.9063187966995345, + "learning_rate": 1.950216222609096e-05, + "loss": 0.8306, + "step": 1178 + }, + { + "epoch": 0.6400651465798045, + "grad_norm": 4.775138114897886, + "learning_rate": 1.9501065821970435e-05, + "loss": 0.9783, + "step": 1179 + }, + { + "epoch": 0.6406080347448425, + "grad_norm": 4.53672284240203, + "learning_rate": 1.9499968242742948e-05, + "loss": 1.3076, + "step": 1180 + }, + { + "epoch": 0.6411509229098805, + "grad_norm": 4.539860050199487, + "learning_rate": 1.949886948854425e-05, + "loss": 1.4212, + "step": 1181 + }, + { + "epoch": 0.6416938110749185, + "grad_norm": 4.251462432005476, + "learning_rate": 1.9497769559510232e-05, + "loss": 1.0274, + "step": 1182 + }, + { + "epoch": 0.6422366992399565, + "grad_norm": 4.49791289564495, + "learning_rate": 1.9496668455776938e-05, + "loss": 1.7531, + "step": 1183 + }, + { + "epoch": 0.6427795874049945, + "grad_norm": 4.996740483321485, + "learning_rate": 1.9495566177480555e-05, + "loss": 1.201, + "step": 1184 + }, + { + "epoch": 0.6433224755700325, + "grad_norm": 5.347352236818569, + "learning_rate": 1.9494462724757413e-05, + "loss": 1.5723, + "step": 1185 + }, + { + "epoch": 0.6438653637350705, + "grad_norm": 4.2242558589374015, + "learning_rate": 1.9493358097743988e-05, + "loss": 1.8014, + "step": 1186 + }, + { + "epoch": 0.6444082519001085, + "grad_norm": 5.042556201755935, + "learning_rate": 1.9492252296576906e-05, + "loss": 1.5466, + "step": 1187 + }, + { + "epoch": 0.6449511400651465, + "grad_norm": 4.284314238035697, + "learning_rate": 1.949114532139293e-05, + "loss": 0.9515, + "step": 1188 + }, + { + "epoch": 0.6454940282301845, + "grad_norm": 4.699826119048575, + "learning_rate": 1.9490037172328974e-05, + "loss": 1.1677, + "step": 1189 + }, + { + "epoch": 0.6460369163952225, + "grad_norm": 6.035247135554916, + "learning_rate": 1.9488927849522095e-05, + "loss": 1.957, + "step": 1190 + }, + { + "epoch": 0.6465798045602605, + "grad_norm": 4.042004235065528, + "learning_rate": 1.94878173531095e-05, + "loss": 1.2182, + "step": 1191 + }, + { + "epoch": 0.6471226927252985, + "grad_norm": 4.114859597350168, + "learning_rate": 1.948670568322853e-05, + "loss": 1.1787, + "step": 1192 + }, + { + "epoch": 0.6476655808903365, + "grad_norm": 4.380177461001504, + "learning_rate": 1.9485592840016682e-05, + "loss": 1.3912, + "step": 1193 + }, + { + "epoch": 0.6482084690553745, + "grad_norm": 5.160710982722015, + "learning_rate": 1.948447882361159e-05, + "loss": 1.2642, + "step": 1194 + }, + { + "epoch": 0.6487513572204126, + "grad_norm": 3.950160510485583, + "learning_rate": 1.9483363634151046e-05, + "loss": 1.1246, + "step": 1195 + }, + { + "epoch": 0.6492942453854506, + "grad_norm": 4.587449618376974, + "learning_rate": 1.9482247271772974e-05, + "loss": 1.3275, + "step": 1196 + }, + { + "epoch": 0.6498371335504886, + "grad_norm": 4.919831513668322, + "learning_rate": 1.9481129736615445e-05, + "loss": 1.4007, + "step": 1197 + }, + { + "epoch": 0.6503800217155266, + "grad_norm": 4.298954423192263, + "learning_rate": 1.948001102881668e-05, + "loss": 0.8076, + "step": 1198 + }, + { + "epoch": 0.6509229098805646, + "grad_norm": 5.525772014179547, + "learning_rate": 1.9478891148515043e-05, + "loss": 1.4558, + "step": 1199 + }, + { + "epoch": 0.6514657980456026, + "grad_norm": 5.473549669240299, + "learning_rate": 1.947777009584904e-05, + "loss": 1.0976, + "step": 1200 + }, + { + "epoch": 0.6520086862106406, + "grad_norm": 5.360010284326786, + "learning_rate": 1.947664787095733e-05, + "loss": 1.0462, + "step": 1201 + }, + { + "epoch": 0.6525515743756786, + "grad_norm": 4.696691081444868, + "learning_rate": 1.9475524473978705e-05, + "loss": 0.933, + "step": 1202 + }, + { + "epoch": 0.6530944625407166, + "grad_norm": 6.0503072087093965, + "learning_rate": 1.947439990505211e-05, + "loss": 0.9628, + "step": 1203 + }, + { + "epoch": 0.6536373507057546, + "grad_norm": 5.476272854511399, + "learning_rate": 1.9473274164316637e-05, + "loss": 0.9716, + "step": 1204 + }, + { + "epoch": 0.6541802388707926, + "grad_norm": 3.910448715873912, + "learning_rate": 1.9472147251911517e-05, + "loss": 1.0732, + "step": 1205 + }, + { + "epoch": 0.6547231270358306, + "grad_norm": 4.596114022138067, + "learning_rate": 1.9471019167976126e-05, + "loss": 0.9399, + "step": 1206 + }, + { + "epoch": 0.6552660152008686, + "grad_norm": 6.764840749864346, + "learning_rate": 1.946988991264999e-05, + "loss": 2.0529, + "step": 1207 + }, + { + "epoch": 0.6558089033659066, + "grad_norm": 5.217328230116913, + "learning_rate": 1.9468759486072778e-05, + "loss": 1.0477, + "step": 1208 + }, + { + "epoch": 0.6563517915309446, + "grad_norm": 4.03429219354135, + "learning_rate": 1.9467627888384303e-05, + "loss": 1.1103, + "step": 1209 + }, + { + "epoch": 0.6568946796959826, + "grad_norm": 4.301388894222403, + "learning_rate": 1.946649511972452e-05, + "loss": 0.8515, + "step": 1210 + }, + { + "epoch": 0.6574375678610206, + "grad_norm": 5.285446092583105, + "learning_rate": 1.9465361180233536e-05, + "loss": 1.2881, + "step": 1211 + }, + { + "epoch": 0.6579804560260586, + "grad_norm": 5.879897026479874, + "learning_rate": 1.9464226070051593e-05, + "loss": 0.9362, + "step": 1212 + }, + { + "epoch": 0.6585233441910966, + "grad_norm": 5.335257600571856, + "learning_rate": 1.9463089789319083e-05, + "loss": 1.4372, + "step": 1213 + }, + { + "epoch": 0.6590662323561346, + "grad_norm": 5.272070125431448, + "learning_rate": 1.9461952338176552e-05, + "loss": 1.2063, + "step": 1214 + }, + { + "epoch": 0.6596091205211726, + "grad_norm": 5.188765715525815, + "learning_rate": 1.946081371676467e-05, + "loss": 0.8912, + "step": 1215 + }, + { + "epoch": 0.6601520086862106, + "grad_norm": 4.747539421286818, + "learning_rate": 1.9459673925224275e-05, + "loss": 1.0381, + "step": 1216 + }, + { + "epoch": 0.6606948968512486, + "grad_norm": 4.486254331013965, + "learning_rate": 1.945853296369633e-05, + "loss": 1.2584, + "step": 1217 + }, + { + "epoch": 0.6612377850162866, + "grad_norm": 4.730012415000151, + "learning_rate": 1.945739083232195e-05, + "loss": 1.1061, + "step": 1218 + }, + { + "epoch": 0.6617806731813246, + "grad_norm": 5.713566521874704, + "learning_rate": 1.9456247531242405e-05, + "loss": 1.4628, + "step": 1219 + }, + { + "epoch": 0.6623235613463626, + "grad_norm": 6.08134803588137, + "learning_rate": 1.9455103060599093e-05, + "loss": 1.6743, + "step": 1220 + }, + { + "epoch": 0.6628664495114006, + "grad_norm": 4.778151760566848, + "learning_rate": 1.9453957420533562e-05, + "loss": 0.8454, + "step": 1221 + }, + { + "epoch": 0.6634093376764386, + "grad_norm": 5.007171333863363, + "learning_rate": 1.945281061118751e-05, + "loss": 1.079, + "step": 1222 + }, + { + "epoch": 0.6639522258414766, + "grad_norm": 4.034543591270571, + "learning_rate": 1.945166263270278e-05, + "loss": 0.8378, + "step": 1223 + }, + { + "epoch": 0.6644951140065146, + "grad_norm": 4.191899722379318, + "learning_rate": 1.9450513485221352e-05, + "loss": 1.3602, + "step": 1224 + }, + { + "epoch": 0.6650380021715526, + "grad_norm": 6.349672663190246, + "learning_rate": 1.944936316888535e-05, + "loss": 1.861, + "step": 1225 + }, + { + "epoch": 0.6655808903365906, + "grad_norm": 5.991894404952759, + "learning_rate": 1.9448211683837055e-05, + "loss": 1.0931, + "step": 1226 + }, + { + "epoch": 0.6661237785016286, + "grad_norm": 4.405247018509644, + "learning_rate": 1.9447059030218876e-05, + "loss": 1.2239, + "step": 1227 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 4.810361623786047, + "learning_rate": 1.9445905208173387e-05, + "loss": 1.1468, + "step": 1228 + }, + { + "epoch": 0.6672095548317046, + "grad_norm": 4.6967843893557015, + "learning_rate": 1.944475021784328e-05, + "loss": 1.5304, + "step": 1229 + }, + { + "epoch": 0.6677524429967426, + "grad_norm": 5.325911382516086, + "learning_rate": 1.9443594059371417e-05, + "loss": 1.2218, + "step": 1230 + }, + { + "epoch": 0.6682953311617806, + "grad_norm": 4.416437363913103, + "learning_rate": 1.9442436732900788e-05, + "loss": 1.223, + "step": 1231 + }, + { + "epoch": 0.6688382193268186, + "grad_norm": 3.720560087272861, + "learning_rate": 1.9441278238574537e-05, + "loss": 0.8155, + "step": 1232 + }, + { + "epoch": 0.6693811074918566, + "grad_norm": 5.273159297494141, + "learning_rate": 1.9440118576535947e-05, + "loss": 1.1958, + "step": 1233 + }, + { + "epoch": 0.6699239956568946, + "grad_norm": 5.684867015184436, + "learning_rate": 1.9438957746928443e-05, + "loss": 1.2384, + "step": 1234 + }, + { + "epoch": 0.6704668838219326, + "grad_norm": 4.952541482610055, + "learning_rate": 1.9437795749895604e-05, + "loss": 1.337, + "step": 1235 + }, + { + "epoch": 0.6710097719869706, + "grad_norm": 4.788863413585636, + "learning_rate": 1.9436632585581145e-05, + "loss": 0.5182, + "step": 1236 + }, + { + "epoch": 0.6715526601520087, + "grad_norm": 5.723878805619219, + "learning_rate": 1.9435468254128925e-05, + "loss": 1.1444, + "step": 1237 + }, + { + "epoch": 0.6720955483170467, + "grad_norm": 4.3068839361839, + "learning_rate": 1.9434302755682958e-05, + "loss": 1.1286, + "step": 1238 + }, + { + "epoch": 0.6726384364820847, + "grad_norm": 3.7588872381668477, + "learning_rate": 1.9433136090387384e-05, + "loss": 0.9718, + "step": 1239 + }, + { + "epoch": 0.6731813246471227, + "grad_norm": 6.802207677110681, + "learning_rate": 1.9431968258386508e-05, + "loss": 1.2389, + "step": 1240 + }, + { + "epoch": 0.6737242128121607, + "grad_norm": 3.6180158875124735, + "learning_rate": 1.9430799259824766e-05, + "loss": 0.644, + "step": 1241 + }, + { + "epoch": 0.6742671009771987, + "grad_norm": 5.559822163300458, + "learning_rate": 1.9429629094846742e-05, + "loss": 1.1157, + "step": 1242 + }, + { + "epoch": 0.6748099891422367, + "grad_norm": 4.084355750224745, + "learning_rate": 1.942845776359716e-05, + "loss": 0.91, + "step": 1243 + }, + { + "epoch": 0.6753528773072747, + "grad_norm": 3.9402358983627583, + "learning_rate": 1.9427285266220895e-05, + "loss": 0.9409, + "step": 1244 + }, + { + "epoch": 0.6758957654723127, + "grad_norm": 5.093172430302549, + "learning_rate": 1.9426111602862968e-05, + "loss": 1.5905, + "step": 1245 + }, + { + "epoch": 0.6764386536373507, + "grad_norm": 4.02536040473524, + "learning_rate": 1.942493677366853e-05, + "loss": 0.9277, + "step": 1246 + }, + { + "epoch": 0.6769815418023887, + "grad_norm": 4.741533426757492, + "learning_rate": 1.942376077878289e-05, + "loss": 1.1027, + "step": 1247 + }, + { + "epoch": 0.6775244299674267, + "grad_norm": 5.365058237804207, + "learning_rate": 1.9422583618351503e-05, + "loss": 1.072, + "step": 1248 + }, + { + "epoch": 0.6780673181324647, + "grad_norm": 4.479339123691231, + "learning_rate": 1.9421405292519956e-05, + "loss": 1.474, + "step": 1249 + }, + { + "epoch": 0.6786102062975027, + "grad_norm": 7.79086989645888, + "learning_rate": 1.942022580143398e-05, + "loss": 1.6973, + "step": 1250 + }, + { + "epoch": 0.6791530944625407, + "grad_norm": 5.621467578044424, + "learning_rate": 1.9419045145239474e-05, + "loss": 1.5432, + "step": 1251 + }, + { + "epoch": 0.6796959826275787, + "grad_norm": 5.041038447258229, + "learning_rate": 1.9417863324082444e-05, + "loss": 1.2977, + "step": 1252 + }, + { + "epoch": 0.6802388707926167, + "grad_norm": 4.717189339301973, + "learning_rate": 1.941668033810907e-05, + "loss": 0.924, + "step": 1253 + }, + { + "epoch": 0.6807817589576547, + "grad_norm": 5.086833595497465, + "learning_rate": 1.9415496187465667e-05, + "loss": 1.5545, + "step": 1254 + }, + { + "epoch": 0.6813246471226927, + "grad_norm": 4.790403842984273, + "learning_rate": 1.941431087229869e-05, + "loss": 1.2545, + "step": 1255 + }, + { + "epoch": 0.6818675352877307, + "grad_norm": 4.470785518257313, + "learning_rate": 1.9413124392754735e-05, + "loss": 1.2036, + "step": 1256 + }, + { + "epoch": 0.6824104234527687, + "grad_norm": 3.820710537112263, + "learning_rate": 1.9411936748980557e-05, + "loss": 1.0856, + "step": 1257 + }, + { + "epoch": 0.6829533116178067, + "grad_norm": 4.453720544384727, + "learning_rate": 1.9410747941123036e-05, + "loss": 1.1645, + "step": 1258 + }, + { + "epoch": 0.6834961997828447, + "grad_norm": 4.692503519890958, + "learning_rate": 1.9409557969329218e-05, + "loss": 1.1174, + "step": 1259 + }, + { + "epoch": 0.6840390879478827, + "grad_norm": 5.666971506629844, + "learning_rate": 1.940836683374627e-05, + "loss": 1.4904, + "step": 1260 + }, + { + "epoch": 0.6845819761129207, + "grad_norm": 5.468580615919992, + "learning_rate": 1.940717453452152e-05, + "loss": 1.2834, + "step": 1261 + }, + { + "epoch": 0.6851248642779587, + "grad_norm": 6.42319990350808, + "learning_rate": 1.9405981071802425e-05, + "loss": 1.1144, + "step": 1262 + }, + { + "epoch": 0.6856677524429967, + "grad_norm": 6.078674980286373, + "learning_rate": 1.9404786445736607e-05, + "loss": 1.4214, + "step": 1263 + }, + { + "epoch": 0.6862106406080347, + "grad_norm": 4.465631588283012, + "learning_rate": 1.9403590656471806e-05, + "loss": 0.7813, + "step": 1264 + }, + { + "epoch": 0.6867535287730727, + "grad_norm": 5.220582351132811, + "learning_rate": 1.940239370415593e-05, + "loss": 1.2112, + "step": 1265 + }, + { + "epoch": 0.6872964169381107, + "grad_norm": 5.1374760487841895, + "learning_rate": 1.9401195588937014e-05, + "loss": 1.2402, + "step": 1266 + }, + { + "epoch": 0.6878393051031487, + "grad_norm": 5.106543779494583, + "learning_rate": 1.9399996310963243e-05, + "loss": 1.1846, + "step": 1267 + }, + { + "epoch": 0.6883821932681867, + "grad_norm": 3.9521299273320363, + "learning_rate": 1.939879587038295e-05, + "loss": 0.8975, + "step": 1268 + }, + { + "epoch": 0.6889250814332247, + "grad_norm": 5.398700470714175, + "learning_rate": 1.9397594267344604e-05, + "loss": 1.4686, + "step": 1269 + }, + { + "epoch": 0.6894679695982627, + "grad_norm": 5.8287973073589265, + "learning_rate": 1.939639150199682e-05, + "loss": 1.011, + "step": 1270 + }, + { + "epoch": 0.6900108577633007, + "grad_norm": 4.812101240482426, + "learning_rate": 1.9395187574488358e-05, + "loss": 0.948, + "step": 1271 + }, + { + "epoch": 0.6905537459283387, + "grad_norm": 5.828731649028842, + "learning_rate": 1.939398248496813e-05, + "loss": 1.2233, + "step": 1272 + }, + { + "epoch": 0.6910966340933767, + "grad_norm": 4.890080173319348, + "learning_rate": 1.9392776233585167e-05, + "loss": 0.8859, + "step": 1273 + }, + { + "epoch": 0.6916395222584147, + "grad_norm": 4.920508250004042, + "learning_rate": 1.9391568820488674e-05, + "loss": 0.73, + "step": 1274 + }, + { + "epoch": 0.6921824104234527, + "grad_norm": 5.635401553567158, + "learning_rate": 1.9390360245827983e-05, + "loss": 1.1432, + "step": 1275 + }, + { + "epoch": 0.6927252985884907, + "grad_norm": 4.588336191443314, + "learning_rate": 1.9389150509752566e-05, + "loss": 1.3326, + "step": 1276 + }, + { + "epoch": 0.6932681867535287, + "grad_norm": 4.631694372660094, + "learning_rate": 1.9387939612412056e-05, + "loss": 1.3762, + "step": 1277 + }, + { + "epoch": 0.6938110749185668, + "grad_norm": 5.585985863722577, + "learning_rate": 1.938672755395621e-05, + "loss": 1.3606, + "step": 1278 + }, + { + "epoch": 0.6943539630836048, + "grad_norm": 4.764783952643876, + "learning_rate": 1.938551433453494e-05, + "loss": 0.9858, + "step": 1279 + }, + { + "epoch": 0.6948968512486428, + "grad_norm": 4.234252918062867, + "learning_rate": 1.9384299954298297e-05, + "loss": 0.7702, + "step": 1280 + }, + { + "epoch": 0.6954397394136808, + "grad_norm": 4.36117548276201, + "learning_rate": 1.938308441339648e-05, + "loss": 1.1846, + "step": 1281 + }, + { + "epoch": 0.6959826275787188, + "grad_norm": 4.311957696864392, + "learning_rate": 1.938186771197983e-05, + "loss": 0.9934, + "step": 1282 + }, + { + "epoch": 0.6965255157437568, + "grad_norm": 5.117807004985933, + "learning_rate": 1.9380649850198824e-05, + "loss": 1.4474, + "step": 1283 + }, + { + "epoch": 0.6970684039087948, + "grad_norm": 5.695329527120276, + "learning_rate": 1.93794308282041e-05, + "loss": 1.3001, + "step": 1284 + }, + { + "epoch": 0.6976112920738328, + "grad_norm": 4.305244126285458, + "learning_rate": 1.937821064614642e-05, + "loss": 1.1421, + "step": 1285 + }, + { + "epoch": 0.6981541802388708, + "grad_norm": 4.574797635236639, + "learning_rate": 1.93769893041767e-05, + "loss": 1.1104, + "step": 1286 + }, + { + "epoch": 0.6986970684039088, + "grad_norm": 5.051193103711606, + "learning_rate": 1.9375766802446002e-05, + "loss": 1.2858, + "step": 1287 + }, + { + "epoch": 0.6992399565689468, + "grad_norm": 5.002235181102662, + "learning_rate": 1.9374543141105518e-05, + "loss": 1.2931, + "step": 1288 + }, + { + "epoch": 0.6997828447339848, + "grad_norm": 5.259386594746194, + "learning_rate": 1.93733183203066e-05, + "loss": 1.0432, + "step": 1289 + }, + { + "epoch": 0.7003257328990228, + "grad_norm": 3.7218449672815117, + "learning_rate": 1.9372092340200736e-05, + "loss": 0.9913, + "step": 1290 + }, + { + "epoch": 0.7008686210640608, + "grad_norm": 4.45109753373239, + "learning_rate": 1.937086520093955e-05, + "loss": 1.2012, + "step": 1291 + }, + { + "epoch": 0.7014115092290988, + "grad_norm": 4.73529586440727, + "learning_rate": 1.9369636902674823e-05, + "loss": 1.3995, + "step": 1292 + }, + { + "epoch": 0.7019543973941368, + "grad_norm": 3.845443933095641, + "learning_rate": 1.936840744555847e-05, + "loss": 1.0971, + "step": 1293 + }, + { + "epoch": 0.7024972855591748, + "grad_norm": 6.67046211091574, + "learning_rate": 1.9367176829742553e-05, + "loss": 1.5273, + "step": 1294 + }, + { + "epoch": 0.7030401737242128, + "grad_norm": 4.454761961182613, + "learning_rate": 1.9365945055379275e-05, + "loss": 1.1282, + "step": 1295 + }, + { + "epoch": 0.7035830618892508, + "grad_norm": 6.552953195473038, + "learning_rate": 1.936471212262099e-05, + "loss": 2.171, + "step": 1296 + }, + { + "epoch": 0.7041259500542888, + "grad_norm": 4.894073730847946, + "learning_rate": 1.9363478031620182e-05, + "loss": 1.2303, + "step": 1297 + }, + { + "epoch": 0.7046688382193268, + "grad_norm": 3.863202427039748, + "learning_rate": 1.936224278252949e-05, + "loss": 0.8515, + "step": 1298 + }, + { + "epoch": 0.7052117263843648, + "grad_norm": 5.503562663348476, + "learning_rate": 1.9361006375501685e-05, + "loss": 1.1834, + "step": 1299 + }, + { + "epoch": 0.7057546145494028, + "grad_norm": 5.318422186143503, + "learning_rate": 1.9359768810689697e-05, + "loss": 1.3378, + "step": 1300 + }, + { + "epoch": 0.7062975027144408, + "grad_norm": 4.46194766163203, + "learning_rate": 1.9358530088246582e-05, + "loss": 1.3318, + "step": 1301 + }, + { + "epoch": 0.7068403908794788, + "grad_norm": 5.6623901137247765, + "learning_rate": 1.9357290208325552e-05, + "loss": 1.3512, + "step": 1302 + }, + { + "epoch": 0.7073832790445168, + "grad_norm": 5.805010217787694, + "learning_rate": 1.9356049171079957e-05, + "loss": 1.2383, + "step": 1303 + }, + { + "epoch": 0.7079261672095548, + "grad_norm": 4.847713425075124, + "learning_rate": 1.9354806976663286e-05, + "loss": 0.7329, + "step": 1304 + }, + { + "epoch": 0.7084690553745928, + "grad_norm": 4.207200307039602, + "learning_rate": 1.935356362522918e-05, + "loss": 1.0119, + "step": 1305 + }, + { + "epoch": 0.7090119435396308, + "grad_norm": 5.76731204400203, + "learning_rate": 1.9352319116931417e-05, + "loss": 1.7383, + "step": 1306 + }, + { + "epoch": 0.7095548317046688, + "grad_norm": 4.989804434976344, + "learning_rate": 1.935107345192392e-05, + "loss": 1.1958, + "step": 1307 + }, + { + "epoch": 0.7100977198697068, + "grad_norm": 4.6099552434159135, + "learning_rate": 1.9349826630360757e-05, + "loss": 0.9933, + "step": 1308 + }, + { + "epoch": 0.7106406080347448, + "grad_norm": 4.693361205108708, + "learning_rate": 1.9348578652396136e-05, + "loss": 1.3553, + "step": 1309 + }, + { + "epoch": 0.7111834961997828, + "grad_norm": 5.473441077550542, + "learning_rate": 1.9347329518184406e-05, + "loss": 1.2202, + "step": 1310 + }, + { + "epoch": 0.7117263843648208, + "grad_norm": 5.885376904454794, + "learning_rate": 1.9346079227880062e-05, + "loss": 1.3698, + "step": 1311 + }, + { + "epoch": 0.7122692725298588, + "grad_norm": 4.115405735194196, + "learning_rate": 1.9344827781637744e-05, + "loss": 1.3361, + "step": 1312 + }, + { + "epoch": 0.7128121606948968, + "grad_norm": 6.745730370060005, + "learning_rate": 1.9343575179612236e-05, + "loss": 1.6411, + "step": 1313 + }, + { + "epoch": 0.7133550488599348, + "grad_norm": 4.363445628534919, + "learning_rate": 1.9342321421958455e-05, + "loss": 0.792, + "step": 1314 + }, + { + "epoch": 0.7138979370249728, + "grad_norm": 6.167581448099132, + "learning_rate": 1.9341066508831472e-05, + "loss": 1.4264, + "step": 1315 + }, + { + "epoch": 0.7144408251900108, + "grad_norm": 4.973571230650737, + "learning_rate": 1.9339810440386495e-05, + "loss": 0.9769, + "step": 1316 + }, + { + "epoch": 0.7149837133550488, + "grad_norm": 5.2637238951774945, + "learning_rate": 1.933855321677888e-05, + "loss": 1.2478, + "step": 1317 + }, + { + "epoch": 0.7155266015200868, + "grad_norm": 5.2296234862723505, + "learning_rate": 1.9337294838164118e-05, + "loss": 1.3999, + "step": 1318 + }, + { + "epoch": 0.7160694896851248, + "grad_norm": 4.720322767524389, + "learning_rate": 1.9336035304697848e-05, + "loss": 1.3561, + "step": 1319 + }, + { + "epoch": 0.7166123778501629, + "grad_norm": 4.2513159313917015, + "learning_rate": 1.9334774616535854e-05, + "loss": 1.176, + "step": 1320 + }, + { + "epoch": 0.7171552660152009, + "grad_norm": 4.5267349508423225, + "learning_rate": 1.9333512773834057e-05, + "loss": 1.1211, + "step": 1321 + }, + { + "epoch": 0.7176981541802389, + "grad_norm": 5.384909301403834, + "learning_rate": 1.9332249776748523e-05, + "loss": 1.3904, + "step": 1322 + }, + { + "epoch": 0.7182410423452769, + "grad_norm": 4.67839864320759, + "learning_rate": 1.9330985625435468e-05, + "loss": 0.8608, + "step": 1323 + }, + { + "epoch": 0.7187839305103149, + "grad_norm": 5.193826637693982, + "learning_rate": 1.9329720320051233e-05, + "loss": 1.3172, + "step": 1324 + }, + { + "epoch": 0.7193268186753529, + "grad_norm": 4.8605293911870096, + "learning_rate": 1.9328453860752324e-05, + "loss": 1.3134, + "step": 1325 + }, + { + "epoch": 0.7198697068403909, + "grad_norm": 5.685265101122876, + "learning_rate": 1.9327186247695377e-05, + "loss": 1.2036, + "step": 1326 + }, + { + "epoch": 0.7204125950054289, + "grad_norm": 5.159715270999642, + "learning_rate": 1.9325917481037164e-05, + "loss": 1.4898, + "step": 1327 + }, + { + "epoch": 0.7209554831704669, + "grad_norm": 5.793178164657722, + "learning_rate": 1.9324647560934613e-05, + "loss": 1.3909, + "step": 1328 + }, + { + "epoch": 0.7214983713355049, + "grad_norm": 3.7442056178189267, + "learning_rate": 1.9323376487544795e-05, + "loss": 0.774, + "step": 1329 + }, + { + "epoch": 0.7220412595005429, + "grad_norm": 4.962252385233631, + "learning_rate": 1.9322104261024912e-05, + "loss": 1.3528, + "step": 1330 + }, + { + "epoch": 0.7225841476655809, + "grad_norm": 5.4534463408495, + "learning_rate": 1.9320830881532316e-05, + "loss": 1.6754, + "step": 1331 + }, + { + "epoch": 0.7231270358306189, + "grad_norm": 6.301262380555263, + "learning_rate": 1.93195563492245e-05, + "loss": 1.5583, + "step": 1332 + }, + { + "epoch": 0.7236699239956569, + "grad_norm": 3.970618104957748, + "learning_rate": 1.9318280664259103e-05, + "loss": 0.8964, + "step": 1333 + }, + { + "epoch": 0.7242128121606949, + "grad_norm": 3.490105469488415, + "learning_rate": 1.9317003826793904e-05, + "loss": 0.8098, + "step": 1334 + }, + { + "epoch": 0.7247557003257329, + "grad_norm": 5.240119278714349, + "learning_rate": 1.9315725836986822e-05, + "loss": 1.1688, + "step": 1335 + }, + { + "epoch": 0.7252985884907709, + "grad_norm": 4.550130688769963, + "learning_rate": 1.931444669499592e-05, + "loss": 1.2569, + "step": 1336 + }, + { + "epoch": 0.7258414766558089, + "grad_norm": 5.058353172472431, + "learning_rate": 1.9313166400979404e-05, + "loss": 1.3471, + "step": 1337 + }, + { + "epoch": 0.7263843648208469, + "grad_norm": 4.528967594836655, + "learning_rate": 1.931188495509563e-05, + "loss": 1.1831, + "step": 1338 + }, + { + "epoch": 0.7269272529858849, + "grad_norm": 4.493564615513058, + "learning_rate": 1.931060235750308e-05, + "loss": 1.4889, + "step": 1339 + }, + { + "epoch": 0.7274701411509229, + "grad_norm": 5.176344629525538, + "learning_rate": 1.9309318608360392e-05, + "loss": 1.6704, + "step": 1340 + }, + { + "epoch": 0.7280130293159609, + "grad_norm": 6.738195156936204, + "learning_rate": 1.930803370782634e-05, + "loss": 1.2147, + "step": 1341 + }, + { + "epoch": 0.7285559174809989, + "grad_norm": 4.248439922934747, + "learning_rate": 1.9306747656059847e-05, + "loss": 1.1943, + "step": 1342 + }, + { + "epoch": 0.7290988056460369, + "grad_norm": 5.175389864390461, + "learning_rate": 1.930546045321997e-05, + "loss": 1.4587, + "step": 1343 + }, + { + "epoch": 0.7296416938110749, + "grad_norm": 5.19934365393201, + "learning_rate": 1.9304172099465914e-05, + "loss": 0.9433, + "step": 1344 + }, + { + "epoch": 0.7301845819761129, + "grad_norm": 6.18681550522946, + "learning_rate": 1.9302882594957025e-05, + "loss": 1.5576, + "step": 1345 + }, + { + "epoch": 0.7307274701411509, + "grad_norm": 4.311510847746277, + "learning_rate": 1.930159193985279e-05, + "loss": 1.327, + "step": 1346 + }, + { + "epoch": 0.7312703583061889, + "grad_norm": 4.044345899710192, + "learning_rate": 1.9300300134312838e-05, + "loss": 1.2106, + "step": 1347 + }, + { + "epoch": 0.7318132464712269, + "grad_norm": 4.181894752624144, + "learning_rate": 1.929900717849694e-05, + "loss": 0.8703, + "step": 1348 + }, + { + "epoch": 0.7323561346362649, + "grad_norm": 5.37566626429507, + "learning_rate": 1.929771307256502e-05, + "loss": 1.446, + "step": 1349 + }, + { + "epoch": 0.7328990228013029, + "grad_norm": 5.37247658500195, + "learning_rate": 1.9296417816677123e-05, + "loss": 0.9239, + "step": 1350 + }, + { + "epoch": 0.7334419109663409, + "grad_norm": 5.749510968066806, + "learning_rate": 1.929512141099346e-05, + "loss": 1.4799, + "step": 1351 + }, + { + "epoch": 0.7339847991313789, + "grad_norm": 5.079635654383841, + "learning_rate": 1.929382385567436e-05, + "loss": 1.2571, + "step": 1352 + }, + { + "epoch": 0.7345276872964169, + "grad_norm": 4.923400923431618, + "learning_rate": 1.929252515088032e-05, + "loss": 1.1967, + "step": 1353 + }, + { + "epoch": 0.7350705754614549, + "grad_norm": 4.397561354835364, + "learning_rate": 1.9291225296771957e-05, + "loss": 0.9088, + "step": 1354 + }, + { + "epoch": 0.7356134636264929, + "grad_norm": 4.13744065040371, + "learning_rate": 1.9289924293510037e-05, + "loss": 1.023, + "step": 1355 + }, + { + "epoch": 0.7361563517915309, + "grad_norm": 4.266553618249588, + "learning_rate": 1.9288622141255477e-05, + "loss": 1.019, + "step": 1356 + }, + { + "epoch": 0.7366992399565689, + "grad_norm": 4.52588883698767, + "learning_rate": 1.928731884016933e-05, + "loss": 1.0935, + "step": 1357 + }, + { + "epoch": 0.7372421281216069, + "grad_norm": 5.46382933116543, + "learning_rate": 1.9286014390412786e-05, + "loss": 1.2833, + "step": 1358 + }, + { + "epoch": 0.737785016286645, + "grad_norm": 4.88651312061201, + "learning_rate": 1.928470879214718e-05, + "loss": 1.6064, + "step": 1359 + }, + { + "epoch": 0.738327904451683, + "grad_norm": 5.403882256899124, + "learning_rate": 1.9283402045533995e-05, + "loss": 1.1537, + "step": 1360 + }, + { + "epoch": 0.738870792616721, + "grad_norm": 6.381532932652649, + "learning_rate": 1.928209415073485e-05, + "loss": 1.527, + "step": 1361 + }, + { + "epoch": 0.739413680781759, + "grad_norm": 4.795244725313847, + "learning_rate": 1.9280785107911505e-05, + "loss": 1.4696, + "step": 1362 + }, + { + "epoch": 0.739956568946797, + "grad_norm": 5.8579061053451955, + "learning_rate": 1.9279474917225866e-05, + "loss": 1.5566, + "step": 1363 + }, + { + "epoch": 0.740499457111835, + "grad_norm": 4.5018596923036185, + "learning_rate": 1.927816357883998e-05, + "loss": 0.9842, + "step": 1364 + }, + { + "epoch": 0.741042345276873, + "grad_norm": 3.596935443077127, + "learning_rate": 1.927685109291604e-05, + "loss": 0.632, + "step": 1365 + }, + { + "epoch": 0.741585233441911, + "grad_norm": 4.535193817325199, + "learning_rate": 1.9275537459616364e-05, + "loss": 1.332, + "step": 1366 + }, + { + "epoch": 0.742128121606949, + "grad_norm": 6.678844775309246, + "learning_rate": 1.9274222679103437e-05, + "loss": 1.6826, + "step": 1367 + }, + { + "epoch": 0.742671009771987, + "grad_norm": 4.444554760823683, + "learning_rate": 1.927290675153987e-05, + "loss": 0.9597, + "step": 1368 + }, + { + "epoch": 0.743213897937025, + "grad_norm": 5.19917061501543, + "learning_rate": 1.927158967708841e-05, + "loss": 1.4333, + "step": 1369 + }, + { + "epoch": 0.743756786102063, + "grad_norm": 4.742331177279521, + "learning_rate": 1.927027145591197e-05, + "loss": 1.1659, + "step": 1370 + }, + { + "epoch": 0.744299674267101, + "grad_norm": 3.713496752478956, + "learning_rate": 1.926895208817358e-05, + "loss": 0.8688, + "step": 1371 + }, + { + "epoch": 0.744842562432139, + "grad_norm": 4.424173231823696, + "learning_rate": 1.9267631574036417e-05, + "loss": 1.0751, + "step": 1372 + }, + { + "epoch": 0.745385450597177, + "grad_norm": 4.7028219143924055, + "learning_rate": 1.9266309913663815e-05, + "loss": 1.3478, + "step": 1373 + }, + { + "epoch": 0.745928338762215, + "grad_norm": 5.498103807374834, + "learning_rate": 1.9264987107219237e-05, + "loss": 1.168, + "step": 1374 + }, + { + "epoch": 0.746471226927253, + "grad_norm": 4.326699254091958, + "learning_rate": 1.9263663154866285e-05, + "loss": 1.0713, + "step": 1375 + }, + { + "epoch": 0.747014115092291, + "grad_norm": 5.58697502498226, + "learning_rate": 1.926233805676871e-05, + "loss": 1.2474, + "step": 1376 + }, + { + "epoch": 0.747557003257329, + "grad_norm": 6.475832093356038, + "learning_rate": 1.92610118130904e-05, + "loss": 1.2805, + "step": 1377 + }, + { + "epoch": 0.748099891422367, + "grad_norm": 5.194166453296571, + "learning_rate": 1.925968442399539e-05, + "loss": 0.9736, + "step": 1378 + }, + { + "epoch": 0.748642779587405, + "grad_norm": 3.867607548943643, + "learning_rate": 1.9258355889647855e-05, + "loss": 1.2325, + "step": 1379 + }, + { + "epoch": 0.749185667752443, + "grad_norm": 5.602424635230674, + "learning_rate": 1.925702621021211e-05, + "loss": 1.0153, + "step": 1380 + }, + { + "epoch": 0.749728555917481, + "grad_norm": 5.011820196133583, + "learning_rate": 1.9255695385852604e-05, + "loss": 1.1823, + "step": 1381 + }, + { + "epoch": 0.750271444082519, + "grad_norm": 5.238515717778492, + "learning_rate": 1.9254363416733944e-05, + "loss": 1.4279, + "step": 1382 + }, + { + "epoch": 0.750814332247557, + "grad_norm": 5.600089025591521, + "learning_rate": 1.925303030302087e-05, + "loss": 0.9212, + "step": 1383 + }, + { + "epoch": 0.751357220412595, + "grad_norm": 4.318444011047328, + "learning_rate": 1.9251696044878255e-05, + "loss": 0.9623, + "step": 1384 + }, + { + "epoch": 0.751900108577633, + "grad_norm": 5.471464995723797, + "learning_rate": 1.925036064247113e-05, + "loss": 1.5446, + "step": 1385 + }, + { + "epoch": 0.752442996742671, + "grad_norm": 5.153911794581426, + "learning_rate": 1.9249024095964663e-05, + "loss": 1.7223, + "step": 1386 + }, + { + "epoch": 0.752985884907709, + "grad_norm": 3.7202924563691537, + "learning_rate": 1.924768640552415e-05, + "loss": 1.1469, + "step": 1387 + }, + { + "epoch": 0.753528773072747, + "grad_norm": 4.612135079041308, + "learning_rate": 1.9246347571315043e-05, + "loss": 0.9572, + "step": 1388 + }, + { + "epoch": 0.754071661237785, + "grad_norm": 4.618898432554816, + "learning_rate": 1.9245007593502937e-05, + "loss": 1.0443, + "step": 1389 + }, + { + "epoch": 0.754614549402823, + "grad_norm": 5.2787203720857345, + "learning_rate": 1.9243666472253554e-05, + "loss": 1.2034, + "step": 1390 + }, + { + "epoch": 0.755157437567861, + "grad_norm": 4.510719578546688, + "learning_rate": 1.9242324207732766e-05, + "loss": 0.8055, + "step": 1391 + }, + { + "epoch": 0.755700325732899, + "grad_norm": 6.256234902874701, + "learning_rate": 1.9240980800106596e-05, + "loss": 1.5985, + "step": 1392 + }, + { + "epoch": 0.756243213897937, + "grad_norm": 9.01213198061367, + "learning_rate": 1.923963624954119e-05, + "loss": 1.7921, + "step": 1393 + }, + { + "epoch": 0.756786102062975, + "grad_norm": 5.149498439254725, + "learning_rate": 1.923829055620285e-05, + "loss": 1.2406, + "step": 1394 + }, + { + "epoch": 0.757328990228013, + "grad_norm": 5.083183084254609, + "learning_rate": 1.9236943720258007e-05, + "loss": 0.9887, + "step": 1395 + }, + { + "epoch": 0.757871878393051, + "grad_norm": 4.5118989088500685, + "learning_rate": 1.9235595741873247e-05, + "loss": 1.3528, + "step": 1396 + }, + { + "epoch": 0.758414766558089, + "grad_norm": 4.583236692880461, + "learning_rate": 1.923424662121528e-05, + "loss": 1.0185, + "step": 1397 + }, + { + "epoch": 0.758957654723127, + "grad_norm": 5.3602373686113625, + "learning_rate": 1.9232896358450976e-05, + "loss": 1.1827, + "step": 1398 + }, + { + "epoch": 0.759500542888165, + "grad_norm": 4.510149132944334, + "learning_rate": 1.9231544953747336e-05, + "loss": 0.9981, + "step": 1399 + }, + { + "epoch": 0.760043431053203, + "grad_norm": 5.008678356958532, + "learning_rate": 1.9230192407271506e-05, + "loss": 1.4957, + "step": 1400 + }, + { + "epoch": 0.760586319218241, + "grad_norm": 4.6938647576746995, + "learning_rate": 1.9228838719190765e-05, + "loss": 1.361, + "step": 1401 + }, + { + "epoch": 0.761129207383279, + "grad_norm": 5.263462060803471, + "learning_rate": 1.9227483889672544e-05, + "loss": 1.1716, + "step": 1402 + }, + { + "epoch": 0.761672095548317, + "grad_norm": 4.737410394333335, + "learning_rate": 1.9226127918884407e-05, + "loss": 1.3924, + "step": 1403 + }, + { + "epoch": 0.762214983713355, + "grad_norm": 5.117476933198257, + "learning_rate": 1.9224770806994066e-05, + "loss": 1.1215, + "step": 1404 + }, + { + "epoch": 0.7627578718783931, + "grad_norm": 4.525492833460315, + "learning_rate": 1.922341255416937e-05, + "loss": 1.0626, + "step": 1405 + }, + { + "epoch": 0.7633007600434311, + "grad_norm": 4.928417801176309, + "learning_rate": 1.9222053160578312e-05, + "loss": 0.9576, + "step": 1406 + }, + { + "epoch": 0.7638436482084691, + "grad_norm": 4.44076437857558, + "learning_rate": 1.9220692626389018e-05, + "loss": 0.9186, + "step": 1407 + }, + { + "epoch": 0.7643865363735071, + "grad_norm": 4.779705057681976, + "learning_rate": 1.9219330951769763e-05, + "loss": 1.3392, + "step": 1408 + }, + { + "epoch": 0.7649294245385451, + "grad_norm": 5.37856562129718, + "learning_rate": 1.9217968136888965e-05, + "loss": 1.043, + "step": 1409 + }, + { + "epoch": 0.7654723127035831, + "grad_norm": 5.596447438931628, + "learning_rate": 1.9216604181915178e-05, + "loss": 1.3223, + "step": 1410 + }, + { + "epoch": 0.7660152008686211, + "grad_norm": 5.267245970848837, + "learning_rate": 1.9215239087017093e-05, + "loss": 1.0484, + "step": 1411 + }, + { + "epoch": 0.7665580890336591, + "grad_norm": 6.124227707475327, + "learning_rate": 1.9213872852363552e-05, + "loss": 1.5361, + "step": 1412 + }, + { + "epoch": 0.7671009771986971, + "grad_norm": 5.632804630433347, + "learning_rate": 1.9212505478123532e-05, + "loss": 1.2227, + "step": 1413 + }, + { + "epoch": 0.7676438653637351, + "grad_norm": 4.903635376409972, + "learning_rate": 1.9211136964466152e-05, + "loss": 0.9045, + "step": 1414 + }, + { + "epoch": 0.7681867535287731, + "grad_norm": 4.995962297168909, + "learning_rate": 1.9209767311560673e-05, + "loss": 1.2364, + "step": 1415 + }, + { + "epoch": 0.7687296416938111, + "grad_norm": 4.76685077713632, + "learning_rate": 1.9208396519576494e-05, + "loss": 1.4849, + "step": 1416 + }, + { + "epoch": 0.7692725298588491, + "grad_norm": 5.66289239913894, + "learning_rate": 1.9207024588683158e-05, + "loss": 1.389, + "step": 1417 + }, + { + "epoch": 0.7698154180238871, + "grad_norm": 4.044195406366437, + "learning_rate": 1.920565151905035e-05, + "loss": 0.5736, + "step": 1418 + }, + { + "epoch": 0.7703583061889251, + "grad_norm": 5.610527750585898, + "learning_rate": 1.9204277310847887e-05, + "loss": 1.5147, + "step": 1419 + }, + { + "epoch": 0.7709011943539631, + "grad_norm": 3.8629108074125424, + "learning_rate": 1.9202901964245734e-05, + "loss": 0.9184, + "step": 1420 + }, + { + "epoch": 0.7714440825190011, + "grad_norm": 4.582445031278247, + "learning_rate": 1.9201525479414e-05, + "loss": 1.071, + "step": 1421 + }, + { + "epoch": 0.7719869706840391, + "grad_norm": 5.016846104390101, + "learning_rate": 1.9200147856522933e-05, + "loss": 1.3673, + "step": 1422 + }, + { + "epoch": 0.7725298588490771, + "grad_norm": 4.798189213061551, + "learning_rate": 1.9198769095742914e-05, + "loss": 1.3483, + "step": 1423 + }, + { + "epoch": 0.7730727470141151, + "grad_norm": 5.871902023790772, + "learning_rate": 1.9197389197244473e-05, + "loss": 1.7625, + "step": 1424 + }, + { + "epoch": 0.7736156351791531, + "grad_norm": 5.895934775040147, + "learning_rate": 1.9196008161198277e-05, + "loss": 0.999, + "step": 1425 + }, + { + "epoch": 0.7741585233441911, + "grad_norm": 5.0199402476408155, + "learning_rate": 1.9194625987775138e-05, + "loss": 1.3251, + "step": 1426 + }, + { + "epoch": 0.7747014115092291, + "grad_norm": 6.650397672217608, + "learning_rate": 1.9193242677146e-05, + "loss": 1.2162, + "step": 1427 + }, + { + "epoch": 0.7752442996742671, + "grad_norm": 5.515267491505962, + "learning_rate": 1.9191858229481958e-05, + "loss": 1.3849, + "step": 1428 + }, + { + "epoch": 0.7757871878393051, + "grad_norm": 5.694611687374825, + "learning_rate": 1.9190472644954236e-05, + "loss": 1.0831, + "step": 1429 + }, + { + "epoch": 0.7763300760043431, + "grad_norm": 4.548114219835821, + "learning_rate": 1.9189085923734215e-05, + "loss": 1.2549, + "step": 1430 + }, + { + "epoch": 0.7768729641693811, + "grad_norm": 6.119997613777156, + "learning_rate": 1.9187698065993398e-05, + "loss": 1.6137, + "step": 1431 + }, + { + "epoch": 0.7774158523344191, + "grad_norm": 4.574150272616086, + "learning_rate": 1.9186309071903445e-05, + "loss": 1.3015, + "step": 1432 + }, + { + "epoch": 0.7779587404994571, + "grad_norm": 4.77854731853541, + "learning_rate": 1.9184918941636142e-05, + "loss": 0.6973, + "step": 1433 + }, + { + "epoch": 0.7785016286644951, + "grad_norm": 6.926122760031406, + "learning_rate": 1.9183527675363425e-05, + "loss": 1.4034, + "step": 1434 + }, + { + "epoch": 0.7790445168295331, + "grad_norm": 4.6748890605309645, + "learning_rate": 1.9182135273257372e-05, + "loss": 0.9854, + "step": 1435 + }, + { + "epoch": 0.7795874049945711, + "grad_norm": 5.072338861625223, + "learning_rate": 1.9180741735490194e-05, + "loss": 1.1604, + "step": 1436 + }, + { + "epoch": 0.7801302931596091, + "grad_norm": 5.995183838581222, + "learning_rate": 1.9179347062234245e-05, + "loss": 1.3913, + "step": 1437 + }, + { + "epoch": 0.7806731813246471, + "grad_norm": 5.494006248546126, + "learning_rate": 1.917795125366202e-05, + "loss": 1.2541, + "step": 1438 + }, + { + "epoch": 0.7812160694896851, + "grad_norm": 3.8223879163574694, + "learning_rate": 1.917655430994616e-05, + "loss": 0.7292, + "step": 1439 + }, + { + "epoch": 0.7817589576547231, + "grad_norm": 4.1476284953657405, + "learning_rate": 1.9175156231259434e-05, + "loss": 0.608, + "step": 1440 + }, + { + "epoch": 0.7823018458197611, + "grad_norm": 5.5144637244676495, + "learning_rate": 1.9173757017774764e-05, + "loss": 1.2674, + "step": 1441 + }, + { + "epoch": 0.7828447339847991, + "grad_norm": 5.4660654663594945, + "learning_rate": 1.9172356669665206e-05, + "loss": 1.3043, + "step": 1442 + }, + { + "epoch": 0.7833876221498371, + "grad_norm": 3.526312966998694, + "learning_rate": 1.9170955187103957e-05, + "loss": 0.8721, + "step": 1443 + }, + { + "epoch": 0.7839305103148752, + "grad_norm": 4.395429767339641, + "learning_rate": 1.9169552570264355e-05, + "loss": 0.985, + "step": 1444 + }, + { + "epoch": 0.7844733984799132, + "grad_norm": 4.608635000415997, + "learning_rate": 1.9168148819319874e-05, + "loss": 1.3492, + "step": 1445 + }, + { + "epoch": 0.7850162866449512, + "grad_norm": 5.729965497961937, + "learning_rate": 1.9166743934444137e-05, + "loss": 1.3405, + "step": 1446 + }, + { + "epoch": 0.7855591748099892, + "grad_norm": 6.60908414105679, + "learning_rate": 1.91653379158109e-05, + "loss": 1.5271, + "step": 1447 + }, + { + "epoch": 0.7861020629750272, + "grad_norm": 4.91386728282805, + "learning_rate": 1.916393076359406e-05, + "loss": 1.475, + "step": 1448 + }, + { + "epoch": 0.7866449511400652, + "grad_norm": 5.286158051966196, + "learning_rate": 1.916252247796766e-05, + "loss": 1.5235, + "step": 1449 + }, + { + "epoch": 0.7871878393051032, + "grad_norm": 4.699459928019414, + "learning_rate": 1.916111305910588e-05, + "loss": 1.151, + "step": 1450 + }, + { + "epoch": 0.7877307274701412, + "grad_norm": 5.125659016648778, + "learning_rate": 1.915970250718303e-05, + "loss": 1.2952, + "step": 1451 + }, + { + "epoch": 0.7882736156351792, + "grad_norm": 4.976170683421487, + "learning_rate": 1.915829082237358e-05, + "loss": 1.3291, + "step": 1452 + }, + { + "epoch": 0.7888165038002172, + "grad_norm": 5.237853683905863, + "learning_rate": 1.9156878004852123e-05, + "loss": 1.4775, + "step": 1453 + }, + { + "epoch": 0.7893593919652552, + "grad_norm": 4.914941394388547, + "learning_rate": 1.9155464054793404e-05, + "loss": 1.2151, + "step": 1454 + }, + { + "epoch": 0.7899022801302932, + "grad_norm": 5.050785426148085, + "learning_rate": 1.9154048972372293e-05, + "loss": 1.163, + "step": 1455 + }, + { + "epoch": 0.7904451682953312, + "grad_norm": 4.2016259300832255, + "learning_rate": 1.915263275776382e-05, + "loss": 0.9601, + "step": 1456 + }, + { + "epoch": 0.7909880564603692, + "grad_norm": 5.48804064939896, + "learning_rate": 1.915121541114314e-05, + "loss": 1.3026, + "step": 1457 + }, + { + "epoch": 0.7915309446254072, + "grad_norm": 5.230495684608947, + "learning_rate": 1.9149796932685552e-05, + "loss": 1.1923, + "step": 1458 + }, + { + "epoch": 0.7920738327904452, + "grad_norm": 4.788655104859546, + "learning_rate": 1.91483773225665e-05, + "loss": 1.2437, + "step": 1459 + }, + { + "epoch": 0.7926167209554832, + "grad_norm": 4.506216689801701, + "learning_rate": 1.9146956580961556e-05, + "loss": 0.9364, + "step": 1460 + }, + { + "epoch": 0.7931596091205212, + "grad_norm": 4.2335852812311865, + "learning_rate": 1.9145534708046446e-05, + "loss": 0.7104, + "step": 1461 + }, + { + "epoch": 0.7937024972855592, + "grad_norm": 6.204688311211956, + "learning_rate": 1.914411170399703e-05, + "loss": 1.0825, + "step": 1462 + }, + { + "epoch": 0.7942453854505972, + "grad_norm": 3.445489329210515, + "learning_rate": 1.91426875689893e-05, + "loss": 0.9921, + "step": 1463 + }, + { + "epoch": 0.7947882736156352, + "grad_norm": 5.204416925095863, + "learning_rate": 1.9141262303199403e-05, + "loss": 1.3043, + "step": 1464 + }, + { + "epoch": 0.7953311617806732, + "grad_norm": 6.3486214559668985, + "learning_rate": 1.9139835906803612e-05, + "loss": 1.3193, + "step": 1465 + }, + { + "epoch": 0.7958740499457112, + "grad_norm": 5.610159156463615, + "learning_rate": 1.913840837997835e-05, + "loss": 1.0455, + "step": 1466 + }, + { + "epoch": 0.7964169381107492, + "grad_norm": 4.625080334899242, + "learning_rate": 1.913697972290018e-05, + "loss": 0.7981, + "step": 1467 + }, + { + "epoch": 0.7969598262757872, + "grad_norm": 5.993005746484773, + "learning_rate": 1.9135549935745792e-05, + "loss": 1.1674, + "step": 1468 + }, + { + "epoch": 0.7975027144408252, + "grad_norm": 5.084136125969368, + "learning_rate": 1.913411901869203e-05, + "loss": 1.0362, + "step": 1469 + }, + { + "epoch": 0.7980456026058632, + "grad_norm": 5.701736254232889, + "learning_rate": 1.913268697191587e-05, + "loss": 1.2159, + "step": 1470 + }, + { + "epoch": 0.7985884907709012, + "grad_norm": 4.715190322082246, + "learning_rate": 1.9131253795594428e-05, + "loss": 0.9848, + "step": 1471 + }, + { + "epoch": 0.7991313789359392, + "grad_norm": 5.375233653105075, + "learning_rate": 1.9129819489904964e-05, + "loss": 1.0476, + "step": 1472 + }, + { + "epoch": 0.7996742671009772, + "grad_norm": 6.352674291337691, + "learning_rate": 1.9128384055024874e-05, + "loss": 1.2362, + "step": 1473 + }, + { + "epoch": 0.8002171552660152, + "grad_norm": 5.46716259791096, + "learning_rate": 1.91269474911317e-05, + "loss": 1.2949, + "step": 1474 + }, + { + "epoch": 0.8007600434310532, + "grad_norm": 5.3150808603597826, + "learning_rate": 1.912550979840311e-05, + "loss": 1.1587, + "step": 1475 + }, + { + "epoch": 0.8013029315960912, + "grad_norm": 4.46814887402293, + "learning_rate": 1.9124070977016926e-05, + "loss": 0.9649, + "step": 1476 + }, + { + "epoch": 0.8018458197611292, + "grad_norm": 4.916267658604107, + "learning_rate": 1.9122631027151103e-05, + "loss": 1.2117, + "step": 1477 + }, + { + "epoch": 0.8023887079261672, + "grad_norm": 5.506988713852874, + "learning_rate": 1.9121189948983733e-05, + "loss": 1.3387, + "step": 1478 + }, + { + "epoch": 0.8029315960912052, + "grad_norm": 4.7704152930487895, + "learning_rate": 1.911974774269305e-05, + "loss": 1.3379, + "step": 1479 + }, + { + "epoch": 0.8034744842562432, + "grad_norm": 5.667769720352476, + "learning_rate": 1.9118304408457435e-05, + "loss": 1.0552, + "step": 1480 + }, + { + "epoch": 0.8040173724212812, + "grad_norm": 4.9471232370904925, + "learning_rate": 1.91168599464554e-05, + "loss": 1.0013, + "step": 1481 + }, + { + "epoch": 0.8045602605863192, + "grad_norm": 4.856623754785127, + "learning_rate": 1.9115414356865594e-05, + "loss": 0.9001, + "step": 1482 + }, + { + "epoch": 0.8051031487513572, + "grad_norm": 4.004976425614515, + "learning_rate": 1.9113967639866815e-05, + "loss": 0.8114, + "step": 1483 + }, + { + "epoch": 0.8056460369163952, + "grad_norm": 5.983237168687411, + "learning_rate": 1.911251979563799e-05, + "loss": 1.1678, + "step": 1484 + }, + { + "epoch": 0.8061889250814332, + "grad_norm": 4.808701495369871, + "learning_rate": 1.9111070824358196e-05, + "loss": 0.9181, + "step": 1485 + }, + { + "epoch": 0.8067318132464713, + "grad_norm": 5.080557830592386, + "learning_rate": 1.910962072620664e-05, + "loss": 1.0751, + "step": 1486 + }, + { + "epoch": 0.8072747014115093, + "grad_norm": 5.730412968009966, + "learning_rate": 1.9108169501362674e-05, + "loss": 1.2727, + "step": 1487 + }, + { + "epoch": 0.8078175895765473, + "grad_norm": 6.47838124684283, + "learning_rate": 1.9106717150005785e-05, + "loss": 1.6491, + "step": 1488 + }, + { + "epoch": 0.8083604777415853, + "grad_norm": 5.311235841870027, + "learning_rate": 1.910526367231561e-05, + "loss": 0.8382, + "step": 1489 + }, + { + "epoch": 0.8089033659066233, + "grad_norm": 5.43072146168114, + "learning_rate": 1.9103809068471914e-05, + "loss": 1.3026, + "step": 1490 + }, + { + "epoch": 0.8094462540716613, + "grad_norm": 4.46156225654375, + "learning_rate": 1.9102353338654597e-05, + "loss": 0.8071, + "step": 1491 + }, + { + "epoch": 0.8099891422366993, + "grad_norm": 5.297520452299967, + "learning_rate": 1.9100896483043714e-05, + "loss": 1.0625, + "step": 1492 + }, + { + "epoch": 0.8105320304017373, + "grad_norm": 6.736735698357962, + "learning_rate": 1.909943850181945e-05, + "loss": 1.6497, + "step": 1493 + }, + { + "epoch": 0.8110749185667753, + "grad_norm": 4.8675910913183955, + "learning_rate": 1.9097979395162132e-05, + "loss": 1.0822, + "step": 1494 + }, + { + "epoch": 0.8116178067318133, + "grad_norm": 4.5489525578867305, + "learning_rate": 1.909651916325222e-05, + "loss": 0.7908, + "step": 1495 + }, + { + "epoch": 0.8121606948968513, + "grad_norm": 6.002121574887706, + "learning_rate": 1.909505780627032e-05, + "loss": 0.5178, + "step": 1496 + }, + { + "epoch": 0.8127035830618893, + "grad_norm": 4.34564271493731, + "learning_rate": 1.9093595324397175e-05, + "loss": 0.9818, + "step": 1497 + }, + { + "epoch": 0.8132464712269273, + "grad_norm": 5.0382562486474525, + "learning_rate": 1.9092131717813668e-05, + "loss": 1.0997, + "step": 1498 + }, + { + "epoch": 0.8137893593919653, + "grad_norm": 6.8289191133749, + "learning_rate": 1.909066698670082e-05, + "loss": 1.0046, + "step": 1499 + }, + { + "epoch": 0.8143322475570033, + "grad_norm": 6.6968746410007585, + "learning_rate": 1.908920113123979e-05, + "loss": 1.5191, + "step": 1500 + }, + { + "epoch": 0.8148751357220413, + "grad_norm": 5.548259576405749, + "learning_rate": 1.9087734151611877e-05, + "loss": 1.3272, + "step": 1501 + }, + { + "epoch": 0.8154180238870793, + "grad_norm": 5.684602861026239, + "learning_rate": 1.9086266047998522e-05, + "loss": 1.0698, + "step": 1502 + }, + { + "epoch": 0.8159609120521173, + "grad_norm": 6.832627276636624, + "learning_rate": 1.90847968205813e-05, + "loss": 1.5796, + "step": 1503 + }, + { + "epoch": 0.8165038002171553, + "grad_norm": 5.605724679057613, + "learning_rate": 1.908332646954193e-05, + "loss": 1.1709, + "step": 1504 + }, + { + "epoch": 0.8170466883821933, + "grad_norm": 4.607345782065147, + "learning_rate": 1.908185499506226e-05, + "loss": 0.8686, + "step": 1505 + }, + { + "epoch": 0.8175895765472313, + "grad_norm": 4.687252800873217, + "learning_rate": 1.9080382397324296e-05, + "loss": 0.8881, + "step": 1506 + }, + { + "epoch": 0.8181324647122693, + "grad_norm": 6.357714385620256, + "learning_rate": 1.907890867651016e-05, + "loss": 1.1948, + "step": 1507 + }, + { + "epoch": 0.8186753528773073, + "grad_norm": 4.8608334992124425, + "learning_rate": 1.9077433832802135e-05, + "loss": 1.0311, + "step": 1508 + }, + { + "epoch": 0.8192182410423453, + "grad_norm": 5.948864135911491, + "learning_rate": 1.9075957866382623e-05, + "loss": 1.1314, + "step": 1509 + }, + { + "epoch": 0.8197611292073833, + "grad_norm": 4.807850657807276, + "learning_rate": 1.9074480777434178e-05, + "loss": 0.9478, + "step": 1510 + }, + { + "epoch": 0.8203040173724213, + "grad_norm": 4.26222909817726, + "learning_rate": 1.9073002566139486e-05, + "loss": 0.8541, + "step": 1511 + }, + { + "epoch": 0.8208469055374593, + "grad_norm": 6.6130092616391005, + "learning_rate": 1.9071523232681382e-05, + "loss": 1.2754, + "step": 1512 + }, + { + "epoch": 0.8213897937024973, + "grad_norm": 5.255155673043404, + "learning_rate": 1.907004277724282e-05, + "loss": 1.0813, + "step": 1513 + }, + { + "epoch": 0.8219326818675353, + "grad_norm": 5.079098182105948, + "learning_rate": 1.9068561200006917e-05, + "loss": 1.0016, + "step": 1514 + }, + { + "epoch": 0.8224755700325733, + "grad_norm": 5.491172879434626, + "learning_rate": 1.906707850115691e-05, + "loss": 1.2884, + "step": 1515 + }, + { + "epoch": 0.8230184581976113, + "grad_norm": 6.718962166599785, + "learning_rate": 1.9065594680876182e-05, + "loss": 1.4973, + "step": 1516 + }, + { + "epoch": 0.8235613463626493, + "grad_norm": 5.348428262646105, + "learning_rate": 1.9064109739348257e-05, + "loss": 1.1113, + "step": 1517 + }, + { + "epoch": 0.8241042345276873, + "grad_norm": 4.604059650726469, + "learning_rate": 1.906262367675679e-05, + "loss": 1.1614, + "step": 1518 + }, + { + "epoch": 0.8246471226927253, + "grad_norm": 5.895860662978225, + "learning_rate": 1.9061136493285586e-05, + "loss": 1.1532, + "step": 1519 + }, + { + "epoch": 0.8251900108577633, + "grad_norm": 5.621847434524929, + "learning_rate": 1.905964818911858e-05, + "loss": 0.9277, + "step": 1520 + }, + { + "epoch": 0.8257328990228013, + "grad_norm": 4.585955492662189, + "learning_rate": 1.9058158764439844e-05, + "loss": 0.7988, + "step": 1521 + }, + { + "epoch": 0.8262757871878393, + "grad_norm": 5.784308925988881, + "learning_rate": 1.9056668219433595e-05, + "loss": 1.6078, + "step": 1522 + }, + { + "epoch": 0.8268186753528773, + "grad_norm": 5.621216253388429, + "learning_rate": 1.905517655428419e-05, + "loss": 1.389, + "step": 1523 + }, + { + "epoch": 0.8273615635179153, + "grad_norm": 5.514208208061458, + "learning_rate": 1.9053683769176115e-05, + "loss": 0.7612, + "step": 1524 + }, + { + "epoch": 0.8279044516829533, + "grad_norm": 4.6572185133859065, + "learning_rate": 1.9052189864294002e-05, + "loss": 0.744, + "step": 1525 + }, + { + "epoch": 0.8284473398479913, + "grad_norm": 5.239632270223703, + "learning_rate": 1.905069483982262e-05, + "loss": 0.939, + "step": 1526 + }, + { + "epoch": 0.8289902280130294, + "grad_norm": 3.7037650901217454, + "learning_rate": 1.9049198695946876e-05, + "loss": 1.0177, + "step": 1527 + }, + { + "epoch": 0.8295331161780674, + "grad_norm": 6.974869726679236, + "learning_rate": 1.9047701432851813e-05, + "loss": 1.2722, + "step": 1528 + }, + { + "epoch": 0.8300760043431054, + "grad_norm": 6.642878483620589, + "learning_rate": 1.904620305072262e-05, + "loss": 1.5369, + "step": 1529 + }, + { + "epoch": 0.8306188925081434, + "grad_norm": 7.99082645392899, + "learning_rate": 1.9044703549744616e-05, + "loss": 1.2245, + "step": 1530 + }, + { + "epoch": 0.8311617806731814, + "grad_norm": 6.5593948883008135, + "learning_rate": 1.904320293010326e-05, + "loss": 1.307, + "step": 1531 + }, + { + "epoch": 0.8317046688382194, + "grad_norm": 4.930764759519961, + "learning_rate": 1.9041701191984155e-05, + "loss": 0.9564, + "step": 1532 + }, + { + "epoch": 0.8322475570032574, + "grad_norm": 3.9621684331427773, + "learning_rate": 1.9040198335573033e-05, + "loss": 0.8153, + "step": 1533 + }, + { + "epoch": 0.8327904451682954, + "grad_norm": 4.859084711241092, + "learning_rate": 1.9038694361055774e-05, + "loss": 1.0967, + "step": 1534 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.9476278172327595, + "learning_rate": 1.903718926861839e-05, + "loss": 1.4743, + "step": 1535 + }, + { + "epoch": 0.8338762214983714, + "grad_norm": 4.312795088472635, + "learning_rate": 1.903568305844704e-05, + "loss": 0.8249, + "step": 1536 + }, + { + "epoch": 0.8344191096634094, + "grad_norm": 6.638027572281857, + "learning_rate": 1.9034175730728e-05, + "loss": 1.6353, + "step": 1537 + }, + { + "epoch": 0.8349619978284474, + "grad_norm": 5.709548912096025, + "learning_rate": 1.9032667285647714e-05, + "loss": 1.8512, + "step": 1538 + }, + { + "epoch": 0.8355048859934854, + "grad_norm": 5.098275426179771, + "learning_rate": 1.9031157723392738e-05, + "loss": 1.0348, + "step": 1539 + }, + { + "epoch": 0.8360477741585234, + "grad_norm": 5.05236269448673, + "learning_rate": 1.9029647044149783e-05, + "loss": 1.108, + "step": 1540 + }, + { + "epoch": 0.8365906623235614, + "grad_norm": 4.674708875536442, + "learning_rate": 1.9028135248105692e-05, + "loss": 1.0453, + "step": 1541 + }, + { + "epoch": 0.8371335504885994, + "grad_norm": 4.717321956389267, + "learning_rate": 1.902662233544744e-05, + "loss": 0.9042, + "step": 1542 + }, + { + "epoch": 0.8376764386536374, + "grad_norm": 6.217040367936212, + "learning_rate": 1.9025108306362158e-05, + "loss": 1.0762, + "step": 1543 + }, + { + "epoch": 0.8382193268186754, + "grad_norm": 5.338744288323046, + "learning_rate": 1.9023593161037094e-05, + "loss": 1.1631, + "step": 1544 + }, + { + "epoch": 0.8387622149837134, + "grad_norm": 4.8224874086687874, + "learning_rate": 1.9022076899659643e-05, + "loss": 1.3907, + "step": 1545 + }, + { + "epoch": 0.8393051031487514, + "grad_norm": 4.699800413179793, + "learning_rate": 1.9020559522417345e-05, + "loss": 0.7682, + "step": 1546 + }, + { + "epoch": 0.8398479913137894, + "grad_norm": 4.271439794022252, + "learning_rate": 1.9019041029497866e-05, + "loss": 0.8475, + "step": 1547 + }, + { + "epoch": 0.8403908794788274, + "grad_norm": 4.240562925498168, + "learning_rate": 1.9017521421089022e-05, + "loss": 0.8201, + "step": 1548 + }, + { + "epoch": 0.8409337676438654, + "grad_norm": 5.24718786534657, + "learning_rate": 1.9016000697378755e-05, + "loss": 1.0728, + "step": 1549 + }, + { + "epoch": 0.8414766558089034, + "grad_norm": 6.492220902503762, + "learning_rate": 1.9014478858555156e-05, + "loss": 0.9432, + "step": 1550 + }, + { + "epoch": 0.8420195439739414, + "grad_norm": 4.917811923935393, + "learning_rate": 1.9012955904806438e-05, + "loss": 1.1672, + "step": 1551 + }, + { + "epoch": 0.8425624321389794, + "grad_norm": 5.66709669173795, + "learning_rate": 1.9011431836320976e-05, + "loss": 1.5058, + "step": 1552 + }, + { + "epoch": 0.8431053203040174, + "grad_norm": 5.658634152413846, + "learning_rate": 1.9009906653287258e-05, + "loss": 1.3653, + "step": 1553 + }, + { + "epoch": 0.8436482084690554, + "grad_norm": 4.956250367213818, + "learning_rate": 1.9008380355893925e-05, + "loss": 0.8309, + "step": 1554 + }, + { + "epoch": 0.8441910966340934, + "grad_norm": 4.329497016180362, + "learning_rate": 1.9006852944329753e-05, + "loss": 1.1141, + "step": 1555 + }, + { + "epoch": 0.8447339847991314, + "grad_norm": 5.717864943326053, + "learning_rate": 1.9005324418783658e-05, + "loss": 1.3274, + "step": 1556 + }, + { + "epoch": 0.8452768729641694, + "grad_norm": 5.657300743895006, + "learning_rate": 1.900379477944468e-05, + "loss": 1.2432, + "step": 1557 + }, + { + "epoch": 0.8458197611292074, + "grad_norm": 5.325761071371062, + "learning_rate": 1.900226402650202e-05, + "loss": 1.3428, + "step": 1558 + }, + { + "epoch": 0.8463626492942454, + "grad_norm": 5.207304704126981, + "learning_rate": 1.9000732160144996e-05, + "loss": 1.455, + "step": 1559 + }, + { + "epoch": 0.8469055374592834, + "grad_norm": 5.248656371435267, + "learning_rate": 1.8999199180563074e-05, + "loss": 0.7851, + "step": 1560 + }, + { + "epoch": 0.8474484256243214, + "grad_norm": 6.92723307445887, + "learning_rate": 1.899766508794585e-05, + "loss": 1.5236, + "step": 1561 + }, + { + "epoch": 0.8479913137893594, + "grad_norm": 4.3434986726191935, + "learning_rate": 1.899612988248307e-05, + "loss": 0.9117, + "step": 1562 + }, + { + "epoch": 0.8485342019543974, + "grad_norm": 5.815800316925401, + "learning_rate": 1.8994593564364612e-05, + "loss": 1.0097, + "step": 1563 + }, + { + "epoch": 0.8490770901194354, + "grad_norm": 5.5494362406379905, + "learning_rate": 1.8993056133780484e-05, + "loss": 1.1829, + "step": 1564 + }, + { + "epoch": 0.8496199782844734, + "grad_norm": 4.323494913553237, + "learning_rate": 1.899151759092084e-05, + "loss": 0.9359, + "step": 1565 + }, + { + "epoch": 0.8501628664495114, + "grad_norm": 4.455911135073797, + "learning_rate": 1.898997793597597e-05, + "loss": 0.9354, + "step": 1566 + }, + { + "epoch": 0.8507057546145494, + "grad_norm": 6.684553981467422, + "learning_rate": 1.8988437169136302e-05, + "loss": 1.3693, + "step": 1567 + }, + { + "epoch": 0.8512486427795874, + "grad_norm": 6.341757024960091, + "learning_rate": 1.89868952905924e-05, + "loss": 0.9441, + "step": 1568 + }, + { + "epoch": 0.8517915309446255, + "grad_norm": 6.017605683279906, + "learning_rate": 1.8985352300534965e-05, + "loss": 1.2747, + "step": 1569 + }, + { + "epoch": 0.8523344191096635, + "grad_norm": 5.501152081837135, + "learning_rate": 1.8983808199154835e-05, + "loss": 0.8414, + "step": 1570 + }, + { + "epoch": 0.8528773072747015, + "grad_norm": 5.075164131477861, + "learning_rate": 1.8982262986642993e-05, + "loss": 1.124, + "step": 1571 + }, + { + "epoch": 0.8534201954397395, + "grad_norm": 5.7750932469486065, + "learning_rate": 1.8980716663190545e-05, + "loss": 1.3968, + "step": 1572 + }, + { + "epoch": 0.8539630836047775, + "grad_norm": 6.967872140444835, + "learning_rate": 1.897916922898875e-05, + "loss": 1.2395, + "step": 1573 + }, + { + "epoch": 0.8545059717698155, + "grad_norm": 5.037804725894728, + "learning_rate": 1.8977620684228994e-05, + "loss": 0.9741, + "step": 1574 + }, + { + "epoch": 0.8550488599348535, + "grad_norm": 4.586000206489282, + "learning_rate": 1.8976071029102802e-05, + "loss": 0.6757, + "step": 1575 + }, + { + "epoch": 0.8555917480998915, + "grad_norm": 4.686417605581938, + "learning_rate": 1.897452026380184e-05, + "loss": 0.8382, + "step": 1576 + }, + { + "epoch": 0.8561346362649295, + "grad_norm": 5.133601528061789, + "learning_rate": 1.8972968388517908e-05, + "loss": 1.2999, + "step": 1577 + }, + { + "epoch": 0.8566775244299675, + "grad_norm": 6.170796021401621, + "learning_rate": 1.8971415403442942e-05, + "loss": 1.1513, + "step": 1578 + }, + { + "epoch": 0.8572204125950055, + "grad_norm": 5.229050623246509, + "learning_rate": 1.8969861308769025e-05, + "loss": 0.9371, + "step": 1579 + }, + { + "epoch": 0.8577633007600435, + "grad_norm": 5.081229057143735, + "learning_rate": 1.8968306104688365e-05, + "loss": 0.9398, + "step": 1580 + }, + { + "epoch": 0.8583061889250815, + "grad_norm": 4.641876540865143, + "learning_rate": 1.896674979139331e-05, + "loss": 1.1401, + "step": 1581 + }, + { + "epoch": 0.8588490770901195, + "grad_norm": 4.149586485324207, + "learning_rate": 1.8965192369076356e-05, + "loss": 0.6768, + "step": 1582 + }, + { + "epoch": 0.8593919652551575, + "grad_norm": 6.422772444923944, + "learning_rate": 1.8963633837930114e-05, + "loss": 1.392, + "step": 1583 + }, + { + "epoch": 0.8599348534201955, + "grad_norm": 4.860635992388028, + "learning_rate": 1.8962074198147357e-05, + "loss": 0.9983, + "step": 1584 + }, + { + "epoch": 0.8604777415852335, + "grad_norm": 4.216877618358879, + "learning_rate": 1.8960513449920982e-05, + "loss": 0.6958, + "step": 1585 + }, + { + "epoch": 0.8610206297502715, + "grad_norm": 5.012340070254638, + "learning_rate": 1.8958951593444017e-05, + "loss": 0.9859, + "step": 1586 + }, + { + "epoch": 0.8615635179153095, + "grad_norm": 5.467321056155664, + "learning_rate": 1.8957388628909644e-05, + "loss": 1.3782, + "step": 1587 + }, + { + "epoch": 0.8621064060803475, + "grad_norm": 5.59589685839463, + "learning_rate": 1.8955824556511168e-05, + "loss": 0.8297, + "step": 1588 + }, + { + "epoch": 0.8626492942453855, + "grad_norm": 6.519624998298618, + "learning_rate": 1.895425937644204e-05, + "loss": 1.7264, + "step": 1589 + }, + { + "epoch": 0.8631921824104235, + "grad_norm": 6.6951270307269795, + "learning_rate": 1.8952693088895837e-05, + "loss": 1.2919, + "step": 1590 + }, + { + "epoch": 0.8637350705754615, + "grad_norm": 4.348741825803296, + "learning_rate": 1.895112569406629e-05, + "loss": 1.0062, + "step": 1591 + }, + { + "epoch": 0.8642779587404995, + "grad_norm": 5.0213162819553565, + "learning_rate": 1.8949557192147243e-05, + "loss": 1.2381, + "step": 1592 + }, + { + "epoch": 0.8648208469055375, + "grad_norm": 6.1637021381056885, + "learning_rate": 1.8947987583332705e-05, + "loss": 1.2561, + "step": 1593 + }, + { + "epoch": 0.8653637350705755, + "grad_norm": 6.638257635344733, + "learning_rate": 1.89464168678168e-05, + "loss": 0.9861, + "step": 1594 + }, + { + "epoch": 0.8659066232356135, + "grad_norm": 4.861237127098975, + "learning_rate": 1.89448450457938e-05, + "loss": 0.9407, + "step": 1595 + }, + { + "epoch": 0.8664495114006515, + "grad_norm": 6.4248430682281565, + "learning_rate": 1.894327211745811e-05, + "loss": 1.3079, + "step": 1596 + }, + { + "epoch": 0.8669923995656895, + "grad_norm": 5.712502208347676, + "learning_rate": 1.8941698083004265e-05, + "loss": 1.4422, + "step": 1597 + }, + { + "epoch": 0.8675352877307275, + "grad_norm": 5.634308307822415, + "learning_rate": 1.8940122942626957e-05, + "loss": 0.7869, + "step": 1598 + }, + { + "epoch": 0.8680781758957655, + "grad_norm": 7.839535783401759, + "learning_rate": 1.893854669652099e-05, + "loss": 1.3397, + "step": 1599 + }, + { + "epoch": 0.8686210640608035, + "grad_norm": 5.156814072355032, + "learning_rate": 1.8936969344881323e-05, + "loss": 1.2541, + "step": 1600 + }, + { + "epoch": 0.8691639522258415, + "grad_norm": 4.912293322601371, + "learning_rate": 1.8935390887903044e-05, + "loss": 0.8418, + "step": 1601 + }, + { + "epoch": 0.8697068403908795, + "grad_norm": 5.436641391280079, + "learning_rate": 1.8933811325781382e-05, + "loss": 1.4456, + "step": 1602 + }, + { + "epoch": 0.8702497285559175, + "grad_norm": 5.221596044730403, + "learning_rate": 1.8932230658711696e-05, + "loss": 0.9626, + "step": 1603 + }, + { + "epoch": 0.8707926167209555, + "grad_norm": 6.103653494209826, + "learning_rate": 1.8930648886889482e-05, + "loss": 1.2338, + "step": 1604 + }, + { + "epoch": 0.8713355048859935, + "grad_norm": 7.003639368124227, + "learning_rate": 1.8929066010510383e-05, + "loss": 1.2216, + "step": 1605 + }, + { + "epoch": 0.8718783930510315, + "grad_norm": 4.911905632269173, + "learning_rate": 1.8927482029770168e-05, + "loss": 0.9049, + "step": 1606 + }, + { + "epoch": 0.8724212812160695, + "grad_norm": 6.079178298711795, + "learning_rate": 1.8925896944864748e-05, + "loss": 1.6408, + "step": 1607 + }, + { + "epoch": 0.8729641693811075, + "grad_norm": 6.045568168649525, + "learning_rate": 1.892431075599017e-05, + "loss": 1.084, + "step": 1608 + }, + { + "epoch": 0.8735070575461455, + "grad_norm": 4.827633470123435, + "learning_rate": 1.892272346334261e-05, + "loss": 0.8121, + "step": 1609 + }, + { + "epoch": 0.8740499457111836, + "grad_norm": 5.314535878915851, + "learning_rate": 1.8921135067118396e-05, + "loss": 1.2657, + "step": 1610 + }, + { + "epoch": 0.8745928338762216, + "grad_norm": 4.786606110240068, + "learning_rate": 1.8919545567513976e-05, + "loss": 1.0534, + "step": 1611 + }, + { + "epoch": 0.8751357220412594, + "grad_norm": 4.075700855026623, + "learning_rate": 1.8917954964725948e-05, + "loss": 0.7757, + "step": 1612 + }, + { + "epoch": 0.8756786102062974, + "grad_norm": 6.286980562376595, + "learning_rate": 1.8916363258951033e-05, + "loss": 0.9782, + "step": 1613 + }, + { + "epoch": 0.8762214983713354, + "grad_norm": 6.745303990081763, + "learning_rate": 1.8914770450386102e-05, + "loss": 0.9488, + "step": 1614 + }, + { + "epoch": 0.8767643865363735, + "grad_norm": 6.2663007227218275, + "learning_rate": 1.8913176539228152e-05, + "loss": 1.9529, + "step": 1615 + }, + { + "epoch": 0.8773072747014115, + "grad_norm": 4.793153881320079, + "learning_rate": 1.8911581525674324e-05, + "loss": 1.0242, + "step": 1616 + }, + { + "epoch": 0.8778501628664495, + "grad_norm": 4.6844946413894, + "learning_rate": 1.890998540992189e-05, + "loss": 0.8565, + "step": 1617 + }, + { + "epoch": 0.8783930510314875, + "grad_norm": 3.993737542716379, + "learning_rate": 1.8908388192168256e-05, + "loss": 0.6107, + "step": 1618 + }, + { + "epoch": 0.8789359391965255, + "grad_norm": 6.099718972513945, + "learning_rate": 1.8906789872610977e-05, + "loss": 0.8043, + "step": 1619 + }, + { + "epoch": 0.8794788273615635, + "grad_norm": 7.113418049135383, + "learning_rate": 1.8905190451447726e-05, + "loss": 1.241, + "step": 1620 + }, + { + "epoch": 0.8800217155266015, + "grad_norm": 4.439254236407197, + "learning_rate": 1.8903589928876337e-05, + "loss": 1.0627, + "step": 1621 + }, + { + "epoch": 0.8805646036916395, + "grad_norm": 5.773948255627926, + "learning_rate": 1.8901988305094746e-05, + "loss": 1.3241, + "step": 1622 + }, + { + "epoch": 0.8811074918566775, + "grad_norm": 6.231870498005159, + "learning_rate": 1.890038558030106e-05, + "loss": 1.5241, + "step": 1623 + }, + { + "epoch": 0.8816503800217155, + "grad_norm": 5.888612961801449, + "learning_rate": 1.8898781754693495e-05, + "loss": 1.109, + "step": 1624 + }, + { + "epoch": 0.8821932681867535, + "grad_norm": 4.880160750019388, + "learning_rate": 1.8897176828470424e-05, + "loss": 1.0124, + "step": 1625 + }, + { + "epoch": 0.8827361563517915, + "grad_norm": 6.365487465081988, + "learning_rate": 1.889557080183034e-05, + "loss": 0.8999, + "step": 1626 + }, + { + "epoch": 0.8832790445168295, + "grad_norm": 4.461166628483842, + "learning_rate": 1.8893963674971883e-05, + "loss": 0.8762, + "step": 1627 + }, + { + "epoch": 0.8838219326818675, + "grad_norm": 5.675449614336963, + "learning_rate": 1.8892355448093825e-05, + "loss": 1.049, + "step": 1628 + }, + { + "epoch": 0.8843648208469055, + "grad_norm": 4.320721965444944, + "learning_rate": 1.8890746121395072e-05, + "loss": 1.1291, + "step": 1629 + }, + { + "epoch": 0.8849077090119435, + "grad_norm": 5.155835351166163, + "learning_rate": 1.8889135695074668e-05, + "loss": 1.1035, + "step": 1630 + }, + { + "epoch": 0.8854505971769815, + "grad_norm": 6.506050711845877, + "learning_rate": 1.8887524169331794e-05, + "loss": 2.1954, + "step": 1631 + }, + { + "epoch": 0.8859934853420195, + "grad_norm": 5.514989348863087, + "learning_rate": 1.8885911544365766e-05, + "loss": 1.0237, + "step": 1632 + }, + { + "epoch": 0.8865363735070575, + "grad_norm": 5.123449587076818, + "learning_rate": 1.8884297820376038e-05, + "loss": 0.9908, + "step": 1633 + }, + { + "epoch": 0.8870792616720955, + "grad_norm": 5.421671567850239, + "learning_rate": 1.8882682997562197e-05, + "loss": 1.7734, + "step": 1634 + }, + { + "epoch": 0.8876221498371335, + "grad_norm": 7.1012436118787186, + "learning_rate": 1.8881067076123963e-05, + "loss": 1.5488, + "step": 1635 + }, + { + "epoch": 0.8881650380021715, + "grad_norm": 5.328780061143714, + "learning_rate": 1.88794500562612e-05, + "loss": 1.3832, + "step": 1636 + }, + { + "epoch": 0.8887079261672095, + "grad_norm": 5.1503624666971595, + "learning_rate": 1.88778319381739e-05, + "loss": 0.8524, + "step": 1637 + }, + { + "epoch": 0.8892508143322475, + "grad_norm": 5.131553278195334, + "learning_rate": 1.88762127220622e-05, + "loss": 0.9769, + "step": 1638 + }, + { + "epoch": 0.8897937024972855, + "grad_norm": 3.1519686008526135, + "learning_rate": 1.8874592408126365e-05, + "loss": 0.472, + "step": 1639 + }, + { + "epoch": 0.8903365906623235, + "grad_norm": 4.164586140054025, + "learning_rate": 1.8872970996566794e-05, + "loss": 0.67, + "step": 1640 + }, + { + "epoch": 0.8908794788273615, + "grad_norm": 7.488659246842809, + "learning_rate": 1.8871348487584028e-05, + "loss": 1.3141, + "step": 1641 + }, + { + "epoch": 0.8914223669923995, + "grad_norm": 5.911298710361154, + "learning_rate": 1.8869724881378743e-05, + "loss": 1.084, + "step": 1642 + }, + { + "epoch": 0.8919652551574375, + "grad_norm": 5.977793625957034, + "learning_rate": 1.886810017815175e-05, + "loss": 0.9273, + "step": 1643 + }, + { + "epoch": 0.8925081433224755, + "grad_norm": 5.2773273690434515, + "learning_rate": 1.8866474378103993e-05, + "loss": 1.2594, + "step": 1644 + }, + { + "epoch": 0.8930510314875135, + "grad_norm": 8.049487951903771, + "learning_rate": 1.8864847481436554e-05, + "loss": 1.0104, + "step": 1645 + }, + { + "epoch": 0.8935939196525515, + "grad_norm": 5.428394606766542, + "learning_rate": 1.886321948835065e-05, + "loss": 0.827, + "step": 1646 + }, + { + "epoch": 0.8941368078175895, + "grad_norm": 5.687133062399138, + "learning_rate": 1.8861590399047635e-05, + "loss": 1.2759, + "step": 1647 + }, + { + "epoch": 0.8946796959826275, + "grad_norm": 5.037591954761082, + "learning_rate": 1.885996021372899e-05, + "loss": 0.8928, + "step": 1648 + }, + { + "epoch": 0.8952225841476655, + "grad_norm": 4.911533421640023, + "learning_rate": 1.8858328932596352e-05, + "loss": 0.705, + "step": 1649 + }, + { + "epoch": 0.8957654723127035, + "grad_norm": 6.054608860222289, + "learning_rate": 1.885669655585147e-05, + "loss": 1.296, + "step": 1650 + }, + { + "epoch": 0.8963083604777415, + "grad_norm": 6.221099773279408, + "learning_rate": 1.8855063083696244e-05, + "loss": 1.5253, + "step": 1651 + }, + { + "epoch": 0.8968512486427795, + "grad_norm": 5.0718329796444035, + "learning_rate": 1.8853428516332702e-05, + "loss": 1.1683, + "step": 1652 + }, + { + "epoch": 0.8973941368078175, + "grad_norm": 5.721985007773526, + "learning_rate": 1.8851792853963015e-05, + "loss": 1.1635, + "step": 1653 + }, + { + "epoch": 0.8979370249728555, + "grad_norm": 7.224045168649458, + "learning_rate": 1.8850156096789473e-05, + "loss": 1.0042, + "step": 1654 + }, + { + "epoch": 0.8984799131378935, + "grad_norm": 4.938883702947416, + "learning_rate": 1.8848518245014526e-05, + "loss": 0.7712, + "step": 1655 + }, + { + "epoch": 0.8990228013029316, + "grad_norm": 5.9261619850877, + "learning_rate": 1.8846879298840735e-05, + "loss": 1.4387, + "step": 1656 + }, + { + "epoch": 0.8995656894679696, + "grad_norm": 5.181216185736725, + "learning_rate": 1.8845239258470817e-05, + "loss": 0.9389, + "step": 1657 + }, + { + "epoch": 0.9001085776330076, + "grad_norm": 5.108249236166185, + "learning_rate": 1.8843598124107608e-05, + "loss": 1.0034, + "step": 1658 + }, + { + "epoch": 0.9006514657980456, + "grad_norm": 4.839799276438571, + "learning_rate": 1.8841955895954088e-05, + "loss": 0.8524, + "step": 1659 + }, + { + "epoch": 0.9011943539630836, + "grad_norm": 5.598018889964705, + "learning_rate": 1.8840312574213372e-05, + "loss": 1.0939, + "step": 1660 + }, + { + "epoch": 0.9017372421281216, + "grad_norm": 5.2496226470868725, + "learning_rate": 1.8838668159088707e-05, + "loss": 1.0812, + "step": 1661 + }, + { + "epoch": 0.9022801302931596, + "grad_norm": 5.2298567821168, + "learning_rate": 1.8837022650783477e-05, + "loss": 0.953, + "step": 1662 + }, + { + "epoch": 0.9028230184581976, + "grad_norm": 4.471282412199049, + "learning_rate": 1.88353760495012e-05, + "loss": 1.2652, + "step": 1663 + }, + { + "epoch": 0.9033659066232356, + "grad_norm": 5.892082884595572, + "learning_rate": 1.8833728355445534e-05, + "loss": 1.1165, + "step": 1664 + }, + { + "epoch": 0.9039087947882736, + "grad_norm": 5.49174731306936, + "learning_rate": 1.8832079568820268e-05, + "loss": 0.9262, + "step": 1665 + }, + { + "epoch": 0.9044516829533116, + "grad_norm": 5.55211658654164, + "learning_rate": 1.883042968982932e-05, + "loss": 1.3112, + "step": 1666 + }, + { + "epoch": 0.9049945711183496, + "grad_norm": 5.597392580982551, + "learning_rate": 1.8828778718676757e-05, + "loss": 0.8511, + "step": 1667 + }, + { + "epoch": 0.9055374592833876, + "grad_norm": 6.183127994440369, + "learning_rate": 1.8827126655566773e-05, + "loss": 1.2784, + "step": 1668 + }, + { + "epoch": 0.9060803474484256, + "grad_norm": 6.991796130593696, + "learning_rate": 1.882547350070369e-05, + "loss": 1.1703, + "step": 1669 + }, + { + "epoch": 0.9066232356134636, + "grad_norm": 5.7794859741652305, + "learning_rate": 1.8823819254291986e-05, + "loss": 1.0974, + "step": 1670 + }, + { + "epoch": 0.9071661237785016, + "grad_norm": 5.388338879165358, + "learning_rate": 1.8822163916536245e-05, + "loss": 1.2087, + "step": 1671 + }, + { + "epoch": 0.9077090119435396, + "grad_norm": 5.072420136637763, + "learning_rate": 1.8820507487641218e-05, + "loss": 0.9244, + "step": 1672 + }, + { + "epoch": 0.9082519001085776, + "grad_norm": 5.9571835605705115, + "learning_rate": 1.8818849967811762e-05, + "loss": 1.3299, + "step": 1673 + }, + { + "epoch": 0.9087947882736156, + "grad_norm": 5.706573186001289, + "learning_rate": 1.8817191357252892e-05, + "loss": 1.1059, + "step": 1674 + }, + { + "epoch": 0.9093376764386536, + "grad_norm": 5.4984127833712435, + "learning_rate": 1.8815531656169737e-05, + "loss": 1.1428, + "step": 1675 + }, + { + "epoch": 0.9098805646036916, + "grad_norm": 4.824306323799433, + "learning_rate": 1.8813870864767582e-05, + "loss": 0.7313, + "step": 1676 + }, + { + "epoch": 0.9104234527687296, + "grad_norm": 6.180183255961245, + "learning_rate": 1.8812208983251828e-05, + "loss": 1.0729, + "step": 1677 + }, + { + "epoch": 0.9109663409337676, + "grad_norm": 6.586598617269493, + "learning_rate": 1.8810546011828024e-05, + "loss": 1.5871, + "step": 1678 + }, + { + "epoch": 0.9115092290988056, + "grad_norm": 5.5611281281711875, + "learning_rate": 1.8808881950701845e-05, + "loss": 0.8133, + "step": 1679 + }, + { + "epoch": 0.9120521172638436, + "grad_norm": 5.214863992160379, + "learning_rate": 1.8807216800079108e-05, + "loss": 1.0437, + "step": 1680 + }, + { + "epoch": 0.9125950054288816, + "grad_norm": 6.631045200232836, + "learning_rate": 1.8805550560165763e-05, + "loss": 1.446, + "step": 1681 + }, + { + "epoch": 0.9131378935939196, + "grad_norm": 6.140822200691373, + "learning_rate": 1.8803883231167887e-05, + "loss": 1.8672, + "step": 1682 + }, + { + "epoch": 0.9136807817589576, + "grad_norm": 4.5606387908479995, + "learning_rate": 1.8802214813291708e-05, + "loss": 1.0662, + "step": 1683 + }, + { + "epoch": 0.9142236699239956, + "grad_norm": 4.166651706834997, + "learning_rate": 1.8800545306743567e-05, + "loss": 0.749, + "step": 1684 + }, + { + "epoch": 0.9147665580890336, + "grad_norm": 4.674238905131916, + "learning_rate": 1.8798874711729957e-05, + "loss": 1.1106, + "step": 1685 + }, + { + "epoch": 0.9153094462540716, + "grad_norm": 5.224600466137879, + "learning_rate": 1.8797203028457497e-05, + "loss": 0.9652, + "step": 1686 + }, + { + "epoch": 0.9158523344191096, + "grad_norm": 5.87706899393542, + "learning_rate": 1.8795530257132947e-05, + "loss": 1.2146, + "step": 1687 + }, + { + "epoch": 0.9163952225841476, + "grad_norm": 7.587853042260599, + "learning_rate": 1.87938563979632e-05, + "loss": 1.9051, + "step": 1688 + }, + { + "epoch": 0.9169381107491856, + "grad_norm": 4.988334377581133, + "learning_rate": 1.8792181451155275e-05, + "loss": 0.8502, + "step": 1689 + }, + { + "epoch": 0.9174809989142236, + "grad_norm": 5.454638378956608, + "learning_rate": 1.8790505416916338e-05, + "loss": 1.0493, + "step": 1690 + }, + { + "epoch": 0.9180238870792616, + "grad_norm": 5.95672873256444, + "learning_rate": 1.878882829545368e-05, + "loss": 1.1296, + "step": 1691 + }, + { + "epoch": 0.9185667752442996, + "grad_norm": 6.921816321230723, + "learning_rate": 1.8787150086974734e-05, + "loss": 1.6388, + "step": 1692 + }, + { + "epoch": 0.9191096634093376, + "grad_norm": 5.423998305707574, + "learning_rate": 1.878547079168706e-05, + "loss": 1.4865, + "step": 1693 + }, + { + "epoch": 0.9196525515743756, + "grad_norm": 4.479238264802495, + "learning_rate": 1.878379040979835e-05, + "loss": 1.0257, + "step": 1694 + }, + { + "epoch": 0.9201954397394136, + "grad_norm": 6.325821606507086, + "learning_rate": 1.8782108941516446e-05, + "loss": 1.1156, + "step": 1695 + }, + { + "epoch": 0.9207383279044516, + "grad_norm": 6.220712298120886, + "learning_rate": 1.8780426387049315e-05, + "loss": 1.375, + "step": 1696 + }, + { + "epoch": 0.9212812160694897, + "grad_norm": 5.813906943099404, + "learning_rate": 1.877874274660505e-05, + "loss": 1.1869, + "step": 1697 + }, + { + "epoch": 0.9218241042345277, + "grad_norm": 5.0160254446208965, + "learning_rate": 1.8777058020391893e-05, + "loss": 1.0443, + "step": 1698 + }, + { + "epoch": 0.9223669923995657, + "grad_norm": 4.273981514908416, + "learning_rate": 1.877537220861821e-05, + "loss": 1.0015, + "step": 1699 + }, + { + "epoch": 0.9229098805646037, + "grad_norm": 4.908796260576538, + "learning_rate": 1.8773685311492513e-05, + "loss": 1.0203, + "step": 1700 + }, + { + "epoch": 0.9234527687296417, + "grad_norm": 5.566794452728763, + "learning_rate": 1.8771997329223425e-05, + "loss": 1.4637, + "step": 1701 + }, + { + "epoch": 0.9239956568946797, + "grad_norm": 4.946565856604512, + "learning_rate": 1.8770308262019733e-05, + "loss": 0.8093, + "step": 1702 + }, + { + "epoch": 0.9245385450597177, + "grad_norm": 4.767566465605265, + "learning_rate": 1.8768618110090334e-05, + "loss": 0.5938, + "step": 1703 + }, + { + "epoch": 0.9250814332247557, + "grad_norm": 6.919371263621627, + "learning_rate": 1.8766926873644272e-05, + "loss": 0.9857, + "step": 1704 + }, + { + "epoch": 0.9256243213897937, + "grad_norm": 3.5297488420709944, + "learning_rate": 1.876523455289072e-05, + "loss": 0.8744, + "step": 1705 + }, + { + "epoch": 0.9261672095548317, + "grad_norm": 4.910386600622229, + "learning_rate": 1.8763541148038994e-05, + "loss": 0.9491, + "step": 1706 + }, + { + "epoch": 0.9267100977198697, + "grad_norm": 5.550892558363115, + "learning_rate": 1.876184665929853e-05, + "loss": 1.2793, + "step": 1707 + }, + { + "epoch": 0.9272529858849077, + "grad_norm": 4.470721855980947, + "learning_rate": 1.8760151086878905e-05, + "loss": 0.9706, + "step": 1708 + }, + { + "epoch": 0.9277958740499457, + "grad_norm": 7.1893201073258926, + "learning_rate": 1.8758454430989833e-05, + "loss": 1.2473, + "step": 1709 + }, + { + "epoch": 0.9283387622149837, + "grad_norm": 5.933304434057961, + "learning_rate": 1.875675669184116e-05, + "loss": 1.4917, + "step": 1710 + }, + { + "epoch": 0.9288816503800217, + "grad_norm": 5.386119916834561, + "learning_rate": 1.8755057869642857e-05, + "loss": 0.9417, + "step": 1711 + }, + { + "epoch": 0.9294245385450597, + "grad_norm": 6.121437468126703, + "learning_rate": 1.875335796460505e-05, + "loss": 1.203, + "step": 1712 + }, + { + "epoch": 0.9299674267100977, + "grad_norm": 5.34957970840911, + "learning_rate": 1.8751656976937974e-05, + "loss": 1.2557, + "step": 1713 + }, + { + "epoch": 0.9305103148751357, + "grad_norm": 7.404283054108375, + "learning_rate": 1.8749954906852023e-05, + "loss": 1.449, + "step": 1714 + }, + { + "epoch": 0.9310532030401737, + "grad_norm": 6.5511839319022585, + "learning_rate": 1.8748251754557696e-05, + "loss": 1.5548, + "step": 1715 + }, + { + "epoch": 0.9315960912052117, + "grad_norm": 5.0254557279034815, + "learning_rate": 1.8746547520265654e-05, + "loss": 0.9777, + "step": 1716 + }, + { + "epoch": 0.9321389793702497, + "grad_norm": 6.385252198006408, + "learning_rate": 1.874484220418667e-05, + "loss": 1.254, + "step": 1717 + }, + { + "epoch": 0.9326818675352877, + "grad_norm": 6.872281478717196, + "learning_rate": 1.874313580653167e-05, + "loss": 1.278, + "step": 1718 + }, + { + "epoch": 0.9332247557003257, + "grad_norm": 8.525842687889295, + "learning_rate": 1.8741428327511696e-05, + "loss": 1.3712, + "step": 1719 + }, + { + "epoch": 0.9337676438653637, + "grad_norm": 7.97420765579221, + "learning_rate": 1.8739719767337933e-05, + "loss": 1.7493, + "step": 1720 + }, + { + "epoch": 0.9343105320304017, + "grad_norm": 6.107109526814546, + "learning_rate": 1.8738010126221705e-05, + "loss": 0.9861, + "step": 1721 + }, + { + "epoch": 0.9348534201954397, + "grad_norm": 6.4666849584489645, + "learning_rate": 1.8736299404374453e-05, + "loss": 0.8041, + "step": 1722 + }, + { + "epoch": 0.9353963083604777, + "grad_norm": 7.489590627219774, + "learning_rate": 1.873458760200777e-05, + "loss": 1.2325, + "step": 1723 + }, + { + "epoch": 0.9359391965255157, + "grad_norm": 7.020182643385788, + "learning_rate": 1.8732874719333373e-05, + "loss": 1.1698, + "step": 1724 + }, + { + "epoch": 0.9364820846905537, + "grad_norm": 6.436449840675483, + "learning_rate": 1.873116075656311e-05, + "loss": 1.084, + "step": 1725 + }, + { + "epoch": 0.9370249728555917, + "grad_norm": 7.27100690341834, + "learning_rate": 1.872944571390897e-05, + "loss": 0.9793, + "step": 1726 + }, + { + "epoch": 0.9375678610206297, + "grad_norm": 6.337617737677888, + "learning_rate": 1.872772959158307e-05, + "loss": 1.0558, + "step": 1727 + }, + { + "epoch": 0.9381107491856677, + "grad_norm": 5.356493873446841, + "learning_rate": 1.8726012389797667e-05, + "loss": 1.1031, + "step": 1728 + }, + { + "epoch": 0.9386536373507057, + "grad_norm": 6.302965546816682, + "learning_rate": 1.8724294108765142e-05, + "loss": 1.1401, + "step": 1729 + }, + { + "epoch": 0.9391965255157437, + "grad_norm": 6.116650787967334, + "learning_rate": 1.872257474869802e-05, + "loss": 1.0848, + "step": 1730 + }, + { + "epoch": 0.9397394136807817, + "grad_norm": 4.3874746967683, + "learning_rate": 1.8720854309808948e-05, + "loss": 0.7463, + "step": 1731 + }, + { + "epoch": 0.9402823018458197, + "grad_norm": 5.00673534942558, + "learning_rate": 1.871913279231072e-05, + "loss": 1.2365, + "step": 1732 + }, + { + "epoch": 0.9408251900108577, + "grad_norm": 5.222845312743783, + "learning_rate": 1.871741019641625e-05, + "loss": 1.062, + "step": 1733 + }, + { + "epoch": 0.9413680781758957, + "grad_norm": 4.624403224726164, + "learning_rate": 1.871568652233859e-05, + "loss": 1.0155, + "step": 1734 + }, + { + "epoch": 0.9419109663409337, + "grad_norm": 5.386412889478532, + "learning_rate": 1.8713961770290936e-05, + "loss": 1.4316, + "step": 1735 + }, + { + "epoch": 0.9424538545059717, + "grad_norm": 6.0899860043030705, + "learning_rate": 1.87122359404866e-05, + "loss": 1.1372, + "step": 1736 + }, + { + "epoch": 0.9429967426710097, + "grad_norm": 5.306362358042622, + "learning_rate": 1.8710509033139037e-05, + "loss": 0.9353, + "step": 1737 + }, + { + "epoch": 0.9435396308360477, + "grad_norm": 6.479420699028343, + "learning_rate": 1.8708781048461832e-05, + "loss": 1.2435, + "step": 1738 + }, + { + "epoch": 0.9440825190010858, + "grad_norm": 6.205746500302448, + "learning_rate": 1.8707051986668712e-05, + "loss": 0.7872, + "step": 1739 + }, + { + "epoch": 0.9446254071661238, + "grad_norm": 4.157411034856641, + "learning_rate": 1.8705321847973523e-05, + "loss": 0.8004, + "step": 1740 + }, + { + "epoch": 0.9451682953311618, + "grad_norm": 5.5814599600578285, + "learning_rate": 1.8703590632590254e-05, + "loss": 0.7384, + "step": 1741 + }, + { + "epoch": 0.9457111834961998, + "grad_norm": 6.040759722497843, + "learning_rate": 1.8701858340733023e-05, + "loss": 1.5941, + "step": 1742 + }, + { + "epoch": 0.9462540716612378, + "grad_norm": 4.592040464655725, + "learning_rate": 1.8700124972616085e-05, + "loss": 1.0662, + "step": 1743 + }, + { + "epoch": 0.9467969598262758, + "grad_norm": 6.077319934267432, + "learning_rate": 1.8698390528453823e-05, + "loss": 1.3726, + "step": 1744 + }, + { + "epoch": 0.9473398479913138, + "grad_norm": 4.255180894541495, + "learning_rate": 1.869665500846076e-05, + "loss": 0.7001, + "step": 1745 + }, + { + "epoch": 0.9478827361563518, + "grad_norm": 6.390593673413393, + "learning_rate": 1.869491841285154e-05, + "loss": 1.2268, + "step": 1746 + }, + { + "epoch": 0.9484256243213898, + "grad_norm": 5.347324129941674, + "learning_rate": 1.8693180741840957e-05, + "loss": 1.2193, + "step": 1747 + }, + { + "epoch": 0.9489685124864278, + "grad_norm": 6.407261723862006, + "learning_rate": 1.8691441995643927e-05, + "loss": 0.868, + "step": 1748 + }, + { + "epoch": 0.9495114006514658, + "grad_norm": 4.400520878629861, + "learning_rate": 1.8689702174475496e-05, + "loss": 0.4679, + "step": 1749 + }, + { + "epoch": 0.9500542888165038, + "grad_norm": 5.6161548427455505, + "learning_rate": 1.8687961278550852e-05, + "loss": 0.8842, + "step": 1750 + }, + { + "epoch": 0.9505971769815418, + "grad_norm": 4.7163538240048, + "learning_rate": 1.8686219308085306e-05, + "loss": 0.9972, + "step": 1751 + }, + { + "epoch": 0.9511400651465798, + "grad_norm": 4.845854095585142, + "learning_rate": 1.8684476263294318e-05, + "loss": 1.1046, + "step": 1752 + }, + { + "epoch": 0.9516829533116178, + "grad_norm": 3.878984502983999, + "learning_rate": 1.8682732144393463e-05, + "loss": 0.7873, + "step": 1753 + }, + { + "epoch": 0.9522258414766558, + "grad_norm": 5.736472248757515, + "learning_rate": 1.8680986951598458e-05, + "loss": 1.2046, + "step": 1754 + }, + { + "epoch": 0.9527687296416938, + "grad_norm": 5.12988223940441, + "learning_rate": 1.867924068512515e-05, + "loss": 0.6293, + "step": 1755 + }, + { + "epoch": 0.9533116178067318, + "grad_norm": 7.822467968073818, + "learning_rate": 1.867749334518952e-05, + "loss": 1.3625, + "step": 1756 + }, + { + "epoch": 0.9538545059717698, + "grad_norm": 5.488388947029871, + "learning_rate": 1.8675744932007687e-05, + "loss": 1.078, + "step": 1757 + }, + { + "epoch": 0.9543973941368078, + "grad_norm": 4.631848438676548, + "learning_rate": 1.8673995445795894e-05, + "loss": 0.69, + "step": 1758 + }, + { + "epoch": 0.9549402823018458, + "grad_norm": 7.078171019993124, + "learning_rate": 1.8672244886770516e-05, + "loss": 1.1036, + "step": 1759 + }, + { + "epoch": 0.9554831704668838, + "grad_norm": 5.9259972832349215, + "learning_rate": 1.8670493255148073e-05, + "loss": 0.7919, + "step": 1760 + }, + { + "epoch": 0.9560260586319218, + "grad_norm": 5.846343833877514, + "learning_rate": 1.8668740551145205e-05, + "loss": 1.1653, + "step": 1761 + }, + { + "epoch": 0.9565689467969598, + "grad_norm": 5.559082125902064, + "learning_rate": 1.8666986774978685e-05, + "loss": 1.4214, + "step": 1762 + }, + { + "epoch": 0.9571118349619978, + "grad_norm": 6.511207981127819, + "learning_rate": 1.8665231926865433e-05, + "loss": 1.3552, + "step": 1763 + }, + { + "epoch": 0.9576547231270358, + "grad_norm": 5.728999474717542, + "learning_rate": 1.8663476007022482e-05, + "loss": 0.7861, + "step": 1764 + }, + { + "epoch": 0.9581976112920738, + "grad_norm": 6.152118564149309, + "learning_rate": 1.8661719015667016e-05, + "loss": 0.8679, + "step": 1765 + }, + { + "epoch": 0.9587404994571118, + "grad_norm": 6.77121947972117, + "learning_rate": 1.8659960953016334e-05, + "loss": 0.8103, + "step": 1766 + }, + { + "epoch": 0.9592833876221498, + "grad_norm": 7.822575173268641, + "learning_rate": 1.865820181928788e-05, + "loss": 1.2558, + "step": 1767 + }, + { + "epoch": 0.9598262757871878, + "grad_norm": 6.389073592671253, + "learning_rate": 1.8656441614699225e-05, + "loss": 1.0314, + "step": 1768 + }, + { + "epoch": 0.9603691639522258, + "grad_norm": 6.141689972496849, + "learning_rate": 1.8654680339468076e-05, + "loss": 1.2452, + "step": 1769 + }, + { + "epoch": 0.9609120521172638, + "grad_norm": 6.934092850250001, + "learning_rate": 1.8652917993812267e-05, + "loss": 1.1186, + "step": 1770 + }, + { + "epoch": 0.9614549402823018, + "grad_norm": 6.129320663785202, + "learning_rate": 1.865115457794977e-05, + "loss": 0.9076, + "step": 1771 + }, + { + "epoch": 0.9619978284473398, + "grad_norm": 7.008261896063745, + "learning_rate": 1.8649390092098693e-05, + "loss": 1.7152, + "step": 1772 + }, + { + "epoch": 0.9625407166123778, + "grad_norm": 6.36833090372477, + "learning_rate": 1.8647624536477255e-05, + "loss": 0.8087, + "step": 1773 + }, + { + "epoch": 0.9630836047774158, + "grad_norm": 4.7595903392838865, + "learning_rate": 1.8645857911303838e-05, + "loss": 0.6374, + "step": 1774 + }, + { + "epoch": 0.9636264929424538, + "grad_norm": 6.410325139984648, + "learning_rate": 1.8644090216796934e-05, + "loss": 1.3611, + "step": 1775 + }, + { + "epoch": 0.9641693811074918, + "grad_norm": 5.900212705071564, + "learning_rate": 1.8642321453175177e-05, + "loss": 0.891, + "step": 1776 + }, + { + "epoch": 0.9647122692725298, + "grad_norm": 6.90562183471938, + "learning_rate": 1.8640551620657326e-05, + "loss": 0.7899, + "step": 1777 + }, + { + "epoch": 0.9652551574375678, + "grad_norm": 6.476066490933977, + "learning_rate": 1.8638780719462278e-05, + "loss": 0.9614, + "step": 1778 + }, + { + "epoch": 0.9657980456026058, + "grad_norm": 6.659586852509603, + "learning_rate": 1.8637008749809065e-05, + "loss": 1.2419, + "step": 1779 + }, + { + "epoch": 0.9663409337676439, + "grad_norm": 6.587754652016677, + "learning_rate": 1.8635235711916847e-05, + "loss": 1.3664, + "step": 1780 + }, + { + "epoch": 0.9668838219326819, + "grad_norm": 6.558578755813245, + "learning_rate": 1.863346160600491e-05, + "loss": 1.1249, + "step": 1781 + }, + { + "epoch": 0.9674267100977199, + "grad_norm": 4.791301716076818, + "learning_rate": 1.8631686432292685e-05, + "loss": 0.6836, + "step": 1782 + }, + { + "epoch": 0.9679695982627579, + "grad_norm": 5.120778292537024, + "learning_rate": 1.862991019099972e-05, + "loss": 0.592, + "step": 1783 + }, + { + "epoch": 0.9685124864277959, + "grad_norm": 8.250755784410368, + "learning_rate": 1.8628132882345713e-05, + "loss": 1.2579, + "step": 1784 + }, + { + "epoch": 0.9690553745928339, + "grad_norm": 5.811263926579046, + "learning_rate": 1.862635450655048e-05, + "loss": 1.2935, + "step": 1785 + }, + { + "epoch": 0.9695982627578719, + "grad_norm": 6.042040633945999, + "learning_rate": 1.862457506383397e-05, + "loss": 0.8784, + "step": 1786 + }, + { + "epoch": 0.9701411509229099, + "grad_norm": 7.115011060082062, + "learning_rate": 1.8622794554416272e-05, + "loss": 1.0717, + "step": 1787 + }, + { + "epoch": 0.9706840390879479, + "grad_norm": 7.520775839533288, + "learning_rate": 1.8621012978517604e-05, + "loss": 1.3468, + "step": 1788 + }, + { + "epoch": 0.9712269272529859, + "grad_norm": 6.949921961354188, + "learning_rate": 1.8619230336358306e-05, + "loss": 1.5609, + "step": 1789 + }, + { + "epoch": 0.9717698154180239, + "grad_norm": 6.396218455762339, + "learning_rate": 1.8617446628158866e-05, + "loss": 1.1807, + "step": 1790 + }, + { + "epoch": 0.9723127035830619, + "grad_norm": 6.314794279037124, + "learning_rate": 1.861566185413989e-05, + "loss": 0.9255, + "step": 1791 + }, + { + "epoch": 0.9728555917480999, + "grad_norm": 5.393752442926871, + "learning_rate": 1.8613876014522128e-05, + "loss": 0.7926, + "step": 1792 + }, + { + "epoch": 0.9733984799131379, + "grad_norm": 5.908313693688933, + "learning_rate": 1.8612089109526453e-05, + "loss": 0.7984, + "step": 1793 + }, + { + "epoch": 0.9739413680781759, + "grad_norm": 4.6952509401002125, + "learning_rate": 1.8610301139373867e-05, + "loss": 1.0344, + "step": 1794 + }, + { + "epoch": 0.9744842562432139, + "grad_norm": 6.81835335105175, + "learning_rate": 1.8608512104285517e-05, + "loss": 1.1532, + "step": 1795 + }, + { + "epoch": 0.9750271444082519, + "grad_norm": 4.159411114912345, + "learning_rate": 1.860672200448267e-05, + "loss": 1.0665, + "step": 1796 + }, + { + "epoch": 0.9755700325732899, + "grad_norm": 4.911404446825004, + "learning_rate": 1.8604930840186726e-05, + "loss": 1.1784, + "step": 1797 + }, + { + "epoch": 0.9761129207383279, + "grad_norm": 5.399453855489055, + "learning_rate": 1.860313861161922e-05, + "loss": 0.8215, + "step": 1798 + }, + { + "epoch": 0.9766558089033659, + "grad_norm": 4.797137239148531, + "learning_rate": 1.860134531900182e-05, + "loss": 0.6405, + "step": 1799 + }, + { + "epoch": 0.9771986970684039, + "grad_norm": 3.937076316066272, + "learning_rate": 1.859955096255633e-05, + "loss": 0.5912, + "step": 1800 + }, + { + "epoch": 0.9777415852334419, + "grad_norm": 5.569110777302129, + "learning_rate": 1.859775554250466e-05, + "loss": 0.8089, + "step": 1801 + }, + { + "epoch": 0.9782844733984799, + "grad_norm": 6.364258901894282, + "learning_rate": 1.859595905906889e-05, + "loss": 1.5181, + "step": 1802 + }, + { + "epoch": 0.9788273615635179, + "grad_norm": 4.995224609073908, + "learning_rate": 1.85941615124712e-05, + "loss": 0.8573, + "step": 1803 + }, + { + "epoch": 0.9793702497285559, + "grad_norm": 5.041992740138782, + "learning_rate": 1.8592362902933918e-05, + "loss": 0.8982, + "step": 1804 + }, + { + "epoch": 0.9799131378935939, + "grad_norm": 5.6054574490953275, + "learning_rate": 1.8590563230679496e-05, + "loss": 0.9285, + "step": 1805 + }, + { + "epoch": 0.9804560260586319, + "grad_norm": 5.269503319178264, + "learning_rate": 1.8588762495930526e-05, + "loss": 1.0963, + "step": 1806 + }, + { + "epoch": 0.9809989142236699, + "grad_norm": 6.184787731593613, + "learning_rate": 1.8586960698909718e-05, + "loss": 1.4005, + "step": 1807 + }, + { + "epoch": 0.9815418023887079, + "grad_norm": 8.50824591479003, + "learning_rate": 1.858515783983993e-05, + "loss": 1.9607, + "step": 1808 + }, + { + "epoch": 0.9820846905537459, + "grad_norm": 5.399464732589782, + "learning_rate": 1.8583353918944134e-05, + "loss": 0.951, + "step": 1809 + }, + { + "epoch": 0.9826275787187839, + "grad_norm": 6.577634699737779, + "learning_rate": 1.8581548936445447e-05, + "loss": 1.5139, + "step": 1810 + }, + { + "epoch": 0.9831704668838219, + "grad_norm": 10.183489163585788, + "learning_rate": 1.8579742892567107e-05, + "loss": 2.0539, + "step": 1811 + }, + { + "epoch": 0.9837133550488599, + "grad_norm": 6.578605968943458, + "learning_rate": 1.8577935787532494e-05, + "loss": 1.1454, + "step": 1812 + }, + { + "epoch": 0.9842562432138979, + "grad_norm": 4.923285404215955, + "learning_rate": 1.8576127621565113e-05, + "loss": 0.8965, + "step": 1813 + }, + { + "epoch": 0.9847991313789359, + "grad_norm": 5.912445932400008, + "learning_rate": 1.85743183948886e-05, + "loss": 1.1311, + "step": 1814 + }, + { + "epoch": 0.9853420195439739, + "grad_norm": 4.5183231221646105, + "learning_rate": 1.8572508107726725e-05, + "loss": 0.7511, + "step": 1815 + }, + { + "epoch": 0.9858849077090119, + "grad_norm": 5.2218543332097935, + "learning_rate": 1.8570696760303378e-05, + "loss": 1.0063, + "step": 1816 + }, + { + "epoch": 0.9864277958740499, + "grad_norm": 7.215748803707923, + "learning_rate": 1.85688843528426e-05, + "loss": 1.3674, + "step": 1817 + }, + { + "epoch": 0.9869706840390879, + "grad_norm": 6.362354074709609, + "learning_rate": 1.8567070885568547e-05, + "loss": 1.1319, + "step": 1818 + }, + { + "epoch": 0.987513572204126, + "grad_norm": 4.680755706450828, + "learning_rate": 1.8565256358705513e-05, + "loss": 0.6613, + "step": 1819 + }, + { + "epoch": 0.988056460369164, + "grad_norm": 4.80241945107321, + "learning_rate": 1.8563440772477922e-05, + "loss": 0.8619, + "step": 1820 + }, + { + "epoch": 0.988599348534202, + "grad_norm": 5.858709563059296, + "learning_rate": 1.856162412711033e-05, + "loss": 0.9432, + "step": 1821 + }, + { + "epoch": 0.98914223669924, + "grad_norm": 6.478119299588854, + "learning_rate": 1.855980642282742e-05, + "loss": 1.1866, + "step": 1822 + }, + { + "epoch": 0.989685124864278, + "grad_norm": 7.081924554034699, + "learning_rate": 1.8557987659854006e-05, + "loss": 1.3657, + "step": 1823 + }, + { + "epoch": 0.990228013029316, + "grad_norm": 5.404410220733361, + "learning_rate": 1.855616783841504e-05, + "loss": 1.1051, + "step": 1824 + }, + { + "epoch": 0.990770901194354, + "grad_norm": 6.742207962837618, + "learning_rate": 1.8554346958735602e-05, + "loss": 1.4667, + "step": 1825 + }, + { + "epoch": 0.991313789359392, + "grad_norm": 6.164178667211635, + "learning_rate": 1.8552525021040895e-05, + "loss": 0.7861, + "step": 1826 + }, + { + "epoch": 0.99185667752443, + "grad_norm": 6.078017237490637, + "learning_rate": 1.8550702025556265e-05, + "loss": 1.3294, + "step": 1827 + }, + { + "epoch": 0.992399565689468, + "grad_norm": 4.7372949568909455, + "learning_rate": 1.8548877972507182e-05, + "loss": 0.9779, + "step": 1828 + }, + { + "epoch": 0.992942453854506, + "grad_norm": 5.25971277777658, + "learning_rate": 1.8547052862119247e-05, + "loss": 1.049, + "step": 1829 + }, + { + "epoch": 0.993485342019544, + "grad_norm": 4.950709970125391, + "learning_rate": 1.854522669461819e-05, + "loss": 0.6665, + "step": 1830 + }, + { + "epoch": 0.994028230184582, + "grad_norm": 5.856428194694733, + "learning_rate": 1.8543399470229876e-05, + "loss": 1.0954, + "step": 1831 + }, + { + "epoch": 0.99457111834962, + "grad_norm": 4.895286613591861, + "learning_rate": 1.85415711891803e-05, + "loss": 0.7069, + "step": 1832 + }, + { + "epoch": 0.995114006514658, + "grad_norm": 5.249228502887773, + "learning_rate": 1.8539741851695586e-05, + "loss": 1.0655, + "step": 1833 + }, + { + "epoch": 0.995656894679696, + "grad_norm": 4.855694799243912, + "learning_rate": 1.8537911458001988e-05, + "loss": 1.1872, + "step": 1834 + }, + { + "epoch": 0.996199782844734, + "grad_norm": 5.180821237913711, + "learning_rate": 1.8536080008325896e-05, + "loss": 0.8281, + "step": 1835 + }, + { + "epoch": 0.996742671009772, + "grad_norm": 5.073254295771654, + "learning_rate": 1.8534247502893823e-05, + "loss": 0.7002, + "step": 1836 + }, + { + "epoch": 0.99728555917481, + "grad_norm": 5.4022604097623255, + "learning_rate": 1.8532413941932416e-05, + "loss": 1.1793, + "step": 1837 + }, + { + "epoch": 0.997828447339848, + "grad_norm": 4.981761324197401, + "learning_rate": 1.8530579325668455e-05, + "loss": 0.8331, + "step": 1838 + }, + { + "epoch": 0.998371335504886, + "grad_norm": 7.325363733761658, + "learning_rate": 1.852874365432885e-05, + "loss": 1.3071, + "step": 1839 + }, + { + "epoch": 0.998914223669924, + "grad_norm": 7.5668251202705505, + "learning_rate": 1.852690692814063e-05, + "loss": 1.4955, + "step": 1840 + }, + { + "epoch": 0.999457111834962, + "grad_norm": 6.706936559053067, + "learning_rate": 1.8525069147330978e-05, + "loss": 1.0286, + "step": 1841 + }, + { + "epoch": 1.0, + "grad_norm": 6.338369283394405, + "learning_rate": 1.8523230312127183e-05, + "loss": 1.0509, + "step": 1842 + } + ], + "logging_steps": 1.0, + "max_steps": 9210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 2197804491472896.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/training_args.bin b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b16de63f342ea53d59c4d97aa8a3bd463a79531f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9801fdd09c6ba798e12b77f77221db3b5747b6e1296785789d44cf2bc492ff9e +size 7160 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/zero_to_fp32.py b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-1842/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/README.md b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/README.md new file mode 100644 index 0000000000000000000000000000000000000000..972a036203d877262d3c6673f4d81814e7409dc5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/README.md @@ -0,0 +1,202 @@ +--- +base_model: llava-hf/LLaVA-NeXT-Video-7B-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/adapter_config.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..719f6e47b29561ff93d519680ccd115b480d825e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/adapter_config.json @@ -0,0 +1,132 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "llava-hf/LLaVA-NeXT-Video-7B-hf", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 8, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [], + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "model.layers.17.self_attn.q_proj", + "model.layers.17.self_attn.k_proj", + "model.layers.13.self_attn.k_proj", + "model.layers.20.self_attn.q_proj", + "model.layers.18.self_attn.k_proj", + "29.self_attn.k_proj", + "model.layers.16.self_attn.v_proj", + "model.layers.14.self_attn.q_proj", + "29.self_attn.v_proj", + "24.self_attn.k_proj", + "model.layers.18.self_attn.v_proj", + "model.layers.5.self_attn.q_proj", + "model.layers.11.self_attn.k_proj", + "27.self_attn.v_proj", + "model.layers.16.self_attn.q_proj", + "model.layers.10.self_attn.q_proj", + "model.layers.12.self_attn.v_proj", + "model.layers.20.self_attn.v_proj", + "model.layers.0.self_attn.v_proj", + "model.layers.7.self_attn.q_proj", + "model.layers.23.self_attn.q_proj", + "model.layers.1.self_attn.q_proj", + "25.self_attn.v_proj", + "model.layers.4.self_attn.v_proj", + "model.layers.19.self_attn.q_proj", + "model.layers.10.self_attn.k_proj", + "model.layers.22.self_attn.k_proj", + "model.layers.14.self_attn.v_proj", + "model.layers.13.self_attn.v_proj", + "25.self_attn.k_proj", + "24.self_attn.v_proj", + "model.layers.21.self_attn.q_proj", + "model.layers.8.self_attn.q_proj", + "model.layers.6.self_attn.q_proj", + "model.layers.6.self_attn.k_proj", + "model.layers.1.self_attn.v_proj", + "model.layers.9.self_attn.k_proj", + "model.layers.13.self_attn.q_proj", + "o_proj", + "model.layers.16.self_attn.k_proj", + "model.layers.6.self_attn.v_proj", + "model.layers.21.self_attn.k_proj", + "model.layers.5.self_attn.k_proj", + "27.self_attn.q_proj", + "model.layers.20.self_attn.k_proj", + "model.layers.11.self_attn.v_proj", + "model.layers.22.self_attn.q_proj", + "model.layers.14.self_attn.k_proj", + "28.self_attn.v_proj", + "model.layers.22.self_attn.v_proj", + "gate_proj", + "model.layers.0.self_attn.k_proj", + "model.layers.5.self_attn.v_proj", + "30.self_attn.q_proj", + "model.layers.3.self_attn.k_proj", + "29.self_attn.q_proj", + "model.layers.19.self_attn.k_proj", + "model.layers.10.self_attn.v_proj", + "model.layers.15.self_attn.k_proj", + "model.layers.7.self_attn.v_proj", + "model.layers.9.self_attn.v_proj", + "up_proj", + "model.layers.4.self_attn.q_proj", + "28.self_attn.q_proj", + "model.layers.3.self_attn.q_proj", + "24.self_attn.q_proj", + "30.self_attn.v_proj", + "model.layers.12.self_attn.k_proj", + "model.layers.23.self_attn.k_proj", + "26.self_attn.q_proj", + "28.self_attn.k_proj", + "31.self_attn.k_proj", + "down_proj", + "model.layers.2.self_attn.q_proj", + "31.self_attn.q_proj", + "model.layers.15.self_attn.q_proj", + "model.layers.12.self_attn.q_proj", + "model.layers.0.self_attn.q_proj", + "model.layers.2.self_attn.k_proj", + "model.layers.3.self_attn.v_proj", + "model.layers.11.self_attn.q_proj", + "25.self_attn.q_proj", + "model.layers.23.self_attn.v_proj", + "30.self_attn.k_proj", + "31.self_attn.v_proj", + "model.layers.15.self_attn.v_proj", + "model.layers.8.self_attn.v_proj", + "26.self_attn.k_proj", + "model.layers.19.self_attn.v_proj", + "model.layers.2.self_attn.v_proj", + "26.self_attn.v_proj", + "model.layers.21.self_attn.v_proj", + "model.layers.17.self_attn.v_proj", + "model.layers.8.self_attn.k_proj", + "model.layers.18.self_attn.q_proj", + "model.layers.7.self_attn.k_proj", + "model.layers.4.self_attn.k_proj", + "model.layers.9.self_attn.q_proj", + "27.self_attn.k_proj", + "model.layers.1.self_attn.k_proj" + ], + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/adapter_model.safetensors b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..494a64df36e4cebd13dca1992c16abce484d6e8e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:865f8810eb39ade0d7fc007d1da35a026c6dfc33d96d9a4392c7e763dd78e6c8 +size 40043208 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..18a6ee180a6f0ee4f44714eead2508963aab351f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1aad915d20ebccd810c3022fbaf55426eb605b56f9c6da5807d4a310df6ac246 +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..1200a2a99506c8447c65d9adbd8ea947901c0743 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:440ec08fd635f88a58f7cc20bc1c1b55f8767c1dc84b22ded2a84255e39e1980 +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/zero_pp_rank_0_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/zero_pp_rank_0_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..37d30d702a2ce51437d81a5c48de81e17b98a2b6 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/zero_pp_rank_0_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:694f6200cdf5d8b137d83f259c71ce8297934d2d596bc232ee0bff8fcff7f49a +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/zero_pp_rank_1_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/zero_pp_rank_1_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..6a30be2b7f69b04e819766c4fbf7135fdf88e644 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/global_step3684/zero_pp_rank_1_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5b8624ce14de8894eef0286de4915ab727a0146f7d58b1fbc8112e43a0005fe +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/latest b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/latest new file mode 100644 index 0000000000000000000000000000000000000000..797a9e2b17b330cc789c3c4984ab2b1b241826d6 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/latest @@ -0,0 +1 @@ +global_step3684 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/rng_state_0.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..de770597ee778fd4322e4d609744d0b02e4670c2 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:663521896818e0e21ea6fc67318759584b5f18e2bb89d80e231d7a0cbe121787 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/rng_state_1.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..a793161fdf717f3e143b8ffc5e033a0a47536705 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f9b2dfac9c04072e8cbab56109ba83aaabc1f20a759819c45f3b82c5ff5c9c6 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/scheduler.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..36e7a5ba60c8a711e46456d81dc617adc3c45b1c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2a87aa7b96e976e81d4bb8ba0c1c0c48eed8f8f3d616170afb74bb57e2dde58 +size 1064 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/trainer_state.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..34f48bf5b7366fb2e4058dbc03deccf5eb434f30 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/trainer_state.json @@ -0,0 +1,25821 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 3684, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005428881650380022, + "grad_norm": 2.1320506281889067, + "learning_rate": 7.220216606498195e-08, + "loss": 5.1936, + "step": 1 + }, + { + "epoch": 0.0010857763300760044, + "grad_norm": 1.8633488766761366, + "learning_rate": 1.444043321299639e-07, + "loss": 4.6995, + "step": 2 + }, + { + "epoch": 0.0016286644951140066, + "grad_norm": 2.063286098879776, + "learning_rate": 2.1660649819494586e-07, + "loss": 4.4902, + "step": 3 + }, + { + "epoch": 0.002171552660152009, + "grad_norm": 1.7123050230907495, + "learning_rate": 2.888086642599278e-07, + "loss": 4.9125, + "step": 4 + }, + { + "epoch": 0.0027144408251900108, + "grad_norm": 1.7499657276799891, + "learning_rate": 3.610108303249098e-07, + "loss": 3.7648, + "step": 5 + }, + { + "epoch": 0.003257328990228013, + "grad_norm": 2.0419045636066886, + "learning_rate": 4.332129963898917e-07, + "loss": 4.7842, + "step": 6 + }, + { + "epoch": 0.003800217155266015, + "grad_norm": 1.4960282529105982, + "learning_rate": 5.054151624548736e-07, + "loss": 4.9333, + "step": 7 + }, + { + "epoch": 0.004343105320304018, + "grad_norm": 1.640954967476452, + "learning_rate": 5.776173285198556e-07, + "loss": 3.9576, + "step": 8 + }, + { + "epoch": 0.004885993485342019, + "grad_norm": 2.0590500459610244, + "learning_rate": 6.498194945848375e-07, + "loss": 4.3833, + "step": 9 + }, + { + "epoch": 0.0054288816503800215, + "grad_norm": 1.07709353257816, + "learning_rate": 7.220216606498196e-07, + "loss": 3.7669, + "step": 10 + }, + { + "epoch": 0.005971769815418024, + "grad_norm": 1.9506235655782926, + "learning_rate": 7.942238267148016e-07, + "loss": 4.5975, + "step": 11 + }, + { + "epoch": 0.006514657980456026, + "grad_norm": 1.411983096954528, + "learning_rate": 8.664259927797834e-07, + "loss": 3.9706, + "step": 12 + }, + { + "epoch": 0.007057546145494028, + "grad_norm": 1.7110053131090028, + "learning_rate": 9.386281588447654e-07, + "loss": 4.9081, + "step": 13 + }, + { + "epoch": 0.00760043431053203, + "grad_norm": 1.3689315757199692, + "learning_rate": 1.0108303249097473e-06, + "loss": 4.2426, + "step": 14 + }, + { + "epoch": 0.008143322475570033, + "grad_norm": 1.5612091671030481, + "learning_rate": 1.0830324909747293e-06, + "loss": 4.5144, + "step": 15 + }, + { + "epoch": 0.008686210640608035, + "grad_norm": 1.3725168953865754, + "learning_rate": 1.1552346570397112e-06, + "loss": 4.9378, + "step": 16 + }, + { + "epoch": 0.009229098805646038, + "grad_norm": 2.027359913808926, + "learning_rate": 1.2274368231046932e-06, + "loss": 3.7477, + "step": 17 + }, + { + "epoch": 0.009771986970684038, + "grad_norm": 2.1695456962160224, + "learning_rate": 1.299638989169675e-06, + "loss": 5.7636, + "step": 18 + }, + { + "epoch": 0.01031487513572204, + "grad_norm": 3.2637002287680428, + "learning_rate": 1.3718411552346572e-06, + "loss": 6.1883, + "step": 19 + }, + { + "epoch": 0.010857763300760043, + "grad_norm": 2.5723616120643857, + "learning_rate": 1.4440433212996392e-06, + "loss": 4.0267, + "step": 20 + }, + { + "epoch": 0.011400651465798045, + "grad_norm": 1.4473754577877487, + "learning_rate": 1.516245487364621e-06, + "loss": 4.526, + "step": 21 + }, + { + "epoch": 0.011943539630836048, + "grad_norm": 1.6098208731761, + "learning_rate": 1.5884476534296031e-06, + "loss": 4.5163, + "step": 22 + }, + { + "epoch": 0.01248642779587405, + "grad_norm": 2.899624682573202, + "learning_rate": 1.6606498194945849e-06, + "loss": 4.8188, + "step": 23 + }, + { + "epoch": 0.013029315960912053, + "grad_norm": 2.273571202553131, + "learning_rate": 1.7328519855595669e-06, + "loss": 5.0074, + "step": 24 + }, + { + "epoch": 0.013572204125950055, + "grad_norm": 1.1877323869171315, + "learning_rate": 1.805054151624549e-06, + "loss": 4.4331, + "step": 25 + }, + { + "epoch": 0.014115092290988056, + "grad_norm": 1.9365501019385494, + "learning_rate": 1.8772563176895308e-06, + "loss": 5.0525, + "step": 26 + }, + { + "epoch": 0.014657980456026058, + "grad_norm": 1.7710485953105986, + "learning_rate": 1.949458483754513e-06, + "loss": 4.1346, + "step": 27 + }, + { + "epoch": 0.01520086862106406, + "grad_norm": 1.231848881292397, + "learning_rate": 2.0216606498194946e-06, + "loss": 4.2431, + "step": 28 + }, + { + "epoch": 0.015743756786102063, + "grad_norm": 1.6372838462475754, + "learning_rate": 2.0938628158844768e-06, + "loss": 5.4317, + "step": 29 + }, + { + "epoch": 0.016286644951140065, + "grad_norm": 1.7326568199523718, + "learning_rate": 2.1660649819494585e-06, + "loss": 4.4774, + "step": 30 + }, + { + "epoch": 0.016829533116178068, + "grad_norm": 2.1315641946923387, + "learning_rate": 2.2382671480144407e-06, + "loss": 4.1418, + "step": 31 + }, + { + "epoch": 0.01737242128121607, + "grad_norm": 1.916538366799816, + "learning_rate": 2.3104693140794225e-06, + "loss": 5.7926, + "step": 32 + }, + { + "epoch": 0.017915309446254073, + "grad_norm": 2.057531274363787, + "learning_rate": 2.3826714801444047e-06, + "loss": 4.2862, + "step": 33 + }, + { + "epoch": 0.018458197611292075, + "grad_norm": 2.6727877745045023, + "learning_rate": 2.4548736462093864e-06, + "loss": 5.1323, + "step": 34 + }, + { + "epoch": 0.019001085776330078, + "grad_norm": 1.5043525736330898, + "learning_rate": 2.527075812274368e-06, + "loss": 4.0634, + "step": 35 + }, + { + "epoch": 0.019543973941368076, + "grad_norm": 2.6324248758076862, + "learning_rate": 2.59927797833935e-06, + "loss": 5.4211, + "step": 36 + }, + { + "epoch": 0.02008686210640608, + "grad_norm": 1.6581905646911415, + "learning_rate": 2.6714801444043326e-06, + "loss": 4.9825, + "step": 37 + }, + { + "epoch": 0.02062975027144408, + "grad_norm": 1.6667970880898455, + "learning_rate": 2.7436823104693144e-06, + "loss": 4.3219, + "step": 38 + }, + { + "epoch": 0.021172638436482084, + "grad_norm": 1.7448150724822071, + "learning_rate": 2.815884476534296e-06, + "loss": 4.3912, + "step": 39 + }, + { + "epoch": 0.021715526601520086, + "grad_norm": 3.720503777726443, + "learning_rate": 2.8880866425992783e-06, + "loss": 6.0592, + "step": 40 + }, + { + "epoch": 0.02225841476655809, + "grad_norm": 1.6853001074010774, + "learning_rate": 2.96028880866426e-06, + "loss": 3.8395, + "step": 41 + }, + { + "epoch": 0.02280130293159609, + "grad_norm": 1.797029067943667, + "learning_rate": 3.032490974729242e-06, + "loss": 4.4651, + "step": 42 + }, + { + "epoch": 0.023344191096634093, + "grad_norm": 1.6298824221707349, + "learning_rate": 3.1046931407942245e-06, + "loss": 4.4507, + "step": 43 + }, + { + "epoch": 0.023887079261672096, + "grad_norm": 1.4793189209059368, + "learning_rate": 3.1768953068592062e-06, + "loss": 4.7849, + "step": 44 + }, + { + "epoch": 0.024429967426710098, + "grad_norm": 1.1741346570472608, + "learning_rate": 3.249097472924188e-06, + "loss": 3.5897, + "step": 45 + }, + { + "epoch": 0.0249728555917481, + "grad_norm": 2.1468867306169064, + "learning_rate": 3.3212996389891698e-06, + "loss": 3.8969, + "step": 46 + }, + { + "epoch": 0.025515743756786103, + "grad_norm": 1.657494000697021, + "learning_rate": 3.393501805054152e-06, + "loss": 4.3412, + "step": 47 + }, + { + "epoch": 0.026058631921824105, + "grad_norm": 1.7125864055710684, + "learning_rate": 3.4657039711191337e-06, + "loss": 3.8059, + "step": 48 + }, + { + "epoch": 0.026601520086862108, + "grad_norm": 2.9515347000925622, + "learning_rate": 3.5379061371841155e-06, + "loss": 4.1069, + "step": 49 + }, + { + "epoch": 0.02714440825190011, + "grad_norm": 3.0185188918098262, + "learning_rate": 3.610108303249098e-06, + "loss": 5.2646, + "step": 50 + }, + { + "epoch": 0.02768729641693811, + "grad_norm": 1.5546972252207274, + "learning_rate": 3.68231046931408e-06, + "loss": 3.7212, + "step": 51 + }, + { + "epoch": 0.02823018458197611, + "grad_norm": 2.6376587426282474, + "learning_rate": 3.7545126353790616e-06, + "loss": 4.2862, + "step": 52 + }, + { + "epoch": 0.028773072747014114, + "grad_norm": 2.053905428064986, + "learning_rate": 3.826714801444043e-06, + "loss": 3.8304, + "step": 53 + }, + { + "epoch": 0.029315960912052116, + "grad_norm": 2.9083183656849725, + "learning_rate": 3.898916967509026e-06, + "loss": 4.8223, + "step": 54 + }, + { + "epoch": 0.02985884907709012, + "grad_norm": 1.9919239162425852, + "learning_rate": 3.971119133574007e-06, + "loss": 3.9054, + "step": 55 + }, + { + "epoch": 0.03040173724212812, + "grad_norm": 2.4076122411223135, + "learning_rate": 4.043321299638989e-06, + "loss": 4.8568, + "step": 56 + }, + { + "epoch": 0.030944625407166124, + "grad_norm": 3.854434602830178, + "learning_rate": 4.115523465703971e-06, + "loss": 5.7771, + "step": 57 + }, + { + "epoch": 0.031487513572204126, + "grad_norm": 2.510303564095041, + "learning_rate": 4.1877256317689535e-06, + "loss": 5.5851, + "step": 58 + }, + { + "epoch": 0.03203040173724213, + "grad_norm": 2.8884448366576305, + "learning_rate": 4.259927797833936e-06, + "loss": 4.1774, + "step": 59 + }, + { + "epoch": 0.03257328990228013, + "grad_norm": 2.0893778739721, + "learning_rate": 4.332129963898917e-06, + "loss": 4.4679, + "step": 60 + }, + { + "epoch": 0.03311617806731813, + "grad_norm": 1.5816439039579229, + "learning_rate": 4.404332129963899e-06, + "loss": 4.1999, + "step": 61 + }, + { + "epoch": 0.033659066232356136, + "grad_norm": 1.9591574137365761, + "learning_rate": 4.4765342960288814e-06, + "loss": 4.1539, + "step": 62 + }, + { + "epoch": 0.03420195439739414, + "grad_norm": 3.1533909413995924, + "learning_rate": 4.548736462093864e-06, + "loss": 4.5842, + "step": 63 + }, + { + "epoch": 0.03474484256243214, + "grad_norm": 2.543715437360663, + "learning_rate": 4.620938628158845e-06, + "loss": 4.1473, + "step": 64 + }, + { + "epoch": 0.03528773072747014, + "grad_norm": 2.7608093188531955, + "learning_rate": 4.693140794223827e-06, + "loss": 4.9354, + "step": 65 + }, + { + "epoch": 0.035830618892508145, + "grad_norm": 4.192163162864289, + "learning_rate": 4.765342960288809e-06, + "loss": 5.4263, + "step": 66 + }, + { + "epoch": 0.03637350705754615, + "grad_norm": 2.809206875027472, + "learning_rate": 4.837545126353791e-06, + "loss": 4.4259, + "step": 67 + }, + { + "epoch": 0.03691639522258415, + "grad_norm": 3.9318073942863534, + "learning_rate": 4.909747292418773e-06, + "loss": 4.1483, + "step": 68 + }, + { + "epoch": 0.03745928338762215, + "grad_norm": 2.935065385226121, + "learning_rate": 4.981949458483755e-06, + "loss": 4.6854, + "step": 69 + }, + { + "epoch": 0.038002171552660155, + "grad_norm": 2.3687140876301305, + "learning_rate": 5.054151624548736e-06, + "loss": 4.5875, + "step": 70 + }, + { + "epoch": 0.03854505971769816, + "grad_norm": 2.987709377109457, + "learning_rate": 5.126353790613719e-06, + "loss": 5.5784, + "step": 71 + }, + { + "epoch": 0.03908794788273615, + "grad_norm": 2.618514747422557, + "learning_rate": 5.1985559566787e-06, + "loss": 3.9901, + "step": 72 + }, + { + "epoch": 0.039630836047774155, + "grad_norm": 3.729239515680902, + "learning_rate": 5.270758122743683e-06, + "loss": 4.6192, + "step": 73 + }, + { + "epoch": 0.04017372421281216, + "grad_norm": 2.488658936956188, + "learning_rate": 5.342960288808665e-06, + "loss": 3.9889, + "step": 74 + }, + { + "epoch": 0.04071661237785016, + "grad_norm": 2.6055640979236223, + "learning_rate": 5.415162454873647e-06, + "loss": 4.3166, + "step": 75 + }, + { + "epoch": 0.04125950054288816, + "grad_norm": 2.091905540809132, + "learning_rate": 5.487364620938629e-06, + "loss": 3.5848, + "step": 76 + }, + { + "epoch": 0.041802388707926165, + "grad_norm": 4.153140315460812, + "learning_rate": 5.559566787003611e-06, + "loss": 4.4889, + "step": 77 + }, + { + "epoch": 0.04234527687296417, + "grad_norm": 4.828033123535034, + "learning_rate": 5.631768953068592e-06, + "loss": 4.886, + "step": 78 + }, + { + "epoch": 0.04288816503800217, + "grad_norm": 3.106146715385438, + "learning_rate": 5.7039711191335744e-06, + "loss": 4.3698, + "step": 79 + }, + { + "epoch": 0.04343105320304017, + "grad_norm": 4.458698539182903, + "learning_rate": 5.776173285198557e-06, + "loss": 4.7018, + "step": 80 + }, + { + "epoch": 0.043973941368078175, + "grad_norm": 4.619902873416999, + "learning_rate": 5.848375451263538e-06, + "loss": 4.3449, + "step": 81 + }, + { + "epoch": 0.04451682953311618, + "grad_norm": 2.411841615517987, + "learning_rate": 5.92057761732852e-06, + "loss": 3.9952, + "step": 82 + }, + { + "epoch": 0.04505971769815418, + "grad_norm": 3.6594837742651705, + "learning_rate": 5.992779783393502e-06, + "loss": 4.1308, + "step": 83 + }, + { + "epoch": 0.04560260586319218, + "grad_norm": 2.3760888273966074, + "learning_rate": 6.064981949458484e-06, + "loss": 4.0315, + "step": 84 + }, + { + "epoch": 0.046145494028230184, + "grad_norm": 4.579655616004927, + "learning_rate": 6.137184115523466e-06, + "loss": 4.1051, + "step": 85 + }, + { + "epoch": 0.04668838219326819, + "grad_norm": 3.5909339610827846, + "learning_rate": 6.209386281588449e-06, + "loss": 4.3126, + "step": 86 + }, + { + "epoch": 0.04723127035830619, + "grad_norm": 3.1030422711991625, + "learning_rate": 6.28158844765343e-06, + "loss": 4.3038, + "step": 87 + }, + { + "epoch": 0.04777415852334419, + "grad_norm": 4.134211324790428, + "learning_rate": 6.3537906137184125e-06, + "loss": 4.906, + "step": 88 + }, + { + "epoch": 0.048317046688382194, + "grad_norm": 3.538946913288816, + "learning_rate": 6.425992779783395e-06, + "loss": 4.2182, + "step": 89 + }, + { + "epoch": 0.048859934853420196, + "grad_norm": 1.9828152543749473, + "learning_rate": 6.498194945848376e-06, + "loss": 3.8487, + "step": 90 + }, + { + "epoch": 0.0494028230184582, + "grad_norm": 2.127969936938475, + "learning_rate": 6.570397111913358e-06, + "loss": 3.9221, + "step": 91 + }, + { + "epoch": 0.0499457111834962, + "grad_norm": 2.740653026382279, + "learning_rate": 6.6425992779783395e-06, + "loss": 4.4661, + "step": 92 + }, + { + "epoch": 0.050488599348534204, + "grad_norm": 3.278496788216742, + "learning_rate": 6.714801444043322e-06, + "loss": 4.1995, + "step": 93 + }, + { + "epoch": 0.051031487513572206, + "grad_norm": 1.5455749978381657, + "learning_rate": 6.787003610108304e-06, + "loss": 3.3941, + "step": 94 + }, + { + "epoch": 0.05157437567861021, + "grad_norm": 2.6938438745410975, + "learning_rate": 6.859205776173285e-06, + "loss": 3.9863, + "step": 95 + }, + { + "epoch": 0.05211726384364821, + "grad_norm": 2.754193515919841, + "learning_rate": 6.9314079422382674e-06, + "loss": 4.1621, + "step": 96 + }, + { + "epoch": 0.05266015200868621, + "grad_norm": 1.787757508585929, + "learning_rate": 7.00361010830325e-06, + "loss": 3.728, + "step": 97 + }, + { + "epoch": 0.053203040173724216, + "grad_norm": 1.6756343581979376, + "learning_rate": 7.075812274368231e-06, + "loss": 3.4483, + "step": 98 + }, + { + "epoch": 0.05374592833876222, + "grad_norm": 3.0418610323064317, + "learning_rate": 7.148014440433214e-06, + "loss": 4.1131, + "step": 99 + }, + { + "epoch": 0.05428881650380022, + "grad_norm": 2.710846119482301, + "learning_rate": 7.220216606498196e-06, + "loss": 5.0449, + "step": 100 + }, + { + "epoch": 0.054831704668838216, + "grad_norm": 1.906382949957388, + "learning_rate": 7.2924187725631776e-06, + "loss": 3.3181, + "step": 101 + }, + { + "epoch": 0.05537459283387622, + "grad_norm": 2.4669878697837, + "learning_rate": 7.36462093862816e-06, + "loss": 3.7996, + "step": 102 + }, + { + "epoch": 0.05591748099891422, + "grad_norm": 2.3755842712374586, + "learning_rate": 7.436823104693142e-06, + "loss": 4.0321, + "step": 103 + }, + { + "epoch": 0.05646036916395222, + "grad_norm": 2.032793864149989, + "learning_rate": 7.509025270758123e-06, + "loss": 3.4121, + "step": 104 + }, + { + "epoch": 0.057003257328990226, + "grad_norm": 2.1561554352810686, + "learning_rate": 7.5812274368231055e-06, + "loss": 4.2585, + "step": 105 + }, + { + "epoch": 0.05754614549402823, + "grad_norm": 1.9627108654626784, + "learning_rate": 7.653429602888087e-06, + "loss": 4.3085, + "step": 106 + }, + { + "epoch": 0.05808903365906623, + "grad_norm": 1.8622753095278954, + "learning_rate": 7.72563176895307e-06, + "loss": 3.8459, + "step": 107 + }, + { + "epoch": 0.05863192182410423, + "grad_norm": 1.4438765470100654, + "learning_rate": 7.797833935018051e-06, + "loss": 3.4914, + "step": 108 + }, + { + "epoch": 0.059174809989142235, + "grad_norm": 1.6164085905580508, + "learning_rate": 7.870036101083033e-06, + "loss": 3.4958, + "step": 109 + }, + { + "epoch": 0.05971769815418024, + "grad_norm": 1.6075014519578008, + "learning_rate": 7.942238267148014e-06, + "loss": 3.202, + "step": 110 + }, + { + "epoch": 0.06026058631921824, + "grad_norm": 1.7630994896021268, + "learning_rate": 8.014440433212997e-06, + "loss": 3.0916, + "step": 111 + }, + { + "epoch": 0.06080347448425624, + "grad_norm": 1.7600666006428431, + "learning_rate": 8.086642599277978e-06, + "loss": 3.5791, + "step": 112 + }, + { + "epoch": 0.061346362649294245, + "grad_norm": 2.324243905399748, + "learning_rate": 8.158844765342961e-06, + "loss": 3.6247, + "step": 113 + }, + { + "epoch": 0.06188925081433225, + "grad_norm": 1.2715449173786617, + "learning_rate": 8.231046931407943e-06, + "loss": 3.2381, + "step": 114 + }, + { + "epoch": 0.06243213897937025, + "grad_norm": 1.9029014476619814, + "learning_rate": 8.303249097472926e-06, + "loss": 3.5269, + "step": 115 + }, + { + "epoch": 0.06297502714440825, + "grad_norm": 1.8782574049467182, + "learning_rate": 8.375451263537907e-06, + "loss": 2.8698, + "step": 116 + }, + { + "epoch": 0.06351791530944625, + "grad_norm": 1.8522383705853351, + "learning_rate": 8.447653429602888e-06, + "loss": 3.5003, + "step": 117 + }, + { + "epoch": 0.06406080347448426, + "grad_norm": 1.6480852413427443, + "learning_rate": 8.519855595667871e-06, + "loss": 3.5665, + "step": 118 + }, + { + "epoch": 0.06460369163952226, + "grad_norm": 2.1628448626122685, + "learning_rate": 8.592057761732853e-06, + "loss": 3.343, + "step": 119 + }, + { + "epoch": 0.06514657980456026, + "grad_norm": 1.568456113423694, + "learning_rate": 8.664259927797834e-06, + "loss": 2.474, + "step": 120 + }, + { + "epoch": 0.06568946796959826, + "grad_norm": 1.8054357065473743, + "learning_rate": 8.736462093862817e-06, + "loss": 2.9228, + "step": 121 + }, + { + "epoch": 0.06623235613463627, + "grad_norm": 1.9418450135728782, + "learning_rate": 8.808664259927798e-06, + "loss": 3.1995, + "step": 122 + }, + { + "epoch": 0.06677524429967427, + "grad_norm": 1.321851521267606, + "learning_rate": 8.88086642599278e-06, + "loss": 3.5515, + "step": 123 + }, + { + "epoch": 0.06731813246471227, + "grad_norm": 1.390198419578882, + "learning_rate": 8.953068592057763e-06, + "loss": 2.7741, + "step": 124 + }, + { + "epoch": 0.06786102062975027, + "grad_norm": 2.119370139523294, + "learning_rate": 9.025270758122744e-06, + "loss": 3.3427, + "step": 125 + }, + { + "epoch": 0.06840390879478828, + "grad_norm": 2.042189734142807, + "learning_rate": 9.097472924187727e-06, + "loss": 3.1414, + "step": 126 + }, + { + "epoch": 0.06894679695982628, + "grad_norm": 2.0399787373405123, + "learning_rate": 9.169675090252709e-06, + "loss": 3.0862, + "step": 127 + }, + { + "epoch": 0.06948968512486428, + "grad_norm": 2.0523727295223915, + "learning_rate": 9.24187725631769e-06, + "loss": 2.6461, + "step": 128 + }, + { + "epoch": 0.07003257328990228, + "grad_norm": 1.459190908283032, + "learning_rate": 9.314079422382673e-06, + "loss": 2.9907, + "step": 129 + }, + { + "epoch": 0.07057546145494029, + "grad_norm": 1.7403521896723462, + "learning_rate": 9.386281588447654e-06, + "loss": 3.0681, + "step": 130 + }, + { + "epoch": 0.07111834961997829, + "grad_norm": 2.2283604183714383, + "learning_rate": 9.458483754512636e-06, + "loss": 3.4619, + "step": 131 + }, + { + "epoch": 0.07166123778501629, + "grad_norm": 1.3291461832292721, + "learning_rate": 9.530685920577619e-06, + "loss": 3.096, + "step": 132 + }, + { + "epoch": 0.0722041259500543, + "grad_norm": 1.7839357457835638, + "learning_rate": 9.6028880866426e-06, + "loss": 2.7512, + "step": 133 + }, + { + "epoch": 0.0727470141150923, + "grad_norm": 1.7836465130823984, + "learning_rate": 9.675090252707581e-06, + "loss": 3.3118, + "step": 134 + }, + { + "epoch": 0.0732899022801303, + "grad_norm": 2.0407172747925135, + "learning_rate": 9.747292418772564e-06, + "loss": 2.1032, + "step": 135 + }, + { + "epoch": 0.0738327904451683, + "grad_norm": 2.335226908691354, + "learning_rate": 9.819494584837546e-06, + "loss": 2.9039, + "step": 136 + }, + { + "epoch": 0.0743756786102063, + "grad_norm": 2.567081845339778, + "learning_rate": 9.891696750902527e-06, + "loss": 3.1071, + "step": 137 + }, + { + "epoch": 0.0749185667752443, + "grad_norm": 1.9549184920601244, + "learning_rate": 9.96389891696751e-06, + "loss": 2.8253, + "step": 138 + }, + { + "epoch": 0.07546145494028231, + "grad_norm": 1.6389815727698025, + "learning_rate": 1.0036101083032491e-05, + "loss": 2.5701, + "step": 139 + }, + { + "epoch": 0.07600434310532031, + "grad_norm": 1.6558924195158018, + "learning_rate": 1.0108303249097473e-05, + "loss": 2.6428, + "step": 140 + }, + { + "epoch": 0.07654723127035831, + "grad_norm": 2.0257343427634087, + "learning_rate": 1.0180505415162456e-05, + "loss": 2.2828, + "step": 141 + }, + { + "epoch": 0.07709011943539631, + "grad_norm": 1.879991343961764, + "learning_rate": 1.0252707581227437e-05, + "loss": 3.1436, + "step": 142 + }, + { + "epoch": 0.07763300760043432, + "grad_norm": 1.8718236498893788, + "learning_rate": 1.032490974729242e-05, + "loss": 3.2434, + "step": 143 + }, + { + "epoch": 0.0781758957654723, + "grad_norm": 1.8252190287923398, + "learning_rate": 1.03971119133574e-05, + "loss": 2.4851, + "step": 144 + }, + { + "epoch": 0.07871878393051031, + "grad_norm": 2.6564413093838213, + "learning_rate": 1.0469314079422383e-05, + "loss": 2.7315, + "step": 145 + }, + { + "epoch": 0.07926167209554831, + "grad_norm": 2.054556112475997, + "learning_rate": 1.0541516245487366e-05, + "loss": 2.647, + "step": 146 + }, + { + "epoch": 0.07980456026058631, + "grad_norm": 1.072343898976271, + "learning_rate": 1.0613718411552347e-05, + "loss": 2.5868, + "step": 147 + }, + { + "epoch": 0.08034744842562432, + "grad_norm": 1.1443585178571194, + "learning_rate": 1.068592057761733e-05, + "loss": 2.3247, + "step": 148 + }, + { + "epoch": 0.08089033659066232, + "grad_norm": 2.0424670761496646, + "learning_rate": 1.0758122743682312e-05, + "loss": 1.989, + "step": 149 + }, + { + "epoch": 0.08143322475570032, + "grad_norm": 1.3715615162918033, + "learning_rate": 1.0830324909747295e-05, + "loss": 2.3989, + "step": 150 + }, + { + "epoch": 0.08197611292073832, + "grad_norm": 1.3740116291532023, + "learning_rate": 1.0902527075812274e-05, + "loss": 1.8726, + "step": 151 + }, + { + "epoch": 0.08251900108577633, + "grad_norm": 2.5824759155435655, + "learning_rate": 1.0974729241877257e-05, + "loss": 2.3019, + "step": 152 + }, + { + "epoch": 0.08306188925081433, + "grad_norm": 1.5584514490971877, + "learning_rate": 1.1046931407942239e-05, + "loss": 2.4209, + "step": 153 + }, + { + "epoch": 0.08360477741585233, + "grad_norm": 1.6184456176727229, + "learning_rate": 1.1119133574007222e-05, + "loss": 2.2377, + "step": 154 + }, + { + "epoch": 0.08414766558089033, + "grad_norm": 2.6665151568462586, + "learning_rate": 1.1191335740072201e-05, + "loss": 2.6381, + "step": 155 + }, + { + "epoch": 0.08469055374592833, + "grad_norm": 1.442929928712006, + "learning_rate": 1.1263537906137184e-05, + "loss": 2.3405, + "step": 156 + }, + { + "epoch": 0.08523344191096634, + "grad_norm": 1.906547769868643, + "learning_rate": 1.1335740072202166e-05, + "loss": 2.4391, + "step": 157 + }, + { + "epoch": 0.08577633007600434, + "grad_norm": 1.2153727696829317, + "learning_rate": 1.1407942238267149e-05, + "loss": 1.8401, + "step": 158 + }, + { + "epoch": 0.08631921824104234, + "grad_norm": 1.4810729195972283, + "learning_rate": 1.1480144404332132e-05, + "loss": 2.7248, + "step": 159 + }, + { + "epoch": 0.08686210640608034, + "grad_norm": 2.9401118999625058, + "learning_rate": 1.1552346570397113e-05, + "loss": 2.6582, + "step": 160 + }, + { + "epoch": 0.08740499457111835, + "grad_norm": 1.145106458573816, + "learning_rate": 1.1624548736462096e-05, + "loss": 2.6648, + "step": 161 + }, + { + "epoch": 0.08794788273615635, + "grad_norm": 1.9228128025198137, + "learning_rate": 1.1696750902527076e-05, + "loss": 2.3702, + "step": 162 + }, + { + "epoch": 0.08849077090119435, + "grad_norm": 1.318978856783607, + "learning_rate": 1.1768953068592059e-05, + "loss": 1.9388, + "step": 163 + }, + { + "epoch": 0.08903365906623235, + "grad_norm": 1.58104775083856, + "learning_rate": 1.184115523465704e-05, + "loss": 2.0774, + "step": 164 + }, + { + "epoch": 0.08957654723127036, + "grad_norm": 1.8206646323260607, + "learning_rate": 1.1913357400722023e-05, + "loss": 2.5221, + "step": 165 + }, + { + "epoch": 0.09011943539630836, + "grad_norm": 1.0815960351491627, + "learning_rate": 1.1985559566787005e-05, + "loss": 2.128, + "step": 166 + }, + { + "epoch": 0.09066232356134636, + "grad_norm": 1.3341486432977188, + "learning_rate": 1.2057761732851988e-05, + "loss": 1.9418, + "step": 167 + }, + { + "epoch": 0.09120521172638436, + "grad_norm": 1.3772486036170464, + "learning_rate": 1.2129963898916967e-05, + "loss": 1.84, + "step": 168 + }, + { + "epoch": 0.09174809989142237, + "grad_norm": 1.42874152479176, + "learning_rate": 1.220216606498195e-05, + "loss": 2.2172, + "step": 169 + }, + { + "epoch": 0.09229098805646037, + "grad_norm": 1.9691394708584433, + "learning_rate": 1.2274368231046932e-05, + "loss": 2.9247, + "step": 170 + }, + { + "epoch": 0.09283387622149837, + "grad_norm": 1.4773974635910447, + "learning_rate": 1.2346570397111915e-05, + "loss": 2.1028, + "step": 171 + }, + { + "epoch": 0.09337676438653637, + "grad_norm": 1.5937344901476087, + "learning_rate": 1.2418772563176898e-05, + "loss": 2.4073, + "step": 172 + }, + { + "epoch": 0.09391965255157438, + "grad_norm": 1.630783959399718, + "learning_rate": 1.2490974729241878e-05, + "loss": 2.6978, + "step": 173 + }, + { + "epoch": 0.09446254071661238, + "grad_norm": 1.44736447594164, + "learning_rate": 1.256317689530686e-05, + "loss": 1.8851, + "step": 174 + }, + { + "epoch": 0.09500542888165038, + "grad_norm": 1.3909459856510267, + "learning_rate": 1.2635379061371842e-05, + "loss": 1.576, + "step": 175 + }, + { + "epoch": 0.09554831704668838, + "grad_norm": 1.7838393677575515, + "learning_rate": 1.2707581227436825e-05, + "loss": 2.3895, + "step": 176 + }, + { + "epoch": 0.09609120521172639, + "grad_norm": 1.1294127385799255, + "learning_rate": 1.2779783393501806e-05, + "loss": 2.4855, + "step": 177 + }, + { + "epoch": 0.09663409337676439, + "grad_norm": 1.259662421149648, + "learning_rate": 1.285198555956679e-05, + "loss": 2.5469, + "step": 178 + }, + { + "epoch": 0.09717698154180239, + "grad_norm": 1.4621367635995284, + "learning_rate": 1.2924187725631769e-05, + "loss": 1.9015, + "step": 179 + }, + { + "epoch": 0.09771986970684039, + "grad_norm": 1.6960126954903645, + "learning_rate": 1.2996389891696752e-05, + "loss": 2.9545, + "step": 180 + }, + { + "epoch": 0.0982627578718784, + "grad_norm": 1.5419022503429414, + "learning_rate": 1.3068592057761733e-05, + "loss": 1.9912, + "step": 181 + }, + { + "epoch": 0.0988056460369164, + "grad_norm": 1.4073422496120207, + "learning_rate": 1.3140794223826716e-05, + "loss": 2.3353, + "step": 182 + }, + { + "epoch": 0.0993485342019544, + "grad_norm": 1.2881659254471862, + "learning_rate": 1.3212996389891696e-05, + "loss": 2.7784, + "step": 183 + }, + { + "epoch": 0.0998914223669924, + "grad_norm": 1.361969663712106, + "learning_rate": 1.3285198555956679e-05, + "loss": 2.523, + "step": 184 + }, + { + "epoch": 0.1004343105320304, + "grad_norm": 2.5841731119141316, + "learning_rate": 1.3357400722021662e-05, + "loss": 2.43, + "step": 185 + }, + { + "epoch": 0.10097719869706841, + "grad_norm": 1.3203537847943392, + "learning_rate": 1.3429602888086643e-05, + "loss": 1.9404, + "step": 186 + }, + { + "epoch": 0.10152008686210641, + "grad_norm": 1.393097901493087, + "learning_rate": 1.3501805054151626e-05, + "loss": 2.6202, + "step": 187 + }, + { + "epoch": 0.10206297502714441, + "grad_norm": 0.9414377380652101, + "learning_rate": 1.3574007220216608e-05, + "loss": 2.0879, + "step": 188 + }, + { + "epoch": 0.10260586319218241, + "grad_norm": 1.6547581806701488, + "learning_rate": 1.3646209386281591e-05, + "loss": 2.8912, + "step": 189 + }, + { + "epoch": 0.10314875135722042, + "grad_norm": 1.1236261864015666, + "learning_rate": 1.371841155234657e-05, + "loss": 2.1802, + "step": 190 + }, + { + "epoch": 0.10369163952225842, + "grad_norm": 1.1766384878462062, + "learning_rate": 1.3790613718411554e-05, + "loss": 1.9127, + "step": 191 + }, + { + "epoch": 0.10423452768729642, + "grad_norm": 2.0453724866488656, + "learning_rate": 1.3862815884476535e-05, + "loss": 1.8522, + "step": 192 + }, + { + "epoch": 0.10477741585233442, + "grad_norm": 1.550964370734908, + "learning_rate": 1.3935018050541518e-05, + "loss": 2.5852, + "step": 193 + }, + { + "epoch": 0.10532030401737243, + "grad_norm": 1.303227180334529, + "learning_rate": 1.40072202166065e-05, + "loss": 2.407, + "step": 194 + }, + { + "epoch": 0.10586319218241043, + "grad_norm": 1.2009193115283474, + "learning_rate": 1.4079422382671482e-05, + "loss": 1.8978, + "step": 195 + }, + { + "epoch": 0.10640608034744843, + "grad_norm": 1.353648963417667, + "learning_rate": 1.4151624548736462e-05, + "loss": 2.283, + "step": 196 + }, + { + "epoch": 0.10694896851248643, + "grad_norm": 1.266154745219424, + "learning_rate": 1.4223826714801445e-05, + "loss": 2.0679, + "step": 197 + }, + { + "epoch": 0.10749185667752444, + "grad_norm": 4.186447043886807, + "learning_rate": 1.4296028880866428e-05, + "loss": 1.9175, + "step": 198 + }, + { + "epoch": 0.10803474484256244, + "grad_norm": 1.147065238667786, + "learning_rate": 1.436823104693141e-05, + "loss": 2.2748, + "step": 199 + }, + { + "epoch": 0.10857763300760044, + "grad_norm": 1.2760699256137895, + "learning_rate": 1.4440433212996392e-05, + "loss": 2.3031, + "step": 200 + }, + { + "epoch": 0.10912052117263844, + "grad_norm": 0.9452152631580967, + "learning_rate": 1.4512635379061372e-05, + "loss": 1.4833, + "step": 201 + }, + { + "epoch": 0.10966340933767643, + "grad_norm": 1.3980490034032074, + "learning_rate": 1.4584837545126355e-05, + "loss": 1.5865, + "step": 202 + }, + { + "epoch": 0.11020629750271443, + "grad_norm": 1.6440838011007093, + "learning_rate": 1.4657039711191336e-05, + "loss": 1.632, + "step": 203 + }, + { + "epoch": 0.11074918566775244, + "grad_norm": 1.3969764794458068, + "learning_rate": 1.472924187725632e-05, + "loss": 1.4301, + "step": 204 + }, + { + "epoch": 0.11129207383279044, + "grad_norm": 1.223663090997016, + "learning_rate": 1.48014440433213e-05, + "loss": 2.3993, + "step": 205 + }, + { + "epoch": 0.11183496199782844, + "grad_norm": 1.183881027275754, + "learning_rate": 1.4873646209386284e-05, + "loss": 1.4987, + "step": 206 + }, + { + "epoch": 0.11237785016286644, + "grad_norm": 1.3788483357578956, + "learning_rate": 1.4945848375451264e-05, + "loss": 2.7742, + "step": 207 + }, + { + "epoch": 0.11292073832790445, + "grad_norm": 1.1260622559595288, + "learning_rate": 1.5018050541516247e-05, + "loss": 1.755, + "step": 208 + }, + { + "epoch": 0.11346362649294245, + "grad_norm": 1.8439396062381286, + "learning_rate": 1.5090252707581228e-05, + "loss": 2.9875, + "step": 209 + }, + { + "epoch": 0.11400651465798045, + "grad_norm": 1.2789617727573501, + "learning_rate": 1.5162454873646211e-05, + "loss": 2.0119, + "step": 210 + }, + { + "epoch": 0.11454940282301845, + "grad_norm": 1.3385913561767113, + "learning_rate": 1.5234657039711192e-05, + "loss": 1.7693, + "step": 211 + }, + { + "epoch": 0.11509229098805646, + "grad_norm": 1.049162702021749, + "learning_rate": 1.5306859205776174e-05, + "loss": 1.7541, + "step": 212 + }, + { + "epoch": 0.11563517915309446, + "grad_norm": 1.0797407129441892, + "learning_rate": 1.537906137184116e-05, + "loss": 2.0427, + "step": 213 + }, + { + "epoch": 0.11617806731813246, + "grad_norm": 1.3661735952693994, + "learning_rate": 1.545126353790614e-05, + "loss": 2.722, + "step": 214 + }, + { + "epoch": 0.11672095548317046, + "grad_norm": 1.7084933140661187, + "learning_rate": 1.552346570397112e-05, + "loss": 2.2125, + "step": 215 + }, + { + "epoch": 0.11726384364820847, + "grad_norm": 1.3814428141700028, + "learning_rate": 1.5595667870036102e-05, + "loss": 2.448, + "step": 216 + }, + { + "epoch": 0.11780673181324647, + "grad_norm": 1.298309196615407, + "learning_rate": 1.5667870036101084e-05, + "loss": 2.2706, + "step": 217 + }, + { + "epoch": 0.11834961997828447, + "grad_norm": 1.4553148411944858, + "learning_rate": 1.5740072202166065e-05, + "loss": 2.0087, + "step": 218 + }, + { + "epoch": 0.11889250814332247, + "grad_norm": 1.5083115470870359, + "learning_rate": 1.581227436823105e-05, + "loss": 1.7179, + "step": 219 + }, + { + "epoch": 0.11943539630836048, + "grad_norm": 1.2335179878839475, + "learning_rate": 1.5884476534296028e-05, + "loss": 2.5337, + "step": 220 + }, + { + "epoch": 0.11997828447339848, + "grad_norm": 1.1729282902419653, + "learning_rate": 1.5956678700361013e-05, + "loss": 1.9604, + "step": 221 + }, + { + "epoch": 0.12052117263843648, + "grad_norm": 1.5700347334422509, + "learning_rate": 1.6028880866425994e-05, + "loss": 2.3655, + "step": 222 + }, + { + "epoch": 0.12106406080347448, + "grad_norm": 1.435764746128967, + "learning_rate": 1.6101083032490975e-05, + "loss": 2.109, + "step": 223 + }, + { + "epoch": 0.12160694896851248, + "grad_norm": 1.0866216936301165, + "learning_rate": 1.6173285198555957e-05, + "loss": 2.3334, + "step": 224 + }, + { + "epoch": 0.12214983713355049, + "grad_norm": 1.8060789361277199, + "learning_rate": 1.624548736462094e-05, + "loss": 1.6605, + "step": 225 + }, + { + "epoch": 0.12269272529858849, + "grad_norm": 1.539101200174871, + "learning_rate": 1.6317689530685923e-05, + "loss": 2.2597, + "step": 226 + }, + { + "epoch": 0.12323561346362649, + "grad_norm": 1.7392984350243024, + "learning_rate": 1.6389891696750904e-05, + "loss": 2.5979, + "step": 227 + }, + { + "epoch": 0.1237785016286645, + "grad_norm": 1.6792982939949113, + "learning_rate": 1.6462093862815885e-05, + "loss": 2.0626, + "step": 228 + }, + { + "epoch": 0.1243213897937025, + "grad_norm": 2.067856070276011, + "learning_rate": 1.6534296028880867e-05, + "loss": 2.056, + "step": 229 + }, + { + "epoch": 0.1248642779587405, + "grad_norm": 1.2773224600362139, + "learning_rate": 1.660649819494585e-05, + "loss": 1.5852, + "step": 230 + }, + { + "epoch": 0.1254071661237785, + "grad_norm": 1.381245750855955, + "learning_rate": 1.6678700361010833e-05, + "loss": 2.0675, + "step": 231 + }, + { + "epoch": 0.1259500542888165, + "grad_norm": 1.4215755121774605, + "learning_rate": 1.6750902527075814e-05, + "loss": 2.1002, + "step": 232 + }, + { + "epoch": 0.1264929424538545, + "grad_norm": 1.4581673093175722, + "learning_rate": 1.6823104693140795e-05, + "loss": 1.9279, + "step": 233 + }, + { + "epoch": 0.1270358306188925, + "grad_norm": 1.2656102558648035, + "learning_rate": 1.6895306859205777e-05, + "loss": 1.4913, + "step": 234 + }, + { + "epoch": 0.1275787187839305, + "grad_norm": 1.3985246778048182, + "learning_rate": 1.6967509025270758e-05, + "loss": 2.1629, + "step": 235 + }, + { + "epoch": 0.1281216069489685, + "grad_norm": 1.757194392177075, + "learning_rate": 1.7039711191335743e-05, + "loss": 2.5688, + "step": 236 + }, + { + "epoch": 0.12866449511400652, + "grad_norm": 1.3766112824147358, + "learning_rate": 1.711191335740072e-05, + "loss": 2.3344, + "step": 237 + }, + { + "epoch": 0.12920738327904452, + "grad_norm": 1.2985043105700416, + "learning_rate": 1.7184115523465706e-05, + "loss": 2.0626, + "step": 238 + }, + { + "epoch": 0.12975027144408252, + "grad_norm": 1.715796674825951, + "learning_rate": 1.7256317689530687e-05, + "loss": 2.1559, + "step": 239 + }, + { + "epoch": 0.13029315960912052, + "grad_norm": 1.6374209306607368, + "learning_rate": 1.7328519855595668e-05, + "loss": 2.0801, + "step": 240 + }, + { + "epoch": 0.13083604777415853, + "grad_norm": 1.244682506367769, + "learning_rate": 1.7400722021660653e-05, + "loss": 1.6125, + "step": 241 + }, + { + "epoch": 0.13137893593919653, + "grad_norm": 1.3001838781366861, + "learning_rate": 1.7472924187725634e-05, + "loss": 2.0934, + "step": 242 + }, + { + "epoch": 0.13192182410423453, + "grad_norm": 1.1600879910417528, + "learning_rate": 1.7545126353790616e-05, + "loss": 1.7862, + "step": 243 + }, + { + "epoch": 0.13246471226927253, + "grad_norm": 1.3503312968411385, + "learning_rate": 1.7617328519855597e-05, + "loss": 2.1636, + "step": 244 + }, + { + "epoch": 0.13300760043431054, + "grad_norm": 2.1299761628528167, + "learning_rate": 1.768953068592058e-05, + "loss": 2.4826, + "step": 245 + }, + { + "epoch": 0.13355048859934854, + "grad_norm": 1.3232052544377066, + "learning_rate": 1.776173285198556e-05, + "loss": 2.0497, + "step": 246 + }, + { + "epoch": 0.13409337676438654, + "grad_norm": 1.5079892467642568, + "learning_rate": 1.7833935018050544e-05, + "loss": 2.4552, + "step": 247 + }, + { + "epoch": 0.13463626492942454, + "grad_norm": 1.716355318914646, + "learning_rate": 1.7906137184115526e-05, + "loss": 1.5309, + "step": 248 + }, + { + "epoch": 0.13517915309446255, + "grad_norm": 2.1523878955969193, + "learning_rate": 1.7978339350180507e-05, + "loss": 2.7403, + "step": 249 + }, + { + "epoch": 0.13572204125950055, + "grad_norm": 1.8169301927997485, + "learning_rate": 1.805054151624549e-05, + "loss": 2.3324, + "step": 250 + }, + { + "epoch": 0.13626492942453855, + "grad_norm": 1.4464924096148426, + "learning_rate": 1.812274368231047e-05, + "loss": 1.8399, + "step": 251 + }, + { + "epoch": 0.13680781758957655, + "grad_norm": 1.9818298103877654, + "learning_rate": 1.8194945848375454e-05, + "loss": 2.4237, + "step": 252 + }, + { + "epoch": 0.13735070575461455, + "grad_norm": 1.6371003739671923, + "learning_rate": 1.8267148014440436e-05, + "loss": 2.0477, + "step": 253 + }, + { + "epoch": 0.13789359391965256, + "grad_norm": 1.2944832148952543, + "learning_rate": 1.8339350180505417e-05, + "loss": 2.2269, + "step": 254 + }, + { + "epoch": 0.13843648208469056, + "grad_norm": 1.4284289887933213, + "learning_rate": 1.84115523465704e-05, + "loss": 2.2465, + "step": 255 + }, + { + "epoch": 0.13897937024972856, + "grad_norm": 1.4633226693151502, + "learning_rate": 1.848375451263538e-05, + "loss": 2.1582, + "step": 256 + }, + { + "epoch": 0.13952225841476656, + "grad_norm": 1.6169546473154692, + "learning_rate": 1.855595667870036e-05, + "loss": 1.9526, + "step": 257 + }, + { + "epoch": 0.14006514657980457, + "grad_norm": 1.47082651679944, + "learning_rate": 1.8628158844765346e-05, + "loss": 2.0314, + "step": 258 + }, + { + "epoch": 0.14060803474484257, + "grad_norm": 1.2279964838302115, + "learning_rate": 1.8700361010830327e-05, + "loss": 1.4345, + "step": 259 + }, + { + "epoch": 0.14115092290988057, + "grad_norm": 1.5727648286246085, + "learning_rate": 1.877256317689531e-05, + "loss": 2.7371, + "step": 260 + }, + { + "epoch": 0.14169381107491857, + "grad_norm": 1.5536143039045338, + "learning_rate": 1.884476534296029e-05, + "loss": 1.6691, + "step": 261 + }, + { + "epoch": 0.14223669923995658, + "grad_norm": 1.504953118413033, + "learning_rate": 1.891696750902527e-05, + "loss": 1.8422, + "step": 262 + }, + { + "epoch": 0.14277958740499458, + "grad_norm": 1.6092868309857633, + "learning_rate": 1.8989169675090253e-05, + "loss": 1.6446, + "step": 263 + }, + { + "epoch": 0.14332247557003258, + "grad_norm": 1.434484546738067, + "learning_rate": 1.9061371841155237e-05, + "loss": 1.5945, + "step": 264 + }, + { + "epoch": 0.14386536373507058, + "grad_norm": 1.3472204852366787, + "learning_rate": 1.913357400722022e-05, + "loss": 1.4511, + "step": 265 + }, + { + "epoch": 0.1444082519001086, + "grad_norm": 2.493842366649246, + "learning_rate": 1.92057761732852e-05, + "loss": 2.1144, + "step": 266 + }, + { + "epoch": 0.1449511400651466, + "grad_norm": 1.3138428233046642, + "learning_rate": 1.927797833935018e-05, + "loss": 1.6592, + "step": 267 + }, + { + "epoch": 0.1454940282301846, + "grad_norm": 1.7201458829291465, + "learning_rate": 1.9350180505415163e-05, + "loss": 1.6895, + "step": 268 + }, + { + "epoch": 0.1460369163952226, + "grad_norm": 1.9777013455620547, + "learning_rate": 1.9422382671480147e-05, + "loss": 2.2628, + "step": 269 + }, + { + "epoch": 0.1465798045602606, + "grad_norm": 1.8559862870685497, + "learning_rate": 1.949458483754513e-05, + "loss": 1.3374, + "step": 270 + }, + { + "epoch": 0.1471226927252986, + "grad_norm": 1.919404449298866, + "learning_rate": 1.956678700361011e-05, + "loss": 2.3805, + "step": 271 + }, + { + "epoch": 0.1476655808903366, + "grad_norm": 1.4760010543412088, + "learning_rate": 1.963898916967509e-05, + "loss": 1.9791, + "step": 272 + }, + { + "epoch": 0.1482084690553746, + "grad_norm": 1.4491842831658432, + "learning_rate": 1.9711191335740073e-05, + "loss": 1.8119, + "step": 273 + }, + { + "epoch": 0.1487513572204126, + "grad_norm": 1.4098796806726177, + "learning_rate": 1.9783393501805054e-05, + "loss": 1.7398, + "step": 274 + }, + { + "epoch": 0.1492942453854506, + "grad_norm": 1.5009369942229867, + "learning_rate": 1.985559566787004e-05, + "loss": 1.5194, + "step": 275 + }, + { + "epoch": 0.1498371335504886, + "grad_norm": 1.5073462671342261, + "learning_rate": 1.992779783393502e-05, + "loss": 1.8125, + "step": 276 + }, + { + "epoch": 0.1503800217155266, + "grad_norm": 1.9315604890805302, + "learning_rate": 2e-05, + "loss": 1.5429, + "step": 277 + }, + { + "epoch": 0.15092290988056462, + "grad_norm": 1.5059106777628657, + "learning_rate": 1.999999938159203e-05, + "loss": 1.6402, + "step": 278 + }, + { + "epoch": 0.15146579804560262, + "grad_norm": 1.7255630320192266, + "learning_rate": 1.9999997526368205e-05, + "loss": 2.2806, + "step": 279 + }, + { + "epoch": 0.15200868621064062, + "grad_norm": 1.7985407961594548, + "learning_rate": 1.999999443432874e-05, + "loss": 2.5213, + "step": 280 + }, + { + "epoch": 0.15255157437567862, + "grad_norm": 1.801471230172198, + "learning_rate": 1.999999010547403e-05, + "loss": 1.8362, + "step": 281 + }, + { + "epoch": 0.15309446254071662, + "grad_norm": 3.236881896883975, + "learning_rate": 1.999998453980461e-05, + "loss": 2.7148, + "step": 282 + }, + { + "epoch": 0.15363735070575463, + "grad_norm": 1.7343786362856708, + "learning_rate": 1.9999977737321156e-05, + "loss": 1.8195, + "step": 283 + }, + { + "epoch": 0.15418023887079263, + "grad_norm": 1.5006942215232695, + "learning_rate": 1.999996969802452e-05, + "loss": 1.8744, + "step": 284 + }, + { + "epoch": 0.15472312703583063, + "grad_norm": 4.171967924853298, + "learning_rate": 1.99999604219157e-05, + "loss": 2.4329, + "step": 285 + }, + { + "epoch": 0.15526601520086863, + "grad_norm": 1.7546486556973808, + "learning_rate": 1.9999949908995832e-05, + "loss": 1.3931, + "step": 286 + }, + { + "epoch": 0.15580890336590664, + "grad_norm": 2.0223702721102375, + "learning_rate": 1.999993815926622e-05, + "loss": 1.6704, + "step": 287 + }, + { + "epoch": 0.1563517915309446, + "grad_norm": 1.8736762651082781, + "learning_rate": 1.9999925172728324e-05, + "loss": 1.9895, + "step": 288 + }, + { + "epoch": 0.15689467969598261, + "grad_norm": 1.8878827291463698, + "learning_rate": 1.9999910949383742e-05, + "loss": 2.4939, + "step": 289 + }, + { + "epoch": 0.15743756786102062, + "grad_norm": 1.7277080078786893, + "learning_rate": 1.9999895489234234e-05, + "loss": 1.9265, + "step": 290 + }, + { + "epoch": 0.15798045602605862, + "grad_norm": 1.7402172969699101, + "learning_rate": 1.999987879228172e-05, + "loss": 1.8101, + "step": 291 + }, + { + "epoch": 0.15852334419109662, + "grad_norm": 2.077278713661093, + "learning_rate": 1.9999860858528257e-05, + "loss": 2.1763, + "step": 292 + }, + { + "epoch": 0.15906623235613462, + "grad_norm": 2.247624069618552, + "learning_rate": 1.9999841687976067e-05, + "loss": 1.9727, + "step": 293 + }, + { + "epoch": 0.15960912052117263, + "grad_norm": 1.6712153667666279, + "learning_rate": 1.999982128062752e-05, + "loss": 1.9921, + "step": 294 + }, + { + "epoch": 0.16015200868621063, + "grad_norm": 1.7126188149003763, + "learning_rate": 1.999979963648514e-05, + "loss": 1.9375, + "step": 295 + }, + { + "epoch": 0.16069489685124863, + "grad_norm": 1.949228534239786, + "learning_rate": 1.99997767555516e-05, + "loss": 2.1628, + "step": 296 + }, + { + "epoch": 0.16123778501628663, + "grad_norm": 1.8253305155157105, + "learning_rate": 1.9999752637829734e-05, + "loss": 1.8293, + "step": 297 + }, + { + "epoch": 0.16178067318132464, + "grad_norm": 1.5881557560095252, + "learning_rate": 1.9999727283322524e-05, + "loss": 1.3688, + "step": 298 + }, + { + "epoch": 0.16232356134636264, + "grad_norm": 2.4008202914116503, + "learning_rate": 1.9999700692033112e-05, + "loss": 1.5919, + "step": 299 + }, + { + "epoch": 0.16286644951140064, + "grad_norm": 1.610839420604411, + "learning_rate": 1.9999672863964778e-05, + "loss": 1.8724, + "step": 300 + }, + { + "epoch": 0.16340933767643864, + "grad_norm": 1.4663439746230646, + "learning_rate": 1.9999643799120964e-05, + "loss": 1.4628, + "step": 301 + }, + { + "epoch": 0.16395222584147665, + "grad_norm": 1.7051101810518976, + "learning_rate": 1.9999613497505272e-05, + "loss": 1.2916, + "step": 302 + }, + { + "epoch": 0.16449511400651465, + "grad_norm": 1.8967399031374104, + "learning_rate": 1.9999581959121443e-05, + "loss": 1.7559, + "step": 303 + }, + { + "epoch": 0.16503800217155265, + "grad_norm": 1.6206520761555416, + "learning_rate": 1.9999549183973382e-05, + "loss": 1.9196, + "step": 304 + }, + { + "epoch": 0.16558089033659065, + "grad_norm": 2.3545652344963544, + "learning_rate": 1.999951517206514e-05, + "loss": 1.2731, + "step": 305 + }, + { + "epoch": 0.16612377850162866, + "grad_norm": 1.7952736257095967, + "learning_rate": 1.9999479923400926e-05, + "loss": 2.0502, + "step": 306 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 1.7199420678903319, + "learning_rate": 1.99994434379851e-05, + "loss": 1.7059, + "step": 307 + }, + { + "epoch": 0.16720955483170466, + "grad_norm": 2.3318553047346, + "learning_rate": 1.9999405715822167e-05, + "loss": 1.4667, + "step": 308 + }, + { + "epoch": 0.16775244299674266, + "grad_norm": 1.618961362672766, + "learning_rate": 1.9999366756916804e-05, + "loss": 1.1866, + "step": 309 + }, + { + "epoch": 0.16829533116178066, + "grad_norm": 2.0805239071290114, + "learning_rate": 1.999932656127382e-05, + "loss": 1.7601, + "step": 310 + }, + { + "epoch": 0.16883821932681867, + "grad_norm": 1.782022796323131, + "learning_rate": 1.9999285128898193e-05, + "loss": 1.6207, + "step": 311 + }, + { + "epoch": 0.16938110749185667, + "grad_norm": 2.2438870289045028, + "learning_rate": 1.9999242459795045e-05, + "loss": 2.1329, + "step": 312 + }, + { + "epoch": 0.16992399565689467, + "grad_norm": 2.2550840823235587, + "learning_rate": 1.9999198553969652e-05, + "loss": 1.6055, + "step": 313 + }, + { + "epoch": 0.17046688382193267, + "grad_norm": 1.8012459587153988, + "learning_rate": 1.9999153411427445e-05, + "loss": 2.018, + "step": 314 + }, + { + "epoch": 0.17100977198697068, + "grad_norm": 1.8498840013582993, + "learning_rate": 1.9999107032174007e-05, + "loss": 1.8113, + "step": 315 + }, + { + "epoch": 0.17155266015200868, + "grad_norm": 2.2799557163953965, + "learning_rate": 1.9999059416215078e-05, + "loss": 1.9494, + "step": 316 + }, + { + "epoch": 0.17209554831704668, + "grad_norm": 2.383891228288174, + "learning_rate": 1.999901056355654e-05, + "loss": 1.8323, + "step": 317 + }, + { + "epoch": 0.17263843648208468, + "grad_norm": 1.9081077971876759, + "learning_rate": 1.9998960474204443e-05, + "loss": 1.5476, + "step": 318 + }, + { + "epoch": 0.1731813246471227, + "grad_norm": 1.9010783437994137, + "learning_rate": 1.999890914816498e-05, + "loss": 1.4602, + "step": 319 + }, + { + "epoch": 0.1737242128121607, + "grad_norm": 2.295573334252978, + "learning_rate": 1.9998856585444493e-05, + "loss": 1.0965, + "step": 320 + }, + { + "epoch": 0.1742671009771987, + "grad_norm": 2.1014332124066954, + "learning_rate": 1.999880278604949e-05, + "loss": 1.3726, + "step": 321 + }, + { + "epoch": 0.1748099891422367, + "grad_norm": 2.0418722317428206, + "learning_rate": 1.9998747749986625e-05, + "loss": 1.8875, + "step": 322 + }, + { + "epoch": 0.1753528773072747, + "grad_norm": 1.768897707524193, + "learning_rate": 1.99986914772627e-05, + "loss": 1.8841, + "step": 323 + }, + { + "epoch": 0.1758957654723127, + "grad_norm": 2.603335158781435, + "learning_rate": 1.9998633967884676e-05, + "loss": 1.4262, + "step": 324 + }, + { + "epoch": 0.1764386536373507, + "grad_norm": 2.1485447383197607, + "learning_rate": 1.999857522185967e-05, + "loss": 1.9283, + "step": 325 + }, + { + "epoch": 0.1769815418023887, + "grad_norm": 1.8398472624384716, + "learning_rate": 1.9998515239194945e-05, + "loss": 1.9089, + "step": 326 + }, + { + "epoch": 0.1775244299674267, + "grad_norm": 2.6988014220525374, + "learning_rate": 1.9998454019897918e-05, + "loss": 1.3636, + "step": 327 + }, + { + "epoch": 0.1780673181324647, + "grad_norm": 2.182794188412918, + "learning_rate": 1.9998391563976166e-05, + "loss": 1.7041, + "step": 328 + }, + { + "epoch": 0.1786102062975027, + "grad_norm": 1.970858895313084, + "learning_rate": 1.9998327871437405e-05, + "loss": 1.4917, + "step": 329 + }, + { + "epoch": 0.1791530944625407, + "grad_norm": 1.829655959511745, + "learning_rate": 1.9998262942289524e-05, + "loss": 1.9694, + "step": 330 + }, + { + "epoch": 0.17969598262757872, + "grad_norm": 1.9990017706312653, + "learning_rate": 1.9998196776540545e-05, + "loss": 2.067, + "step": 331 + }, + { + "epoch": 0.18023887079261672, + "grad_norm": 2.2829395740777234, + "learning_rate": 1.9998129374198655e-05, + "loss": 1.6707, + "step": 332 + }, + { + "epoch": 0.18078175895765472, + "grad_norm": 2.174263401724357, + "learning_rate": 1.9998060735272186e-05, + "loss": 1.6897, + "step": 333 + }, + { + "epoch": 0.18132464712269272, + "grad_norm": 1.8438064757899841, + "learning_rate": 1.9997990859769633e-05, + "loss": 1.4446, + "step": 334 + }, + { + "epoch": 0.18186753528773072, + "grad_norm": 2.0988396037795507, + "learning_rate": 1.9997919747699638e-05, + "loss": 1.8837, + "step": 335 + }, + { + "epoch": 0.18241042345276873, + "grad_norm": 2.582616969632433, + "learning_rate": 1.999784739907099e-05, + "loss": 2.459, + "step": 336 + }, + { + "epoch": 0.18295331161780673, + "grad_norm": 2.5198677585207707, + "learning_rate": 1.9997773813892644e-05, + "loss": 1.5117, + "step": 337 + }, + { + "epoch": 0.18349619978284473, + "grad_norm": 2.1361317951692635, + "learning_rate": 1.9997698992173697e-05, + "loss": 1.5892, + "step": 338 + }, + { + "epoch": 0.18403908794788273, + "grad_norm": 1.7284990063627486, + "learning_rate": 1.9997622933923406e-05, + "loss": 1.5151, + "step": 339 + }, + { + "epoch": 0.18458197611292074, + "grad_norm": 2.076752898704788, + "learning_rate": 1.9997545639151176e-05, + "loss": 2.3424, + "step": 340 + }, + { + "epoch": 0.18512486427795874, + "grad_norm": 2.479369276757794, + "learning_rate": 1.999746710786657e-05, + "loss": 1.5923, + "step": 341 + }, + { + "epoch": 0.18566775244299674, + "grad_norm": 1.9193986648366674, + "learning_rate": 1.9997387340079294e-05, + "loss": 1.6222, + "step": 342 + }, + { + "epoch": 0.18621064060803474, + "grad_norm": 1.8622986153150223, + "learning_rate": 1.999730633579922e-05, + "loss": 1.8402, + "step": 343 + }, + { + "epoch": 0.18675352877307275, + "grad_norm": 2.4012194427686238, + "learning_rate": 1.999722409503637e-05, + "loss": 1.9406, + "step": 344 + }, + { + "epoch": 0.18729641693811075, + "grad_norm": 2.3641518463127538, + "learning_rate": 1.9997140617800907e-05, + "loss": 2.0108, + "step": 345 + }, + { + "epoch": 0.18783930510314875, + "grad_norm": 2.406749148558192, + "learning_rate": 1.9997055904103156e-05, + "loss": 1.1323, + "step": 346 + }, + { + "epoch": 0.18838219326818675, + "grad_norm": 2.002785642517938, + "learning_rate": 1.99969699539536e-05, + "loss": 1.8745, + "step": 347 + }, + { + "epoch": 0.18892508143322476, + "grad_norm": 2.0002882431712816, + "learning_rate": 1.9996882767362874e-05, + "loss": 1.6311, + "step": 348 + }, + { + "epoch": 0.18946796959826276, + "grad_norm": 2.14418043520414, + "learning_rate": 1.9996794344341744e-05, + "loss": 1.7789, + "step": 349 + }, + { + "epoch": 0.19001085776330076, + "grad_norm": 2.0976983604919144, + "learning_rate": 1.9996704684901163e-05, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.19055374592833876, + "grad_norm": 2.017704706993719, + "learning_rate": 1.9996613789052214e-05, + "loss": 1.7974, + "step": 351 + }, + { + "epoch": 0.19109663409337677, + "grad_norm": 1.9907769296520095, + "learning_rate": 1.999652165680614e-05, + "loss": 1.9079, + "step": 352 + }, + { + "epoch": 0.19163952225841477, + "grad_norm": 2.3031451756435746, + "learning_rate": 1.999642828817433e-05, + "loss": 1.7022, + "step": 353 + }, + { + "epoch": 0.19218241042345277, + "grad_norm": 1.6722822353766713, + "learning_rate": 1.9996333683168342e-05, + "loss": 1.6655, + "step": 354 + }, + { + "epoch": 0.19272529858849077, + "grad_norm": 2.2974087830498027, + "learning_rate": 1.9996237841799874e-05, + "loss": 1.761, + "step": 355 + }, + { + "epoch": 0.19326818675352878, + "grad_norm": 1.9181254747298617, + "learning_rate": 1.9996140764080777e-05, + "loss": 1.8259, + "step": 356 + }, + { + "epoch": 0.19381107491856678, + "grad_norm": 2.07933834321742, + "learning_rate": 1.9996042450023053e-05, + "loss": 1.3783, + "step": 357 + }, + { + "epoch": 0.19435396308360478, + "grad_norm": 2.8332787295887436, + "learning_rate": 1.9995942899638875e-05, + "loss": 1.9796, + "step": 358 + }, + { + "epoch": 0.19489685124864278, + "grad_norm": 1.7924418183793758, + "learning_rate": 1.9995842112940545e-05, + "loss": 1.5063, + "step": 359 + }, + { + "epoch": 0.19543973941368079, + "grad_norm": 2.78627522830663, + "learning_rate": 1.9995740089940532e-05, + "loss": 1.891, + "step": 360 + }, + { + "epoch": 0.1959826275787188, + "grad_norm": 2.067311381839208, + "learning_rate": 1.9995636830651453e-05, + "loss": 1.9235, + "step": 361 + }, + { + "epoch": 0.1965255157437568, + "grad_norm": 2.539112412358468, + "learning_rate": 1.9995532335086078e-05, + "loss": 2.3062, + "step": 362 + }, + { + "epoch": 0.1970684039087948, + "grad_norm": 1.9040915545393904, + "learning_rate": 1.999542660325734e-05, + "loss": 1.9263, + "step": 363 + }, + { + "epoch": 0.1976112920738328, + "grad_norm": 1.7664296640025938, + "learning_rate": 1.9995319635178305e-05, + "loss": 1.5562, + "step": 364 + }, + { + "epoch": 0.1981541802388708, + "grad_norm": 1.687746217581272, + "learning_rate": 1.9995211430862206e-05, + "loss": 1.738, + "step": 365 + }, + { + "epoch": 0.1986970684039088, + "grad_norm": 1.9292930460598687, + "learning_rate": 1.9995101990322428e-05, + "loss": 1.5446, + "step": 366 + }, + { + "epoch": 0.1992399565689468, + "grad_norm": 1.9890390069401638, + "learning_rate": 1.9994991313572508e-05, + "loss": 1.9086, + "step": 367 + }, + { + "epoch": 0.1997828447339848, + "grad_norm": 1.8632430045841502, + "learning_rate": 1.999487940062613e-05, + "loss": 1.9747, + "step": 368 + }, + { + "epoch": 0.2003257328990228, + "grad_norm": 2.0677818303742237, + "learning_rate": 1.999476625149714e-05, + "loss": 1.4377, + "step": 369 + }, + { + "epoch": 0.2008686210640608, + "grad_norm": 1.7412205305358945, + "learning_rate": 1.9994651866199527e-05, + "loss": 1.9352, + "step": 370 + }, + { + "epoch": 0.2014115092290988, + "grad_norm": 1.8562175126383418, + "learning_rate": 1.9994536244747448e-05, + "loss": 1.5469, + "step": 371 + }, + { + "epoch": 0.20195439739413681, + "grad_norm": 1.8688345857586197, + "learning_rate": 1.9994419387155194e-05, + "loss": 1.8327, + "step": 372 + }, + { + "epoch": 0.20249728555917482, + "grad_norm": 2.1538407513899647, + "learning_rate": 1.9994301293437223e-05, + "loss": 1.9621, + "step": 373 + }, + { + "epoch": 0.20304017372421282, + "grad_norm": 2.4849664096044752, + "learning_rate": 1.999418196360814e-05, + "loss": 1.8179, + "step": 374 + }, + { + "epoch": 0.20358306188925082, + "grad_norm": 2.098949652063271, + "learning_rate": 1.99940613976827e-05, + "loss": 1.8719, + "step": 375 + }, + { + "epoch": 0.20412595005428882, + "grad_norm": 1.8802547433747465, + "learning_rate": 1.999393959567582e-05, + "loss": 1.5697, + "step": 376 + }, + { + "epoch": 0.20466883821932683, + "grad_norm": 2.090861017905227, + "learning_rate": 1.9993816557602567e-05, + "loss": 1.8249, + "step": 377 + }, + { + "epoch": 0.20521172638436483, + "grad_norm": 1.775481861127165, + "learning_rate": 1.999369228347815e-05, + "loss": 1.3095, + "step": 378 + }, + { + "epoch": 0.20575461454940283, + "grad_norm": 2.260330600420479, + "learning_rate": 1.999356677331794e-05, + "loss": 1.9153, + "step": 379 + }, + { + "epoch": 0.20629750271444083, + "grad_norm": 2.0769016143551484, + "learning_rate": 1.999344002713747e-05, + "loss": 1.958, + "step": 380 + }, + { + "epoch": 0.20684039087947884, + "grad_norm": 2.063584775140761, + "learning_rate": 1.9993312044952408e-05, + "loss": 1.7887, + "step": 381 + }, + { + "epoch": 0.20738327904451684, + "grad_norm": 1.9638736053910157, + "learning_rate": 1.9993182826778588e-05, + "loss": 1.3178, + "step": 382 + }, + { + "epoch": 0.20792616720955484, + "grad_norm": 1.7752329715548703, + "learning_rate": 1.9993052372631988e-05, + "loss": 1.4473, + "step": 383 + }, + { + "epoch": 0.20846905537459284, + "grad_norm": 2.1002641405528952, + "learning_rate": 1.999292068252874e-05, + "loss": 1.9698, + "step": 384 + }, + { + "epoch": 0.20901194353963085, + "grad_norm": 2.757803188261632, + "learning_rate": 1.999278775648514e-05, + "loss": 1.1517, + "step": 385 + }, + { + "epoch": 0.20955483170466885, + "grad_norm": 3.0341284449704378, + "learning_rate": 1.9992653594517624e-05, + "loss": 1.987, + "step": 386 + }, + { + "epoch": 0.21009771986970685, + "grad_norm": 2.445830333821562, + "learning_rate": 1.9992518196642786e-05, + "loss": 1.7656, + "step": 387 + }, + { + "epoch": 0.21064060803474485, + "grad_norm": 2.737783962568244, + "learning_rate": 1.9992381562877368e-05, + "loss": 2.3012, + "step": 388 + }, + { + "epoch": 0.21118349619978286, + "grad_norm": 2.0108432149096793, + "learning_rate": 1.9992243693238275e-05, + "loss": 2.1096, + "step": 389 + }, + { + "epoch": 0.21172638436482086, + "grad_norm": 2.1448465445226397, + "learning_rate": 1.9992104587742558e-05, + "loss": 1.6912, + "step": 390 + }, + { + "epoch": 0.21226927252985886, + "grad_norm": 2.842739011518123, + "learning_rate": 1.999196424640742e-05, + "loss": 1.9987, + "step": 391 + }, + { + "epoch": 0.21281216069489686, + "grad_norm": 2.026900017823922, + "learning_rate": 1.9991822669250216e-05, + "loss": 1.6251, + "step": 392 + }, + { + "epoch": 0.21335504885993486, + "grad_norm": 2.139818571931489, + "learning_rate": 1.9991679856288462e-05, + "loss": 1.4181, + "step": 393 + }, + { + "epoch": 0.21389793702497287, + "grad_norm": 2.061327407258535, + "learning_rate": 1.999153580753982e-05, + "loss": 2.0976, + "step": 394 + }, + { + "epoch": 0.21444082519001087, + "grad_norm": 1.9510646261617408, + "learning_rate": 1.9991390523022105e-05, + "loss": 1.3091, + "step": 395 + }, + { + "epoch": 0.21498371335504887, + "grad_norm": 2.0892427860245664, + "learning_rate": 1.9991244002753287e-05, + "loss": 1.3693, + "step": 396 + }, + { + "epoch": 0.21552660152008687, + "grad_norm": 1.9528708238712815, + "learning_rate": 1.9991096246751483e-05, + "loss": 1.2807, + "step": 397 + }, + { + "epoch": 0.21606948968512488, + "grad_norm": 2.0084632909809983, + "learning_rate": 1.9990947255034977e-05, + "loss": 1.7429, + "step": 398 + }, + { + "epoch": 0.21661237785016288, + "grad_norm": 2.19414527352725, + "learning_rate": 1.999079702762219e-05, + "loss": 1.5962, + "step": 399 + }, + { + "epoch": 0.21715526601520088, + "grad_norm": 2.184484203373308, + "learning_rate": 1.9990645564531702e-05, + "loss": 1.6887, + "step": 400 + }, + { + "epoch": 0.21769815418023888, + "grad_norm": 2.214473402714986, + "learning_rate": 1.9990492865782248e-05, + "loss": 1.5226, + "step": 401 + }, + { + "epoch": 0.2182410423452769, + "grad_norm": 2.1467748369867414, + "learning_rate": 1.9990338931392714e-05, + "loss": 1.5634, + "step": 402 + }, + { + "epoch": 0.21878393051031486, + "grad_norm": 2.1969353910959684, + "learning_rate": 1.999018376138214e-05, + "loss": 1.8652, + "step": 403 + }, + { + "epoch": 0.21932681867535286, + "grad_norm": 2.245031544698939, + "learning_rate": 1.9990027355769715e-05, + "loss": 2.1811, + "step": 404 + }, + { + "epoch": 0.21986970684039087, + "grad_norm": 2.381275606702572, + "learning_rate": 1.9989869714574784e-05, + "loss": 1.6356, + "step": 405 + }, + { + "epoch": 0.22041259500542887, + "grad_norm": 2.8847531540936053, + "learning_rate": 1.9989710837816846e-05, + "loss": 1.9953, + "step": 406 + }, + { + "epoch": 0.22095548317046687, + "grad_norm": 2.383074182943421, + "learning_rate": 1.9989550725515553e-05, + "loss": 1.5589, + "step": 407 + }, + { + "epoch": 0.22149837133550487, + "grad_norm": 2.3867744628533987, + "learning_rate": 1.99893893776907e-05, + "loss": 1.5351, + "step": 408 + }, + { + "epoch": 0.22204125950054288, + "grad_norm": 2.4139767232435463, + "learning_rate": 1.998922679436225e-05, + "loss": 1.5949, + "step": 409 + }, + { + "epoch": 0.22258414766558088, + "grad_norm": 3.034840719531937, + "learning_rate": 1.9989062975550313e-05, + "loss": 1.7994, + "step": 410 + }, + { + "epoch": 0.22312703583061888, + "grad_norm": 2.6245673280102544, + "learning_rate": 1.9988897921275144e-05, + "loss": 2.2593, + "step": 411 + }, + { + "epoch": 0.22366992399565688, + "grad_norm": 2.403934048820897, + "learning_rate": 1.998873163155716e-05, + "loss": 1.3179, + "step": 412 + }, + { + "epoch": 0.22421281216069489, + "grad_norm": 2.0174357446052524, + "learning_rate": 1.998856410641693e-05, + "loss": 1.8698, + "step": 413 + }, + { + "epoch": 0.2247557003257329, + "grad_norm": 2.25869172419699, + "learning_rate": 1.998839534587517e-05, + "loss": 1.596, + "step": 414 + }, + { + "epoch": 0.2252985884907709, + "grad_norm": 3.384596068816443, + "learning_rate": 1.9988225349952758e-05, + "loss": 1.5232, + "step": 415 + }, + { + "epoch": 0.2258414766558089, + "grad_norm": 2.2109577343653246, + "learning_rate": 1.9988054118670712e-05, + "loss": 1.2403, + "step": 416 + }, + { + "epoch": 0.2263843648208469, + "grad_norm": 2.1825505717325053, + "learning_rate": 1.9987881652050215e-05, + "loss": 1.7205, + "step": 417 + }, + { + "epoch": 0.2269272529858849, + "grad_norm": 2.7992033856118583, + "learning_rate": 1.99877079501126e-05, + "loss": 1.7486, + "step": 418 + }, + { + "epoch": 0.2274701411509229, + "grad_norm": 2.6351781282432793, + "learning_rate": 1.9987533012879344e-05, + "loss": 1.6693, + "step": 419 + }, + { + "epoch": 0.2280130293159609, + "grad_norm": 2.2385207734696384, + "learning_rate": 1.9987356840372088e-05, + "loss": 1.1296, + "step": 420 + }, + { + "epoch": 0.2285559174809989, + "grad_norm": 2.4921881643709107, + "learning_rate": 1.998717943261262e-05, + "loss": 1.1379, + "step": 421 + }, + { + "epoch": 0.2290988056460369, + "grad_norm": 2.842685080646979, + "learning_rate": 1.9987000789622884e-05, + "loss": 1.7963, + "step": 422 + }, + { + "epoch": 0.2296416938110749, + "grad_norm": 2.4776027353414145, + "learning_rate": 1.9986820911424972e-05, + "loss": 1.6939, + "step": 423 + }, + { + "epoch": 0.2301845819761129, + "grad_norm": 2.688785422525701, + "learning_rate": 1.9986639798041134e-05, + "loss": 1.2282, + "step": 424 + }, + { + "epoch": 0.23072747014115091, + "grad_norm": 3.4758865457335495, + "learning_rate": 1.998645744949377e-05, + "loss": 1.7664, + "step": 425 + }, + { + "epoch": 0.23127035830618892, + "grad_norm": 2.5586740557561605, + "learning_rate": 1.9986273865805432e-05, + "loss": 1.5637, + "step": 426 + }, + { + "epoch": 0.23181324647122692, + "grad_norm": 2.645306116903878, + "learning_rate": 1.9986089046998827e-05, + "loss": 1.5331, + "step": 427 + }, + { + "epoch": 0.23235613463626492, + "grad_norm": 2.748145883999443, + "learning_rate": 1.998590299309681e-05, + "loss": 1.7316, + "step": 428 + }, + { + "epoch": 0.23289902280130292, + "grad_norm": 2.4357923833746438, + "learning_rate": 1.99857157041224e-05, + "loss": 1.1479, + "step": 429 + }, + { + "epoch": 0.23344191096634093, + "grad_norm": 2.7322633978331377, + "learning_rate": 1.9985527180098755e-05, + "loss": 1.6089, + "step": 430 + }, + { + "epoch": 0.23398479913137893, + "grad_norm": 2.625531215380387, + "learning_rate": 1.9985337421049193e-05, + "loss": 1.8973, + "step": 431 + }, + { + "epoch": 0.23452768729641693, + "grad_norm": 2.1586487666504754, + "learning_rate": 1.9985146426997185e-05, + "loss": 1.664, + "step": 432 + }, + { + "epoch": 0.23507057546145493, + "grad_norm": 2.748704307086674, + "learning_rate": 1.9984954197966355e-05, + "loss": 1.677, + "step": 433 + }, + { + "epoch": 0.23561346362649294, + "grad_norm": 2.345532526959197, + "learning_rate": 1.9984760733980476e-05, + "loss": 1.5133, + "step": 434 + }, + { + "epoch": 0.23615635179153094, + "grad_norm": 2.1655585346308848, + "learning_rate": 1.9984566035063473e-05, + "loss": 1.4206, + "step": 435 + }, + { + "epoch": 0.23669923995656894, + "grad_norm": 2.697584333049271, + "learning_rate": 1.9984370101239434e-05, + "loss": 1.8131, + "step": 436 + }, + { + "epoch": 0.23724212812160694, + "grad_norm": 2.235069158786981, + "learning_rate": 1.9984172932532583e-05, + "loss": 1.7839, + "step": 437 + }, + { + "epoch": 0.23778501628664495, + "grad_norm": 2.548562657856099, + "learning_rate": 1.998397452896731e-05, + "loss": 1.4876, + "step": 438 + }, + { + "epoch": 0.23832790445168295, + "grad_norm": 2.1698874138883673, + "learning_rate": 1.9983774890568163e-05, + "loss": 1.2916, + "step": 439 + }, + { + "epoch": 0.23887079261672095, + "grad_norm": 2.3237954527681084, + "learning_rate": 1.998357401735982e-05, + "loss": 1.398, + "step": 440 + }, + { + "epoch": 0.23941368078175895, + "grad_norm": 2.448364938314344, + "learning_rate": 1.9983371909367135e-05, + "loss": 1.2663, + "step": 441 + }, + { + "epoch": 0.23995656894679696, + "grad_norm": 1.935899726785714, + "learning_rate": 1.99831685666151e-05, + "loss": 1.2448, + "step": 442 + }, + { + "epoch": 0.24049945711183496, + "grad_norm": 2.14357016947003, + "learning_rate": 1.9982963989128864e-05, + "loss": 1.4786, + "step": 443 + }, + { + "epoch": 0.24104234527687296, + "grad_norm": 2.5830135196717148, + "learning_rate": 1.998275817693373e-05, + "loss": 1.8324, + "step": 444 + }, + { + "epoch": 0.24158523344191096, + "grad_norm": 3.159289299734819, + "learning_rate": 1.9982551130055157e-05, + "loss": 1.9091, + "step": 445 + }, + { + "epoch": 0.24212812160694897, + "grad_norm": 3.2712239674501755, + "learning_rate": 1.9982342848518753e-05, + "loss": 1.7869, + "step": 446 + }, + { + "epoch": 0.24267100977198697, + "grad_norm": 3.35834841128443, + "learning_rate": 1.998213333235027e-05, + "loss": 1.2772, + "step": 447 + }, + { + "epoch": 0.24321389793702497, + "grad_norm": 2.9379666136057354, + "learning_rate": 1.998192258157563e-05, + "loss": 1.3497, + "step": 448 + }, + { + "epoch": 0.24375678610206297, + "grad_norm": 2.334119075845609, + "learning_rate": 1.9981710596220897e-05, + "loss": 1.5541, + "step": 449 + }, + { + "epoch": 0.24429967426710097, + "grad_norm": 2.781469856763384, + "learning_rate": 1.998149737631229e-05, + "loss": 1.9413, + "step": 450 + }, + { + "epoch": 0.24484256243213898, + "grad_norm": 3.0196781534130452, + "learning_rate": 1.9981282921876177e-05, + "loss": 1.3238, + "step": 451 + }, + { + "epoch": 0.24538545059717698, + "grad_norm": 3.547169321727429, + "learning_rate": 1.9981067232939086e-05, + "loss": 1.9952, + "step": 452 + }, + { + "epoch": 0.24592833876221498, + "grad_norm": 3.2411499530913535, + "learning_rate": 1.9980850309527693e-05, + "loss": 1.8244, + "step": 453 + }, + { + "epoch": 0.24647122692725298, + "grad_norm": 2.670313260104859, + "learning_rate": 1.9980632151668822e-05, + "loss": 1.607, + "step": 454 + }, + { + "epoch": 0.247014115092291, + "grad_norm": 2.8822100638306143, + "learning_rate": 1.9980412759389468e-05, + "loss": 1.8868, + "step": 455 + }, + { + "epoch": 0.247557003257329, + "grad_norm": 2.4776152417583317, + "learning_rate": 1.9980192132716748e-05, + "loss": 1.778, + "step": 456 + }, + { + "epoch": 0.248099891422367, + "grad_norm": 2.2001923672712076, + "learning_rate": 1.9979970271677967e-05, + "loss": 1.3544, + "step": 457 + }, + { + "epoch": 0.248642779587405, + "grad_norm": 2.7694932683911837, + "learning_rate": 1.9979747176300553e-05, + "loss": 1.6521, + "step": 458 + }, + { + "epoch": 0.249185667752443, + "grad_norm": 2.8464573075472845, + "learning_rate": 1.99795228466121e-05, + "loss": 1.8803, + "step": 459 + }, + { + "epoch": 0.249728555917481, + "grad_norm": 2.577989994947286, + "learning_rate": 1.9979297282640365e-05, + "loss": 1.8838, + "step": 460 + }, + { + "epoch": 0.250271444082519, + "grad_norm": 2.9859834231033164, + "learning_rate": 1.997907048441323e-05, + "loss": 1.5122, + "step": 461 + }, + { + "epoch": 0.250814332247557, + "grad_norm": 3.0857580735568098, + "learning_rate": 1.9978842451958757e-05, + "loss": 1.7789, + "step": 462 + }, + { + "epoch": 0.251357220412595, + "grad_norm": 2.523783848348806, + "learning_rate": 1.9978613185305145e-05, + "loss": 1.5815, + "step": 463 + }, + { + "epoch": 0.251900108577633, + "grad_norm": 2.699094362013039, + "learning_rate": 1.9978382684480747e-05, + "loss": 1.7448, + "step": 464 + }, + { + "epoch": 0.252442996742671, + "grad_norm": 2.736268550628698, + "learning_rate": 1.997815094951408e-05, + "loss": 1.7052, + "step": 465 + }, + { + "epoch": 0.252985884907709, + "grad_norm": 3.1150828428090014, + "learning_rate": 1.99779179804338e-05, + "loss": 1.7743, + "step": 466 + }, + { + "epoch": 0.253528773072747, + "grad_norm": 3.3513164619888482, + "learning_rate": 1.997768377726872e-05, + "loss": 2.3905, + "step": 467 + }, + { + "epoch": 0.254071661237785, + "grad_norm": 2.400886982379507, + "learning_rate": 1.9977448340047808e-05, + "loss": 1.6096, + "step": 468 + }, + { + "epoch": 0.254614549402823, + "grad_norm": 2.633533044966171, + "learning_rate": 1.9977211668800186e-05, + "loss": 1.7796, + "step": 469 + }, + { + "epoch": 0.255157437567861, + "grad_norm": 3.9297139007235042, + "learning_rate": 1.997697376355512e-05, + "loss": 1.4602, + "step": 470 + }, + { + "epoch": 0.255700325732899, + "grad_norm": 5.588260619709643, + "learning_rate": 1.9976734624342044e-05, + "loss": 2.0389, + "step": 471 + }, + { + "epoch": 0.256243213897937, + "grad_norm": 2.600723171476426, + "learning_rate": 1.9976494251190522e-05, + "loss": 1.3676, + "step": 472 + }, + { + "epoch": 0.25678610206297503, + "grad_norm": 2.61945002649116, + "learning_rate": 1.9976252644130297e-05, + "loss": 1.7902, + "step": 473 + }, + { + "epoch": 0.25732899022801303, + "grad_norm": 3.7513561790803838, + "learning_rate": 1.997600980319124e-05, + "loss": 1.5997, + "step": 474 + }, + { + "epoch": 0.25787187839305103, + "grad_norm": 2.8832359552778737, + "learning_rate": 1.9975765728403395e-05, + "loss": 1.9636, + "step": 475 + }, + { + "epoch": 0.25841476655808904, + "grad_norm": 3.0975018752600243, + "learning_rate": 1.9975520419796942e-05, + "loss": 1.1165, + "step": 476 + }, + { + "epoch": 0.25895765472312704, + "grad_norm": 2.121708502818221, + "learning_rate": 1.9975273877402227e-05, + "loss": 1.5108, + "step": 477 + }, + { + "epoch": 0.25950054288816504, + "grad_norm": 2.4073592870530116, + "learning_rate": 1.997502610124974e-05, + "loss": 1.7828, + "step": 478 + }, + { + "epoch": 0.26004343105320304, + "grad_norm": 2.6858679986632974, + "learning_rate": 1.997477709137013e-05, + "loss": 1.8483, + "step": 479 + }, + { + "epoch": 0.26058631921824105, + "grad_norm": 3.283752190131325, + "learning_rate": 1.997452684779419e-05, + "loss": 1.6105, + "step": 480 + }, + { + "epoch": 0.26112920738327905, + "grad_norm": 2.500181185675909, + "learning_rate": 1.997427537055287e-05, + "loss": 1.5475, + "step": 481 + }, + { + "epoch": 0.26167209554831705, + "grad_norm": 2.3992324550953885, + "learning_rate": 1.9974022659677278e-05, + "loss": 1.6062, + "step": 482 + }, + { + "epoch": 0.26221498371335505, + "grad_norm": 2.7692293822867837, + "learning_rate": 1.9973768715198667e-05, + "loss": 1.4995, + "step": 483 + }, + { + "epoch": 0.26275787187839306, + "grad_norm": 2.971423364277874, + "learning_rate": 1.9973513537148447e-05, + "loss": 1.7904, + "step": 484 + }, + { + "epoch": 0.26330076004343106, + "grad_norm": 2.2769736321644105, + "learning_rate": 1.9973257125558177e-05, + "loss": 1.3121, + "step": 485 + }, + { + "epoch": 0.26384364820846906, + "grad_norm": 2.5949046051899254, + "learning_rate": 1.997299948045957e-05, + "loss": 1.4555, + "step": 486 + }, + { + "epoch": 0.26438653637350706, + "grad_norm": 3.0210593683445204, + "learning_rate": 1.997274060188449e-05, + "loss": 1.8211, + "step": 487 + }, + { + "epoch": 0.26492942453854507, + "grad_norm": 3.076626224616319, + "learning_rate": 1.9972480489864962e-05, + "loss": 1.6366, + "step": 488 + }, + { + "epoch": 0.26547231270358307, + "grad_norm": 3.1369728365663536, + "learning_rate": 1.9972219144433148e-05, + "loss": 1.5027, + "step": 489 + }, + { + "epoch": 0.26601520086862107, + "grad_norm": 2.2870450349164635, + "learning_rate": 1.9971956565621383e-05, + "loss": 1.2784, + "step": 490 + }, + { + "epoch": 0.2665580890336591, + "grad_norm": 3.0130036065633776, + "learning_rate": 1.9971692753462134e-05, + "loss": 1.2083, + "step": 491 + }, + { + "epoch": 0.2671009771986971, + "grad_norm": 2.6063157323029733, + "learning_rate": 1.9971427707988034e-05, + "loss": 1.0083, + "step": 492 + }, + { + "epoch": 0.2676438653637351, + "grad_norm": 2.579371053895234, + "learning_rate": 1.997116142923186e-05, + "loss": 1.1937, + "step": 493 + }, + { + "epoch": 0.2681867535287731, + "grad_norm": 2.532537971800688, + "learning_rate": 1.9970893917226554e-05, + "loss": 1.4735, + "step": 494 + }, + { + "epoch": 0.2687296416938111, + "grad_norm": 2.3483150144294105, + "learning_rate": 1.997062517200519e-05, + "loss": 1.7269, + "step": 495 + }, + { + "epoch": 0.2692725298588491, + "grad_norm": 2.594809867192747, + "learning_rate": 1.997035519360102e-05, + "loss": 1.8283, + "step": 496 + }, + { + "epoch": 0.2698154180238871, + "grad_norm": 2.613813750609998, + "learning_rate": 1.9970083982047428e-05, + "loss": 1.2302, + "step": 497 + }, + { + "epoch": 0.2703583061889251, + "grad_norm": 2.648279162964909, + "learning_rate": 1.9969811537377956e-05, + "loss": 1.6225, + "step": 498 + }, + { + "epoch": 0.2709011943539631, + "grad_norm": 2.3790190706794325, + "learning_rate": 1.9969537859626308e-05, + "loss": 1.5172, + "step": 499 + }, + { + "epoch": 0.2714440825190011, + "grad_norm": 2.7054998578606364, + "learning_rate": 1.9969262948826326e-05, + "loss": 1.4525, + "step": 500 + }, + { + "epoch": 0.2719869706840391, + "grad_norm": 2.412151508264948, + "learning_rate": 1.9968986805012012e-05, + "loss": 1.3299, + "step": 501 + }, + { + "epoch": 0.2725298588490771, + "grad_norm": 2.4836460319285414, + "learning_rate": 1.9968709428217525e-05, + "loss": 1.5217, + "step": 502 + }, + { + "epoch": 0.2730727470141151, + "grad_norm": 3.3112944949184606, + "learning_rate": 1.9968430818477168e-05, + "loss": 2.0643, + "step": 503 + }, + { + "epoch": 0.2736156351791531, + "grad_norm": 3.266395629954733, + "learning_rate": 1.9968150975825397e-05, + "loss": 1.1719, + "step": 504 + }, + { + "epoch": 0.2741585233441911, + "grad_norm": 2.3487351103507073, + "learning_rate": 1.996786990029683e-05, + "loss": 1.6876, + "step": 505 + }, + { + "epoch": 0.2747014115092291, + "grad_norm": 3.0672442719402673, + "learning_rate": 1.9967587591926227e-05, + "loss": 1.3946, + "step": 506 + }, + { + "epoch": 0.2752442996742671, + "grad_norm": 2.84320621483769, + "learning_rate": 1.99673040507485e-05, + "loss": 1.3259, + "step": 507 + }, + { + "epoch": 0.2757871878393051, + "grad_norm": 2.5078449617010707, + "learning_rate": 1.9967019276798728e-05, + "loss": 0.8478, + "step": 508 + }, + { + "epoch": 0.2763300760043431, + "grad_norm": 3.137784699454796, + "learning_rate": 1.9966733270112126e-05, + "loss": 1.1688, + "step": 509 + }, + { + "epoch": 0.2768729641693811, + "grad_norm": 3.7277826491955017, + "learning_rate": 1.996644603072407e-05, + "loss": 1.1091, + "step": 510 + }, + { + "epoch": 0.2774158523344191, + "grad_norm": 2.5276781604415635, + "learning_rate": 1.996615755867008e-05, + "loss": 1.1299, + "step": 511 + }, + { + "epoch": 0.2779587404994571, + "grad_norm": 3.1012700661738744, + "learning_rate": 1.996586785398584e-05, + "loss": 1.7218, + "step": 512 + }, + { + "epoch": 0.2785016286644951, + "grad_norm": 2.285166712515903, + "learning_rate": 1.9965576916707182e-05, + "loss": 1.2868, + "step": 513 + }, + { + "epoch": 0.27904451682953313, + "grad_norm": 2.33097906349044, + "learning_rate": 1.9965284746870088e-05, + "loss": 0.9887, + "step": 514 + }, + { + "epoch": 0.27958740499457113, + "grad_norm": 2.6473787082237927, + "learning_rate": 1.9964991344510697e-05, + "loss": 1.8543, + "step": 515 + }, + { + "epoch": 0.28013029315960913, + "grad_norm": 2.4628160599533366, + "learning_rate": 1.996469670966529e-05, + "loss": 1.2263, + "step": 516 + }, + { + "epoch": 0.28067318132464714, + "grad_norm": 3.2897847068350905, + "learning_rate": 1.9964400842370314e-05, + "loss": 1.6338, + "step": 517 + }, + { + "epoch": 0.28121606948968514, + "grad_norm": 2.4439319341540324, + "learning_rate": 1.9964103742662363e-05, + "loss": 1.0836, + "step": 518 + }, + { + "epoch": 0.28175895765472314, + "grad_norm": 2.3221991020412003, + "learning_rate": 1.996380541057818e-05, + "loss": 1.2331, + "step": 519 + }, + { + "epoch": 0.28230184581976114, + "grad_norm": 2.9571040634251564, + "learning_rate": 1.9963505846154662e-05, + "loss": 1.3066, + "step": 520 + }, + { + "epoch": 0.28284473398479915, + "grad_norm": 3.7512706020225624, + "learning_rate": 1.996320504942886e-05, + "loss": 1.7482, + "step": 521 + }, + { + "epoch": 0.28338762214983715, + "grad_norm": 2.4620109793388267, + "learning_rate": 1.9962903020437983e-05, + "loss": 1.5334, + "step": 522 + }, + { + "epoch": 0.28393051031487515, + "grad_norm": 3.030374272795485, + "learning_rate": 1.9962599759219383e-05, + "loss": 1.8957, + "step": 523 + }, + { + "epoch": 0.28447339847991315, + "grad_norm": 2.452389821491403, + "learning_rate": 1.9962295265810563e-05, + "loss": 1.5438, + "step": 524 + }, + { + "epoch": 0.28501628664495116, + "grad_norm": 2.713028369466205, + "learning_rate": 1.996198954024919e-05, + "loss": 1.4272, + "step": 525 + }, + { + "epoch": 0.28555917480998916, + "grad_norm": 3.810321275175567, + "learning_rate": 1.996168258257307e-05, + "loss": 1.9028, + "step": 526 + }, + { + "epoch": 0.28610206297502716, + "grad_norm": 2.7774100977441236, + "learning_rate": 1.9961374392820173e-05, + "loss": 1.5644, + "step": 527 + }, + { + "epoch": 0.28664495114006516, + "grad_norm": 2.6798712089104186, + "learning_rate": 1.9961064971028616e-05, + "loss": 1.188, + "step": 528 + }, + { + "epoch": 0.28718783930510317, + "grad_norm": 2.889465990486677, + "learning_rate": 1.9960754317236666e-05, + "loss": 1.5393, + "step": 529 + }, + { + "epoch": 0.28773072747014117, + "grad_norm": 3.465251366831076, + "learning_rate": 1.996044243148275e-05, + "loss": 2.1899, + "step": 530 + }, + { + "epoch": 0.28827361563517917, + "grad_norm": 2.645941940974219, + "learning_rate": 1.9960129313805437e-05, + "loss": 1.3691, + "step": 531 + }, + { + "epoch": 0.2888165038002172, + "grad_norm": 2.3914199977194293, + "learning_rate": 1.9959814964243455e-05, + "loss": 1.5219, + "step": 532 + }, + { + "epoch": 0.2893593919652552, + "grad_norm": 2.3023810529281343, + "learning_rate": 1.995949938283569e-05, + "loss": 1.5147, + "step": 533 + }, + { + "epoch": 0.2899022801302932, + "grad_norm": 2.7362205671791155, + "learning_rate": 1.9959182569621164e-05, + "loss": 1.7571, + "step": 534 + }, + { + "epoch": 0.2904451682953312, + "grad_norm": 3.971162331076012, + "learning_rate": 1.9958864524639066e-05, + "loss": 1.3425, + "step": 535 + }, + { + "epoch": 0.2909880564603692, + "grad_norm": 3.4144928239616514, + "learning_rate": 1.9958545247928727e-05, + "loss": 1.6962, + "step": 536 + }, + { + "epoch": 0.2915309446254072, + "grad_norm": 3.5063126675319043, + "learning_rate": 1.9958224739529647e-05, + "loss": 1.6406, + "step": 537 + }, + { + "epoch": 0.2920738327904452, + "grad_norm": 2.9013783116047547, + "learning_rate": 1.995790299948146e-05, + "loss": 1.6376, + "step": 538 + }, + { + "epoch": 0.2926167209554832, + "grad_norm": 2.4827691033904693, + "learning_rate": 1.9957580027823957e-05, + "loss": 1.8672, + "step": 539 + }, + { + "epoch": 0.2931596091205212, + "grad_norm": 3.1245563354940242, + "learning_rate": 1.9957255824597087e-05, + "loss": 1.885, + "step": 540 + }, + { + "epoch": 0.2937024972855592, + "grad_norm": 2.7941860466759896, + "learning_rate": 1.9956930389840945e-05, + "loss": 1.0903, + "step": 541 + }, + { + "epoch": 0.2942453854505972, + "grad_norm": 3.1348104146717772, + "learning_rate": 1.9956603723595784e-05, + "loss": 2.1446, + "step": 542 + }, + { + "epoch": 0.2947882736156352, + "grad_norm": 2.4248222663066747, + "learning_rate": 1.995627582590201e-05, + "loss": 1.3503, + "step": 543 + }, + { + "epoch": 0.2953311617806732, + "grad_norm": 3.1538245864476337, + "learning_rate": 1.995594669680017e-05, + "loss": 1.3275, + "step": 544 + }, + { + "epoch": 0.2958740499457112, + "grad_norm": 3.436905240480997, + "learning_rate": 1.9955616336330976e-05, + "loss": 1.7249, + "step": 545 + }, + { + "epoch": 0.2964169381107492, + "grad_norm": 3.9145352507266393, + "learning_rate": 1.9955284744535287e-05, + "loss": 2.0089, + "step": 546 + }, + { + "epoch": 0.2969598262757872, + "grad_norm": 2.8600509325831185, + "learning_rate": 1.9954951921454113e-05, + "loss": 1.4527, + "step": 547 + }, + { + "epoch": 0.2975027144408252, + "grad_norm": 2.865760632888347, + "learning_rate": 1.995461786712862e-05, + "loss": 1.4553, + "step": 548 + }, + { + "epoch": 0.2980456026058632, + "grad_norm": 3.3559882259900706, + "learning_rate": 1.9954282581600127e-05, + "loss": 2.0456, + "step": 549 + }, + { + "epoch": 0.2985884907709012, + "grad_norm": 3.3424656149019008, + "learning_rate": 1.9953946064910098e-05, + "loss": 1.8253, + "step": 550 + }, + { + "epoch": 0.2991313789359392, + "grad_norm": 2.569119318410074, + "learning_rate": 1.9953608317100153e-05, + "loss": 1.3623, + "step": 551 + }, + { + "epoch": 0.2996742671009772, + "grad_norm": 2.8918953822102424, + "learning_rate": 1.995326933821207e-05, + "loss": 1.7521, + "step": 552 + }, + { + "epoch": 0.3002171552660152, + "grad_norm": 2.7627900710192246, + "learning_rate": 1.995292912828777e-05, + "loss": 1.775, + "step": 553 + }, + { + "epoch": 0.3007600434310532, + "grad_norm": 2.8053609300694804, + "learning_rate": 1.9952587687369334e-05, + "loss": 1.7536, + "step": 554 + }, + { + "epoch": 0.30130293159609123, + "grad_norm": 2.9775715301146803, + "learning_rate": 1.995224501549899e-05, + "loss": 1.6715, + "step": 555 + }, + { + "epoch": 0.30184581976112923, + "grad_norm": 3.146995410263436, + "learning_rate": 1.9951901112719123e-05, + "loss": 1.1032, + "step": 556 + }, + { + "epoch": 0.30238870792616723, + "grad_norm": 3.219197817112143, + "learning_rate": 1.9951555979072266e-05, + "loss": 1.6326, + "step": 557 + }, + { + "epoch": 0.30293159609120524, + "grad_norm": 2.676508071644292, + "learning_rate": 1.99512096146011e-05, + "loss": 1.4836, + "step": 558 + }, + { + "epoch": 0.30347448425624324, + "grad_norm": 3.8806900857620374, + "learning_rate": 1.9950862019348474e-05, + "loss": 1.7794, + "step": 559 + }, + { + "epoch": 0.30401737242128124, + "grad_norm": 2.547409032322543, + "learning_rate": 1.995051319335737e-05, + "loss": 1.3263, + "step": 560 + }, + { + "epoch": 0.30456026058631924, + "grad_norm": 3.025189850713409, + "learning_rate": 1.995016313667094e-05, + "loss": 1.2409, + "step": 561 + }, + { + "epoch": 0.30510314875135724, + "grad_norm": 3.3644665856402614, + "learning_rate": 1.9949811849332476e-05, + "loss": 1.2988, + "step": 562 + }, + { + "epoch": 0.30564603691639525, + "grad_norm": 3.4126932100522755, + "learning_rate": 1.9949459331385422e-05, + "loss": 1.6126, + "step": 563 + }, + { + "epoch": 0.30618892508143325, + "grad_norm": 3.3751971677759416, + "learning_rate": 1.994910558287338e-05, + "loss": 1.6243, + "step": 564 + }, + { + "epoch": 0.30673181324647125, + "grad_norm": 3.0948952637866105, + "learning_rate": 1.9948750603840102e-05, + "loss": 1.6553, + "step": 565 + }, + { + "epoch": 0.30727470141150925, + "grad_norm": 3.1116929004314224, + "learning_rate": 1.9948394394329494e-05, + "loss": 1.0466, + "step": 566 + }, + { + "epoch": 0.30781758957654726, + "grad_norm": 2.649244034262683, + "learning_rate": 1.9948036954385613e-05, + "loss": 1.2914, + "step": 567 + }, + { + "epoch": 0.30836047774158526, + "grad_norm": 3.9338023936000965, + "learning_rate": 1.9947678284052667e-05, + "loss": 1.7532, + "step": 568 + }, + { + "epoch": 0.30890336590662326, + "grad_norm": 2.887133065759567, + "learning_rate": 1.9947318383375017e-05, + "loss": 1.8001, + "step": 569 + }, + { + "epoch": 0.30944625407166126, + "grad_norm": 3.6996119456889915, + "learning_rate": 1.9946957252397173e-05, + "loss": 2.4852, + "step": 570 + }, + { + "epoch": 0.30998914223669927, + "grad_norm": 2.829855772438557, + "learning_rate": 1.9946594891163808e-05, + "loss": 1.6048, + "step": 571 + }, + { + "epoch": 0.31053203040173727, + "grad_norm": 4.288772356209683, + "learning_rate": 1.9946231299719732e-05, + "loss": 1.4841, + "step": 572 + }, + { + "epoch": 0.31107491856677527, + "grad_norm": 2.37958986026152, + "learning_rate": 1.9945866478109914e-05, + "loss": 1.1797, + "step": 573 + }, + { + "epoch": 0.3116178067318133, + "grad_norm": 2.9977970906442932, + "learning_rate": 1.9945500426379483e-05, + "loss": 1.224, + "step": 574 + }, + { + "epoch": 0.3121606948968513, + "grad_norm": 3.406596416686285, + "learning_rate": 1.9945133144573705e-05, + "loss": 1.4793, + "step": 575 + }, + { + "epoch": 0.3127035830618892, + "grad_norm": 3.3772239188722244, + "learning_rate": 1.994476463273801e-05, + "loss": 1.3696, + "step": 576 + }, + { + "epoch": 0.3132464712269272, + "grad_norm": 2.729299768057245, + "learning_rate": 1.9944394890917977e-05, + "loss": 1.448, + "step": 577 + }, + { + "epoch": 0.31378935939196523, + "grad_norm": 2.6526406591248297, + "learning_rate": 1.9944023919159335e-05, + "loss": 1.6905, + "step": 578 + }, + { + "epoch": 0.31433224755700323, + "grad_norm": 2.9512501195242944, + "learning_rate": 1.9943651717507965e-05, + "loss": 1.6277, + "step": 579 + }, + { + "epoch": 0.31487513572204123, + "grad_norm": 2.8292191327941723, + "learning_rate": 1.9943278286009903e-05, + "loss": 1.0532, + "step": 580 + }, + { + "epoch": 0.31541802388707924, + "grad_norm": 3.2105870294745436, + "learning_rate": 1.9942903624711335e-05, + "loss": 1.1823, + "step": 581 + }, + { + "epoch": 0.31596091205211724, + "grad_norm": 3.224137115744835, + "learning_rate": 1.9942527733658602e-05, + "loss": 1.5409, + "step": 582 + }, + { + "epoch": 0.31650380021715524, + "grad_norm": 3.126134123334164, + "learning_rate": 1.9942150612898194e-05, + "loss": 1.2423, + "step": 583 + }, + { + "epoch": 0.31704668838219324, + "grad_norm": 2.709954974374804, + "learning_rate": 1.994177226247675e-05, + "loss": 1.3419, + "step": 584 + }, + { + "epoch": 0.31758957654723124, + "grad_norm": 2.9030367366777927, + "learning_rate": 1.9941392682441066e-05, + "loss": 1.427, + "step": 585 + }, + { + "epoch": 0.31813246471226925, + "grad_norm": 3.272875981776567, + "learning_rate": 1.9941011872838092e-05, + "loss": 2.0196, + "step": 586 + }, + { + "epoch": 0.31867535287730725, + "grad_norm": 3.3480185179867, + "learning_rate": 1.994062983371493e-05, + "loss": 1.6038, + "step": 587 + }, + { + "epoch": 0.31921824104234525, + "grad_norm": 2.801738772846361, + "learning_rate": 1.9940246565118822e-05, + "loss": 1.7505, + "step": 588 + }, + { + "epoch": 0.31976112920738325, + "grad_norm": 2.743550967049156, + "learning_rate": 1.993986206709718e-05, + "loss": 1.3914, + "step": 589 + }, + { + "epoch": 0.32030401737242126, + "grad_norm": 2.481845489278486, + "learning_rate": 1.9939476339697555e-05, + "loss": 1.6927, + "step": 590 + }, + { + "epoch": 0.32084690553745926, + "grad_norm": 3.145019330878407, + "learning_rate": 1.993908938296765e-05, + "loss": 1.42, + "step": 591 + }, + { + "epoch": 0.32138979370249726, + "grad_norm": 3.326427662456394, + "learning_rate": 1.9938701196955335e-05, + "loss": 1.2516, + "step": 592 + }, + { + "epoch": 0.32193268186753526, + "grad_norm": 3.195815443331326, + "learning_rate": 1.9938311781708616e-05, + "loss": 2.1428, + "step": 593 + }, + { + "epoch": 0.32247557003257327, + "grad_norm": 2.461395361556941, + "learning_rate": 1.9937921137275657e-05, + "loss": 1.4448, + "step": 594 + }, + { + "epoch": 0.32301845819761127, + "grad_norm": 3.1793351214549794, + "learning_rate": 1.993752926370477e-05, + "loss": 1.4609, + "step": 595 + }, + { + "epoch": 0.32356134636264927, + "grad_norm": 2.783909288864463, + "learning_rate": 1.9937136161044427e-05, + "loss": 1.3355, + "step": 596 + }, + { + "epoch": 0.3241042345276873, + "grad_norm": 2.8156929696256734, + "learning_rate": 1.9936741829343247e-05, + "loss": 2.101, + "step": 597 + }, + { + "epoch": 0.3246471226927253, + "grad_norm": 3.793352093788154, + "learning_rate": 1.993634626865e-05, + "loss": 2.19, + "step": 598 + }, + { + "epoch": 0.3251900108577633, + "grad_norm": 2.6680863350639545, + "learning_rate": 1.993594947901361e-05, + "loss": 1.5199, + "step": 599 + }, + { + "epoch": 0.3257328990228013, + "grad_norm": 2.6498586813134297, + "learning_rate": 1.9935551460483155e-05, + "loss": 0.9282, + "step": 600 + }, + { + "epoch": 0.3262757871878393, + "grad_norm": 3.443063689484479, + "learning_rate": 1.993515221310786e-05, + "loss": 1.9107, + "step": 601 + }, + { + "epoch": 0.3268186753528773, + "grad_norm": 2.45686855619251, + "learning_rate": 1.9934751736937103e-05, + "loss": 1.2929, + "step": 602 + }, + { + "epoch": 0.3273615635179153, + "grad_norm": 3.3974155688994077, + "learning_rate": 1.9934350032020417e-05, + "loss": 1.546, + "step": 603 + }, + { + "epoch": 0.3279044516829533, + "grad_norm": 2.3692191456624783, + "learning_rate": 1.993394709840749e-05, + "loss": 1.2239, + "step": 604 + }, + { + "epoch": 0.3284473398479913, + "grad_norm": 4.044352925058249, + "learning_rate": 1.993354293614815e-05, + "loss": 1.801, + "step": 605 + }, + { + "epoch": 0.3289902280130293, + "grad_norm": 4.051192057617293, + "learning_rate": 1.993313754529239e-05, + "loss": 1.8043, + "step": 606 + }, + { + "epoch": 0.3295331161780673, + "grad_norm": 3.3680281079594634, + "learning_rate": 1.9932730925890344e-05, + "loss": 1.4915, + "step": 607 + }, + { + "epoch": 0.3300760043431053, + "grad_norm": 2.993677842102555, + "learning_rate": 1.9932323077992312e-05, + "loss": 1.3457, + "step": 608 + }, + { + "epoch": 0.3306188925081433, + "grad_norm": 3.325666927669253, + "learning_rate": 1.9931914001648726e-05, + "loss": 1.4221, + "step": 609 + }, + { + "epoch": 0.3311617806731813, + "grad_norm": 2.459382097232459, + "learning_rate": 1.993150369691019e-05, + "loss": 1.2551, + "step": 610 + }, + { + "epoch": 0.3317046688382193, + "grad_norm": 5.281638549560053, + "learning_rate": 1.993109216382745e-05, + "loss": 1.142, + "step": 611 + }, + { + "epoch": 0.3322475570032573, + "grad_norm": 2.6931033883715374, + "learning_rate": 1.99306794024514e-05, + "loss": 1.2573, + "step": 612 + }, + { + "epoch": 0.3327904451682953, + "grad_norm": 2.8066393604642714, + "learning_rate": 1.9930265412833097e-05, + "loss": 0.9847, + "step": 613 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 3.682172765988149, + "learning_rate": 1.992985019502374e-05, + "loss": 1.4482, + "step": 614 + }, + { + "epoch": 0.3338762214983713, + "grad_norm": 3.7190248655227873, + "learning_rate": 1.9929433749074684e-05, + "loss": 1.3682, + "step": 615 + }, + { + "epoch": 0.3344191096634093, + "grad_norm": 2.8594798198555527, + "learning_rate": 1.9929016075037438e-05, + "loss": 1.2396, + "step": 616 + }, + { + "epoch": 0.3349619978284473, + "grad_norm": 3.5439348644341035, + "learning_rate": 1.992859717296366e-05, + "loss": 1.5843, + "step": 617 + }, + { + "epoch": 0.3355048859934853, + "grad_norm": 2.3447575030783283, + "learning_rate": 1.992817704290516e-05, + "loss": 1.2025, + "step": 618 + }, + { + "epoch": 0.3360477741585233, + "grad_norm": 3.027039186876229, + "learning_rate": 1.99277556849139e-05, + "loss": 1.7133, + "step": 619 + }, + { + "epoch": 0.33659066232356133, + "grad_norm": 2.9868315770124343, + "learning_rate": 1.9927333099041992e-05, + "loss": 1.2309, + "step": 620 + }, + { + "epoch": 0.33713355048859933, + "grad_norm": 3.215889472723888, + "learning_rate": 1.9926909285341705e-05, + "loss": 1.2217, + "step": 621 + }, + { + "epoch": 0.33767643865363733, + "grad_norm": 3.293119071439456, + "learning_rate": 1.992648424386546e-05, + "loss": 1.7828, + "step": 622 + }, + { + "epoch": 0.33821932681867534, + "grad_norm": 2.7023629491691623, + "learning_rate": 1.992605797466582e-05, + "loss": 1.3745, + "step": 623 + }, + { + "epoch": 0.33876221498371334, + "grad_norm": 4.234534944070885, + "learning_rate": 1.9925630477795514e-05, + "loss": 1.3415, + "step": 624 + }, + { + "epoch": 0.33930510314875134, + "grad_norm": 3.6994706259321815, + "learning_rate": 1.9925201753307406e-05, + "loss": 1.4234, + "step": 625 + }, + { + "epoch": 0.33984799131378934, + "grad_norm": 3.156992028368943, + "learning_rate": 1.992477180125453e-05, + "loss": 1.4845, + "step": 626 + }, + { + "epoch": 0.34039087947882735, + "grad_norm": 3.1286393814611397, + "learning_rate": 1.9924340621690058e-05, + "loss": 0.9685, + "step": 627 + }, + { + "epoch": 0.34093376764386535, + "grad_norm": 3.336970766050665, + "learning_rate": 1.9923908214667323e-05, + "loss": 1.3268, + "step": 628 + }, + { + "epoch": 0.34147665580890335, + "grad_norm": 4.290496076646803, + "learning_rate": 1.99234745802398e-05, + "loss": 1.2785, + "step": 629 + }, + { + "epoch": 0.34201954397394135, + "grad_norm": 2.7491777959300023, + "learning_rate": 1.9923039718461127e-05, + "loss": 1.3798, + "step": 630 + }, + { + "epoch": 0.34256243213897936, + "grad_norm": 5.312529540414799, + "learning_rate": 1.992260362938509e-05, + "loss": 2.2512, + "step": 631 + }, + { + "epoch": 0.34310532030401736, + "grad_norm": 3.9557409230221676, + "learning_rate": 1.9922166313065618e-05, + "loss": 1.849, + "step": 632 + }, + { + "epoch": 0.34364820846905536, + "grad_norm": 3.7327574664165786, + "learning_rate": 1.9921727769556805e-05, + "loss": 1.4078, + "step": 633 + }, + { + "epoch": 0.34419109663409336, + "grad_norm": 3.7644055867642594, + "learning_rate": 1.992128799891289e-05, + "loss": 1.714, + "step": 634 + }, + { + "epoch": 0.34473398479913137, + "grad_norm": 3.923952524278675, + "learning_rate": 1.9920847001188258e-05, + "loss": 1.7613, + "step": 635 + }, + { + "epoch": 0.34527687296416937, + "grad_norm": 3.1523646374054155, + "learning_rate": 1.992040477643746e-05, + "loss": 1.2563, + "step": 636 + }, + { + "epoch": 0.34581976112920737, + "grad_norm": 3.6157151039083444, + "learning_rate": 1.991996132471519e-05, + "loss": 1.1795, + "step": 637 + }, + { + "epoch": 0.3463626492942454, + "grad_norm": 4.222007576307266, + "learning_rate": 1.991951664607629e-05, + "loss": 1.6605, + "step": 638 + }, + { + "epoch": 0.3469055374592834, + "grad_norm": 3.1555785318795544, + "learning_rate": 1.9919070740575764e-05, + "loss": 1.526, + "step": 639 + }, + { + "epoch": 0.3474484256243214, + "grad_norm": 3.2645269107818677, + "learning_rate": 1.991862360826876e-05, + "loss": 1.3182, + "step": 640 + }, + { + "epoch": 0.3479913137893594, + "grad_norm": 3.53044955128402, + "learning_rate": 1.991817524921058e-05, + "loss": 1.5583, + "step": 641 + }, + { + "epoch": 0.3485342019543974, + "grad_norm": 3.24710260787145, + "learning_rate": 1.9917725663456678e-05, + "loss": 0.9878, + "step": 642 + }, + { + "epoch": 0.3490770901194354, + "grad_norm": 3.9154742975791654, + "learning_rate": 1.991727485106266e-05, + "loss": 1.6034, + "step": 643 + }, + { + "epoch": 0.3496199782844734, + "grad_norm": 3.8199965805178318, + "learning_rate": 1.9916822812084282e-05, + "loss": 1.3768, + "step": 644 + }, + { + "epoch": 0.3501628664495114, + "grad_norm": 3.198793670499035, + "learning_rate": 1.9916369546577455e-05, + "loss": 1.4308, + "step": 645 + }, + { + "epoch": 0.3507057546145494, + "grad_norm": 2.908177540585374, + "learning_rate": 1.9915915054598237e-05, + "loss": 1.3964, + "step": 646 + }, + { + "epoch": 0.3512486427795874, + "grad_norm": 2.920889003436479, + "learning_rate": 1.9915459336202844e-05, + "loss": 1.056, + "step": 647 + }, + { + "epoch": 0.3517915309446254, + "grad_norm": 3.7046400158340864, + "learning_rate": 1.991500239144763e-05, + "loss": 1.9052, + "step": 648 + }, + { + "epoch": 0.3523344191096634, + "grad_norm": 4.412988121206581, + "learning_rate": 1.9914544220389124e-05, + "loss": 1.944, + "step": 649 + }, + { + "epoch": 0.3528773072747014, + "grad_norm": 3.87124697251994, + "learning_rate": 1.9914084823083988e-05, + "loss": 1.4951, + "step": 650 + }, + { + "epoch": 0.3534201954397394, + "grad_norm": 3.8734663848108584, + "learning_rate": 1.9913624199589037e-05, + "loss": 1.5462, + "step": 651 + }, + { + "epoch": 0.3539630836047774, + "grad_norm": 2.7283359997443126, + "learning_rate": 1.9913162349961248e-05, + "loss": 0.9188, + "step": 652 + }, + { + "epoch": 0.3545059717698154, + "grad_norm": 2.885333878853398, + "learning_rate": 1.991269927425774e-05, + "loss": 1.2292, + "step": 653 + }, + { + "epoch": 0.3550488599348534, + "grad_norm": 3.2425772460053257, + "learning_rate": 1.9912234972535788e-05, + "loss": 1.2863, + "step": 654 + }, + { + "epoch": 0.3555917480998914, + "grad_norm": 3.060966260943164, + "learning_rate": 1.991176944485281e-05, + "loss": 1.5056, + "step": 655 + }, + { + "epoch": 0.3561346362649294, + "grad_norm": 4.533553635387312, + "learning_rate": 1.99113026912664e-05, + "loss": 1.8229, + "step": 656 + }, + { + "epoch": 0.3566775244299674, + "grad_norm": 2.8582150809054045, + "learning_rate": 1.9910834711834267e-05, + "loss": 1.3233, + "step": 657 + }, + { + "epoch": 0.3572204125950054, + "grad_norm": 4.930680768753159, + "learning_rate": 1.9910365506614308e-05, + "loss": 1.5997, + "step": 658 + }, + { + "epoch": 0.3577633007600434, + "grad_norm": 3.0733335483268083, + "learning_rate": 1.9909895075664545e-05, + "loss": 0.9206, + "step": 659 + }, + { + "epoch": 0.3583061889250814, + "grad_norm": 3.4659704922712686, + "learning_rate": 1.990942341904317e-05, + "loss": 1.2752, + "step": 660 + }, + { + "epoch": 0.35884907709011943, + "grad_norm": 3.493850558129149, + "learning_rate": 1.9908950536808508e-05, + "loss": 1.588, + "step": 661 + }, + { + "epoch": 0.35939196525515743, + "grad_norm": 3.61256229272583, + "learning_rate": 1.9908476429019056e-05, + "loss": 1.469, + "step": 662 + }, + { + "epoch": 0.35993485342019543, + "grad_norm": 3.3430037514089803, + "learning_rate": 1.9908001095733445e-05, + "loss": 1.3038, + "step": 663 + }, + { + "epoch": 0.36047774158523344, + "grad_norm": 3.8918915484374557, + "learning_rate": 1.9907524537010467e-05, + "loss": 1.3683, + "step": 664 + }, + { + "epoch": 0.36102062975027144, + "grad_norm": 3.6032272045245053, + "learning_rate": 1.9907046752909064e-05, + "loss": 1.8694, + "step": 665 + }, + { + "epoch": 0.36156351791530944, + "grad_norm": 4.546268377661869, + "learning_rate": 1.9906567743488326e-05, + "loss": 1.2871, + "step": 666 + }, + { + "epoch": 0.36210640608034744, + "grad_norm": 3.5283169402860777, + "learning_rate": 1.9906087508807504e-05, + "loss": 1.5334, + "step": 667 + }, + { + "epoch": 0.36264929424538545, + "grad_norm": 2.7984191262784504, + "learning_rate": 1.9905606048925993e-05, + "loss": 1.1924, + "step": 668 + }, + { + "epoch": 0.36319218241042345, + "grad_norm": 3.54703835950394, + "learning_rate": 1.9905123363903335e-05, + "loss": 1.5972, + "step": 669 + }, + { + "epoch": 0.36373507057546145, + "grad_norm": 3.0126938064772855, + "learning_rate": 1.9904639453799236e-05, + "loss": 1.2294, + "step": 670 + }, + { + "epoch": 0.36427795874049945, + "grad_norm": 3.16358553989926, + "learning_rate": 1.990415431867354e-05, + "loss": 1.3564, + "step": 671 + }, + { + "epoch": 0.36482084690553745, + "grad_norm": 4.118910539945259, + "learning_rate": 1.990366795858626e-05, + "loss": 1.6838, + "step": 672 + }, + { + "epoch": 0.36536373507057546, + "grad_norm": 3.8466533902335502, + "learning_rate": 1.9903180373597534e-05, + "loss": 1.7986, + "step": 673 + }, + { + "epoch": 0.36590662323561346, + "grad_norm": 3.6195384409794684, + "learning_rate": 1.990269156376768e-05, + "loss": 1.6113, + "step": 674 + }, + { + "epoch": 0.36644951140065146, + "grad_norm": 3.509724379422402, + "learning_rate": 1.9902201529157152e-05, + "loss": 1.5496, + "step": 675 + }, + { + "epoch": 0.36699239956568946, + "grad_norm": 3.579737463592409, + "learning_rate": 1.9901710269826554e-05, + "loss": 1.4856, + "step": 676 + }, + { + "epoch": 0.36753528773072747, + "grad_norm": 3.780431786449365, + "learning_rate": 1.9901217785836655e-05, + "loss": 1.8519, + "step": 677 + }, + { + "epoch": 0.36807817589576547, + "grad_norm": 4.0659352627131735, + "learning_rate": 1.9900724077248354e-05, + "loss": 1.4382, + "step": 678 + }, + { + "epoch": 0.36862106406080347, + "grad_norm": 2.9524656376446274, + "learning_rate": 1.9900229144122723e-05, + "loss": 1.4541, + "step": 679 + }, + { + "epoch": 0.3691639522258415, + "grad_norm": 2.388269642575342, + "learning_rate": 1.989973298652097e-05, + "loss": 0.9567, + "step": 680 + }, + { + "epoch": 0.3697068403908795, + "grad_norm": 2.8322419431529453, + "learning_rate": 1.9899235604504467e-05, + "loss": 1.1187, + "step": 681 + }, + { + "epoch": 0.3702497285559175, + "grad_norm": 3.0469435898757613, + "learning_rate": 1.9898736998134726e-05, + "loss": 1.5571, + "step": 682 + }, + { + "epoch": 0.3707926167209555, + "grad_norm": 3.623982705749655, + "learning_rate": 1.9898237167473416e-05, + "loss": 1.2047, + "step": 683 + }, + { + "epoch": 0.3713355048859935, + "grad_norm": 3.669362267695381, + "learning_rate": 1.9897736112582357e-05, + "loss": 1.7747, + "step": 684 + }, + { + "epoch": 0.3718783930510315, + "grad_norm": 2.9109325710576353, + "learning_rate": 1.989723383352352e-05, + "loss": 1.0044, + "step": 685 + }, + { + "epoch": 0.3724212812160695, + "grad_norm": 3.4478434843760146, + "learning_rate": 1.9896730330359032e-05, + "loss": 1.1245, + "step": 686 + }, + { + "epoch": 0.3729641693811075, + "grad_norm": 3.2957514959337275, + "learning_rate": 1.989622560315116e-05, + "loss": 1.3243, + "step": 687 + }, + { + "epoch": 0.3735070575461455, + "grad_norm": 3.5500423086355988, + "learning_rate": 1.989571965196234e-05, + "loss": 2.0478, + "step": 688 + }, + { + "epoch": 0.3740499457111835, + "grad_norm": 2.621987663839103, + "learning_rate": 1.9895212476855136e-05, + "loss": 1.3135, + "step": 689 + }, + { + "epoch": 0.3745928338762215, + "grad_norm": 4.055390555970487, + "learning_rate": 1.989470407789228e-05, + "loss": 1.5165, + "step": 690 + }, + { + "epoch": 0.3751357220412595, + "grad_norm": 2.94204110816024, + "learning_rate": 1.989419445513666e-05, + "loss": 1.3458, + "step": 691 + }, + { + "epoch": 0.3756786102062975, + "grad_norm": 3.5714132672245977, + "learning_rate": 1.98936836086513e-05, + "loss": 1.3034, + "step": 692 + }, + { + "epoch": 0.3762214983713355, + "grad_norm": 3.659137782783938, + "learning_rate": 1.9893171538499382e-05, + "loss": 1.6203, + "step": 693 + }, + { + "epoch": 0.3767643865363735, + "grad_norm": 5.116097300755018, + "learning_rate": 1.9892658244744236e-05, + "loss": 2.1071, + "step": 694 + }, + { + "epoch": 0.3773072747014115, + "grad_norm": 2.8449662018307005, + "learning_rate": 1.9892143727449357e-05, + "loss": 1.0477, + "step": 695 + }, + { + "epoch": 0.3778501628664495, + "grad_norm": 4.021030963638319, + "learning_rate": 1.989162798667838e-05, + "loss": 1.9528, + "step": 696 + }, + { + "epoch": 0.3783930510314875, + "grad_norm": 3.377368666687089, + "learning_rate": 1.989111102249508e-05, + "loss": 1.2481, + "step": 697 + }, + { + "epoch": 0.3789359391965255, + "grad_norm": 3.403268816169458, + "learning_rate": 1.9890592834963406e-05, + "loss": 1.1864, + "step": 698 + }, + { + "epoch": 0.3794788273615635, + "grad_norm": 3.082879601892987, + "learning_rate": 1.9890073424147453e-05, + "loss": 1.7365, + "step": 699 + }, + { + "epoch": 0.3800217155266015, + "grad_norm": 4.122610250215441, + "learning_rate": 1.988955279011145e-05, + "loss": 1.4139, + "step": 700 + }, + { + "epoch": 0.3805646036916395, + "grad_norm": 3.298087120576001, + "learning_rate": 1.98890309329198e-05, + "loss": 1.5904, + "step": 701 + }, + { + "epoch": 0.3811074918566775, + "grad_norm": 3.1183850173888703, + "learning_rate": 1.9888507852637043e-05, + "loss": 1.2146, + "step": 702 + }, + { + "epoch": 0.38165038002171553, + "grad_norm": 3.128042971411966, + "learning_rate": 1.9887983549327873e-05, + "loss": 1.2751, + "step": 703 + }, + { + "epoch": 0.38219326818675353, + "grad_norm": 3.362555156234201, + "learning_rate": 1.988745802305714e-05, + "loss": 1.1363, + "step": 704 + }, + { + "epoch": 0.38273615635179153, + "grad_norm": 3.29574344438116, + "learning_rate": 1.988693127388984e-05, + "loss": 1.227, + "step": 705 + }, + { + "epoch": 0.38327904451682954, + "grad_norm": 4.2418853227437205, + "learning_rate": 1.9886403301891123e-05, + "loss": 1.7091, + "step": 706 + }, + { + "epoch": 0.38382193268186754, + "grad_norm": 4.165071133964158, + "learning_rate": 1.9885874107126287e-05, + "loss": 1.9403, + "step": 707 + }, + { + "epoch": 0.38436482084690554, + "grad_norm": 4.639609386348944, + "learning_rate": 1.9885343689660787e-05, + "loss": 1.221, + "step": 708 + }, + { + "epoch": 0.38490770901194354, + "grad_norm": 3.0746775894794034, + "learning_rate": 1.9884812049560226e-05, + "loss": 1.4958, + "step": 709 + }, + { + "epoch": 0.38545059717698155, + "grad_norm": 4.822804859325608, + "learning_rate": 1.9884279186890357e-05, + "loss": 1.6928, + "step": 710 + }, + { + "epoch": 0.38599348534201955, + "grad_norm": 4.299811491105524, + "learning_rate": 1.9883745101717084e-05, + "loss": 1.3075, + "step": 711 + }, + { + "epoch": 0.38653637350705755, + "grad_norm": 3.882170994492157, + "learning_rate": 1.9883209794106464e-05, + "loss": 1.9712, + "step": 712 + }, + { + "epoch": 0.38707926167209555, + "grad_norm": 4.164019029255246, + "learning_rate": 1.9882673264124705e-05, + "loss": 1.6109, + "step": 713 + }, + { + "epoch": 0.38762214983713356, + "grad_norm": 3.691226317297577, + "learning_rate": 1.9882135511838167e-05, + "loss": 1.3225, + "step": 714 + }, + { + "epoch": 0.38816503800217156, + "grad_norm": 5.777282505586183, + "learning_rate": 1.988159653731336e-05, + "loss": 1.7945, + "step": 715 + }, + { + "epoch": 0.38870792616720956, + "grad_norm": 3.3926512133724165, + "learning_rate": 1.9881056340616944e-05, + "loss": 1.5797, + "step": 716 + }, + { + "epoch": 0.38925081433224756, + "grad_norm": 3.7756965718668467, + "learning_rate": 1.988051492181573e-05, + "loss": 1.5515, + "step": 717 + }, + { + "epoch": 0.38979370249728557, + "grad_norm": 3.5509658648264613, + "learning_rate": 1.987997228097668e-05, + "loss": 1.2004, + "step": 718 + }, + { + "epoch": 0.39033659066232357, + "grad_norm": 4.682493032945398, + "learning_rate": 1.987942841816692e-05, + "loss": 1.2957, + "step": 719 + }, + { + "epoch": 0.39087947882736157, + "grad_norm": 2.923319971884201, + "learning_rate": 1.9878883333453704e-05, + "loss": 0.753, + "step": 720 + }, + { + "epoch": 0.3914223669923996, + "grad_norm": 2.8706465146507845, + "learning_rate": 1.987833702690445e-05, + "loss": 1.1057, + "step": 721 + }, + { + "epoch": 0.3919652551574376, + "grad_norm": 3.4516078574834057, + "learning_rate": 1.987778949858673e-05, + "loss": 1.1275, + "step": 722 + }, + { + "epoch": 0.3925081433224756, + "grad_norm": 3.3400553957950567, + "learning_rate": 1.9877240748568263e-05, + "loss": 1.1538, + "step": 723 + }, + { + "epoch": 0.3930510314875136, + "grad_norm": 3.155540544638446, + "learning_rate": 1.987669077691692e-05, + "loss": 1.2486, + "step": 724 + }, + { + "epoch": 0.3935939196525516, + "grad_norm": 3.1450276486107054, + "learning_rate": 1.987613958370072e-05, + "loss": 1.265, + "step": 725 + }, + { + "epoch": 0.3941368078175896, + "grad_norm": 3.0493327058864406, + "learning_rate": 1.9875587168987834e-05, + "loss": 1.2097, + "step": 726 + }, + { + "epoch": 0.3946796959826276, + "grad_norm": 3.716259137065701, + "learning_rate": 1.987503353284659e-05, + "loss": 1.5386, + "step": 727 + }, + { + "epoch": 0.3952225841476656, + "grad_norm": 3.441725300410296, + "learning_rate": 1.9874478675345458e-05, + "loss": 1.3936, + "step": 728 + }, + { + "epoch": 0.3957654723127036, + "grad_norm": 3.773353353497139, + "learning_rate": 1.9873922596553067e-05, + "loss": 1.731, + "step": 729 + }, + { + "epoch": 0.3963083604777416, + "grad_norm": 3.527645597828566, + "learning_rate": 1.987336529653819e-05, + "loss": 1.708, + "step": 730 + }, + { + "epoch": 0.3968512486427796, + "grad_norm": 3.8440953147603643, + "learning_rate": 1.9872806775369762e-05, + "loss": 1.6102, + "step": 731 + }, + { + "epoch": 0.3973941368078176, + "grad_norm": 3.4036622029265966, + "learning_rate": 1.9872247033116855e-05, + "loss": 0.9298, + "step": 732 + }, + { + "epoch": 0.3979370249728556, + "grad_norm": 4.034027632823603, + "learning_rate": 1.98716860698487e-05, + "loss": 1.78, + "step": 733 + }, + { + "epoch": 0.3984799131378936, + "grad_norm": 4.10524690369694, + "learning_rate": 1.987112388563468e-05, + "loss": 1.2354, + "step": 734 + }, + { + "epoch": 0.3990228013029316, + "grad_norm": 4.059289261823474, + "learning_rate": 1.9870560480544325e-05, + "loss": 1.804, + "step": 735 + }, + { + "epoch": 0.3995656894679696, + "grad_norm": 3.2544839104292667, + "learning_rate": 1.986999585464732e-05, + "loss": 1.1228, + "step": 736 + }, + { + "epoch": 0.4001085776330076, + "grad_norm": 3.741117309135591, + "learning_rate": 1.9869430008013496e-05, + "loss": 1.329, + "step": 737 + }, + { + "epoch": 0.4006514657980456, + "grad_norm": 4.446330579981585, + "learning_rate": 1.9868862940712838e-05, + "loss": 1.6506, + "step": 738 + }, + { + "epoch": 0.4011943539630836, + "grad_norm": 3.139296317304318, + "learning_rate": 1.9868294652815483e-05, + "loss": 1.144, + "step": 739 + }, + { + "epoch": 0.4017372421281216, + "grad_norm": 4.025270331554587, + "learning_rate": 1.986772514439172e-05, + "loss": 1.336, + "step": 740 + }, + { + "epoch": 0.4022801302931596, + "grad_norm": 3.4545518889756557, + "learning_rate": 1.986715441551198e-05, + "loss": 1.6006, + "step": 741 + }, + { + "epoch": 0.4028230184581976, + "grad_norm": 3.748598610621174, + "learning_rate": 1.986658246624686e-05, + "loss": 1.4812, + "step": 742 + }, + { + "epoch": 0.4033659066232356, + "grad_norm": 5.126058311631635, + "learning_rate": 1.9866009296667093e-05, + "loss": 2.0129, + "step": 743 + }, + { + "epoch": 0.40390879478827363, + "grad_norm": 8.145692299678405, + "learning_rate": 1.9865434906843574e-05, + "loss": 1.8321, + "step": 744 + }, + { + "epoch": 0.40445168295331163, + "grad_norm": 4.23918362592579, + "learning_rate": 1.9864859296847343e-05, + "loss": 1.5765, + "step": 745 + }, + { + "epoch": 0.40499457111834963, + "grad_norm": 4.85145039296263, + "learning_rate": 1.986428246674959e-05, + "loss": 1.6486, + "step": 746 + }, + { + "epoch": 0.40553745928338764, + "grad_norm": 3.7681072947070415, + "learning_rate": 1.986370441662166e-05, + "loss": 1.4466, + "step": 747 + }, + { + "epoch": 0.40608034744842564, + "grad_norm": 4.206136090938125, + "learning_rate": 1.986312514653505e-05, + "loss": 1.6717, + "step": 748 + }, + { + "epoch": 0.40662323561346364, + "grad_norm": 3.7336223866799965, + "learning_rate": 1.9862544656561403e-05, + "loss": 0.9599, + "step": 749 + }, + { + "epoch": 0.40716612377850164, + "grad_norm": 4.037759008136602, + "learning_rate": 1.986196294677251e-05, + "loss": 1.9387, + "step": 750 + }, + { + "epoch": 0.40770901194353965, + "grad_norm": 3.8832525632606156, + "learning_rate": 1.9861380017240324e-05, + "loss": 2.0184, + "step": 751 + }, + { + "epoch": 0.40825190010857765, + "grad_norm": 3.9521376069923546, + "learning_rate": 1.986079586803694e-05, + "loss": 1.0926, + "step": 752 + }, + { + "epoch": 0.40879478827361565, + "grad_norm": 5.128973903987411, + "learning_rate": 1.986021049923461e-05, + "loss": 1.5943, + "step": 753 + }, + { + "epoch": 0.40933767643865365, + "grad_norm": 4.5372892013581865, + "learning_rate": 1.9859623910905728e-05, + "loss": 1.643, + "step": 754 + }, + { + "epoch": 0.40988056460369166, + "grad_norm": 4.2137978897196415, + "learning_rate": 1.985903610312285e-05, + "loss": 1.7237, + "step": 755 + }, + { + "epoch": 0.41042345276872966, + "grad_norm": 2.713265124864733, + "learning_rate": 1.985844707595867e-05, + "loss": 1.0521, + "step": 756 + }, + { + "epoch": 0.41096634093376766, + "grad_norm": 3.362872283003281, + "learning_rate": 1.9857856829486045e-05, + "loss": 1.2879, + "step": 757 + }, + { + "epoch": 0.41150922909880566, + "grad_norm": 4.022235417044952, + "learning_rate": 1.9857265363777975e-05, + "loss": 1.398, + "step": 758 + }, + { + "epoch": 0.41205211726384366, + "grad_norm": 4.083280210123792, + "learning_rate": 1.9856672678907616e-05, + "loss": 1.6652, + "step": 759 + }, + { + "epoch": 0.41259500542888167, + "grad_norm": 4.469771898345589, + "learning_rate": 1.985607877494827e-05, + "loss": 1.6242, + "step": 760 + }, + { + "epoch": 0.41313789359391967, + "grad_norm": 3.0490808076449016, + "learning_rate": 1.9855483651973396e-05, + "loss": 1.0783, + "step": 761 + }, + { + "epoch": 0.41368078175895767, + "grad_norm": 3.488230005104396, + "learning_rate": 1.9854887310056593e-05, + "loss": 1.3953, + "step": 762 + }, + { + "epoch": 0.4142236699239957, + "grad_norm": 3.5610333355465653, + "learning_rate": 1.9854289749271624e-05, + "loss": 1.3692, + "step": 763 + }, + { + "epoch": 0.4147665580890337, + "grad_norm": 3.5572639096784533, + "learning_rate": 1.9853690969692393e-05, + "loss": 1.646, + "step": 764 + }, + { + "epoch": 0.4153094462540717, + "grad_norm": 3.811741925703646, + "learning_rate": 1.9853090971392953e-05, + "loss": 1.6637, + "step": 765 + }, + { + "epoch": 0.4158523344191097, + "grad_norm": 2.912609699248709, + "learning_rate": 1.9852489754447526e-05, + "loss": 0.9086, + "step": 766 + }, + { + "epoch": 0.4163952225841477, + "grad_norm": 3.7608762982374646, + "learning_rate": 1.985188731893046e-05, + "loss": 1.5131, + "step": 767 + }, + { + "epoch": 0.4169381107491857, + "grad_norm": 3.776000457813264, + "learning_rate": 1.985128366491627e-05, + "loss": 1.5929, + "step": 768 + }, + { + "epoch": 0.4174809989142237, + "grad_norm": 4.129873624081247, + "learning_rate": 1.9850678792479613e-05, + "loss": 1.5461, + "step": 769 + }, + { + "epoch": 0.4180238870792617, + "grad_norm": 3.2233425858963645, + "learning_rate": 1.9850072701695306e-05, + "loss": 0.9614, + "step": 770 + }, + { + "epoch": 0.4185667752442997, + "grad_norm": 3.8215876727885054, + "learning_rate": 1.984946539263831e-05, + "loss": 1.2549, + "step": 771 + }, + { + "epoch": 0.4191096634093377, + "grad_norm": 3.0062072101032125, + "learning_rate": 1.9848856865383732e-05, + "loss": 0.9796, + "step": 772 + }, + { + "epoch": 0.4196525515743757, + "grad_norm": 2.937513473620578, + "learning_rate": 1.984824712000684e-05, + "loss": 1.0204, + "step": 773 + }, + { + "epoch": 0.4201954397394137, + "grad_norm": 2.949279428268976, + "learning_rate": 1.984763615658305e-05, + "loss": 1.332, + "step": 774 + }, + { + "epoch": 0.4207383279044517, + "grad_norm": 3.455520053658832, + "learning_rate": 1.9847023975187925e-05, + "loss": 1.3786, + "step": 775 + }, + { + "epoch": 0.4212812160694897, + "grad_norm": 3.8352233275650733, + "learning_rate": 1.9846410575897183e-05, + "loss": 1.4897, + "step": 776 + }, + { + "epoch": 0.4218241042345277, + "grad_norm": 3.262791090623995, + "learning_rate": 1.984579595878669e-05, + "loss": 1.5097, + "step": 777 + }, + { + "epoch": 0.4223669923995657, + "grad_norm": 3.603809745585289, + "learning_rate": 1.9845180123932456e-05, + "loss": 1.8419, + "step": 778 + }, + { + "epoch": 0.4229098805646037, + "grad_norm": 3.9648489260279196, + "learning_rate": 1.9844563071410656e-05, + "loss": 1.5159, + "step": 779 + }, + { + "epoch": 0.4234527687296417, + "grad_norm": 4.576987136894564, + "learning_rate": 1.9843944801297605e-05, + "loss": 1.6866, + "step": 780 + }, + { + "epoch": 0.4239956568946797, + "grad_norm": 3.4164331578089406, + "learning_rate": 1.9843325313669774e-05, + "loss": 1.5533, + "step": 781 + }, + { + "epoch": 0.4245385450597177, + "grad_norm": 3.2121441901059367, + "learning_rate": 1.9842704608603774e-05, + "loss": 0.8879, + "step": 782 + }, + { + "epoch": 0.4250814332247557, + "grad_norm": 3.6213877563860537, + "learning_rate": 1.9842082686176388e-05, + "loss": 1.3098, + "step": 783 + }, + { + "epoch": 0.4256243213897937, + "grad_norm": 3.719993086277978, + "learning_rate": 1.9841459546464527e-05, + "loss": 1.331, + "step": 784 + }, + { + "epoch": 0.4261672095548317, + "grad_norm": 2.7952357248876516, + "learning_rate": 1.9840835189545266e-05, + "loss": 1.1298, + "step": 785 + }, + { + "epoch": 0.42671009771986973, + "grad_norm": 3.450597586563723, + "learning_rate": 1.9840209615495822e-05, + "loss": 1.0023, + "step": 786 + }, + { + "epoch": 0.42725298588490773, + "grad_norm": 4.338480682869939, + "learning_rate": 1.983958282439357e-05, + "loss": 1.7044, + "step": 787 + }, + { + "epoch": 0.42779587404994573, + "grad_norm": 3.4831324579282765, + "learning_rate": 1.983895481631603e-05, + "loss": 1.4948, + "step": 788 + }, + { + "epoch": 0.42833876221498374, + "grad_norm": 6.382810905837672, + "learning_rate": 1.9838325591340885e-05, + "loss": 2.3008, + "step": 789 + }, + { + "epoch": 0.42888165038002174, + "grad_norm": 3.063848408601439, + "learning_rate": 1.9837695149545945e-05, + "loss": 1.2045, + "step": 790 + }, + { + "epoch": 0.42942453854505974, + "grad_norm": 4.80361465821882, + "learning_rate": 1.9837063491009193e-05, + "loss": 1.2552, + "step": 791 + }, + { + "epoch": 0.42996742671009774, + "grad_norm": 4.16884813133452, + "learning_rate": 1.9836430615808745e-05, + "loss": 1.6618, + "step": 792 + }, + { + "epoch": 0.43051031487513575, + "grad_norm": 4.215747683493222, + "learning_rate": 1.9835796524022886e-05, + "loss": 1.2571, + "step": 793 + }, + { + "epoch": 0.43105320304017375, + "grad_norm": 4.894853560761769, + "learning_rate": 1.9835161215730038e-05, + "loss": 0.793, + "step": 794 + }, + { + "epoch": 0.43159609120521175, + "grad_norm": 3.264854272979086, + "learning_rate": 1.983452469100877e-05, + "loss": 1.0721, + "step": 795 + }, + { + "epoch": 0.43213897937024975, + "grad_norm": 3.2346237132359335, + "learning_rate": 1.9833886949937823e-05, + "loss": 1.2396, + "step": 796 + }, + { + "epoch": 0.43268186753528776, + "grad_norm": 3.7265266370894023, + "learning_rate": 1.9833247992596058e-05, + "loss": 1.8845, + "step": 797 + }, + { + "epoch": 0.43322475570032576, + "grad_norm": 3.0655010730082126, + "learning_rate": 1.9832607819062513e-05, + "loss": 1.556, + "step": 798 + }, + { + "epoch": 0.43376764386536376, + "grad_norm": 3.209417077142042, + "learning_rate": 1.983196642941636e-05, + "loss": 1.2183, + "step": 799 + }, + { + "epoch": 0.43431053203040176, + "grad_norm": 4.384697511143271, + "learning_rate": 1.9831323823736933e-05, + "loss": 2.0889, + "step": 800 + }, + { + "epoch": 0.43485342019543977, + "grad_norm": 3.2773759507081084, + "learning_rate": 1.9830680002103703e-05, + "loss": 0.919, + "step": 801 + }, + { + "epoch": 0.43539630836047777, + "grad_norm": 4.1436406822638805, + "learning_rate": 1.9830034964596304e-05, + "loss": 1.7759, + "step": 802 + }, + { + "epoch": 0.43593919652551577, + "grad_norm": 5.715436881791243, + "learning_rate": 1.9829388711294512e-05, + "loss": 1.4277, + "step": 803 + }, + { + "epoch": 0.4364820846905538, + "grad_norm": 3.166985124043999, + "learning_rate": 1.982874124227826e-05, + "loss": 1.56, + "step": 804 + }, + { + "epoch": 0.4370249728555918, + "grad_norm": 4.237249076660815, + "learning_rate": 1.9828092557627626e-05, + "loss": 1.2366, + "step": 805 + }, + { + "epoch": 0.4375678610206297, + "grad_norm": 3.073398749866342, + "learning_rate": 1.982744265742284e-05, + "loss": 1.1475, + "step": 806 + }, + { + "epoch": 0.4381107491856677, + "grad_norm": 3.356777818996544, + "learning_rate": 1.9826791541744285e-05, + "loss": 1.2666, + "step": 807 + }, + { + "epoch": 0.4386536373507057, + "grad_norm": 3.171383453112648, + "learning_rate": 1.982613921067249e-05, + "loss": 1.1407, + "step": 808 + }, + { + "epoch": 0.43919652551574373, + "grad_norm": 3.334493176451811, + "learning_rate": 1.9825485664288138e-05, + "loss": 1.3309, + "step": 809 + }, + { + "epoch": 0.43973941368078173, + "grad_norm": 3.199045603289457, + "learning_rate": 1.982483090267206e-05, + "loss": 1.1634, + "step": 810 + }, + { + "epoch": 0.44028230184581973, + "grad_norm": 2.9970299057571856, + "learning_rate": 1.9824174925905235e-05, + "loss": 0.9642, + "step": 811 + }, + { + "epoch": 0.44082519001085774, + "grad_norm": 3.9851247078132994, + "learning_rate": 1.98235177340688e-05, + "loss": 1.4926, + "step": 812 + }, + { + "epoch": 0.44136807817589574, + "grad_norm": 3.586604601363014, + "learning_rate": 1.9822859327244034e-05, + "loss": 1.3105, + "step": 813 + }, + { + "epoch": 0.44191096634093374, + "grad_norm": 4.198634753759357, + "learning_rate": 1.9822199705512372e-05, + "loss": 1.4437, + "step": 814 + }, + { + "epoch": 0.44245385450597174, + "grad_norm": 3.6499375531563096, + "learning_rate": 1.9821538868955394e-05, + "loss": 1.2063, + "step": 815 + }, + { + "epoch": 0.44299674267100975, + "grad_norm": 4.3973905860052716, + "learning_rate": 1.9820876817654836e-05, + "loss": 1.5002, + "step": 816 + }, + { + "epoch": 0.44353963083604775, + "grad_norm": 2.7871960857121287, + "learning_rate": 1.9820213551692585e-05, + "loss": 0.6589, + "step": 817 + }, + { + "epoch": 0.44408251900108575, + "grad_norm": 3.4126665543063184, + "learning_rate": 1.981954907115067e-05, + "loss": 1.1651, + "step": 818 + }, + { + "epoch": 0.44462540716612375, + "grad_norm": 4.229651941034759, + "learning_rate": 1.981888337611127e-05, + "loss": 1.6444, + "step": 819 + }, + { + "epoch": 0.44516829533116176, + "grad_norm": 3.4135993440364056, + "learning_rate": 1.981821646665673e-05, + "loss": 1.2614, + "step": 820 + }, + { + "epoch": 0.44571118349619976, + "grad_norm": 4.274078711060124, + "learning_rate": 1.9817548342869527e-05, + "loss": 1.66, + "step": 821 + }, + { + "epoch": 0.44625407166123776, + "grad_norm": 2.9406863812447903, + "learning_rate": 1.98168790048323e-05, + "loss": 1.0412, + "step": 822 + }, + { + "epoch": 0.44679695982627576, + "grad_norm": 4.300279758003301, + "learning_rate": 1.981620845262783e-05, + "loss": 1.3689, + "step": 823 + }, + { + "epoch": 0.44733984799131377, + "grad_norm": 3.130525530503507, + "learning_rate": 1.9815536686339056e-05, + "loss": 1.1771, + "step": 824 + }, + { + "epoch": 0.44788273615635177, + "grad_norm": 3.6606030046920597, + "learning_rate": 1.981486370604906e-05, + "loss": 1.0323, + "step": 825 + }, + { + "epoch": 0.44842562432138977, + "grad_norm": 3.495865966877518, + "learning_rate": 1.981418951184108e-05, + "loss": 1.4493, + "step": 826 + }, + { + "epoch": 0.4489685124864278, + "grad_norm": 3.621229661878418, + "learning_rate": 1.9813514103798498e-05, + "loss": 1.5392, + "step": 827 + }, + { + "epoch": 0.4495114006514658, + "grad_norm": 3.6798995423063485, + "learning_rate": 1.9812837482004853e-05, + "loss": 1.0593, + "step": 828 + }, + { + "epoch": 0.4500542888165038, + "grad_norm": 3.4514040085362954, + "learning_rate": 1.9812159646543824e-05, + "loss": 0.9262, + "step": 829 + }, + { + "epoch": 0.4505971769815418, + "grad_norm": 3.195454430712678, + "learning_rate": 1.9811480597499257e-05, + "loss": 1.1469, + "step": 830 + }, + { + "epoch": 0.4511400651465798, + "grad_norm": 4.590874916307442, + "learning_rate": 1.981080033495513e-05, + "loss": 1.6534, + "step": 831 + }, + { + "epoch": 0.4516829533116178, + "grad_norm": 3.668702667937113, + "learning_rate": 1.981011885899558e-05, + "loss": 1.0597, + "step": 832 + }, + { + "epoch": 0.4522258414766558, + "grad_norm": 2.7669639542025046, + "learning_rate": 1.98094361697049e-05, + "loss": 1.0387, + "step": 833 + }, + { + "epoch": 0.4527687296416938, + "grad_norm": 4.734510249253564, + "learning_rate": 1.9808752267167515e-05, + "loss": 1.6757, + "step": 834 + }, + { + "epoch": 0.4533116178067318, + "grad_norm": 9.006078348758516, + "learning_rate": 1.9808067151468018e-05, + "loss": 2.1559, + "step": 835 + }, + { + "epoch": 0.4538545059717698, + "grad_norm": 4.169764802563779, + "learning_rate": 1.9807380822691146e-05, + "loss": 1.2557, + "step": 836 + }, + { + "epoch": 0.4543973941368078, + "grad_norm": 3.970541817152741, + "learning_rate": 1.9806693280921783e-05, + "loss": 1.5586, + "step": 837 + }, + { + "epoch": 0.4549402823018458, + "grad_norm": 5.042448582457387, + "learning_rate": 1.9806004526244966e-05, + "loss": 1.3191, + "step": 838 + }, + { + "epoch": 0.4554831704668838, + "grad_norm": 3.7467475303782236, + "learning_rate": 1.9805314558745876e-05, + "loss": 1.3318, + "step": 839 + }, + { + "epoch": 0.4560260586319218, + "grad_norm": 4.214820131429537, + "learning_rate": 1.980462337850986e-05, + "loss": 1.3958, + "step": 840 + }, + { + "epoch": 0.4565689467969598, + "grad_norm": 3.9655409992603965, + "learning_rate": 1.980393098562239e-05, + "loss": 0.9978, + "step": 841 + }, + { + "epoch": 0.4571118349619978, + "grad_norm": 5.522638140748038, + "learning_rate": 1.980323738016912e-05, + "loss": 1.4151, + "step": 842 + }, + { + "epoch": 0.4576547231270358, + "grad_norm": 3.84359618797871, + "learning_rate": 1.9802542562235822e-05, + "loss": 1.3567, + "step": 843 + }, + { + "epoch": 0.4581976112920738, + "grad_norm": 5.115138943098397, + "learning_rate": 1.9801846531908437e-05, + "loss": 1.6437, + "step": 844 + }, + { + "epoch": 0.4587404994571118, + "grad_norm": 3.811847477591368, + "learning_rate": 1.9801149289273054e-05, + "loss": 1.2335, + "step": 845 + }, + { + "epoch": 0.4592833876221498, + "grad_norm": 4.4009848681180115, + "learning_rate": 1.98004508344159e-05, + "loss": 1.2147, + "step": 846 + }, + { + "epoch": 0.4598262757871878, + "grad_norm": 3.850991850744113, + "learning_rate": 1.9799751167423376e-05, + "loss": 1.4865, + "step": 847 + }, + { + "epoch": 0.4603691639522258, + "grad_norm": 3.1303042065802917, + "learning_rate": 1.9799050288382004e-05, + "loss": 1.4926, + "step": 848 + }, + { + "epoch": 0.4609120521172638, + "grad_norm": 4.225513214699549, + "learning_rate": 1.9798348197378475e-05, + "loss": 1.2044, + "step": 849 + }, + { + "epoch": 0.46145494028230183, + "grad_norm": 4.139979880818486, + "learning_rate": 1.9797644894499625e-05, + "loss": 1.5682, + "step": 850 + }, + { + "epoch": 0.46199782844733983, + "grad_norm": 4.414209735978709, + "learning_rate": 1.979694037983244e-05, + "loss": 1.9033, + "step": 851 + }, + { + "epoch": 0.46254071661237783, + "grad_norm": 5.079031167256292, + "learning_rate": 1.9796234653464057e-05, + "loss": 1.8745, + "step": 852 + }, + { + "epoch": 0.46308360477741584, + "grad_norm": 6.59971413476838, + "learning_rate": 1.9795527715481755e-05, + "loss": 1.2449, + "step": 853 + }, + { + "epoch": 0.46362649294245384, + "grad_norm": 3.863740127637264, + "learning_rate": 1.9794819565972973e-05, + "loss": 1.3912, + "step": 854 + }, + { + "epoch": 0.46416938110749184, + "grad_norm": 3.8578186204961327, + "learning_rate": 1.9794110205025302e-05, + "loss": 1.3487, + "step": 855 + }, + { + "epoch": 0.46471226927252984, + "grad_norm": 4.19751015099396, + "learning_rate": 1.9793399632726466e-05, + "loss": 1.129, + "step": 856 + }, + { + "epoch": 0.46525515743756785, + "grad_norm": 4.261483190159906, + "learning_rate": 1.979268784916436e-05, + "loss": 1.7298, + "step": 857 + }, + { + "epoch": 0.46579804560260585, + "grad_norm": 4.1774094914047435, + "learning_rate": 1.9791974854427008e-05, + "loss": 0.8716, + "step": 858 + }, + { + "epoch": 0.46634093376764385, + "grad_norm": 3.845723101011648, + "learning_rate": 1.9791260648602603e-05, + "loss": 0.9086, + "step": 859 + }, + { + "epoch": 0.46688382193268185, + "grad_norm": 3.4922208632687393, + "learning_rate": 1.9790545231779476e-05, + "loss": 1.3688, + "step": 860 + }, + { + "epoch": 0.46742671009771986, + "grad_norm": 2.976761133296452, + "learning_rate": 1.978982860404611e-05, + "loss": 0.7865, + "step": 861 + }, + { + "epoch": 0.46796959826275786, + "grad_norm": 3.934765402497862, + "learning_rate": 1.978911076549114e-05, + "loss": 1.2993, + "step": 862 + }, + { + "epoch": 0.46851248642779586, + "grad_norm": 3.029198772099951, + "learning_rate": 1.978839171620335e-05, + "loss": 1.1737, + "step": 863 + }, + { + "epoch": 0.46905537459283386, + "grad_norm": 3.5637545602445946, + "learning_rate": 1.978767145627167e-05, + "loss": 1.1932, + "step": 864 + }, + { + "epoch": 0.46959826275787186, + "grad_norm": 3.131289093388952, + "learning_rate": 1.9786949985785187e-05, + "loss": 0.69, + "step": 865 + }, + { + "epoch": 0.47014115092290987, + "grad_norm": 3.522468470152659, + "learning_rate": 1.978622730483313e-05, + "loss": 1.0277, + "step": 866 + }, + { + "epoch": 0.47068403908794787, + "grad_norm": 4.046439598363532, + "learning_rate": 1.9785503413504883e-05, + "loss": 1.2917, + "step": 867 + }, + { + "epoch": 0.47122692725298587, + "grad_norm": 5.55259287782346, + "learning_rate": 1.978477831188998e-05, + "loss": 1.866, + "step": 868 + }, + { + "epoch": 0.4717698154180239, + "grad_norm": 5.874766516228883, + "learning_rate": 1.97840520000781e-05, + "loss": 1.6221, + "step": 869 + }, + { + "epoch": 0.4723127035830619, + "grad_norm": 5.530080877073305, + "learning_rate": 1.9783324478159074e-05, + "loss": 1.7617, + "step": 870 + }, + { + "epoch": 0.4728555917480999, + "grad_norm": 4.996268624326209, + "learning_rate": 1.9782595746222886e-05, + "loss": 1.5794, + "step": 871 + }, + { + "epoch": 0.4733984799131379, + "grad_norm": 3.934049315300115, + "learning_rate": 1.9781865804359663e-05, + "loss": 1.2603, + "step": 872 + }, + { + "epoch": 0.4739413680781759, + "grad_norm": 4.534039775579426, + "learning_rate": 1.978113465265969e-05, + "loss": 1.2766, + "step": 873 + }, + { + "epoch": 0.4744842562432139, + "grad_norm": 4.281791787646871, + "learning_rate": 1.9780402291213393e-05, + "loss": 1.6424, + "step": 874 + }, + { + "epoch": 0.4750271444082519, + "grad_norm": 4.443571747878234, + "learning_rate": 1.977966872011135e-05, + "loss": 1.514, + "step": 875 + }, + { + "epoch": 0.4755700325732899, + "grad_norm": 4.432653652742011, + "learning_rate": 1.9778933939444298e-05, + "loss": 1.2405, + "step": 876 + }, + { + "epoch": 0.4761129207383279, + "grad_norm": 3.408851551186198, + "learning_rate": 1.9778197949303107e-05, + "loss": 0.9597, + "step": 877 + }, + { + "epoch": 0.4766558089033659, + "grad_norm": 3.13258933410851, + "learning_rate": 1.9777460749778812e-05, + "loss": 0.8015, + "step": 878 + }, + { + "epoch": 0.4771986970684039, + "grad_norm": 3.535707436018844, + "learning_rate": 1.977672234096259e-05, + "loss": 1.3375, + "step": 879 + }, + { + "epoch": 0.4777415852334419, + "grad_norm": 5.210333639552661, + "learning_rate": 1.9775982722945764e-05, + "loss": 1.5039, + "step": 880 + }, + { + "epoch": 0.4782844733984799, + "grad_norm": 4.14115837125392, + "learning_rate": 1.9775241895819818e-05, + "loss": 1.0267, + "step": 881 + }, + { + "epoch": 0.4788273615635179, + "grad_norm": 4.390843049257208, + "learning_rate": 1.977449985967637e-05, + "loss": 1.6639, + "step": 882 + }, + { + "epoch": 0.4793702497285559, + "grad_norm": 3.5906251649150405, + "learning_rate": 1.9773756614607205e-05, + "loss": 1.2382, + "step": 883 + }, + { + "epoch": 0.4799131378935939, + "grad_norm": 4.78880370244169, + "learning_rate": 1.9773012160704242e-05, + "loss": 1.9436, + "step": 884 + }, + { + "epoch": 0.4804560260586319, + "grad_norm": 4.024679917575935, + "learning_rate": 1.977226649805956e-05, + "loss": 1.7126, + "step": 885 + }, + { + "epoch": 0.4809989142236699, + "grad_norm": 4.789782875875197, + "learning_rate": 1.9771519626765384e-05, + "loss": 1.4639, + "step": 886 + }, + { + "epoch": 0.4815418023887079, + "grad_norm": 4.188325588019175, + "learning_rate": 1.9770771546914088e-05, + "loss": 0.9698, + "step": 887 + }, + { + "epoch": 0.4820846905537459, + "grad_norm": 4.491436069698424, + "learning_rate": 1.9770022258598192e-05, + "loss": 1.5504, + "step": 888 + }, + { + "epoch": 0.4826275787187839, + "grad_norm": 4.889270563723842, + "learning_rate": 1.9769271761910373e-05, + "loss": 1.6052, + "step": 889 + }, + { + "epoch": 0.4831704668838219, + "grad_norm": 3.7615089990618493, + "learning_rate": 1.9768520056943454e-05, + "loss": 1.2445, + "step": 890 + }, + { + "epoch": 0.4837133550488599, + "grad_norm": 4.153654659347469, + "learning_rate": 1.9767767143790403e-05, + "loss": 1.4304, + "step": 891 + }, + { + "epoch": 0.48425624321389793, + "grad_norm": 4.040032964648755, + "learning_rate": 1.9767013022544346e-05, + "loss": 1.2906, + "step": 892 + }, + { + "epoch": 0.48479913137893593, + "grad_norm": 4.6559235863058275, + "learning_rate": 1.976625769329855e-05, + "loss": 1.5214, + "step": 893 + }, + { + "epoch": 0.48534201954397393, + "grad_norm": 4.224494142526306, + "learning_rate": 1.976550115614644e-05, + "loss": 1.2299, + "step": 894 + }, + { + "epoch": 0.48588490770901194, + "grad_norm": 5.225621473809677, + "learning_rate": 1.9764743411181585e-05, + "loss": 1.3489, + "step": 895 + }, + { + "epoch": 0.48642779587404994, + "grad_norm": 3.298440710545404, + "learning_rate": 1.97639844584977e-05, + "loss": 1.2107, + "step": 896 + }, + { + "epoch": 0.48697068403908794, + "grad_norm": 4.8373671778516485, + "learning_rate": 1.9763224298188652e-05, + "loss": 1.7748, + "step": 897 + }, + { + "epoch": 0.48751357220412594, + "grad_norm": 3.9332967347674224, + "learning_rate": 1.9762462930348465e-05, + "loss": 1.1036, + "step": 898 + }, + { + "epoch": 0.48805646036916395, + "grad_norm": 3.555513906222738, + "learning_rate": 1.9761700355071306e-05, + "loss": 1.5988, + "step": 899 + }, + { + "epoch": 0.48859934853420195, + "grad_norm": 4.115001448738943, + "learning_rate": 1.9760936572451487e-05, + "loss": 0.9789, + "step": 900 + }, + { + "epoch": 0.48914223669923995, + "grad_norm": 3.955269044937872, + "learning_rate": 1.9760171582583476e-05, + "loss": 1.3833, + "step": 901 + }, + { + "epoch": 0.48968512486427795, + "grad_norm": 2.718909991804504, + "learning_rate": 1.975940538556189e-05, + "loss": 1.1708, + "step": 902 + }, + { + "epoch": 0.49022801302931596, + "grad_norm": 4.234679090051759, + "learning_rate": 1.975863798148149e-05, + "loss": 1.3002, + "step": 903 + }, + { + "epoch": 0.49077090119435396, + "grad_norm": 4.162638017444828, + "learning_rate": 1.975786937043719e-05, + "loss": 1.246, + "step": 904 + }, + { + "epoch": 0.49131378935939196, + "grad_norm": 4.466027820294199, + "learning_rate": 1.975709955252406e-05, + "loss": 1.7345, + "step": 905 + }, + { + "epoch": 0.49185667752442996, + "grad_norm": 4.120456111509664, + "learning_rate": 1.97563285278373e-05, + "loss": 1.1339, + "step": 906 + }, + { + "epoch": 0.49239956568946797, + "grad_norm": 3.7932999801551346, + "learning_rate": 1.9755556296472286e-05, + "loss": 1.241, + "step": 907 + }, + { + "epoch": 0.49294245385450597, + "grad_norm": 3.3311865780622107, + "learning_rate": 1.9754782858524515e-05, + "loss": 1.2329, + "step": 908 + }, + { + "epoch": 0.49348534201954397, + "grad_norm": 4.21189840914184, + "learning_rate": 1.9754008214089654e-05, + "loss": 1.089, + "step": 909 + }, + { + "epoch": 0.494028230184582, + "grad_norm": 4.882048591403942, + "learning_rate": 1.9753232363263513e-05, + "loss": 1.3579, + "step": 910 + }, + { + "epoch": 0.49457111834962, + "grad_norm": 3.8495412115497833, + "learning_rate": 1.9752455306142052e-05, + "loss": 0.876, + "step": 911 + }, + { + "epoch": 0.495114006514658, + "grad_norm": 5.35378911928701, + "learning_rate": 1.975167704282137e-05, + "loss": 1.5025, + "step": 912 + }, + { + "epoch": 0.495656894679696, + "grad_norm": 3.4776107222093176, + "learning_rate": 1.9750897573397733e-05, + "loss": 1.1576, + "step": 913 + }, + { + "epoch": 0.496199782844734, + "grad_norm": 4.3680755226003525, + "learning_rate": 1.9750116897967545e-05, + "loss": 1.2519, + "step": 914 + }, + { + "epoch": 0.496742671009772, + "grad_norm": 4.154395835162648, + "learning_rate": 1.9749335016627355e-05, + "loss": 1.3397, + "step": 915 + }, + { + "epoch": 0.49728555917481, + "grad_norm": 5.465210053229008, + "learning_rate": 1.9748551929473873e-05, + "loss": 1.7555, + "step": 916 + }, + { + "epoch": 0.497828447339848, + "grad_norm": 4.391089608848125, + "learning_rate": 1.9747767636603953e-05, + "loss": 0.9685, + "step": 917 + }, + { + "epoch": 0.498371335504886, + "grad_norm": 4.215127655197256, + "learning_rate": 1.9746982138114597e-05, + "loss": 1.3889, + "step": 918 + }, + { + "epoch": 0.498914223669924, + "grad_norm": 4.559291789845839, + "learning_rate": 1.9746195434102956e-05, + "loss": 1.2411, + "step": 919 + }, + { + "epoch": 0.499457111834962, + "grad_norm": 4.226267506705079, + "learning_rate": 1.9745407524666327e-05, + "loss": 0.9517, + "step": 920 + }, + { + "epoch": 0.5, + "grad_norm": 4.289151336000736, + "learning_rate": 1.9744618409902164e-05, + "loss": 0.9123, + "step": 921 + }, + { + "epoch": 0.500542888165038, + "grad_norm": 5.356043714827742, + "learning_rate": 1.9743828089908067e-05, + "loss": 1.4039, + "step": 922 + }, + { + "epoch": 0.501085776330076, + "grad_norm": 4.867388642606108, + "learning_rate": 1.9743036564781785e-05, + "loss": 1.877, + "step": 923 + }, + { + "epoch": 0.501628664495114, + "grad_norm": 4.034273796102648, + "learning_rate": 1.9742243834621207e-05, + "loss": 1.6151, + "step": 924 + }, + { + "epoch": 0.502171552660152, + "grad_norm": 4.67131234305618, + "learning_rate": 1.9741449899524393e-05, + "loss": 1.6415, + "step": 925 + }, + { + "epoch": 0.50271444082519, + "grad_norm": 4.609211834325983, + "learning_rate": 1.9740654759589524e-05, + "loss": 1.2702, + "step": 926 + }, + { + "epoch": 0.503257328990228, + "grad_norm": 3.8693712873725064, + "learning_rate": 1.973985841491495e-05, + "loss": 1.0518, + "step": 927 + }, + { + "epoch": 0.503800217155266, + "grad_norm": 4.224951197105749, + "learning_rate": 1.9739060865599163e-05, + "loss": 1.1367, + "step": 928 + }, + { + "epoch": 0.504343105320304, + "grad_norm": 4.851656117293343, + "learning_rate": 1.973826211174081e-05, + "loss": 1.2753, + "step": 929 + }, + { + "epoch": 0.504885993485342, + "grad_norm": 4.255261482914816, + "learning_rate": 1.9737462153438675e-05, + "loss": 1.1723, + "step": 930 + }, + { + "epoch": 0.50542888165038, + "grad_norm": 3.8038614228711616, + "learning_rate": 1.9736660990791705e-05, + "loss": 1.1919, + "step": 931 + }, + { + "epoch": 0.505971769815418, + "grad_norm": 3.934227455395002, + "learning_rate": 1.9735858623898984e-05, + "loss": 1.1311, + "step": 932 + }, + { + "epoch": 0.506514657980456, + "grad_norm": 6.087081162565673, + "learning_rate": 1.9735055052859752e-05, + "loss": 1.4226, + "step": 933 + }, + { + "epoch": 0.507057546145494, + "grad_norm": 3.901680463400142, + "learning_rate": 1.9734250277773398e-05, + "loss": 0.8336, + "step": 934 + }, + { + "epoch": 0.507600434310532, + "grad_norm": 4.29804205692996, + "learning_rate": 1.9733444298739454e-05, + "loss": 1.5767, + "step": 935 + }, + { + "epoch": 0.50814332247557, + "grad_norm": 4.114842645062852, + "learning_rate": 1.9732637115857606e-05, + "loss": 1.3353, + "step": 936 + }, + { + "epoch": 0.508686210640608, + "grad_norm": 3.8525228530533706, + "learning_rate": 1.9731828729227686e-05, + "loss": 0.8208, + "step": 937 + }, + { + "epoch": 0.509229098805646, + "grad_norm": 4.592070280997817, + "learning_rate": 1.973101913894968e-05, + "loss": 1.3908, + "step": 938 + }, + { + "epoch": 0.509771986970684, + "grad_norm": 3.5842386277465628, + "learning_rate": 1.9730208345123718e-05, + "loss": 1.1996, + "step": 939 + }, + { + "epoch": 0.510314875135722, + "grad_norm": 6.053647281932304, + "learning_rate": 1.9729396347850082e-05, + "loss": 1.565, + "step": 940 + }, + { + "epoch": 0.51085776330076, + "grad_norm": 4.515664489508703, + "learning_rate": 1.9728583147229196e-05, + "loss": 1.4823, + "step": 941 + }, + { + "epoch": 0.511400651465798, + "grad_norm": 4.7752750595730475, + "learning_rate": 1.9727768743361644e-05, + "loss": 1.3867, + "step": 942 + }, + { + "epoch": 0.511943539630836, + "grad_norm": 3.983262894008015, + "learning_rate": 1.972695313634815e-05, + "loss": 1.1327, + "step": 943 + }, + { + "epoch": 0.512486427795874, + "grad_norm": 3.8510372454489326, + "learning_rate": 1.9726136326289586e-05, + "loss": 1.0233, + "step": 944 + }, + { + "epoch": 0.5130293159609121, + "grad_norm": 3.7881258264140634, + "learning_rate": 1.9725318313286985e-05, + "loss": 1.0239, + "step": 945 + }, + { + "epoch": 0.5135722041259501, + "grad_norm": 5.532602386318144, + "learning_rate": 1.9724499097441513e-05, + "loss": 1.8325, + "step": 946 + }, + { + "epoch": 0.5141150922909881, + "grad_norm": 5.978249809716352, + "learning_rate": 1.9723678678854493e-05, + "loss": 1.4899, + "step": 947 + }, + { + "epoch": 0.5146579804560261, + "grad_norm": 4.164217442125778, + "learning_rate": 1.9722857057627398e-05, + "loss": 1.5428, + "step": 948 + }, + { + "epoch": 0.5152008686210641, + "grad_norm": 3.315839086509451, + "learning_rate": 1.9722034233861848e-05, + "loss": 1.161, + "step": 949 + }, + { + "epoch": 0.5157437567861021, + "grad_norm": 4.362448119896709, + "learning_rate": 1.9721210207659608e-05, + "loss": 1.3402, + "step": 950 + }, + { + "epoch": 0.5162866449511401, + "grad_norm": 4.196652705893848, + "learning_rate": 1.9720384979122594e-05, + "loss": 1.4782, + "step": 951 + }, + { + "epoch": 0.5168295331161781, + "grad_norm": 4.093486911730782, + "learning_rate": 1.9719558548352876e-05, + "loss": 0.8508, + "step": 952 + }, + { + "epoch": 0.5173724212812161, + "grad_norm": 3.623412868803193, + "learning_rate": 1.9718730915452664e-05, + "loss": 1.4979, + "step": 953 + }, + { + "epoch": 0.5179153094462541, + "grad_norm": 3.4964911024452845, + "learning_rate": 1.9717902080524324e-05, + "loss": 1.0495, + "step": 954 + }, + { + "epoch": 0.5184581976112921, + "grad_norm": 4.021867363378223, + "learning_rate": 1.9717072043670367e-05, + "loss": 1.5284, + "step": 955 + }, + { + "epoch": 0.5190010857763301, + "grad_norm": 4.383918034859983, + "learning_rate": 1.9716240804993454e-05, + "loss": 1.3092, + "step": 956 + }, + { + "epoch": 0.5195439739413681, + "grad_norm": 4.61200677033352, + "learning_rate": 1.971540836459639e-05, + "loss": 1.8809, + "step": 957 + }, + { + "epoch": 0.5200868621064061, + "grad_norm": 4.490685725956694, + "learning_rate": 1.9714574722582142e-05, + "loss": 1.4004, + "step": 958 + }, + { + "epoch": 0.5206297502714441, + "grad_norm": 3.4814367857792736, + "learning_rate": 1.9713739879053802e-05, + "loss": 0.7918, + "step": 959 + }, + { + "epoch": 0.5211726384364821, + "grad_norm": 4.751597566480749, + "learning_rate": 1.9712903834114635e-05, + "loss": 1.6595, + "step": 960 + }, + { + "epoch": 0.5217155266015201, + "grad_norm": 4.535333205317335, + "learning_rate": 1.9712066587868042e-05, + "loss": 1.5676, + "step": 961 + }, + { + "epoch": 0.5222584147665581, + "grad_norm": 4.285001993552699, + "learning_rate": 1.9711228140417577e-05, + "loss": 1.0156, + "step": 962 + }, + { + "epoch": 0.5228013029315961, + "grad_norm": 4.61361421766516, + "learning_rate": 1.9710388491866934e-05, + "loss": 0.9552, + "step": 963 + }, + { + "epoch": 0.5233441910966341, + "grad_norm": 5.760482951437034, + "learning_rate": 1.9709547642319968e-05, + "loss": 2.2573, + "step": 964 + }, + { + "epoch": 0.5238870792616721, + "grad_norm": 4.38974180509626, + "learning_rate": 1.9708705591880674e-05, + "loss": 0.9939, + "step": 965 + }, + { + "epoch": 0.5244299674267101, + "grad_norm": 4.856793718321527, + "learning_rate": 1.97078623406532e-05, + "loss": 1.6161, + "step": 966 + }, + { + "epoch": 0.5249728555917481, + "grad_norm": 4.164767885626303, + "learning_rate": 1.9707017888741838e-05, + "loss": 1.4327, + "step": 967 + }, + { + "epoch": 0.5255157437567861, + "grad_norm": 5.163022804287371, + "learning_rate": 1.970617223625104e-05, + "loss": 1.2063, + "step": 968 + }, + { + "epoch": 0.5260586319218241, + "grad_norm": 4.826390195584734, + "learning_rate": 1.9705325383285384e-05, + "loss": 1.609, + "step": 969 + }, + { + "epoch": 0.5266015200868621, + "grad_norm": 6.3570394668292405, + "learning_rate": 1.9704477329949617e-05, + "loss": 1.5845, + "step": 970 + }, + { + "epoch": 0.5271444082519001, + "grad_norm": 5.176208926963624, + "learning_rate": 1.9703628076348628e-05, + "loss": 1.5792, + "step": 971 + }, + { + "epoch": 0.5276872964169381, + "grad_norm": 4.968725075311641, + "learning_rate": 1.9702777622587452e-05, + "loss": 1.774, + "step": 972 + }, + { + "epoch": 0.5282301845819761, + "grad_norm": 3.733580423042607, + "learning_rate": 1.9701925968771277e-05, + "loss": 0.8424, + "step": 973 + }, + { + "epoch": 0.5287730727470141, + "grad_norm": 4.330993358163065, + "learning_rate": 1.9701073115005437e-05, + "loss": 1.6198, + "step": 974 + }, + { + "epoch": 0.5293159609120521, + "grad_norm": 4.64032700965104, + "learning_rate": 1.9700219061395408e-05, + "loss": 1.2958, + "step": 975 + }, + { + "epoch": 0.5298588490770901, + "grad_norm": 4.08538215893539, + "learning_rate": 1.969936380804683e-05, + "loss": 1.5371, + "step": 976 + }, + { + "epoch": 0.5304017372421281, + "grad_norm": 3.7672102191405012, + "learning_rate": 1.9698507355065478e-05, + "loss": 1.4831, + "step": 977 + }, + { + "epoch": 0.5309446254071661, + "grad_norm": 4.237599031471779, + "learning_rate": 1.969764970255728e-05, + "loss": 1.4561, + "step": 978 + }, + { + "epoch": 0.5314875135722041, + "grad_norm": 4.296489590459447, + "learning_rate": 1.9696790850628308e-05, + "loss": 1.3813, + "step": 979 + }, + { + "epoch": 0.5320304017372421, + "grad_norm": 4.037479773042692, + "learning_rate": 1.9695930799384787e-05, + "loss": 1.103, + "step": 980 + }, + { + "epoch": 0.5325732899022801, + "grad_norm": 3.388491089813824, + "learning_rate": 1.9695069548933097e-05, + "loss": 1.0416, + "step": 981 + }, + { + "epoch": 0.5331161780673181, + "grad_norm": 3.6808593992848873, + "learning_rate": 1.969420709937975e-05, + "loss": 1.3158, + "step": 982 + }, + { + "epoch": 0.5336590662323561, + "grad_norm": 3.574434479685402, + "learning_rate": 1.969334345083142e-05, + "loss": 0.9542, + "step": 983 + }, + { + "epoch": 0.5342019543973942, + "grad_norm": 4.205886917864195, + "learning_rate": 1.9692478603394926e-05, + "loss": 1.1542, + "step": 984 + }, + { + "epoch": 0.5347448425624322, + "grad_norm": 3.89660120144171, + "learning_rate": 1.9691612557177225e-05, + "loss": 1.2374, + "step": 985 + }, + { + "epoch": 0.5352877307274702, + "grad_norm": 5.508260629607227, + "learning_rate": 1.9690745312285443e-05, + "loss": 2.0836, + "step": 986 + }, + { + "epoch": 0.5358306188925082, + "grad_norm": 5.482641917586033, + "learning_rate": 1.968987686882683e-05, + "loss": 1.5608, + "step": 987 + }, + { + "epoch": 0.5363735070575462, + "grad_norm": 5.4640893579957845, + "learning_rate": 1.9689007226908807e-05, + "loss": 1.3632, + "step": 988 + }, + { + "epoch": 0.5369163952225842, + "grad_norm": 4.163793092253212, + "learning_rate": 1.9688136386638926e-05, + "loss": 1.0509, + "step": 989 + }, + { + "epoch": 0.5374592833876222, + "grad_norm": 7.431185891169344, + "learning_rate": 1.96872643481249e-05, + "loss": 2.2147, + "step": 990 + }, + { + "epoch": 0.5380021715526602, + "grad_norm": 4.243102861037138, + "learning_rate": 1.9686391111474574e-05, + "loss": 0.9978, + "step": 991 + }, + { + "epoch": 0.5385450597176982, + "grad_norm": 4.250454886057412, + "learning_rate": 1.968551667679596e-05, + "loss": 1.5265, + "step": 992 + }, + { + "epoch": 0.5390879478827362, + "grad_norm": 5.989466882905416, + "learning_rate": 1.9684641044197207e-05, + "loss": 1.3344, + "step": 993 + }, + { + "epoch": 0.5396308360477742, + "grad_norm": 6.003524822087355, + "learning_rate": 1.9683764213786617e-05, + "loss": 2.0785, + "step": 994 + }, + { + "epoch": 0.5401737242128122, + "grad_norm": 4.535858561115482, + "learning_rate": 1.9682886185672633e-05, + "loss": 1.7654, + "step": 995 + }, + { + "epoch": 0.5407166123778502, + "grad_norm": 3.304287633610362, + "learning_rate": 1.9682006959963854e-05, + "loss": 0.9151, + "step": 996 + }, + { + "epoch": 0.5412595005428882, + "grad_norm": 3.6154194334954797, + "learning_rate": 1.9681126536769022e-05, + "loss": 1.1871, + "step": 997 + }, + { + "epoch": 0.5418023887079262, + "grad_norm": 5.596808777173772, + "learning_rate": 1.968024491619703e-05, + "loss": 1.1773, + "step": 998 + }, + { + "epoch": 0.5423452768729642, + "grad_norm": 4.486728043355676, + "learning_rate": 1.9679362098356923e-05, + "loss": 1.6495, + "step": 999 + }, + { + "epoch": 0.5428881650380022, + "grad_norm": 5.254525660355539, + "learning_rate": 1.9678478083357882e-05, + "loss": 1.7632, + "step": 1000 + }, + { + "epoch": 0.5434310532030402, + "grad_norm": 4.084222439189464, + "learning_rate": 1.9677592871309248e-05, + "loss": 1.4108, + "step": 1001 + }, + { + "epoch": 0.5439739413680782, + "grad_norm": 3.7647794378542248, + "learning_rate": 1.9676706462320504e-05, + "loss": 1.0893, + "step": 1002 + }, + { + "epoch": 0.5445168295331162, + "grad_norm": 3.658128786211824, + "learning_rate": 1.967581885650128e-05, + "loss": 1.0626, + "step": 1003 + }, + { + "epoch": 0.5450597176981542, + "grad_norm": 4.111525087755858, + "learning_rate": 1.967493005396136e-05, + "loss": 1.1915, + "step": 1004 + }, + { + "epoch": 0.5456026058631922, + "grad_norm": 4.130761699023899, + "learning_rate": 1.967404005481067e-05, + "loss": 1.1007, + "step": 1005 + }, + { + "epoch": 0.5461454940282302, + "grad_norm": 4.610164776763781, + "learning_rate": 1.9673148859159292e-05, + "loss": 1.3626, + "step": 1006 + }, + { + "epoch": 0.5466883821932682, + "grad_norm": 3.3516729554956686, + "learning_rate": 1.9672256467117445e-05, + "loss": 1.1138, + "step": 1007 + }, + { + "epoch": 0.5472312703583062, + "grad_norm": 5.185808250719794, + "learning_rate": 1.9671362878795502e-05, + "loss": 1.5639, + "step": 1008 + }, + { + "epoch": 0.5477741585233442, + "grad_norm": 3.9549787550673323, + "learning_rate": 1.9670468094303983e-05, + "loss": 1.0094, + "step": 1009 + }, + { + "epoch": 0.5483170466883822, + "grad_norm": 5.025417088271204, + "learning_rate": 1.966957211375356e-05, + "loss": 1.9741, + "step": 1010 + }, + { + "epoch": 0.5488599348534202, + "grad_norm": 5.1560673391175, + "learning_rate": 1.9668674937255044e-05, + "loss": 0.9653, + "step": 1011 + }, + { + "epoch": 0.5494028230184582, + "grad_norm": 4.843540796107876, + "learning_rate": 1.9667776564919404e-05, + "loss": 1.4745, + "step": 1012 + }, + { + "epoch": 0.5499457111834962, + "grad_norm": 3.8379573247832193, + "learning_rate": 1.966687699685775e-05, + "loss": 1.537, + "step": 1013 + }, + { + "epoch": 0.5504885993485342, + "grad_norm": 4.291804816024517, + "learning_rate": 1.9665976233181342e-05, + "loss": 1.6859, + "step": 1014 + }, + { + "epoch": 0.5510314875135722, + "grad_norm": 4.827630960232388, + "learning_rate": 1.966507427400159e-05, + "loss": 1.3205, + "step": 1015 + }, + { + "epoch": 0.5515743756786102, + "grad_norm": 4.251540690047392, + "learning_rate": 1.9664171119430044e-05, + "loss": 0.8213, + "step": 1016 + }, + { + "epoch": 0.5521172638436482, + "grad_norm": 3.7693342275615476, + "learning_rate": 1.9663266769578414e-05, + "loss": 1.0349, + "step": 1017 + }, + { + "epoch": 0.5526601520086862, + "grad_norm": 4.075472936638236, + "learning_rate": 1.966236122455855e-05, + "loss": 1.0853, + "step": 1018 + }, + { + "epoch": 0.5532030401737242, + "grad_norm": 4.771167333111926, + "learning_rate": 1.9661454484482448e-05, + "loss": 1.2135, + "step": 1019 + }, + { + "epoch": 0.5537459283387622, + "grad_norm": 3.834854303591261, + "learning_rate": 1.9660546549462262e-05, + "loss": 1.0972, + "step": 1020 + }, + { + "epoch": 0.5542888165038002, + "grad_norm": 4.712560042125338, + "learning_rate": 1.9659637419610278e-05, + "loss": 0.8779, + "step": 1021 + }, + { + "epoch": 0.5548317046688382, + "grad_norm": 3.8174143787104566, + "learning_rate": 1.9658727095038942e-05, + "loss": 0.7395, + "step": 1022 + }, + { + "epoch": 0.5553745928338762, + "grad_norm": 5.970646625648977, + "learning_rate": 1.965781557586085e-05, + "loss": 1.4233, + "step": 1023 + }, + { + "epoch": 0.5559174809989142, + "grad_norm": 3.872798187838392, + "learning_rate": 1.9656902862188732e-05, + "loss": 0.8147, + "step": 1024 + }, + { + "epoch": 0.5564603691639523, + "grad_norm": 5.632609234869673, + "learning_rate": 1.9655988954135473e-05, + "loss": 1.1581, + "step": 1025 + }, + { + "epoch": 0.5570032573289903, + "grad_norm": 4.141188191911578, + "learning_rate": 1.9655073851814117e-05, + "loss": 1.1707, + "step": 1026 + }, + { + "epoch": 0.5575461454940283, + "grad_norm": 5.7835323867879564, + "learning_rate": 1.9654157555337837e-05, + "loss": 1.2383, + "step": 1027 + }, + { + "epoch": 0.5580890336590663, + "grad_norm": 6.410603262158535, + "learning_rate": 1.9653240064819965e-05, + "loss": 2.3461, + "step": 1028 + }, + { + "epoch": 0.5586319218241043, + "grad_norm": 4.661944777700264, + "learning_rate": 1.9652321380373974e-05, + "loss": 1.1718, + "step": 1029 + }, + { + "epoch": 0.5591748099891423, + "grad_norm": 4.224334614804915, + "learning_rate": 1.9651401502113497e-05, + "loss": 1.0937, + "step": 1030 + }, + { + "epoch": 0.5597176981541803, + "grad_norm": 4.292561821167698, + "learning_rate": 1.9650480430152295e-05, + "loss": 0.8306, + "step": 1031 + }, + { + "epoch": 0.5602605863192183, + "grad_norm": 4.515070808084738, + "learning_rate": 1.9649558164604293e-05, + "loss": 1.4042, + "step": 1032 + }, + { + "epoch": 0.5608034744842563, + "grad_norm": 4.906478484057087, + "learning_rate": 1.964863470558356e-05, + "loss": 1.3396, + "step": 1033 + }, + { + "epoch": 0.5613463626492943, + "grad_norm": 3.9530548146664306, + "learning_rate": 1.9647710053204307e-05, + "loss": 0.917, + "step": 1034 + }, + { + "epoch": 0.5618892508143323, + "grad_norm": 4.640160859146004, + "learning_rate": 1.96467842075809e-05, + "loss": 1.6625, + "step": 1035 + }, + { + "epoch": 0.5624321389793703, + "grad_norm": 4.921887921875984, + "learning_rate": 1.964585716882785e-05, + "loss": 1.5133, + "step": 1036 + }, + { + "epoch": 0.5629750271444083, + "grad_norm": 4.667679101301949, + "learning_rate": 1.964492893705981e-05, + "loss": 0.9713, + "step": 1037 + }, + { + "epoch": 0.5635179153094463, + "grad_norm": 6.0457527064758825, + "learning_rate": 1.9643999512391586e-05, + "loss": 1.1101, + "step": 1038 + }, + { + "epoch": 0.5640608034744843, + "grad_norm": 5.188403972218186, + "learning_rate": 1.964306889493813e-05, + "loss": 1.8276, + "step": 1039 + }, + { + "epoch": 0.5646036916395223, + "grad_norm": 4.449370010668479, + "learning_rate": 1.9642137084814548e-05, + "loss": 1.0797, + "step": 1040 + }, + { + "epoch": 0.5651465798045603, + "grad_norm": 5.97660491495462, + "learning_rate": 1.9641204082136085e-05, + "loss": 1.913, + "step": 1041 + }, + { + "epoch": 0.5656894679695983, + "grad_norm": 5.611282387277076, + "learning_rate": 1.9640269887018135e-05, + "loss": 2.4593, + "step": 1042 + }, + { + "epoch": 0.5662323561346363, + "grad_norm": 3.6226559029227996, + "learning_rate": 1.9639334499576237e-05, + "loss": 0.7939, + "step": 1043 + }, + { + "epoch": 0.5667752442996743, + "grad_norm": 4.701569628633213, + "learning_rate": 1.963839791992609e-05, + "loss": 1.4092, + "step": 1044 + }, + { + "epoch": 0.5673181324647123, + "grad_norm": 5.185222096624812, + "learning_rate": 1.9637460148183525e-05, + "loss": 1.3851, + "step": 1045 + }, + { + "epoch": 0.5678610206297503, + "grad_norm": 4.094772946058271, + "learning_rate": 1.963652118446453e-05, + "loss": 1.1872, + "step": 1046 + }, + { + "epoch": 0.5684039087947883, + "grad_norm": 4.653748340189311, + "learning_rate": 1.9635581028885233e-05, + "loss": 1.1546, + "step": 1047 + }, + { + "epoch": 0.5689467969598263, + "grad_norm": 3.9426831069951844, + "learning_rate": 1.9634639681561924e-05, + "loss": 0.9324, + "step": 1048 + }, + { + "epoch": 0.5694896851248643, + "grad_norm": 4.949304407204856, + "learning_rate": 1.9633697142611017e-05, + "loss": 1.2372, + "step": 1049 + }, + { + "epoch": 0.5700325732899023, + "grad_norm": 5.364689744345368, + "learning_rate": 1.9632753412149096e-05, + "loss": 1.5104, + "step": 1050 + }, + { + "epoch": 0.5705754614549403, + "grad_norm": 5.307969223857254, + "learning_rate": 1.9631808490292884e-05, + "loss": 1.212, + "step": 1051 + }, + { + "epoch": 0.5711183496199783, + "grad_norm": 5.197310781163775, + "learning_rate": 1.963086237715924e-05, + "loss": 1.0175, + "step": 1052 + }, + { + "epoch": 0.5716612377850163, + "grad_norm": 4.095170540464123, + "learning_rate": 1.9629915072865194e-05, + "loss": 1.2711, + "step": 1053 + }, + { + "epoch": 0.5722041259500543, + "grad_norm": 4.816474093418903, + "learning_rate": 1.9628966577527902e-05, + "loss": 0.7274, + "step": 1054 + }, + { + "epoch": 0.5727470141150923, + "grad_norm": 4.000604469041167, + "learning_rate": 1.962801689126468e-05, + "loss": 1.2913, + "step": 1055 + }, + { + "epoch": 0.5732899022801303, + "grad_norm": 4.31668064977018, + "learning_rate": 1.962706601419298e-05, + "loss": 1.2598, + "step": 1056 + }, + { + "epoch": 0.5738327904451683, + "grad_norm": 5.033983318277113, + "learning_rate": 1.9626113946430414e-05, + "loss": 1.5149, + "step": 1057 + }, + { + "epoch": 0.5743756786102063, + "grad_norm": 4.51738649586986, + "learning_rate": 1.9625160688094733e-05, + "loss": 1.2389, + "step": 1058 + }, + { + "epoch": 0.5749185667752443, + "grad_norm": 3.6899120907028853, + "learning_rate": 1.9624206239303837e-05, + "loss": 1.1632, + "step": 1059 + }, + { + "epoch": 0.5754614549402823, + "grad_norm": 4.423539955787714, + "learning_rate": 1.9623250600175775e-05, + "loss": 0.9526, + "step": 1060 + }, + { + "epoch": 0.5760043431053203, + "grad_norm": 5.062986942342047, + "learning_rate": 1.962229377082874e-05, + "loss": 1.6768, + "step": 1061 + }, + { + "epoch": 0.5765472312703583, + "grad_norm": 5.044415439938055, + "learning_rate": 1.962133575138108e-05, + "loss": 1.3808, + "step": 1062 + }, + { + "epoch": 0.5770901194353963, + "grad_norm": 3.1405283495786587, + "learning_rate": 1.962037654195128e-05, + "loss": 0.7588, + "step": 1063 + }, + { + "epoch": 0.5776330076004343, + "grad_norm": 5.449873922495929, + "learning_rate": 1.9619416142657974e-05, + "loss": 1.1544, + "step": 1064 + }, + { + "epoch": 0.5781758957654723, + "grad_norm": 4.268527750000251, + "learning_rate": 1.961845455361995e-05, + "loss": 1.1941, + "step": 1065 + }, + { + "epoch": 0.5787187839305103, + "grad_norm": 4.6641878464575095, + "learning_rate": 1.9617491774956137e-05, + "loss": 1.7162, + "step": 1066 + }, + { + "epoch": 0.5792616720955484, + "grad_norm": 4.583539478633983, + "learning_rate": 1.9616527806785612e-05, + "loss": 1.1415, + "step": 1067 + }, + { + "epoch": 0.5798045602605864, + "grad_norm": 4.325093393868848, + "learning_rate": 1.96155626492276e-05, + "loss": 1.3284, + "step": 1068 + }, + { + "epoch": 0.5803474484256244, + "grad_norm": 4.180109104666535, + "learning_rate": 1.9614596302401478e-05, + "loss": 1.136, + "step": 1069 + }, + { + "epoch": 0.5808903365906624, + "grad_norm": 4.108700840520527, + "learning_rate": 1.9613628766426762e-05, + "loss": 1.6142, + "step": 1070 + }, + { + "epoch": 0.5814332247557004, + "grad_norm": 5.1931558422083555, + "learning_rate": 1.9612660041423116e-05, + "loss": 0.8479, + "step": 1071 + }, + { + "epoch": 0.5819761129207384, + "grad_norm": 6.008601497593343, + "learning_rate": 1.9611690127510358e-05, + "loss": 1.4743, + "step": 1072 + }, + { + "epoch": 0.5825190010857764, + "grad_norm": 4.531622764164067, + "learning_rate": 1.9610719024808444e-05, + "loss": 1.5755, + "step": 1073 + }, + { + "epoch": 0.5830618892508144, + "grad_norm": 6.18647874774261, + "learning_rate": 1.9609746733437486e-05, + "loss": 1.7369, + "step": 1074 + }, + { + "epoch": 0.5836047774158524, + "grad_norm": 5.5880249590216025, + "learning_rate": 1.9608773253517738e-05, + "loss": 1.4402, + "step": 1075 + }, + { + "epoch": 0.5841476655808904, + "grad_norm": 4.798692544278573, + "learning_rate": 1.9607798585169595e-05, + "loss": 1.2257, + "step": 1076 + }, + { + "epoch": 0.5846905537459284, + "grad_norm": 5.635786862796798, + "learning_rate": 1.960682272851361e-05, + "loss": 1.183, + "step": 1077 + }, + { + "epoch": 0.5852334419109664, + "grad_norm": 3.3979344506068037, + "learning_rate": 1.9605845683670484e-05, + "loss": 0.6463, + "step": 1078 + }, + { + "epoch": 0.5857763300760044, + "grad_norm": 4.5152680576801405, + "learning_rate": 1.960486745076105e-05, + "loss": 0.8476, + "step": 1079 + }, + { + "epoch": 0.5863192182410424, + "grad_norm": 4.4606140059612835, + "learning_rate": 1.9603888029906305e-05, + "loss": 1.185, + "step": 1080 + }, + { + "epoch": 0.5868621064060804, + "grad_norm": 4.229219175605004, + "learning_rate": 1.960290742122738e-05, + "loss": 1.3672, + "step": 1081 + }, + { + "epoch": 0.5874049945711184, + "grad_norm": 4.964172517040214, + "learning_rate": 1.960192562484556e-05, + "loss": 1.0561, + "step": 1082 + }, + { + "epoch": 0.5879478827361564, + "grad_norm": 4.8348336861653785, + "learning_rate": 1.960094264088228e-05, + "loss": 1.1005, + "step": 1083 + }, + { + "epoch": 0.5884907709011944, + "grad_norm": 4.3153540955607514, + "learning_rate": 1.959995846945911e-05, + "loss": 1.6513, + "step": 1084 + }, + { + "epoch": 0.5890336590662324, + "grad_norm": 4.733352525701168, + "learning_rate": 1.9598973110697773e-05, + "loss": 1.6162, + "step": 1085 + }, + { + "epoch": 0.5895765472312704, + "grad_norm": 5.402815147148315, + "learning_rate": 1.959798656472015e-05, + "loss": 1.4178, + "step": 1086 + }, + { + "epoch": 0.5901194353963084, + "grad_norm": 3.8276911935691147, + "learning_rate": 1.9596998831648247e-05, + "loss": 1.1554, + "step": 1087 + }, + { + "epoch": 0.5906623235613464, + "grad_norm": 4.499779496087119, + "learning_rate": 1.9596009911604232e-05, + "loss": 0.9786, + "step": 1088 + }, + { + "epoch": 0.5912052117263844, + "grad_norm": 4.3705352636049595, + "learning_rate": 1.959501980471042e-05, + "loss": 1.0827, + "step": 1089 + }, + { + "epoch": 0.5917480998914224, + "grad_norm": 4.9770508472387025, + "learning_rate": 1.9594028511089264e-05, + "loss": 1.2512, + "step": 1090 + }, + { + "epoch": 0.5922909880564604, + "grad_norm": 4.23068807802663, + "learning_rate": 1.9593036030863376e-05, + "loss": 0.7528, + "step": 1091 + }, + { + "epoch": 0.5928338762214984, + "grad_norm": 4.321323716192747, + "learning_rate": 1.9592042364155496e-05, + "loss": 1.0773, + "step": 1092 + }, + { + "epoch": 0.5933767643865364, + "grad_norm": 3.380036112638087, + "learning_rate": 1.9591047511088535e-05, + "loss": 0.7238, + "step": 1093 + }, + { + "epoch": 0.5939196525515744, + "grad_norm": 3.9722637658815447, + "learning_rate": 1.959005147178553e-05, + "loss": 0.8132, + "step": 1094 + }, + { + "epoch": 0.5944625407166124, + "grad_norm": 4.354488865716506, + "learning_rate": 1.9589054246369673e-05, + "loss": 1.1214, + "step": 1095 + }, + { + "epoch": 0.5950054288816504, + "grad_norm": 4.337355429292092, + "learning_rate": 1.9588055834964307e-05, + "loss": 0.9039, + "step": 1096 + }, + { + "epoch": 0.5955483170466884, + "grad_norm": 5.864047590425705, + "learning_rate": 1.9587056237692912e-05, + "loss": 1.6648, + "step": 1097 + }, + { + "epoch": 0.5960912052117264, + "grad_norm": 4.904173131706714, + "learning_rate": 1.9586055454679123e-05, + "loss": 1.1465, + "step": 1098 + }, + { + "epoch": 0.5966340933767644, + "grad_norm": 4.411877930208604, + "learning_rate": 1.9585053486046717e-05, + "loss": 1.4579, + "step": 1099 + }, + { + "epoch": 0.5971769815418024, + "grad_norm": 4.985171448881133, + "learning_rate": 1.9584050331919616e-05, + "loss": 1.5497, + "step": 1100 + }, + { + "epoch": 0.5977198697068404, + "grad_norm": 4.950532301400581, + "learning_rate": 1.9583045992421902e-05, + "loss": 1.3811, + "step": 1101 + }, + { + "epoch": 0.5982627578718784, + "grad_norm": 4.154119544907876, + "learning_rate": 1.9582040467677782e-05, + "loss": 1.186, + "step": 1102 + }, + { + "epoch": 0.5988056460369164, + "grad_norm": 4.069018604381756, + "learning_rate": 1.9581033757811628e-05, + "loss": 0.9312, + "step": 1103 + }, + { + "epoch": 0.5993485342019544, + "grad_norm": 5.33070083129983, + "learning_rate": 1.9580025862947948e-05, + "loss": 1.4274, + "step": 1104 + }, + { + "epoch": 0.5998914223669924, + "grad_norm": 4.855606632567906, + "learning_rate": 1.95790167832114e-05, + "loss": 1.0875, + "step": 1105 + }, + { + "epoch": 0.6004343105320304, + "grad_norm": 5.30182267546013, + "learning_rate": 1.957800651872679e-05, + "loss": 1.6501, + "step": 1106 + }, + { + "epoch": 0.6009771986970684, + "grad_norm": 4.433903391255917, + "learning_rate": 1.957699506961907e-05, + "loss": 1.3897, + "step": 1107 + }, + { + "epoch": 0.6015200868621065, + "grad_norm": 4.687653782723915, + "learning_rate": 1.9575982436013335e-05, + "loss": 0.8484, + "step": 1108 + }, + { + "epoch": 0.6020629750271445, + "grad_norm": 5.650426140105992, + "learning_rate": 1.957496861803483e-05, + "loss": 1.923, + "step": 1109 + }, + { + "epoch": 0.6026058631921825, + "grad_norm": 4.024650772512973, + "learning_rate": 1.957395361580895e-05, + "loss": 1.2713, + "step": 1110 + }, + { + "epoch": 0.6031487513572205, + "grad_norm": 4.343921133786845, + "learning_rate": 1.9572937429461223e-05, + "loss": 1.3152, + "step": 1111 + }, + { + "epoch": 0.6036916395222585, + "grad_norm": 4.1467176317930265, + "learning_rate": 1.957192005911734e-05, + "loss": 1.3325, + "step": 1112 + }, + { + "epoch": 0.6042345276872965, + "grad_norm": 4.305465221711973, + "learning_rate": 1.9570901504903128e-05, + "loss": 0.9237, + "step": 1113 + }, + { + "epoch": 0.6047774158523345, + "grad_norm": 4.101045495900469, + "learning_rate": 1.9569881766944564e-05, + "loss": 1.3932, + "step": 1114 + }, + { + "epoch": 0.6053203040173725, + "grad_norm": 4.0920748344339835, + "learning_rate": 1.956886084536777e-05, + "loss": 1.0067, + "step": 1115 + }, + { + "epoch": 0.6058631921824105, + "grad_norm": 4.844431281019285, + "learning_rate": 1.956783874029902e-05, + "loss": 1.2726, + "step": 1116 + }, + { + "epoch": 0.6064060803474485, + "grad_norm": 4.329792546128024, + "learning_rate": 1.9566815451864723e-05, + "loss": 1.1144, + "step": 1117 + }, + { + "epoch": 0.6069489685124865, + "grad_norm": 5.430322578642178, + "learning_rate": 1.9565790980191447e-05, + "loss": 1.3307, + "step": 1118 + }, + { + "epoch": 0.6074918566775245, + "grad_norm": 3.672369600344093, + "learning_rate": 1.9564765325405895e-05, + "loss": 0.8514, + "step": 1119 + }, + { + "epoch": 0.6080347448425625, + "grad_norm": 3.6433282223345334, + "learning_rate": 1.9563738487634924e-05, + "loss": 1.1972, + "step": 1120 + }, + { + "epoch": 0.6085776330076005, + "grad_norm": 4.23249445053715, + "learning_rate": 1.956271046700553e-05, + "loss": 1.0638, + "step": 1121 + }, + { + "epoch": 0.6091205211726385, + "grad_norm": 3.9162807435580134, + "learning_rate": 1.956168126364487e-05, + "loss": 1.121, + "step": 1122 + }, + { + "epoch": 0.6096634093376765, + "grad_norm": 3.3651337591577812, + "learning_rate": 1.956065087768023e-05, + "loss": 0.9604, + "step": 1123 + }, + { + "epoch": 0.6102062975027145, + "grad_norm": 5.5905307606778685, + "learning_rate": 1.955961930923905e-05, + "loss": 1.4509, + "step": 1124 + }, + { + "epoch": 0.6107491856677525, + "grad_norm": 3.6120233107886404, + "learning_rate": 1.955858655844892e-05, + "loss": 1.0106, + "step": 1125 + }, + { + "epoch": 0.6112920738327905, + "grad_norm": 3.6242967727445468, + "learning_rate": 1.9557552625437574e-05, + "loss": 1.0768, + "step": 1126 + }, + { + "epoch": 0.6118349619978285, + "grad_norm": 4.856763858492221, + "learning_rate": 1.9556517510332883e-05, + "loss": 1.6372, + "step": 1127 + }, + { + "epoch": 0.6123778501628665, + "grad_norm": 4.648427809898363, + "learning_rate": 1.9555481213262873e-05, + "loss": 1.1622, + "step": 1128 + }, + { + "epoch": 0.6129207383279045, + "grad_norm": 4.228089396781667, + "learning_rate": 1.9554443734355723e-05, + "loss": 1.067, + "step": 1129 + }, + { + "epoch": 0.6134636264929425, + "grad_norm": 4.069786144676624, + "learning_rate": 1.9553405073739743e-05, + "loss": 1.1427, + "step": 1130 + }, + { + "epoch": 0.6140065146579805, + "grad_norm": 3.112182391585594, + "learning_rate": 1.9552365231543395e-05, + "loss": 0.7862, + "step": 1131 + }, + { + "epoch": 0.6145494028230185, + "grad_norm": 4.783518734299315, + "learning_rate": 1.955132420789529e-05, + "loss": 1.0366, + "step": 1132 + }, + { + "epoch": 0.6150922909880565, + "grad_norm": 4.905128048667836, + "learning_rate": 1.9550282002924187e-05, + "loss": 1.2566, + "step": 1133 + }, + { + "epoch": 0.6156351791530945, + "grad_norm": 4.643678450407587, + "learning_rate": 1.9549238616758987e-05, + "loss": 1.2749, + "step": 1134 + }, + { + "epoch": 0.6161780673181325, + "grad_norm": 4.18363664236292, + "learning_rate": 1.954819404952873e-05, + "loss": 1.1523, + "step": 1135 + }, + { + "epoch": 0.6167209554831705, + "grad_norm": 3.9776065520060335, + "learning_rate": 1.9547148301362623e-05, + "loss": 0.8672, + "step": 1136 + }, + { + "epoch": 0.6172638436482085, + "grad_norm": 5.417170242762362, + "learning_rate": 1.9546101372389994e-05, + "loss": 1.6434, + "step": 1137 + }, + { + "epoch": 0.6178067318132465, + "grad_norm": 3.3325133642660068, + "learning_rate": 1.9545053262740335e-05, + "loss": 1.1447, + "step": 1138 + }, + { + "epoch": 0.6183496199782845, + "grad_norm": 5.4727757064949545, + "learning_rate": 1.9544003972543273e-05, + "loss": 1.5013, + "step": 1139 + }, + { + "epoch": 0.6188925081433225, + "grad_norm": 4.619714236951128, + "learning_rate": 1.954295350192859e-05, + "loss": 0.9758, + "step": 1140 + }, + { + "epoch": 0.6194353963083605, + "grad_norm": 5.189787937143953, + "learning_rate": 1.954190185102621e-05, + "loss": 1.2771, + "step": 1141 + }, + { + "epoch": 0.6199782844733985, + "grad_norm": 6.571114312185869, + "learning_rate": 1.9540849019966198e-05, + "loss": 1.2805, + "step": 1142 + }, + { + "epoch": 0.6205211726384365, + "grad_norm": 5.6343515388196455, + "learning_rate": 1.9539795008878774e-05, + "loss": 1.4061, + "step": 1143 + }, + { + "epoch": 0.6210640608034745, + "grad_norm": 5.582702517092888, + "learning_rate": 1.9538739817894302e-05, + "loss": 1.613, + "step": 1144 + }, + { + "epoch": 0.6216069489685125, + "grad_norm": 6.194616996988272, + "learning_rate": 1.9537683447143287e-05, + "loss": 1.114, + "step": 1145 + }, + { + "epoch": 0.6221498371335505, + "grad_norm": 6.0395907363668515, + "learning_rate": 1.9536625896756377e-05, + "loss": 1.5233, + "step": 1146 + }, + { + "epoch": 0.6226927252985885, + "grad_norm": 4.672511820840566, + "learning_rate": 1.9535567166864382e-05, + "loss": 1.1491, + "step": 1147 + }, + { + "epoch": 0.6232356134636265, + "grad_norm": 4.954133576744497, + "learning_rate": 1.9534507257598244e-05, + "loss": 1.1974, + "step": 1148 + }, + { + "epoch": 0.6237785016286646, + "grad_norm": 4.447205012745297, + "learning_rate": 1.953344616908905e-05, + "loss": 1.019, + "step": 1149 + }, + { + "epoch": 0.6243213897937026, + "grad_norm": 4.284087771325503, + "learning_rate": 1.9532383901468038e-05, + "loss": 1.0814, + "step": 1150 + }, + { + "epoch": 0.6248642779587406, + "grad_norm": 3.576067836847051, + "learning_rate": 1.9531320454866595e-05, + "loss": 1.1025, + "step": 1151 + }, + { + "epoch": 0.6254071661237784, + "grad_norm": 3.6831292979487777, + "learning_rate": 1.9530255829416246e-05, + "loss": 0.967, + "step": 1152 + }, + { + "epoch": 0.6259500542888164, + "grad_norm": 3.842496728015195, + "learning_rate": 1.952919002524867e-05, + "loss": 1.0903, + "step": 1153 + }, + { + "epoch": 0.6264929424538545, + "grad_norm": 4.077633480381933, + "learning_rate": 1.952812304249568e-05, + "loss": 0.9227, + "step": 1154 + }, + { + "epoch": 0.6270358306188925, + "grad_norm": 4.157004955121598, + "learning_rate": 1.952705488128925e-05, + "loss": 0.9427, + "step": 1155 + }, + { + "epoch": 0.6275787187839305, + "grad_norm": 5.267614104282471, + "learning_rate": 1.952598554176149e-05, + "loss": 1.9875, + "step": 1156 + }, + { + "epoch": 0.6281216069489685, + "grad_norm": 3.8938350314376735, + "learning_rate": 1.952491502404465e-05, + "loss": 1.1662, + "step": 1157 + }, + { + "epoch": 0.6286644951140065, + "grad_norm": 4.851409026807682, + "learning_rate": 1.9523843328271144e-05, + "loss": 1.9559, + "step": 1158 + }, + { + "epoch": 0.6292073832790445, + "grad_norm": 4.4104613077358845, + "learning_rate": 1.9522770454573513e-05, + "loss": 0.8011, + "step": 1159 + }, + { + "epoch": 0.6297502714440825, + "grad_norm": 3.6883335641947856, + "learning_rate": 1.952169640308446e-05, + "loss": 1.1838, + "step": 1160 + }, + { + "epoch": 0.6302931596091205, + "grad_norm": 3.728203039459734, + "learning_rate": 1.9520621173936818e-05, + "loss": 1.1423, + "step": 1161 + }, + { + "epoch": 0.6308360477741585, + "grad_norm": 3.6272612276564296, + "learning_rate": 1.9519544767263574e-05, + "loss": 0.8376, + "step": 1162 + }, + { + "epoch": 0.6313789359391965, + "grad_norm": 3.928276333701112, + "learning_rate": 1.951846718319786e-05, + "loss": 1.0749, + "step": 1163 + }, + { + "epoch": 0.6319218241042345, + "grad_norm": 5.941949883397345, + "learning_rate": 1.951738842187296e-05, + "loss": 1.8535, + "step": 1164 + }, + { + "epoch": 0.6324647122692725, + "grad_norm": 3.4741200080360644, + "learning_rate": 1.951630848342229e-05, + "loss": 0.7537, + "step": 1165 + }, + { + "epoch": 0.6330076004343105, + "grad_norm": 5.392480994365815, + "learning_rate": 1.9515227367979416e-05, + "loss": 0.8776, + "step": 1166 + }, + { + "epoch": 0.6335504885993485, + "grad_norm": 5.203169892005258, + "learning_rate": 1.951414507567806e-05, + "loss": 1.2678, + "step": 1167 + }, + { + "epoch": 0.6340933767643865, + "grad_norm": 7.499226723318992, + "learning_rate": 1.9513061606652076e-05, + "loss": 1.3295, + "step": 1168 + }, + { + "epoch": 0.6346362649294245, + "grad_norm": 4.355172848230307, + "learning_rate": 1.9511976961035474e-05, + "loss": 0.9188, + "step": 1169 + }, + { + "epoch": 0.6351791530944625, + "grad_norm": 4.983220943925928, + "learning_rate": 1.9510891138962398e-05, + "loss": 1.1951, + "step": 1170 + }, + { + "epoch": 0.6357220412595005, + "grad_norm": 5.3098934312394395, + "learning_rate": 1.950980414056715e-05, + "loss": 0.855, + "step": 1171 + }, + { + "epoch": 0.6362649294245385, + "grad_norm": 5.125683116898323, + "learning_rate": 1.950871596598417e-05, + "loss": 1.1371, + "step": 1172 + }, + { + "epoch": 0.6368078175895765, + "grad_norm": 4.3010814264850215, + "learning_rate": 1.950762661534804e-05, + "loss": 1.7401, + "step": 1173 + }, + { + "epoch": 0.6373507057546145, + "grad_norm": 5.860625964437982, + "learning_rate": 1.95065360887935e-05, + "loss": 1.3557, + "step": 1174 + }, + { + "epoch": 0.6378935939196525, + "grad_norm": 6.3356570754666155, + "learning_rate": 1.9505444386455426e-05, + "loss": 1.2281, + "step": 1175 + }, + { + "epoch": 0.6384364820846905, + "grad_norm": 4.788366086551355, + "learning_rate": 1.9504351508468842e-05, + "loss": 1.3458, + "step": 1176 + }, + { + "epoch": 0.6389793702497285, + "grad_norm": 4.5553889359557065, + "learning_rate": 1.9503257454968914e-05, + "loss": 0.803, + "step": 1177 + }, + { + "epoch": 0.6395222584147665, + "grad_norm": 4.9063187966995345, + "learning_rate": 1.950216222609096e-05, + "loss": 0.8306, + "step": 1178 + }, + { + "epoch": 0.6400651465798045, + "grad_norm": 4.775138114897886, + "learning_rate": 1.9501065821970435e-05, + "loss": 0.9783, + "step": 1179 + }, + { + "epoch": 0.6406080347448425, + "grad_norm": 4.53672284240203, + "learning_rate": 1.9499968242742948e-05, + "loss": 1.3076, + "step": 1180 + }, + { + "epoch": 0.6411509229098805, + "grad_norm": 4.539860050199487, + "learning_rate": 1.949886948854425e-05, + "loss": 1.4212, + "step": 1181 + }, + { + "epoch": 0.6416938110749185, + "grad_norm": 4.251462432005476, + "learning_rate": 1.9497769559510232e-05, + "loss": 1.0274, + "step": 1182 + }, + { + "epoch": 0.6422366992399565, + "grad_norm": 4.49791289564495, + "learning_rate": 1.9496668455776938e-05, + "loss": 1.7531, + "step": 1183 + }, + { + "epoch": 0.6427795874049945, + "grad_norm": 4.996740483321485, + "learning_rate": 1.9495566177480555e-05, + "loss": 1.201, + "step": 1184 + }, + { + "epoch": 0.6433224755700325, + "grad_norm": 5.347352236818569, + "learning_rate": 1.9494462724757413e-05, + "loss": 1.5723, + "step": 1185 + }, + { + "epoch": 0.6438653637350705, + "grad_norm": 4.2242558589374015, + "learning_rate": 1.9493358097743988e-05, + "loss": 1.8014, + "step": 1186 + }, + { + "epoch": 0.6444082519001085, + "grad_norm": 5.042556201755935, + "learning_rate": 1.9492252296576906e-05, + "loss": 1.5466, + "step": 1187 + }, + { + "epoch": 0.6449511400651465, + "grad_norm": 4.284314238035697, + "learning_rate": 1.949114532139293e-05, + "loss": 0.9515, + "step": 1188 + }, + { + "epoch": 0.6454940282301845, + "grad_norm": 4.699826119048575, + "learning_rate": 1.9490037172328974e-05, + "loss": 1.1677, + "step": 1189 + }, + { + "epoch": 0.6460369163952225, + "grad_norm": 6.035247135554916, + "learning_rate": 1.9488927849522095e-05, + "loss": 1.957, + "step": 1190 + }, + { + "epoch": 0.6465798045602605, + "grad_norm": 4.042004235065528, + "learning_rate": 1.94878173531095e-05, + "loss": 1.2182, + "step": 1191 + }, + { + "epoch": 0.6471226927252985, + "grad_norm": 4.114859597350168, + "learning_rate": 1.948670568322853e-05, + "loss": 1.1787, + "step": 1192 + }, + { + "epoch": 0.6476655808903365, + "grad_norm": 4.380177461001504, + "learning_rate": 1.9485592840016682e-05, + "loss": 1.3912, + "step": 1193 + }, + { + "epoch": 0.6482084690553745, + "grad_norm": 5.160710982722015, + "learning_rate": 1.948447882361159e-05, + "loss": 1.2642, + "step": 1194 + }, + { + "epoch": 0.6487513572204126, + "grad_norm": 3.950160510485583, + "learning_rate": 1.9483363634151046e-05, + "loss": 1.1246, + "step": 1195 + }, + { + "epoch": 0.6492942453854506, + "grad_norm": 4.587449618376974, + "learning_rate": 1.9482247271772974e-05, + "loss": 1.3275, + "step": 1196 + }, + { + "epoch": 0.6498371335504886, + "grad_norm": 4.919831513668322, + "learning_rate": 1.9481129736615445e-05, + "loss": 1.4007, + "step": 1197 + }, + { + "epoch": 0.6503800217155266, + "grad_norm": 4.298954423192263, + "learning_rate": 1.948001102881668e-05, + "loss": 0.8076, + "step": 1198 + }, + { + "epoch": 0.6509229098805646, + "grad_norm": 5.525772014179547, + "learning_rate": 1.9478891148515043e-05, + "loss": 1.4558, + "step": 1199 + }, + { + "epoch": 0.6514657980456026, + "grad_norm": 5.473549669240299, + "learning_rate": 1.947777009584904e-05, + "loss": 1.0976, + "step": 1200 + }, + { + "epoch": 0.6520086862106406, + "grad_norm": 5.360010284326786, + "learning_rate": 1.947664787095733e-05, + "loss": 1.0462, + "step": 1201 + }, + { + "epoch": 0.6525515743756786, + "grad_norm": 4.696691081444868, + "learning_rate": 1.9475524473978705e-05, + "loss": 0.933, + "step": 1202 + }, + { + "epoch": 0.6530944625407166, + "grad_norm": 6.0503072087093965, + "learning_rate": 1.947439990505211e-05, + "loss": 0.9628, + "step": 1203 + }, + { + "epoch": 0.6536373507057546, + "grad_norm": 5.476272854511399, + "learning_rate": 1.9473274164316637e-05, + "loss": 0.9716, + "step": 1204 + }, + { + "epoch": 0.6541802388707926, + "grad_norm": 3.910448715873912, + "learning_rate": 1.9472147251911517e-05, + "loss": 1.0732, + "step": 1205 + }, + { + "epoch": 0.6547231270358306, + "grad_norm": 4.596114022138067, + "learning_rate": 1.9471019167976126e-05, + "loss": 0.9399, + "step": 1206 + }, + { + "epoch": 0.6552660152008686, + "grad_norm": 6.764840749864346, + "learning_rate": 1.946988991264999e-05, + "loss": 2.0529, + "step": 1207 + }, + { + "epoch": 0.6558089033659066, + "grad_norm": 5.217328230116913, + "learning_rate": 1.9468759486072778e-05, + "loss": 1.0477, + "step": 1208 + }, + { + "epoch": 0.6563517915309446, + "grad_norm": 4.03429219354135, + "learning_rate": 1.9467627888384303e-05, + "loss": 1.1103, + "step": 1209 + }, + { + "epoch": 0.6568946796959826, + "grad_norm": 4.301388894222403, + "learning_rate": 1.946649511972452e-05, + "loss": 0.8515, + "step": 1210 + }, + { + "epoch": 0.6574375678610206, + "grad_norm": 5.285446092583105, + "learning_rate": 1.9465361180233536e-05, + "loss": 1.2881, + "step": 1211 + }, + { + "epoch": 0.6579804560260586, + "grad_norm": 5.879897026479874, + "learning_rate": 1.9464226070051593e-05, + "loss": 0.9362, + "step": 1212 + }, + { + "epoch": 0.6585233441910966, + "grad_norm": 5.335257600571856, + "learning_rate": 1.9463089789319083e-05, + "loss": 1.4372, + "step": 1213 + }, + { + "epoch": 0.6590662323561346, + "grad_norm": 5.272070125431448, + "learning_rate": 1.9461952338176552e-05, + "loss": 1.2063, + "step": 1214 + }, + { + "epoch": 0.6596091205211726, + "grad_norm": 5.188765715525815, + "learning_rate": 1.946081371676467e-05, + "loss": 0.8912, + "step": 1215 + }, + { + "epoch": 0.6601520086862106, + "grad_norm": 4.747539421286818, + "learning_rate": 1.9459673925224275e-05, + "loss": 1.0381, + "step": 1216 + }, + { + "epoch": 0.6606948968512486, + "grad_norm": 4.486254331013965, + "learning_rate": 1.945853296369633e-05, + "loss": 1.2584, + "step": 1217 + }, + { + "epoch": 0.6612377850162866, + "grad_norm": 4.730012415000151, + "learning_rate": 1.945739083232195e-05, + "loss": 1.1061, + "step": 1218 + }, + { + "epoch": 0.6617806731813246, + "grad_norm": 5.713566521874704, + "learning_rate": 1.9456247531242405e-05, + "loss": 1.4628, + "step": 1219 + }, + { + "epoch": 0.6623235613463626, + "grad_norm": 6.08134803588137, + "learning_rate": 1.9455103060599093e-05, + "loss": 1.6743, + "step": 1220 + }, + { + "epoch": 0.6628664495114006, + "grad_norm": 4.778151760566848, + "learning_rate": 1.9453957420533562e-05, + "loss": 0.8454, + "step": 1221 + }, + { + "epoch": 0.6634093376764386, + "grad_norm": 5.007171333863363, + "learning_rate": 1.945281061118751e-05, + "loss": 1.079, + "step": 1222 + }, + { + "epoch": 0.6639522258414766, + "grad_norm": 4.034543591270571, + "learning_rate": 1.945166263270278e-05, + "loss": 0.8378, + "step": 1223 + }, + { + "epoch": 0.6644951140065146, + "grad_norm": 4.191899722379318, + "learning_rate": 1.9450513485221352e-05, + "loss": 1.3602, + "step": 1224 + }, + { + "epoch": 0.6650380021715526, + "grad_norm": 6.349672663190246, + "learning_rate": 1.944936316888535e-05, + "loss": 1.861, + "step": 1225 + }, + { + "epoch": 0.6655808903365906, + "grad_norm": 5.991894404952759, + "learning_rate": 1.9448211683837055e-05, + "loss": 1.0931, + "step": 1226 + }, + { + "epoch": 0.6661237785016286, + "grad_norm": 4.405247018509644, + "learning_rate": 1.9447059030218876e-05, + "loss": 1.2239, + "step": 1227 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 4.810361623786047, + "learning_rate": 1.9445905208173387e-05, + "loss": 1.1468, + "step": 1228 + }, + { + "epoch": 0.6672095548317046, + "grad_norm": 4.6967843893557015, + "learning_rate": 1.944475021784328e-05, + "loss": 1.5304, + "step": 1229 + }, + { + "epoch": 0.6677524429967426, + "grad_norm": 5.325911382516086, + "learning_rate": 1.9443594059371417e-05, + "loss": 1.2218, + "step": 1230 + }, + { + "epoch": 0.6682953311617806, + "grad_norm": 4.416437363913103, + "learning_rate": 1.9442436732900788e-05, + "loss": 1.223, + "step": 1231 + }, + { + "epoch": 0.6688382193268186, + "grad_norm": 3.720560087272861, + "learning_rate": 1.9441278238574537e-05, + "loss": 0.8155, + "step": 1232 + }, + { + "epoch": 0.6693811074918566, + "grad_norm": 5.273159297494141, + "learning_rate": 1.9440118576535947e-05, + "loss": 1.1958, + "step": 1233 + }, + { + "epoch": 0.6699239956568946, + "grad_norm": 5.684867015184436, + "learning_rate": 1.9438957746928443e-05, + "loss": 1.2384, + "step": 1234 + }, + { + "epoch": 0.6704668838219326, + "grad_norm": 4.952541482610055, + "learning_rate": 1.9437795749895604e-05, + "loss": 1.337, + "step": 1235 + }, + { + "epoch": 0.6710097719869706, + "grad_norm": 4.788863413585636, + "learning_rate": 1.9436632585581145e-05, + "loss": 0.5182, + "step": 1236 + }, + { + "epoch": 0.6715526601520087, + "grad_norm": 5.723878805619219, + "learning_rate": 1.9435468254128925e-05, + "loss": 1.1444, + "step": 1237 + }, + { + "epoch": 0.6720955483170467, + "grad_norm": 4.3068839361839, + "learning_rate": 1.9434302755682958e-05, + "loss": 1.1286, + "step": 1238 + }, + { + "epoch": 0.6726384364820847, + "grad_norm": 3.7588872381668477, + "learning_rate": 1.9433136090387384e-05, + "loss": 0.9718, + "step": 1239 + }, + { + "epoch": 0.6731813246471227, + "grad_norm": 6.802207677110681, + "learning_rate": 1.9431968258386508e-05, + "loss": 1.2389, + "step": 1240 + }, + { + "epoch": 0.6737242128121607, + "grad_norm": 3.6180158875124735, + "learning_rate": 1.9430799259824766e-05, + "loss": 0.644, + "step": 1241 + }, + { + "epoch": 0.6742671009771987, + "grad_norm": 5.559822163300458, + "learning_rate": 1.9429629094846742e-05, + "loss": 1.1157, + "step": 1242 + }, + { + "epoch": 0.6748099891422367, + "grad_norm": 4.084355750224745, + "learning_rate": 1.942845776359716e-05, + "loss": 0.91, + "step": 1243 + }, + { + "epoch": 0.6753528773072747, + "grad_norm": 3.9402358983627583, + "learning_rate": 1.9427285266220895e-05, + "loss": 0.9409, + "step": 1244 + }, + { + "epoch": 0.6758957654723127, + "grad_norm": 5.093172430302549, + "learning_rate": 1.9426111602862968e-05, + "loss": 1.5905, + "step": 1245 + }, + { + "epoch": 0.6764386536373507, + "grad_norm": 4.02536040473524, + "learning_rate": 1.942493677366853e-05, + "loss": 0.9277, + "step": 1246 + }, + { + "epoch": 0.6769815418023887, + "grad_norm": 4.741533426757492, + "learning_rate": 1.942376077878289e-05, + "loss": 1.1027, + "step": 1247 + }, + { + "epoch": 0.6775244299674267, + "grad_norm": 5.365058237804207, + "learning_rate": 1.9422583618351503e-05, + "loss": 1.072, + "step": 1248 + }, + { + "epoch": 0.6780673181324647, + "grad_norm": 4.479339123691231, + "learning_rate": 1.9421405292519956e-05, + "loss": 1.474, + "step": 1249 + }, + { + "epoch": 0.6786102062975027, + "grad_norm": 7.79086989645888, + "learning_rate": 1.942022580143398e-05, + "loss": 1.6973, + "step": 1250 + }, + { + "epoch": 0.6791530944625407, + "grad_norm": 5.621467578044424, + "learning_rate": 1.9419045145239474e-05, + "loss": 1.5432, + "step": 1251 + }, + { + "epoch": 0.6796959826275787, + "grad_norm": 5.041038447258229, + "learning_rate": 1.9417863324082444e-05, + "loss": 1.2977, + "step": 1252 + }, + { + "epoch": 0.6802388707926167, + "grad_norm": 4.717189339301973, + "learning_rate": 1.941668033810907e-05, + "loss": 0.924, + "step": 1253 + }, + { + "epoch": 0.6807817589576547, + "grad_norm": 5.086833595497465, + "learning_rate": 1.9415496187465667e-05, + "loss": 1.5545, + "step": 1254 + }, + { + "epoch": 0.6813246471226927, + "grad_norm": 4.790403842984273, + "learning_rate": 1.941431087229869e-05, + "loss": 1.2545, + "step": 1255 + }, + { + "epoch": 0.6818675352877307, + "grad_norm": 4.470785518257313, + "learning_rate": 1.9413124392754735e-05, + "loss": 1.2036, + "step": 1256 + }, + { + "epoch": 0.6824104234527687, + "grad_norm": 3.820710537112263, + "learning_rate": 1.9411936748980557e-05, + "loss": 1.0856, + "step": 1257 + }, + { + "epoch": 0.6829533116178067, + "grad_norm": 4.453720544384727, + "learning_rate": 1.9410747941123036e-05, + "loss": 1.1645, + "step": 1258 + }, + { + "epoch": 0.6834961997828447, + "grad_norm": 4.692503519890958, + "learning_rate": 1.9409557969329218e-05, + "loss": 1.1174, + "step": 1259 + }, + { + "epoch": 0.6840390879478827, + "grad_norm": 5.666971506629844, + "learning_rate": 1.940836683374627e-05, + "loss": 1.4904, + "step": 1260 + }, + { + "epoch": 0.6845819761129207, + "grad_norm": 5.468580615919992, + "learning_rate": 1.940717453452152e-05, + "loss": 1.2834, + "step": 1261 + }, + { + "epoch": 0.6851248642779587, + "grad_norm": 6.42319990350808, + "learning_rate": 1.9405981071802425e-05, + "loss": 1.1144, + "step": 1262 + }, + { + "epoch": 0.6856677524429967, + "grad_norm": 6.078674980286373, + "learning_rate": 1.9404786445736607e-05, + "loss": 1.4214, + "step": 1263 + }, + { + "epoch": 0.6862106406080347, + "grad_norm": 4.465631588283012, + "learning_rate": 1.9403590656471806e-05, + "loss": 0.7813, + "step": 1264 + }, + { + "epoch": 0.6867535287730727, + "grad_norm": 5.220582351132811, + "learning_rate": 1.940239370415593e-05, + "loss": 1.2112, + "step": 1265 + }, + { + "epoch": 0.6872964169381107, + "grad_norm": 5.1374760487841895, + "learning_rate": 1.9401195588937014e-05, + "loss": 1.2402, + "step": 1266 + }, + { + "epoch": 0.6878393051031487, + "grad_norm": 5.106543779494583, + "learning_rate": 1.9399996310963243e-05, + "loss": 1.1846, + "step": 1267 + }, + { + "epoch": 0.6883821932681867, + "grad_norm": 3.9521299273320363, + "learning_rate": 1.939879587038295e-05, + "loss": 0.8975, + "step": 1268 + }, + { + "epoch": 0.6889250814332247, + "grad_norm": 5.398700470714175, + "learning_rate": 1.9397594267344604e-05, + "loss": 1.4686, + "step": 1269 + }, + { + "epoch": 0.6894679695982627, + "grad_norm": 5.8287973073589265, + "learning_rate": 1.939639150199682e-05, + "loss": 1.011, + "step": 1270 + }, + { + "epoch": 0.6900108577633007, + "grad_norm": 4.812101240482426, + "learning_rate": 1.9395187574488358e-05, + "loss": 0.948, + "step": 1271 + }, + { + "epoch": 0.6905537459283387, + "grad_norm": 5.828731649028842, + "learning_rate": 1.939398248496813e-05, + "loss": 1.2233, + "step": 1272 + }, + { + "epoch": 0.6910966340933767, + "grad_norm": 4.890080173319348, + "learning_rate": 1.9392776233585167e-05, + "loss": 0.8859, + "step": 1273 + }, + { + "epoch": 0.6916395222584147, + "grad_norm": 4.920508250004042, + "learning_rate": 1.9391568820488674e-05, + "loss": 0.73, + "step": 1274 + }, + { + "epoch": 0.6921824104234527, + "grad_norm": 5.635401553567158, + "learning_rate": 1.9390360245827983e-05, + "loss": 1.1432, + "step": 1275 + }, + { + "epoch": 0.6927252985884907, + "grad_norm": 4.588336191443314, + "learning_rate": 1.9389150509752566e-05, + "loss": 1.3326, + "step": 1276 + }, + { + "epoch": 0.6932681867535287, + "grad_norm": 4.631694372660094, + "learning_rate": 1.9387939612412056e-05, + "loss": 1.3762, + "step": 1277 + }, + { + "epoch": 0.6938110749185668, + "grad_norm": 5.585985863722577, + "learning_rate": 1.938672755395621e-05, + "loss": 1.3606, + "step": 1278 + }, + { + "epoch": 0.6943539630836048, + "grad_norm": 4.764783952643876, + "learning_rate": 1.938551433453494e-05, + "loss": 0.9858, + "step": 1279 + }, + { + "epoch": 0.6948968512486428, + "grad_norm": 4.234252918062867, + "learning_rate": 1.9384299954298297e-05, + "loss": 0.7702, + "step": 1280 + }, + { + "epoch": 0.6954397394136808, + "grad_norm": 4.36117548276201, + "learning_rate": 1.938308441339648e-05, + "loss": 1.1846, + "step": 1281 + }, + { + "epoch": 0.6959826275787188, + "grad_norm": 4.311957696864392, + "learning_rate": 1.938186771197983e-05, + "loss": 0.9934, + "step": 1282 + }, + { + "epoch": 0.6965255157437568, + "grad_norm": 5.117807004985933, + "learning_rate": 1.9380649850198824e-05, + "loss": 1.4474, + "step": 1283 + }, + { + "epoch": 0.6970684039087948, + "grad_norm": 5.695329527120276, + "learning_rate": 1.93794308282041e-05, + "loss": 1.3001, + "step": 1284 + }, + { + "epoch": 0.6976112920738328, + "grad_norm": 4.305244126285458, + "learning_rate": 1.937821064614642e-05, + "loss": 1.1421, + "step": 1285 + }, + { + "epoch": 0.6981541802388708, + "grad_norm": 4.574797635236639, + "learning_rate": 1.93769893041767e-05, + "loss": 1.1104, + "step": 1286 + }, + { + "epoch": 0.6986970684039088, + "grad_norm": 5.051193103711606, + "learning_rate": 1.9375766802446002e-05, + "loss": 1.2858, + "step": 1287 + }, + { + "epoch": 0.6992399565689468, + "grad_norm": 5.002235181102662, + "learning_rate": 1.9374543141105518e-05, + "loss": 1.2931, + "step": 1288 + }, + { + "epoch": 0.6997828447339848, + "grad_norm": 5.259386594746194, + "learning_rate": 1.93733183203066e-05, + "loss": 1.0432, + "step": 1289 + }, + { + "epoch": 0.7003257328990228, + "grad_norm": 3.7218449672815117, + "learning_rate": 1.9372092340200736e-05, + "loss": 0.9913, + "step": 1290 + }, + { + "epoch": 0.7008686210640608, + "grad_norm": 4.45109753373239, + "learning_rate": 1.937086520093955e-05, + "loss": 1.2012, + "step": 1291 + }, + { + "epoch": 0.7014115092290988, + "grad_norm": 4.73529586440727, + "learning_rate": 1.9369636902674823e-05, + "loss": 1.3995, + "step": 1292 + }, + { + "epoch": 0.7019543973941368, + "grad_norm": 3.845443933095641, + "learning_rate": 1.936840744555847e-05, + "loss": 1.0971, + "step": 1293 + }, + { + "epoch": 0.7024972855591748, + "grad_norm": 6.67046211091574, + "learning_rate": 1.9367176829742553e-05, + "loss": 1.5273, + "step": 1294 + }, + { + "epoch": 0.7030401737242128, + "grad_norm": 4.454761961182613, + "learning_rate": 1.9365945055379275e-05, + "loss": 1.1282, + "step": 1295 + }, + { + "epoch": 0.7035830618892508, + "grad_norm": 6.552953195473038, + "learning_rate": 1.936471212262099e-05, + "loss": 2.171, + "step": 1296 + }, + { + "epoch": 0.7041259500542888, + "grad_norm": 4.894073730847946, + "learning_rate": 1.9363478031620182e-05, + "loss": 1.2303, + "step": 1297 + }, + { + "epoch": 0.7046688382193268, + "grad_norm": 3.863202427039748, + "learning_rate": 1.936224278252949e-05, + "loss": 0.8515, + "step": 1298 + }, + { + "epoch": 0.7052117263843648, + "grad_norm": 5.503562663348476, + "learning_rate": 1.9361006375501685e-05, + "loss": 1.1834, + "step": 1299 + }, + { + "epoch": 0.7057546145494028, + "grad_norm": 5.318422186143503, + "learning_rate": 1.9359768810689697e-05, + "loss": 1.3378, + "step": 1300 + }, + { + "epoch": 0.7062975027144408, + "grad_norm": 4.46194766163203, + "learning_rate": 1.9358530088246582e-05, + "loss": 1.3318, + "step": 1301 + }, + { + "epoch": 0.7068403908794788, + "grad_norm": 5.6623901137247765, + "learning_rate": 1.9357290208325552e-05, + "loss": 1.3512, + "step": 1302 + }, + { + "epoch": 0.7073832790445168, + "grad_norm": 5.805010217787694, + "learning_rate": 1.9356049171079957e-05, + "loss": 1.2383, + "step": 1303 + }, + { + "epoch": 0.7079261672095548, + "grad_norm": 4.847713425075124, + "learning_rate": 1.9354806976663286e-05, + "loss": 0.7329, + "step": 1304 + }, + { + "epoch": 0.7084690553745928, + "grad_norm": 4.207200307039602, + "learning_rate": 1.935356362522918e-05, + "loss": 1.0119, + "step": 1305 + }, + { + "epoch": 0.7090119435396308, + "grad_norm": 5.76731204400203, + "learning_rate": 1.9352319116931417e-05, + "loss": 1.7383, + "step": 1306 + }, + { + "epoch": 0.7095548317046688, + "grad_norm": 4.989804434976344, + "learning_rate": 1.935107345192392e-05, + "loss": 1.1958, + "step": 1307 + }, + { + "epoch": 0.7100977198697068, + "grad_norm": 4.6099552434159135, + "learning_rate": 1.9349826630360757e-05, + "loss": 0.9933, + "step": 1308 + }, + { + "epoch": 0.7106406080347448, + "grad_norm": 4.693361205108708, + "learning_rate": 1.9348578652396136e-05, + "loss": 1.3553, + "step": 1309 + }, + { + "epoch": 0.7111834961997828, + "grad_norm": 5.473441077550542, + "learning_rate": 1.9347329518184406e-05, + "loss": 1.2202, + "step": 1310 + }, + { + "epoch": 0.7117263843648208, + "grad_norm": 5.885376904454794, + "learning_rate": 1.9346079227880062e-05, + "loss": 1.3698, + "step": 1311 + }, + { + "epoch": 0.7122692725298588, + "grad_norm": 4.115405735194196, + "learning_rate": 1.9344827781637744e-05, + "loss": 1.3361, + "step": 1312 + }, + { + "epoch": 0.7128121606948968, + "grad_norm": 6.745730370060005, + "learning_rate": 1.9343575179612236e-05, + "loss": 1.6411, + "step": 1313 + }, + { + "epoch": 0.7133550488599348, + "grad_norm": 4.363445628534919, + "learning_rate": 1.9342321421958455e-05, + "loss": 0.792, + "step": 1314 + }, + { + "epoch": 0.7138979370249728, + "grad_norm": 6.167581448099132, + "learning_rate": 1.9341066508831472e-05, + "loss": 1.4264, + "step": 1315 + }, + { + "epoch": 0.7144408251900108, + "grad_norm": 4.973571230650737, + "learning_rate": 1.9339810440386495e-05, + "loss": 0.9769, + "step": 1316 + }, + { + "epoch": 0.7149837133550488, + "grad_norm": 5.2637238951774945, + "learning_rate": 1.933855321677888e-05, + "loss": 1.2478, + "step": 1317 + }, + { + "epoch": 0.7155266015200868, + "grad_norm": 5.2296234862723505, + "learning_rate": 1.9337294838164118e-05, + "loss": 1.3999, + "step": 1318 + }, + { + "epoch": 0.7160694896851248, + "grad_norm": 4.720322767524389, + "learning_rate": 1.9336035304697848e-05, + "loss": 1.3561, + "step": 1319 + }, + { + "epoch": 0.7166123778501629, + "grad_norm": 4.2513159313917015, + "learning_rate": 1.9334774616535854e-05, + "loss": 1.176, + "step": 1320 + }, + { + "epoch": 0.7171552660152009, + "grad_norm": 4.5267349508423225, + "learning_rate": 1.9333512773834057e-05, + "loss": 1.1211, + "step": 1321 + }, + { + "epoch": 0.7176981541802389, + "grad_norm": 5.384909301403834, + "learning_rate": 1.9332249776748523e-05, + "loss": 1.3904, + "step": 1322 + }, + { + "epoch": 0.7182410423452769, + "grad_norm": 4.67839864320759, + "learning_rate": 1.9330985625435468e-05, + "loss": 0.8608, + "step": 1323 + }, + { + "epoch": 0.7187839305103149, + "grad_norm": 5.193826637693982, + "learning_rate": 1.9329720320051233e-05, + "loss": 1.3172, + "step": 1324 + }, + { + "epoch": 0.7193268186753529, + "grad_norm": 4.8605293911870096, + "learning_rate": 1.9328453860752324e-05, + "loss": 1.3134, + "step": 1325 + }, + { + "epoch": 0.7198697068403909, + "grad_norm": 5.685265101122876, + "learning_rate": 1.9327186247695377e-05, + "loss": 1.2036, + "step": 1326 + }, + { + "epoch": 0.7204125950054289, + "grad_norm": 5.159715270999642, + "learning_rate": 1.9325917481037164e-05, + "loss": 1.4898, + "step": 1327 + }, + { + "epoch": 0.7209554831704669, + "grad_norm": 5.793178164657722, + "learning_rate": 1.9324647560934613e-05, + "loss": 1.3909, + "step": 1328 + }, + { + "epoch": 0.7214983713355049, + "grad_norm": 3.7442056178189267, + "learning_rate": 1.9323376487544795e-05, + "loss": 0.774, + "step": 1329 + }, + { + "epoch": 0.7220412595005429, + "grad_norm": 4.962252385233631, + "learning_rate": 1.9322104261024912e-05, + "loss": 1.3528, + "step": 1330 + }, + { + "epoch": 0.7225841476655809, + "grad_norm": 5.4534463408495, + "learning_rate": 1.9320830881532316e-05, + "loss": 1.6754, + "step": 1331 + }, + { + "epoch": 0.7231270358306189, + "grad_norm": 6.301262380555263, + "learning_rate": 1.93195563492245e-05, + "loss": 1.5583, + "step": 1332 + }, + { + "epoch": 0.7236699239956569, + "grad_norm": 3.970618104957748, + "learning_rate": 1.9318280664259103e-05, + "loss": 0.8964, + "step": 1333 + }, + { + "epoch": 0.7242128121606949, + "grad_norm": 3.490105469488415, + "learning_rate": 1.9317003826793904e-05, + "loss": 0.8098, + "step": 1334 + }, + { + "epoch": 0.7247557003257329, + "grad_norm": 5.240119278714349, + "learning_rate": 1.9315725836986822e-05, + "loss": 1.1688, + "step": 1335 + }, + { + "epoch": 0.7252985884907709, + "grad_norm": 4.550130688769963, + "learning_rate": 1.931444669499592e-05, + "loss": 1.2569, + "step": 1336 + }, + { + "epoch": 0.7258414766558089, + "grad_norm": 5.058353172472431, + "learning_rate": 1.9313166400979404e-05, + "loss": 1.3471, + "step": 1337 + }, + { + "epoch": 0.7263843648208469, + "grad_norm": 4.528967594836655, + "learning_rate": 1.931188495509563e-05, + "loss": 1.1831, + "step": 1338 + }, + { + "epoch": 0.7269272529858849, + "grad_norm": 4.493564615513058, + "learning_rate": 1.931060235750308e-05, + "loss": 1.4889, + "step": 1339 + }, + { + "epoch": 0.7274701411509229, + "grad_norm": 5.176344629525538, + "learning_rate": 1.9309318608360392e-05, + "loss": 1.6704, + "step": 1340 + }, + { + "epoch": 0.7280130293159609, + "grad_norm": 6.738195156936204, + "learning_rate": 1.930803370782634e-05, + "loss": 1.2147, + "step": 1341 + }, + { + "epoch": 0.7285559174809989, + "grad_norm": 4.248439922934747, + "learning_rate": 1.9306747656059847e-05, + "loss": 1.1943, + "step": 1342 + }, + { + "epoch": 0.7290988056460369, + "grad_norm": 5.175389864390461, + "learning_rate": 1.930546045321997e-05, + "loss": 1.4587, + "step": 1343 + }, + { + "epoch": 0.7296416938110749, + "grad_norm": 5.19934365393201, + "learning_rate": 1.9304172099465914e-05, + "loss": 0.9433, + "step": 1344 + }, + { + "epoch": 0.7301845819761129, + "grad_norm": 6.18681550522946, + "learning_rate": 1.9302882594957025e-05, + "loss": 1.5576, + "step": 1345 + }, + { + "epoch": 0.7307274701411509, + "grad_norm": 4.311510847746277, + "learning_rate": 1.930159193985279e-05, + "loss": 1.327, + "step": 1346 + }, + { + "epoch": 0.7312703583061889, + "grad_norm": 4.044345899710192, + "learning_rate": 1.9300300134312838e-05, + "loss": 1.2106, + "step": 1347 + }, + { + "epoch": 0.7318132464712269, + "grad_norm": 4.181894752624144, + "learning_rate": 1.929900717849694e-05, + "loss": 0.8703, + "step": 1348 + }, + { + "epoch": 0.7323561346362649, + "grad_norm": 5.37566626429507, + "learning_rate": 1.929771307256502e-05, + "loss": 1.446, + "step": 1349 + }, + { + "epoch": 0.7328990228013029, + "grad_norm": 5.37247658500195, + "learning_rate": 1.9296417816677123e-05, + "loss": 0.9239, + "step": 1350 + }, + { + "epoch": 0.7334419109663409, + "grad_norm": 5.749510968066806, + "learning_rate": 1.929512141099346e-05, + "loss": 1.4799, + "step": 1351 + }, + { + "epoch": 0.7339847991313789, + "grad_norm": 5.079635654383841, + "learning_rate": 1.929382385567436e-05, + "loss": 1.2571, + "step": 1352 + }, + { + "epoch": 0.7345276872964169, + "grad_norm": 4.923400923431618, + "learning_rate": 1.929252515088032e-05, + "loss": 1.1967, + "step": 1353 + }, + { + "epoch": 0.7350705754614549, + "grad_norm": 4.397561354835364, + "learning_rate": 1.9291225296771957e-05, + "loss": 0.9088, + "step": 1354 + }, + { + "epoch": 0.7356134636264929, + "grad_norm": 4.13744065040371, + "learning_rate": 1.9289924293510037e-05, + "loss": 1.023, + "step": 1355 + }, + { + "epoch": 0.7361563517915309, + "grad_norm": 4.266553618249588, + "learning_rate": 1.9288622141255477e-05, + "loss": 1.019, + "step": 1356 + }, + { + "epoch": 0.7366992399565689, + "grad_norm": 4.52588883698767, + "learning_rate": 1.928731884016933e-05, + "loss": 1.0935, + "step": 1357 + }, + { + "epoch": 0.7372421281216069, + "grad_norm": 5.46382933116543, + "learning_rate": 1.9286014390412786e-05, + "loss": 1.2833, + "step": 1358 + }, + { + "epoch": 0.737785016286645, + "grad_norm": 4.88651312061201, + "learning_rate": 1.928470879214718e-05, + "loss": 1.6064, + "step": 1359 + }, + { + "epoch": 0.738327904451683, + "grad_norm": 5.403882256899124, + "learning_rate": 1.9283402045533995e-05, + "loss": 1.1537, + "step": 1360 + }, + { + "epoch": 0.738870792616721, + "grad_norm": 6.381532932652649, + "learning_rate": 1.928209415073485e-05, + "loss": 1.527, + "step": 1361 + }, + { + "epoch": 0.739413680781759, + "grad_norm": 4.795244725313847, + "learning_rate": 1.9280785107911505e-05, + "loss": 1.4696, + "step": 1362 + }, + { + "epoch": 0.739956568946797, + "grad_norm": 5.8579061053451955, + "learning_rate": 1.9279474917225866e-05, + "loss": 1.5566, + "step": 1363 + }, + { + "epoch": 0.740499457111835, + "grad_norm": 4.5018596923036185, + "learning_rate": 1.927816357883998e-05, + "loss": 0.9842, + "step": 1364 + }, + { + "epoch": 0.741042345276873, + "grad_norm": 3.596935443077127, + "learning_rate": 1.927685109291604e-05, + "loss": 0.632, + "step": 1365 + }, + { + "epoch": 0.741585233441911, + "grad_norm": 4.535193817325199, + "learning_rate": 1.9275537459616364e-05, + "loss": 1.332, + "step": 1366 + }, + { + "epoch": 0.742128121606949, + "grad_norm": 6.678844775309246, + "learning_rate": 1.9274222679103437e-05, + "loss": 1.6826, + "step": 1367 + }, + { + "epoch": 0.742671009771987, + "grad_norm": 4.444554760823683, + "learning_rate": 1.927290675153987e-05, + "loss": 0.9597, + "step": 1368 + }, + { + "epoch": 0.743213897937025, + "grad_norm": 5.19917061501543, + "learning_rate": 1.927158967708841e-05, + "loss": 1.4333, + "step": 1369 + }, + { + "epoch": 0.743756786102063, + "grad_norm": 4.742331177279521, + "learning_rate": 1.927027145591197e-05, + "loss": 1.1659, + "step": 1370 + }, + { + "epoch": 0.744299674267101, + "grad_norm": 3.713496752478956, + "learning_rate": 1.926895208817358e-05, + "loss": 0.8688, + "step": 1371 + }, + { + "epoch": 0.744842562432139, + "grad_norm": 4.424173231823696, + "learning_rate": 1.9267631574036417e-05, + "loss": 1.0751, + "step": 1372 + }, + { + "epoch": 0.745385450597177, + "grad_norm": 4.7028219143924055, + "learning_rate": 1.9266309913663815e-05, + "loss": 1.3478, + "step": 1373 + }, + { + "epoch": 0.745928338762215, + "grad_norm": 5.498103807374834, + "learning_rate": 1.9264987107219237e-05, + "loss": 1.168, + "step": 1374 + }, + { + "epoch": 0.746471226927253, + "grad_norm": 4.326699254091958, + "learning_rate": 1.9263663154866285e-05, + "loss": 1.0713, + "step": 1375 + }, + { + "epoch": 0.747014115092291, + "grad_norm": 5.58697502498226, + "learning_rate": 1.926233805676871e-05, + "loss": 1.2474, + "step": 1376 + }, + { + "epoch": 0.747557003257329, + "grad_norm": 6.475832093356038, + "learning_rate": 1.92610118130904e-05, + "loss": 1.2805, + "step": 1377 + }, + { + "epoch": 0.748099891422367, + "grad_norm": 5.194166453296571, + "learning_rate": 1.925968442399539e-05, + "loss": 0.9736, + "step": 1378 + }, + { + "epoch": 0.748642779587405, + "grad_norm": 3.867607548943643, + "learning_rate": 1.9258355889647855e-05, + "loss": 1.2325, + "step": 1379 + }, + { + "epoch": 0.749185667752443, + "grad_norm": 5.602424635230674, + "learning_rate": 1.925702621021211e-05, + "loss": 1.0153, + "step": 1380 + }, + { + "epoch": 0.749728555917481, + "grad_norm": 5.011820196133583, + "learning_rate": 1.9255695385852604e-05, + "loss": 1.1823, + "step": 1381 + }, + { + "epoch": 0.750271444082519, + "grad_norm": 5.238515717778492, + "learning_rate": 1.9254363416733944e-05, + "loss": 1.4279, + "step": 1382 + }, + { + "epoch": 0.750814332247557, + "grad_norm": 5.600089025591521, + "learning_rate": 1.925303030302087e-05, + "loss": 0.9212, + "step": 1383 + }, + { + "epoch": 0.751357220412595, + "grad_norm": 4.318444011047328, + "learning_rate": 1.9251696044878255e-05, + "loss": 0.9623, + "step": 1384 + }, + { + "epoch": 0.751900108577633, + "grad_norm": 5.471464995723797, + "learning_rate": 1.925036064247113e-05, + "loss": 1.5446, + "step": 1385 + }, + { + "epoch": 0.752442996742671, + "grad_norm": 5.153911794581426, + "learning_rate": 1.9249024095964663e-05, + "loss": 1.7223, + "step": 1386 + }, + { + "epoch": 0.752985884907709, + "grad_norm": 3.7202924563691537, + "learning_rate": 1.924768640552415e-05, + "loss": 1.1469, + "step": 1387 + }, + { + "epoch": 0.753528773072747, + "grad_norm": 4.612135079041308, + "learning_rate": 1.9246347571315043e-05, + "loss": 0.9572, + "step": 1388 + }, + { + "epoch": 0.754071661237785, + "grad_norm": 4.618898432554816, + "learning_rate": 1.9245007593502937e-05, + "loss": 1.0443, + "step": 1389 + }, + { + "epoch": 0.754614549402823, + "grad_norm": 5.2787203720857345, + "learning_rate": 1.9243666472253554e-05, + "loss": 1.2034, + "step": 1390 + }, + { + "epoch": 0.755157437567861, + "grad_norm": 4.510719578546688, + "learning_rate": 1.9242324207732766e-05, + "loss": 0.8055, + "step": 1391 + }, + { + "epoch": 0.755700325732899, + "grad_norm": 6.256234902874701, + "learning_rate": 1.9240980800106596e-05, + "loss": 1.5985, + "step": 1392 + }, + { + "epoch": 0.756243213897937, + "grad_norm": 9.01213198061367, + "learning_rate": 1.923963624954119e-05, + "loss": 1.7921, + "step": 1393 + }, + { + "epoch": 0.756786102062975, + "grad_norm": 5.149498439254725, + "learning_rate": 1.923829055620285e-05, + "loss": 1.2406, + "step": 1394 + }, + { + "epoch": 0.757328990228013, + "grad_norm": 5.083183084254609, + "learning_rate": 1.9236943720258007e-05, + "loss": 0.9887, + "step": 1395 + }, + { + "epoch": 0.757871878393051, + "grad_norm": 4.5118989088500685, + "learning_rate": 1.9235595741873247e-05, + "loss": 1.3528, + "step": 1396 + }, + { + "epoch": 0.758414766558089, + "grad_norm": 4.583236692880461, + "learning_rate": 1.923424662121528e-05, + "loss": 1.0185, + "step": 1397 + }, + { + "epoch": 0.758957654723127, + "grad_norm": 5.3602373686113625, + "learning_rate": 1.9232896358450976e-05, + "loss": 1.1827, + "step": 1398 + }, + { + "epoch": 0.759500542888165, + "grad_norm": 4.510149132944334, + "learning_rate": 1.9231544953747336e-05, + "loss": 0.9981, + "step": 1399 + }, + { + "epoch": 0.760043431053203, + "grad_norm": 5.008678356958532, + "learning_rate": 1.9230192407271506e-05, + "loss": 1.4957, + "step": 1400 + }, + { + "epoch": 0.760586319218241, + "grad_norm": 4.6938647576746995, + "learning_rate": 1.9228838719190765e-05, + "loss": 1.361, + "step": 1401 + }, + { + "epoch": 0.761129207383279, + "grad_norm": 5.263462060803471, + "learning_rate": 1.9227483889672544e-05, + "loss": 1.1716, + "step": 1402 + }, + { + "epoch": 0.761672095548317, + "grad_norm": 4.737410394333335, + "learning_rate": 1.9226127918884407e-05, + "loss": 1.3924, + "step": 1403 + }, + { + "epoch": 0.762214983713355, + "grad_norm": 5.117476933198257, + "learning_rate": 1.9224770806994066e-05, + "loss": 1.1215, + "step": 1404 + }, + { + "epoch": 0.7627578718783931, + "grad_norm": 4.525492833460315, + "learning_rate": 1.922341255416937e-05, + "loss": 1.0626, + "step": 1405 + }, + { + "epoch": 0.7633007600434311, + "grad_norm": 4.928417801176309, + "learning_rate": 1.9222053160578312e-05, + "loss": 0.9576, + "step": 1406 + }, + { + "epoch": 0.7638436482084691, + "grad_norm": 4.44076437857558, + "learning_rate": 1.9220692626389018e-05, + "loss": 0.9186, + "step": 1407 + }, + { + "epoch": 0.7643865363735071, + "grad_norm": 4.779705057681976, + "learning_rate": 1.9219330951769763e-05, + "loss": 1.3392, + "step": 1408 + }, + { + "epoch": 0.7649294245385451, + "grad_norm": 5.37856562129718, + "learning_rate": 1.9217968136888965e-05, + "loss": 1.043, + "step": 1409 + }, + { + "epoch": 0.7654723127035831, + "grad_norm": 5.596447438931628, + "learning_rate": 1.9216604181915178e-05, + "loss": 1.3223, + "step": 1410 + }, + { + "epoch": 0.7660152008686211, + "grad_norm": 5.267245970848837, + "learning_rate": 1.9215239087017093e-05, + "loss": 1.0484, + "step": 1411 + }, + { + "epoch": 0.7665580890336591, + "grad_norm": 6.124227707475327, + "learning_rate": 1.9213872852363552e-05, + "loss": 1.5361, + "step": 1412 + }, + { + "epoch": 0.7671009771986971, + "grad_norm": 5.632804630433347, + "learning_rate": 1.9212505478123532e-05, + "loss": 1.2227, + "step": 1413 + }, + { + "epoch": 0.7676438653637351, + "grad_norm": 4.903635376409972, + "learning_rate": 1.9211136964466152e-05, + "loss": 0.9045, + "step": 1414 + }, + { + "epoch": 0.7681867535287731, + "grad_norm": 4.995962297168909, + "learning_rate": 1.9209767311560673e-05, + "loss": 1.2364, + "step": 1415 + }, + { + "epoch": 0.7687296416938111, + "grad_norm": 4.76685077713632, + "learning_rate": 1.9208396519576494e-05, + "loss": 1.4849, + "step": 1416 + }, + { + "epoch": 0.7692725298588491, + "grad_norm": 5.66289239913894, + "learning_rate": 1.9207024588683158e-05, + "loss": 1.389, + "step": 1417 + }, + { + "epoch": 0.7698154180238871, + "grad_norm": 4.044195406366437, + "learning_rate": 1.920565151905035e-05, + "loss": 0.5736, + "step": 1418 + }, + { + "epoch": 0.7703583061889251, + "grad_norm": 5.610527750585898, + "learning_rate": 1.9204277310847887e-05, + "loss": 1.5147, + "step": 1419 + }, + { + "epoch": 0.7709011943539631, + "grad_norm": 3.8629108074125424, + "learning_rate": 1.9202901964245734e-05, + "loss": 0.9184, + "step": 1420 + }, + { + "epoch": 0.7714440825190011, + "grad_norm": 4.582445031278247, + "learning_rate": 1.9201525479414e-05, + "loss": 1.071, + "step": 1421 + }, + { + "epoch": 0.7719869706840391, + "grad_norm": 5.016846104390101, + "learning_rate": 1.9200147856522933e-05, + "loss": 1.3673, + "step": 1422 + }, + { + "epoch": 0.7725298588490771, + "grad_norm": 4.798189213061551, + "learning_rate": 1.9198769095742914e-05, + "loss": 1.3483, + "step": 1423 + }, + { + "epoch": 0.7730727470141151, + "grad_norm": 5.871902023790772, + "learning_rate": 1.9197389197244473e-05, + "loss": 1.7625, + "step": 1424 + }, + { + "epoch": 0.7736156351791531, + "grad_norm": 5.895934775040147, + "learning_rate": 1.9196008161198277e-05, + "loss": 0.999, + "step": 1425 + }, + { + "epoch": 0.7741585233441911, + "grad_norm": 5.0199402476408155, + "learning_rate": 1.9194625987775138e-05, + "loss": 1.3251, + "step": 1426 + }, + { + "epoch": 0.7747014115092291, + "grad_norm": 6.650397672217608, + "learning_rate": 1.9193242677146e-05, + "loss": 1.2162, + "step": 1427 + }, + { + "epoch": 0.7752442996742671, + "grad_norm": 5.515267491505962, + "learning_rate": 1.9191858229481958e-05, + "loss": 1.3849, + "step": 1428 + }, + { + "epoch": 0.7757871878393051, + "grad_norm": 5.694611687374825, + "learning_rate": 1.9190472644954236e-05, + "loss": 1.0831, + "step": 1429 + }, + { + "epoch": 0.7763300760043431, + "grad_norm": 4.548114219835821, + "learning_rate": 1.9189085923734215e-05, + "loss": 1.2549, + "step": 1430 + }, + { + "epoch": 0.7768729641693811, + "grad_norm": 6.119997613777156, + "learning_rate": 1.9187698065993398e-05, + "loss": 1.6137, + "step": 1431 + }, + { + "epoch": 0.7774158523344191, + "grad_norm": 4.574150272616086, + "learning_rate": 1.9186309071903445e-05, + "loss": 1.3015, + "step": 1432 + }, + { + "epoch": 0.7779587404994571, + "grad_norm": 4.77854731853541, + "learning_rate": 1.9184918941636142e-05, + "loss": 0.6973, + "step": 1433 + }, + { + "epoch": 0.7785016286644951, + "grad_norm": 6.926122760031406, + "learning_rate": 1.9183527675363425e-05, + "loss": 1.4034, + "step": 1434 + }, + { + "epoch": 0.7790445168295331, + "grad_norm": 4.6748890605309645, + "learning_rate": 1.9182135273257372e-05, + "loss": 0.9854, + "step": 1435 + }, + { + "epoch": 0.7795874049945711, + "grad_norm": 5.072338861625223, + "learning_rate": 1.9180741735490194e-05, + "loss": 1.1604, + "step": 1436 + }, + { + "epoch": 0.7801302931596091, + "grad_norm": 5.995183838581222, + "learning_rate": 1.9179347062234245e-05, + "loss": 1.3913, + "step": 1437 + }, + { + "epoch": 0.7806731813246471, + "grad_norm": 5.494006248546126, + "learning_rate": 1.917795125366202e-05, + "loss": 1.2541, + "step": 1438 + }, + { + "epoch": 0.7812160694896851, + "grad_norm": 3.8223879163574694, + "learning_rate": 1.917655430994616e-05, + "loss": 0.7292, + "step": 1439 + }, + { + "epoch": 0.7817589576547231, + "grad_norm": 4.1476284953657405, + "learning_rate": 1.9175156231259434e-05, + "loss": 0.608, + "step": 1440 + }, + { + "epoch": 0.7823018458197611, + "grad_norm": 5.5144637244676495, + "learning_rate": 1.9173757017774764e-05, + "loss": 1.2674, + "step": 1441 + }, + { + "epoch": 0.7828447339847991, + "grad_norm": 5.4660654663594945, + "learning_rate": 1.9172356669665206e-05, + "loss": 1.3043, + "step": 1442 + }, + { + "epoch": 0.7833876221498371, + "grad_norm": 3.526312966998694, + "learning_rate": 1.9170955187103957e-05, + "loss": 0.8721, + "step": 1443 + }, + { + "epoch": 0.7839305103148752, + "grad_norm": 4.395429767339641, + "learning_rate": 1.9169552570264355e-05, + "loss": 0.985, + "step": 1444 + }, + { + "epoch": 0.7844733984799132, + "grad_norm": 4.608635000415997, + "learning_rate": 1.9168148819319874e-05, + "loss": 1.3492, + "step": 1445 + }, + { + "epoch": 0.7850162866449512, + "grad_norm": 5.729965497961937, + "learning_rate": 1.9166743934444137e-05, + "loss": 1.3405, + "step": 1446 + }, + { + "epoch": 0.7855591748099892, + "grad_norm": 6.60908414105679, + "learning_rate": 1.91653379158109e-05, + "loss": 1.5271, + "step": 1447 + }, + { + "epoch": 0.7861020629750272, + "grad_norm": 4.91386728282805, + "learning_rate": 1.916393076359406e-05, + "loss": 1.475, + "step": 1448 + }, + { + "epoch": 0.7866449511400652, + "grad_norm": 5.286158051966196, + "learning_rate": 1.916252247796766e-05, + "loss": 1.5235, + "step": 1449 + }, + { + "epoch": 0.7871878393051032, + "grad_norm": 4.699459928019414, + "learning_rate": 1.916111305910588e-05, + "loss": 1.151, + "step": 1450 + }, + { + "epoch": 0.7877307274701412, + "grad_norm": 5.125659016648778, + "learning_rate": 1.915970250718303e-05, + "loss": 1.2952, + "step": 1451 + }, + { + "epoch": 0.7882736156351792, + "grad_norm": 4.976170683421487, + "learning_rate": 1.915829082237358e-05, + "loss": 1.3291, + "step": 1452 + }, + { + "epoch": 0.7888165038002172, + "grad_norm": 5.237853683905863, + "learning_rate": 1.9156878004852123e-05, + "loss": 1.4775, + "step": 1453 + }, + { + "epoch": 0.7893593919652552, + "grad_norm": 4.914941394388547, + "learning_rate": 1.9155464054793404e-05, + "loss": 1.2151, + "step": 1454 + }, + { + "epoch": 0.7899022801302932, + "grad_norm": 5.050785426148085, + "learning_rate": 1.9154048972372293e-05, + "loss": 1.163, + "step": 1455 + }, + { + "epoch": 0.7904451682953312, + "grad_norm": 4.2016259300832255, + "learning_rate": 1.915263275776382e-05, + "loss": 0.9601, + "step": 1456 + }, + { + "epoch": 0.7909880564603692, + "grad_norm": 5.48804064939896, + "learning_rate": 1.915121541114314e-05, + "loss": 1.3026, + "step": 1457 + }, + { + "epoch": 0.7915309446254072, + "grad_norm": 5.230495684608947, + "learning_rate": 1.9149796932685552e-05, + "loss": 1.1923, + "step": 1458 + }, + { + "epoch": 0.7920738327904452, + "grad_norm": 4.788655104859546, + "learning_rate": 1.91483773225665e-05, + "loss": 1.2437, + "step": 1459 + }, + { + "epoch": 0.7926167209554832, + "grad_norm": 4.506216689801701, + "learning_rate": 1.9146956580961556e-05, + "loss": 0.9364, + "step": 1460 + }, + { + "epoch": 0.7931596091205212, + "grad_norm": 4.2335852812311865, + "learning_rate": 1.9145534708046446e-05, + "loss": 0.7104, + "step": 1461 + }, + { + "epoch": 0.7937024972855592, + "grad_norm": 6.204688311211956, + "learning_rate": 1.914411170399703e-05, + "loss": 1.0825, + "step": 1462 + }, + { + "epoch": 0.7942453854505972, + "grad_norm": 3.445489329210515, + "learning_rate": 1.91426875689893e-05, + "loss": 0.9921, + "step": 1463 + }, + { + "epoch": 0.7947882736156352, + "grad_norm": 5.204416925095863, + "learning_rate": 1.9141262303199403e-05, + "loss": 1.3043, + "step": 1464 + }, + { + "epoch": 0.7953311617806732, + "grad_norm": 6.3486214559668985, + "learning_rate": 1.9139835906803612e-05, + "loss": 1.3193, + "step": 1465 + }, + { + "epoch": 0.7958740499457112, + "grad_norm": 5.610159156463615, + "learning_rate": 1.913840837997835e-05, + "loss": 1.0455, + "step": 1466 + }, + { + "epoch": 0.7964169381107492, + "grad_norm": 4.625080334899242, + "learning_rate": 1.913697972290018e-05, + "loss": 0.7981, + "step": 1467 + }, + { + "epoch": 0.7969598262757872, + "grad_norm": 5.993005746484773, + "learning_rate": 1.9135549935745792e-05, + "loss": 1.1674, + "step": 1468 + }, + { + "epoch": 0.7975027144408252, + "grad_norm": 5.084136125969368, + "learning_rate": 1.913411901869203e-05, + "loss": 1.0362, + "step": 1469 + }, + { + "epoch": 0.7980456026058632, + "grad_norm": 5.701736254232889, + "learning_rate": 1.913268697191587e-05, + "loss": 1.2159, + "step": 1470 + }, + { + "epoch": 0.7985884907709012, + "grad_norm": 4.715190322082246, + "learning_rate": 1.9131253795594428e-05, + "loss": 0.9848, + "step": 1471 + }, + { + "epoch": 0.7991313789359392, + "grad_norm": 5.375233653105075, + "learning_rate": 1.9129819489904964e-05, + "loss": 1.0476, + "step": 1472 + }, + { + "epoch": 0.7996742671009772, + "grad_norm": 6.352674291337691, + "learning_rate": 1.9128384055024874e-05, + "loss": 1.2362, + "step": 1473 + }, + { + "epoch": 0.8002171552660152, + "grad_norm": 5.46716259791096, + "learning_rate": 1.91269474911317e-05, + "loss": 1.2949, + "step": 1474 + }, + { + "epoch": 0.8007600434310532, + "grad_norm": 5.3150808603597826, + "learning_rate": 1.912550979840311e-05, + "loss": 1.1587, + "step": 1475 + }, + { + "epoch": 0.8013029315960912, + "grad_norm": 4.46814887402293, + "learning_rate": 1.9124070977016926e-05, + "loss": 0.9649, + "step": 1476 + }, + { + "epoch": 0.8018458197611292, + "grad_norm": 4.916267658604107, + "learning_rate": 1.9122631027151103e-05, + "loss": 1.2117, + "step": 1477 + }, + { + "epoch": 0.8023887079261672, + "grad_norm": 5.506988713852874, + "learning_rate": 1.9121189948983733e-05, + "loss": 1.3387, + "step": 1478 + }, + { + "epoch": 0.8029315960912052, + "grad_norm": 4.7704152930487895, + "learning_rate": 1.911974774269305e-05, + "loss": 1.3379, + "step": 1479 + }, + { + "epoch": 0.8034744842562432, + "grad_norm": 5.667769720352476, + "learning_rate": 1.9118304408457435e-05, + "loss": 1.0552, + "step": 1480 + }, + { + "epoch": 0.8040173724212812, + "grad_norm": 4.9471232370904925, + "learning_rate": 1.91168599464554e-05, + "loss": 1.0013, + "step": 1481 + }, + { + "epoch": 0.8045602605863192, + "grad_norm": 4.856623754785127, + "learning_rate": 1.9115414356865594e-05, + "loss": 0.9001, + "step": 1482 + }, + { + "epoch": 0.8051031487513572, + "grad_norm": 4.004976425614515, + "learning_rate": 1.9113967639866815e-05, + "loss": 0.8114, + "step": 1483 + }, + { + "epoch": 0.8056460369163952, + "grad_norm": 5.983237168687411, + "learning_rate": 1.911251979563799e-05, + "loss": 1.1678, + "step": 1484 + }, + { + "epoch": 0.8061889250814332, + "grad_norm": 4.808701495369871, + "learning_rate": 1.9111070824358196e-05, + "loss": 0.9181, + "step": 1485 + }, + { + "epoch": 0.8067318132464713, + "grad_norm": 5.080557830592386, + "learning_rate": 1.910962072620664e-05, + "loss": 1.0751, + "step": 1486 + }, + { + "epoch": 0.8072747014115093, + "grad_norm": 5.730412968009966, + "learning_rate": 1.9108169501362674e-05, + "loss": 1.2727, + "step": 1487 + }, + { + "epoch": 0.8078175895765473, + "grad_norm": 6.47838124684283, + "learning_rate": 1.9106717150005785e-05, + "loss": 1.6491, + "step": 1488 + }, + { + "epoch": 0.8083604777415853, + "grad_norm": 5.311235841870027, + "learning_rate": 1.910526367231561e-05, + "loss": 0.8382, + "step": 1489 + }, + { + "epoch": 0.8089033659066233, + "grad_norm": 5.43072146168114, + "learning_rate": 1.9103809068471914e-05, + "loss": 1.3026, + "step": 1490 + }, + { + "epoch": 0.8094462540716613, + "grad_norm": 4.46156225654375, + "learning_rate": 1.9102353338654597e-05, + "loss": 0.8071, + "step": 1491 + }, + { + "epoch": 0.8099891422366993, + "grad_norm": 5.297520452299967, + "learning_rate": 1.9100896483043714e-05, + "loss": 1.0625, + "step": 1492 + }, + { + "epoch": 0.8105320304017373, + "grad_norm": 6.736735698357962, + "learning_rate": 1.909943850181945e-05, + "loss": 1.6497, + "step": 1493 + }, + { + "epoch": 0.8110749185667753, + "grad_norm": 4.8675910913183955, + "learning_rate": 1.9097979395162132e-05, + "loss": 1.0822, + "step": 1494 + }, + { + "epoch": 0.8116178067318133, + "grad_norm": 4.5489525578867305, + "learning_rate": 1.909651916325222e-05, + "loss": 0.7908, + "step": 1495 + }, + { + "epoch": 0.8121606948968513, + "grad_norm": 6.002121574887706, + "learning_rate": 1.909505780627032e-05, + "loss": 0.5178, + "step": 1496 + }, + { + "epoch": 0.8127035830618893, + "grad_norm": 4.34564271493731, + "learning_rate": 1.9093595324397175e-05, + "loss": 0.9818, + "step": 1497 + }, + { + "epoch": 0.8132464712269273, + "grad_norm": 5.0382562486474525, + "learning_rate": 1.9092131717813668e-05, + "loss": 1.0997, + "step": 1498 + }, + { + "epoch": 0.8137893593919653, + "grad_norm": 6.8289191133749, + "learning_rate": 1.909066698670082e-05, + "loss": 1.0046, + "step": 1499 + }, + { + "epoch": 0.8143322475570033, + "grad_norm": 6.6968746410007585, + "learning_rate": 1.908920113123979e-05, + "loss": 1.5191, + "step": 1500 + }, + { + "epoch": 0.8148751357220413, + "grad_norm": 5.548259576405749, + "learning_rate": 1.9087734151611877e-05, + "loss": 1.3272, + "step": 1501 + }, + { + "epoch": 0.8154180238870793, + "grad_norm": 5.684602861026239, + "learning_rate": 1.9086266047998522e-05, + "loss": 1.0698, + "step": 1502 + }, + { + "epoch": 0.8159609120521173, + "grad_norm": 6.832627276636624, + "learning_rate": 1.90847968205813e-05, + "loss": 1.5796, + "step": 1503 + }, + { + "epoch": 0.8165038002171553, + "grad_norm": 5.605724679057613, + "learning_rate": 1.908332646954193e-05, + "loss": 1.1709, + "step": 1504 + }, + { + "epoch": 0.8170466883821933, + "grad_norm": 4.607345782065147, + "learning_rate": 1.908185499506226e-05, + "loss": 0.8686, + "step": 1505 + }, + { + "epoch": 0.8175895765472313, + "grad_norm": 4.687252800873217, + "learning_rate": 1.9080382397324296e-05, + "loss": 0.8881, + "step": 1506 + }, + { + "epoch": 0.8181324647122693, + "grad_norm": 6.357714385620256, + "learning_rate": 1.907890867651016e-05, + "loss": 1.1948, + "step": 1507 + }, + { + "epoch": 0.8186753528773073, + "grad_norm": 4.8608334992124425, + "learning_rate": 1.9077433832802135e-05, + "loss": 1.0311, + "step": 1508 + }, + { + "epoch": 0.8192182410423453, + "grad_norm": 5.948864135911491, + "learning_rate": 1.9075957866382623e-05, + "loss": 1.1314, + "step": 1509 + }, + { + "epoch": 0.8197611292073833, + "grad_norm": 4.807850657807276, + "learning_rate": 1.9074480777434178e-05, + "loss": 0.9478, + "step": 1510 + }, + { + "epoch": 0.8203040173724213, + "grad_norm": 4.26222909817726, + "learning_rate": 1.9073002566139486e-05, + "loss": 0.8541, + "step": 1511 + }, + { + "epoch": 0.8208469055374593, + "grad_norm": 6.6130092616391005, + "learning_rate": 1.9071523232681382e-05, + "loss": 1.2754, + "step": 1512 + }, + { + "epoch": 0.8213897937024973, + "grad_norm": 5.255155673043404, + "learning_rate": 1.907004277724282e-05, + "loss": 1.0813, + "step": 1513 + }, + { + "epoch": 0.8219326818675353, + "grad_norm": 5.079098182105948, + "learning_rate": 1.9068561200006917e-05, + "loss": 1.0016, + "step": 1514 + }, + { + "epoch": 0.8224755700325733, + "grad_norm": 5.491172879434626, + "learning_rate": 1.906707850115691e-05, + "loss": 1.2884, + "step": 1515 + }, + { + "epoch": 0.8230184581976113, + "grad_norm": 6.718962166599785, + "learning_rate": 1.9065594680876182e-05, + "loss": 1.4973, + "step": 1516 + }, + { + "epoch": 0.8235613463626493, + "grad_norm": 5.348428262646105, + "learning_rate": 1.9064109739348257e-05, + "loss": 1.1113, + "step": 1517 + }, + { + "epoch": 0.8241042345276873, + "grad_norm": 4.604059650726469, + "learning_rate": 1.906262367675679e-05, + "loss": 1.1614, + "step": 1518 + }, + { + "epoch": 0.8246471226927253, + "grad_norm": 5.895860662978225, + "learning_rate": 1.9061136493285586e-05, + "loss": 1.1532, + "step": 1519 + }, + { + "epoch": 0.8251900108577633, + "grad_norm": 5.621847434524929, + "learning_rate": 1.905964818911858e-05, + "loss": 0.9277, + "step": 1520 + }, + { + "epoch": 0.8257328990228013, + "grad_norm": 4.585955492662189, + "learning_rate": 1.9058158764439844e-05, + "loss": 0.7988, + "step": 1521 + }, + { + "epoch": 0.8262757871878393, + "grad_norm": 5.784308925988881, + "learning_rate": 1.9056668219433595e-05, + "loss": 1.6078, + "step": 1522 + }, + { + "epoch": 0.8268186753528773, + "grad_norm": 5.621216253388429, + "learning_rate": 1.905517655428419e-05, + "loss": 1.389, + "step": 1523 + }, + { + "epoch": 0.8273615635179153, + "grad_norm": 5.514208208061458, + "learning_rate": 1.9053683769176115e-05, + "loss": 0.7612, + "step": 1524 + }, + { + "epoch": 0.8279044516829533, + "grad_norm": 4.6572185133859065, + "learning_rate": 1.9052189864294002e-05, + "loss": 0.744, + "step": 1525 + }, + { + "epoch": 0.8284473398479913, + "grad_norm": 5.239632270223703, + "learning_rate": 1.905069483982262e-05, + "loss": 0.939, + "step": 1526 + }, + { + "epoch": 0.8289902280130294, + "grad_norm": 3.7037650901217454, + "learning_rate": 1.9049198695946876e-05, + "loss": 1.0177, + "step": 1527 + }, + { + "epoch": 0.8295331161780674, + "grad_norm": 6.974869726679236, + "learning_rate": 1.9047701432851813e-05, + "loss": 1.2722, + "step": 1528 + }, + { + "epoch": 0.8300760043431054, + "grad_norm": 6.642878483620589, + "learning_rate": 1.904620305072262e-05, + "loss": 1.5369, + "step": 1529 + }, + { + "epoch": 0.8306188925081434, + "grad_norm": 7.99082645392899, + "learning_rate": 1.9044703549744616e-05, + "loss": 1.2245, + "step": 1530 + }, + { + "epoch": 0.8311617806731814, + "grad_norm": 6.5593948883008135, + "learning_rate": 1.904320293010326e-05, + "loss": 1.307, + "step": 1531 + }, + { + "epoch": 0.8317046688382194, + "grad_norm": 4.930764759519961, + "learning_rate": 1.9041701191984155e-05, + "loss": 0.9564, + "step": 1532 + }, + { + "epoch": 0.8322475570032574, + "grad_norm": 3.9621684331427773, + "learning_rate": 1.9040198335573033e-05, + "loss": 0.8153, + "step": 1533 + }, + { + "epoch": 0.8327904451682954, + "grad_norm": 4.859084711241092, + "learning_rate": 1.9038694361055774e-05, + "loss": 1.0967, + "step": 1534 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.9476278172327595, + "learning_rate": 1.903718926861839e-05, + "loss": 1.4743, + "step": 1535 + }, + { + "epoch": 0.8338762214983714, + "grad_norm": 4.312795088472635, + "learning_rate": 1.903568305844704e-05, + "loss": 0.8249, + "step": 1536 + }, + { + "epoch": 0.8344191096634094, + "grad_norm": 6.638027572281857, + "learning_rate": 1.9034175730728e-05, + "loss": 1.6353, + "step": 1537 + }, + { + "epoch": 0.8349619978284474, + "grad_norm": 5.709548912096025, + "learning_rate": 1.9032667285647714e-05, + "loss": 1.8512, + "step": 1538 + }, + { + "epoch": 0.8355048859934854, + "grad_norm": 5.098275426179771, + "learning_rate": 1.9031157723392738e-05, + "loss": 1.0348, + "step": 1539 + }, + { + "epoch": 0.8360477741585234, + "grad_norm": 5.05236269448673, + "learning_rate": 1.9029647044149783e-05, + "loss": 1.108, + "step": 1540 + }, + { + "epoch": 0.8365906623235614, + "grad_norm": 4.674708875536442, + "learning_rate": 1.9028135248105692e-05, + "loss": 1.0453, + "step": 1541 + }, + { + "epoch": 0.8371335504885994, + "grad_norm": 4.717321956389267, + "learning_rate": 1.902662233544744e-05, + "loss": 0.9042, + "step": 1542 + }, + { + "epoch": 0.8376764386536374, + "grad_norm": 6.217040367936212, + "learning_rate": 1.9025108306362158e-05, + "loss": 1.0762, + "step": 1543 + }, + { + "epoch": 0.8382193268186754, + "grad_norm": 5.338744288323046, + "learning_rate": 1.9023593161037094e-05, + "loss": 1.1631, + "step": 1544 + }, + { + "epoch": 0.8387622149837134, + "grad_norm": 4.8224874086687874, + "learning_rate": 1.9022076899659643e-05, + "loss": 1.3907, + "step": 1545 + }, + { + "epoch": 0.8393051031487514, + "grad_norm": 4.699800413179793, + "learning_rate": 1.9020559522417345e-05, + "loss": 0.7682, + "step": 1546 + }, + { + "epoch": 0.8398479913137894, + "grad_norm": 4.271439794022252, + "learning_rate": 1.9019041029497866e-05, + "loss": 0.8475, + "step": 1547 + }, + { + "epoch": 0.8403908794788274, + "grad_norm": 4.240562925498168, + "learning_rate": 1.9017521421089022e-05, + "loss": 0.8201, + "step": 1548 + }, + { + "epoch": 0.8409337676438654, + "grad_norm": 5.24718786534657, + "learning_rate": 1.9016000697378755e-05, + "loss": 1.0728, + "step": 1549 + }, + { + "epoch": 0.8414766558089034, + "grad_norm": 6.492220902503762, + "learning_rate": 1.9014478858555156e-05, + "loss": 0.9432, + "step": 1550 + }, + { + "epoch": 0.8420195439739414, + "grad_norm": 4.917811923935393, + "learning_rate": 1.9012955904806438e-05, + "loss": 1.1672, + "step": 1551 + }, + { + "epoch": 0.8425624321389794, + "grad_norm": 5.66709669173795, + "learning_rate": 1.9011431836320976e-05, + "loss": 1.5058, + "step": 1552 + }, + { + "epoch": 0.8431053203040174, + "grad_norm": 5.658634152413846, + "learning_rate": 1.9009906653287258e-05, + "loss": 1.3653, + "step": 1553 + }, + { + "epoch": 0.8436482084690554, + "grad_norm": 4.956250367213818, + "learning_rate": 1.9008380355893925e-05, + "loss": 0.8309, + "step": 1554 + }, + { + "epoch": 0.8441910966340934, + "grad_norm": 4.329497016180362, + "learning_rate": 1.9006852944329753e-05, + "loss": 1.1141, + "step": 1555 + }, + { + "epoch": 0.8447339847991314, + "grad_norm": 5.717864943326053, + "learning_rate": 1.9005324418783658e-05, + "loss": 1.3274, + "step": 1556 + }, + { + "epoch": 0.8452768729641694, + "grad_norm": 5.657300743895006, + "learning_rate": 1.900379477944468e-05, + "loss": 1.2432, + "step": 1557 + }, + { + "epoch": 0.8458197611292074, + "grad_norm": 5.325761071371062, + "learning_rate": 1.900226402650202e-05, + "loss": 1.3428, + "step": 1558 + }, + { + "epoch": 0.8463626492942454, + "grad_norm": 5.207304704126981, + "learning_rate": 1.9000732160144996e-05, + "loss": 1.455, + "step": 1559 + }, + { + "epoch": 0.8469055374592834, + "grad_norm": 5.248656371435267, + "learning_rate": 1.8999199180563074e-05, + "loss": 0.7851, + "step": 1560 + }, + { + "epoch": 0.8474484256243214, + "grad_norm": 6.92723307445887, + "learning_rate": 1.899766508794585e-05, + "loss": 1.5236, + "step": 1561 + }, + { + "epoch": 0.8479913137893594, + "grad_norm": 4.3434986726191935, + "learning_rate": 1.899612988248307e-05, + "loss": 0.9117, + "step": 1562 + }, + { + "epoch": 0.8485342019543974, + "grad_norm": 5.815800316925401, + "learning_rate": 1.8994593564364612e-05, + "loss": 1.0097, + "step": 1563 + }, + { + "epoch": 0.8490770901194354, + "grad_norm": 5.5494362406379905, + "learning_rate": 1.8993056133780484e-05, + "loss": 1.1829, + "step": 1564 + }, + { + "epoch": 0.8496199782844734, + "grad_norm": 4.323494913553237, + "learning_rate": 1.899151759092084e-05, + "loss": 0.9359, + "step": 1565 + }, + { + "epoch": 0.8501628664495114, + "grad_norm": 4.455911135073797, + "learning_rate": 1.898997793597597e-05, + "loss": 0.9354, + "step": 1566 + }, + { + "epoch": 0.8507057546145494, + "grad_norm": 6.684553981467422, + "learning_rate": 1.8988437169136302e-05, + "loss": 1.3693, + "step": 1567 + }, + { + "epoch": 0.8512486427795874, + "grad_norm": 6.341757024960091, + "learning_rate": 1.89868952905924e-05, + "loss": 0.9441, + "step": 1568 + }, + { + "epoch": 0.8517915309446255, + "grad_norm": 6.017605683279906, + "learning_rate": 1.8985352300534965e-05, + "loss": 1.2747, + "step": 1569 + }, + { + "epoch": 0.8523344191096635, + "grad_norm": 5.501152081837135, + "learning_rate": 1.8983808199154835e-05, + "loss": 0.8414, + "step": 1570 + }, + { + "epoch": 0.8528773072747015, + "grad_norm": 5.075164131477861, + "learning_rate": 1.8982262986642993e-05, + "loss": 1.124, + "step": 1571 + }, + { + "epoch": 0.8534201954397395, + "grad_norm": 5.7750932469486065, + "learning_rate": 1.8980716663190545e-05, + "loss": 1.3968, + "step": 1572 + }, + { + "epoch": 0.8539630836047775, + "grad_norm": 6.967872140444835, + "learning_rate": 1.897916922898875e-05, + "loss": 1.2395, + "step": 1573 + }, + { + "epoch": 0.8545059717698155, + "grad_norm": 5.037804725894728, + "learning_rate": 1.8977620684228994e-05, + "loss": 0.9741, + "step": 1574 + }, + { + "epoch": 0.8550488599348535, + "grad_norm": 4.586000206489282, + "learning_rate": 1.8976071029102802e-05, + "loss": 0.6757, + "step": 1575 + }, + { + "epoch": 0.8555917480998915, + "grad_norm": 4.686417605581938, + "learning_rate": 1.897452026380184e-05, + "loss": 0.8382, + "step": 1576 + }, + { + "epoch": 0.8561346362649295, + "grad_norm": 5.133601528061789, + "learning_rate": 1.8972968388517908e-05, + "loss": 1.2999, + "step": 1577 + }, + { + "epoch": 0.8566775244299675, + "grad_norm": 6.170796021401621, + "learning_rate": 1.8971415403442942e-05, + "loss": 1.1513, + "step": 1578 + }, + { + "epoch": 0.8572204125950055, + "grad_norm": 5.229050623246509, + "learning_rate": 1.8969861308769025e-05, + "loss": 0.9371, + "step": 1579 + }, + { + "epoch": 0.8577633007600435, + "grad_norm": 5.081229057143735, + "learning_rate": 1.8968306104688365e-05, + "loss": 0.9398, + "step": 1580 + }, + { + "epoch": 0.8583061889250815, + "grad_norm": 4.641876540865143, + "learning_rate": 1.896674979139331e-05, + "loss": 1.1401, + "step": 1581 + }, + { + "epoch": 0.8588490770901195, + "grad_norm": 4.149586485324207, + "learning_rate": 1.8965192369076356e-05, + "loss": 0.6768, + "step": 1582 + }, + { + "epoch": 0.8593919652551575, + "grad_norm": 6.422772444923944, + "learning_rate": 1.8963633837930114e-05, + "loss": 1.392, + "step": 1583 + }, + { + "epoch": 0.8599348534201955, + "grad_norm": 4.860635992388028, + "learning_rate": 1.8962074198147357e-05, + "loss": 0.9983, + "step": 1584 + }, + { + "epoch": 0.8604777415852335, + "grad_norm": 4.216877618358879, + "learning_rate": 1.8960513449920982e-05, + "loss": 0.6958, + "step": 1585 + }, + { + "epoch": 0.8610206297502715, + "grad_norm": 5.012340070254638, + "learning_rate": 1.8958951593444017e-05, + "loss": 0.9859, + "step": 1586 + }, + { + "epoch": 0.8615635179153095, + "grad_norm": 5.467321056155664, + "learning_rate": 1.8957388628909644e-05, + "loss": 1.3782, + "step": 1587 + }, + { + "epoch": 0.8621064060803475, + "grad_norm": 5.59589685839463, + "learning_rate": 1.8955824556511168e-05, + "loss": 0.8297, + "step": 1588 + }, + { + "epoch": 0.8626492942453855, + "grad_norm": 6.519624998298618, + "learning_rate": 1.895425937644204e-05, + "loss": 1.7264, + "step": 1589 + }, + { + "epoch": 0.8631921824104235, + "grad_norm": 6.6951270307269795, + "learning_rate": 1.8952693088895837e-05, + "loss": 1.2919, + "step": 1590 + }, + { + "epoch": 0.8637350705754615, + "grad_norm": 4.348741825803296, + "learning_rate": 1.895112569406629e-05, + "loss": 1.0062, + "step": 1591 + }, + { + "epoch": 0.8642779587404995, + "grad_norm": 5.0213162819553565, + "learning_rate": 1.8949557192147243e-05, + "loss": 1.2381, + "step": 1592 + }, + { + "epoch": 0.8648208469055375, + "grad_norm": 6.1637021381056885, + "learning_rate": 1.8947987583332705e-05, + "loss": 1.2561, + "step": 1593 + }, + { + "epoch": 0.8653637350705755, + "grad_norm": 6.638257635344733, + "learning_rate": 1.89464168678168e-05, + "loss": 0.9861, + "step": 1594 + }, + { + "epoch": 0.8659066232356135, + "grad_norm": 4.861237127098975, + "learning_rate": 1.89448450457938e-05, + "loss": 0.9407, + "step": 1595 + }, + { + "epoch": 0.8664495114006515, + "grad_norm": 6.4248430682281565, + "learning_rate": 1.894327211745811e-05, + "loss": 1.3079, + "step": 1596 + }, + { + "epoch": 0.8669923995656895, + "grad_norm": 5.712502208347676, + "learning_rate": 1.8941698083004265e-05, + "loss": 1.4422, + "step": 1597 + }, + { + "epoch": 0.8675352877307275, + "grad_norm": 5.634308307822415, + "learning_rate": 1.8940122942626957e-05, + "loss": 0.7869, + "step": 1598 + }, + { + "epoch": 0.8680781758957655, + "grad_norm": 7.839535783401759, + "learning_rate": 1.893854669652099e-05, + "loss": 1.3397, + "step": 1599 + }, + { + "epoch": 0.8686210640608035, + "grad_norm": 5.156814072355032, + "learning_rate": 1.8936969344881323e-05, + "loss": 1.2541, + "step": 1600 + }, + { + "epoch": 0.8691639522258415, + "grad_norm": 4.912293322601371, + "learning_rate": 1.8935390887903044e-05, + "loss": 0.8418, + "step": 1601 + }, + { + "epoch": 0.8697068403908795, + "grad_norm": 5.436641391280079, + "learning_rate": 1.8933811325781382e-05, + "loss": 1.4456, + "step": 1602 + }, + { + "epoch": 0.8702497285559175, + "grad_norm": 5.221596044730403, + "learning_rate": 1.8932230658711696e-05, + "loss": 0.9626, + "step": 1603 + }, + { + "epoch": 0.8707926167209555, + "grad_norm": 6.103653494209826, + "learning_rate": 1.8930648886889482e-05, + "loss": 1.2338, + "step": 1604 + }, + { + "epoch": 0.8713355048859935, + "grad_norm": 7.003639368124227, + "learning_rate": 1.8929066010510383e-05, + "loss": 1.2216, + "step": 1605 + }, + { + "epoch": 0.8718783930510315, + "grad_norm": 4.911905632269173, + "learning_rate": 1.8927482029770168e-05, + "loss": 0.9049, + "step": 1606 + }, + { + "epoch": 0.8724212812160695, + "grad_norm": 6.079178298711795, + "learning_rate": 1.8925896944864748e-05, + "loss": 1.6408, + "step": 1607 + }, + { + "epoch": 0.8729641693811075, + "grad_norm": 6.045568168649525, + "learning_rate": 1.892431075599017e-05, + "loss": 1.084, + "step": 1608 + }, + { + "epoch": 0.8735070575461455, + "grad_norm": 4.827633470123435, + "learning_rate": 1.892272346334261e-05, + "loss": 0.8121, + "step": 1609 + }, + { + "epoch": 0.8740499457111836, + "grad_norm": 5.314535878915851, + "learning_rate": 1.8921135067118396e-05, + "loss": 1.2657, + "step": 1610 + }, + { + "epoch": 0.8745928338762216, + "grad_norm": 4.786606110240068, + "learning_rate": 1.8919545567513976e-05, + "loss": 1.0534, + "step": 1611 + }, + { + "epoch": 0.8751357220412594, + "grad_norm": 4.075700855026623, + "learning_rate": 1.8917954964725948e-05, + "loss": 0.7757, + "step": 1612 + }, + { + "epoch": 0.8756786102062974, + "grad_norm": 6.286980562376595, + "learning_rate": 1.8916363258951033e-05, + "loss": 0.9782, + "step": 1613 + }, + { + "epoch": 0.8762214983713354, + "grad_norm": 6.745303990081763, + "learning_rate": 1.8914770450386102e-05, + "loss": 0.9488, + "step": 1614 + }, + { + "epoch": 0.8767643865363735, + "grad_norm": 6.2663007227218275, + "learning_rate": 1.8913176539228152e-05, + "loss": 1.9529, + "step": 1615 + }, + { + "epoch": 0.8773072747014115, + "grad_norm": 4.793153881320079, + "learning_rate": 1.8911581525674324e-05, + "loss": 1.0242, + "step": 1616 + }, + { + "epoch": 0.8778501628664495, + "grad_norm": 4.6844946413894, + "learning_rate": 1.890998540992189e-05, + "loss": 0.8565, + "step": 1617 + }, + { + "epoch": 0.8783930510314875, + "grad_norm": 3.993737542716379, + "learning_rate": 1.8908388192168256e-05, + "loss": 0.6107, + "step": 1618 + }, + { + "epoch": 0.8789359391965255, + "grad_norm": 6.099718972513945, + "learning_rate": 1.8906789872610977e-05, + "loss": 0.8043, + "step": 1619 + }, + { + "epoch": 0.8794788273615635, + "grad_norm": 7.113418049135383, + "learning_rate": 1.8905190451447726e-05, + "loss": 1.241, + "step": 1620 + }, + { + "epoch": 0.8800217155266015, + "grad_norm": 4.439254236407197, + "learning_rate": 1.8903589928876337e-05, + "loss": 1.0627, + "step": 1621 + }, + { + "epoch": 0.8805646036916395, + "grad_norm": 5.773948255627926, + "learning_rate": 1.8901988305094746e-05, + "loss": 1.3241, + "step": 1622 + }, + { + "epoch": 0.8811074918566775, + "grad_norm": 6.231870498005159, + "learning_rate": 1.890038558030106e-05, + "loss": 1.5241, + "step": 1623 + }, + { + "epoch": 0.8816503800217155, + "grad_norm": 5.888612961801449, + "learning_rate": 1.8898781754693495e-05, + "loss": 1.109, + "step": 1624 + }, + { + "epoch": 0.8821932681867535, + "grad_norm": 4.880160750019388, + "learning_rate": 1.8897176828470424e-05, + "loss": 1.0124, + "step": 1625 + }, + { + "epoch": 0.8827361563517915, + "grad_norm": 6.365487465081988, + "learning_rate": 1.889557080183034e-05, + "loss": 0.8999, + "step": 1626 + }, + { + "epoch": 0.8832790445168295, + "grad_norm": 4.461166628483842, + "learning_rate": 1.8893963674971883e-05, + "loss": 0.8762, + "step": 1627 + }, + { + "epoch": 0.8838219326818675, + "grad_norm": 5.675449614336963, + "learning_rate": 1.8892355448093825e-05, + "loss": 1.049, + "step": 1628 + }, + { + "epoch": 0.8843648208469055, + "grad_norm": 4.320721965444944, + "learning_rate": 1.8890746121395072e-05, + "loss": 1.1291, + "step": 1629 + }, + { + "epoch": 0.8849077090119435, + "grad_norm": 5.155835351166163, + "learning_rate": 1.8889135695074668e-05, + "loss": 1.1035, + "step": 1630 + }, + { + "epoch": 0.8854505971769815, + "grad_norm": 6.506050711845877, + "learning_rate": 1.8887524169331794e-05, + "loss": 2.1954, + "step": 1631 + }, + { + "epoch": 0.8859934853420195, + "grad_norm": 5.514989348863087, + "learning_rate": 1.8885911544365766e-05, + "loss": 1.0237, + "step": 1632 + }, + { + "epoch": 0.8865363735070575, + "grad_norm": 5.123449587076818, + "learning_rate": 1.8884297820376038e-05, + "loss": 0.9908, + "step": 1633 + }, + { + "epoch": 0.8870792616720955, + "grad_norm": 5.421671567850239, + "learning_rate": 1.8882682997562197e-05, + "loss": 1.7734, + "step": 1634 + }, + { + "epoch": 0.8876221498371335, + "grad_norm": 7.1012436118787186, + "learning_rate": 1.8881067076123963e-05, + "loss": 1.5488, + "step": 1635 + }, + { + "epoch": 0.8881650380021715, + "grad_norm": 5.328780061143714, + "learning_rate": 1.88794500562612e-05, + "loss": 1.3832, + "step": 1636 + }, + { + "epoch": 0.8887079261672095, + "grad_norm": 5.1503624666971595, + "learning_rate": 1.88778319381739e-05, + "loss": 0.8524, + "step": 1637 + }, + { + "epoch": 0.8892508143322475, + "grad_norm": 5.131553278195334, + "learning_rate": 1.88762127220622e-05, + "loss": 0.9769, + "step": 1638 + }, + { + "epoch": 0.8897937024972855, + "grad_norm": 3.1519686008526135, + "learning_rate": 1.8874592408126365e-05, + "loss": 0.472, + "step": 1639 + }, + { + "epoch": 0.8903365906623235, + "grad_norm": 4.164586140054025, + "learning_rate": 1.8872970996566794e-05, + "loss": 0.67, + "step": 1640 + }, + { + "epoch": 0.8908794788273615, + "grad_norm": 7.488659246842809, + "learning_rate": 1.8871348487584028e-05, + "loss": 1.3141, + "step": 1641 + }, + { + "epoch": 0.8914223669923995, + "grad_norm": 5.911298710361154, + "learning_rate": 1.8869724881378743e-05, + "loss": 1.084, + "step": 1642 + }, + { + "epoch": 0.8919652551574375, + "grad_norm": 5.977793625957034, + "learning_rate": 1.886810017815175e-05, + "loss": 0.9273, + "step": 1643 + }, + { + "epoch": 0.8925081433224755, + "grad_norm": 5.2773273690434515, + "learning_rate": 1.8866474378103993e-05, + "loss": 1.2594, + "step": 1644 + }, + { + "epoch": 0.8930510314875135, + "grad_norm": 8.049487951903771, + "learning_rate": 1.8864847481436554e-05, + "loss": 1.0104, + "step": 1645 + }, + { + "epoch": 0.8935939196525515, + "grad_norm": 5.428394606766542, + "learning_rate": 1.886321948835065e-05, + "loss": 0.827, + "step": 1646 + }, + { + "epoch": 0.8941368078175895, + "grad_norm": 5.687133062399138, + "learning_rate": 1.8861590399047635e-05, + "loss": 1.2759, + "step": 1647 + }, + { + "epoch": 0.8946796959826275, + "grad_norm": 5.037591954761082, + "learning_rate": 1.885996021372899e-05, + "loss": 0.8928, + "step": 1648 + }, + { + "epoch": 0.8952225841476655, + "grad_norm": 4.911533421640023, + "learning_rate": 1.8858328932596352e-05, + "loss": 0.705, + "step": 1649 + }, + { + "epoch": 0.8957654723127035, + "grad_norm": 6.054608860222289, + "learning_rate": 1.885669655585147e-05, + "loss": 1.296, + "step": 1650 + }, + { + "epoch": 0.8963083604777415, + "grad_norm": 6.221099773279408, + "learning_rate": 1.8855063083696244e-05, + "loss": 1.5253, + "step": 1651 + }, + { + "epoch": 0.8968512486427795, + "grad_norm": 5.0718329796444035, + "learning_rate": 1.8853428516332702e-05, + "loss": 1.1683, + "step": 1652 + }, + { + "epoch": 0.8973941368078175, + "grad_norm": 5.721985007773526, + "learning_rate": 1.8851792853963015e-05, + "loss": 1.1635, + "step": 1653 + }, + { + "epoch": 0.8979370249728555, + "grad_norm": 7.224045168649458, + "learning_rate": 1.8850156096789473e-05, + "loss": 1.0042, + "step": 1654 + }, + { + "epoch": 0.8984799131378935, + "grad_norm": 4.938883702947416, + "learning_rate": 1.8848518245014526e-05, + "loss": 0.7712, + "step": 1655 + }, + { + "epoch": 0.8990228013029316, + "grad_norm": 5.9261619850877, + "learning_rate": 1.8846879298840735e-05, + "loss": 1.4387, + "step": 1656 + }, + { + "epoch": 0.8995656894679696, + "grad_norm": 5.181216185736725, + "learning_rate": 1.8845239258470817e-05, + "loss": 0.9389, + "step": 1657 + }, + { + "epoch": 0.9001085776330076, + "grad_norm": 5.108249236166185, + "learning_rate": 1.8843598124107608e-05, + "loss": 1.0034, + "step": 1658 + }, + { + "epoch": 0.9006514657980456, + "grad_norm": 4.839799276438571, + "learning_rate": 1.8841955895954088e-05, + "loss": 0.8524, + "step": 1659 + }, + { + "epoch": 0.9011943539630836, + "grad_norm": 5.598018889964705, + "learning_rate": 1.8840312574213372e-05, + "loss": 1.0939, + "step": 1660 + }, + { + "epoch": 0.9017372421281216, + "grad_norm": 5.2496226470868725, + "learning_rate": 1.8838668159088707e-05, + "loss": 1.0812, + "step": 1661 + }, + { + "epoch": 0.9022801302931596, + "grad_norm": 5.2298567821168, + "learning_rate": 1.8837022650783477e-05, + "loss": 0.953, + "step": 1662 + }, + { + "epoch": 0.9028230184581976, + "grad_norm": 4.471282412199049, + "learning_rate": 1.88353760495012e-05, + "loss": 1.2652, + "step": 1663 + }, + { + "epoch": 0.9033659066232356, + "grad_norm": 5.892082884595572, + "learning_rate": 1.8833728355445534e-05, + "loss": 1.1165, + "step": 1664 + }, + { + "epoch": 0.9039087947882736, + "grad_norm": 5.49174731306936, + "learning_rate": 1.8832079568820268e-05, + "loss": 0.9262, + "step": 1665 + }, + { + "epoch": 0.9044516829533116, + "grad_norm": 5.55211658654164, + "learning_rate": 1.883042968982932e-05, + "loss": 1.3112, + "step": 1666 + }, + { + "epoch": 0.9049945711183496, + "grad_norm": 5.597392580982551, + "learning_rate": 1.8828778718676757e-05, + "loss": 0.8511, + "step": 1667 + }, + { + "epoch": 0.9055374592833876, + "grad_norm": 6.183127994440369, + "learning_rate": 1.8827126655566773e-05, + "loss": 1.2784, + "step": 1668 + }, + { + "epoch": 0.9060803474484256, + "grad_norm": 6.991796130593696, + "learning_rate": 1.882547350070369e-05, + "loss": 1.1703, + "step": 1669 + }, + { + "epoch": 0.9066232356134636, + "grad_norm": 5.7794859741652305, + "learning_rate": 1.8823819254291986e-05, + "loss": 1.0974, + "step": 1670 + }, + { + "epoch": 0.9071661237785016, + "grad_norm": 5.388338879165358, + "learning_rate": 1.8822163916536245e-05, + "loss": 1.2087, + "step": 1671 + }, + { + "epoch": 0.9077090119435396, + "grad_norm": 5.072420136637763, + "learning_rate": 1.8820507487641218e-05, + "loss": 0.9244, + "step": 1672 + }, + { + "epoch": 0.9082519001085776, + "grad_norm": 5.9571835605705115, + "learning_rate": 1.8818849967811762e-05, + "loss": 1.3299, + "step": 1673 + }, + { + "epoch": 0.9087947882736156, + "grad_norm": 5.706573186001289, + "learning_rate": 1.8817191357252892e-05, + "loss": 1.1059, + "step": 1674 + }, + { + "epoch": 0.9093376764386536, + "grad_norm": 5.4984127833712435, + "learning_rate": 1.8815531656169737e-05, + "loss": 1.1428, + "step": 1675 + }, + { + "epoch": 0.9098805646036916, + "grad_norm": 4.824306323799433, + "learning_rate": 1.8813870864767582e-05, + "loss": 0.7313, + "step": 1676 + }, + { + "epoch": 0.9104234527687296, + "grad_norm": 6.180183255961245, + "learning_rate": 1.8812208983251828e-05, + "loss": 1.0729, + "step": 1677 + }, + { + "epoch": 0.9109663409337676, + "grad_norm": 6.586598617269493, + "learning_rate": 1.8810546011828024e-05, + "loss": 1.5871, + "step": 1678 + }, + { + "epoch": 0.9115092290988056, + "grad_norm": 5.5611281281711875, + "learning_rate": 1.8808881950701845e-05, + "loss": 0.8133, + "step": 1679 + }, + { + "epoch": 0.9120521172638436, + "grad_norm": 5.214863992160379, + "learning_rate": 1.8807216800079108e-05, + "loss": 1.0437, + "step": 1680 + }, + { + "epoch": 0.9125950054288816, + "grad_norm": 6.631045200232836, + "learning_rate": 1.8805550560165763e-05, + "loss": 1.446, + "step": 1681 + }, + { + "epoch": 0.9131378935939196, + "grad_norm": 6.140822200691373, + "learning_rate": 1.8803883231167887e-05, + "loss": 1.8672, + "step": 1682 + }, + { + "epoch": 0.9136807817589576, + "grad_norm": 4.5606387908479995, + "learning_rate": 1.8802214813291708e-05, + "loss": 1.0662, + "step": 1683 + }, + { + "epoch": 0.9142236699239956, + "grad_norm": 4.166651706834997, + "learning_rate": 1.8800545306743567e-05, + "loss": 0.749, + "step": 1684 + }, + { + "epoch": 0.9147665580890336, + "grad_norm": 4.674238905131916, + "learning_rate": 1.8798874711729957e-05, + "loss": 1.1106, + "step": 1685 + }, + { + "epoch": 0.9153094462540716, + "grad_norm": 5.224600466137879, + "learning_rate": 1.8797203028457497e-05, + "loss": 0.9652, + "step": 1686 + }, + { + "epoch": 0.9158523344191096, + "grad_norm": 5.87706899393542, + "learning_rate": 1.8795530257132947e-05, + "loss": 1.2146, + "step": 1687 + }, + { + "epoch": 0.9163952225841476, + "grad_norm": 7.587853042260599, + "learning_rate": 1.87938563979632e-05, + "loss": 1.9051, + "step": 1688 + }, + { + "epoch": 0.9169381107491856, + "grad_norm": 4.988334377581133, + "learning_rate": 1.8792181451155275e-05, + "loss": 0.8502, + "step": 1689 + }, + { + "epoch": 0.9174809989142236, + "grad_norm": 5.454638378956608, + "learning_rate": 1.8790505416916338e-05, + "loss": 1.0493, + "step": 1690 + }, + { + "epoch": 0.9180238870792616, + "grad_norm": 5.95672873256444, + "learning_rate": 1.878882829545368e-05, + "loss": 1.1296, + "step": 1691 + }, + { + "epoch": 0.9185667752442996, + "grad_norm": 6.921816321230723, + "learning_rate": 1.8787150086974734e-05, + "loss": 1.6388, + "step": 1692 + }, + { + "epoch": 0.9191096634093376, + "grad_norm": 5.423998305707574, + "learning_rate": 1.878547079168706e-05, + "loss": 1.4865, + "step": 1693 + }, + { + "epoch": 0.9196525515743756, + "grad_norm": 4.479238264802495, + "learning_rate": 1.878379040979835e-05, + "loss": 1.0257, + "step": 1694 + }, + { + "epoch": 0.9201954397394136, + "grad_norm": 6.325821606507086, + "learning_rate": 1.8782108941516446e-05, + "loss": 1.1156, + "step": 1695 + }, + { + "epoch": 0.9207383279044516, + "grad_norm": 6.220712298120886, + "learning_rate": 1.8780426387049315e-05, + "loss": 1.375, + "step": 1696 + }, + { + "epoch": 0.9212812160694897, + "grad_norm": 5.813906943099404, + "learning_rate": 1.877874274660505e-05, + "loss": 1.1869, + "step": 1697 + }, + { + "epoch": 0.9218241042345277, + "grad_norm": 5.0160254446208965, + "learning_rate": 1.8777058020391893e-05, + "loss": 1.0443, + "step": 1698 + }, + { + "epoch": 0.9223669923995657, + "grad_norm": 4.273981514908416, + "learning_rate": 1.877537220861821e-05, + "loss": 1.0015, + "step": 1699 + }, + { + "epoch": 0.9229098805646037, + "grad_norm": 4.908796260576538, + "learning_rate": 1.8773685311492513e-05, + "loss": 1.0203, + "step": 1700 + }, + { + "epoch": 0.9234527687296417, + "grad_norm": 5.566794452728763, + "learning_rate": 1.8771997329223425e-05, + "loss": 1.4637, + "step": 1701 + }, + { + "epoch": 0.9239956568946797, + "grad_norm": 4.946565856604512, + "learning_rate": 1.8770308262019733e-05, + "loss": 0.8093, + "step": 1702 + }, + { + "epoch": 0.9245385450597177, + "grad_norm": 4.767566465605265, + "learning_rate": 1.8768618110090334e-05, + "loss": 0.5938, + "step": 1703 + }, + { + "epoch": 0.9250814332247557, + "grad_norm": 6.919371263621627, + "learning_rate": 1.8766926873644272e-05, + "loss": 0.9857, + "step": 1704 + }, + { + "epoch": 0.9256243213897937, + "grad_norm": 3.5297488420709944, + "learning_rate": 1.876523455289072e-05, + "loss": 0.8744, + "step": 1705 + }, + { + "epoch": 0.9261672095548317, + "grad_norm": 4.910386600622229, + "learning_rate": 1.8763541148038994e-05, + "loss": 0.9491, + "step": 1706 + }, + { + "epoch": 0.9267100977198697, + "grad_norm": 5.550892558363115, + "learning_rate": 1.876184665929853e-05, + "loss": 1.2793, + "step": 1707 + }, + { + "epoch": 0.9272529858849077, + "grad_norm": 4.470721855980947, + "learning_rate": 1.8760151086878905e-05, + "loss": 0.9706, + "step": 1708 + }, + { + "epoch": 0.9277958740499457, + "grad_norm": 7.1893201073258926, + "learning_rate": 1.8758454430989833e-05, + "loss": 1.2473, + "step": 1709 + }, + { + "epoch": 0.9283387622149837, + "grad_norm": 5.933304434057961, + "learning_rate": 1.875675669184116e-05, + "loss": 1.4917, + "step": 1710 + }, + { + "epoch": 0.9288816503800217, + "grad_norm": 5.386119916834561, + "learning_rate": 1.8755057869642857e-05, + "loss": 0.9417, + "step": 1711 + }, + { + "epoch": 0.9294245385450597, + "grad_norm": 6.121437468126703, + "learning_rate": 1.875335796460505e-05, + "loss": 1.203, + "step": 1712 + }, + { + "epoch": 0.9299674267100977, + "grad_norm": 5.34957970840911, + "learning_rate": 1.8751656976937974e-05, + "loss": 1.2557, + "step": 1713 + }, + { + "epoch": 0.9305103148751357, + "grad_norm": 7.404283054108375, + "learning_rate": 1.8749954906852023e-05, + "loss": 1.449, + "step": 1714 + }, + { + "epoch": 0.9310532030401737, + "grad_norm": 6.5511839319022585, + "learning_rate": 1.8748251754557696e-05, + "loss": 1.5548, + "step": 1715 + }, + { + "epoch": 0.9315960912052117, + "grad_norm": 5.0254557279034815, + "learning_rate": 1.8746547520265654e-05, + "loss": 0.9777, + "step": 1716 + }, + { + "epoch": 0.9321389793702497, + "grad_norm": 6.385252198006408, + "learning_rate": 1.874484220418667e-05, + "loss": 1.254, + "step": 1717 + }, + { + "epoch": 0.9326818675352877, + "grad_norm": 6.872281478717196, + "learning_rate": 1.874313580653167e-05, + "loss": 1.278, + "step": 1718 + }, + { + "epoch": 0.9332247557003257, + "grad_norm": 8.525842687889295, + "learning_rate": 1.8741428327511696e-05, + "loss": 1.3712, + "step": 1719 + }, + { + "epoch": 0.9337676438653637, + "grad_norm": 7.97420765579221, + "learning_rate": 1.8739719767337933e-05, + "loss": 1.7493, + "step": 1720 + }, + { + "epoch": 0.9343105320304017, + "grad_norm": 6.107109526814546, + "learning_rate": 1.8738010126221705e-05, + "loss": 0.9861, + "step": 1721 + }, + { + "epoch": 0.9348534201954397, + "grad_norm": 6.4666849584489645, + "learning_rate": 1.8736299404374453e-05, + "loss": 0.8041, + "step": 1722 + }, + { + "epoch": 0.9353963083604777, + "grad_norm": 7.489590627219774, + "learning_rate": 1.873458760200777e-05, + "loss": 1.2325, + "step": 1723 + }, + { + "epoch": 0.9359391965255157, + "grad_norm": 7.020182643385788, + "learning_rate": 1.8732874719333373e-05, + "loss": 1.1698, + "step": 1724 + }, + { + "epoch": 0.9364820846905537, + "grad_norm": 6.436449840675483, + "learning_rate": 1.873116075656311e-05, + "loss": 1.084, + "step": 1725 + }, + { + "epoch": 0.9370249728555917, + "grad_norm": 7.27100690341834, + "learning_rate": 1.872944571390897e-05, + "loss": 0.9793, + "step": 1726 + }, + { + "epoch": 0.9375678610206297, + "grad_norm": 6.337617737677888, + "learning_rate": 1.872772959158307e-05, + "loss": 1.0558, + "step": 1727 + }, + { + "epoch": 0.9381107491856677, + "grad_norm": 5.356493873446841, + "learning_rate": 1.8726012389797667e-05, + "loss": 1.1031, + "step": 1728 + }, + { + "epoch": 0.9386536373507057, + "grad_norm": 6.302965546816682, + "learning_rate": 1.8724294108765142e-05, + "loss": 1.1401, + "step": 1729 + }, + { + "epoch": 0.9391965255157437, + "grad_norm": 6.116650787967334, + "learning_rate": 1.872257474869802e-05, + "loss": 1.0848, + "step": 1730 + }, + { + "epoch": 0.9397394136807817, + "grad_norm": 4.3874746967683, + "learning_rate": 1.8720854309808948e-05, + "loss": 0.7463, + "step": 1731 + }, + { + "epoch": 0.9402823018458197, + "grad_norm": 5.00673534942558, + "learning_rate": 1.871913279231072e-05, + "loss": 1.2365, + "step": 1732 + }, + { + "epoch": 0.9408251900108577, + "grad_norm": 5.222845312743783, + "learning_rate": 1.871741019641625e-05, + "loss": 1.062, + "step": 1733 + }, + { + "epoch": 0.9413680781758957, + "grad_norm": 4.624403224726164, + "learning_rate": 1.871568652233859e-05, + "loss": 1.0155, + "step": 1734 + }, + { + "epoch": 0.9419109663409337, + "grad_norm": 5.386412889478532, + "learning_rate": 1.8713961770290936e-05, + "loss": 1.4316, + "step": 1735 + }, + { + "epoch": 0.9424538545059717, + "grad_norm": 6.0899860043030705, + "learning_rate": 1.87122359404866e-05, + "loss": 1.1372, + "step": 1736 + }, + { + "epoch": 0.9429967426710097, + "grad_norm": 5.306362358042622, + "learning_rate": 1.8710509033139037e-05, + "loss": 0.9353, + "step": 1737 + }, + { + "epoch": 0.9435396308360477, + "grad_norm": 6.479420699028343, + "learning_rate": 1.8708781048461832e-05, + "loss": 1.2435, + "step": 1738 + }, + { + "epoch": 0.9440825190010858, + "grad_norm": 6.205746500302448, + "learning_rate": 1.8707051986668712e-05, + "loss": 0.7872, + "step": 1739 + }, + { + "epoch": 0.9446254071661238, + "grad_norm": 4.157411034856641, + "learning_rate": 1.8705321847973523e-05, + "loss": 0.8004, + "step": 1740 + }, + { + "epoch": 0.9451682953311618, + "grad_norm": 5.5814599600578285, + "learning_rate": 1.8703590632590254e-05, + "loss": 0.7384, + "step": 1741 + }, + { + "epoch": 0.9457111834961998, + "grad_norm": 6.040759722497843, + "learning_rate": 1.8701858340733023e-05, + "loss": 1.5941, + "step": 1742 + }, + { + "epoch": 0.9462540716612378, + "grad_norm": 4.592040464655725, + "learning_rate": 1.8700124972616085e-05, + "loss": 1.0662, + "step": 1743 + }, + { + "epoch": 0.9467969598262758, + "grad_norm": 6.077319934267432, + "learning_rate": 1.8698390528453823e-05, + "loss": 1.3726, + "step": 1744 + }, + { + "epoch": 0.9473398479913138, + "grad_norm": 4.255180894541495, + "learning_rate": 1.869665500846076e-05, + "loss": 0.7001, + "step": 1745 + }, + { + "epoch": 0.9478827361563518, + "grad_norm": 6.390593673413393, + "learning_rate": 1.869491841285154e-05, + "loss": 1.2268, + "step": 1746 + }, + { + "epoch": 0.9484256243213898, + "grad_norm": 5.347324129941674, + "learning_rate": 1.8693180741840957e-05, + "loss": 1.2193, + "step": 1747 + }, + { + "epoch": 0.9489685124864278, + "grad_norm": 6.407261723862006, + "learning_rate": 1.8691441995643927e-05, + "loss": 0.868, + "step": 1748 + }, + { + "epoch": 0.9495114006514658, + "grad_norm": 4.400520878629861, + "learning_rate": 1.8689702174475496e-05, + "loss": 0.4679, + "step": 1749 + }, + { + "epoch": 0.9500542888165038, + "grad_norm": 5.6161548427455505, + "learning_rate": 1.8687961278550852e-05, + "loss": 0.8842, + "step": 1750 + }, + { + "epoch": 0.9505971769815418, + "grad_norm": 4.7163538240048, + "learning_rate": 1.8686219308085306e-05, + "loss": 0.9972, + "step": 1751 + }, + { + "epoch": 0.9511400651465798, + "grad_norm": 4.845854095585142, + "learning_rate": 1.8684476263294318e-05, + "loss": 1.1046, + "step": 1752 + }, + { + "epoch": 0.9516829533116178, + "grad_norm": 3.878984502983999, + "learning_rate": 1.8682732144393463e-05, + "loss": 0.7873, + "step": 1753 + }, + { + "epoch": 0.9522258414766558, + "grad_norm": 5.736472248757515, + "learning_rate": 1.8680986951598458e-05, + "loss": 1.2046, + "step": 1754 + }, + { + "epoch": 0.9527687296416938, + "grad_norm": 5.12988223940441, + "learning_rate": 1.867924068512515e-05, + "loss": 0.6293, + "step": 1755 + }, + { + "epoch": 0.9533116178067318, + "grad_norm": 7.822467968073818, + "learning_rate": 1.867749334518952e-05, + "loss": 1.3625, + "step": 1756 + }, + { + "epoch": 0.9538545059717698, + "grad_norm": 5.488388947029871, + "learning_rate": 1.8675744932007687e-05, + "loss": 1.078, + "step": 1757 + }, + { + "epoch": 0.9543973941368078, + "grad_norm": 4.631848438676548, + "learning_rate": 1.8673995445795894e-05, + "loss": 0.69, + "step": 1758 + }, + { + "epoch": 0.9549402823018458, + "grad_norm": 7.078171019993124, + "learning_rate": 1.8672244886770516e-05, + "loss": 1.1036, + "step": 1759 + }, + { + "epoch": 0.9554831704668838, + "grad_norm": 5.9259972832349215, + "learning_rate": 1.8670493255148073e-05, + "loss": 0.7919, + "step": 1760 + }, + { + "epoch": 0.9560260586319218, + "grad_norm": 5.846343833877514, + "learning_rate": 1.8668740551145205e-05, + "loss": 1.1653, + "step": 1761 + }, + { + "epoch": 0.9565689467969598, + "grad_norm": 5.559082125902064, + "learning_rate": 1.8666986774978685e-05, + "loss": 1.4214, + "step": 1762 + }, + { + "epoch": 0.9571118349619978, + "grad_norm": 6.511207981127819, + "learning_rate": 1.8665231926865433e-05, + "loss": 1.3552, + "step": 1763 + }, + { + "epoch": 0.9576547231270358, + "grad_norm": 5.728999474717542, + "learning_rate": 1.8663476007022482e-05, + "loss": 0.7861, + "step": 1764 + }, + { + "epoch": 0.9581976112920738, + "grad_norm": 6.152118564149309, + "learning_rate": 1.8661719015667016e-05, + "loss": 0.8679, + "step": 1765 + }, + { + "epoch": 0.9587404994571118, + "grad_norm": 6.77121947972117, + "learning_rate": 1.8659960953016334e-05, + "loss": 0.8103, + "step": 1766 + }, + { + "epoch": 0.9592833876221498, + "grad_norm": 7.822575173268641, + "learning_rate": 1.865820181928788e-05, + "loss": 1.2558, + "step": 1767 + }, + { + "epoch": 0.9598262757871878, + "grad_norm": 6.389073592671253, + "learning_rate": 1.8656441614699225e-05, + "loss": 1.0314, + "step": 1768 + }, + { + "epoch": 0.9603691639522258, + "grad_norm": 6.141689972496849, + "learning_rate": 1.8654680339468076e-05, + "loss": 1.2452, + "step": 1769 + }, + { + "epoch": 0.9609120521172638, + "grad_norm": 6.934092850250001, + "learning_rate": 1.8652917993812267e-05, + "loss": 1.1186, + "step": 1770 + }, + { + "epoch": 0.9614549402823018, + "grad_norm": 6.129320663785202, + "learning_rate": 1.865115457794977e-05, + "loss": 0.9076, + "step": 1771 + }, + { + "epoch": 0.9619978284473398, + "grad_norm": 7.008261896063745, + "learning_rate": 1.8649390092098693e-05, + "loss": 1.7152, + "step": 1772 + }, + { + "epoch": 0.9625407166123778, + "grad_norm": 6.36833090372477, + "learning_rate": 1.8647624536477255e-05, + "loss": 0.8087, + "step": 1773 + }, + { + "epoch": 0.9630836047774158, + "grad_norm": 4.7595903392838865, + "learning_rate": 1.8645857911303838e-05, + "loss": 0.6374, + "step": 1774 + }, + { + "epoch": 0.9636264929424538, + "grad_norm": 6.410325139984648, + "learning_rate": 1.8644090216796934e-05, + "loss": 1.3611, + "step": 1775 + }, + { + "epoch": 0.9641693811074918, + "grad_norm": 5.900212705071564, + "learning_rate": 1.8642321453175177e-05, + "loss": 0.891, + "step": 1776 + }, + { + "epoch": 0.9647122692725298, + "grad_norm": 6.90562183471938, + "learning_rate": 1.8640551620657326e-05, + "loss": 0.7899, + "step": 1777 + }, + { + "epoch": 0.9652551574375678, + "grad_norm": 6.476066490933977, + "learning_rate": 1.8638780719462278e-05, + "loss": 0.9614, + "step": 1778 + }, + { + "epoch": 0.9657980456026058, + "grad_norm": 6.659586852509603, + "learning_rate": 1.8637008749809065e-05, + "loss": 1.2419, + "step": 1779 + }, + { + "epoch": 0.9663409337676439, + "grad_norm": 6.587754652016677, + "learning_rate": 1.8635235711916847e-05, + "loss": 1.3664, + "step": 1780 + }, + { + "epoch": 0.9668838219326819, + "grad_norm": 6.558578755813245, + "learning_rate": 1.863346160600491e-05, + "loss": 1.1249, + "step": 1781 + }, + { + "epoch": 0.9674267100977199, + "grad_norm": 4.791301716076818, + "learning_rate": 1.8631686432292685e-05, + "loss": 0.6836, + "step": 1782 + }, + { + "epoch": 0.9679695982627579, + "grad_norm": 5.120778292537024, + "learning_rate": 1.862991019099972e-05, + "loss": 0.592, + "step": 1783 + }, + { + "epoch": 0.9685124864277959, + "grad_norm": 8.250755784410368, + "learning_rate": 1.8628132882345713e-05, + "loss": 1.2579, + "step": 1784 + }, + { + "epoch": 0.9690553745928339, + "grad_norm": 5.811263926579046, + "learning_rate": 1.862635450655048e-05, + "loss": 1.2935, + "step": 1785 + }, + { + "epoch": 0.9695982627578719, + "grad_norm": 6.042040633945999, + "learning_rate": 1.862457506383397e-05, + "loss": 0.8784, + "step": 1786 + }, + { + "epoch": 0.9701411509229099, + "grad_norm": 7.115011060082062, + "learning_rate": 1.8622794554416272e-05, + "loss": 1.0717, + "step": 1787 + }, + { + "epoch": 0.9706840390879479, + "grad_norm": 7.520775839533288, + "learning_rate": 1.8621012978517604e-05, + "loss": 1.3468, + "step": 1788 + }, + { + "epoch": 0.9712269272529859, + "grad_norm": 6.949921961354188, + "learning_rate": 1.8619230336358306e-05, + "loss": 1.5609, + "step": 1789 + }, + { + "epoch": 0.9717698154180239, + "grad_norm": 6.396218455762339, + "learning_rate": 1.8617446628158866e-05, + "loss": 1.1807, + "step": 1790 + }, + { + "epoch": 0.9723127035830619, + "grad_norm": 6.314794279037124, + "learning_rate": 1.861566185413989e-05, + "loss": 0.9255, + "step": 1791 + }, + { + "epoch": 0.9728555917480999, + "grad_norm": 5.393752442926871, + "learning_rate": 1.8613876014522128e-05, + "loss": 0.7926, + "step": 1792 + }, + { + "epoch": 0.9733984799131379, + "grad_norm": 5.908313693688933, + "learning_rate": 1.8612089109526453e-05, + "loss": 0.7984, + "step": 1793 + }, + { + "epoch": 0.9739413680781759, + "grad_norm": 4.6952509401002125, + "learning_rate": 1.8610301139373867e-05, + "loss": 1.0344, + "step": 1794 + }, + { + "epoch": 0.9744842562432139, + "grad_norm": 6.81835335105175, + "learning_rate": 1.8608512104285517e-05, + "loss": 1.1532, + "step": 1795 + }, + { + "epoch": 0.9750271444082519, + "grad_norm": 4.159411114912345, + "learning_rate": 1.860672200448267e-05, + "loss": 1.0665, + "step": 1796 + }, + { + "epoch": 0.9755700325732899, + "grad_norm": 4.911404446825004, + "learning_rate": 1.8604930840186726e-05, + "loss": 1.1784, + "step": 1797 + }, + { + "epoch": 0.9761129207383279, + "grad_norm": 5.399453855489055, + "learning_rate": 1.860313861161922e-05, + "loss": 0.8215, + "step": 1798 + }, + { + "epoch": 0.9766558089033659, + "grad_norm": 4.797137239148531, + "learning_rate": 1.860134531900182e-05, + "loss": 0.6405, + "step": 1799 + }, + { + "epoch": 0.9771986970684039, + "grad_norm": 3.937076316066272, + "learning_rate": 1.859955096255633e-05, + "loss": 0.5912, + "step": 1800 + }, + { + "epoch": 0.9777415852334419, + "grad_norm": 5.569110777302129, + "learning_rate": 1.859775554250466e-05, + "loss": 0.8089, + "step": 1801 + }, + { + "epoch": 0.9782844733984799, + "grad_norm": 6.364258901894282, + "learning_rate": 1.859595905906889e-05, + "loss": 1.5181, + "step": 1802 + }, + { + "epoch": 0.9788273615635179, + "grad_norm": 4.995224609073908, + "learning_rate": 1.85941615124712e-05, + "loss": 0.8573, + "step": 1803 + }, + { + "epoch": 0.9793702497285559, + "grad_norm": 5.041992740138782, + "learning_rate": 1.8592362902933918e-05, + "loss": 0.8982, + "step": 1804 + }, + { + "epoch": 0.9799131378935939, + "grad_norm": 5.6054574490953275, + "learning_rate": 1.8590563230679496e-05, + "loss": 0.9285, + "step": 1805 + }, + { + "epoch": 0.9804560260586319, + "grad_norm": 5.269503319178264, + "learning_rate": 1.8588762495930526e-05, + "loss": 1.0963, + "step": 1806 + }, + { + "epoch": 0.9809989142236699, + "grad_norm": 6.184787731593613, + "learning_rate": 1.8586960698909718e-05, + "loss": 1.4005, + "step": 1807 + }, + { + "epoch": 0.9815418023887079, + "grad_norm": 8.50824591479003, + "learning_rate": 1.858515783983993e-05, + "loss": 1.9607, + "step": 1808 + }, + { + "epoch": 0.9820846905537459, + "grad_norm": 5.399464732589782, + "learning_rate": 1.8583353918944134e-05, + "loss": 0.951, + "step": 1809 + }, + { + "epoch": 0.9826275787187839, + "grad_norm": 6.577634699737779, + "learning_rate": 1.8581548936445447e-05, + "loss": 1.5139, + "step": 1810 + }, + { + "epoch": 0.9831704668838219, + "grad_norm": 10.183489163585788, + "learning_rate": 1.8579742892567107e-05, + "loss": 2.0539, + "step": 1811 + }, + { + "epoch": 0.9837133550488599, + "grad_norm": 6.578605968943458, + "learning_rate": 1.8577935787532494e-05, + "loss": 1.1454, + "step": 1812 + }, + { + "epoch": 0.9842562432138979, + "grad_norm": 4.923285404215955, + "learning_rate": 1.8576127621565113e-05, + "loss": 0.8965, + "step": 1813 + }, + { + "epoch": 0.9847991313789359, + "grad_norm": 5.912445932400008, + "learning_rate": 1.85743183948886e-05, + "loss": 1.1311, + "step": 1814 + }, + { + "epoch": 0.9853420195439739, + "grad_norm": 4.5183231221646105, + "learning_rate": 1.8572508107726725e-05, + "loss": 0.7511, + "step": 1815 + }, + { + "epoch": 0.9858849077090119, + "grad_norm": 5.2218543332097935, + "learning_rate": 1.8570696760303378e-05, + "loss": 1.0063, + "step": 1816 + }, + { + "epoch": 0.9864277958740499, + "grad_norm": 7.215748803707923, + "learning_rate": 1.85688843528426e-05, + "loss": 1.3674, + "step": 1817 + }, + { + "epoch": 0.9869706840390879, + "grad_norm": 6.362354074709609, + "learning_rate": 1.8567070885568547e-05, + "loss": 1.1319, + "step": 1818 + }, + { + "epoch": 0.987513572204126, + "grad_norm": 4.680755706450828, + "learning_rate": 1.8565256358705513e-05, + "loss": 0.6613, + "step": 1819 + }, + { + "epoch": 0.988056460369164, + "grad_norm": 4.80241945107321, + "learning_rate": 1.8563440772477922e-05, + "loss": 0.8619, + "step": 1820 + }, + { + "epoch": 0.988599348534202, + "grad_norm": 5.858709563059296, + "learning_rate": 1.856162412711033e-05, + "loss": 0.9432, + "step": 1821 + }, + { + "epoch": 0.98914223669924, + "grad_norm": 6.478119299588854, + "learning_rate": 1.855980642282742e-05, + "loss": 1.1866, + "step": 1822 + }, + { + "epoch": 0.989685124864278, + "grad_norm": 7.081924554034699, + "learning_rate": 1.8557987659854006e-05, + "loss": 1.3657, + "step": 1823 + }, + { + "epoch": 0.990228013029316, + "grad_norm": 5.404410220733361, + "learning_rate": 1.855616783841504e-05, + "loss": 1.1051, + "step": 1824 + }, + { + "epoch": 0.990770901194354, + "grad_norm": 6.742207962837618, + "learning_rate": 1.8554346958735602e-05, + "loss": 1.4667, + "step": 1825 + }, + { + "epoch": 0.991313789359392, + "grad_norm": 6.164178667211635, + "learning_rate": 1.8552525021040895e-05, + "loss": 0.7861, + "step": 1826 + }, + { + "epoch": 0.99185667752443, + "grad_norm": 6.078017237490637, + "learning_rate": 1.8550702025556265e-05, + "loss": 1.3294, + "step": 1827 + }, + { + "epoch": 0.992399565689468, + "grad_norm": 4.7372949568909455, + "learning_rate": 1.8548877972507182e-05, + "loss": 0.9779, + "step": 1828 + }, + { + "epoch": 0.992942453854506, + "grad_norm": 5.25971277777658, + "learning_rate": 1.8547052862119247e-05, + "loss": 1.049, + "step": 1829 + }, + { + "epoch": 0.993485342019544, + "grad_norm": 4.950709970125391, + "learning_rate": 1.854522669461819e-05, + "loss": 0.6665, + "step": 1830 + }, + { + "epoch": 0.994028230184582, + "grad_norm": 5.856428194694733, + "learning_rate": 1.8543399470229876e-05, + "loss": 1.0954, + "step": 1831 + }, + { + "epoch": 0.99457111834962, + "grad_norm": 4.895286613591861, + "learning_rate": 1.85415711891803e-05, + "loss": 0.7069, + "step": 1832 + }, + { + "epoch": 0.995114006514658, + "grad_norm": 5.249228502887773, + "learning_rate": 1.8539741851695586e-05, + "loss": 1.0655, + "step": 1833 + }, + { + "epoch": 0.995656894679696, + "grad_norm": 4.855694799243912, + "learning_rate": 1.8537911458001988e-05, + "loss": 1.1872, + "step": 1834 + }, + { + "epoch": 0.996199782844734, + "grad_norm": 5.180821237913711, + "learning_rate": 1.8536080008325896e-05, + "loss": 0.8281, + "step": 1835 + }, + { + "epoch": 0.996742671009772, + "grad_norm": 5.073254295771654, + "learning_rate": 1.8534247502893823e-05, + "loss": 0.7002, + "step": 1836 + }, + { + "epoch": 0.99728555917481, + "grad_norm": 5.4022604097623255, + "learning_rate": 1.8532413941932416e-05, + "loss": 1.1793, + "step": 1837 + }, + { + "epoch": 0.997828447339848, + "grad_norm": 4.981761324197401, + "learning_rate": 1.8530579325668455e-05, + "loss": 0.8331, + "step": 1838 + }, + { + "epoch": 0.998371335504886, + "grad_norm": 7.325363733761658, + "learning_rate": 1.852874365432885e-05, + "loss": 1.3071, + "step": 1839 + }, + { + "epoch": 0.998914223669924, + "grad_norm": 7.5668251202705505, + "learning_rate": 1.852690692814063e-05, + "loss": 1.4955, + "step": 1840 + }, + { + "epoch": 0.999457111834962, + "grad_norm": 6.706936559053067, + "learning_rate": 1.8525069147330978e-05, + "loss": 1.0286, + "step": 1841 + }, + { + "epoch": 1.0, + "grad_norm": 6.338369283394405, + "learning_rate": 1.8523230312127183e-05, + "loss": 1.0509, + "step": 1842 + }, + { + "epoch": 1.000542888165038, + "grad_norm": 6.119502670416093, + "learning_rate": 1.8521390422756683e-05, + "loss": 1.0805, + "step": 1843 + }, + { + "epoch": 1.001085776330076, + "grad_norm": 5.074420967885639, + "learning_rate": 1.851954947944703e-05, + "loss": 1.0931, + "step": 1844 + }, + { + "epoch": 1.001628664495114, + "grad_norm": 8.194067497007946, + "learning_rate": 1.851770748242592e-05, + "loss": 2.0732, + "step": 1845 + }, + { + "epoch": 1.002171552660152, + "grad_norm": 4.389522929363591, + "learning_rate": 1.8515864431921177e-05, + "loss": 0.8684, + "step": 1846 + }, + { + "epoch": 1.00271444082519, + "grad_norm": 5.720373954211297, + "learning_rate": 1.8514020328160748e-05, + "loss": 1.0223, + "step": 1847 + }, + { + "epoch": 1.003257328990228, + "grad_norm": 5.163678744409467, + "learning_rate": 1.8512175171372713e-05, + "loss": 0.9611, + "step": 1848 + }, + { + "epoch": 1.003800217155266, + "grad_norm": 5.786056891749039, + "learning_rate": 1.8510328961785286e-05, + "loss": 1.3473, + "step": 1849 + }, + { + "epoch": 1.004343105320304, + "grad_norm": 5.9194250471944985, + "learning_rate": 1.850848169962681e-05, + "loss": 0.8218, + "step": 1850 + }, + { + "epoch": 1.004885993485342, + "grad_norm": 5.890241796399565, + "learning_rate": 1.850663338512576e-05, + "loss": 0.842, + "step": 1851 + }, + { + "epoch": 1.00542888165038, + "grad_norm": 6.939341519595904, + "learning_rate": 1.8504784018510732e-05, + "loss": 0.9263, + "step": 1852 + }, + { + "epoch": 1.005971769815418, + "grad_norm": 4.592014594714093, + "learning_rate": 1.850293360001046e-05, + "loss": 1.015, + "step": 1853 + }, + { + "epoch": 1.006514657980456, + "grad_norm": 4.575145990552055, + "learning_rate": 1.8501082129853816e-05, + "loss": 1.0505, + "step": 1854 + }, + { + "epoch": 1.007057546145494, + "grad_norm": 5.999218616951103, + "learning_rate": 1.849922960826978e-05, + "loss": 1.0945, + "step": 1855 + }, + { + "epoch": 1.007600434310532, + "grad_norm": 5.459227552994586, + "learning_rate": 1.8497376035487483e-05, + "loss": 1.0167, + "step": 1856 + }, + { + "epoch": 1.00814332247557, + "grad_norm": 4.785837605715454, + "learning_rate": 1.8495521411736173e-05, + "loss": 1.0559, + "step": 1857 + }, + { + "epoch": 1.008686210640608, + "grad_norm": 5.813312187108195, + "learning_rate": 1.8493665737245236e-05, + "loss": 1.3477, + "step": 1858 + }, + { + "epoch": 1.009229098805646, + "grad_norm": 6.512945450475729, + "learning_rate": 1.8491809012244182e-05, + "loss": 1.4416, + "step": 1859 + }, + { + "epoch": 1.009771986970684, + "grad_norm": 5.599962385405486, + "learning_rate": 1.8489951236962658e-05, + "loss": 0.8961, + "step": 1860 + }, + { + "epoch": 1.010314875135722, + "grad_norm": 4.863989016796961, + "learning_rate": 1.848809241163043e-05, + "loss": 0.639, + "step": 1861 + }, + { + "epoch": 1.01085776330076, + "grad_norm": 5.937471136238718, + "learning_rate": 1.848623253647741e-05, + "loss": 1.0548, + "step": 1862 + }, + { + "epoch": 1.011400651465798, + "grad_norm": 4.75172125273428, + "learning_rate": 1.8484371611733625e-05, + "loss": 0.6493, + "step": 1863 + }, + { + "epoch": 1.011943539630836, + "grad_norm": 5.867263430143688, + "learning_rate": 1.848250963762923e-05, + "loss": 0.8569, + "step": 1864 + }, + { + "epoch": 1.012486427795874, + "grad_norm": 5.932410925593062, + "learning_rate": 1.848064661439453e-05, + "loss": 0.8351, + "step": 1865 + }, + { + "epoch": 1.013029315960912, + "grad_norm": 5.8320858990674, + "learning_rate": 1.847878254225994e-05, + "loss": 0.8916, + "step": 1866 + }, + { + "epoch": 1.01357220412595, + "grad_norm": 5.836721791741835, + "learning_rate": 1.847691742145601e-05, + "loss": 1.1742, + "step": 1867 + }, + { + "epoch": 1.014115092290988, + "grad_norm": 6.806668722890977, + "learning_rate": 1.8475051252213423e-05, + "loss": 1.0714, + "step": 1868 + }, + { + "epoch": 1.014657980456026, + "grad_norm": 7.026347641082525, + "learning_rate": 1.8473184034762992e-05, + "loss": 1.4101, + "step": 1869 + }, + { + "epoch": 1.015200868621064, + "grad_norm": 5.9532414283206005, + "learning_rate": 1.8471315769335657e-05, + "loss": 0.7855, + "step": 1870 + }, + { + "epoch": 1.015743756786102, + "grad_norm": 5.7578706206451855, + "learning_rate": 1.846944645616248e-05, + "loss": 0.7644, + "step": 1871 + }, + { + "epoch": 1.01628664495114, + "grad_norm": 6.340541706204924, + "learning_rate": 1.846757609547467e-05, + "loss": 1.1607, + "step": 1872 + }, + { + "epoch": 1.016829533116178, + "grad_norm": 4.942991579941027, + "learning_rate": 1.8465704687503558e-05, + "loss": 0.7578, + "step": 1873 + }, + { + "epoch": 1.017372421281216, + "grad_norm": 6.42121491667965, + "learning_rate": 1.846383223248059e-05, + "loss": 0.9924, + "step": 1874 + }, + { + "epoch": 1.017915309446254, + "grad_norm": 6.252498990648505, + "learning_rate": 1.8461958730637368e-05, + "loss": 0.811, + "step": 1875 + }, + { + "epoch": 1.018458197611292, + "grad_norm": 5.941336187529878, + "learning_rate": 1.84600841822056e-05, + "loss": 0.772, + "step": 1876 + }, + { + "epoch": 1.01900108577633, + "grad_norm": 5.323895762793836, + "learning_rate": 1.845820858741714e-05, + "loss": 0.8812, + "step": 1877 + }, + { + "epoch": 1.019543973941368, + "grad_norm": 7.2574721612646105, + "learning_rate": 1.845633194650396e-05, + "loss": 1.2759, + "step": 1878 + }, + { + "epoch": 1.020086862106406, + "grad_norm": 4.307176114235153, + "learning_rate": 1.8454454259698165e-05, + "loss": 0.569, + "step": 1879 + }, + { + "epoch": 1.020629750271444, + "grad_norm": 6.504543473682945, + "learning_rate": 1.8452575527231997e-05, + "loss": 0.9355, + "step": 1880 + }, + { + "epoch": 1.021172638436482, + "grad_norm": 5.822065969327671, + "learning_rate": 1.8450695749337816e-05, + "loss": 1.0287, + "step": 1881 + }, + { + "epoch": 1.02171552660152, + "grad_norm": 8.90668319984316, + "learning_rate": 1.8448814926248112e-05, + "loss": 1.5348, + "step": 1882 + }, + { + "epoch": 1.022258414766558, + "grad_norm": 5.233217719773852, + "learning_rate": 1.844693305819552e-05, + "loss": 0.564, + "step": 1883 + }, + { + "epoch": 1.022801302931596, + "grad_norm": 5.806961114054563, + "learning_rate": 1.844505014541278e-05, + "loss": 0.9349, + "step": 1884 + }, + { + "epoch": 1.023344191096634, + "grad_norm": 6.239514346807307, + "learning_rate": 1.8443166188132777e-05, + "loss": 0.9677, + "step": 1885 + }, + { + "epoch": 1.023887079261672, + "grad_norm": 5.596402826374599, + "learning_rate": 1.8441281186588528e-05, + "loss": 0.695, + "step": 1886 + }, + { + "epoch": 1.02442996742671, + "grad_norm": 6.531429104200087, + "learning_rate": 1.8439395141013165e-05, + "loss": 1.1543, + "step": 1887 + }, + { + "epoch": 1.024972855591748, + "grad_norm": 5.129461327017003, + "learning_rate": 1.843750805163996e-05, + "loss": 1.2624, + "step": 1888 + }, + { + "epoch": 1.0255157437567861, + "grad_norm": 6.862993011679228, + "learning_rate": 1.8435619918702318e-05, + "loss": 1.2673, + "step": 1889 + }, + { + "epoch": 1.0260586319218241, + "grad_norm": 5.134671966877655, + "learning_rate": 1.8433730742433755e-05, + "loss": 0.6032, + "step": 1890 + }, + { + "epoch": 1.0266015200868621, + "grad_norm": 6.9199078267443195, + "learning_rate": 1.8431840523067932e-05, + "loss": 1.1255, + "step": 1891 + }, + { + "epoch": 1.0271444082519001, + "grad_norm": 5.808796921918892, + "learning_rate": 1.8429949260838635e-05, + "loss": 0.8488, + "step": 1892 + }, + { + "epoch": 1.0276872964169381, + "grad_norm": 7.102540931353532, + "learning_rate": 1.842805695597978e-05, + "loss": 1.0634, + "step": 1893 + }, + { + "epoch": 1.0282301845819761, + "grad_norm": 9.932212718124791, + "learning_rate": 1.8426163608725403e-05, + "loss": 1.3909, + "step": 1894 + }, + { + "epoch": 1.0287730727470141, + "grad_norm": 6.999752698080143, + "learning_rate": 1.8424269219309686e-05, + "loss": 1.1106, + "step": 1895 + }, + { + "epoch": 1.0293159609120521, + "grad_norm": 6.569873028691542, + "learning_rate": 1.842237378796693e-05, + "loss": 0.9455, + "step": 1896 + }, + { + "epoch": 1.0298588490770901, + "grad_norm": 6.418746090533204, + "learning_rate": 1.8420477314931554e-05, + "loss": 0.934, + "step": 1897 + }, + { + "epoch": 1.0304017372421281, + "grad_norm": 7.7235490740904815, + "learning_rate": 1.8418579800438125e-05, + "loss": 1.1582, + "step": 1898 + }, + { + "epoch": 1.0309446254071661, + "grad_norm": 5.4350436368749895, + "learning_rate": 1.841668124472133e-05, + "loss": 0.7726, + "step": 1899 + }, + { + "epoch": 1.0314875135722041, + "grad_norm": 5.34390520812136, + "learning_rate": 1.8414781648015983e-05, + "loss": 0.7497, + "step": 1900 + }, + { + "epoch": 1.0320304017372421, + "grad_norm": 6.563688266956639, + "learning_rate": 1.841288101055703e-05, + "loss": 1.3352, + "step": 1901 + }, + { + "epoch": 1.0325732899022801, + "grad_norm": 7.710283890330315, + "learning_rate": 1.841097933257955e-05, + "loss": 0.8804, + "step": 1902 + }, + { + "epoch": 1.0331161780673181, + "grad_norm": 5.31500431944041, + "learning_rate": 1.840907661431874e-05, + "loss": 0.8667, + "step": 1903 + }, + { + "epoch": 1.0336590662323561, + "grad_norm": 7.476982099026401, + "learning_rate": 1.840717285600993e-05, + "loss": 0.8317, + "step": 1904 + }, + { + "epoch": 1.0342019543973942, + "grad_norm": 5.1563433597221415, + "learning_rate": 1.840526805788858e-05, + "loss": 1.0108, + "step": 1905 + }, + { + "epoch": 1.0347448425624322, + "grad_norm": 6.411832164370409, + "learning_rate": 1.8403362220190284e-05, + "loss": 0.7778, + "step": 1906 + }, + { + "epoch": 1.0352877307274702, + "grad_norm": 6.688312152010462, + "learning_rate": 1.8401455343150757e-05, + "loss": 1.2438, + "step": 1907 + }, + { + "epoch": 1.0358306188925082, + "grad_norm": 6.412659783246144, + "learning_rate": 1.839954742700584e-05, + "loss": 0.8191, + "step": 1908 + }, + { + "epoch": 1.0363735070575462, + "grad_norm": 6.399618830323376, + "learning_rate": 1.839763847199151e-05, + "loss": 1.4485, + "step": 1909 + }, + { + "epoch": 1.0369163952225842, + "grad_norm": 8.041718562623554, + "learning_rate": 1.8395728478343873e-05, + "loss": 1.0043, + "step": 1910 + }, + { + "epoch": 1.0374592833876222, + "grad_norm": 6.4118854777282905, + "learning_rate": 1.8393817446299152e-05, + "loss": 0.9183, + "step": 1911 + }, + { + "epoch": 1.0380021715526602, + "grad_norm": 6.400834366436549, + "learning_rate": 1.8391905376093717e-05, + "loss": 1.1517, + "step": 1912 + }, + { + "epoch": 1.0385450597176982, + "grad_norm": 6.160364431172616, + "learning_rate": 1.8389992267964046e-05, + "loss": 0.5949, + "step": 1913 + }, + { + "epoch": 1.0390879478827362, + "grad_norm": 7.298564934931577, + "learning_rate": 1.8388078122146763e-05, + "loss": 1.0234, + "step": 1914 + }, + { + "epoch": 1.0396308360477742, + "grad_norm": 6.132796013684569, + "learning_rate": 1.838616293887861e-05, + "loss": 0.9417, + "step": 1915 + }, + { + "epoch": 1.0401737242128122, + "grad_norm": 8.083369310536504, + "learning_rate": 1.8384246718396458e-05, + "loss": 1.2617, + "step": 1916 + }, + { + "epoch": 1.0407166123778502, + "grad_norm": 5.492413210001038, + "learning_rate": 1.8382329460937306e-05, + "loss": 0.7906, + "step": 1917 + }, + { + "epoch": 1.0412595005428882, + "grad_norm": 5.936786319330656, + "learning_rate": 1.838041116673829e-05, + "loss": 0.8254, + "step": 1918 + }, + { + "epoch": 1.0418023887079262, + "grad_norm": 6.637068024744901, + "learning_rate": 1.8378491836036666e-05, + "loss": 1.1779, + "step": 1919 + }, + { + "epoch": 1.0423452768729642, + "grad_norm": 5.481354066687985, + "learning_rate": 1.8376571469069814e-05, + "loss": 0.9586, + "step": 1920 + }, + { + "epoch": 1.0428881650380022, + "grad_norm": 7.669719057080363, + "learning_rate": 1.8374650066075257e-05, + "loss": 1.2058, + "step": 1921 + }, + { + "epoch": 1.0434310532030402, + "grad_norm": 7.387160499244698, + "learning_rate": 1.8372727627290627e-05, + "loss": 1.3861, + "step": 1922 + }, + { + "epoch": 1.0439739413680782, + "grad_norm": 4.975886150156898, + "learning_rate": 1.8370804152953704e-05, + "loss": 0.5849, + "step": 1923 + }, + { + "epoch": 1.0445168295331162, + "grad_norm": 5.413290989228805, + "learning_rate": 1.8368879643302383e-05, + "loss": 0.8425, + "step": 1924 + }, + { + "epoch": 1.0450597176981542, + "grad_norm": 6.564098001301886, + "learning_rate": 1.836695409857469e-05, + "loss": 0.9658, + "step": 1925 + }, + { + "epoch": 1.0456026058631922, + "grad_norm": 8.050721283259312, + "learning_rate": 1.8365027519008774e-05, + "loss": 1.684, + "step": 1926 + }, + { + "epoch": 1.0461454940282302, + "grad_norm": 5.409117554986433, + "learning_rate": 1.836309990484293e-05, + "loss": 0.8247, + "step": 1927 + }, + { + "epoch": 1.0466883821932682, + "grad_norm": 5.108392155413988, + "learning_rate": 1.8361171256315555e-05, + "loss": 0.8244, + "step": 1928 + }, + { + "epoch": 1.0472312703583062, + "grad_norm": 6.806129539780182, + "learning_rate": 1.8359241573665194e-05, + "loss": 1.0678, + "step": 1929 + }, + { + "epoch": 1.0477741585233442, + "grad_norm": 4.43019983774416, + "learning_rate": 1.8357310857130514e-05, + "loss": 0.6213, + "step": 1930 + }, + { + "epoch": 1.0483170466883822, + "grad_norm": 6.194875830854703, + "learning_rate": 1.835537910695031e-05, + "loss": 0.8774, + "step": 1931 + }, + { + "epoch": 1.0488599348534202, + "grad_norm": 4.540451693265499, + "learning_rate": 1.8353446323363496e-05, + "loss": 0.8508, + "step": 1932 + }, + { + "epoch": 1.0494028230184582, + "grad_norm": 5.9491142341914225, + "learning_rate": 1.8351512506609133e-05, + "loss": 1.0139, + "step": 1933 + }, + { + "epoch": 1.0499457111834962, + "grad_norm": 6.436571178648067, + "learning_rate": 1.834957765692639e-05, + "loss": 0.9272, + "step": 1934 + }, + { + "epoch": 1.0504885993485342, + "grad_norm": 5.6538429161280535, + "learning_rate": 1.8347641774554573e-05, + "loss": 0.7834, + "step": 1935 + }, + { + "epoch": 1.0510314875135722, + "grad_norm": 6.481163094620501, + "learning_rate": 1.8345704859733123e-05, + "loss": 1.0916, + "step": 1936 + }, + { + "epoch": 1.0515743756786102, + "grad_norm": 6.165149195118496, + "learning_rate": 1.8343766912701588e-05, + "loss": 1.1054, + "step": 1937 + }, + { + "epoch": 1.0521172638436482, + "grad_norm": 7.778083794032891, + "learning_rate": 1.834182793369967e-05, + "loss": 1.297, + "step": 1938 + }, + { + "epoch": 1.0526601520086862, + "grad_norm": 8.871857523729659, + "learning_rate": 1.8339887922967176e-05, + "loss": 1.5749, + "step": 1939 + }, + { + "epoch": 1.0532030401737242, + "grad_norm": 6.563413115451867, + "learning_rate": 1.8337946880744047e-05, + "loss": 1.0271, + "step": 1940 + }, + { + "epoch": 1.0537459283387622, + "grad_norm": 6.5304770661428275, + "learning_rate": 1.833600480727036e-05, + "loss": 0.9236, + "step": 1941 + }, + { + "epoch": 1.0542888165038002, + "grad_norm": 6.802479270907137, + "learning_rate": 1.8334061702786317e-05, + "loss": 0.9076, + "step": 1942 + }, + { + "epoch": 1.0548317046688382, + "grad_norm": 5.92560525391674, + "learning_rate": 1.833211756753224e-05, + "loss": 0.9601, + "step": 1943 + }, + { + "epoch": 1.0553745928338762, + "grad_norm": 7.436926375817581, + "learning_rate": 1.8330172401748584e-05, + "loss": 1.1758, + "step": 1944 + }, + { + "epoch": 1.0559174809989142, + "grad_norm": 6.369742715679364, + "learning_rate": 1.8328226205675927e-05, + "loss": 0.8424, + "step": 1945 + }, + { + "epoch": 1.0564603691639523, + "grad_norm": 7.60632781901004, + "learning_rate": 1.8326278979554976e-05, + "loss": 1.7815, + "step": 1946 + }, + { + "epoch": 1.0570032573289903, + "grad_norm": 8.499242837541805, + "learning_rate": 1.8324330723626578e-05, + "loss": 1.1341, + "step": 1947 + }, + { + "epoch": 1.0575461454940283, + "grad_norm": 6.469082998862602, + "learning_rate": 1.8322381438131686e-05, + "loss": 0.7798, + "step": 1948 + }, + { + "epoch": 1.0580890336590663, + "grad_norm": 5.923622010018089, + "learning_rate": 1.8320431123311388e-05, + "loss": 0.7971, + "step": 1949 + }, + { + "epoch": 1.0586319218241043, + "grad_norm": 6.617488371294151, + "learning_rate": 1.8318479779406914e-05, + "loss": 1.1551, + "step": 1950 + }, + { + "epoch": 1.0591748099891423, + "grad_norm": 6.813827391520987, + "learning_rate": 1.8316527406659604e-05, + "loss": 1.0234, + "step": 1951 + }, + { + "epoch": 1.0597176981541803, + "grad_norm": 7.580867549521637, + "learning_rate": 1.831457400531093e-05, + "loss": 0.7896, + "step": 1952 + }, + { + "epoch": 1.0602605863192183, + "grad_norm": 5.975929968595557, + "learning_rate": 1.8312619575602486e-05, + "loss": 1.1248, + "step": 1953 + }, + { + "epoch": 1.0608034744842563, + "grad_norm": 6.2878616306091955, + "learning_rate": 1.831066411777601e-05, + "loss": 1.0833, + "step": 1954 + }, + { + "epoch": 1.0613463626492943, + "grad_norm": 6.284963948133933, + "learning_rate": 1.8308707632073345e-05, + "loss": 1.0773, + "step": 1955 + }, + { + "epoch": 1.0618892508143323, + "grad_norm": 4.009278799804313, + "learning_rate": 1.830675011873648e-05, + "loss": 0.4782, + "step": 1956 + }, + { + "epoch": 1.0624321389793703, + "grad_norm": 6.648653088879531, + "learning_rate": 1.8304791578007524e-05, + "loss": 0.9761, + "step": 1957 + }, + { + "epoch": 1.0629750271444083, + "grad_norm": 6.867797713608168, + "learning_rate": 1.830283201012871e-05, + "loss": 0.9315, + "step": 1958 + }, + { + "epoch": 1.0635179153094463, + "grad_norm": 5.989268423819992, + "learning_rate": 1.83008714153424e-05, + "loss": 0.8312, + "step": 1959 + }, + { + "epoch": 1.0640608034744843, + "grad_norm": 6.207518262073172, + "learning_rate": 1.8298909793891083e-05, + "loss": 1.1003, + "step": 1960 + }, + { + "epoch": 1.0646036916395223, + "grad_norm": 5.5435710943195815, + "learning_rate": 1.8296947146017373e-05, + "loss": 0.8012, + "step": 1961 + }, + { + "epoch": 1.0651465798045603, + "grad_norm": 8.000995164961113, + "learning_rate": 1.829498347196402e-05, + "loss": 0.9864, + "step": 1962 + }, + { + "epoch": 1.0656894679695983, + "grad_norm": 6.483090321755992, + "learning_rate": 1.829301877197389e-05, + "loss": 1.2823, + "step": 1963 + }, + { + "epoch": 1.0662323561346363, + "grad_norm": 7.147142303433649, + "learning_rate": 1.8291053046289985e-05, + "loss": 0.9185, + "step": 1964 + }, + { + "epoch": 1.0667752442996743, + "grad_norm": 6.000816189695904, + "learning_rate": 1.828908629515542e-05, + "loss": 1.0284, + "step": 1965 + }, + { + "epoch": 1.0673181324647123, + "grad_norm": 6.188793151434744, + "learning_rate": 1.8287118518813453e-05, + "loss": 0.7041, + "step": 1966 + }, + { + "epoch": 1.0678610206297503, + "grad_norm": 7.353722352349769, + "learning_rate": 1.828514971750746e-05, + "loss": 1.0707, + "step": 1967 + }, + { + "epoch": 1.0684039087947883, + "grad_norm": 7.553378116857428, + "learning_rate": 1.8283179891480944e-05, + "loss": 1.1426, + "step": 1968 + }, + { + "epoch": 1.0689467969598263, + "grad_norm": 7.134594743515001, + "learning_rate": 1.828120904097754e-05, + "loss": 0.9583, + "step": 1969 + }, + { + "epoch": 1.0694896851248643, + "grad_norm": 5.680617024729498, + "learning_rate": 1.8279237166241004e-05, + "loss": 0.9632, + "step": 1970 + }, + { + "epoch": 1.0700325732899023, + "grad_norm": 5.932910663063727, + "learning_rate": 1.8277264267515218e-05, + "loss": 0.9984, + "step": 1971 + }, + { + "epoch": 1.0705754614549403, + "grad_norm": 7.833159555663921, + "learning_rate": 1.8275290345044198e-05, + "loss": 0.8038, + "step": 1972 + }, + { + "epoch": 1.0711183496199783, + "grad_norm": 6.429988603781813, + "learning_rate": 1.8273315399072076e-05, + "loss": 0.9125, + "step": 1973 + }, + { + "epoch": 1.0716612377850163, + "grad_norm": 7.955595430340166, + "learning_rate": 1.827133942984312e-05, + "loss": 0.8697, + "step": 1974 + }, + { + "epoch": 1.0722041259500543, + "grad_norm": 5.983514211221058, + "learning_rate": 1.826936243760172e-05, + "loss": 0.8155, + "step": 1975 + }, + { + "epoch": 1.0727470141150923, + "grad_norm": 6.19913176090901, + "learning_rate": 1.8267384422592398e-05, + "loss": 1.3207, + "step": 1976 + }, + { + "epoch": 1.0732899022801303, + "grad_norm": 6.467400855860047, + "learning_rate": 1.8265405385059792e-05, + "loss": 1.1279, + "step": 1977 + }, + { + "epoch": 1.0738327904451683, + "grad_norm": 6.515150741776156, + "learning_rate": 1.8263425325248675e-05, + "loss": 1.372, + "step": 1978 + }, + { + "epoch": 1.0743756786102063, + "grad_norm": 6.181497786496646, + "learning_rate": 1.8261444243403945e-05, + "loss": 0.8571, + "step": 1979 + }, + { + "epoch": 1.0749185667752443, + "grad_norm": 7.637816753034266, + "learning_rate": 1.8259462139770624e-05, + "loss": 1.4861, + "step": 1980 + }, + { + "epoch": 1.0754614549402823, + "grad_norm": 7.0094523855293485, + "learning_rate": 1.825747901459386e-05, + "loss": 1.0113, + "step": 1981 + }, + { + "epoch": 1.0760043431053203, + "grad_norm": 7.6230550634556655, + "learning_rate": 1.8255494868118933e-05, + "loss": 1.3818, + "step": 1982 + }, + { + "epoch": 1.0765472312703583, + "grad_norm": 6.946643864799888, + "learning_rate": 1.8253509700591242e-05, + "loss": 1.1384, + "step": 1983 + }, + { + "epoch": 1.0770901194353963, + "grad_norm": 7.288235795183878, + "learning_rate": 1.825152351225632e-05, + "loss": 1.1288, + "step": 1984 + }, + { + "epoch": 1.0776330076004343, + "grad_norm": 8.872287367521855, + "learning_rate": 1.8249536303359816e-05, + "loss": 1.6016, + "step": 1985 + }, + { + "epoch": 1.0781758957654723, + "grad_norm": 5.75562973378046, + "learning_rate": 1.8247548074147515e-05, + "loss": 1.1803, + "step": 1986 + }, + { + "epoch": 1.0787187839305103, + "grad_norm": 6.377752945158632, + "learning_rate": 1.824555882486532e-05, + "loss": 0.6991, + "step": 1987 + }, + { + "epoch": 1.0792616720955484, + "grad_norm": 7.178965937765225, + "learning_rate": 1.8243568555759274e-05, + "loss": 0.988, + "step": 1988 + }, + { + "epoch": 1.0798045602605864, + "grad_norm": 6.0869015160234525, + "learning_rate": 1.824157726707553e-05, + "loss": 0.8587, + "step": 1989 + }, + { + "epoch": 1.0803474484256244, + "grad_norm": 5.9113747723637395, + "learning_rate": 1.823958495906037e-05, + "loss": 0.8815, + "step": 1990 + }, + { + "epoch": 1.0808903365906624, + "grad_norm": 7.84172234151508, + "learning_rate": 1.8237591631960218e-05, + "loss": 1.2621, + "step": 1991 + }, + { + "epoch": 1.0814332247557004, + "grad_norm": 7.188505727546616, + "learning_rate": 1.8235597286021597e-05, + "loss": 0.5991, + "step": 1992 + }, + { + "epoch": 1.0819761129207384, + "grad_norm": 6.358900329219305, + "learning_rate": 1.823360192149118e-05, + "loss": 1.3325, + "step": 1993 + }, + { + "epoch": 1.0825190010857764, + "grad_norm": 6.545296131842583, + "learning_rate": 1.8231605538615756e-05, + "loss": 1.1091, + "step": 1994 + }, + { + "epoch": 1.0830618892508144, + "grad_norm": 6.708966367161533, + "learning_rate": 1.8229608137642238e-05, + "loss": 1.1352, + "step": 1995 + }, + { + "epoch": 1.0836047774158524, + "grad_norm": 6.651776068492304, + "learning_rate": 1.822760971881767e-05, + "loss": 0.6818, + "step": 1996 + }, + { + "epoch": 1.0841476655808904, + "grad_norm": 4.745518059898734, + "learning_rate": 1.8225610282389222e-05, + "loss": 0.9403, + "step": 1997 + }, + { + "epoch": 1.0846905537459284, + "grad_norm": 6.526942367542791, + "learning_rate": 1.8223609828604184e-05, + "loss": 1.0543, + "step": 1998 + }, + { + "epoch": 1.0852334419109664, + "grad_norm": 5.331318824902818, + "learning_rate": 1.8221608357709973e-05, + "loss": 0.6626, + "step": 1999 + }, + { + "epoch": 1.0857763300760044, + "grad_norm": 6.640324946305343, + "learning_rate": 1.8219605869954134e-05, + "loss": 1.0166, + "step": 2000 + }, + { + "epoch": 1.0863192182410424, + "grad_norm": 5.684199608008352, + "learning_rate": 1.8217602365584352e-05, + "loss": 0.6796, + "step": 2001 + }, + { + "epoch": 1.0868621064060804, + "grad_norm": 5.429738239817468, + "learning_rate": 1.8215597844848403e-05, + "loss": 0.6783, + "step": 2002 + }, + { + "epoch": 1.0874049945711184, + "grad_norm": 6.1832795338939235, + "learning_rate": 1.821359230799422e-05, + "loss": 1.4244, + "step": 2003 + }, + { + "epoch": 1.0879478827361564, + "grad_norm": 7.118410329483546, + "learning_rate": 1.8211585755269852e-05, + "loss": 1.0274, + "step": 2004 + }, + { + "epoch": 1.0884907709011944, + "grad_norm": 5.5061218005078265, + "learning_rate": 1.820957818692347e-05, + "loss": 0.7143, + "step": 2005 + }, + { + "epoch": 1.0890336590662324, + "grad_norm": 5.518412565715972, + "learning_rate": 1.8207569603203373e-05, + "loss": 1.3158, + "step": 2006 + }, + { + "epoch": 1.0895765472312704, + "grad_norm": 5.47000403426222, + "learning_rate": 1.820556000435799e-05, + "loss": 1.0694, + "step": 2007 + }, + { + "epoch": 1.0901194353963084, + "grad_norm": 7.469961154104622, + "learning_rate": 1.820354939063586e-05, + "loss": 1.0702, + "step": 2008 + }, + { + "epoch": 1.0906623235613464, + "grad_norm": 5.889961089255285, + "learning_rate": 1.8201537762285674e-05, + "loss": 1.0598, + "step": 2009 + }, + { + "epoch": 1.0912052117263844, + "grad_norm": 6.659595389259375, + "learning_rate": 1.8199525119556226e-05, + "loss": 0.8456, + "step": 2010 + }, + { + "epoch": 1.0917480998914224, + "grad_norm": 6.328157257432415, + "learning_rate": 1.8197511462696443e-05, + "loss": 0.9763, + "step": 2011 + }, + { + "epoch": 1.0922909880564604, + "grad_norm": 5.916734142619132, + "learning_rate": 1.8195496791955373e-05, + "loss": 1.1194, + "step": 2012 + }, + { + "epoch": 1.0928338762214984, + "grad_norm": 5.904087624332602, + "learning_rate": 1.8193481107582203e-05, + "loss": 0.756, + "step": 2013 + }, + { + "epoch": 1.0933767643865364, + "grad_norm": 6.284745154848233, + "learning_rate": 1.8191464409826227e-05, + "loss": 0.7931, + "step": 2014 + }, + { + "epoch": 1.0939196525515744, + "grad_norm": 7.838306243079945, + "learning_rate": 1.8189446698936878e-05, + "loss": 1.3635, + "step": 2015 + }, + { + "epoch": 1.0944625407166124, + "grad_norm": 6.3405758352546995, + "learning_rate": 1.818742797516371e-05, + "loss": 0.5365, + "step": 2016 + }, + { + "epoch": 1.0950054288816504, + "grad_norm": 8.315910084823138, + "learning_rate": 1.8185408238756405e-05, + "loss": 1.6592, + "step": 2017 + }, + { + "epoch": 1.0955483170466884, + "grad_norm": 5.205053142209431, + "learning_rate": 1.8183387489964762e-05, + "loss": 0.6416, + "step": 2018 + }, + { + "epoch": 1.0960912052117264, + "grad_norm": 7.018373673519811, + "learning_rate": 1.8181365729038706e-05, + "loss": 1.0814, + "step": 2019 + }, + { + "epoch": 1.0966340933767644, + "grad_norm": 5.705860306884454, + "learning_rate": 1.8179342956228307e-05, + "loss": 0.5806, + "step": 2020 + }, + { + "epoch": 1.0971769815418024, + "grad_norm": 7.077662290720525, + "learning_rate": 1.8177319171783728e-05, + "loss": 0.8969, + "step": 2021 + }, + { + "epoch": 1.0977198697068404, + "grad_norm": 6.540901861387866, + "learning_rate": 1.8175294375955284e-05, + "loss": 0.7982, + "step": 2022 + }, + { + "epoch": 1.0982627578718784, + "grad_norm": 7.283213833491601, + "learning_rate": 1.81732685689934e-05, + "loss": 1.2659, + "step": 2023 + }, + { + "epoch": 1.0988056460369164, + "grad_norm": 7.016589270114781, + "learning_rate": 1.8171241751148633e-05, + "loss": 1.0011, + "step": 2024 + }, + { + "epoch": 1.0993485342019544, + "grad_norm": 5.724817250167612, + "learning_rate": 1.8169213922671666e-05, + "loss": 0.8967, + "step": 2025 + }, + { + "epoch": 1.0998914223669924, + "grad_norm": 7.431207902407107, + "learning_rate": 1.81671850838133e-05, + "loss": 0.9445, + "step": 2026 + }, + { + "epoch": 1.1004343105320304, + "grad_norm": 6.9152342645144955, + "learning_rate": 1.816515523482447e-05, + "loss": 0.8187, + "step": 2027 + }, + { + "epoch": 1.1009771986970684, + "grad_norm": 7.045028624610342, + "learning_rate": 1.816312437595622e-05, + "loss": 0.7323, + "step": 2028 + }, + { + "epoch": 1.1015200868621065, + "grad_norm": 6.63265573559803, + "learning_rate": 1.816109250745974e-05, + "loss": 0.6689, + "step": 2029 + }, + { + "epoch": 1.1020629750271445, + "grad_norm": 8.605675554379312, + "learning_rate": 1.8159059629586333e-05, + "loss": 1.3733, + "step": 2030 + }, + { + "epoch": 1.1026058631921825, + "grad_norm": 7.481808301387036, + "learning_rate": 1.8157025742587426e-05, + "loss": 0.9767, + "step": 2031 + }, + { + "epoch": 1.1031487513572205, + "grad_norm": 7.113214270917786, + "learning_rate": 1.8154990846714575e-05, + "loss": 0.6365, + "step": 2032 + }, + { + "epoch": 1.1036916395222585, + "grad_norm": 7.672152203903941, + "learning_rate": 1.8152954942219462e-05, + "loss": 1.3314, + "step": 2033 + }, + { + "epoch": 1.1042345276872965, + "grad_norm": 7.533257156973554, + "learning_rate": 1.8150918029353885e-05, + "loss": 0.967, + "step": 2034 + }, + { + "epoch": 1.1047774158523345, + "grad_norm": 6.469759228237502, + "learning_rate": 1.8148880108369775e-05, + "loss": 1.348, + "step": 2035 + }, + { + "epoch": 1.1053203040173725, + "grad_norm": 6.620692148711997, + "learning_rate": 1.8146841179519186e-05, + "loss": 0.7538, + "step": 2036 + }, + { + "epoch": 1.1058631921824105, + "grad_norm": 6.974705446587608, + "learning_rate": 1.8144801243054297e-05, + "loss": 1.3664, + "step": 2037 + }, + { + "epoch": 1.1064060803474485, + "grad_norm": 8.047700738112871, + "learning_rate": 1.8142760299227408e-05, + "loss": 1.0466, + "step": 2038 + }, + { + "epoch": 1.1069489685124865, + "grad_norm": 5.469412128245213, + "learning_rate": 1.814071834829095e-05, + "loss": 0.5464, + "step": 2039 + }, + { + "epoch": 1.1074918566775245, + "grad_norm": 5.732997692484992, + "learning_rate": 1.813867539049747e-05, + "loss": 0.7307, + "step": 2040 + }, + { + "epoch": 1.1080347448425625, + "grad_norm": 8.89935072196041, + "learning_rate": 1.8136631426099646e-05, + "loss": 1.8231, + "step": 2041 + }, + { + "epoch": 1.1085776330076005, + "grad_norm": 7.0036806828899465, + "learning_rate": 1.813458645535028e-05, + "loss": 1.063, + "step": 2042 + }, + { + "epoch": 1.1091205211726385, + "grad_norm": 6.013891951698058, + "learning_rate": 1.8132540478502297e-05, + "loss": 0.9836, + "step": 2043 + }, + { + "epoch": 1.1096634093376765, + "grad_norm": 4.234811309290956, + "learning_rate": 1.813049349580875e-05, + "loss": 0.3708, + "step": 2044 + }, + { + "epoch": 1.1102062975027145, + "grad_norm": 6.259459841046158, + "learning_rate": 1.8128445507522806e-05, + "loss": 0.8463, + "step": 2045 + }, + { + "epoch": 1.1107491856677525, + "grad_norm": 7.9568342619649775, + "learning_rate": 1.8126396513897764e-05, + "loss": 1.0563, + "step": 2046 + }, + { + "epoch": 1.1112920738327905, + "grad_norm": 6.937848428371049, + "learning_rate": 1.8124346515187056e-05, + "loss": 0.6359, + "step": 2047 + }, + { + "epoch": 1.1118349619978285, + "grad_norm": 7.672632972051955, + "learning_rate": 1.8122295511644218e-05, + "loss": 0.657, + "step": 2048 + }, + { + "epoch": 1.1123778501628665, + "grad_norm": 6.450214026534184, + "learning_rate": 1.8120243503522924e-05, + "loss": 0.7279, + "step": 2049 + }, + { + "epoch": 1.1129207383279045, + "grad_norm": 6.628216836060006, + "learning_rate": 1.8118190491076978e-05, + "loss": 0.5804, + "step": 2050 + }, + { + "epoch": 1.1134636264929425, + "grad_norm": 6.698074316145226, + "learning_rate": 1.8116136474560288e-05, + "loss": 1.011, + "step": 2051 + }, + { + "epoch": 1.1140065146579805, + "grad_norm": 5.4126983291571795, + "learning_rate": 1.8114081454226905e-05, + "loss": 0.8564, + "step": 2052 + }, + { + "epoch": 1.1145494028230185, + "grad_norm": 8.596797841329312, + "learning_rate": 1.8112025430331e-05, + "loss": 1.1924, + "step": 2053 + }, + { + "epoch": 1.1150922909880565, + "grad_norm": 7.416912053391948, + "learning_rate": 1.8109968403126856e-05, + "loss": 0.9894, + "step": 2054 + }, + { + "epoch": 1.1156351791530945, + "grad_norm": 7.434308789484601, + "learning_rate": 1.8107910372868898e-05, + "loss": 1.1544, + "step": 2055 + }, + { + "epoch": 1.1161780673181325, + "grad_norm": 8.325689788099726, + "learning_rate": 1.8105851339811663e-05, + "loss": 0.9994, + "step": 2056 + }, + { + "epoch": 1.1167209554831705, + "grad_norm": 8.384964960696278, + "learning_rate": 1.8103791304209813e-05, + "loss": 1.1768, + "step": 2057 + }, + { + "epoch": 1.1172638436482085, + "grad_norm": 7.044535285708613, + "learning_rate": 1.810173026631814e-05, + "loss": 1.1801, + "step": 2058 + }, + { + "epoch": 1.1178067318132465, + "grad_norm": 6.665810359686943, + "learning_rate": 1.8099668226391552e-05, + "loss": 0.7865, + "step": 2059 + }, + { + "epoch": 1.1183496199782845, + "grad_norm": 7.395697719004186, + "learning_rate": 1.8097605184685093e-05, + "loss": 1.1357, + "step": 2060 + }, + { + "epoch": 1.1188925081433225, + "grad_norm": 7.1495559268215185, + "learning_rate": 1.809554114145392e-05, + "loss": 1.1507, + "step": 2061 + }, + { + "epoch": 1.1194353963083605, + "grad_norm": 7.877063869804663, + "learning_rate": 1.8093476096953315e-05, + "loss": 1.634, + "step": 2062 + }, + { + "epoch": 1.1199782844733985, + "grad_norm": 6.528060830825324, + "learning_rate": 1.809141005143869e-05, + "loss": 0.697, + "step": 2063 + }, + { + "epoch": 1.1205211726384365, + "grad_norm": 6.5949576624011, + "learning_rate": 1.808934300516557e-05, + "loss": 1.1506, + "step": 2064 + }, + { + "epoch": 1.1210640608034745, + "grad_norm": 6.787348272435354, + "learning_rate": 1.8087274958389612e-05, + "loss": 0.9986, + "step": 2065 + }, + { + "epoch": 1.1216069489685125, + "grad_norm": 6.310955809912108, + "learning_rate": 1.8085205911366602e-05, + "loss": 0.89, + "step": 2066 + }, + { + "epoch": 1.1221498371335505, + "grad_norm": 6.8487503881448895, + "learning_rate": 1.8083135864352442e-05, + "loss": 1.1061, + "step": 2067 + }, + { + "epoch": 1.1226927252985885, + "grad_norm": 5.329494375939551, + "learning_rate": 1.808106481760315e-05, + "loss": 1.0011, + "step": 2068 + }, + { + "epoch": 1.1232356134636265, + "grad_norm": 6.286005313228098, + "learning_rate": 1.8078992771374886e-05, + "loss": 0.8399, + "step": 2069 + }, + { + "epoch": 1.1237785016286646, + "grad_norm": 6.324896503266866, + "learning_rate": 1.8076919725923917e-05, + "loss": 1.2716, + "step": 2070 + }, + { + "epoch": 1.1243213897937026, + "grad_norm": 5.663105266452551, + "learning_rate": 1.8074845681506644e-05, + "loss": 0.9943, + "step": 2071 + }, + { + "epoch": 1.1248642779587406, + "grad_norm": 5.175572461557621, + "learning_rate": 1.807277063837959e-05, + "loss": 0.6319, + "step": 2072 + }, + { + "epoch": 1.1254071661237786, + "grad_norm": 7.297487212372251, + "learning_rate": 1.8070694596799397e-05, + "loss": 0.7533, + "step": 2073 + }, + { + "epoch": 1.1259500542888166, + "grad_norm": 9.046578233072974, + "learning_rate": 1.806861755702283e-05, + "loss": 0.9991, + "step": 2074 + }, + { + "epoch": 1.1264929424538546, + "grad_norm": 4.516511292735119, + "learning_rate": 1.8066539519306786e-05, + "loss": 0.6393, + "step": 2075 + }, + { + "epoch": 1.1270358306188926, + "grad_norm": 5.16623460840236, + "learning_rate": 1.8064460483908283e-05, + "loss": 0.7083, + "step": 2076 + }, + { + "epoch": 1.1275787187839306, + "grad_norm": 10.7323945119629, + "learning_rate": 1.8062380451084445e-05, + "loss": 2.4193, + "step": 2077 + }, + { + "epoch": 1.1281216069489686, + "grad_norm": 6.438876475868993, + "learning_rate": 1.8060299421092554e-05, + "loss": 0.8468, + "step": 2078 + }, + { + "epoch": 1.1286644951140066, + "grad_norm": 8.437791617086905, + "learning_rate": 1.8058217394189976e-05, + "loss": 1.4895, + "step": 2079 + }, + { + "epoch": 1.1292073832790446, + "grad_norm": 5.237528459736948, + "learning_rate": 1.805613437063423e-05, + "loss": 0.643, + "step": 2080 + }, + { + "epoch": 1.1297502714440826, + "grad_norm": 6.010151293762264, + "learning_rate": 1.8054050350682947e-05, + "loss": 0.9001, + "step": 2081 + }, + { + "epoch": 1.1302931596091206, + "grad_norm": 4.816849804237718, + "learning_rate": 1.805196533459388e-05, + "loss": 0.6349, + "step": 2082 + }, + { + "epoch": 1.1308360477741586, + "grad_norm": 6.872496495387926, + "learning_rate": 1.8049879322624906e-05, + "loss": 0.8808, + "step": 2083 + }, + { + "epoch": 1.1313789359391966, + "grad_norm": 5.86067647581003, + "learning_rate": 1.804779231503403e-05, + "loss": 0.6811, + "step": 2084 + }, + { + "epoch": 1.1319218241042346, + "grad_norm": 6.073217718554074, + "learning_rate": 1.8045704312079376e-05, + "loss": 1.0086, + "step": 2085 + }, + { + "epoch": 1.1324647122692726, + "grad_norm": 7.519215180241719, + "learning_rate": 1.804361531401918e-05, + "loss": 0.8815, + "step": 2086 + }, + { + "epoch": 1.1330076004343106, + "grad_norm": 6.49012361077388, + "learning_rate": 1.8041525321111835e-05, + "loss": 1.0129, + "step": 2087 + }, + { + "epoch": 1.1335504885993486, + "grad_norm": 5.938773374319115, + "learning_rate": 1.8039434333615814e-05, + "loss": 0.8485, + "step": 2088 + }, + { + "epoch": 1.1340933767643866, + "grad_norm": 5.607636698742016, + "learning_rate": 1.8037342351789743e-05, + "loss": 0.6209, + "step": 2089 + }, + { + "epoch": 1.1346362649294246, + "grad_norm": 5.522935966563286, + "learning_rate": 1.803524937589236e-05, + "loss": 0.6576, + "step": 2090 + }, + { + "epoch": 1.1351791530944626, + "grad_norm": 8.130974168982206, + "learning_rate": 1.8033155406182533e-05, + "loss": 1.0676, + "step": 2091 + }, + { + "epoch": 1.1357220412595006, + "grad_norm": 6.027276275319545, + "learning_rate": 1.803106044291924e-05, + "loss": 0.9853, + "step": 2092 + }, + { + "epoch": 1.1362649294245386, + "grad_norm": 4.645985748184023, + "learning_rate": 1.8028964486361586e-05, + "loss": 0.7505, + "step": 2093 + }, + { + "epoch": 1.1368078175895766, + "grad_norm": 6.820894579082682, + "learning_rate": 1.8026867536768816e-05, + "loss": 0.9358, + "step": 2094 + }, + { + "epoch": 1.1373507057546146, + "grad_norm": 3.9530727606720872, + "learning_rate": 1.802476959440027e-05, + "loss": 0.7736, + "step": 2095 + }, + { + "epoch": 1.1378935939196526, + "grad_norm": 7.650275740883737, + "learning_rate": 1.8022670659515432e-05, + "loss": 1.4832, + "step": 2096 + }, + { + "epoch": 1.1384364820846906, + "grad_norm": 6.94341164115752, + "learning_rate": 1.80205707323739e-05, + "loss": 1.0336, + "step": 2097 + }, + { + "epoch": 1.1389793702497286, + "grad_norm": 5.608528053150328, + "learning_rate": 1.8018469813235403e-05, + "loss": 0.829, + "step": 2098 + }, + { + "epoch": 1.1395222584147666, + "grad_norm": 5.2191336182398995, + "learning_rate": 1.8016367902359776e-05, + "loss": 0.9708, + "step": 2099 + }, + { + "epoch": 1.1400651465798046, + "grad_norm": 6.773346810380977, + "learning_rate": 1.801426500000699e-05, + "loss": 0.6755, + "step": 2100 + }, + { + "epoch": 1.1406080347448426, + "grad_norm": 7.292992073718992, + "learning_rate": 1.8012161106437137e-05, + "loss": 0.8842, + "step": 2101 + }, + { + "epoch": 1.1411509229098806, + "grad_norm": 6.699434744454233, + "learning_rate": 1.8010056221910427e-05, + "loss": 1.0385, + "step": 2102 + }, + { + "epoch": 1.1416938110749186, + "grad_norm": 6.565299107509569, + "learning_rate": 1.8007950346687198e-05, + "loss": 0.6557, + "step": 2103 + }, + { + "epoch": 1.1422366992399566, + "grad_norm": 7.2741514068627025, + "learning_rate": 1.800584348102791e-05, + "loss": 1.0347, + "step": 2104 + }, + { + "epoch": 1.1427795874049946, + "grad_norm": 6.467241522225306, + "learning_rate": 1.800373562519314e-05, + "loss": 0.8026, + "step": 2105 + }, + { + "epoch": 1.1433224755700326, + "grad_norm": 6.109318081193457, + "learning_rate": 1.800162677944359e-05, + "loss": 0.9182, + "step": 2106 + }, + { + "epoch": 1.1438653637350706, + "grad_norm": 7.136347555199431, + "learning_rate": 1.7999516944040087e-05, + "loss": 1.2743, + "step": 2107 + }, + { + "epoch": 1.1444082519001086, + "grad_norm": 7.791284332349926, + "learning_rate": 1.7997406119243582e-05, + "loss": 0.9738, + "step": 2108 + }, + { + "epoch": 1.1449511400651466, + "grad_norm": 7.315957477694666, + "learning_rate": 1.7995294305315137e-05, + "loss": 1.0674, + "step": 2109 + }, + { + "epoch": 1.1454940282301846, + "grad_norm": 6.305984776426275, + "learning_rate": 1.7993181502515957e-05, + "loss": 0.8204, + "step": 2110 + }, + { + "epoch": 1.1460369163952226, + "grad_norm": 5.824114865699788, + "learning_rate": 1.7991067711107345e-05, + "loss": 0.7953, + "step": 2111 + }, + { + "epoch": 1.1465798045602607, + "grad_norm": 7.4284096250312786, + "learning_rate": 1.798895293135074e-05, + "loss": 0.7349, + "step": 2112 + }, + { + "epoch": 1.1471226927252987, + "grad_norm": 6.415949338549894, + "learning_rate": 1.798683716350771e-05, + "loss": 0.8346, + "step": 2113 + }, + { + "epoch": 1.1476655808903367, + "grad_norm": 6.360883520037482, + "learning_rate": 1.7984720407839925e-05, + "loss": 0.7966, + "step": 2114 + }, + { + "epoch": 1.1482084690553747, + "grad_norm": 6.703196393264975, + "learning_rate": 1.7982602664609198e-05, + "loss": 1.3341, + "step": 2115 + }, + { + "epoch": 1.1487513572204127, + "grad_norm": 8.115384707506347, + "learning_rate": 1.798048393407745e-05, + "loss": 1.0284, + "step": 2116 + }, + { + "epoch": 1.1492942453854507, + "grad_norm": 8.519166176194911, + "learning_rate": 1.797836421650673e-05, + "loss": 1.4075, + "step": 2117 + }, + { + "epoch": 1.1498371335504887, + "grad_norm": 7.120345871314324, + "learning_rate": 1.7976243512159207e-05, + "loss": 1.1933, + "step": 2118 + }, + { + "epoch": 1.1503800217155267, + "grad_norm": 5.887743000051609, + "learning_rate": 1.7974121821297178e-05, + "loss": 0.9143, + "step": 2119 + }, + { + "epoch": 1.1509229098805647, + "grad_norm": 6.264158167802545, + "learning_rate": 1.797199914418305e-05, + "loss": 0.805, + "step": 2120 + }, + { + "epoch": 1.1514657980456027, + "grad_norm": 7.747345696360016, + "learning_rate": 1.7969875481079363e-05, + "loss": 1.1374, + "step": 2121 + }, + { + "epoch": 1.1520086862106407, + "grad_norm": 5.932381639003466, + "learning_rate": 1.7967750832248774e-05, + "loss": 0.7801, + "step": 2122 + }, + { + "epoch": 1.1525515743756787, + "grad_norm": 5.646019326027894, + "learning_rate": 1.7965625197954064e-05, + "loss": 0.5053, + "step": 2123 + }, + { + "epoch": 1.1530944625407167, + "grad_norm": 6.085604242778934, + "learning_rate": 1.7963498578458135e-05, + "loss": 0.6711, + "step": 2124 + }, + { + "epoch": 1.1536373507057547, + "grad_norm": 6.390132028289734, + "learning_rate": 1.7961370974024008e-05, + "loss": 0.9897, + "step": 2125 + }, + { + "epoch": 1.1541802388707927, + "grad_norm": 5.052558465006266, + "learning_rate": 1.795924238491483e-05, + "loss": 0.8742, + "step": 2126 + }, + { + "epoch": 1.1547231270358307, + "grad_norm": 5.86713723099216, + "learning_rate": 1.7957112811393867e-05, + "loss": 0.965, + "step": 2127 + }, + { + "epoch": 1.1552660152008687, + "grad_norm": 6.092033075473227, + "learning_rate": 1.7954982253724515e-05, + "loss": 1.104, + "step": 2128 + }, + { + "epoch": 1.1558089033659067, + "grad_norm": 7.051795930424219, + "learning_rate": 1.7952850712170278e-05, + "loss": 1.0918, + "step": 2129 + }, + { + "epoch": 1.1563517915309447, + "grad_norm": 5.613114568974867, + "learning_rate": 1.7950718186994788e-05, + "loss": 0.6698, + "step": 2130 + }, + { + "epoch": 1.1568946796959827, + "grad_norm": 5.533439816776461, + "learning_rate": 1.7948584678461802e-05, + "loss": 1.4395, + "step": 2131 + }, + { + "epoch": 1.1574375678610207, + "grad_norm": 7.73758559868109, + "learning_rate": 1.7946450186835195e-05, + "loss": 1.0806, + "step": 2132 + }, + { + "epoch": 1.1579804560260587, + "grad_norm": 5.982378084297589, + "learning_rate": 1.7944314712378966e-05, + "loss": 1.0428, + "step": 2133 + }, + { + "epoch": 1.1585233441910967, + "grad_norm": 5.604421383798244, + "learning_rate": 1.7942178255357227e-05, + "loss": 0.9411, + "step": 2134 + }, + { + "epoch": 1.1590662323561347, + "grad_norm": 6.5637069585312435, + "learning_rate": 1.794004081603423e-05, + "loss": 0.8285, + "step": 2135 + }, + { + "epoch": 1.1596091205211727, + "grad_norm": 5.567535422744373, + "learning_rate": 1.7937902394674326e-05, + "loss": 1.0185, + "step": 2136 + }, + { + "epoch": 1.1601520086862107, + "grad_norm": 7.898198443147329, + "learning_rate": 1.7935762991542004e-05, + "loss": 0.8878, + "step": 2137 + }, + { + "epoch": 1.1606948968512487, + "grad_norm": 5.12407458777131, + "learning_rate": 1.7933622606901865e-05, + "loss": 0.9174, + "step": 2138 + }, + { + "epoch": 1.1612377850162867, + "grad_norm": 4.480165604953252, + "learning_rate": 1.7931481241018642e-05, + "loss": 0.9424, + "step": 2139 + }, + { + "epoch": 1.1617806731813247, + "grad_norm": 7.6708161194386495, + "learning_rate": 1.7929338894157173e-05, + "loss": 0.9875, + "step": 2140 + }, + { + "epoch": 1.1623235613463627, + "grad_norm": 9.749916559657342, + "learning_rate": 1.7927195566582435e-05, + "loss": 1.7168, + "step": 2141 + }, + { + "epoch": 1.1628664495114007, + "grad_norm": 5.50355689991458, + "learning_rate": 1.7925051258559516e-05, + "loss": 0.7957, + "step": 2142 + }, + { + "epoch": 1.1634093376764387, + "grad_norm": 5.910752363178181, + "learning_rate": 1.7922905970353627e-05, + "loss": 1.0525, + "step": 2143 + }, + { + "epoch": 1.1639522258414767, + "grad_norm": 6.523454927461212, + "learning_rate": 1.7920759702230098e-05, + "loss": 0.9938, + "step": 2144 + }, + { + "epoch": 1.1644951140065147, + "grad_norm": 6.963057538162332, + "learning_rate": 1.7918612454454387e-05, + "loss": 0.8677, + "step": 2145 + }, + { + "epoch": 1.1650380021715527, + "grad_norm": 6.55011455398606, + "learning_rate": 1.7916464227292067e-05, + "loss": 0.7268, + "step": 2146 + }, + { + "epoch": 1.1655808903365907, + "grad_norm": 6.47957363507043, + "learning_rate": 1.7914315021008836e-05, + "loss": 0.7449, + "step": 2147 + }, + { + "epoch": 1.1661237785016287, + "grad_norm": 6.574809242111077, + "learning_rate": 1.7912164835870506e-05, + "loss": 0.6999, + "step": 2148 + }, + { + "epoch": 1.1666666666666667, + "grad_norm": 8.307297080129953, + "learning_rate": 1.791001367214302e-05, + "loss": 1.0402, + "step": 2149 + }, + { + "epoch": 1.1672095548317047, + "grad_norm": 8.01469917888599, + "learning_rate": 1.7907861530092438e-05, + "loss": 1.1024, + "step": 2150 + }, + { + "epoch": 1.1677524429967427, + "grad_norm": 7.336381076899337, + "learning_rate": 1.7905708409984938e-05, + "loss": 0.9112, + "step": 2151 + }, + { + "epoch": 1.1682953311617807, + "grad_norm": 7.872105686379811, + "learning_rate": 1.7903554312086823e-05, + "loss": 0.9002, + "step": 2152 + }, + { + "epoch": 1.1688382193268188, + "grad_norm": 5.842761146463113, + "learning_rate": 1.7901399236664514e-05, + "loss": 0.9457, + "step": 2153 + }, + { + "epoch": 1.1693811074918568, + "grad_norm": 5.8100183378475005, + "learning_rate": 1.7899243183984552e-05, + "loss": 0.9675, + "step": 2154 + }, + { + "epoch": 1.1699239956568948, + "grad_norm": 5.208689262592845, + "learning_rate": 1.7897086154313604e-05, + "loss": 0.7562, + "step": 2155 + }, + { + "epoch": 1.1704668838219328, + "grad_norm": 7.695958654055186, + "learning_rate": 1.7894928147918458e-05, + "loss": 0.9995, + "step": 2156 + }, + { + "epoch": 1.1710097719869708, + "grad_norm": 5.8108200610880525, + "learning_rate": 1.789276916506601e-05, + "loss": 0.8179, + "step": 2157 + }, + { + "epoch": 1.1715526601520088, + "grad_norm": 5.776078861908261, + "learning_rate": 1.7890609206023298e-05, + "loss": 0.6612, + "step": 2158 + }, + { + "epoch": 1.1720955483170468, + "grad_norm": 6.458218841014727, + "learning_rate": 1.7888448271057463e-05, + "loss": 0.8419, + "step": 2159 + }, + { + "epoch": 1.1726384364820848, + "grad_norm": 7.236506499419343, + "learning_rate": 1.7886286360435772e-05, + "loss": 0.6336, + "step": 2160 + }, + { + "epoch": 1.1731813246471228, + "grad_norm": 7.20552304926134, + "learning_rate": 1.7884123474425614e-05, + "loss": 0.9157, + "step": 2161 + }, + { + "epoch": 1.1737242128121608, + "grad_norm": 8.090329675463652, + "learning_rate": 1.78819596132945e-05, + "loss": 1.6894, + "step": 2162 + }, + { + "epoch": 1.1742671009771988, + "grad_norm": 7.500510561892478, + "learning_rate": 1.787979477731006e-05, + "loss": 0.9469, + "step": 2163 + }, + { + "epoch": 1.1748099891422368, + "grad_norm": 7.6126459328076415, + "learning_rate": 1.7877628966740044e-05, + "loss": 0.9179, + "step": 2164 + }, + { + "epoch": 1.1753528773072748, + "grad_norm": 7.065751172246334, + "learning_rate": 1.787546218185232e-05, + "loss": 0.9649, + "step": 2165 + }, + { + "epoch": 1.1758957654723128, + "grad_norm": 6.923487399525626, + "learning_rate": 1.787329442291488e-05, + "loss": 1.0791, + "step": 2166 + }, + { + "epoch": 1.1764386536373508, + "grad_norm": 7.614292041669914, + "learning_rate": 1.7871125690195843e-05, + "loss": 1.2367, + "step": 2167 + }, + { + "epoch": 1.1769815418023888, + "grad_norm": 8.560468932577576, + "learning_rate": 1.7868955983963434e-05, + "loss": 1.0876, + "step": 2168 + }, + { + "epoch": 1.1775244299674268, + "grad_norm": 6.209836424501219, + "learning_rate": 1.786678530448601e-05, + "loss": 0.6251, + "step": 2169 + }, + { + "epoch": 1.1780673181324648, + "grad_norm": 5.926645140708837, + "learning_rate": 1.7864613652032035e-05, + "loss": 0.5939, + "step": 2170 + }, + { + "epoch": 1.1786102062975028, + "grad_norm": 5.460593849347875, + "learning_rate": 1.7862441026870114e-05, + "loss": 0.8769, + "step": 2171 + }, + { + "epoch": 1.1791530944625408, + "grad_norm": 8.063368903057139, + "learning_rate": 1.7860267429268954e-05, + "loss": 1.1492, + "step": 2172 + }, + { + "epoch": 1.1796959826275788, + "grad_norm": 6.246398264149539, + "learning_rate": 1.7858092859497392e-05, + "loss": 0.9146, + "step": 2173 + }, + { + "epoch": 1.1802388707926168, + "grad_norm": 9.483282452267977, + "learning_rate": 1.7855917317824383e-05, + "loss": 1.3714, + "step": 2174 + }, + { + "epoch": 1.1807817589576548, + "grad_norm": 7.377757446359213, + "learning_rate": 1.7853740804519e-05, + "loss": 0.7914, + "step": 2175 + }, + { + "epoch": 1.1813246471226928, + "grad_norm": 6.555194448443568, + "learning_rate": 1.7851563319850435e-05, + "loss": 1.0025, + "step": 2176 + }, + { + "epoch": 1.1818675352877308, + "grad_norm": 5.968820734002013, + "learning_rate": 1.7849384864088005e-05, + "loss": 1.0449, + "step": 2177 + }, + { + "epoch": 1.1824104234527688, + "grad_norm": 8.050731398373367, + "learning_rate": 1.7847205437501145e-05, + "loss": 1.2944, + "step": 2178 + }, + { + "epoch": 1.1829533116178068, + "grad_norm": 6.479093872062473, + "learning_rate": 1.784502504035941e-05, + "loss": 0.8312, + "step": 2179 + }, + { + "epoch": 1.1834961997828448, + "grad_norm": 9.645172625032446, + "learning_rate": 1.7842843672932473e-05, + "loss": 1.4877, + "step": 2180 + }, + { + "epoch": 1.1840390879478828, + "grad_norm": 7.811721210097083, + "learning_rate": 1.7840661335490133e-05, + "loss": 1.0861, + "step": 2181 + }, + { + "epoch": 1.1845819761129208, + "grad_norm": 5.939133632197409, + "learning_rate": 1.7838478028302303e-05, + "loss": 0.7359, + "step": 2182 + }, + { + "epoch": 1.1851248642779588, + "grad_norm": 7.121597254982603, + "learning_rate": 1.7836293751639017e-05, + "loss": 0.626, + "step": 2183 + }, + { + "epoch": 1.1856677524429968, + "grad_norm": 7.225734005859657, + "learning_rate": 1.783410850577043e-05, + "loss": 1.0651, + "step": 2184 + }, + { + "epoch": 1.1862106406080348, + "grad_norm": 6.3733995711762965, + "learning_rate": 1.783192229096682e-05, + "loss": 1.0399, + "step": 2185 + }, + { + "epoch": 1.1867535287730728, + "grad_norm": 7.215840642904109, + "learning_rate": 1.7829735107498576e-05, + "loss": 1.0958, + "step": 2186 + }, + { + "epoch": 1.1872964169381108, + "grad_norm": 7.2349637750528135, + "learning_rate": 1.7827546955636216e-05, + "loss": 0.8005, + "step": 2187 + }, + { + "epoch": 1.1878393051031488, + "grad_norm": 7.592927799049675, + "learning_rate": 1.7825357835650376e-05, + "loss": 1.0981, + "step": 2188 + }, + { + "epoch": 1.1883821932681868, + "grad_norm": 8.010034188435748, + "learning_rate": 1.7823167747811805e-05, + "loss": 1.3088, + "step": 2189 + }, + { + "epoch": 1.1889250814332248, + "grad_norm": 5.674359355487917, + "learning_rate": 1.7820976692391377e-05, + "loss": 0.8725, + "step": 2190 + }, + { + "epoch": 1.1894679695982628, + "grad_norm": 5.236415159647034, + "learning_rate": 1.781878466966009e-05, + "loss": 0.7162, + "step": 2191 + }, + { + "epoch": 1.1900108577633008, + "grad_norm": 5.677268305787393, + "learning_rate": 1.781659167988905e-05, + "loss": 0.7101, + "step": 2192 + }, + { + "epoch": 1.1905537459283388, + "grad_norm": 6.370730170332413, + "learning_rate": 1.7814397723349496e-05, + "loss": 0.9087, + "step": 2193 + }, + { + "epoch": 1.1910966340933768, + "grad_norm": 6.045624314512159, + "learning_rate": 1.7812202800312776e-05, + "loss": 1.1714, + "step": 2194 + }, + { + "epoch": 1.1916395222584149, + "grad_norm": 10.798495850925244, + "learning_rate": 1.7810006911050366e-05, + "loss": 1.1997, + "step": 2195 + }, + { + "epoch": 1.1921824104234529, + "grad_norm": 5.89010386817935, + "learning_rate": 1.780781005583385e-05, + "loss": 1.0702, + "step": 2196 + }, + { + "epoch": 1.1927252985884909, + "grad_norm": 6.09421937705502, + "learning_rate": 1.7805612234934946e-05, + "loss": 0.825, + "step": 2197 + }, + { + "epoch": 1.1932681867535289, + "grad_norm": 5.582107046729241, + "learning_rate": 1.780341344862548e-05, + "loss": 0.8548, + "step": 2198 + }, + { + "epoch": 1.1938110749185669, + "grad_norm": 7.598179521038406, + "learning_rate": 1.78012136971774e-05, + "loss": 1.0015, + "step": 2199 + }, + { + "epoch": 1.1943539630836049, + "grad_norm": 7.4068666684121585, + "learning_rate": 1.7799012980862777e-05, + "loss": 1.0645, + "step": 2200 + }, + { + "epoch": 1.1948968512486429, + "grad_norm": 6.087565282375561, + "learning_rate": 1.7796811299953796e-05, + "loss": 0.9169, + "step": 2201 + }, + { + "epoch": 1.1954397394136809, + "grad_norm": 6.440325481735819, + "learning_rate": 1.7794608654722772e-05, + "loss": 1.0707, + "step": 2202 + }, + { + "epoch": 1.1959826275787189, + "grad_norm": 7.562903613167714, + "learning_rate": 1.7792405045442125e-05, + "loss": 1.0086, + "step": 2203 + }, + { + "epoch": 1.1965255157437569, + "grad_norm": 6.2377346284888, + "learning_rate": 1.77902004723844e-05, + "loss": 0.6111, + "step": 2204 + }, + { + "epoch": 1.1970684039087949, + "grad_norm": 5.978299124706355, + "learning_rate": 1.7787994935822268e-05, + "loss": 0.5782, + "step": 2205 + }, + { + "epoch": 1.1976112920738329, + "grad_norm": 5.9766366336106715, + "learning_rate": 1.778578843602851e-05, + "loss": 0.8409, + "step": 2206 + }, + { + "epoch": 1.1981541802388709, + "grad_norm": 7.700010929432555, + "learning_rate": 1.7783580973276027e-05, + "loss": 1.091, + "step": 2207 + }, + { + "epoch": 1.1986970684039089, + "grad_norm": 8.043903647823342, + "learning_rate": 1.778137254783785e-05, + "loss": 1.2894, + "step": 2208 + }, + { + "epoch": 1.1992399565689469, + "grad_norm": 8.709529168595635, + "learning_rate": 1.777916315998711e-05, + "loss": 1.5601, + "step": 2209 + }, + { + "epoch": 1.1997828447339849, + "grad_norm": 6.234520197542074, + "learning_rate": 1.7776952809997073e-05, + "loss": 1.0414, + "step": 2210 + }, + { + "epoch": 1.200325732899023, + "grad_norm": 5.8108372705767, + "learning_rate": 1.7774741498141116e-05, + "loss": 0.8811, + "step": 2211 + }, + { + "epoch": 1.200868621064061, + "grad_norm": 6.423388338643525, + "learning_rate": 1.7772529224692744e-05, + "loss": 0.8371, + "step": 2212 + }, + { + "epoch": 1.201411509229099, + "grad_norm": 7.740996825838731, + "learning_rate": 1.7770315989925565e-05, + "loss": 0.9993, + "step": 2213 + }, + { + "epoch": 1.201954397394137, + "grad_norm": 6.33257469290407, + "learning_rate": 1.7768101794113323e-05, + "loss": 0.8563, + "step": 2214 + }, + { + "epoch": 1.202497285559175, + "grad_norm": 8.126954888789392, + "learning_rate": 1.776588663752987e-05, + "loss": 0.9479, + "step": 2215 + }, + { + "epoch": 1.203040173724213, + "grad_norm": 7.781450888823224, + "learning_rate": 1.7763670520449178e-05, + "loss": 1.0559, + "step": 2216 + }, + { + "epoch": 1.203583061889251, + "grad_norm": 6.615763983931515, + "learning_rate": 1.7761453443145348e-05, + "loss": 0.7463, + "step": 2217 + }, + { + "epoch": 1.204125950054289, + "grad_norm": 10.193394757123674, + "learning_rate": 1.7759235405892584e-05, + "loss": 1.3597, + "step": 2218 + }, + { + "epoch": 1.204668838219327, + "grad_norm": 7.365896137130193, + "learning_rate": 1.7757016408965217e-05, + "loss": 0.5822, + "step": 2219 + }, + { + "epoch": 1.205211726384365, + "grad_norm": 8.530190774406401, + "learning_rate": 1.77547964526377e-05, + "loss": 1.0075, + "step": 2220 + }, + { + "epoch": 1.205754614549403, + "grad_norm": 7.143571088845206, + "learning_rate": 1.7752575537184597e-05, + "loss": 0.9395, + "step": 2221 + }, + { + "epoch": 1.206297502714441, + "grad_norm": 7.790375269151535, + "learning_rate": 1.7750353662880595e-05, + "loss": 1.0556, + "step": 2222 + }, + { + "epoch": 1.206840390879479, + "grad_norm": 6.5033471136457255, + "learning_rate": 1.77481308300005e-05, + "loss": 0.7464, + "step": 2223 + }, + { + "epoch": 1.207383279044517, + "grad_norm": 7.06026002468489, + "learning_rate": 1.774590703881924e-05, + "loss": 0.6315, + "step": 2224 + }, + { + "epoch": 1.207926167209555, + "grad_norm": 8.773109114641178, + "learning_rate": 1.774368228961185e-05, + "loss": 0.9884, + "step": 2225 + }, + { + "epoch": 1.208469055374593, + "grad_norm": 7.343364313617542, + "learning_rate": 1.774145658265349e-05, + "loss": 1.1542, + "step": 2226 + }, + { + "epoch": 1.209011943539631, + "grad_norm": 5.953950803365371, + "learning_rate": 1.773922991821944e-05, + "loss": 0.8485, + "step": 2227 + }, + { + "epoch": 1.209554831704669, + "grad_norm": 8.905484324028157, + "learning_rate": 1.7737002296585105e-05, + "loss": 1.0171, + "step": 2228 + }, + { + "epoch": 1.210097719869707, + "grad_norm": 7.158929117808072, + "learning_rate": 1.7734773718025992e-05, + "loss": 0.6134, + "step": 2229 + }, + { + "epoch": 1.210640608034745, + "grad_norm": 8.342029051568577, + "learning_rate": 1.7732544182817737e-05, + "loss": 0.9775, + "step": 2230 + }, + { + "epoch": 1.211183496199783, + "grad_norm": 5.587278578636139, + "learning_rate": 1.7730313691236098e-05, + "loss": 0.9395, + "step": 2231 + }, + { + "epoch": 1.211726384364821, + "grad_norm": 6.1948883535007475, + "learning_rate": 1.7728082243556936e-05, + "loss": 0.9977, + "step": 2232 + }, + { + "epoch": 1.212269272529859, + "grad_norm": 6.095377323439518, + "learning_rate": 1.7725849840056248e-05, + "loss": 0.9141, + "step": 2233 + }, + { + "epoch": 1.212812160694897, + "grad_norm": 8.71522968216411, + "learning_rate": 1.7723616481010137e-05, + "loss": 1.4356, + "step": 2234 + }, + { + "epoch": 1.213355048859935, + "grad_norm": 5.416863970522961, + "learning_rate": 1.7721382166694834e-05, + "loss": 0.6594, + "step": 2235 + }, + { + "epoch": 1.213897937024973, + "grad_norm": 8.046753742934493, + "learning_rate": 1.7719146897386674e-05, + "loss": 1.1187, + "step": 2236 + }, + { + "epoch": 1.214440825190011, + "grad_norm": 6.625446441699582, + "learning_rate": 1.7716910673362123e-05, + "loss": 0.8731, + "step": 2237 + }, + { + "epoch": 1.214983713355049, + "grad_norm": 6.47539429072121, + "learning_rate": 1.7714673494897767e-05, + "loss": 0.9473, + "step": 2238 + }, + { + "epoch": 1.215526601520087, + "grad_norm": 7.458434209050833, + "learning_rate": 1.7712435362270292e-05, + "loss": 0.7844, + "step": 2239 + }, + { + "epoch": 1.216069489685125, + "grad_norm": 6.935847404573923, + "learning_rate": 1.7710196275756524e-05, + "loss": 0.8431, + "step": 2240 + }, + { + "epoch": 1.216612377850163, + "grad_norm": 8.015017144687471, + "learning_rate": 1.770795623563339e-05, + "loss": 1.2329, + "step": 2241 + }, + { + "epoch": 1.217155266015201, + "grad_norm": 6.914919733505818, + "learning_rate": 1.7705715242177944e-05, + "loss": 1.1335, + "step": 2242 + }, + { + "epoch": 1.217698154180239, + "grad_norm": 5.900061304870694, + "learning_rate": 1.7703473295667354e-05, + "loss": 0.7453, + "step": 2243 + }, + { + "epoch": 1.218241042345277, + "grad_norm": 4.959614765477869, + "learning_rate": 1.770123039637891e-05, + "loss": 0.6093, + "step": 2244 + }, + { + "epoch": 1.2187839305103148, + "grad_norm": 8.038465432918807, + "learning_rate": 1.769898654459002e-05, + "loss": 1.0029, + "step": 2245 + }, + { + "epoch": 1.219326818675353, + "grad_norm": 8.08635156220498, + "learning_rate": 1.7696741740578204e-05, + "loss": 1.0839, + "step": 2246 + }, + { + "epoch": 1.2198697068403908, + "grad_norm": 7.985335036181301, + "learning_rate": 1.7694495984621097e-05, + "loss": 1.262, + "step": 2247 + }, + { + "epoch": 1.220412595005429, + "grad_norm": 8.31037377117703, + "learning_rate": 1.769224927699647e-05, + "loss": 1.0265, + "step": 2248 + }, + { + "epoch": 1.2209554831704668, + "grad_norm": 6.241752286447289, + "learning_rate": 1.769000161798219e-05, + "loss": 0.6179, + "step": 2249 + }, + { + "epoch": 1.221498371335505, + "grad_norm": 6.767956890107702, + "learning_rate": 1.7687753007856253e-05, + "loss": 0.9897, + "step": 2250 + }, + { + "epoch": 1.2220412595005428, + "grad_norm": 7.119621854687504, + "learning_rate": 1.7685503446896772e-05, + "loss": 1.0529, + "step": 2251 + }, + { + "epoch": 1.222584147665581, + "grad_norm": 6.37786350129091, + "learning_rate": 1.7683252935381976e-05, + "loss": 0.8575, + "step": 2252 + }, + { + "epoch": 1.2231270358306188, + "grad_norm": 9.280258739154695, + "learning_rate": 1.768100147359021e-05, + "loss": 0.959, + "step": 2253 + }, + { + "epoch": 1.223669923995657, + "grad_norm": 7.172898793975323, + "learning_rate": 1.7678749061799942e-05, + "loss": 1.3745, + "step": 2254 + }, + { + "epoch": 1.2242128121606948, + "grad_norm": 5.3297508420409825, + "learning_rate": 1.7676495700289753e-05, + "loss": 0.6697, + "step": 2255 + }, + { + "epoch": 1.224755700325733, + "grad_norm": 6.518884308754971, + "learning_rate": 1.767424138933834e-05, + "loss": 0.9854, + "step": 2256 + }, + { + "epoch": 1.2252985884907708, + "grad_norm": 7.939498291049385, + "learning_rate": 1.767198612922452e-05, + "loss": 0.8918, + "step": 2257 + }, + { + "epoch": 1.225841476655809, + "grad_norm": 6.309602610605377, + "learning_rate": 1.7669729920227226e-05, + "loss": 0.9939, + "step": 2258 + }, + { + "epoch": 1.2263843648208468, + "grad_norm": 7.462197154770194, + "learning_rate": 1.7667472762625518e-05, + "loss": 1.1299, + "step": 2259 + }, + { + "epoch": 1.226927252985885, + "grad_norm": 6.700031938849371, + "learning_rate": 1.7665214656698555e-05, + "loss": 0.8655, + "step": 2260 + }, + { + "epoch": 1.2274701411509228, + "grad_norm": 6.72124660471967, + "learning_rate": 1.7662955602725627e-05, + "loss": 0.7383, + "step": 2261 + }, + { + "epoch": 1.228013029315961, + "grad_norm": 6.071348581517431, + "learning_rate": 1.7660695600986137e-05, + "loss": 0.8571, + "step": 2262 + }, + { + "epoch": 1.2285559174809988, + "grad_norm": 5.914184858865564, + "learning_rate": 1.7658434651759605e-05, + "loss": 0.7494, + "step": 2263 + }, + { + "epoch": 1.229098805646037, + "grad_norm": 12.685614201985125, + "learning_rate": 1.7656172755325675e-05, + "loss": 1.608, + "step": 2264 + }, + { + "epoch": 1.2296416938110748, + "grad_norm": 7.001237249554534, + "learning_rate": 1.765390991196409e-05, + "loss": 1.0437, + "step": 2265 + }, + { + "epoch": 1.230184581976113, + "grad_norm": 8.232094967642231, + "learning_rate": 1.765164612195473e-05, + "loss": 0.8574, + "step": 2266 + }, + { + "epoch": 1.2307274701411508, + "grad_norm": 6.367083948330669, + "learning_rate": 1.7649381385577582e-05, + "loss": 0.7106, + "step": 2267 + }, + { + "epoch": 1.231270358306189, + "grad_norm": 6.2062819181758195, + "learning_rate": 1.7647115703112756e-05, + "loss": 0.6838, + "step": 2268 + }, + { + "epoch": 1.2318132464712268, + "grad_norm": 6.896096048797596, + "learning_rate": 1.7644849074840472e-05, + "loss": 0.9595, + "step": 2269 + }, + { + "epoch": 1.232356134636265, + "grad_norm": 5.915145487902616, + "learning_rate": 1.7642581501041067e-05, + "loss": 0.5741, + "step": 2270 + }, + { + "epoch": 1.2328990228013028, + "grad_norm": 5.966871119524657, + "learning_rate": 1.7640312981995004e-05, + "loss": 0.8686, + "step": 2271 + }, + { + "epoch": 1.233441910966341, + "grad_norm": 6.835680775363237, + "learning_rate": 1.7638043517982854e-05, + "loss": 0.9532, + "step": 2272 + }, + { + "epoch": 1.2339847991313788, + "grad_norm": 8.228029492129801, + "learning_rate": 1.7635773109285306e-05, + "loss": 1.0834, + "step": 2273 + }, + { + "epoch": 1.234527687296417, + "grad_norm": 8.272927259623488, + "learning_rate": 1.7633501756183174e-05, + "loss": 1.2591, + "step": 2274 + }, + { + "epoch": 1.2350705754614548, + "grad_norm": 6.755748928056399, + "learning_rate": 1.7631229458957377e-05, + "loss": 0.997, + "step": 2275 + }, + { + "epoch": 1.235613463626493, + "grad_norm": 7.335027842642818, + "learning_rate": 1.762895621788896e-05, + "loss": 0.8695, + "step": 2276 + }, + { + "epoch": 1.2361563517915308, + "grad_norm": 5.237971036796901, + "learning_rate": 1.7626682033259077e-05, + "loss": 0.6846, + "step": 2277 + }, + { + "epoch": 1.236699239956569, + "grad_norm": 6.333455514649705, + "learning_rate": 1.7624406905349004e-05, + "loss": 0.7532, + "step": 2278 + }, + { + "epoch": 1.2372421281216068, + "grad_norm": 6.255556726847127, + "learning_rate": 1.7622130834440136e-05, + "loss": 0.7801, + "step": 2279 + }, + { + "epoch": 1.237785016286645, + "grad_norm": 5.602832960919912, + "learning_rate": 1.761985382081398e-05, + "loss": 0.5105, + "step": 2280 + }, + { + "epoch": 1.2383279044516828, + "grad_norm": 5.814619917758832, + "learning_rate": 1.7617575864752155e-05, + "loss": 0.9034, + "step": 2281 + }, + { + "epoch": 1.238870792616721, + "grad_norm": 6.743402665368447, + "learning_rate": 1.7615296966536407e-05, + "loss": 0.7713, + "step": 2282 + }, + { + "epoch": 1.2394136807817588, + "grad_norm": 7.176469994889228, + "learning_rate": 1.7613017126448597e-05, + "loss": 0.957, + "step": 2283 + }, + { + "epoch": 1.239956568946797, + "grad_norm": 8.5427155590311, + "learning_rate": 1.7610736344770693e-05, + "loss": 0.8789, + "step": 2284 + }, + { + "epoch": 1.2404994571118348, + "grad_norm": 6.674205629131359, + "learning_rate": 1.7608454621784786e-05, + "loss": 1.0239, + "step": 2285 + }, + { + "epoch": 1.241042345276873, + "grad_norm": 7.687894188099568, + "learning_rate": 1.7606171957773088e-05, + "loss": 1.3164, + "step": 2286 + }, + { + "epoch": 1.2415852334419109, + "grad_norm": 6.331613648212601, + "learning_rate": 1.760388835301792e-05, + "loss": 0.8047, + "step": 2287 + }, + { + "epoch": 1.242128121606949, + "grad_norm": 7.0776252458357725, + "learning_rate": 1.760160380780172e-05, + "loss": 1.1893, + "step": 2288 + }, + { + "epoch": 1.2426710097719869, + "grad_norm": 5.434346705496786, + "learning_rate": 1.7599318322407047e-05, + "loss": 0.763, + "step": 2289 + }, + { + "epoch": 1.243213897937025, + "grad_norm": 6.0271106768536775, + "learning_rate": 1.759703189711657e-05, + "loss": 1.1129, + "step": 2290 + }, + { + "epoch": 1.2437567861020629, + "grad_norm": 7.286292908201965, + "learning_rate": 1.7594744532213088e-05, + "loss": 1.1136, + "step": 2291 + }, + { + "epoch": 1.244299674267101, + "grad_norm": 6.685086643433707, + "learning_rate": 1.7592456227979488e-05, + "loss": 0.724, + "step": 2292 + }, + { + "epoch": 1.2448425624321389, + "grad_norm": 7.184174802374185, + "learning_rate": 1.7590166984698807e-05, + "loss": 1.1471, + "step": 2293 + }, + { + "epoch": 1.245385450597177, + "grad_norm": 7.787563610771557, + "learning_rate": 1.7587876802654176e-05, + "loss": 1.01, + "step": 2294 + }, + { + "epoch": 1.2459283387622149, + "grad_norm": 7.036440767647583, + "learning_rate": 1.7585585682128848e-05, + "loss": 0.8612, + "step": 2295 + }, + { + "epoch": 1.246471226927253, + "grad_norm": 6.6372154433448705, + "learning_rate": 1.7583293623406192e-05, + "loss": 1.3394, + "step": 2296 + }, + { + "epoch": 1.2470141150922909, + "grad_norm": 7.707986168491473, + "learning_rate": 1.7581000626769697e-05, + "loss": 0.8688, + "step": 2297 + }, + { + "epoch": 1.247557003257329, + "grad_norm": 7.862359368033938, + "learning_rate": 1.757870669250296e-05, + "loss": 1.1999, + "step": 2298 + }, + { + "epoch": 1.2480998914223669, + "grad_norm": 7.942280078814929, + "learning_rate": 1.75764118208897e-05, + "loss": 1.241, + "step": 2299 + }, + { + "epoch": 1.248642779587405, + "grad_norm": 7.258825765708901, + "learning_rate": 1.7574116012213753e-05, + "loss": 0.8336, + "step": 2300 + }, + { + "epoch": 1.2491856677524429, + "grad_norm": 7.0149158036864065, + "learning_rate": 1.7571819266759066e-05, + "loss": 0.8637, + "step": 2301 + }, + { + "epoch": 1.249728555917481, + "grad_norm": 7.069351546856077, + "learning_rate": 1.7569521584809703e-05, + "loss": 0.8146, + "step": 2302 + }, + { + "epoch": 1.250271444082519, + "grad_norm": 4.6424312647112425, + "learning_rate": 1.756722296664985e-05, + "loss": 0.5056, + "step": 2303 + }, + { + "epoch": 1.2508143322475571, + "grad_norm": 7.857412677197525, + "learning_rate": 1.7564923412563797e-05, + "loss": 1.3072, + "step": 2304 + }, + { + "epoch": 1.251357220412595, + "grad_norm": 6.081845076053873, + "learning_rate": 1.7562622922835962e-05, + "loss": 1.0212, + "step": 2305 + }, + { + "epoch": 1.2519001085776331, + "grad_norm": 5.244697868381976, + "learning_rate": 1.7560321497750867e-05, + "loss": 0.9677, + "step": 2306 + }, + { + "epoch": 1.252442996742671, + "grad_norm": 11.027915321544773, + "learning_rate": 1.7558019137593162e-05, + "loss": 0.8984, + "step": 2307 + }, + { + "epoch": 1.2529858849077091, + "grad_norm": 5.955816485468965, + "learning_rate": 1.7555715842647603e-05, + "loss": 0.6282, + "step": 2308 + }, + { + "epoch": 1.253528773072747, + "grad_norm": 8.925991341226974, + "learning_rate": 1.755341161319907e-05, + "loss": 1.2121, + "step": 2309 + }, + { + "epoch": 1.2540716612377851, + "grad_norm": 7.947406679428915, + "learning_rate": 1.7551106449532547e-05, + "loss": 1.0545, + "step": 2310 + }, + { + "epoch": 1.254614549402823, + "grad_norm": 7.329950391674132, + "learning_rate": 1.7548800351933147e-05, + "loss": 1.0585, + "step": 2311 + }, + { + "epoch": 1.2551574375678611, + "grad_norm": 5.0931347466751955, + "learning_rate": 1.7546493320686084e-05, + "loss": 0.5533, + "step": 2312 + }, + { + "epoch": 1.255700325732899, + "grad_norm": 7.125279667319334, + "learning_rate": 1.7544185356076703e-05, + "loss": 1.1011, + "step": 2313 + }, + { + "epoch": 1.2562432138979371, + "grad_norm": 8.060697991979964, + "learning_rate": 1.7541876458390453e-05, + "loss": 0.8732, + "step": 2314 + }, + { + "epoch": 1.256786102062975, + "grad_norm": 9.732797197135078, + "learning_rate": 1.75395666279129e-05, + "loss": 1.5663, + "step": 2315 + }, + { + "epoch": 1.2573289902280131, + "grad_norm": 5.3673655570709435, + "learning_rate": 1.7537255864929733e-05, + "loss": 0.7057, + "step": 2316 + }, + { + "epoch": 1.257871878393051, + "grad_norm": 5.997561584131186, + "learning_rate": 1.7534944169726752e-05, + "loss": 0.8526, + "step": 2317 + }, + { + "epoch": 1.2584147665580891, + "grad_norm": 7.236766116786611, + "learning_rate": 1.753263154258986e-05, + "loss": 0.8921, + "step": 2318 + }, + { + "epoch": 1.258957654723127, + "grad_norm": 7.663917187467161, + "learning_rate": 1.7530317983805096e-05, + "loss": 0.9309, + "step": 2319 + }, + { + "epoch": 1.2595005428881652, + "grad_norm": 5.704193869412544, + "learning_rate": 1.75280034936586e-05, + "loss": 0.6291, + "step": 2320 + }, + { + "epoch": 1.260043431053203, + "grad_norm": 7.365455665485428, + "learning_rate": 1.7525688072436636e-05, + "loss": 0.9946, + "step": 2321 + }, + { + "epoch": 1.2605863192182412, + "grad_norm": 4.8267670121648205, + "learning_rate": 1.752337172042558e-05, + "loss": 0.8254, + "step": 2322 + }, + { + "epoch": 1.261129207383279, + "grad_norm": 7.231116660749922, + "learning_rate": 1.752105443791192e-05, + "loss": 1.3835, + "step": 2323 + }, + { + "epoch": 1.2616720955483172, + "grad_norm": 6.02849395913157, + "learning_rate": 1.7518736225182253e-05, + "loss": 0.6229, + "step": 2324 + }, + { + "epoch": 1.262214983713355, + "grad_norm": 6.458823304206971, + "learning_rate": 1.7516417082523314e-05, + "loss": 0.7707, + "step": 2325 + }, + { + "epoch": 1.2627578718783932, + "grad_norm": 7.163598438660781, + "learning_rate": 1.7514097010221927e-05, + "loss": 1.0007, + "step": 2326 + }, + { + "epoch": 1.263300760043431, + "grad_norm": 6.607394766457111, + "learning_rate": 1.7511776008565043e-05, + "loss": 1.0289, + "step": 2327 + }, + { + "epoch": 1.2638436482084692, + "grad_norm": 8.20658895752514, + "learning_rate": 1.7509454077839736e-05, + "loss": 0.7553, + "step": 2328 + }, + { + "epoch": 1.264386536373507, + "grad_norm": 8.963409997263136, + "learning_rate": 1.750713121833318e-05, + "loss": 0.9558, + "step": 2329 + }, + { + "epoch": 1.2649294245385452, + "grad_norm": 6.463632425962391, + "learning_rate": 1.7504807430332668e-05, + "loss": 0.6553, + "step": 2330 + }, + { + "epoch": 1.265472312703583, + "grad_norm": 6.936173778617146, + "learning_rate": 1.7502482714125612e-05, + "loss": 0.7861, + "step": 2331 + }, + { + "epoch": 1.2660152008686212, + "grad_norm": 6.975529559460763, + "learning_rate": 1.7500157069999536e-05, + "loss": 0.665, + "step": 2332 + }, + { + "epoch": 1.266558089033659, + "grad_norm": 6.768480922885786, + "learning_rate": 1.7497830498242082e-05, + "loss": 1.0163, + "step": 2333 + }, + { + "epoch": 1.2671009771986972, + "grad_norm": 6.357421024100016, + "learning_rate": 1.7495502999141004e-05, + "loss": 0.5897, + "step": 2334 + }, + { + "epoch": 1.267643865363735, + "grad_norm": 7.82539792030223, + "learning_rate": 1.7493174572984168e-05, + "loss": 1.0509, + "step": 2335 + }, + { + "epoch": 1.2681867535287732, + "grad_norm": 7.6523935004894215, + "learning_rate": 1.7490845220059554e-05, + "loss": 0.7603, + "step": 2336 + }, + { + "epoch": 1.268729641693811, + "grad_norm": 6.677360283026087, + "learning_rate": 1.7488514940655267e-05, + "loss": 1.1797, + "step": 2337 + }, + { + "epoch": 1.2692725298588492, + "grad_norm": 7.837606738097467, + "learning_rate": 1.7486183735059517e-05, + "loss": 1.1995, + "step": 2338 + }, + { + "epoch": 1.269815418023887, + "grad_norm": 8.913928805014676, + "learning_rate": 1.7483851603560634e-05, + "loss": 1.6007, + "step": 2339 + }, + { + "epoch": 1.2703583061889252, + "grad_norm": 8.467363918664116, + "learning_rate": 1.7481518546447058e-05, + "loss": 0.9515, + "step": 2340 + }, + { + "epoch": 1.270901194353963, + "grad_norm": 6.758882762498143, + "learning_rate": 1.747918456400734e-05, + "loss": 0.7935, + "step": 2341 + }, + { + "epoch": 1.2714440825190012, + "grad_norm": 5.455764070482765, + "learning_rate": 1.7476849656530158e-05, + "loss": 0.4623, + "step": 2342 + }, + { + "epoch": 1.271986970684039, + "grad_norm": 8.417334291043444, + "learning_rate": 1.7474513824304294e-05, + "loss": 1.1563, + "step": 2343 + }, + { + "epoch": 1.2725298588490772, + "grad_norm": 5.61084844590966, + "learning_rate": 1.7472177067618646e-05, + "loss": 0.8566, + "step": 2344 + }, + { + "epoch": 1.273072747014115, + "grad_norm": 7.726478476188633, + "learning_rate": 1.746983938676223e-05, + "loss": 1.0568, + "step": 2345 + }, + { + "epoch": 1.2736156351791532, + "grad_norm": 8.882013629978706, + "learning_rate": 1.7467500782024177e-05, + "loss": 1.0921, + "step": 2346 + }, + { + "epoch": 1.274158523344191, + "grad_norm": 6.976574380776494, + "learning_rate": 1.746516125369372e-05, + "loss": 1.1222, + "step": 2347 + }, + { + "epoch": 1.2747014115092292, + "grad_norm": 7.641347465211616, + "learning_rate": 1.7462820802060224e-05, + "loss": 1.0672, + "step": 2348 + }, + { + "epoch": 1.275244299674267, + "grad_norm": 6.473243122755322, + "learning_rate": 1.7460479427413156e-05, + "loss": 0.9019, + "step": 2349 + }, + { + "epoch": 1.2757871878393052, + "grad_norm": 6.494958002593054, + "learning_rate": 1.7458137130042105e-05, + "loss": 0.7862, + "step": 2350 + }, + { + "epoch": 1.276330076004343, + "grad_norm": 6.54459648554074, + "learning_rate": 1.7455793910236764e-05, + "loss": 0.9572, + "step": 2351 + }, + { + "epoch": 1.2768729641693812, + "grad_norm": 6.249218254368162, + "learning_rate": 1.7453449768286952e-05, + "loss": 0.9415, + "step": 2352 + }, + { + "epoch": 1.277415852334419, + "grad_norm": 6.335956299458767, + "learning_rate": 1.7451104704482592e-05, + "loss": 0.7647, + "step": 2353 + }, + { + "epoch": 1.2779587404994572, + "grad_norm": 6.296362619304004, + "learning_rate": 1.744875871911373e-05, + "loss": 0.6293, + "step": 2354 + }, + { + "epoch": 1.278501628664495, + "grad_norm": 6.629983181103369, + "learning_rate": 1.7446411812470516e-05, + "loss": 0.7816, + "step": 2355 + }, + { + "epoch": 1.2790445168295332, + "grad_norm": 6.921460108918273, + "learning_rate": 1.744406398484322e-05, + "loss": 0.6373, + "step": 2356 + }, + { + "epoch": 1.279587404994571, + "grad_norm": 9.4421099603369, + "learning_rate": 1.7441715236522228e-05, + "loss": 1.5291, + "step": 2357 + }, + { + "epoch": 1.2801302931596092, + "grad_norm": 7.838644313955947, + "learning_rate": 1.7439365567798035e-05, + "loss": 0.7934, + "step": 2358 + }, + { + "epoch": 1.280673181324647, + "grad_norm": 5.7659414273416, + "learning_rate": 1.743701497896125e-05, + "loss": 0.8358, + "step": 2359 + }, + { + "epoch": 1.2812160694896852, + "grad_norm": 4.731198780322981, + "learning_rate": 1.7434663470302602e-05, + "loss": 0.7357, + "step": 2360 + }, + { + "epoch": 1.281758957654723, + "grad_norm": 8.993227122901354, + "learning_rate": 1.7432311042112926e-05, + "loss": 1.0679, + "step": 2361 + }, + { + "epoch": 1.2823018458197613, + "grad_norm": 7.193753575449645, + "learning_rate": 1.7429957694683175e-05, + "loss": 0.9257, + "step": 2362 + }, + { + "epoch": 1.282844733984799, + "grad_norm": 7.1027186942584635, + "learning_rate": 1.7427603428304416e-05, + "loss": 0.64, + "step": 2363 + }, + { + "epoch": 1.2833876221498373, + "grad_norm": 9.441739719572269, + "learning_rate": 1.7425248243267824e-05, + "loss": 1.3236, + "step": 2364 + }, + { + "epoch": 1.283930510314875, + "grad_norm": 6.478965109994284, + "learning_rate": 1.7422892139864696e-05, + "loss": 0.7726, + "step": 2365 + }, + { + "epoch": 1.2844733984799133, + "grad_norm": 6.372951649068815, + "learning_rate": 1.742053511838644e-05, + "loss": 0.7785, + "step": 2366 + }, + { + "epoch": 1.285016286644951, + "grad_norm": 10.052356612127076, + "learning_rate": 1.7418177179124574e-05, + "loss": 1.7877, + "step": 2367 + }, + { + "epoch": 1.2855591748099893, + "grad_norm": 5.24633801429042, + "learning_rate": 1.741581832237073e-05, + "loss": 0.4195, + "step": 2368 + }, + { + "epoch": 1.286102062975027, + "grad_norm": 7.958362658884473, + "learning_rate": 1.7413458548416656e-05, + "loss": 1.1782, + "step": 2369 + }, + { + "epoch": 1.2866449511400653, + "grad_norm": 7.923031881378813, + "learning_rate": 1.7411097857554216e-05, + "loss": 1.0861, + "step": 2370 + }, + { + "epoch": 1.287187839305103, + "grad_norm": 6.53099588563486, + "learning_rate": 1.7408736250075378e-05, + "loss": 0.8717, + "step": 2371 + }, + { + "epoch": 1.2877307274701413, + "grad_norm": 7.364911313605863, + "learning_rate": 1.740637372627224e-05, + "loss": 0.9341, + "step": 2372 + }, + { + "epoch": 1.288273615635179, + "grad_norm": 8.083622859612968, + "learning_rate": 1.740401028643699e-05, + "loss": 1.1654, + "step": 2373 + }, + { + "epoch": 1.2888165038002173, + "grad_norm": 6.561315990238234, + "learning_rate": 1.7401645930861944e-05, + "loss": 0.8088, + "step": 2374 + }, + { + "epoch": 1.289359391965255, + "grad_norm": 6.926193447339149, + "learning_rate": 1.739928065983954e-05, + "loss": 0.8335, + "step": 2375 + }, + { + "epoch": 1.2899022801302933, + "grad_norm": 7.876730774519387, + "learning_rate": 1.7396914473662306e-05, + "loss": 0.9969, + "step": 2376 + }, + { + "epoch": 1.290445168295331, + "grad_norm": 6.513645878383447, + "learning_rate": 1.73945473726229e-05, + "loss": 0.9302, + "step": 2377 + }, + { + "epoch": 1.2909880564603693, + "grad_norm": 6.441562626276137, + "learning_rate": 1.73921793570141e-05, + "loss": 1.0318, + "step": 2378 + }, + { + "epoch": 1.291530944625407, + "grad_norm": 5.226443422910506, + "learning_rate": 1.7389810427128766e-05, + "loss": 0.6142, + "step": 2379 + }, + { + "epoch": 1.2920738327904453, + "grad_norm": 7.436131751002477, + "learning_rate": 1.7387440583259906e-05, + "loss": 0.9907, + "step": 2380 + }, + { + "epoch": 1.292616720955483, + "grad_norm": 6.335395136223753, + "learning_rate": 1.7385069825700615e-05, + "loss": 0.8431, + "step": 2381 + }, + { + "epoch": 1.2931596091205213, + "grad_norm": 8.93013417703781, + "learning_rate": 1.738269815474412e-05, + "loss": 1.3162, + "step": 2382 + }, + { + "epoch": 1.293702497285559, + "grad_norm": 6.616000665685319, + "learning_rate": 1.7380325570683754e-05, + "loss": 0.81, + "step": 2383 + }, + { + "epoch": 1.2942453854505973, + "grad_norm": 9.126678309827314, + "learning_rate": 1.737795207381296e-05, + "loss": 1.4976, + "step": 2384 + }, + { + "epoch": 1.294788273615635, + "grad_norm": 9.726273416685123, + "learning_rate": 1.737557766442529e-05, + "loss": 1.0284, + "step": 2385 + }, + { + "epoch": 1.2953311617806733, + "grad_norm": 6.207571191669611, + "learning_rate": 1.737320234281442e-05, + "loss": 0.9409, + "step": 2386 + }, + { + "epoch": 1.295874049945711, + "grad_norm": 6.150632184961992, + "learning_rate": 1.7370826109274136e-05, + "loss": 0.7558, + "step": 2387 + }, + { + "epoch": 1.2964169381107493, + "grad_norm": 6.555861715157727, + "learning_rate": 1.736844896409833e-05, + "loss": 1.4217, + "step": 2388 + }, + { + "epoch": 1.296959826275787, + "grad_norm": 7.291607947091272, + "learning_rate": 1.7366070907581015e-05, + "loss": 0.9969, + "step": 2389 + }, + { + "epoch": 1.2975027144408253, + "grad_norm": 7.003873752430508, + "learning_rate": 1.7363691940016307e-05, + "loss": 0.9569, + "step": 2390 + }, + { + "epoch": 1.298045602605863, + "grad_norm": 6.355975398274633, + "learning_rate": 1.7361312061698444e-05, + "loss": 0.9244, + "step": 2391 + }, + { + "epoch": 1.2985884907709013, + "grad_norm": 5.093818720596728, + "learning_rate": 1.7358931272921773e-05, + "loss": 0.5581, + "step": 2392 + }, + { + "epoch": 1.299131378935939, + "grad_norm": 4.532047411652702, + "learning_rate": 1.7356549573980753e-05, + "loss": 0.5153, + "step": 2393 + }, + { + "epoch": 1.2996742671009773, + "grad_norm": 7.378796483754393, + "learning_rate": 1.735416696516996e-05, + "loss": 0.9631, + "step": 2394 + }, + { + "epoch": 1.3002171552660151, + "grad_norm": 6.186232096125507, + "learning_rate": 1.7351783446784075e-05, + "loss": 0.7876, + "step": 2395 + }, + { + "epoch": 1.3007600434310533, + "grad_norm": 7.851387526494335, + "learning_rate": 1.7349399019117897e-05, + "loss": 0.8983, + "step": 2396 + }, + { + "epoch": 1.3013029315960911, + "grad_norm": 7.275506442499215, + "learning_rate": 1.7347013682466335e-05, + "loss": 0.8334, + "step": 2397 + }, + { + "epoch": 1.3018458197611293, + "grad_norm": 6.264072188984264, + "learning_rate": 1.7344627437124407e-05, + "loss": 0.6585, + "step": 2398 + }, + { + "epoch": 1.3023887079261671, + "grad_norm": 7.537073303434296, + "learning_rate": 1.734224028338726e-05, + "loss": 1.1033, + "step": 2399 + }, + { + "epoch": 1.3029315960912053, + "grad_norm": 6.850281208830108, + "learning_rate": 1.7339852221550126e-05, + "loss": 0.8133, + "step": 2400 + }, + { + "epoch": 1.3034744842562431, + "grad_norm": 8.089656198012262, + "learning_rate": 1.7337463251908374e-05, + "loss": 1.1342, + "step": 2401 + }, + { + "epoch": 1.3040173724212814, + "grad_norm": 6.2919884156176105, + "learning_rate": 1.733507337475747e-05, + "loss": 0.6533, + "step": 2402 + }, + { + "epoch": 1.3045602605863191, + "grad_norm": 5.114953151143873, + "learning_rate": 1.7332682590393008e-05, + "loss": 0.7451, + "step": 2403 + }, + { + "epoch": 1.3051031487513574, + "grad_norm": 6.18470588828327, + "learning_rate": 1.733029089911067e-05, + "loss": 1.1299, + "step": 2404 + }, + { + "epoch": 1.3056460369163951, + "grad_norm": 5.950129534147113, + "learning_rate": 1.7327898301206273e-05, + "loss": 0.7329, + "step": 2405 + }, + { + "epoch": 1.3061889250814334, + "grad_norm": 6.293661255114301, + "learning_rate": 1.7325504796975732e-05, + "loss": 0.7535, + "step": 2406 + }, + { + "epoch": 1.3067318132464711, + "grad_norm": 7.682262802579054, + "learning_rate": 1.732311038671509e-05, + "loss": 1.2399, + "step": 2407 + }, + { + "epoch": 1.3072747014115094, + "grad_norm": 7.314299514485556, + "learning_rate": 1.7320715070720478e-05, + "loss": 0.9298, + "step": 2408 + }, + { + "epoch": 1.3078175895765471, + "grad_norm": 8.921126356976833, + "learning_rate": 1.7318318849288158e-05, + "loss": 1.4254, + "step": 2409 + }, + { + "epoch": 1.3083604777415854, + "grad_norm": 7.060288459365394, + "learning_rate": 1.7315921722714503e-05, + "loss": 0.9428, + "step": 2410 + }, + { + "epoch": 1.3089033659066232, + "grad_norm": 8.568251970472561, + "learning_rate": 1.7313523691295988e-05, + "loss": 0.9011, + "step": 2411 + }, + { + "epoch": 1.3094462540716614, + "grad_norm": 6.049363733159396, + "learning_rate": 1.7311124755329206e-05, + "loss": 1.0718, + "step": 2412 + }, + { + "epoch": 1.3099891422366992, + "grad_norm": 8.287809242433424, + "learning_rate": 1.7308724915110864e-05, + "loss": 0.9137, + "step": 2413 + }, + { + "epoch": 1.3105320304017374, + "grad_norm": 8.436233795197621, + "learning_rate": 1.7306324170937774e-05, + "loss": 1.2427, + "step": 2414 + }, + { + "epoch": 1.3110749185667752, + "grad_norm": 7.035710404708695, + "learning_rate": 1.7303922523106863e-05, + "loss": 0.8206, + "step": 2415 + }, + { + "epoch": 1.3116178067318134, + "grad_norm": 5.9728457186472275, + "learning_rate": 1.730151997191518e-05, + "loss": 0.7307, + "step": 2416 + }, + { + "epoch": 1.3121606948968512, + "grad_norm": 6.296494925379207, + "learning_rate": 1.7299116517659865e-05, + "loss": 0.9808, + "step": 2417 + }, + { + "epoch": 1.3127035830618892, + "grad_norm": 7.672949954931739, + "learning_rate": 1.7296712160638192e-05, + "loss": 1.0289, + "step": 2418 + }, + { + "epoch": 1.3132464712269272, + "grad_norm": 5.958844838187463, + "learning_rate": 1.7294306901147525e-05, + "loss": 0.9657, + "step": 2419 + }, + { + "epoch": 1.3137893593919652, + "grad_norm": 6.5874736925476745, + "learning_rate": 1.7291900739485356e-05, + "loss": 0.8701, + "step": 2420 + }, + { + "epoch": 1.3143322475570032, + "grad_norm": 8.273467490572564, + "learning_rate": 1.7289493675949282e-05, + "loss": 0.9353, + "step": 2421 + }, + { + "epoch": 1.3148751357220412, + "grad_norm": 8.115143215098843, + "learning_rate": 1.7287085710837013e-05, + "loss": 0.6241, + "step": 2422 + }, + { + "epoch": 1.3154180238870792, + "grad_norm": 7.004703854660018, + "learning_rate": 1.7284676844446368e-05, + "loss": 0.801, + "step": 2423 + }, + { + "epoch": 1.3159609120521172, + "grad_norm": 6.039976470078682, + "learning_rate": 1.728226707707528e-05, + "loss": 1.085, + "step": 2424 + }, + { + "epoch": 1.3165038002171552, + "grad_norm": 5.318299059101813, + "learning_rate": 1.72798564090218e-05, + "loss": 0.4945, + "step": 2425 + }, + { + "epoch": 1.3170466883821932, + "grad_norm": 6.095578475774351, + "learning_rate": 1.727744484058407e-05, + "loss": 0.7696, + "step": 2426 + }, + { + "epoch": 1.3175895765472312, + "grad_norm": 6.4886080927006695, + "learning_rate": 1.7275032372060368e-05, + "loss": 0.5364, + "step": 2427 + }, + { + "epoch": 1.3181324647122692, + "grad_norm": 8.681669485105683, + "learning_rate": 1.7272619003749066e-05, + "loss": 0.9132, + "step": 2428 + }, + { + "epoch": 1.3186753528773072, + "grad_norm": 6.819144122945338, + "learning_rate": 1.7270204735948653e-05, + "loss": 0.7867, + "step": 2429 + }, + { + "epoch": 1.3192182410423452, + "grad_norm": 6.65338033222052, + "learning_rate": 1.7267789568957734e-05, + "loss": 0.7236, + "step": 2430 + }, + { + "epoch": 1.3197611292073832, + "grad_norm": 6.622862530496735, + "learning_rate": 1.7265373503075014e-05, + "loss": 0.9038, + "step": 2431 + }, + { + "epoch": 1.3203040173724212, + "grad_norm": 6.284216416307323, + "learning_rate": 1.7262956538599323e-05, + "loss": 0.677, + "step": 2432 + }, + { + "epoch": 1.3208469055374592, + "grad_norm": 9.150803536423062, + "learning_rate": 1.7260538675829593e-05, + "loss": 1.7953, + "step": 2433 + }, + { + "epoch": 1.3213897937024972, + "grad_norm": 9.236250071710021, + "learning_rate": 1.7258119915064867e-05, + "loss": 0.8568, + "step": 2434 + }, + { + "epoch": 1.3219326818675352, + "grad_norm": 7.061272817814159, + "learning_rate": 1.72557002566043e-05, + "loss": 0.9355, + "step": 2435 + }, + { + "epoch": 1.3224755700325732, + "grad_norm": 6.786187626307571, + "learning_rate": 1.7253279700747164e-05, + "loss": 0.7707, + "step": 2436 + }, + { + "epoch": 1.3230184581976112, + "grad_norm": 5.592273858747665, + "learning_rate": 1.725085824779283e-05, + "loss": 0.5932, + "step": 2437 + }, + { + "epoch": 1.3235613463626492, + "grad_norm": 8.137496643840409, + "learning_rate": 1.72484358980408e-05, + "loss": 1.0244, + "step": 2438 + }, + { + "epoch": 1.3241042345276872, + "grad_norm": 5.981122138792478, + "learning_rate": 1.724601265179066e-05, + "loss": 0.9807, + "step": 2439 + }, + { + "epoch": 1.3246471226927252, + "grad_norm": 6.912027532729551, + "learning_rate": 1.7243588509342127e-05, + "loss": 0.8673, + "step": 2440 + }, + { + "epoch": 1.3251900108577632, + "grad_norm": 5.836398731450734, + "learning_rate": 1.7241163470995024e-05, + "loss": 0.5738, + "step": 2441 + }, + { + "epoch": 1.3257328990228012, + "grad_norm": 5.0689326588185795, + "learning_rate": 1.7238737537049283e-05, + "loss": 0.4449, + "step": 2442 + }, + { + "epoch": 1.3262757871878392, + "grad_norm": 6.628337387569613, + "learning_rate": 1.7236310707804943e-05, + "loss": 0.7975, + "step": 2443 + }, + { + "epoch": 1.3268186753528772, + "grad_norm": 9.359654484357105, + "learning_rate": 1.7233882983562168e-05, + "loss": 0.8996, + "step": 2444 + }, + { + "epoch": 1.3273615635179152, + "grad_norm": 5.979606493030731, + "learning_rate": 1.723145436462121e-05, + "loss": 0.732, + "step": 2445 + }, + { + "epoch": 1.3279044516829532, + "grad_norm": 6.393853922089353, + "learning_rate": 1.7229024851282453e-05, + "loss": 0.9713, + "step": 2446 + }, + { + "epoch": 1.3284473398479912, + "grad_norm": 9.783386290725945, + "learning_rate": 1.722659444384638e-05, + "loss": 1.2991, + "step": 2447 + }, + { + "epoch": 1.3289902280130292, + "grad_norm": 8.471557152983001, + "learning_rate": 1.722416314261359e-05, + "loss": 0.8517, + "step": 2448 + }, + { + "epoch": 1.3295331161780672, + "grad_norm": 7.20996107804807, + "learning_rate": 1.7221730947884793e-05, + "loss": 1.0892, + "step": 2449 + }, + { + "epoch": 1.3300760043431052, + "grad_norm": 9.97079093211024, + "learning_rate": 1.7219297859960796e-05, + "loss": 0.8866, + "step": 2450 + }, + { + "epoch": 1.3306188925081432, + "grad_norm": 5.88899239355285, + "learning_rate": 1.7216863879142536e-05, + "loss": 0.7286, + "step": 2451 + }, + { + "epoch": 1.3311617806731812, + "grad_norm": 8.317968838525326, + "learning_rate": 1.7214429005731054e-05, + "loss": 0.9374, + "step": 2452 + }, + { + "epoch": 1.3317046688382193, + "grad_norm": 8.278980033154179, + "learning_rate": 1.721199324002749e-05, + "loss": 1.2977, + "step": 2453 + }, + { + "epoch": 1.3322475570032573, + "grad_norm": 7.50073001311901, + "learning_rate": 1.7209556582333106e-05, + "loss": 0.9479, + "step": 2454 + }, + { + "epoch": 1.3327904451682953, + "grad_norm": 9.34140520001984, + "learning_rate": 1.720711903294928e-05, + "loss": 1.1834, + "step": 2455 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 6.155163014555799, + "learning_rate": 1.720468059217748e-05, + "loss": 1.0674, + "step": 2456 + }, + { + "epoch": 1.3338762214983713, + "grad_norm": 8.207152122618293, + "learning_rate": 1.7202241260319305e-05, + "loss": 1.1623, + "step": 2457 + }, + { + "epoch": 1.3344191096634093, + "grad_norm": 8.477127177193351, + "learning_rate": 1.719980103767645e-05, + "loss": 1.0812, + "step": 2458 + }, + { + "epoch": 1.3349619978284473, + "grad_norm": 9.352009366934986, + "learning_rate": 1.7197359924550726e-05, + "loss": 0.9593, + "step": 2459 + }, + { + "epoch": 1.3355048859934853, + "grad_norm": 6.377220189803284, + "learning_rate": 1.719491792124406e-05, + "loss": 0.8569, + "step": 2460 + }, + { + "epoch": 1.3360477741585233, + "grad_norm": 10.372722210417326, + "learning_rate": 1.7192475028058475e-05, + "loss": 0.8778, + "step": 2461 + }, + { + "epoch": 1.3365906623235613, + "grad_norm": 6.943692981514271, + "learning_rate": 1.7190031245296118e-05, + "loss": 0.8143, + "step": 2462 + }, + { + "epoch": 1.3371335504885993, + "grad_norm": 8.53588364120296, + "learning_rate": 1.7187586573259237e-05, + "loss": 1.2608, + "step": 2463 + }, + { + "epoch": 1.3376764386536373, + "grad_norm": 6.85497949710551, + "learning_rate": 1.7185141012250195e-05, + "loss": 0.621, + "step": 2464 + }, + { + "epoch": 1.3382193268186753, + "grad_norm": 9.575890345813113, + "learning_rate": 1.7182694562571458e-05, + "loss": 1.5686, + "step": 2465 + }, + { + "epoch": 1.3387622149837133, + "grad_norm": 8.637844905126164, + "learning_rate": 1.718024722452561e-05, + "loss": 0.8287, + "step": 2466 + }, + { + "epoch": 1.3393051031487513, + "grad_norm": 7.464168837468054, + "learning_rate": 1.7177798998415344e-05, + "loss": 0.9837, + "step": 2467 + }, + { + "epoch": 1.3398479913137893, + "grad_norm": 9.27927006463804, + "learning_rate": 1.7175349884543458e-05, + "loss": 1.3479, + "step": 2468 + }, + { + "epoch": 1.3403908794788273, + "grad_norm": 6.9150180643503765, + "learning_rate": 1.717289988321286e-05, + "loss": 0.7731, + "step": 2469 + }, + { + "epoch": 1.3409337676438653, + "grad_norm": 8.264286643294954, + "learning_rate": 1.7170448994726574e-05, + "loss": 1.1252, + "step": 2470 + }, + { + "epoch": 1.3414766558089033, + "grad_norm": 6.1769106952218715, + "learning_rate": 1.7167997219387728e-05, + "loss": 0.6077, + "step": 2471 + }, + { + "epoch": 1.3420195439739413, + "grad_norm": 6.00046554758586, + "learning_rate": 1.7165544557499563e-05, + "loss": 0.9807, + "step": 2472 + }, + { + "epoch": 1.3425624321389793, + "grad_norm": 5.684026968717126, + "learning_rate": 1.7163091009365427e-05, + "loss": 0.8755, + "step": 2473 + }, + { + "epoch": 1.3431053203040173, + "grad_norm": 8.910694398982464, + "learning_rate": 1.7160636575288777e-05, + "loss": 1.1781, + "step": 2474 + }, + { + "epoch": 1.3436482084690553, + "grad_norm": 6.090059050660724, + "learning_rate": 1.7158181255573184e-05, + "loss": 0.8757, + "step": 2475 + }, + { + "epoch": 1.3441910966340933, + "grad_norm": 5.3693699493967575, + "learning_rate": 1.7155725050522325e-05, + "loss": 0.4111, + "step": 2476 + }, + { + "epoch": 1.3447339847991313, + "grad_norm": 6.4883376045754675, + "learning_rate": 1.7153267960439987e-05, + "loss": 1.1667, + "step": 2477 + }, + { + "epoch": 1.3452768729641693, + "grad_norm": 6.9649527868325585, + "learning_rate": 1.7150809985630065e-05, + "loss": 1.2642, + "step": 2478 + }, + { + "epoch": 1.3458197611292073, + "grad_norm": 7.458967124353787, + "learning_rate": 1.714835112639657e-05, + "loss": 1.1198, + "step": 2479 + }, + { + "epoch": 1.3463626492942453, + "grad_norm": 7.4820540779231015, + "learning_rate": 1.7145891383043613e-05, + "loss": 1.069, + "step": 2480 + }, + { + "epoch": 1.3469055374592833, + "grad_norm": 6.8442083181868165, + "learning_rate": 1.7143430755875422e-05, + "loss": 1.1187, + "step": 2481 + }, + { + "epoch": 1.3474484256243213, + "grad_norm": 8.389465670773259, + "learning_rate": 1.7140969245196332e-05, + "loss": 1.216, + "step": 2482 + }, + { + "epoch": 1.3479913137893593, + "grad_norm": 8.099481746156716, + "learning_rate": 1.713850685131078e-05, + "loss": 1.1365, + "step": 2483 + }, + { + "epoch": 1.3485342019543973, + "grad_norm": 7.005684437567773, + "learning_rate": 1.713604357452333e-05, + "loss": 0.7665, + "step": 2484 + }, + { + "epoch": 1.3490770901194353, + "grad_norm": 6.032904961736961, + "learning_rate": 1.7133579415138634e-05, + "loss": 0.7527, + "step": 2485 + }, + { + "epoch": 1.3496199782844733, + "grad_norm": 7.644359802639278, + "learning_rate": 1.7131114373461466e-05, + "loss": 0.7293, + "step": 2486 + }, + { + "epoch": 1.3501628664495113, + "grad_norm": 9.681466854129468, + "learning_rate": 1.7128648449796706e-05, + "loss": 1.0718, + "step": 2487 + }, + { + "epoch": 1.3507057546145493, + "grad_norm": 7.703820713192501, + "learning_rate": 1.7126181644449348e-05, + "loss": 1.0049, + "step": 2488 + }, + { + "epoch": 1.3512486427795873, + "grad_norm": 8.313943023532504, + "learning_rate": 1.7123713957724482e-05, + "loss": 1.1647, + "step": 2489 + }, + { + "epoch": 1.3517915309446253, + "grad_norm": 6.869318843676937, + "learning_rate": 1.7121245389927326e-05, + "loss": 0.7056, + "step": 2490 + }, + { + "epoch": 1.3523344191096633, + "grad_norm": 8.354366894281007, + "learning_rate": 1.7118775941363186e-05, + "loss": 1.4107, + "step": 2491 + }, + { + "epoch": 1.3528773072747013, + "grad_norm": 7.594996874862703, + "learning_rate": 1.7116305612337493e-05, + "loss": 1.3947, + "step": 2492 + }, + { + "epoch": 1.3534201954397393, + "grad_norm": 5.912640331169303, + "learning_rate": 1.7113834403155782e-05, + "loss": 0.678, + "step": 2493 + }, + { + "epoch": 1.3539630836047774, + "grad_norm": 8.603089873660597, + "learning_rate": 1.7111362314123693e-05, + "loss": 1.1846, + "step": 2494 + }, + { + "epoch": 1.3545059717698154, + "grad_norm": 5.899587064063632, + "learning_rate": 1.710888934554698e-05, + "loss": 0.7241, + "step": 2495 + }, + { + "epoch": 1.3550488599348534, + "grad_norm": 5.1366064982711315, + "learning_rate": 1.7106415497731502e-05, + "loss": 0.6714, + "step": 2496 + }, + { + "epoch": 1.3555917480998914, + "grad_norm": 6.55484035818926, + "learning_rate": 1.710394077098323e-05, + "loss": 1.0163, + "step": 2497 + }, + { + "epoch": 1.3561346362649294, + "grad_norm": 7.060748735507699, + "learning_rate": 1.710146516560824e-05, + "loss": 0.8407, + "step": 2498 + }, + { + "epoch": 1.3566775244299674, + "grad_norm": 7.488417378829399, + "learning_rate": 1.709898868191272e-05, + "loss": 1.0936, + "step": 2499 + }, + { + "epoch": 1.3572204125950054, + "grad_norm": 5.616560879457259, + "learning_rate": 1.7096511320202965e-05, + "loss": 0.6962, + "step": 2500 + }, + { + "epoch": 1.3577633007600434, + "grad_norm": 6.5457002772857775, + "learning_rate": 1.7094033080785384e-05, + "loss": 0.6338, + "step": 2501 + }, + { + "epoch": 1.3583061889250814, + "grad_norm": 5.788382651374094, + "learning_rate": 1.709155396396648e-05, + "loss": 0.9251, + "step": 2502 + }, + { + "epoch": 1.3588490770901194, + "grad_norm": 5.94278314769549, + "learning_rate": 1.7089073970052883e-05, + "loss": 0.5738, + "step": 2503 + }, + { + "epoch": 1.3593919652551574, + "grad_norm": 8.987615230603502, + "learning_rate": 1.7086593099351318e-05, + "loss": 1.1561, + "step": 2504 + }, + { + "epoch": 1.3599348534201954, + "grad_norm": 6.25509019167934, + "learning_rate": 1.7084111352168627e-05, + "loss": 0.5584, + "step": 2505 + }, + { + "epoch": 1.3604777415852334, + "grad_norm": 6.42626784615777, + "learning_rate": 1.708162872881175e-05, + "loss": 0.6385, + "step": 2506 + }, + { + "epoch": 1.3610206297502714, + "grad_norm": 7.342680004921898, + "learning_rate": 1.707914522958775e-05, + "loss": 1.0729, + "step": 2507 + }, + { + "epoch": 1.3615635179153094, + "grad_norm": 6.858864236878158, + "learning_rate": 1.707666085480378e-05, + "loss": 0.8641, + "step": 2508 + }, + { + "epoch": 1.3621064060803474, + "grad_norm": 6.176457712037886, + "learning_rate": 1.707417560476712e-05, + "loss": 0.9472, + "step": 2509 + }, + { + "epoch": 1.3626492942453854, + "grad_norm": 7.346954574146698, + "learning_rate": 1.7071689479785145e-05, + "loss": 1.0253, + "step": 2510 + }, + { + "epoch": 1.3631921824104234, + "grad_norm": 4.9331066104603085, + "learning_rate": 1.7069202480165344e-05, + "loss": 0.7307, + "step": 2511 + }, + { + "epoch": 1.3637350705754614, + "grad_norm": 8.316711829847868, + "learning_rate": 1.7066714606215316e-05, + "loss": 0.8078, + "step": 2512 + }, + { + "epoch": 1.3642779587404994, + "grad_norm": 8.079444940270628, + "learning_rate": 1.706422585824276e-05, + "loss": 1.2922, + "step": 2513 + }, + { + "epoch": 1.3648208469055374, + "grad_norm": 7.503741374806181, + "learning_rate": 1.7061736236555494e-05, + "loss": 0.7943, + "step": 2514 + }, + { + "epoch": 1.3653637350705754, + "grad_norm": 6.4023293864360635, + "learning_rate": 1.7059245741461435e-05, + "loss": 1.0111, + "step": 2515 + }, + { + "epoch": 1.3659066232356134, + "grad_norm": 5.7325920409358595, + "learning_rate": 1.705675437326861e-05, + "loss": 1.1201, + "step": 2516 + }, + { + "epoch": 1.3664495114006514, + "grad_norm": 8.452660578831315, + "learning_rate": 1.705426213228516e-05, + "loss": 0.8818, + "step": 2517 + }, + { + "epoch": 1.3669923995656894, + "grad_norm": 7.8301031574049755, + "learning_rate": 1.7051769018819327e-05, + "loss": 0.8803, + "step": 2518 + }, + { + "epoch": 1.3675352877307274, + "grad_norm": 10.77827779179992, + "learning_rate": 1.704927503317946e-05, + "loss": 1.6667, + "step": 2519 + }, + { + "epoch": 1.3680781758957654, + "grad_norm": 4.579120603442138, + "learning_rate": 1.7046780175674023e-05, + "loss": 0.6101, + "step": 2520 + }, + { + "epoch": 1.3686210640608034, + "grad_norm": 7.443733277623912, + "learning_rate": 1.7044284446611584e-05, + "loss": 0.8174, + "step": 2521 + }, + { + "epoch": 1.3691639522258414, + "grad_norm": 10.860855146774309, + "learning_rate": 1.7041787846300817e-05, + "loss": 1.2849, + "step": 2522 + }, + { + "epoch": 1.3697068403908794, + "grad_norm": 9.311283837684355, + "learning_rate": 1.7039290375050503e-05, + "loss": 1.4923, + "step": 2523 + }, + { + "epoch": 1.3702497285559174, + "grad_norm": 7.171226901795553, + "learning_rate": 1.703679203316954e-05, + "loss": 0.9493, + "step": 2524 + }, + { + "epoch": 1.3707926167209554, + "grad_norm": 8.658053683953712, + "learning_rate": 1.7034292820966923e-05, + "loss": 1.1366, + "step": 2525 + }, + { + "epoch": 1.3713355048859934, + "grad_norm": 6.400723985129287, + "learning_rate": 1.703179273875176e-05, + "loss": 1.0484, + "step": 2526 + }, + { + "epoch": 1.3718783930510314, + "grad_norm": 6.2215780206112274, + "learning_rate": 1.702929178683326e-05, + "loss": 1.1058, + "step": 2527 + }, + { + "epoch": 1.3724212812160694, + "grad_norm": 7.548758658264992, + "learning_rate": 1.7026789965520755e-05, + "loss": 1.094, + "step": 2528 + }, + { + "epoch": 1.3729641693811074, + "grad_norm": 7.250992484079327, + "learning_rate": 1.7024287275123664e-05, + "loss": 0.8424, + "step": 2529 + }, + { + "epoch": 1.3735070575461454, + "grad_norm": 6.664109303085317, + "learning_rate": 1.702178371595153e-05, + "loss": 1.0498, + "step": 2530 + }, + { + "epoch": 1.3740499457111834, + "grad_norm": 6.379180216401721, + "learning_rate": 1.7019279288313994e-05, + "loss": 0.9038, + "step": 2531 + }, + { + "epoch": 1.3745928338762214, + "grad_norm": 7.524913368583668, + "learning_rate": 1.7016773992520806e-05, + "loss": 0.8787, + "step": 2532 + }, + { + "epoch": 1.3751357220412594, + "grad_norm": 5.722778911929788, + "learning_rate": 1.701426782888183e-05, + "loss": 0.8199, + "step": 2533 + }, + { + "epoch": 1.3756786102062974, + "grad_norm": 6.26381521799836, + "learning_rate": 1.701176079770703e-05, + "loss": 0.6989, + "step": 2534 + }, + { + "epoch": 1.3762214983713354, + "grad_norm": 7.2300935887481765, + "learning_rate": 1.700925289930648e-05, + "loss": 1.094, + "step": 2535 + }, + { + "epoch": 1.3767643865363735, + "grad_norm": 7.385614391862118, + "learning_rate": 1.7006744133990358e-05, + "loss": 0.7753, + "step": 2536 + }, + { + "epoch": 1.3773072747014115, + "grad_norm": 5.52011787904346, + "learning_rate": 1.7004234502068952e-05, + "loss": 0.9232, + "step": 2537 + }, + { + "epoch": 1.3778501628664495, + "grad_norm": 6.523028453950784, + "learning_rate": 1.7001724003852668e-05, + "loss": 1.1181, + "step": 2538 + }, + { + "epoch": 1.3783930510314875, + "grad_norm": 6.334090281128193, + "learning_rate": 1.6999212639651995e-05, + "loss": 0.9384, + "step": 2539 + }, + { + "epoch": 1.3789359391965255, + "grad_norm": 7.3687722051890265, + "learning_rate": 1.6996700409777548e-05, + "loss": 0.9678, + "step": 2540 + }, + { + "epoch": 1.3794788273615635, + "grad_norm": 9.49113126498042, + "learning_rate": 1.6994187314540042e-05, + "loss": 0.7875, + "step": 2541 + }, + { + "epoch": 1.3800217155266015, + "grad_norm": 4.885613246464195, + "learning_rate": 1.69916733542503e-05, + "loss": 0.4488, + "step": 2542 + }, + { + "epoch": 1.3805646036916395, + "grad_norm": 7.362521236541499, + "learning_rate": 1.6989158529219262e-05, + "loss": 0.9808, + "step": 2543 + }, + { + "epoch": 1.3811074918566775, + "grad_norm": 8.6488909724118, + "learning_rate": 1.6986642839757953e-05, + "loss": 1.0599, + "step": 2544 + }, + { + "epoch": 1.3816503800217155, + "grad_norm": 6.054541159551389, + "learning_rate": 1.698412628617752e-05, + "loss": 0.8252, + "step": 2545 + }, + { + "epoch": 1.3821932681867535, + "grad_norm": 7.77360294326381, + "learning_rate": 1.698160886878922e-05, + "loss": 1.0718, + "step": 2546 + }, + { + "epoch": 1.3827361563517915, + "grad_norm": 9.493714306510151, + "learning_rate": 1.697909058790441e-05, + "loss": 1.2784, + "step": 2547 + }, + { + "epoch": 1.3832790445168295, + "grad_norm": 7.7307988823000455, + "learning_rate": 1.6976571443834555e-05, + "loss": 1.0308, + "step": 2548 + }, + { + "epoch": 1.3838219326818675, + "grad_norm": 7.149534093427325, + "learning_rate": 1.697405143689122e-05, + "loss": 0.9702, + "step": 2549 + }, + { + "epoch": 1.3843648208469055, + "grad_norm": 6.16619254042592, + "learning_rate": 1.6971530567386087e-05, + "loss": 0.6922, + "step": 2550 + }, + { + "epoch": 1.3849077090119435, + "grad_norm": 6.555255328428672, + "learning_rate": 1.6969008835630947e-05, + "loss": 0.8771, + "step": 2551 + }, + { + "epoch": 1.3854505971769815, + "grad_norm": 7.652524555180243, + "learning_rate": 1.6966486241937685e-05, + "loss": 1.0635, + "step": 2552 + }, + { + "epoch": 1.3859934853420195, + "grad_norm": 9.675328775672968, + "learning_rate": 1.69639627866183e-05, + "loss": 1.1616, + "step": 2553 + }, + { + "epoch": 1.3865363735070575, + "grad_norm": 6.5667057730439256, + "learning_rate": 1.69614384699849e-05, + "loss": 0.7181, + "step": 2554 + }, + { + "epoch": 1.3870792616720955, + "grad_norm": 7.374476144001139, + "learning_rate": 1.6958913292349698e-05, + "loss": 1.123, + "step": 2555 + }, + { + "epoch": 1.3876221498371335, + "grad_norm": 8.873033944832025, + "learning_rate": 1.6956387254025005e-05, + "loss": 1.2651, + "step": 2556 + }, + { + "epoch": 1.3881650380021715, + "grad_norm": 6.877978468812787, + "learning_rate": 1.695386035532325e-05, + "loss": 1.0039, + "step": 2557 + }, + { + "epoch": 1.3887079261672095, + "grad_norm": 4.746919583421645, + "learning_rate": 1.6951332596556966e-05, + "loss": 0.7222, + "step": 2558 + }, + { + "epoch": 1.3892508143322475, + "grad_norm": 5.538693020619686, + "learning_rate": 1.6948803978038787e-05, + "loss": 0.8463, + "step": 2559 + }, + { + "epoch": 1.3897937024972855, + "grad_norm": 6.636448326582328, + "learning_rate": 1.6946274500081455e-05, + "loss": 0.7594, + "step": 2560 + }, + { + "epoch": 1.3903365906623235, + "grad_norm": 8.74658695496399, + "learning_rate": 1.6943744162997825e-05, + "loss": 1.224, + "step": 2561 + }, + { + "epoch": 1.3908794788273615, + "grad_norm": 7.305739264312674, + "learning_rate": 1.694121296710085e-05, + "loss": 1.0755, + "step": 2562 + }, + { + "epoch": 1.3914223669923995, + "grad_norm": 4.6475508332537165, + "learning_rate": 1.693868091270359e-05, + "loss": 0.6127, + "step": 2563 + }, + { + "epoch": 1.3919652551574375, + "grad_norm": 8.484116710765257, + "learning_rate": 1.6936148000119218e-05, + "loss": 1.029, + "step": 2564 + }, + { + "epoch": 1.3925081433224755, + "grad_norm": 6.532276109161297, + "learning_rate": 1.6933614229661008e-05, + "loss": 1.0517, + "step": 2565 + }, + { + "epoch": 1.3930510314875135, + "grad_norm": 5.43723324019315, + "learning_rate": 1.6931079601642336e-05, + "loss": 0.6915, + "step": 2566 + }, + { + "epoch": 1.3935939196525515, + "grad_norm": 7.562306350008547, + "learning_rate": 1.69285441163767e-05, + "loss": 0.5726, + "step": 2567 + }, + { + "epoch": 1.3941368078175895, + "grad_norm": 8.500602954507755, + "learning_rate": 1.6926007774177678e-05, + "loss": 1.0297, + "step": 2568 + }, + { + "epoch": 1.3946796959826275, + "grad_norm": 6.432151204429844, + "learning_rate": 1.6923470575358977e-05, + "loss": 1.0016, + "step": 2569 + }, + { + "epoch": 1.3952225841476655, + "grad_norm": 9.552257208599151, + "learning_rate": 1.6920932520234402e-05, + "loss": 1.0814, + "step": 2570 + }, + { + "epoch": 1.3957654723127035, + "grad_norm": 8.075918082294125, + "learning_rate": 1.6918393609117863e-05, + "loss": 1.1699, + "step": 2571 + }, + { + "epoch": 1.3963083604777415, + "grad_norm": 6.296577538864887, + "learning_rate": 1.6915853842323373e-05, + "loss": 1.0211, + "step": 2572 + }, + { + "epoch": 1.3968512486427795, + "grad_norm": 7.785873884786678, + "learning_rate": 1.691331322016506e-05, + "loss": 0.9225, + "step": 2573 + }, + { + "epoch": 1.3973941368078175, + "grad_norm": 6.768130802749829, + "learning_rate": 1.691077174295715e-05, + "loss": 0.8092, + "step": 2574 + }, + { + "epoch": 1.3979370249728555, + "grad_norm": 5.566346022945414, + "learning_rate": 1.6908229411013977e-05, + "loss": 0.8312, + "step": 2575 + }, + { + "epoch": 1.3984799131378935, + "grad_norm": 7.2860384645139815, + "learning_rate": 1.6905686224649978e-05, + "loss": 0.9357, + "step": 2576 + }, + { + "epoch": 1.3990228013029316, + "grad_norm": 7.072468614016554, + "learning_rate": 1.6903142184179704e-05, + "loss": 0.7362, + "step": 2577 + }, + { + "epoch": 1.3995656894679696, + "grad_norm": 6.421225685489366, + "learning_rate": 1.6900597289917803e-05, + "loss": 0.6828, + "step": 2578 + }, + { + "epoch": 1.4001085776330076, + "grad_norm": 8.167489592155093, + "learning_rate": 1.689805154217903e-05, + "loss": 1.1237, + "step": 2579 + }, + { + "epoch": 1.4006514657980456, + "grad_norm": 11.676833397587332, + "learning_rate": 1.6895504941278246e-05, + "loss": 0.9532, + "step": 2580 + }, + { + "epoch": 1.4011943539630836, + "grad_norm": 6.04293054204543, + "learning_rate": 1.689295748753042e-05, + "loss": 0.6541, + "step": 2581 + }, + { + "epoch": 1.4017372421281216, + "grad_norm": 8.432023560966908, + "learning_rate": 1.6890409181250632e-05, + "loss": 1.0167, + "step": 2582 + }, + { + "epoch": 1.4022801302931596, + "grad_norm": 7.494983940960417, + "learning_rate": 1.688786002275405e-05, + "loss": 0.8913, + "step": 2583 + }, + { + "epoch": 1.4028230184581976, + "grad_norm": 6.475912046078606, + "learning_rate": 1.6885310012355964e-05, + "loss": 0.8599, + "step": 2584 + }, + { + "epoch": 1.4033659066232356, + "grad_norm": 5.145369710998311, + "learning_rate": 1.6882759150371765e-05, + "loss": 0.8079, + "step": 2585 + }, + { + "epoch": 1.4039087947882736, + "grad_norm": 6.1855027756804155, + "learning_rate": 1.688020743711694e-05, + "loss": 0.8046, + "step": 2586 + }, + { + "epoch": 1.4044516829533116, + "grad_norm": 8.409570707749053, + "learning_rate": 1.68776548729071e-05, + "loss": 0.75, + "step": 2587 + }, + { + "epoch": 1.4049945711183496, + "grad_norm": 6.150666283054011, + "learning_rate": 1.687510145805794e-05, + "loss": 0.8339, + "step": 2588 + }, + { + "epoch": 1.4055374592833876, + "grad_norm": 5.4082205122977625, + "learning_rate": 1.6872547192885272e-05, + "loss": 0.7878, + "step": 2589 + }, + { + "epoch": 1.4060803474484256, + "grad_norm": 7.011727228130495, + "learning_rate": 1.686999207770502e-05, + "loss": 0.6415, + "step": 2590 + }, + { + "epoch": 1.4066232356134636, + "grad_norm": 7.384524419917869, + "learning_rate": 1.6867436112833193e-05, + "loss": 0.8489, + "step": 2591 + }, + { + "epoch": 1.4071661237785016, + "grad_norm": 6.846659366372164, + "learning_rate": 1.6864879298585925e-05, + "loss": 1.128, + "step": 2592 + }, + { + "epoch": 1.4077090119435396, + "grad_norm": 6.142948047083128, + "learning_rate": 1.6862321635279444e-05, + "loss": 0.7886, + "step": 2593 + }, + { + "epoch": 1.4082519001085776, + "grad_norm": 10.23557881641297, + "learning_rate": 1.6859763123230086e-05, + "loss": 0.8323, + "step": 2594 + }, + { + "epoch": 1.4087947882736156, + "grad_norm": 6.995135696669204, + "learning_rate": 1.6857203762754294e-05, + "loss": 1.0283, + "step": 2595 + }, + { + "epoch": 1.4093376764386536, + "grad_norm": 7.709014920487576, + "learning_rate": 1.685464355416861e-05, + "loss": 0.8586, + "step": 2596 + }, + { + "epoch": 1.4098805646036916, + "grad_norm": 8.153794053960306, + "learning_rate": 1.6852082497789684e-05, + "loss": 1.1797, + "step": 2597 + }, + { + "epoch": 1.4104234527687296, + "grad_norm": 7.721286185320884, + "learning_rate": 1.684952059393428e-05, + "loss": 0.9196, + "step": 2598 + }, + { + "epoch": 1.4109663409337676, + "grad_norm": 8.024349951964256, + "learning_rate": 1.684695784291925e-05, + "loss": 0.7769, + "step": 2599 + }, + { + "epoch": 1.4115092290988056, + "grad_norm": 8.135395328882872, + "learning_rate": 1.684439424506156e-05, + "loss": 1.1642, + "step": 2600 + }, + { + "epoch": 1.4120521172638436, + "grad_norm": 6.057801451822921, + "learning_rate": 1.684182980067828e-05, + "loss": 0.7352, + "step": 2601 + }, + { + "epoch": 1.4125950054288816, + "grad_norm": 7.808755828672067, + "learning_rate": 1.683926451008659e-05, + "loss": 0.9788, + "step": 2602 + }, + { + "epoch": 1.4131378935939196, + "grad_norm": 7.426450515083944, + "learning_rate": 1.6836698373603765e-05, + "loss": 0.7696, + "step": 2603 + }, + { + "epoch": 1.4136807817589576, + "grad_norm": 9.382792907740685, + "learning_rate": 1.6834131391547187e-05, + "loss": 1.0872, + "step": 2604 + }, + { + "epoch": 1.4142236699239956, + "grad_norm": 6.171700145892427, + "learning_rate": 1.6831563564234347e-05, + "loss": 0.7233, + "step": 2605 + }, + { + "epoch": 1.4147665580890336, + "grad_norm": 11.270930151197227, + "learning_rate": 1.682899489198284e-05, + "loss": 0.9739, + "step": 2606 + }, + { + "epoch": 1.4153094462540716, + "grad_norm": 9.050544540368543, + "learning_rate": 1.6826425375110357e-05, + "loss": 0.9436, + "step": 2607 + }, + { + "epoch": 1.4158523344191096, + "grad_norm": 6.969930349987723, + "learning_rate": 1.6823855013934705e-05, + "loss": 0.7632, + "step": 2608 + }, + { + "epoch": 1.4163952225841476, + "grad_norm": 5.791480821515424, + "learning_rate": 1.682128380877379e-05, + "loss": 0.8202, + "step": 2609 + }, + { + "epoch": 1.4169381107491856, + "grad_norm": 9.905757175497309, + "learning_rate": 1.6818711759945623e-05, + "loss": 1.2161, + "step": 2610 + }, + { + "epoch": 1.4174809989142236, + "grad_norm": 6.082684402096193, + "learning_rate": 1.6816138867768318e-05, + "loss": 0.7245, + "step": 2611 + }, + { + "epoch": 1.4180238870792616, + "grad_norm": 10.15499139340414, + "learning_rate": 1.6813565132560092e-05, + "loss": 1.1514, + "step": 2612 + }, + { + "epoch": 1.4185667752442996, + "grad_norm": 6.978513243855861, + "learning_rate": 1.6810990554639276e-05, + "loss": 0.5331, + "step": 2613 + }, + { + "epoch": 1.4191096634093376, + "grad_norm": 8.037739371135425, + "learning_rate": 1.6808415134324288e-05, + "loss": 0.7089, + "step": 2614 + }, + { + "epoch": 1.4196525515743756, + "grad_norm": 6.925791024229076, + "learning_rate": 1.6805838871933664e-05, + "loss": 0.8981, + "step": 2615 + }, + { + "epoch": 1.4201954397394136, + "grad_norm": 6.352165521713976, + "learning_rate": 1.6803261767786048e-05, + "loss": 0.5644, + "step": 2616 + }, + { + "epoch": 1.4207383279044516, + "grad_norm": 10.267497627922179, + "learning_rate": 1.680068382220017e-05, + "loss": 0.9607, + "step": 2617 + }, + { + "epoch": 1.4212812160694897, + "grad_norm": 7.696468630335302, + "learning_rate": 1.679810503549488e-05, + "loss": 0.729, + "step": 2618 + }, + { + "epoch": 1.4218241042345277, + "grad_norm": 6.868981955235214, + "learning_rate": 1.679552540798912e-05, + "loss": 0.9152, + "step": 2619 + }, + { + "epoch": 1.4223669923995657, + "grad_norm": 8.909790381587262, + "learning_rate": 1.6792944940001952e-05, + "loss": 0.9335, + "step": 2620 + }, + { + "epoch": 1.4229098805646037, + "grad_norm": 9.62415635421133, + "learning_rate": 1.6790363631852524e-05, + "loss": 1.1012, + "step": 2621 + }, + { + "epoch": 1.4234527687296417, + "grad_norm": 7.307385111665146, + "learning_rate": 1.67877814838601e-05, + "loss": 0.6963, + "step": 2622 + }, + { + "epoch": 1.4239956568946797, + "grad_norm": 9.018642559605246, + "learning_rate": 1.678519849634405e-05, + "loss": 1.0612, + "step": 2623 + }, + { + "epoch": 1.4245385450597177, + "grad_norm": 6.929244675286548, + "learning_rate": 1.6782614669623827e-05, + "loss": 0.9462, + "step": 2624 + }, + { + "epoch": 1.4250814332247557, + "grad_norm": 8.577419731292203, + "learning_rate": 1.6780030004019016e-05, + "loss": 1.112, + "step": 2625 + }, + { + "epoch": 1.4256243213897937, + "grad_norm": 7.125694710248634, + "learning_rate": 1.677744449984929e-05, + "loss": 0.851, + "step": 2626 + }, + { + "epoch": 1.4261672095548317, + "grad_norm": 9.488635353984538, + "learning_rate": 1.6774858157434425e-05, + "loss": 1.3064, + "step": 2627 + }, + { + "epoch": 1.4267100977198697, + "grad_norm": 7.500971938576216, + "learning_rate": 1.6772270977094307e-05, + "loss": 1.3675, + "step": 2628 + }, + { + "epoch": 1.4272529858849077, + "grad_norm": 6.979855088071626, + "learning_rate": 1.676968295914892e-05, + "loss": 0.8896, + "step": 2629 + }, + { + "epoch": 1.4277958740499457, + "grad_norm": 7.795843207695432, + "learning_rate": 1.6767094103918357e-05, + "loss": 0.7758, + "step": 2630 + }, + { + "epoch": 1.4283387622149837, + "grad_norm": 6.131884468994572, + "learning_rate": 1.6764504411722806e-05, + "loss": 0.7744, + "step": 2631 + }, + { + "epoch": 1.4288816503800217, + "grad_norm": 10.33693974618067, + "learning_rate": 1.676191388288257e-05, + "loss": 1.5364, + "step": 2632 + }, + { + "epoch": 1.4294245385450597, + "grad_norm": 10.005426546410897, + "learning_rate": 1.6759322517718048e-05, + "loss": 0.9713, + "step": 2633 + }, + { + "epoch": 1.4299674267100977, + "grad_norm": 9.480236119835816, + "learning_rate": 1.6756730316549745e-05, + "loss": 1.2573, + "step": 2634 + }, + { + "epoch": 1.4305103148751357, + "grad_norm": 10.874101086577593, + "learning_rate": 1.675413727969827e-05, + "loss": 1.0308, + "step": 2635 + }, + { + "epoch": 1.4310532030401737, + "grad_norm": 5.34984499382639, + "learning_rate": 1.675154340748433e-05, + "loss": 0.6013, + "step": 2636 + }, + { + "epoch": 1.4315960912052117, + "grad_norm": 7.464386702227026, + "learning_rate": 1.674894870022874e-05, + "loss": 0.9918, + "step": 2637 + }, + { + "epoch": 1.4321389793702497, + "grad_norm": 8.280480153626241, + "learning_rate": 1.674635315825242e-05, + "loss": 0.8745, + "step": 2638 + }, + { + "epoch": 1.4326818675352877, + "grad_norm": 6.169575731939732, + "learning_rate": 1.6743756781876385e-05, + "loss": 0.6984, + "step": 2639 + }, + { + "epoch": 1.4332247557003257, + "grad_norm": 5.390920772379431, + "learning_rate": 1.6741159571421768e-05, + "loss": 0.8886, + "step": 2640 + }, + { + "epoch": 1.4337676438653637, + "grad_norm": 7.003336357340687, + "learning_rate": 1.6738561527209792e-05, + "loss": 1.0335, + "step": 2641 + }, + { + "epoch": 1.4343105320304017, + "grad_norm": 7.722166856340297, + "learning_rate": 1.6735962649561784e-05, + "loss": 0.9804, + "step": 2642 + }, + { + "epoch": 1.4348534201954397, + "grad_norm": 7.328242752900802, + "learning_rate": 1.673336293879918e-05, + "loss": 0.8675, + "step": 2643 + }, + { + "epoch": 1.4353963083604777, + "grad_norm": 7.344577484351048, + "learning_rate": 1.6730762395243515e-05, + "loss": 1.2518, + "step": 2644 + }, + { + "epoch": 1.4359391965255157, + "grad_norm": 7.560893408391802, + "learning_rate": 1.6728161019216433e-05, + "loss": 1.0196, + "step": 2645 + }, + { + "epoch": 1.4364820846905537, + "grad_norm": 7.509821569775923, + "learning_rate": 1.6725558811039674e-05, + "loss": 1.1112, + "step": 2646 + }, + { + "epoch": 1.4370249728555917, + "grad_norm": 7.672412677248201, + "learning_rate": 1.672295577103508e-05, + "loss": 0.9075, + "step": 2647 + }, + { + "epoch": 1.4375678610206297, + "grad_norm": 8.562389123203614, + "learning_rate": 1.67203518995246e-05, + "loss": 1.2123, + "step": 2648 + }, + { + "epoch": 1.4381107491856677, + "grad_norm": 8.172008773566091, + "learning_rate": 1.671774719683029e-05, + "loss": 0.8064, + "step": 2649 + }, + { + "epoch": 1.4386536373507057, + "grad_norm": 7.5145061670971405, + "learning_rate": 1.6715141663274297e-05, + "loss": 1.3513, + "step": 2650 + }, + { + "epoch": 1.4391965255157437, + "grad_norm": 5.675981880534199, + "learning_rate": 1.6712535299178883e-05, + "loss": 0.5029, + "step": 2651 + }, + { + "epoch": 1.4397394136807817, + "grad_norm": 8.195158801834715, + "learning_rate": 1.6709928104866403e-05, + "loss": 0.9414, + "step": 2652 + }, + { + "epoch": 1.4402823018458197, + "grad_norm": 7.374510239535514, + "learning_rate": 1.6707320080659322e-05, + "loss": 0.8103, + "step": 2653 + }, + { + "epoch": 1.4408251900108577, + "grad_norm": 5.703232221762084, + "learning_rate": 1.6704711226880204e-05, + "loss": 1.1646, + "step": 2654 + }, + { + "epoch": 1.4413680781758957, + "grad_norm": 6.981749095787225, + "learning_rate": 1.6702101543851714e-05, + "loss": 0.7174, + "step": 2655 + }, + { + "epoch": 1.4419109663409337, + "grad_norm": 6.456866833246808, + "learning_rate": 1.6699491031896625e-05, + "loss": 0.9795, + "step": 2656 + }, + { + "epoch": 1.4424538545059717, + "grad_norm": 10.193228844116039, + "learning_rate": 1.6696879691337807e-05, + "loss": 1.1288, + "step": 2657 + }, + { + "epoch": 1.4429967426710097, + "grad_norm": 9.920534268729181, + "learning_rate": 1.6694267522498237e-05, + "loss": 0.8379, + "step": 2658 + }, + { + "epoch": 1.4435396308360477, + "grad_norm": 6.153776792682404, + "learning_rate": 1.669165452570099e-05, + "loss": 0.8818, + "step": 2659 + }, + { + "epoch": 1.4440825190010858, + "grad_norm": 5.650871346064418, + "learning_rate": 1.6689040701269245e-05, + "loss": 0.732, + "step": 2660 + }, + { + "epoch": 1.4446254071661238, + "grad_norm": 7.829649689315363, + "learning_rate": 1.668642604952629e-05, + "loss": 0.8119, + "step": 2661 + }, + { + "epoch": 1.4451682953311618, + "grad_norm": 5.959350361834891, + "learning_rate": 1.6683810570795498e-05, + "loss": 0.6829, + "step": 2662 + }, + { + "epoch": 1.4457111834961998, + "grad_norm": 8.76687690585514, + "learning_rate": 1.6681194265400365e-05, + "loss": 1.4888, + "step": 2663 + }, + { + "epoch": 1.4462540716612378, + "grad_norm": 6.129763847493397, + "learning_rate": 1.6678577133664476e-05, + "loss": 0.8372, + "step": 2664 + }, + { + "epoch": 1.4467969598262758, + "grad_norm": 6.529235674005995, + "learning_rate": 1.6675959175911527e-05, + "loss": 0.9655, + "step": 2665 + }, + { + "epoch": 1.4473398479913138, + "grad_norm": 7.741547370588863, + "learning_rate": 1.6673340392465304e-05, + "loss": 1.0772, + "step": 2666 + }, + { + "epoch": 1.4478827361563518, + "grad_norm": 6.674042204787818, + "learning_rate": 1.6670720783649706e-05, + "loss": 0.9681, + "step": 2667 + }, + { + "epoch": 1.4484256243213898, + "grad_norm": 7.920578917847642, + "learning_rate": 1.666810034978873e-05, + "loss": 1.1068, + "step": 2668 + }, + { + "epoch": 1.4489685124864278, + "grad_norm": 7.249023283604718, + "learning_rate": 1.6665479091206476e-05, + "loss": 1.1025, + "step": 2669 + }, + { + "epoch": 1.4495114006514658, + "grad_norm": 6.407638163624444, + "learning_rate": 1.6662857008227145e-05, + "loss": 0.8236, + "step": 2670 + }, + { + "epoch": 1.4500542888165038, + "grad_norm": 7.289336395940614, + "learning_rate": 1.6660234101175036e-05, + "loss": 0.9386, + "step": 2671 + }, + { + "epoch": 1.4505971769815418, + "grad_norm": 6.973959594036107, + "learning_rate": 1.665761037037456e-05, + "loss": 0.8435, + "step": 2672 + }, + { + "epoch": 1.4511400651465798, + "grad_norm": 6.971928597160326, + "learning_rate": 1.665498581615023e-05, + "loss": 0.9065, + "step": 2673 + }, + { + "epoch": 1.4516829533116178, + "grad_norm": 6.1608850694063735, + "learning_rate": 1.665236043882664e-05, + "loss": 0.9289, + "step": 2674 + }, + { + "epoch": 1.4522258414766558, + "grad_norm": 6.62873967155635, + "learning_rate": 1.6649734238728512e-05, + "loss": 0.7797, + "step": 2675 + }, + { + "epoch": 1.4527687296416938, + "grad_norm": 7.089586063125632, + "learning_rate": 1.6647107216180655e-05, + "loss": 0.6306, + "step": 2676 + }, + { + "epoch": 1.4533116178067318, + "grad_norm": 10.986990416571043, + "learning_rate": 1.6644479371507985e-05, + "loss": 0.8764, + "step": 2677 + }, + { + "epoch": 1.4538545059717698, + "grad_norm": 5.865852703001031, + "learning_rate": 1.664185070503551e-05, + "loss": 0.9106, + "step": 2678 + }, + { + "epoch": 1.4543973941368078, + "grad_norm": 7.303323841843, + "learning_rate": 1.663922121708836e-05, + "loss": 0.8188, + "step": 2679 + }, + { + "epoch": 1.4549402823018458, + "grad_norm": 5.715280365147904, + "learning_rate": 1.663659090799175e-05, + "loss": 1.0573, + "step": 2680 + }, + { + "epoch": 1.4554831704668838, + "grad_norm": 7.084313061900368, + "learning_rate": 1.6633959778070992e-05, + "loss": 1.0749, + "step": 2681 + }, + { + "epoch": 1.4560260586319218, + "grad_norm": 7.902311581251679, + "learning_rate": 1.6631327827651524e-05, + "loss": 0.8677, + "step": 2682 + }, + { + "epoch": 1.4565689467969598, + "grad_norm": 8.1594384957164, + "learning_rate": 1.6628695057058855e-05, + "loss": 0.9999, + "step": 2683 + }, + { + "epoch": 1.4571118349619978, + "grad_norm": 5.580141751242927, + "learning_rate": 1.6626061466618623e-05, + "loss": 0.6387, + "step": 2684 + }, + { + "epoch": 1.4576547231270358, + "grad_norm": 8.145754034329466, + "learning_rate": 1.6623427056656544e-05, + "loss": 0.8984, + "step": 2685 + }, + { + "epoch": 1.4581976112920738, + "grad_norm": 6.453637820501138, + "learning_rate": 1.6620791827498454e-05, + "loss": 0.6922, + "step": 2686 + }, + { + "epoch": 1.4587404994571118, + "grad_norm": 7.3511834666776075, + "learning_rate": 1.6618155779470275e-05, + "loss": 0.8462, + "step": 2687 + }, + { + "epoch": 1.4592833876221498, + "grad_norm": 7.962874909840381, + "learning_rate": 1.6615518912898043e-05, + "loss": 0.9639, + "step": 2688 + }, + { + "epoch": 1.4598262757871878, + "grad_norm": 6.975602073560851, + "learning_rate": 1.6612881228107886e-05, + "loss": 0.9364, + "step": 2689 + }, + { + "epoch": 1.4603691639522258, + "grad_norm": 6.916712660881488, + "learning_rate": 1.6610242725426044e-05, + "loss": 0.8033, + "step": 2690 + }, + { + "epoch": 1.4609120521172638, + "grad_norm": 6.268208953592866, + "learning_rate": 1.6607603405178842e-05, + "loss": 0.7542, + "step": 2691 + }, + { + "epoch": 1.4614549402823018, + "grad_norm": 9.047138051843492, + "learning_rate": 1.660496326769272e-05, + "loss": 1.5956, + "step": 2692 + }, + { + "epoch": 1.4619978284473398, + "grad_norm": 7.228556514923675, + "learning_rate": 1.6602322313294216e-05, + "loss": 0.8461, + "step": 2693 + }, + { + "epoch": 1.4625407166123778, + "grad_norm": 6.195288408674038, + "learning_rate": 1.659968054230997e-05, + "loss": 0.9851, + "step": 2694 + }, + { + "epoch": 1.4630836047774158, + "grad_norm": 7.413116367868534, + "learning_rate": 1.6597037955066713e-05, + "loss": 0.9693, + "step": 2695 + }, + { + "epoch": 1.4636264929424538, + "grad_norm": 7.014031390984585, + "learning_rate": 1.6594394551891288e-05, + "loss": 1.0539, + "step": 2696 + }, + { + "epoch": 1.4641693811074918, + "grad_norm": 6.8196013668965225, + "learning_rate": 1.6591750333110634e-05, + "loss": 0.712, + "step": 2697 + }, + { + "epoch": 1.4647122692725298, + "grad_norm": 8.160690900976675, + "learning_rate": 1.658910529905179e-05, + "loss": 0.6551, + "step": 2698 + }, + { + "epoch": 1.4652551574375678, + "grad_norm": 5.174705288640052, + "learning_rate": 1.6586459450041906e-05, + "loss": 0.8285, + "step": 2699 + }, + { + "epoch": 1.4657980456026058, + "grad_norm": 7.029482771999071, + "learning_rate": 1.6583812786408216e-05, + "loss": 0.923, + "step": 2700 + }, + { + "epoch": 1.4663409337676439, + "grad_norm": 7.8877107275923946, + "learning_rate": 1.658116530847807e-05, + "loss": 1.1915, + "step": 2701 + }, + { + "epoch": 1.4668838219326819, + "grad_norm": 9.269305925663183, + "learning_rate": 1.657851701657891e-05, + "loss": 1.098, + "step": 2702 + }, + { + "epoch": 1.4674267100977199, + "grad_norm": 10.478546362437942, + "learning_rate": 1.657586791103828e-05, + "loss": 1.2416, + "step": 2703 + }, + { + "epoch": 1.4679695982627579, + "grad_norm": 8.207786471433751, + "learning_rate": 1.6573217992183826e-05, + "loss": 0.8658, + "step": 2704 + }, + { + "epoch": 1.4685124864277959, + "grad_norm": 7.862118410069405, + "learning_rate": 1.6570567260343294e-05, + "loss": 0.8714, + "step": 2705 + }, + { + "epoch": 1.4690553745928339, + "grad_norm": 7.744957075186954, + "learning_rate": 1.6567915715844534e-05, + "loss": 0.9374, + "step": 2706 + }, + { + "epoch": 1.4695982627578719, + "grad_norm": 6.97424297060321, + "learning_rate": 1.6565263359015488e-05, + "loss": 0.7183, + "step": 2707 + }, + { + "epoch": 1.4701411509229099, + "grad_norm": 7.349900880491086, + "learning_rate": 1.6562610190184206e-05, + "loss": 0.6066, + "step": 2708 + }, + { + "epoch": 1.4706840390879479, + "grad_norm": 6.982564531951699, + "learning_rate": 1.655995620967884e-05, + "loss": 1.2218, + "step": 2709 + }, + { + "epoch": 1.4712269272529859, + "grad_norm": 7.572476514419708, + "learning_rate": 1.6557301417827632e-05, + "loss": 0.89, + "step": 2710 + }, + { + "epoch": 1.4717698154180239, + "grad_norm": 8.654751087453503, + "learning_rate": 1.6554645814958932e-05, + "loss": 0.7188, + "step": 2711 + }, + { + "epoch": 1.4723127035830619, + "grad_norm": 6.282711052010393, + "learning_rate": 1.6551989401401196e-05, + "loss": 0.7586, + "step": 2712 + }, + { + "epoch": 1.4728555917480999, + "grad_norm": 8.732262593936317, + "learning_rate": 1.6549332177482966e-05, + "loss": 1.0178, + "step": 2713 + }, + { + "epoch": 1.4733984799131379, + "grad_norm": 6.989470691687101, + "learning_rate": 1.6546674143532895e-05, + "loss": 0.6981, + "step": 2714 + }, + { + "epoch": 1.4739413680781759, + "grad_norm": 7.179372740296565, + "learning_rate": 1.6544015299879734e-05, + "loss": 0.7617, + "step": 2715 + }, + { + "epoch": 1.4744842562432139, + "grad_norm": 9.117294639273645, + "learning_rate": 1.6541355646852327e-05, + "loss": 0.8822, + "step": 2716 + }, + { + "epoch": 1.475027144408252, + "grad_norm": 7.875727317127435, + "learning_rate": 1.653869518477963e-05, + "loss": 0.9149, + "step": 2717 + }, + { + "epoch": 1.47557003257329, + "grad_norm": 7.926910105501724, + "learning_rate": 1.6536033913990687e-05, + "loss": 0.8797, + "step": 2718 + }, + { + "epoch": 1.476112920738328, + "grad_norm": 5.14878192977047, + "learning_rate": 1.6533371834814657e-05, + "loss": 0.4491, + "step": 2719 + }, + { + "epoch": 1.476655808903366, + "grad_norm": 5.327596982546826, + "learning_rate": 1.6530708947580785e-05, + "loss": 0.4335, + "step": 2720 + }, + { + "epoch": 1.477198697068404, + "grad_norm": 9.158336314582998, + "learning_rate": 1.6528045252618423e-05, + "loss": 1.1208, + "step": 2721 + }, + { + "epoch": 1.477741585233442, + "grad_norm": 10.630652314970348, + "learning_rate": 1.6525380750257022e-05, + "loss": 1.097, + "step": 2722 + }, + { + "epoch": 1.47828447339848, + "grad_norm": 8.97220182564547, + "learning_rate": 1.652271544082613e-05, + "loss": 0.7838, + "step": 2723 + }, + { + "epoch": 1.478827361563518, + "grad_norm": 10.609251261516084, + "learning_rate": 1.652004932465539e-05, + "loss": 0.8861, + "step": 2724 + }, + { + "epoch": 1.479370249728556, + "grad_norm": 7.303801635012744, + "learning_rate": 1.6517382402074563e-05, + "loss": 0.5437, + "step": 2725 + }, + { + "epoch": 1.479913137893594, + "grad_norm": 11.770584674671793, + "learning_rate": 1.651471467341349e-05, + "loss": 1.4947, + "step": 2726 + }, + { + "epoch": 1.480456026058632, + "grad_norm": 6.824299484966605, + "learning_rate": 1.6512046139002128e-05, + "loss": 0.7144, + "step": 2727 + }, + { + "epoch": 1.48099891422367, + "grad_norm": 7.595224046166877, + "learning_rate": 1.650937679917052e-05, + "loss": 1.0259, + "step": 2728 + }, + { + "epoch": 1.481541802388708, + "grad_norm": 8.979875171863954, + "learning_rate": 1.6506706654248813e-05, + "loss": 1.0951, + "step": 2729 + }, + { + "epoch": 1.482084690553746, + "grad_norm": 7.700209331252355, + "learning_rate": 1.650403570456726e-05, + "loss": 1.0228, + "step": 2730 + }, + { + "epoch": 1.482627578718784, + "grad_norm": 9.090566362046857, + "learning_rate": 1.65013639504562e-05, + "loss": 1.0199, + "step": 2731 + }, + { + "epoch": 1.483170466883822, + "grad_norm": 6.644791741009562, + "learning_rate": 1.6498691392246088e-05, + "loss": 0.5478, + "step": 2732 + }, + { + "epoch": 1.48371335504886, + "grad_norm": 7.067718277939746, + "learning_rate": 1.6496018030267467e-05, + "loss": 0.864, + "step": 2733 + }, + { + "epoch": 1.484256243213898, + "grad_norm": 8.017737461642762, + "learning_rate": 1.6493343864850984e-05, + "loss": 1.1688, + "step": 2734 + }, + { + "epoch": 1.484799131378936, + "grad_norm": 6.479930584435578, + "learning_rate": 1.6490668896327382e-05, + "loss": 0.7227, + "step": 2735 + }, + { + "epoch": 1.485342019543974, + "grad_norm": 8.392289588009671, + "learning_rate": 1.6487993125027504e-05, + "loss": 0.7213, + "step": 2736 + }, + { + "epoch": 1.485884907709012, + "grad_norm": 7.817891248616158, + "learning_rate": 1.64853165512823e-05, + "loss": 0.7521, + "step": 2737 + }, + { + "epoch": 1.48642779587405, + "grad_norm": 8.824885297850575, + "learning_rate": 1.6482639175422804e-05, + "loss": 1.016, + "step": 2738 + }, + { + "epoch": 1.486970684039088, + "grad_norm": 11.48655873033237, + "learning_rate": 1.6479960997780165e-05, + "loss": 0.9814, + "step": 2739 + }, + { + "epoch": 1.487513572204126, + "grad_norm": 10.88221555712033, + "learning_rate": 1.6477282018685628e-05, + "loss": 1.464, + "step": 2740 + }, + { + "epoch": 1.488056460369164, + "grad_norm": 8.701247869952772, + "learning_rate": 1.6474602238470524e-05, + "loss": 0.918, + "step": 2741 + }, + { + "epoch": 1.488599348534202, + "grad_norm": 6.4327269926453505, + "learning_rate": 1.6471921657466294e-05, + "loss": 0.7949, + "step": 2742 + }, + { + "epoch": 1.48914223669924, + "grad_norm": 7.592315420975657, + "learning_rate": 1.6469240276004477e-05, + "loss": 0.9485, + "step": 2743 + }, + { + "epoch": 1.489685124864278, + "grad_norm": 7.682698616258038, + "learning_rate": 1.6466558094416717e-05, + "loss": 0.8439, + "step": 2744 + }, + { + "epoch": 1.490228013029316, + "grad_norm": 6.127626205001148, + "learning_rate": 1.6463875113034743e-05, + "loss": 0.8652, + "step": 2745 + }, + { + "epoch": 1.490770901194354, + "grad_norm": 9.65248236696942, + "learning_rate": 1.6461191332190397e-05, + "loss": 1.6278, + "step": 2746 + }, + { + "epoch": 1.491313789359392, + "grad_norm": 8.46309710838681, + "learning_rate": 1.6458506752215603e-05, + "loss": 1.5697, + "step": 2747 + }, + { + "epoch": 1.49185667752443, + "grad_norm": 6.578253971678976, + "learning_rate": 1.6455821373442407e-05, + "loss": 0.8836, + "step": 2748 + }, + { + "epoch": 1.492399565689468, + "grad_norm": 5.927853369454149, + "learning_rate": 1.645313519620293e-05, + "loss": 0.7182, + "step": 2749 + }, + { + "epoch": 1.492942453854506, + "grad_norm": 7.725891960915116, + "learning_rate": 1.645044822082941e-05, + "loss": 1.1794, + "step": 2750 + }, + { + "epoch": 1.493485342019544, + "grad_norm": 8.975220195142995, + "learning_rate": 1.644776044765417e-05, + "loss": 0.8909, + "step": 2751 + }, + { + "epoch": 1.494028230184582, + "grad_norm": 7.34643792387778, + "learning_rate": 1.6445071877009643e-05, + "loss": 1.0047, + "step": 2752 + }, + { + "epoch": 1.49457111834962, + "grad_norm": 7.561064198594243, + "learning_rate": 1.6442382509228355e-05, + "loss": 0.7533, + "step": 2753 + }, + { + "epoch": 1.495114006514658, + "grad_norm": 9.327085390245353, + "learning_rate": 1.6439692344642933e-05, + "loss": 0.8048, + "step": 2754 + }, + { + "epoch": 1.495656894679696, + "grad_norm": 8.662401438640156, + "learning_rate": 1.6437001383586095e-05, + "loss": 1.2513, + "step": 2755 + }, + { + "epoch": 1.496199782844734, + "grad_norm": 7.683989985812991, + "learning_rate": 1.6434309626390667e-05, + "loss": 0.8005, + "step": 2756 + }, + { + "epoch": 1.496742671009772, + "grad_norm": 6.295413547214335, + "learning_rate": 1.6431617073389574e-05, + "loss": 0.8778, + "step": 2757 + }, + { + "epoch": 1.49728555917481, + "grad_norm": 10.390849259618786, + "learning_rate": 1.6428923724915825e-05, + "loss": 1.6234, + "step": 2758 + }, + { + "epoch": 1.497828447339848, + "grad_norm": 6.815996692584369, + "learning_rate": 1.6426229581302545e-05, + "loss": 0.8142, + "step": 2759 + }, + { + "epoch": 1.498371335504886, + "grad_norm": 7.203388917066917, + "learning_rate": 1.642353464288295e-05, + "loss": 1.1273, + "step": 2760 + }, + { + "epoch": 1.498914223669924, + "grad_norm": 6.45381693158517, + "learning_rate": 1.6420838909990356e-05, + "loss": 0.7989, + "step": 2761 + }, + { + "epoch": 1.499457111834962, + "grad_norm": 7.135095705156932, + "learning_rate": 1.6418142382958167e-05, + "loss": 0.6753, + "step": 2762 + }, + { + "epoch": 1.5, + "grad_norm": 8.263302638252904, + "learning_rate": 1.64154450621199e-05, + "loss": 1.0357, + "step": 2763 + }, + { + "epoch": 1.500542888165038, + "grad_norm": 8.19537793615059, + "learning_rate": 1.6412746947809165e-05, + "loss": 0.8135, + "step": 2764 + }, + { + "epoch": 1.501085776330076, + "grad_norm": 5.084112201229359, + "learning_rate": 1.6410048040359665e-05, + "loss": 0.3466, + "step": 2765 + }, + { + "epoch": 1.501628664495114, + "grad_norm": 6.621661501175659, + "learning_rate": 1.6407348340105208e-05, + "loss": 1.1468, + "step": 2766 + }, + { + "epoch": 1.502171552660152, + "grad_norm": 7.5339066223692495, + "learning_rate": 1.6404647847379696e-05, + "loss": 0.8808, + "step": 2767 + }, + { + "epoch": 1.50271444082519, + "grad_norm": 6.501647846893153, + "learning_rate": 1.6401946562517134e-05, + "loss": 0.7589, + "step": 2768 + }, + { + "epoch": 1.503257328990228, + "grad_norm": 8.29356168583173, + "learning_rate": 1.6399244485851614e-05, + "loss": 1.3969, + "step": 2769 + }, + { + "epoch": 1.503800217155266, + "grad_norm": 6.518146966456392, + "learning_rate": 1.6396541617717337e-05, + "loss": 0.7951, + "step": 2770 + }, + { + "epoch": 1.504343105320304, + "grad_norm": 6.51134736256057, + "learning_rate": 1.63938379584486e-05, + "loss": 0.9073, + "step": 2771 + }, + { + "epoch": 1.504885993485342, + "grad_norm": 6.682459044312825, + "learning_rate": 1.6391133508379797e-05, + "loss": 0.733, + "step": 2772 + }, + { + "epoch": 1.50542888165038, + "grad_norm": 6.390687464891663, + "learning_rate": 1.638842826784541e-05, + "loss": 0.6934, + "step": 2773 + }, + { + "epoch": 1.505971769815418, + "grad_norm": 8.130197298030478, + "learning_rate": 1.6385722237180038e-05, + "loss": 0.8567, + "step": 2774 + }, + { + "epoch": 1.506514657980456, + "grad_norm": 7.959636836831885, + "learning_rate": 1.6383015416718356e-05, + "loss": 0.9926, + "step": 2775 + }, + { + "epoch": 1.507057546145494, + "grad_norm": 7.732366564125716, + "learning_rate": 1.638030780679516e-05, + "loss": 1.2228, + "step": 2776 + }, + { + "epoch": 1.507600434310532, + "grad_norm": 4.633651796670195, + "learning_rate": 1.6377599407745324e-05, + "loss": 0.4633, + "step": 2777 + }, + { + "epoch": 1.50814332247557, + "grad_norm": 7.016748827947926, + "learning_rate": 1.6374890219903828e-05, + "loss": 0.9593, + "step": 2778 + }, + { + "epoch": 1.508686210640608, + "grad_norm": 8.428067394567327, + "learning_rate": 1.637218024360575e-05, + "loss": 1.0493, + "step": 2779 + }, + { + "epoch": 1.509229098805646, + "grad_norm": 8.297567044395613, + "learning_rate": 1.6369469479186266e-05, + "loss": 1.2335, + "step": 2780 + }, + { + "epoch": 1.509771986970684, + "grad_norm": 5.8755950768313205, + "learning_rate": 1.6366757926980643e-05, + "loss": 0.6556, + "step": 2781 + }, + { + "epoch": 1.510314875135722, + "grad_norm": 6.837923725489964, + "learning_rate": 1.6364045587324254e-05, + "loss": 0.7031, + "step": 2782 + }, + { + "epoch": 1.51085776330076, + "grad_norm": 6.718777411432131, + "learning_rate": 1.6361332460552565e-05, + "loss": 0.8982, + "step": 2783 + }, + { + "epoch": 1.511400651465798, + "grad_norm": 8.788667184000067, + "learning_rate": 1.6358618547001137e-05, + "loss": 0.8548, + "step": 2784 + }, + { + "epoch": 1.511943539630836, + "grad_norm": 6.559897969720367, + "learning_rate": 1.635590384700563e-05, + "loss": 0.9627, + "step": 2785 + }, + { + "epoch": 1.512486427795874, + "grad_norm": 6.538968751920423, + "learning_rate": 1.635318836090181e-05, + "loss": 1.0731, + "step": 2786 + }, + { + "epoch": 1.513029315960912, + "grad_norm": 8.149646454407613, + "learning_rate": 1.6350472089025523e-05, + "loss": 0.9898, + "step": 2787 + }, + { + "epoch": 1.51357220412595, + "grad_norm": 8.544281989077668, + "learning_rate": 1.6347755031712734e-05, + "loss": 1.1125, + "step": 2788 + }, + { + "epoch": 1.514115092290988, + "grad_norm": 6.930736082842491, + "learning_rate": 1.634503718929948e-05, + "loss": 0.7617, + "step": 2789 + }, + { + "epoch": 1.514657980456026, + "grad_norm": 8.527931500643358, + "learning_rate": 1.6342318562121916e-05, + "loss": 1.2179, + "step": 2790 + }, + { + "epoch": 1.515200868621064, + "grad_norm": 7.509610271006853, + "learning_rate": 1.6339599150516283e-05, + "loss": 0.908, + "step": 2791 + }, + { + "epoch": 1.515743756786102, + "grad_norm": 6.0902820346277045, + "learning_rate": 1.6336878954818926e-05, + "loss": 0.829, + "step": 2792 + }, + { + "epoch": 1.51628664495114, + "grad_norm": 7.3274444131512935, + "learning_rate": 1.6334157975366278e-05, + "loss": 0.9667, + "step": 2793 + }, + { + "epoch": 1.516829533116178, + "grad_norm": 7.753641450951333, + "learning_rate": 1.633143621249488e-05, + "loss": 0.7961, + "step": 2794 + }, + { + "epoch": 1.517372421281216, + "grad_norm": 14.786266283108317, + "learning_rate": 1.6328713666541357e-05, + "loss": 1.4119, + "step": 2795 + }, + { + "epoch": 1.517915309446254, + "grad_norm": 6.1388168060810875, + "learning_rate": 1.632599033784244e-05, + "loss": 0.7726, + "step": 2796 + }, + { + "epoch": 1.518458197611292, + "grad_norm": 9.134190897524304, + "learning_rate": 1.632326622673496e-05, + "loss": 1.208, + "step": 2797 + }, + { + "epoch": 1.51900108577633, + "grad_norm": 6.898277731717493, + "learning_rate": 1.632054133355583e-05, + "loss": 0.6929, + "step": 2798 + }, + { + "epoch": 1.519543973941368, + "grad_norm": 6.90513711625334, + "learning_rate": 1.631781565864208e-05, + "loss": 0.7279, + "step": 2799 + }, + { + "epoch": 1.520086862106406, + "grad_norm": 8.332173251791465, + "learning_rate": 1.6315089202330817e-05, + "loss": 0.987, + "step": 2800 + }, + { + "epoch": 1.520629750271444, + "grad_norm": 6.783078863499027, + "learning_rate": 1.631236196495926e-05, + "loss": 0.6793, + "step": 2801 + }, + { + "epoch": 1.521172638436482, + "grad_norm": 7.841204229725534, + "learning_rate": 1.6309633946864712e-05, + "loss": 0.832, + "step": 2802 + }, + { + "epoch": 1.52171552660152, + "grad_norm": 6.4931904342190405, + "learning_rate": 1.630690514838458e-05, + "loss": 0.6662, + "step": 2803 + }, + { + "epoch": 1.522258414766558, + "grad_norm": 8.50187086749661, + "learning_rate": 1.6304175569856368e-05, + "loss": 0.8359, + "step": 2804 + }, + { + "epoch": 1.522801302931596, + "grad_norm": 6.196174007418195, + "learning_rate": 1.6301445211617676e-05, + "loss": 0.8437, + "step": 2805 + }, + { + "epoch": 1.523344191096634, + "grad_norm": 8.232047070633076, + "learning_rate": 1.6298714074006196e-05, + "loss": 0.9134, + "step": 2806 + }, + { + "epoch": 1.523887079261672, + "grad_norm": 6.594428796101092, + "learning_rate": 1.629598215735972e-05, + "loss": 0.6703, + "step": 2807 + }, + { + "epoch": 1.52442996742671, + "grad_norm": 7.634546429927066, + "learning_rate": 1.629324946201614e-05, + "loss": 0.8208, + "step": 2808 + }, + { + "epoch": 1.524972855591748, + "grad_norm": 8.225245989940067, + "learning_rate": 1.6290515988313432e-05, + "loss": 1.0002, + "step": 2809 + }, + { + "epoch": 1.5255157437567861, + "grad_norm": 8.334512060802671, + "learning_rate": 1.628778173658968e-05, + "loss": 0.8883, + "step": 2810 + }, + { + "epoch": 1.5260586319218241, + "grad_norm": 7.733244019009635, + "learning_rate": 1.6285046707183068e-05, + "loss": 0.7155, + "step": 2811 + }, + { + "epoch": 1.5266015200868621, + "grad_norm": 9.607081900344873, + "learning_rate": 1.628231090043186e-05, + "loss": 0.8321, + "step": 2812 + }, + { + "epoch": 1.5271444082519001, + "grad_norm": 8.390046905361096, + "learning_rate": 1.6279574316674426e-05, + "loss": 0.8818, + "step": 2813 + }, + { + "epoch": 1.5276872964169381, + "grad_norm": 7.608191285481891, + "learning_rate": 1.6276836956249235e-05, + "loss": 0.7382, + "step": 2814 + }, + { + "epoch": 1.5282301845819761, + "grad_norm": 8.998269112751723, + "learning_rate": 1.6274098819494844e-05, + "loss": 0.9285, + "step": 2815 + }, + { + "epoch": 1.5287730727470141, + "grad_norm": 5.877268783539307, + "learning_rate": 1.627135990674991e-05, + "loss": 0.8434, + "step": 2816 + }, + { + "epoch": 1.5293159609120521, + "grad_norm": 6.956730702284698, + "learning_rate": 1.6268620218353188e-05, + "loss": 0.5463, + "step": 2817 + }, + { + "epoch": 1.5298588490770901, + "grad_norm": 6.516430093047641, + "learning_rate": 1.626587975464353e-05, + "loss": 0.4886, + "step": 2818 + }, + { + "epoch": 1.5304017372421281, + "grad_norm": 6.751333342823053, + "learning_rate": 1.626313851595987e-05, + "loss": 0.6279, + "step": 2819 + }, + { + "epoch": 1.5309446254071661, + "grad_norm": 8.120808986494415, + "learning_rate": 1.6260396502641264e-05, + "loss": 0.8192, + "step": 2820 + }, + { + "epoch": 1.5314875135722041, + "grad_norm": 7.570228095805488, + "learning_rate": 1.6257653715026837e-05, + "loss": 0.5813, + "step": 2821 + }, + { + "epoch": 1.5320304017372421, + "grad_norm": 8.436330907807918, + "learning_rate": 1.625491015345583e-05, + "loss": 0.9115, + "step": 2822 + }, + { + "epoch": 1.5325732899022801, + "grad_norm": 8.384048301051982, + "learning_rate": 1.6252165818267564e-05, + "loss": 1.0131, + "step": 2823 + }, + { + "epoch": 1.5331161780673181, + "grad_norm": 6.665532827172479, + "learning_rate": 1.6249420709801462e-05, + "loss": 0.5454, + "step": 2824 + }, + { + "epoch": 1.5336590662323561, + "grad_norm": 7.096056294730765, + "learning_rate": 1.624667482839705e-05, + "loss": 1.043, + "step": 2825 + }, + { + "epoch": 1.5342019543973942, + "grad_norm": 5.384161272714602, + "learning_rate": 1.6243928174393935e-05, + "loss": 0.8345, + "step": 2826 + }, + { + "epoch": 1.5347448425624322, + "grad_norm": 8.398527018753486, + "learning_rate": 1.6241180748131834e-05, + "loss": 0.6911, + "step": 2827 + }, + { + "epoch": 1.5352877307274702, + "grad_norm": 6.880924484125559, + "learning_rate": 1.6238432549950552e-05, + "loss": 0.6039, + "step": 2828 + }, + { + "epoch": 1.5358306188925082, + "grad_norm": 9.23669884568822, + "learning_rate": 1.623568358018999e-05, + "loss": 1.1145, + "step": 2829 + }, + { + "epoch": 1.5363735070575462, + "grad_norm": 9.838376476241027, + "learning_rate": 1.6232933839190146e-05, + "loss": 1.0435, + "step": 2830 + }, + { + "epoch": 1.5369163952225842, + "grad_norm": 8.64475205930121, + "learning_rate": 1.6230183327291108e-05, + "loss": 1.0886, + "step": 2831 + }, + { + "epoch": 1.5374592833876222, + "grad_norm": 8.446847258487077, + "learning_rate": 1.6227432044833072e-05, + "loss": 0.8922, + "step": 2832 + }, + { + "epoch": 1.5380021715526602, + "grad_norm": 7.373724851347831, + "learning_rate": 1.622467999215631e-05, + "loss": 0.7823, + "step": 2833 + }, + { + "epoch": 1.5385450597176982, + "grad_norm": 9.661226436708493, + "learning_rate": 1.622192716960121e-05, + "loss": 1.2651, + "step": 2834 + }, + { + "epoch": 1.5390879478827362, + "grad_norm": 9.316115050127708, + "learning_rate": 1.6219173577508237e-05, + "loss": 0.897, + "step": 2835 + }, + { + "epoch": 1.5396308360477742, + "grad_norm": 12.011440258586251, + "learning_rate": 1.621641921621797e-05, + "loss": 0.868, + "step": 2836 + }, + { + "epoch": 1.5401737242128122, + "grad_norm": 7.737580132404622, + "learning_rate": 1.6213664086071058e-05, + "loss": 1.0727, + "step": 2837 + }, + { + "epoch": 1.5407166123778502, + "grad_norm": 6.696935187657473, + "learning_rate": 1.6210908187408275e-05, + "loss": 0.7404, + "step": 2838 + }, + { + "epoch": 1.5412595005428882, + "grad_norm": 6.460846604904407, + "learning_rate": 1.6208151520570465e-05, + "loss": 0.5892, + "step": 2839 + }, + { + "epoch": 1.5418023887079262, + "grad_norm": 8.244871763303658, + "learning_rate": 1.6205394085898586e-05, + "loss": 1.0692, + "step": 2840 + }, + { + "epoch": 1.5423452768729642, + "grad_norm": 6.140143281113326, + "learning_rate": 1.620263588373367e-05, + "loss": 0.7648, + "step": 2841 + }, + { + "epoch": 1.5428881650380022, + "grad_norm": 6.3369705303866235, + "learning_rate": 1.619987691441687e-05, + "loss": 0.6665, + "step": 2842 + }, + { + "epoch": 1.5434310532030402, + "grad_norm": 8.864356355873657, + "learning_rate": 1.6197117178289405e-05, + "loss": 0.9855, + "step": 2843 + }, + { + "epoch": 1.5439739413680782, + "grad_norm": 8.487808912288623, + "learning_rate": 1.6194356675692614e-05, + "loss": 1.3813, + "step": 2844 + }, + { + "epoch": 1.5445168295331162, + "grad_norm": 6.5782618837581435, + "learning_rate": 1.619159540696792e-05, + "loss": 1.1941, + "step": 2845 + }, + { + "epoch": 1.5450597176981542, + "grad_norm": 8.668580644259926, + "learning_rate": 1.6188833372456833e-05, + "loss": 1.1044, + "step": 2846 + }, + { + "epoch": 1.5456026058631922, + "grad_norm": 8.677506532765584, + "learning_rate": 1.6186070572500972e-05, + "loss": 1.3984, + "step": 2847 + }, + { + "epoch": 1.5461454940282302, + "grad_norm": 6.822422879140709, + "learning_rate": 1.6183307007442046e-05, + "loss": 0.6903, + "step": 2848 + }, + { + "epoch": 1.5466883821932682, + "grad_norm": 8.480828664812217, + "learning_rate": 1.6180542677621852e-05, + "loss": 1.2466, + "step": 2849 + }, + { + "epoch": 1.5472312703583062, + "grad_norm": 7.804300661419004, + "learning_rate": 1.617777758338229e-05, + "loss": 0.8833, + "step": 2850 + }, + { + "epoch": 1.5477741585233442, + "grad_norm": 6.418809144902363, + "learning_rate": 1.617501172506535e-05, + "loss": 0.9174, + "step": 2851 + }, + { + "epoch": 1.5483170466883822, + "grad_norm": 6.627083812958508, + "learning_rate": 1.617224510301312e-05, + "loss": 0.7867, + "step": 2852 + }, + { + "epoch": 1.5488599348534202, + "grad_norm": 7.792114190076133, + "learning_rate": 1.616947771756778e-05, + "loss": 0.6242, + "step": 2853 + }, + { + "epoch": 1.5494028230184582, + "grad_norm": 6.965108246318713, + "learning_rate": 1.6166709569071598e-05, + "loss": 0.5844, + "step": 2854 + }, + { + "epoch": 1.5499457111834962, + "grad_norm": 6.9966373703206655, + "learning_rate": 1.616394065786695e-05, + "loss": 1.1688, + "step": 2855 + }, + { + "epoch": 1.5504885993485342, + "grad_norm": 8.466603441198636, + "learning_rate": 1.6161170984296298e-05, + "loss": 0.9506, + "step": 2856 + }, + { + "epoch": 1.5510314875135722, + "grad_norm": 6.361127259105106, + "learning_rate": 1.61584005487022e-05, + "loss": 0.5573, + "step": 2857 + }, + { + "epoch": 1.5515743756786102, + "grad_norm": 6.137955961198825, + "learning_rate": 1.6155629351427306e-05, + "loss": 1.0474, + "step": 2858 + }, + { + "epoch": 1.5521172638436482, + "grad_norm": 6.262211837195136, + "learning_rate": 1.6152857392814367e-05, + "loss": 0.6786, + "step": 2859 + }, + { + "epoch": 1.5526601520086862, + "grad_norm": 9.3980797837583, + "learning_rate": 1.6150084673206214e-05, + "loss": 1.2893, + "step": 2860 + }, + { + "epoch": 1.5532030401737242, + "grad_norm": 9.600537200283373, + "learning_rate": 1.614731119294579e-05, + "loss": 0.8668, + "step": 2861 + }, + { + "epoch": 1.5537459283387622, + "grad_norm": 8.65304941049683, + "learning_rate": 1.614453695237612e-05, + "loss": 1.0377, + "step": 2862 + }, + { + "epoch": 1.5542888165038002, + "grad_norm": 6.075747510405264, + "learning_rate": 1.6141761951840327e-05, + "loss": 0.8203, + "step": 2863 + }, + { + "epoch": 1.5548317046688382, + "grad_norm": 6.578770531031146, + "learning_rate": 1.6138986191681626e-05, + "loss": 0.6808, + "step": 2864 + }, + { + "epoch": 1.5553745928338762, + "grad_norm": 9.206515322081769, + "learning_rate": 1.6136209672243332e-05, + "loss": 0.9111, + "step": 2865 + }, + { + "epoch": 1.5559174809989142, + "grad_norm": 8.829723041453656, + "learning_rate": 1.613343239386884e-05, + "loss": 1.7392, + "step": 2866 + }, + { + "epoch": 1.5564603691639523, + "grad_norm": 9.489367814866748, + "learning_rate": 1.613065435690166e-05, + "loss": 0.9186, + "step": 2867 + }, + { + "epoch": 1.5570032573289903, + "grad_norm": 7.7196736231335175, + "learning_rate": 1.6127875561685376e-05, + "loss": 0.769, + "step": 2868 + }, + { + "epoch": 1.5575461454940283, + "grad_norm": 7.668511006453606, + "learning_rate": 1.6125096008563677e-05, + "loss": 0.7572, + "step": 2869 + }, + { + "epoch": 1.5580890336590663, + "grad_norm": 8.31551289254719, + "learning_rate": 1.6122315697880343e-05, + "loss": 1.0657, + "step": 2870 + }, + { + "epoch": 1.5586319218241043, + "grad_norm": 11.828690633927348, + "learning_rate": 1.6119534629979244e-05, + "loss": 1.0095, + "step": 2871 + }, + { + "epoch": 1.5591748099891423, + "grad_norm": 7.137823181609677, + "learning_rate": 1.611675280520435e-05, + "loss": 0.7246, + "step": 2872 + }, + { + "epoch": 1.5597176981541803, + "grad_norm": 5.051313327297567, + "learning_rate": 1.611397022389972e-05, + "loss": 0.5905, + "step": 2873 + }, + { + "epoch": 1.5602605863192183, + "grad_norm": 8.165805991671068, + "learning_rate": 1.6111186886409504e-05, + "loss": 1.0678, + "step": 2874 + }, + { + "epoch": 1.5608034744842563, + "grad_norm": 11.597649604311457, + "learning_rate": 1.6108402793077957e-05, + "loss": 1.4836, + "step": 2875 + }, + { + "epoch": 1.5613463626492943, + "grad_norm": 7.953203184026488, + "learning_rate": 1.610561794424942e-05, + "loss": 0.9216, + "step": 2876 + }, + { + "epoch": 1.5618892508143323, + "grad_norm": 7.397930684647979, + "learning_rate": 1.6102832340268322e-05, + "loss": 0.6688, + "step": 2877 + }, + { + "epoch": 1.5624321389793703, + "grad_norm": 8.999084750858366, + "learning_rate": 1.6100045981479195e-05, + "loss": 1.254, + "step": 2878 + }, + { + "epoch": 1.5629750271444083, + "grad_norm": 7.445594920489225, + "learning_rate": 1.6097258868226658e-05, + "loss": 0.6406, + "step": 2879 + }, + { + "epoch": 1.5635179153094463, + "grad_norm": 10.469355169581702, + "learning_rate": 1.609447100085543e-05, + "loss": 0.7928, + "step": 2880 + }, + { + "epoch": 1.5640608034744843, + "grad_norm": 8.853634937458189, + "learning_rate": 1.6091682379710313e-05, + "loss": 1.0788, + "step": 2881 + }, + { + "epoch": 1.5646036916395223, + "grad_norm": 7.909620213759866, + "learning_rate": 1.6088893005136206e-05, + "loss": 1.011, + "step": 2882 + }, + { + "epoch": 1.5651465798045603, + "grad_norm": 10.379252728708625, + "learning_rate": 1.6086102877478117e-05, + "loss": 0.8675, + "step": 2883 + }, + { + "epoch": 1.5656894679695983, + "grad_norm": 8.143888785034273, + "learning_rate": 1.6083311997081116e-05, + "loss": 0.8935, + "step": 2884 + }, + { + "epoch": 1.5662323561346363, + "grad_norm": 8.577520797942114, + "learning_rate": 1.6080520364290396e-05, + "loss": 0.6763, + "step": 2885 + }, + { + "epoch": 1.5667752442996743, + "grad_norm": 9.226094142154105, + "learning_rate": 1.6077727979451228e-05, + "loss": 0.6679, + "step": 2886 + }, + { + "epoch": 1.5673181324647123, + "grad_norm": 11.197536431412058, + "learning_rate": 1.607493484290897e-05, + "loss": 1.0525, + "step": 2887 + }, + { + "epoch": 1.5678610206297503, + "grad_norm": 8.43012225245798, + "learning_rate": 1.6072140955009093e-05, + "loss": 1.0317, + "step": 2888 + }, + { + "epoch": 1.5684039087947883, + "grad_norm": 6.69358812320454, + "learning_rate": 1.606934631609715e-05, + "loss": 0.7703, + "step": 2889 + }, + { + "epoch": 1.5689467969598263, + "grad_norm": 9.36327291463899, + "learning_rate": 1.6066550926518776e-05, + "loss": 1.0844, + "step": 2890 + }, + { + "epoch": 1.5694896851248643, + "grad_norm": 8.40518653624583, + "learning_rate": 1.6063754786619716e-05, + "loss": 0.6863, + "step": 2891 + }, + { + "epoch": 1.5700325732899023, + "grad_norm": 9.552486022227878, + "learning_rate": 1.60609578967458e-05, + "loss": 1.0342, + "step": 2892 + }, + { + "epoch": 1.5705754614549403, + "grad_norm": 6.76445711513327, + "learning_rate": 1.6058160257242953e-05, + "loss": 0.8777, + "step": 2893 + }, + { + "epoch": 1.5711183496199783, + "grad_norm": 9.040602694092291, + "learning_rate": 1.6055361868457188e-05, + "loss": 0.9634, + "step": 2894 + }, + { + "epoch": 1.5716612377850163, + "grad_norm": 7.00278619777266, + "learning_rate": 1.6052562730734614e-05, + "loss": 0.8995, + "step": 2895 + }, + { + "epoch": 1.5722041259500543, + "grad_norm": 8.533327622490077, + "learning_rate": 1.604976284442144e-05, + "loss": 1.0896, + "step": 2896 + }, + { + "epoch": 1.5727470141150923, + "grad_norm": 9.895218231842426, + "learning_rate": 1.6046962209863953e-05, + "loss": 0.9309, + "step": 2897 + }, + { + "epoch": 1.5732899022801303, + "grad_norm": 5.945823958044379, + "learning_rate": 1.604416082740854e-05, + "loss": 0.6512, + "step": 2898 + }, + { + "epoch": 1.5738327904451683, + "grad_norm": 10.822429137471252, + "learning_rate": 1.6041358697401687e-05, + "loss": 1.0744, + "step": 2899 + }, + { + "epoch": 1.5743756786102063, + "grad_norm": 7.728147258511819, + "learning_rate": 1.603855582018996e-05, + "loss": 0.9566, + "step": 2900 + }, + { + "epoch": 1.5749185667752443, + "grad_norm": 8.643748761880973, + "learning_rate": 1.603575219612003e-05, + "loss": 0.9188, + "step": 2901 + }, + { + "epoch": 1.5754614549402823, + "grad_norm": 9.357572097357403, + "learning_rate": 1.603294782553864e-05, + "loss": 0.9045, + "step": 2902 + }, + { + "epoch": 1.5760043431053203, + "grad_norm": 6.76207345266047, + "learning_rate": 1.6030142708792653e-05, + "loss": 0.7658, + "step": 2903 + }, + { + "epoch": 1.5765472312703583, + "grad_norm": 6.47603683733189, + "learning_rate": 1.6027336846229005e-05, + "loss": 0.5406, + "step": 2904 + }, + { + "epoch": 1.5770901194353963, + "grad_norm": 6.030033648190992, + "learning_rate": 1.602453023819473e-05, + "loss": 0.5141, + "step": 2905 + }, + { + "epoch": 1.5776330076004343, + "grad_norm": 8.344821553681937, + "learning_rate": 1.6021722885036954e-05, + "loss": 1.255, + "step": 2906 + }, + { + "epoch": 1.5781758957654723, + "grad_norm": 7.2392259468937885, + "learning_rate": 1.601891478710289e-05, + "loss": 0.8358, + "step": 2907 + }, + { + "epoch": 1.5787187839305103, + "grad_norm": 8.011046574978268, + "learning_rate": 1.6016105944739856e-05, + "loss": 1.0316, + "step": 2908 + }, + { + "epoch": 1.5792616720955484, + "grad_norm": 6.609140115647256, + "learning_rate": 1.601329635829525e-05, + "loss": 0.7324, + "step": 2909 + }, + { + "epoch": 1.5798045602605864, + "grad_norm": 6.672864395642082, + "learning_rate": 1.6010486028116568e-05, + "loss": 0.6386, + "step": 2910 + }, + { + "epoch": 1.5803474484256244, + "grad_norm": 7.270447246601273, + "learning_rate": 1.600767495455139e-05, + "loss": 0.6699, + "step": 2911 + }, + { + "epoch": 1.5808903365906624, + "grad_norm": 9.7186017734922, + "learning_rate": 1.6004863137947405e-05, + "loss": 0.9604, + "step": 2912 + }, + { + "epoch": 1.5814332247557004, + "grad_norm": 7.18305646407285, + "learning_rate": 1.6002050578652374e-05, + "loss": 0.7286, + "step": 2913 + }, + { + "epoch": 1.5819761129207384, + "grad_norm": 8.171408196684215, + "learning_rate": 1.5999237277014162e-05, + "loss": 0.9121, + "step": 2914 + }, + { + "epoch": 1.5825190010857764, + "grad_norm": 7.710673421134947, + "learning_rate": 1.599642323338072e-05, + "loss": 0.7287, + "step": 2915 + }, + { + "epoch": 1.5830618892508144, + "grad_norm": 8.264934044661224, + "learning_rate": 1.5993608448100095e-05, + "loss": 0.6537, + "step": 2916 + }, + { + "epoch": 1.5836047774158524, + "grad_norm": 7.874914961028911, + "learning_rate": 1.599079292152043e-05, + "loss": 0.6511, + "step": 2917 + }, + { + "epoch": 1.5841476655808904, + "grad_norm": 7.399378371785218, + "learning_rate": 1.5987976653989945e-05, + "loss": 0.6236, + "step": 2918 + }, + { + "epoch": 1.5846905537459284, + "grad_norm": 7.837764322959507, + "learning_rate": 1.5985159645856966e-05, + "loss": 0.8184, + "step": 2919 + }, + { + "epoch": 1.5852334419109664, + "grad_norm": 8.64272103945922, + "learning_rate": 1.5982341897469903e-05, + "loss": 0.9182, + "step": 2920 + }, + { + "epoch": 1.5857763300760044, + "grad_norm": 9.994158036669248, + "learning_rate": 1.5979523409177254e-05, + "loss": 1.3466, + "step": 2921 + }, + { + "epoch": 1.5863192182410424, + "grad_norm": 6.671268632879107, + "learning_rate": 1.5976704181327626e-05, + "loss": 0.7506, + "step": 2922 + }, + { + "epoch": 1.5868621064060804, + "grad_norm": 8.161895609948308, + "learning_rate": 1.59738842142697e-05, + "loss": 0.9683, + "step": 2923 + }, + { + "epoch": 1.5874049945711184, + "grad_norm": 7.236619510875819, + "learning_rate": 1.597106350835225e-05, + "loss": 0.798, + "step": 2924 + }, + { + "epoch": 1.5879478827361564, + "grad_norm": 10.263209688443808, + "learning_rate": 1.5968242063924152e-05, + "loss": 1.0799, + "step": 2925 + }, + { + "epoch": 1.5884907709011944, + "grad_norm": 9.433308260958457, + "learning_rate": 1.596541988133436e-05, + "loss": 0.8755, + "step": 2926 + }, + { + "epoch": 1.5890336590662324, + "grad_norm": 9.48995441422385, + "learning_rate": 1.5962596960931927e-05, + "loss": 1.4712, + "step": 2927 + }, + { + "epoch": 1.5895765472312704, + "grad_norm": 7.317297215230638, + "learning_rate": 1.5959773303066005e-05, + "loss": 1.0421, + "step": 2928 + }, + { + "epoch": 1.5901194353963084, + "grad_norm": 7.491426794362114, + "learning_rate": 1.595694890808582e-05, + "loss": 0.9468, + "step": 2929 + }, + { + "epoch": 1.5906623235613464, + "grad_norm": 7.320440852081977, + "learning_rate": 1.5954123776340702e-05, + "loss": 0.7459, + "step": 2930 + }, + { + "epoch": 1.5912052117263844, + "grad_norm": 8.334158307527314, + "learning_rate": 1.5951297908180062e-05, + "loss": 1.3182, + "step": 2931 + }, + { + "epoch": 1.5917480998914224, + "grad_norm": 7.02613629768153, + "learning_rate": 1.5948471303953418e-05, + "loss": 0.919, + "step": 2932 + }, + { + "epoch": 1.5922909880564604, + "grad_norm": 6.904443487786306, + "learning_rate": 1.594564396401036e-05, + "loss": 0.5737, + "step": 2933 + }, + { + "epoch": 1.5928338762214984, + "grad_norm": 8.041576521446402, + "learning_rate": 1.594281588870058e-05, + "loss": 0.7191, + "step": 2934 + }, + { + "epoch": 1.5933767643865364, + "grad_norm": 8.557472225459021, + "learning_rate": 1.5939987078373856e-05, + "loss": 0.8936, + "step": 2935 + }, + { + "epoch": 1.5939196525515744, + "grad_norm": 10.051674589973098, + "learning_rate": 1.5937157533380065e-05, + "loss": 1.1777, + "step": 2936 + }, + { + "epoch": 1.5944625407166124, + "grad_norm": 7.1580018473739235, + "learning_rate": 1.5934327254069167e-05, + "loss": 0.8628, + "step": 2937 + }, + { + "epoch": 1.5950054288816504, + "grad_norm": 8.689338133459966, + "learning_rate": 1.593149624079122e-05, + "loss": 1.1877, + "step": 2938 + }, + { + "epoch": 1.5955483170466884, + "grad_norm": 8.40904043481009, + "learning_rate": 1.5928664493896364e-05, + "loss": 0.9815, + "step": 2939 + }, + { + "epoch": 1.5960912052117264, + "grad_norm": 7.778305354024487, + "learning_rate": 1.5925832013734832e-05, + "loss": 0.8282, + "step": 2940 + }, + { + "epoch": 1.5966340933767644, + "grad_norm": 7.074281482828989, + "learning_rate": 1.5922998800656956e-05, + "loss": 0.8347, + "step": 2941 + }, + { + "epoch": 1.5971769815418024, + "grad_norm": 9.188840060350078, + "learning_rate": 1.5920164855013145e-05, + "loss": 0.8566, + "step": 2942 + }, + { + "epoch": 1.5977198697068404, + "grad_norm": 7.021116417649334, + "learning_rate": 1.591733017715391e-05, + "loss": 0.8367, + "step": 2943 + }, + { + "epoch": 1.5982627578718784, + "grad_norm": 8.039041768789302, + "learning_rate": 1.5914494767429846e-05, + "loss": 0.9889, + "step": 2944 + }, + { + "epoch": 1.5988056460369164, + "grad_norm": 6.252701385804125, + "learning_rate": 1.5911658626191645e-05, + "loss": 1.0478, + "step": 2945 + }, + { + "epoch": 1.5993485342019544, + "grad_norm": 8.223409826369243, + "learning_rate": 1.5908821753790083e-05, + "loss": 1.2458, + "step": 2946 + }, + { + "epoch": 1.5998914223669924, + "grad_norm": 8.972690403999096, + "learning_rate": 1.590598415057603e-05, + "loss": 0.8471, + "step": 2947 + }, + { + "epoch": 1.6004343105320304, + "grad_norm": 8.163032020074416, + "learning_rate": 1.5903145816900445e-05, + "loss": 0.8213, + "step": 2948 + }, + { + "epoch": 1.6009771986970684, + "grad_norm": 5.667868413023322, + "learning_rate": 1.5900306753114375e-05, + "loss": 0.8426, + "step": 2949 + }, + { + "epoch": 1.6015200868621065, + "grad_norm": 6.960744978365833, + "learning_rate": 1.5897466959568967e-05, + "loss": 0.7224, + "step": 2950 + }, + { + "epoch": 1.6020629750271445, + "grad_norm": 7.648210267430006, + "learning_rate": 1.589462643661544e-05, + "loss": 0.7451, + "step": 2951 + }, + { + "epoch": 1.6026058631921825, + "grad_norm": 5.9568596208122875, + "learning_rate": 1.5891785184605123e-05, + "loss": 0.5252, + "step": 2952 + }, + { + "epoch": 1.6031487513572205, + "grad_norm": 6.41676218105273, + "learning_rate": 1.5888943203889427e-05, + "loss": 0.6502, + "step": 2953 + }, + { + "epoch": 1.6036916395222585, + "grad_norm": 7.704871607753619, + "learning_rate": 1.5886100494819846e-05, + "loss": 1.1762, + "step": 2954 + }, + { + "epoch": 1.6042345276872965, + "grad_norm": 7.656219757403813, + "learning_rate": 1.5883257057747975e-05, + "loss": 0.8076, + "step": 2955 + }, + { + "epoch": 1.6047774158523345, + "grad_norm": 6.897030921112048, + "learning_rate": 1.58804128930255e-05, + "loss": 0.883, + "step": 2956 + }, + { + "epoch": 1.6053203040173725, + "grad_norm": 6.718724904599124, + "learning_rate": 1.5877568001004182e-05, + "loss": 0.8533, + "step": 2957 + }, + { + "epoch": 1.6058631921824105, + "grad_norm": 8.458861423777194, + "learning_rate": 1.5874722382035887e-05, + "loss": 1.2017, + "step": 2958 + }, + { + "epoch": 1.6064060803474485, + "grad_norm": 9.16622725543938, + "learning_rate": 1.5871876036472565e-05, + "loss": 1.0074, + "step": 2959 + }, + { + "epoch": 1.6069489685124865, + "grad_norm": 7.053786350768386, + "learning_rate": 1.5869028964666254e-05, + "loss": 0.8193, + "step": 2960 + }, + { + "epoch": 1.6074918566775245, + "grad_norm": 7.670976634760575, + "learning_rate": 1.5866181166969088e-05, + "loss": 0.8331, + "step": 2961 + }, + { + "epoch": 1.6080347448425625, + "grad_norm": 9.942186867470006, + "learning_rate": 1.586333264373329e-05, + "loss": 0.8672, + "step": 2962 + }, + { + "epoch": 1.6085776330076005, + "grad_norm": 7.744463789862724, + "learning_rate": 1.586048339531116e-05, + "loss": 0.8859, + "step": 2963 + }, + { + "epoch": 1.6091205211726385, + "grad_norm": 7.181553947846623, + "learning_rate": 1.5857633422055104e-05, + "loss": 1.0376, + "step": 2964 + }, + { + "epoch": 1.6096634093376765, + "grad_norm": 7.184586780377298, + "learning_rate": 1.5854782724317616e-05, + "loss": 0.694, + "step": 2965 + }, + { + "epoch": 1.6102062975027145, + "grad_norm": 6.9154974167720855, + "learning_rate": 1.5851931302451262e-05, + "loss": 0.8573, + "step": 2966 + }, + { + "epoch": 1.6107491856677525, + "grad_norm": 5.056944401416877, + "learning_rate": 1.5849079156808726e-05, + "loss": 0.4972, + "step": 2967 + }, + { + "epoch": 1.6112920738327905, + "grad_norm": 6.98635366168859, + "learning_rate": 1.584622628774275e-05, + "loss": 0.9002, + "step": 2968 + }, + { + "epoch": 1.6118349619978285, + "grad_norm": 9.22418255248469, + "learning_rate": 1.5843372695606196e-05, + "loss": 1.4143, + "step": 2969 + }, + { + "epoch": 1.6123778501628665, + "grad_norm": 9.49766654897752, + "learning_rate": 1.584051838075199e-05, + "loss": 1.0145, + "step": 2970 + }, + { + "epoch": 1.6129207383279045, + "grad_norm": 9.681135461557052, + "learning_rate": 1.5837663343533166e-05, + "loss": 0.8841, + "step": 2971 + }, + { + "epoch": 1.6134636264929425, + "grad_norm": 6.00661489729409, + "learning_rate": 1.583480758430283e-05, + "loss": 0.677, + "step": 2972 + }, + { + "epoch": 1.6140065146579805, + "grad_norm": 7.843885336692957, + "learning_rate": 1.5831951103414194e-05, + "loss": 0.4729, + "step": 2973 + }, + { + "epoch": 1.6145494028230185, + "grad_norm": 9.391971866087287, + "learning_rate": 1.5829093901220557e-05, + "loss": 1.0833, + "step": 2974 + }, + { + "epoch": 1.6150922909880565, + "grad_norm": 7.202123614778415, + "learning_rate": 1.582623597807529e-05, + "loss": 0.9755, + "step": 2975 + }, + { + "epoch": 1.6156351791530945, + "grad_norm": 8.050976157508988, + "learning_rate": 1.5823377334331875e-05, + "loss": 0.8981, + "step": 2976 + }, + { + "epoch": 1.6161780673181325, + "grad_norm": 6.409986462897389, + "learning_rate": 1.5820517970343867e-05, + "loss": 0.3994, + "step": 2977 + }, + { + "epoch": 1.6167209554831705, + "grad_norm": 6.001133896600598, + "learning_rate": 1.581765788646492e-05, + "loss": 0.4701, + "step": 2978 + }, + { + "epoch": 1.6172638436482085, + "grad_norm": 12.858193971885504, + "learning_rate": 1.581479708304878e-05, + "loss": 1.4327, + "step": 2979 + }, + { + "epoch": 1.6178067318132465, + "grad_norm": 9.559044098763156, + "learning_rate": 1.5811935560449262e-05, + "loss": 0.9058, + "step": 2980 + }, + { + "epoch": 1.6183496199782845, + "grad_norm": 8.041859101089647, + "learning_rate": 1.5809073319020293e-05, + "loss": 0.6251, + "step": 2981 + }, + { + "epoch": 1.6188925081433225, + "grad_norm": 7.413714357196275, + "learning_rate": 1.580621035911588e-05, + "loss": 0.7887, + "step": 2982 + }, + { + "epoch": 1.6194353963083605, + "grad_norm": 11.941805543277479, + "learning_rate": 1.5803346681090113e-05, + "loss": 1.2547, + "step": 2983 + }, + { + "epoch": 1.6199782844733985, + "grad_norm": 5.728003242952451, + "learning_rate": 1.580048228529718e-05, + "loss": 0.4941, + "step": 2984 + }, + { + "epoch": 1.6205211726384365, + "grad_norm": 8.67398306592389, + "learning_rate": 1.5797617172091354e-05, + "loss": 0.7758, + "step": 2985 + }, + { + "epoch": 1.6210640608034745, + "grad_norm": 6.198163559825324, + "learning_rate": 1.5794751341826996e-05, + "loss": 0.4601, + "step": 2986 + }, + { + "epoch": 1.6216069489685125, + "grad_norm": 8.257469675141675, + "learning_rate": 1.5791884794858557e-05, + "loss": 0.8679, + "step": 2987 + }, + { + "epoch": 1.6221498371335505, + "grad_norm": 12.953868836673164, + "learning_rate": 1.5789017531540575e-05, + "loss": 1.572, + "step": 2988 + }, + { + "epoch": 1.6226927252985885, + "grad_norm": 9.232767459675694, + "learning_rate": 1.5786149552227682e-05, + "loss": 0.7511, + "step": 2989 + }, + { + "epoch": 1.6232356134636265, + "grad_norm": 7.465499623540017, + "learning_rate": 1.5783280857274586e-05, + "loss": 0.5754, + "step": 2990 + }, + { + "epoch": 1.6237785016286646, + "grad_norm": 5.087952971939767, + "learning_rate": 1.5780411447036097e-05, + "loss": 0.532, + "step": 2991 + }, + { + "epoch": 1.6243213897937026, + "grad_norm": 11.064786965172743, + "learning_rate": 1.577754132186711e-05, + "loss": 0.7195, + "step": 2992 + }, + { + "epoch": 1.6248642779587406, + "grad_norm": 10.268274657097892, + "learning_rate": 1.57746704821226e-05, + "loss": 1.1258, + "step": 2993 + }, + { + "epoch": 1.6254071661237783, + "grad_norm": 7.650078703788778, + "learning_rate": 1.5771798928157645e-05, + "loss": 0.5949, + "step": 2994 + }, + { + "epoch": 1.6259500542888166, + "grad_norm": 9.330939436907839, + "learning_rate": 1.5768926660327396e-05, + "loss": 0.9243, + "step": 2995 + }, + { + "epoch": 1.6264929424538543, + "grad_norm": 11.566902464205237, + "learning_rate": 1.576605367898711e-05, + "loss": 1.5588, + "step": 2996 + }, + { + "epoch": 1.6270358306188926, + "grad_norm": 9.91827217636312, + "learning_rate": 1.576317998449211e-05, + "loss": 1.1418, + "step": 2997 + }, + { + "epoch": 1.6275787187839303, + "grad_norm": 5.842910545051031, + "learning_rate": 1.5760305577197824e-05, + "loss": 0.5064, + "step": 2998 + }, + { + "epoch": 1.6281216069489686, + "grad_norm": 7.20637691224012, + "learning_rate": 1.5757430457459765e-05, + "loss": 0.8129, + "step": 2999 + }, + { + "epoch": 1.6286644951140063, + "grad_norm": 7.5338669003432965, + "learning_rate": 1.5754554625633535e-05, + "loss": 0.7244, + "step": 3000 + }, + { + "epoch": 1.6292073832790446, + "grad_norm": 7.489311206471124, + "learning_rate": 1.5751678082074813e-05, + "loss": 1.1409, + "step": 3001 + }, + { + "epoch": 1.6297502714440824, + "grad_norm": 7.26299222123934, + "learning_rate": 1.574880082713938e-05, + "loss": 0.7061, + "step": 3002 + }, + { + "epoch": 1.6302931596091206, + "grad_norm": 6.292246912680897, + "learning_rate": 1.5745922861183095e-05, + "loss": 0.6383, + "step": 3003 + }, + { + "epoch": 1.6308360477741584, + "grad_norm": 7.483086031387996, + "learning_rate": 1.574304418456192e-05, + "loss": 0.8194, + "step": 3004 + }, + { + "epoch": 1.6313789359391966, + "grad_norm": 8.062451557552306, + "learning_rate": 1.5740164797631882e-05, + "loss": 0.9474, + "step": 3005 + }, + { + "epoch": 1.6319218241042344, + "grad_norm": 7.62904254588456, + "learning_rate": 1.5737284700749116e-05, + "loss": 1.2324, + "step": 3006 + }, + { + "epoch": 1.6324647122692726, + "grad_norm": 8.533361938490982, + "learning_rate": 1.573440389426983e-05, + "loss": 0.6182, + "step": 3007 + }, + { + "epoch": 1.6330076004343104, + "grad_norm": 7.0451736765954855, + "learning_rate": 1.5731522378550337e-05, + "loss": 0.7464, + "step": 3008 + }, + { + "epoch": 1.6335504885993486, + "grad_norm": 8.121336222871497, + "learning_rate": 1.572864015394702e-05, + "loss": 1.1087, + "step": 3009 + }, + { + "epoch": 1.6340933767643864, + "grad_norm": 7.2112540180194955, + "learning_rate": 1.5725757220816356e-05, + "loss": 0.7756, + "step": 3010 + }, + { + "epoch": 1.6346362649294246, + "grad_norm": 8.092755806422884, + "learning_rate": 1.5722873579514915e-05, + "loss": 1.1184, + "step": 3011 + }, + { + "epoch": 1.6351791530944624, + "grad_norm": 7.485157971844339, + "learning_rate": 1.5719989230399347e-05, + "loss": 0.6839, + "step": 3012 + }, + { + "epoch": 1.6357220412595006, + "grad_norm": 8.516980380341119, + "learning_rate": 1.5717104173826397e-05, + "loss": 0.948, + "step": 3013 + }, + { + "epoch": 1.6362649294245384, + "grad_norm": 7.602547351124631, + "learning_rate": 1.571421841015289e-05, + "loss": 0.8479, + "step": 3014 + }, + { + "epoch": 1.6368078175895766, + "grad_norm": 7.0470960880026245, + "learning_rate": 1.5711331939735744e-05, + "loss": 0.9014, + "step": 3015 + }, + { + "epoch": 1.6373507057546144, + "grad_norm": 6.760810209806174, + "learning_rate": 1.570844476293196e-05, + "loss": 0.9316, + "step": 3016 + }, + { + "epoch": 1.6378935939196526, + "grad_norm": 8.02053578699558, + "learning_rate": 1.570555688009863e-05, + "loss": 0.7837, + "step": 3017 + }, + { + "epoch": 1.6384364820846904, + "grad_norm": 4.624578610614632, + "learning_rate": 1.5702668291592936e-05, + "loss": 0.4706, + "step": 3018 + }, + { + "epoch": 1.6389793702497286, + "grad_norm": 8.011712265192347, + "learning_rate": 1.569977899777213e-05, + "loss": 1.0338, + "step": 3019 + }, + { + "epoch": 1.6395222584147664, + "grad_norm": 7.773921575637797, + "learning_rate": 1.569688899899358e-05, + "loss": 0.7512, + "step": 3020 + }, + { + "epoch": 1.6400651465798046, + "grad_norm": 12.19433508686472, + "learning_rate": 1.569399829561472e-05, + "loss": 1.3502, + "step": 3021 + }, + { + "epoch": 1.6406080347448424, + "grad_norm": 11.319107865931453, + "learning_rate": 1.569110688799307e-05, + "loss": 0.9996, + "step": 3022 + }, + { + "epoch": 1.6411509229098806, + "grad_norm": 7.1911299711947585, + "learning_rate": 1.5688214776486255e-05, + "loss": 0.9037, + "step": 3023 + }, + { + "epoch": 1.6416938110749184, + "grad_norm": 8.326843836043436, + "learning_rate": 1.5685321961451968e-05, + "loss": 1.1179, + "step": 3024 + }, + { + "epoch": 1.6422366992399566, + "grad_norm": 8.257534801333424, + "learning_rate": 1.5682428443248002e-05, + "loss": 0.7197, + "step": 3025 + }, + { + "epoch": 1.6427795874049944, + "grad_norm": 6.669062512754087, + "learning_rate": 1.567953422223223e-05, + "loss": 0.6292, + "step": 3026 + }, + { + "epoch": 1.6433224755700326, + "grad_norm": 7.99477616152165, + "learning_rate": 1.567663929876261e-05, + "loss": 0.926, + "step": 3027 + }, + { + "epoch": 1.6438653637350704, + "grad_norm": 5.3717184947082615, + "learning_rate": 1.56737436731972e-05, + "loss": 0.5378, + "step": 3028 + }, + { + "epoch": 1.6444082519001086, + "grad_norm": 6.522826762176589, + "learning_rate": 1.5670847345894125e-05, + "loss": 0.7393, + "step": 3029 + }, + { + "epoch": 1.6449511400651464, + "grad_norm": 9.503371738808127, + "learning_rate": 1.5667950317211612e-05, + "loss": 1.2727, + "step": 3030 + }, + { + "epoch": 1.6454940282301846, + "grad_norm": 8.208077816222845, + "learning_rate": 1.5665052587507974e-05, + "loss": 0.7032, + "step": 3031 + }, + { + "epoch": 1.6460369163952224, + "grad_norm": 9.101101018529372, + "learning_rate": 1.56621541571416e-05, + "loss": 0.6983, + "step": 3032 + }, + { + "epoch": 1.6465798045602607, + "grad_norm": 7.254188682380924, + "learning_rate": 1.565925502647098e-05, + "loss": 0.7196, + "step": 3033 + }, + { + "epoch": 1.6471226927252984, + "grad_norm": 7.837018641144827, + "learning_rate": 1.5656355195854676e-05, + "loss": 1.2035, + "step": 3034 + }, + { + "epoch": 1.6476655808903367, + "grad_norm": 8.949191823273495, + "learning_rate": 1.5653454665651344e-05, + "loss": 1.0169, + "step": 3035 + }, + { + "epoch": 1.6482084690553744, + "grad_norm": 8.166317934132502, + "learning_rate": 1.5650553436219732e-05, + "loss": 0.672, + "step": 3036 + }, + { + "epoch": 1.6487513572204127, + "grad_norm": 6.443983188038021, + "learning_rate": 1.564765150791866e-05, + "loss": 0.5552, + "step": 3037 + }, + { + "epoch": 1.6492942453854504, + "grad_norm": 10.403060348162544, + "learning_rate": 1.5644748881107057e-05, + "loss": 1.0709, + "step": 3038 + }, + { + "epoch": 1.6498371335504887, + "grad_norm": 8.319421686790951, + "learning_rate": 1.564184555614391e-05, + "loss": 0.7923, + "step": 3039 + }, + { + "epoch": 1.6503800217155264, + "grad_norm": 9.65474819234395, + "learning_rate": 1.5638941533388318e-05, + "loss": 0.8519, + "step": 3040 + }, + { + "epoch": 1.6509229098805647, + "grad_norm": 8.635959327913778, + "learning_rate": 1.5636036813199445e-05, + "loss": 0.8447, + "step": 3041 + }, + { + "epoch": 1.6514657980456025, + "grad_norm": 7.674155625456285, + "learning_rate": 1.563313139593656e-05, + "loss": 0.7205, + "step": 3042 + }, + { + "epoch": 1.6520086862106407, + "grad_norm": 7.249478120616782, + "learning_rate": 1.5630225281959003e-05, + "loss": 0.6632, + "step": 3043 + }, + { + "epoch": 1.6525515743756785, + "grad_norm": 8.90384172612564, + "learning_rate": 1.5627318471626208e-05, + "loss": 0.8243, + "step": 3044 + }, + { + "epoch": 1.6530944625407167, + "grad_norm": 8.274462585020828, + "learning_rate": 1.5624410965297703e-05, + "loss": 0.7381, + "step": 3045 + }, + { + "epoch": 1.6536373507057545, + "grad_norm": 7.266215459792981, + "learning_rate": 1.562150276333308e-05, + "loss": 0.4989, + "step": 3046 + }, + { + "epoch": 1.6541802388707927, + "grad_norm": 4.816150056433193, + "learning_rate": 1.5618593866092036e-05, + "loss": 0.5382, + "step": 3047 + }, + { + "epoch": 1.6547231270358305, + "grad_norm": 6.835417235425346, + "learning_rate": 1.561568427393435e-05, + "loss": 1.044, + "step": 3048 + }, + { + "epoch": 1.6552660152008687, + "grad_norm": 7.2539344971377435, + "learning_rate": 1.5612773987219885e-05, + "loss": 0.5901, + "step": 3049 + }, + { + "epoch": 1.6558089033659065, + "grad_norm": 11.003646663954257, + "learning_rate": 1.5609863006308586e-05, + "loss": 0.9407, + "step": 3050 + }, + { + "epoch": 1.6563517915309447, + "grad_norm": 8.268114443948154, + "learning_rate": 1.560695133156049e-05, + "loss": 1.0164, + "step": 3051 + }, + { + "epoch": 1.6568946796959825, + "grad_norm": 7.824779774533612, + "learning_rate": 1.5604038963335716e-05, + "loss": 0.7111, + "step": 3052 + }, + { + "epoch": 1.6574375678610207, + "grad_norm": 8.542948361290183, + "learning_rate": 1.560112590199447e-05, + "loss": 0.6501, + "step": 3053 + }, + { + "epoch": 1.6579804560260585, + "grad_norm": 7.85824626614946, + "learning_rate": 1.5598212147897047e-05, + "loss": 0.752, + "step": 3054 + }, + { + "epoch": 1.6585233441910967, + "grad_norm": 6.588621135611242, + "learning_rate": 1.559529770140382e-05, + "loss": 0.507, + "step": 3055 + }, + { + "epoch": 1.6590662323561345, + "grad_norm": 6.418764525989286, + "learning_rate": 1.559238256287526e-05, + "loss": 0.5947, + "step": 3056 + }, + { + "epoch": 1.6596091205211727, + "grad_norm": 7.681071515192806, + "learning_rate": 1.5589466732671913e-05, + "loss": 0.8854, + "step": 3057 + }, + { + "epoch": 1.6601520086862105, + "grad_norm": 5.918230957225734, + "learning_rate": 1.558655021115441e-05, + "loss": 0.5546, + "step": 3058 + }, + { + "epoch": 1.6606948968512487, + "grad_norm": 10.669803193569846, + "learning_rate": 1.5583632998683475e-05, + "loss": 1.1548, + "step": 3059 + }, + { + "epoch": 1.6612377850162865, + "grad_norm": 9.271579004183549, + "learning_rate": 1.558071509561991e-05, + "loss": 0.8049, + "step": 3060 + }, + { + "epoch": 1.6617806731813247, + "grad_norm": 8.337085337830533, + "learning_rate": 1.557779650232461e-05, + "loss": 0.954, + "step": 3061 + }, + { + "epoch": 1.6623235613463625, + "grad_norm": 11.145075141347627, + "learning_rate": 1.5574877219158543e-05, + "loss": 1.1592, + "step": 3062 + }, + { + "epoch": 1.6628664495114007, + "grad_norm": 13.953049521423079, + "learning_rate": 1.557195724648278e-05, + "loss": 1.5699, + "step": 3063 + }, + { + "epoch": 1.6634093376764385, + "grad_norm": 8.654584332416588, + "learning_rate": 1.5569036584658466e-05, + "loss": 1.0185, + "step": 3064 + }, + { + "epoch": 1.6639522258414767, + "grad_norm": 7.520260829640502, + "learning_rate": 1.556611523404683e-05, + "loss": 0.7898, + "step": 3065 + }, + { + "epoch": 1.6644951140065145, + "grad_norm": 6.7373805901465875, + "learning_rate": 1.5563193195009188e-05, + "loss": 0.8392, + "step": 3066 + }, + { + "epoch": 1.6650380021715527, + "grad_norm": 6.732598633253056, + "learning_rate": 1.556027046790695e-05, + "loss": 0.6672, + "step": 3067 + }, + { + "epoch": 1.6655808903365905, + "grad_norm": 9.217725100287737, + "learning_rate": 1.55573470531016e-05, + "loss": 0.9874, + "step": 3068 + }, + { + "epoch": 1.6661237785016287, + "grad_norm": 6.963043776606882, + "learning_rate": 1.5554422950954706e-05, + "loss": 0.812, + "step": 3069 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 6.449074540999346, + "learning_rate": 1.5551498161827928e-05, + "loss": 0.5035, + "step": 3070 + }, + { + "epoch": 1.6672095548317047, + "grad_norm": 7.793693453975787, + "learning_rate": 1.554857268608301e-05, + "loss": 0.7974, + "step": 3071 + }, + { + "epoch": 1.6677524429967425, + "grad_norm": 10.124707886106192, + "learning_rate": 1.554564652408178e-05, + "loss": 1.4946, + "step": 3072 + }, + { + "epoch": 1.6682953311617807, + "grad_norm": 5.481775138355068, + "learning_rate": 1.5542719676186153e-05, + "loss": 0.502, + "step": 3073 + }, + { + "epoch": 1.6688382193268185, + "grad_norm": 6.055974601989459, + "learning_rate": 1.5539792142758114e-05, + "loss": 0.7463, + "step": 3074 + }, + { + "epoch": 1.6693811074918568, + "grad_norm": 7.452121005302326, + "learning_rate": 1.5536863924159762e-05, + "loss": 0.7282, + "step": 3075 + }, + { + "epoch": 1.6699239956568945, + "grad_norm": 8.991453230181806, + "learning_rate": 1.5533935020753252e-05, + "loss": 0.9327, + "step": 3076 + }, + { + "epoch": 1.6704668838219328, + "grad_norm": 8.138196256677595, + "learning_rate": 1.5531005432900838e-05, + "loss": 0.658, + "step": 3077 + }, + { + "epoch": 1.6710097719869705, + "grad_norm": 6.674313540160648, + "learning_rate": 1.552807516096486e-05, + "loss": 0.569, + "step": 3078 + }, + { + "epoch": 1.6715526601520088, + "grad_norm": 7.2303438904479105, + "learning_rate": 1.5525144205307734e-05, + "loss": 0.5186, + "step": 3079 + }, + { + "epoch": 1.6720955483170465, + "grad_norm": 6.872945037718312, + "learning_rate": 1.5522212566291966e-05, + "loss": 0.7282, + "step": 3080 + }, + { + "epoch": 1.6726384364820848, + "grad_norm": 7.913543579080126, + "learning_rate": 1.5519280244280146e-05, + "loss": 0.7015, + "step": 3081 + }, + { + "epoch": 1.6731813246471225, + "grad_norm": 6.144062932600708, + "learning_rate": 1.551634723963495e-05, + "loss": 0.7829, + "step": 3082 + }, + { + "epoch": 1.6737242128121608, + "grad_norm": 7.111349951384809, + "learning_rate": 1.5513413552719137e-05, + "loss": 0.6904, + "step": 3083 + }, + { + "epoch": 1.6742671009771986, + "grad_norm": 10.453360973997349, + "learning_rate": 1.5510479183895545e-05, + "loss": 1.4584, + "step": 3084 + }, + { + "epoch": 1.6748099891422368, + "grad_norm": 13.976555611119428, + "learning_rate": 1.5507544133527108e-05, + "loss": 1.542, + "step": 3085 + }, + { + "epoch": 1.6753528773072746, + "grad_norm": 5.894489690173581, + "learning_rate": 1.5504608401976835e-05, + "loss": 0.6792, + "step": 3086 + }, + { + "epoch": 1.6758957654723128, + "grad_norm": 8.937162315762803, + "learning_rate": 1.550167198960782e-05, + "loss": 0.8243, + "step": 3087 + }, + { + "epoch": 1.6764386536373506, + "grad_norm": 9.787723535687237, + "learning_rate": 1.549873489678325e-05, + "loss": 1.0071, + "step": 3088 + }, + { + "epoch": 1.6769815418023888, + "grad_norm": 6.0371827348978755, + "learning_rate": 1.5495797123866378e-05, + "loss": 0.6407, + "step": 3089 + }, + { + "epoch": 1.6775244299674266, + "grad_norm": 5.997673025227573, + "learning_rate": 1.5492858671220563e-05, + "loss": 0.7129, + "step": 3090 + }, + { + "epoch": 1.6780673181324648, + "grad_norm": 11.08813931574187, + "learning_rate": 1.5489919539209232e-05, + "loss": 0.8435, + "step": 3091 + }, + { + "epoch": 1.6786102062975026, + "grad_norm": 8.730951161091724, + "learning_rate": 1.54869797281959e-05, + "loss": 0.9159, + "step": 3092 + }, + { + "epoch": 1.6791530944625408, + "grad_norm": 7.293119927853279, + "learning_rate": 1.5484039238544173e-05, + "loss": 1.1001, + "step": 3093 + }, + { + "epoch": 1.6796959826275786, + "grad_norm": 7.091810627268002, + "learning_rate": 1.5481098070617734e-05, + "loss": 0.747, + "step": 3094 + }, + { + "epoch": 1.6802388707926168, + "grad_norm": 9.079124871383598, + "learning_rate": 1.547815622478035e-05, + "loss": 1.2292, + "step": 3095 + }, + { + "epoch": 1.6807817589576546, + "grad_norm": 7.238161650182196, + "learning_rate": 1.5475213701395867e-05, + "loss": 0.6657, + "step": 3096 + }, + { + "epoch": 1.6813246471226928, + "grad_norm": 9.385885763398004, + "learning_rate": 1.5472270500828236e-05, + "loss": 1.305, + "step": 3097 + }, + { + "epoch": 1.6818675352877306, + "grad_norm": 8.98902696621929, + "learning_rate": 1.5469326623441463e-05, + "loss": 0.8641, + "step": 3098 + }, + { + "epoch": 1.6824104234527688, + "grad_norm": 6.519773416264212, + "learning_rate": 1.5466382069599656e-05, + "loss": 0.4567, + "step": 3099 + }, + { + "epoch": 1.6829533116178066, + "grad_norm": 6.775917495491388, + "learning_rate": 1.5463436839667007e-05, + "loss": 0.6874, + "step": 3100 + }, + { + "epoch": 1.6834961997828448, + "grad_norm": 7.82569199258335, + "learning_rate": 1.5460490934007776e-05, + "loss": 0.819, + "step": 3101 + }, + { + "epoch": 1.6840390879478826, + "grad_norm": 6.885216490352139, + "learning_rate": 1.5457544352986326e-05, + "loss": 0.7891, + "step": 3102 + }, + { + "epoch": 1.6845819761129208, + "grad_norm": 6.733247427994312, + "learning_rate": 1.5454597096967093e-05, + "loss": 0.6664, + "step": 3103 + }, + { + "epoch": 1.6851248642779586, + "grad_norm": 6.96769902934235, + "learning_rate": 1.5451649166314598e-05, + "loss": 0.6578, + "step": 3104 + }, + { + "epoch": 1.6856677524429968, + "grad_norm": 6.941713803652074, + "learning_rate": 1.5448700561393444e-05, + "loss": 0.5855, + "step": 3105 + }, + { + "epoch": 1.6862106406080346, + "grad_norm": 8.203921934361416, + "learning_rate": 1.5445751282568324e-05, + "loss": 0.9843, + "step": 3106 + }, + { + "epoch": 1.6867535287730728, + "grad_norm": 8.822160140984163, + "learning_rate": 1.5442801330204004e-05, + "loss": 1.1003, + "step": 3107 + }, + { + "epoch": 1.6872964169381106, + "grad_norm": 8.330851912949297, + "learning_rate": 1.5439850704665338e-05, + "loss": 1.0992, + "step": 3108 + }, + { + "epoch": 1.6878393051031488, + "grad_norm": 6.552171568470756, + "learning_rate": 1.543689940631727e-05, + "loss": 0.5434, + "step": 3109 + }, + { + "epoch": 1.6883821932681866, + "grad_norm": 5.772049074992556, + "learning_rate": 1.5433947435524822e-05, + "loss": 0.7515, + "step": 3110 + }, + { + "epoch": 1.6889250814332248, + "grad_norm": 9.575440058135785, + "learning_rate": 1.543099479265309e-05, + "loss": 0.9691, + "step": 3111 + }, + { + "epoch": 1.6894679695982626, + "grad_norm": 6.008162314259628, + "learning_rate": 1.5428041478067263e-05, + "loss": 0.6137, + "step": 3112 + }, + { + "epoch": 1.6900108577633008, + "grad_norm": 8.993171311918207, + "learning_rate": 1.542508749213262e-05, + "loss": 0.7608, + "step": 3113 + }, + { + "epoch": 1.6905537459283386, + "grad_norm": 8.042583538721042, + "learning_rate": 1.542213283521451e-05, + "loss": 0.8188, + "step": 3114 + }, + { + "epoch": 1.6910966340933768, + "grad_norm": 9.290959640283983, + "learning_rate": 1.541917750767837e-05, + "loss": 1.1737, + "step": 3115 + }, + { + "epoch": 1.6916395222584146, + "grad_norm": 6.3883435583657535, + "learning_rate": 1.5416221509889718e-05, + "loss": 0.4376, + "step": 3116 + }, + { + "epoch": 1.6921824104234529, + "grad_norm": 8.943225081508276, + "learning_rate": 1.541326484221416e-05, + "loss": 0.9983, + "step": 3117 + }, + { + "epoch": 1.6927252985884906, + "grad_norm": 7.583355336491315, + "learning_rate": 1.541030750501737e-05, + "loss": 0.6078, + "step": 3118 + }, + { + "epoch": 1.6932681867535289, + "grad_norm": 7.318011543277784, + "learning_rate": 1.5407349498665133e-05, + "loss": 0.745, + "step": 3119 + }, + { + "epoch": 1.6938110749185666, + "grad_norm": 6.212465291912028, + "learning_rate": 1.5404390823523287e-05, + "loss": 0.4739, + "step": 3120 + }, + { + "epoch": 1.6943539630836049, + "grad_norm": 8.223585888775384, + "learning_rate": 1.5401431479957775e-05, + "loss": 1.1819, + "step": 3121 + }, + { + "epoch": 1.6948968512486426, + "grad_norm": 7.957232407413796, + "learning_rate": 1.5398471468334605e-05, + "loss": 0.7071, + "step": 3122 + }, + { + "epoch": 1.6954397394136809, + "grad_norm": 5.52597775233848, + "learning_rate": 1.5395510789019884e-05, + "loss": 0.6171, + "step": 3123 + }, + { + "epoch": 1.6959826275787186, + "grad_norm": 8.015216509333372, + "learning_rate": 1.5392549442379785e-05, + "loss": 0.9051, + "step": 3124 + }, + { + "epoch": 1.6965255157437569, + "grad_norm": 8.330470177699917, + "learning_rate": 1.538958742878058e-05, + "loss": 0.5037, + "step": 3125 + }, + { + "epoch": 1.6970684039087947, + "grad_norm": 6.1063862074204245, + "learning_rate": 1.538662474858861e-05, + "loss": 0.6509, + "step": 3126 + }, + { + "epoch": 1.6976112920738329, + "grad_norm": 8.11378323370098, + "learning_rate": 1.5383661402170308e-05, + "loss": 1.0356, + "step": 3127 + }, + { + "epoch": 1.6981541802388707, + "grad_norm": 5.995828293111682, + "learning_rate": 1.5380697389892185e-05, + "loss": 0.5297, + "step": 3128 + }, + { + "epoch": 1.6986970684039089, + "grad_norm": 6.962108385144619, + "learning_rate": 1.537773271212083e-05, + "loss": 0.8711, + "step": 3129 + }, + { + "epoch": 1.6992399565689467, + "grad_norm": 8.69611973474616, + "learning_rate": 1.5374767369222922e-05, + "loss": 0.911, + "step": 3130 + }, + { + "epoch": 1.6997828447339849, + "grad_norm": 11.047832664292853, + "learning_rate": 1.5371801361565223e-05, + "loss": 0.8499, + "step": 3131 + }, + { + "epoch": 1.7003257328990227, + "grad_norm": 10.490928296052365, + "learning_rate": 1.5368834689514568e-05, + "loss": 0.7617, + "step": 3132 + }, + { + "epoch": 1.700868621064061, + "grad_norm": 8.723324926351989, + "learning_rate": 1.536586735343788e-05, + "loss": 0.7158, + "step": 3133 + }, + { + "epoch": 1.7014115092290987, + "grad_norm": 7.780646168464086, + "learning_rate": 1.536289935370217e-05, + "loss": 0.6452, + "step": 3134 + }, + { + "epoch": 1.701954397394137, + "grad_norm": 8.037748179528911, + "learning_rate": 1.5359930690674518e-05, + "loss": 0.8368, + "step": 3135 + }, + { + "epoch": 1.7024972855591747, + "grad_norm": 7.539822669872962, + "learning_rate": 1.5356961364722096e-05, + "loss": 0.7224, + "step": 3136 + }, + { + "epoch": 1.703040173724213, + "grad_norm": 10.929660828744602, + "learning_rate": 1.5353991376212155e-05, + "loss": 1.2459, + "step": 3137 + }, + { + "epoch": 1.7035830618892507, + "grad_norm": 11.118929306326729, + "learning_rate": 1.5351020725512028e-05, + "loss": 1.2769, + "step": 3138 + }, + { + "epoch": 1.704125950054289, + "grad_norm": 8.820097629927263, + "learning_rate": 1.534804941298913e-05, + "loss": 1.0722, + "step": 3139 + }, + { + "epoch": 1.7046688382193267, + "grad_norm": 8.79991079654506, + "learning_rate": 1.5345077439010956e-05, + "loss": 0.7596, + "step": 3140 + }, + { + "epoch": 1.705211726384365, + "grad_norm": 7.5009896900263096, + "learning_rate": 1.5342104803945087e-05, + "loss": 0.6457, + "step": 3141 + }, + { + "epoch": 1.7057546145494027, + "grad_norm": 9.190121352339093, + "learning_rate": 1.533913150815918e-05, + "loss": 0.7152, + "step": 3142 + }, + { + "epoch": 1.706297502714441, + "grad_norm": 7.209026897095485, + "learning_rate": 1.5336157552020977e-05, + "loss": 0.534, + "step": 3143 + }, + { + "epoch": 1.7068403908794787, + "grad_norm": 8.752247270964865, + "learning_rate": 1.5333182935898306e-05, + "loss": 1.0953, + "step": 3144 + }, + { + "epoch": 1.707383279044517, + "grad_norm": 8.714563086288697, + "learning_rate": 1.5330207660159068e-05, + "loss": 0.8016, + "step": 3145 + }, + { + "epoch": 1.7079261672095547, + "grad_norm": 9.540320185420228, + "learning_rate": 1.5327231725171255e-05, + "loss": 1.2679, + "step": 3146 + }, + { + "epoch": 1.708469055374593, + "grad_norm": 8.091263210809633, + "learning_rate": 1.532425513130293e-05, + "loss": 0.7521, + "step": 3147 + }, + { + "epoch": 1.7090119435396307, + "grad_norm": 10.062104841108379, + "learning_rate": 1.5321277878922246e-05, + "loss": 1.0696, + "step": 3148 + }, + { + "epoch": 1.709554831704669, + "grad_norm": 8.184716628635337, + "learning_rate": 1.531829996839743e-05, + "loss": 0.643, + "step": 3149 + }, + { + "epoch": 1.7100977198697067, + "grad_norm": 11.130045309860368, + "learning_rate": 1.53153214000968e-05, + "loss": 1.2565, + "step": 3150 + }, + { + "epoch": 1.710640608034745, + "grad_norm": 8.856470164658607, + "learning_rate": 1.5312342174388746e-05, + "loss": 0.8376, + "step": 3151 + }, + { + "epoch": 1.7111834961997827, + "grad_norm": 11.958863336707164, + "learning_rate": 1.5309362291641747e-05, + "loss": 0.978, + "step": 3152 + }, + { + "epoch": 1.711726384364821, + "grad_norm": 10.369392130776264, + "learning_rate": 1.5306381752224357e-05, + "loss": 1.1822, + "step": 3153 + }, + { + "epoch": 1.7122692725298587, + "grad_norm": 7.346196971786586, + "learning_rate": 1.5303400556505213e-05, + "loss": 0.9284, + "step": 3154 + }, + { + "epoch": 1.712812160694897, + "grad_norm": 6.652184209108613, + "learning_rate": 1.5300418704853042e-05, + "loss": 0.5787, + "step": 3155 + }, + { + "epoch": 1.7133550488599347, + "grad_norm": 9.745597378440513, + "learning_rate": 1.5297436197636634e-05, + "loss": 1.0194, + "step": 3156 + }, + { + "epoch": 1.713897937024973, + "grad_norm": 6.17200933356452, + "learning_rate": 1.5294453035224874e-05, + "loss": 0.6315, + "step": 3157 + }, + { + "epoch": 1.7144408251900107, + "grad_norm": 8.373044157038668, + "learning_rate": 1.5291469217986724e-05, + "loss": 0.9127, + "step": 3158 + }, + { + "epoch": 1.714983713355049, + "grad_norm": 9.811623309823384, + "learning_rate": 1.5288484746291227e-05, + "loss": 0.7791, + "step": 3159 + }, + { + "epoch": 1.7155266015200867, + "grad_norm": 9.264521529810194, + "learning_rate": 1.5285499620507513e-05, + "loss": 0.6764, + "step": 3160 + }, + { + "epoch": 1.716069489685125, + "grad_norm": 8.688562646594697, + "learning_rate": 1.5282513841004777e-05, + "loss": 1.0202, + "step": 3161 + }, + { + "epoch": 1.7166123778501627, + "grad_norm": 8.522452395448054, + "learning_rate": 1.527952740815231e-05, + "loss": 1.0043, + "step": 3162 + }, + { + "epoch": 1.717155266015201, + "grad_norm": 6.330008909603884, + "learning_rate": 1.527654032231948e-05, + "loss": 0.6651, + "step": 3163 + }, + { + "epoch": 1.7176981541802387, + "grad_norm": 9.314495623946685, + "learning_rate": 1.5273552583875736e-05, + "loss": 0.8312, + "step": 3164 + }, + { + "epoch": 1.718241042345277, + "grad_norm": 9.157480690732486, + "learning_rate": 1.52705641931906e-05, + "loss": 1.0017, + "step": 3165 + }, + { + "epoch": 1.7187839305103148, + "grad_norm": 8.300251884517943, + "learning_rate": 1.5267575150633687e-05, + "loss": 0.9929, + "step": 3166 + }, + { + "epoch": 1.719326818675353, + "grad_norm": 8.328644452020212, + "learning_rate": 1.5264585456574684e-05, + "loss": 0.8105, + "step": 3167 + }, + { + "epoch": 1.7198697068403908, + "grad_norm": 9.499775808882656, + "learning_rate": 1.526159511138336e-05, + "loss": 0.8593, + "step": 3168 + }, + { + "epoch": 1.720412595005429, + "grad_norm": 7.684886832360556, + "learning_rate": 1.5258604115429567e-05, + "loss": 0.5919, + "step": 3169 + }, + { + "epoch": 1.7209554831704668, + "grad_norm": 8.666869336691208, + "learning_rate": 1.5255612469083239e-05, + "loss": 0.6537, + "step": 3170 + }, + { + "epoch": 1.721498371335505, + "grad_norm": 11.871617471990923, + "learning_rate": 1.5252620172714378e-05, + "loss": 1.3826, + "step": 3171 + }, + { + "epoch": 1.7220412595005428, + "grad_norm": 6.088995555337644, + "learning_rate": 1.5249627226693089e-05, + "loss": 0.5085, + "step": 3172 + }, + { + "epoch": 1.722584147665581, + "grad_norm": 9.255760547122362, + "learning_rate": 1.5246633631389536e-05, + "loss": 0.8823, + "step": 3173 + }, + { + "epoch": 1.7231270358306188, + "grad_norm": 9.91965748862537, + "learning_rate": 1.5243639387173974e-05, + "loss": 0.8752, + "step": 3174 + }, + { + "epoch": 1.723669923995657, + "grad_norm": 6.397561552165108, + "learning_rate": 1.5240644494416734e-05, + "loss": 0.6673, + "step": 3175 + }, + { + "epoch": 1.7242128121606948, + "grad_norm": 7.066817871358342, + "learning_rate": 1.523764895348823e-05, + "loss": 1.041, + "step": 3176 + }, + { + "epoch": 1.724755700325733, + "grad_norm": 8.282830238008364, + "learning_rate": 1.5234652764758959e-05, + "loss": 0.9521, + "step": 3177 + }, + { + "epoch": 1.7252985884907708, + "grad_norm": 7.611547937947443, + "learning_rate": 1.523165592859949e-05, + "loss": 0.7182, + "step": 3178 + }, + { + "epoch": 1.725841476655809, + "grad_norm": 8.068698205768472, + "learning_rate": 1.5228658445380475e-05, + "loss": 0.9078, + "step": 3179 + }, + { + "epoch": 1.7263843648208468, + "grad_norm": 8.731772184014808, + "learning_rate": 1.5225660315472652e-05, + "loss": 1.2148, + "step": 3180 + }, + { + "epoch": 1.726927252985885, + "grad_norm": 7.785982141134894, + "learning_rate": 1.5222661539246832e-05, + "loss": 0.8653, + "step": 3181 + }, + { + "epoch": 1.7274701411509228, + "grad_norm": 7.6267653051373285, + "learning_rate": 1.521966211707391e-05, + "loss": 0.7454, + "step": 3182 + }, + { + "epoch": 1.728013029315961, + "grad_norm": 7.440711922406893, + "learning_rate": 1.521666204932486e-05, + "loss": 0.6576, + "step": 3183 + }, + { + "epoch": 1.7285559174809988, + "grad_norm": 7.69035147644333, + "learning_rate": 1.521366133637073e-05, + "loss": 0.8269, + "step": 3184 + }, + { + "epoch": 1.729098805646037, + "grad_norm": 8.387844627258719, + "learning_rate": 1.5210659978582662e-05, + "loss": 1.1049, + "step": 3185 + }, + { + "epoch": 1.7296416938110748, + "grad_norm": 9.08085381726355, + "learning_rate": 1.5207657976331862e-05, + "loss": 0.97, + "step": 3186 + }, + { + "epoch": 1.730184581976113, + "grad_norm": 7.621613071814458, + "learning_rate": 1.5204655329989617e-05, + "loss": 0.7113, + "step": 3187 + }, + { + "epoch": 1.7307274701411508, + "grad_norm": 6.743838306403499, + "learning_rate": 1.5201652039927313e-05, + "loss": 0.7307, + "step": 3188 + }, + { + "epoch": 1.731270358306189, + "grad_norm": 9.602658786805504, + "learning_rate": 1.5198648106516392e-05, + "loss": 1.0515, + "step": 3189 + }, + { + "epoch": 1.7318132464712268, + "grad_norm": 8.153765987898234, + "learning_rate": 1.5195643530128387e-05, + "loss": 1.0879, + "step": 3190 + }, + { + "epoch": 1.732356134636265, + "grad_norm": 9.745244927474419, + "learning_rate": 1.519263831113491e-05, + "loss": 0.9542, + "step": 3191 + }, + { + "epoch": 1.7328990228013028, + "grad_norm": 8.908425729214914, + "learning_rate": 1.5189632449907654e-05, + "loss": 0.8936, + "step": 3192 + }, + { + "epoch": 1.733441910966341, + "grad_norm": 7.409065356665995, + "learning_rate": 1.5186625946818382e-05, + "loss": 0.7645, + "step": 3193 + }, + { + "epoch": 1.7339847991313788, + "grad_norm": 7.958784484679066, + "learning_rate": 1.5183618802238949e-05, + "loss": 0.8512, + "step": 3194 + }, + { + "epoch": 1.734527687296417, + "grad_norm": 5.808082449761269, + "learning_rate": 1.5180611016541278e-05, + "loss": 0.7241, + "step": 3195 + }, + { + "epoch": 1.7350705754614548, + "grad_norm": 8.88976614296984, + "learning_rate": 1.5177602590097382e-05, + "loss": 1.0506, + "step": 3196 + }, + { + "epoch": 1.735613463626493, + "grad_norm": 8.197925091983134, + "learning_rate": 1.5174593523279346e-05, + "loss": 0.7959, + "step": 3197 + }, + { + "epoch": 1.7361563517915308, + "grad_norm": 10.2752921216035, + "learning_rate": 1.5171583816459334e-05, + "loss": 1.2859, + "step": 3198 + }, + { + "epoch": 1.736699239956569, + "grad_norm": 9.43498245507836, + "learning_rate": 1.5168573470009596e-05, + "loss": 1.0459, + "step": 3199 + }, + { + "epoch": 1.7372421281216068, + "grad_norm": 9.033995794822129, + "learning_rate": 1.516556248430245e-05, + "loss": 1.0003, + "step": 3200 + }, + { + "epoch": 1.737785016286645, + "grad_norm": 9.577836032406582, + "learning_rate": 1.5162550859710306e-05, + "loss": 1.2394, + "step": 3201 + }, + { + "epoch": 1.7383279044516828, + "grad_norm": 5.2063230201693775, + "learning_rate": 1.5159538596605642e-05, + "loss": 0.3625, + "step": 3202 + }, + { + "epoch": 1.738870792616721, + "grad_norm": 7.741298457702819, + "learning_rate": 1.5156525695361022e-05, + "loss": 0.6744, + "step": 3203 + }, + { + "epoch": 1.7394136807817588, + "grad_norm": 8.718382461609222, + "learning_rate": 1.5153512156349086e-05, + "loss": 1.0423, + "step": 3204 + }, + { + "epoch": 1.739956568946797, + "grad_norm": 6.706656901306365, + "learning_rate": 1.5150497979942556e-05, + "loss": 0.5121, + "step": 3205 + }, + { + "epoch": 1.7404994571118348, + "grad_norm": 9.934397785927217, + "learning_rate": 1.5147483166514225e-05, + "loss": 1.5269, + "step": 3206 + }, + { + "epoch": 1.741042345276873, + "grad_norm": 8.069224944733923, + "learning_rate": 1.514446771643697e-05, + "loss": 0.7801, + "step": 3207 + }, + { + "epoch": 1.7415852334419109, + "grad_norm": 9.193161021474104, + "learning_rate": 1.514145163008375e-05, + "loss": 1.2144, + "step": 3208 + }, + { + "epoch": 1.742128121606949, + "grad_norm": 6.5647281089901, + "learning_rate": 1.5138434907827599e-05, + "loss": 0.6692, + "step": 3209 + }, + { + "epoch": 1.7426710097719869, + "grad_norm": 9.387942500806554, + "learning_rate": 1.513541755004163e-05, + "loss": 1.1655, + "step": 3210 + }, + { + "epoch": 1.743213897937025, + "grad_norm": 6.936398185140981, + "learning_rate": 1.5132399557099031e-05, + "loss": 0.7761, + "step": 3211 + }, + { + "epoch": 1.7437567861020629, + "grad_norm": 8.699280554503527, + "learning_rate": 1.5129380929373076e-05, + "loss": 1.2891, + "step": 3212 + }, + { + "epoch": 1.744299674267101, + "grad_norm": 11.932192326740894, + "learning_rate": 1.5126361667237116e-05, + "loss": 0.9868, + "step": 3213 + }, + { + "epoch": 1.7448425624321389, + "grad_norm": 8.456694582402795, + "learning_rate": 1.512334177106457e-05, + "loss": 1.199, + "step": 3214 + }, + { + "epoch": 1.745385450597177, + "grad_norm": 8.284979161281417, + "learning_rate": 1.512032124122895e-05, + "loss": 0.658, + "step": 3215 + }, + { + "epoch": 1.7459283387622149, + "grad_norm": 8.490755188379255, + "learning_rate": 1.5117300078103841e-05, + "loss": 0.6974, + "step": 3216 + }, + { + "epoch": 1.746471226927253, + "grad_norm": 14.621603534181931, + "learning_rate": 1.5114278282062898e-05, + "loss": 1.4386, + "step": 3217 + }, + { + "epoch": 1.7470141150922909, + "grad_norm": 8.35853855400652, + "learning_rate": 1.511125585347987e-05, + "loss": 0.7991, + "step": 3218 + }, + { + "epoch": 1.747557003257329, + "grad_norm": 10.14192231306983, + "learning_rate": 1.5108232792728567e-05, + "loss": 0.9942, + "step": 3219 + }, + { + "epoch": 1.7480998914223669, + "grad_norm": 6.003113555943506, + "learning_rate": 1.5105209100182893e-05, + "loss": 0.5614, + "step": 3220 + }, + { + "epoch": 1.748642779587405, + "grad_norm": 7.420823995167257, + "learning_rate": 1.5102184776216824e-05, + "loss": 0.6918, + "step": 3221 + }, + { + "epoch": 1.7491856677524429, + "grad_norm": 7.3724702610256, + "learning_rate": 1.5099159821204406e-05, + "loss": 0.815, + "step": 3222 + }, + { + "epoch": 1.749728555917481, + "grad_norm": 8.539056209113738, + "learning_rate": 1.509613423551978e-05, + "loss": 1.0859, + "step": 3223 + }, + { + "epoch": 1.750271444082519, + "grad_norm": 10.518635534148203, + "learning_rate": 1.5093108019537147e-05, + "loss": 1.1873, + "step": 3224 + }, + { + "epoch": 1.7508143322475571, + "grad_norm": 8.049010262043616, + "learning_rate": 1.5090081173630795e-05, + "loss": 0.971, + "step": 3225 + }, + { + "epoch": 1.751357220412595, + "grad_norm": 7.874724964543097, + "learning_rate": 1.5087053698175092e-05, + "loss": 1.06, + "step": 3226 + }, + { + "epoch": 1.7519001085776331, + "grad_norm": 8.018813791724389, + "learning_rate": 1.508402559354448e-05, + "loss": 0.7192, + "step": 3227 + }, + { + "epoch": 1.752442996742671, + "grad_norm": 11.748757549078345, + "learning_rate": 1.508099686011348e-05, + "loss": 1.4476, + "step": 3228 + }, + { + "epoch": 1.7529858849077091, + "grad_norm": 7.885344513181394, + "learning_rate": 1.507796749825669e-05, + "loss": 0.7493, + "step": 3229 + }, + { + "epoch": 1.753528773072747, + "grad_norm": 6.808213775155858, + "learning_rate": 1.507493750834879e-05, + "loss": 0.6646, + "step": 3230 + }, + { + "epoch": 1.7540716612377851, + "grad_norm": 6.304489454276083, + "learning_rate": 1.5071906890764527e-05, + "loss": 0.7375, + "step": 3231 + }, + { + "epoch": 1.754614549402823, + "grad_norm": 8.217636823407403, + "learning_rate": 1.5068875645878739e-05, + "loss": 1.0069, + "step": 3232 + }, + { + "epoch": 1.7551574375678611, + "grad_norm": 10.706707510999644, + "learning_rate": 1.5065843774066329e-05, + "loss": 0.689, + "step": 3233 + }, + { + "epoch": 1.755700325732899, + "grad_norm": 7.600175930285094, + "learning_rate": 1.5062811275702291e-05, + "loss": 0.8645, + "step": 3234 + }, + { + "epoch": 1.7562432138979371, + "grad_norm": 8.463001355349205, + "learning_rate": 1.5059778151161684e-05, + "loss": 0.7185, + "step": 3235 + }, + { + "epoch": 1.756786102062975, + "grad_norm": 5.834156873881953, + "learning_rate": 1.5056744400819651e-05, + "loss": 0.6486, + "step": 3236 + }, + { + "epoch": 1.7573289902280131, + "grad_norm": 7.152665898997485, + "learning_rate": 1.5053710025051411e-05, + "loss": 0.9116, + "step": 3237 + }, + { + "epoch": 1.757871878393051, + "grad_norm": 8.037482464894001, + "learning_rate": 1.5050675024232262e-05, + "loss": 0.8231, + "step": 3238 + }, + { + "epoch": 1.7584147665580891, + "grad_norm": 10.416359782719013, + "learning_rate": 1.5047639398737573e-05, + "loss": 1.2555, + "step": 3239 + }, + { + "epoch": 1.758957654723127, + "grad_norm": 6.789336798229921, + "learning_rate": 1.50446031489428e-05, + "loss": 0.7295, + "step": 3240 + }, + { + "epoch": 1.7595005428881652, + "grad_norm": 9.927450801451293, + "learning_rate": 1.5041566275223472e-05, + "loss": 1.5417, + "step": 3241 + }, + { + "epoch": 1.760043431053203, + "grad_norm": 7.040644062299268, + "learning_rate": 1.5038528777955188e-05, + "loss": 0.4495, + "step": 3242 + }, + { + "epoch": 1.7605863192182412, + "grad_norm": 8.848733486454524, + "learning_rate": 1.5035490657513638e-05, + "loss": 1.0123, + "step": 3243 + }, + { + "epoch": 1.761129207383279, + "grad_norm": 8.109736914130819, + "learning_rate": 1.503245191427458e-05, + "loss": 0.8984, + "step": 3244 + }, + { + "epoch": 1.7616720955483172, + "grad_norm": 7.210785707597705, + "learning_rate": 1.5029412548613845e-05, + "loss": 0.6947, + "step": 3245 + }, + { + "epoch": 1.762214983713355, + "grad_norm": 10.368251730052483, + "learning_rate": 1.502637256090735e-05, + "loss": 0.8413, + "step": 3246 + }, + { + "epoch": 1.7627578718783932, + "grad_norm": 6.770955437649876, + "learning_rate": 1.5023331951531086e-05, + "loss": 0.6386, + "step": 3247 + }, + { + "epoch": 1.763300760043431, + "grad_norm": 10.818384833705863, + "learning_rate": 1.5020290720861122e-05, + "loss": 1.2544, + "step": 3248 + }, + { + "epoch": 1.7638436482084692, + "grad_norm": 7.9104385792951195, + "learning_rate": 1.50172488692736e-05, + "loss": 0.6538, + "step": 3249 + }, + { + "epoch": 1.764386536373507, + "grad_norm": 9.565689306609565, + "learning_rate": 1.5014206397144742e-05, + "loss": 1.1232, + "step": 3250 + }, + { + "epoch": 1.7649294245385452, + "grad_norm": 11.651444088862487, + "learning_rate": 1.5011163304850844e-05, + "loss": 1.2054, + "step": 3251 + }, + { + "epoch": 1.765472312703583, + "grad_norm": 10.326063989708704, + "learning_rate": 1.500811959276828e-05, + "loss": 0.9441, + "step": 3252 + }, + { + "epoch": 1.7660152008686212, + "grad_norm": 8.0529225996154, + "learning_rate": 1.5005075261273506e-05, + "loss": 0.9431, + "step": 3253 + }, + { + "epoch": 1.766558089033659, + "grad_norm": 7.391183931496241, + "learning_rate": 1.500203031074305e-05, + "loss": 0.8619, + "step": 3254 + }, + { + "epoch": 1.7671009771986972, + "grad_norm": 6.892445188811786, + "learning_rate": 1.4998984741553508e-05, + "loss": 0.7039, + "step": 3255 + }, + { + "epoch": 1.767643865363735, + "grad_norm": 6.541771639151763, + "learning_rate": 1.4995938554081568e-05, + "loss": 0.6359, + "step": 3256 + }, + { + "epoch": 1.7681867535287732, + "grad_norm": 7.690176361123399, + "learning_rate": 1.4992891748703985e-05, + "loss": 0.8998, + "step": 3257 + }, + { + "epoch": 1.768729641693811, + "grad_norm": 6.497301879748706, + "learning_rate": 1.498984432579759e-05, + "loss": 0.5743, + "step": 3258 + }, + { + "epoch": 1.7692725298588492, + "grad_norm": 6.656723951695686, + "learning_rate": 1.4986796285739298e-05, + "loss": 0.7002, + "step": 3259 + }, + { + "epoch": 1.769815418023887, + "grad_norm": 7.9129397206513685, + "learning_rate": 1.4983747628906095e-05, + "loss": 0.9239, + "step": 3260 + }, + { + "epoch": 1.7703583061889252, + "grad_norm": 4.544544755639526, + "learning_rate": 1.4980698355675043e-05, + "loss": 0.3431, + "step": 3261 + }, + { + "epoch": 1.770901194353963, + "grad_norm": 8.243743590575315, + "learning_rate": 1.4977648466423278e-05, + "loss": 1.1386, + "step": 3262 + }, + { + "epoch": 1.7714440825190012, + "grad_norm": 9.404041237838763, + "learning_rate": 1.4974597961528021e-05, + "loss": 0.9087, + "step": 3263 + }, + { + "epoch": 1.771986970684039, + "grad_norm": 8.534670509056685, + "learning_rate": 1.4971546841366556e-05, + "loss": 0.8769, + "step": 3264 + }, + { + "epoch": 1.7725298588490772, + "grad_norm": 9.442139570892468, + "learning_rate": 1.4968495106316254e-05, + "loss": 0.6942, + "step": 3265 + }, + { + "epoch": 1.773072747014115, + "grad_norm": 8.449365139837813, + "learning_rate": 1.496544275675456e-05, + "loss": 1.0743, + "step": 3266 + }, + { + "epoch": 1.7736156351791532, + "grad_norm": 9.461545898435599, + "learning_rate": 1.4962389793058994e-05, + "loss": 0.7473, + "step": 3267 + }, + { + "epoch": 1.774158523344191, + "grad_norm": 11.871672085414737, + "learning_rate": 1.4959336215607147e-05, + "loss": 1.5433, + "step": 3268 + }, + { + "epoch": 1.7747014115092292, + "grad_norm": 6.417686676489462, + "learning_rate": 1.4956282024776693e-05, + "loss": 0.5006, + "step": 3269 + }, + { + "epoch": 1.775244299674267, + "grad_norm": 10.49161190534651, + "learning_rate": 1.4953227220945381e-05, + "loss": 0.9918, + "step": 3270 + }, + { + "epoch": 1.7757871878393052, + "grad_norm": 9.298743913365943, + "learning_rate": 1.4950171804491031e-05, + "loss": 1.3149, + "step": 3271 + }, + { + "epoch": 1.776330076004343, + "grad_norm": 7.775473503246929, + "learning_rate": 1.4947115775791541e-05, + "loss": 0.6873, + "step": 3272 + }, + { + "epoch": 1.7768729641693812, + "grad_norm": 6.75591368497042, + "learning_rate": 1.4944059135224891e-05, + "loss": 0.7799, + "step": 3273 + }, + { + "epoch": 1.777415852334419, + "grad_norm": 8.234180005996857, + "learning_rate": 1.4941001883169124e-05, + "loss": 1.0393, + "step": 3274 + }, + { + "epoch": 1.7779587404994572, + "grad_norm": 8.419005181659063, + "learning_rate": 1.4937944020002371e-05, + "loss": 0.6798, + "step": 3275 + }, + { + "epoch": 1.778501628664495, + "grad_norm": 7.0982600255490205, + "learning_rate": 1.4934885546102833e-05, + "loss": 0.4936, + "step": 3276 + }, + { + "epoch": 1.7790445168295332, + "grad_norm": 8.404626943584326, + "learning_rate": 1.4931826461848785e-05, + "loss": 1.0507, + "step": 3277 + }, + { + "epoch": 1.779587404994571, + "grad_norm": 5.861882830901275, + "learning_rate": 1.4928766767618576e-05, + "loss": 0.6785, + "step": 3278 + }, + { + "epoch": 1.7801302931596092, + "grad_norm": 8.114055102204162, + "learning_rate": 1.4925706463790642e-05, + "loss": 0.765, + "step": 3279 + }, + { + "epoch": 1.780673181324647, + "grad_norm": 10.033269670592238, + "learning_rate": 1.492264555074348e-05, + "loss": 0.9163, + "step": 3280 + }, + { + "epoch": 1.7812160694896852, + "grad_norm": 7.040755628619449, + "learning_rate": 1.4919584028855671e-05, + "loss": 0.5962, + "step": 3281 + }, + { + "epoch": 1.781758957654723, + "grad_norm": 12.093987616559675, + "learning_rate": 1.4916521898505872e-05, + "loss": 0.8999, + "step": 3282 + }, + { + "epoch": 1.7823018458197613, + "grad_norm": 9.400200287574112, + "learning_rate": 1.4913459160072805e-05, + "loss": 1.0203, + "step": 3283 + }, + { + "epoch": 1.782844733984799, + "grad_norm": 8.605701710844945, + "learning_rate": 1.4910395813935279e-05, + "loss": 0.7874, + "step": 3284 + }, + { + "epoch": 1.7833876221498373, + "grad_norm": 10.279849472300688, + "learning_rate": 1.4907331860472174e-05, + "loss": 1.4306, + "step": 3285 + }, + { + "epoch": 1.783930510314875, + "grad_norm": 10.582881297295359, + "learning_rate": 1.4904267300062443e-05, + "loss": 0.6536, + "step": 3286 + }, + { + "epoch": 1.7844733984799133, + "grad_norm": 6.960885594125151, + "learning_rate": 1.4901202133085115e-05, + "loss": 0.7256, + "step": 3287 + }, + { + "epoch": 1.785016286644951, + "grad_norm": 9.088589913670967, + "learning_rate": 1.4898136359919298e-05, + "loss": 0.9299, + "step": 3288 + }, + { + "epoch": 1.7855591748099893, + "grad_norm": 8.922355254191798, + "learning_rate": 1.4895069980944168e-05, + "loss": 0.8159, + "step": 3289 + }, + { + "epoch": 1.786102062975027, + "grad_norm": 9.459478678151763, + "learning_rate": 1.4892002996538982e-05, + "loss": 1.132, + "step": 3290 + }, + { + "epoch": 1.7866449511400653, + "grad_norm": 9.746458814865415, + "learning_rate": 1.4888935407083065e-05, + "loss": 0.9874, + "step": 3291 + }, + { + "epoch": 1.787187839305103, + "grad_norm": 8.13065950431162, + "learning_rate": 1.4885867212955827e-05, + "loss": 0.8013, + "step": 3292 + }, + { + "epoch": 1.7877307274701413, + "grad_norm": 8.469962208509646, + "learning_rate": 1.4882798414536749e-05, + "loss": 1.0408, + "step": 3293 + }, + { + "epoch": 1.788273615635179, + "grad_norm": 9.018264603768202, + "learning_rate": 1.4879729012205379e-05, + "loss": 0.8324, + "step": 3294 + }, + { + "epoch": 1.7888165038002173, + "grad_norm": 8.934092671110381, + "learning_rate": 1.4876659006341347e-05, + "loss": 0.8842, + "step": 3295 + }, + { + "epoch": 1.789359391965255, + "grad_norm": 7.612464888658943, + "learning_rate": 1.4873588397324356e-05, + "loss": 0.9356, + "step": 3296 + }, + { + "epoch": 1.7899022801302933, + "grad_norm": 7.303452266512586, + "learning_rate": 1.4870517185534184e-05, + "loss": 1.0291, + "step": 3297 + }, + { + "epoch": 1.790445168295331, + "grad_norm": 9.805172223903986, + "learning_rate": 1.4867445371350687e-05, + "loss": 1.0243, + "step": 3298 + }, + { + "epoch": 1.7909880564603693, + "grad_norm": 7.742224062288237, + "learning_rate": 1.4864372955153785e-05, + "loss": 0.7173, + "step": 3299 + }, + { + "epoch": 1.791530944625407, + "grad_norm": 9.174714938761847, + "learning_rate": 1.4861299937323485e-05, + "loss": 0.9721, + "step": 3300 + }, + { + "epoch": 1.7920738327904453, + "grad_norm": 8.116980821679304, + "learning_rate": 1.4858226318239864e-05, + "loss": 0.7776, + "step": 3301 + }, + { + "epoch": 1.792616720955483, + "grad_norm": 5.885386861657401, + "learning_rate": 1.4855152098283068e-05, + "loss": 0.3833, + "step": 3302 + }, + { + "epoch": 1.7931596091205213, + "grad_norm": 6.804177263707199, + "learning_rate": 1.4852077277833319e-05, + "loss": 0.7972, + "step": 3303 + }, + { + "epoch": 1.793702497285559, + "grad_norm": 8.874804659745028, + "learning_rate": 1.4849001857270922e-05, + "loss": 1.063, + "step": 3304 + }, + { + "epoch": 1.7942453854505973, + "grad_norm": 10.231899453600786, + "learning_rate": 1.4845925836976247e-05, + "loss": 1.6617, + "step": 3305 + }, + { + "epoch": 1.794788273615635, + "grad_norm": 9.067079935451718, + "learning_rate": 1.4842849217329742e-05, + "loss": 1.1062, + "step": 3306 + }, + { + "epoch": 1.7953311617806733, + "grad_norm": 9.5261315220339, + "learning_rate": 1.4839771998711925e-05, + "loss": 1.5991, + "step": 3307 + }, + { + "epoch": 1.795874049945711, + "grad_norm": 9.53442504374803, + "learning_rate": 1.4836694181503397e-05, + "loss": 0.9267, + "step": 3308 + }, + { + "epoch": 1.7964169381107493, + "grad_norm": 8.231497735162955, + "learning_rate": 1.4833615766084821e-05, + "loss": 0.9802, + "step": 3309 + }, + { + "epoch": 1.796959826275787, + "grad_norm": 7.2171054452276815, + "learning_rate": 1.4830536752836944e-05, + "loss": 0.6116, + "step": 3310 + }, + { + "epoch": 1.7975027144408253, + "grad_norm": 8.412637332813022, + "learning_rate": 1.4827457142140583e-05, + "loss": 0.7482, + "step": 3311 + }, + { + "epoch": 1.798045602605863, + "grad_norm": 6.24239912297862, + "learning_rate": 1.482437693437663e-05, + "loss": 0.4333, + "step": 3312 + }, + { + "epoch": 1.7985884907709013, + "grad_norm": 9.706212881828902, + "learning_rate": 1.4821296129926049e-05, + "loss": 1.3862, + "step": 3313 + }, + { + "epoch": 1.799131378935939, + "grad_norm": 8.275124350006633, + "learning_rate": 1.4818214729169878e-05, + "loss": 0.812, + "step": 3314 + }, + { + "epoch": 1.7996742671009773, + "grad_norm": 7.0810692617964275, + "learning_rate": 1.4815132732489228e-05, + "loss": 0.8919, + "step": 3315 + }, + { + "epoch": 1.8002171552660151, + "grad_norm": 7.225811411306636, + "learning_rate": 1.4812050140265288e-05, + "loss": 0.7512, + "step": 3316 + }, + { + "epoch": 1.8007600434310533, + "grad_norm": 6.458874811050634, + "learning_rate": 1.4808966952879318e-05, + "loss": 0.6914, + "step": 3317 + }, + { + "epoch": 1.8013029315960911, + "grad_norm": 6.537680447766501, + "learning_rate": 1.4805883170712653e-05, + "loss": 0.5511, + "step": 3318 + }, + { + "epoch": 1.8018458197611293, + "grad_norm": 8.023777979063889, + "learning_rate": 1.4802798794146695e-05, + "loss": 0.6814, + "step": 3319 + }, + { + "epoch": 1.8023887079261671, + "grad_norm": 13.062126505855263, + "learning_rate": 1.4799713823562932e-05, + "loss": 1.0526, + "step": 3320 + }, + { + "epoch": 1.8029315960912053, + "grad_norm": 8.173027381817493, + "learning_rate": 1.479662825934291e-05, + "loss": 0.8939, + "step": 3321 + }, + { + "epoch": 1.8034744842562431, + "grad_norm": 7.465360921787222, + "learning_rate": 1.4793542101868261e-05, + "loss": 0.5965, + "step": 3322 + }, + { + "epoch": 1.8040173724212814, + "grad_norm": 6.5098081549027835, + "learning_rate": 1.4790455351520684e-05, + "loss": 0.5963, + "step": 3323 + }, + { + "epoch": 1.8045602605863191, + "grad_norm": 11.071700718231854, + "learning_rate": 1.4787368008681956e-05, + "loss": 0.9728, + "step": 3324 + }, + { + "epoch": 1.8051031487513574, + "grad_norm": 9.251168051579464, + "learning_rate": 1.4784280073733924e-05, + "loss": 0.8816, + "step": 3325 + }, + { + "epoch": 1.8056460369163951, + "grad_norm": 7.876824650963893, + "learning_rate": 1.4781191547058505e-05, + "loss": 1.0365, + "step": 3326 + }, + { + "epoch": 1.8061889250814334, + "grad_norm": 7.984254709859244, + "learning_rate": 1.4778102429037696e-05, + "loss": 0.7252, + "step": 3327 + }, + { + "epoch": 1.8067318132464711, + "grad_norm": 8.162664097932598, + "learning_rate": 1.4775012720053563e-05, + "loss": 0.8297, + "step": 3328 + }, + { + "epoch": 1.8072747014115094, + "grad_norm": 6.631733087479191, + "learning_rate": 1.4771922420488245e-05, + "loss": 0.6962, + "step": 3329 + }, + { + "epoch": 1.8078175895765471, + "grad_norm": 8.80250228426254, + "learning_rate": 1.476883153072396e-05, + "loss": 1.0576, + "step": 3330 + }, + { + "epoch": 1.8083604777415854, + "grad_norm": 10.637063152950418, + "learning_rate": 1.476574005114299e-05, + "loss": 0.8866, + "step": 3331 + }, + { + "epoch": 1.8089033659066232, + "grad_norm": 9.58332512337787, + "learning_rate": 1.4762647982127696e-05, + "loss": 1.6148, + "step": 3332 + }, + { + "epoch": 1.8094462540716614, + "grad_norm": 7.826567769007031, + "learning_rate": 1.4759555324060508e-05, + "loss": 0.9256, + "step": 3333 + }, + { + "epoch": 1.8099891422366992, + "grad_norm": 9.445307614493151, + "learning_rate": 1.4756462077323931e-05, + "loss": 1.1389, + "step": 3334 + }, + { + "epoch": 1.8105320304017374, + "grad_norm": 7.504505604214132, + "learning_rate": 1.4753368242300543e-05, + "loss": 1.1333, + "step": 3335 + }, + { + "epoch": 1.8110749185667752, + "grad_norm": 6.996150668976886, + "learning_rate": 1.4750273819372996e-05, + "loss": 0.6721, + "step": 3336 + }, + { + "epoch": 1.8116178067318134, + "grad_norm": 7.593185686378431, + "learning_rate": 1.4747178808924012e-05, + "loss": 0.6856, + "step": 3337 + }, + { + "epoch": 1.8121606948968512, + "grad_norm": 7.277948846492979, + "learning_rate": 1.474408321133639e-05, + "loss": 0.5487, + "step": 3338 + }, + { + "epoch": 1.8127035830618894, + "grad_norm": 9.500801329454344, + "learning_rate": 1.474098702699299e-05, + "loss": 1.063, + "step": 3339 + }, + { + "epoch": 1.8132464712269272, + "grad_norm": 7.732920511283457, + "learning_rate": 1.4737890256276763e-05, + "loss": 0.8875, + "step": 3340 + }, + { + "epoch": 1.8137893593919654, + "grad_norm": 9.56546309643246, + "learning_rate": 1.4734792899570716e-05, + "loss": 0.7724, + "step": 3341 + }, + { + "epoch": 1.8143322475570032, + "grad_norm": 7.735046028758256, + "learning_rate": 1.4731694957257938e-05, + "loss": 0.8498, + "step": 3342 + }, + { + "epoch": 1.8148751357220414, + "grad_norm": 9.077413763001829, + "learning_rate": 1.4728596429721587e-05, + "loss": 1.1604, + "step": 3343 + }, + { + "epoch": 1.8154180238870792, + "grad_norm": 8.33530084886159, + "learning_rate": 1.4725497317344892e-05, + "loss": 1.0126, + "step": 3344 + }, + { + "epoch": 1.8159609120521174, + "grad_norm": 11.938749223535789, + "learning_rate": 1.4722397620511158e-05, + "loss": 0.9538, + "step": 3345 + }, + { + "epoch": 1.8165038002171552, + "grad_norm": 4.946988727080576, + "learning_rate": 1.471929733960376e-05, + "loss": 0.3007, + "step": 3346 + }, + { + "epoch": 1.8170466883821934, + "grad_norm": 6.934881579248824, + "learning_rate": 1.4716196475006145e-05, + "loss": 0.5286, + "step": 3347 + }, + { + "epoch": 1.8175895765472312, + "grad_norm": 12.504257989709076, + "learning_rate": 1.4713095027101833e-05, + "loss": 0.9948, + "step": 3348 + }, + { + "epoch": 1.8181324647122694, + "grad_norm": 6.299717892798866, + "learning_rate": 1.4709992996274416e-05, + "loss": 0.6503, + "step": 3349 + }, + { + "epoch": 1.8186753528773072, + "grad_norm": 8.48347618555589, + "learning_rate": 1.4706890382907562e-05, + "loss": 0.7195, + "step": 3350 + }, + { + "epoch": 1.8192182410423454, + "grad_norm": 7.992565784639017, + "learning_rate": 1.4703787187385002e-05, + "loss": 0.7349, + "step": 3351 + }, + { + "epoch": 1.8197611292073832, + "grad_norm": 7.567197193038979, + "learning_rate": 1.4700683410090546e-05, + "loss": 0.4719, + "step": 3352 + }, + { + "epoch": 1.8203040173724214, + "grad_norm": 7.276389893333727, + "learning_rate": 1.4697579051408073e-05, + "loss": 0.93, + "step": 3353 + }, + { + "epoch": 1.8208469055374592, + "grad_norm": 8.13251629417801, + "learning_rate": 1.4694474111721537e-05, + "loss": 0.9865, + "step": 3354 + }, + { + "epoch": 1.8213897937024974, + "grad_norm": 6.788010338042944, + "learning_rate": 1.469136859141496e-05, + "loss": 0.5671, + "step": 3355 + }, + { + "epoch": 1.8219326818675352, + "grad_norm": 6.83194614525194, + "learning_rate": 1.4688262490872438e-05, + "loss": 0.5486, + "step": 3356 + }, + { + "epoch": 1.8224755700325734, + "grad_norm": 12.201391760355081, + "learning_rate": 1.4685155810478143e-05, + "loss": 0.9178, + "step": 3357 + }, + { + "epoch": 1.8230184581976112, + "grad_norm": 7.955977866907339, + "learning_rate": 1.4682048550616306e-05, + "loss": 0.9671, + "step": 3358 + }, + { + "epoch": 1.8235613463626494, + "grad_norm": 9.699881747283202, + "learning_rate": 1.4678940711671245e-05, + "loss": 0.7699, + "step": 3359 + }, + { + "epoch": 1.8241042345276872, + "grad_norm": 7.703855587772515, + "learning_rate": 1.4675832294027339e-05, + "loss": 0.6471, + "step": 3360 + }, + { + "epoch": 1.8246471226927254, + "grad_norm": 11.230217201340418, + "learning_rate": 1.4672723298069043e-05, + "loss": 1.1256, + "step": 3361 + }, + { + "epoch": 1.8251900108577632, + "grad_norm": 11.964156332533788, + "learning_rate": 1.4669613724180883e-05, + "loss": 1.1829, + "step": 3362 + }, + { + "epoch": 1.8257328990228014, + "grad_norm": 8.221701868139569, + "learning_rate": 1.4666503572747452e-05, + "loss": 0.6004, + "step": 3363 + }, + { + "epoch": 1.8262757871878392, + "grad_norm": 6.179895131215982, + "learning_rate": 1.4663392844153426e-05, + "loss": 0.6923, + "step": 3364 + }, + { + "epoch": 1.8268186753528775, + "grad_norm": 8.333959237389427, + "learning_rate": 1.4660281538783538e-05, + "loss": 0.8785, + "step": 3365 + }, + { + "epoch": 1.8273615635179152, + "grad_norm": 9.742872979143373, + "learning_rate": 1.4657169657022603e-05, + "loss": 1.0129, + "step": 3366 + }, + { + "epoch": 1.8279044516829535, + "grad_norm": 8.019435507225802, + "learning_rate": 1.46540571992555e-05, + "loss": 1.0615, + "step": 3367 + }, + { + "epoch": 1.8284473398479912, + "grad_norm": 5.872559756186675, + "learning_rate": 1.4650944165867187e-05, + "loss": 0.6391, + "step": 3368 + }, + { + "epoch": 1.8289902280130295, + "grad_norm": 7.878073300625923, + "learning_rate": 1.4647830557242685e-05, + "loss": 1.0317, + "step": 3369 + }, + { + "epoch": 1.8295331161780672, + "grad_norm": 10.290400619980408, + "learning_rate": 1.4644716373767096e-05, + "loss": 1.3108, + "step": 3370 + }, + { + "epoch": 1.8300760043431055, + "grad_norm": 7.655417729944032, + "learning_rate": 1.4641601615825582e-05, + "loss": 0.8434, + "step": 3371 + }, + { + "epoch": 1.8306188925081432, + "grad_norm": 8.894875963607664, + "learning_rate": 1.4638486283803384e-05, + "loss": 1.4973, + "step": 3372 + }, + { + "epoch": 1.8311617806731815, + "grad_norm": 7.70346849256546, + "learning_rate": 1.4635370378085809e-05, + "loss": 0.9202, + "step": 3373 + }, + { + "epoch": 1.8317046688382193, + "grad_norm": 7.428334777443758, + "learning_rate": 1.4632253899058233e-05, + "loss": 1.005, + "step": 3374 + }, + { + "epoch": 1.8322475570032575, + "grad_norm": 6.905061230055496, + "learning_rate": 1.4629136847106118e-05, + "loss": 0.7045, + "step": 3375 + }, + { + "epoch": 1.8327904451682953, + "grad_norm": 9.092059147160219, + "learning_rate": 1.4626019222614977e-05, + "loss": 1.1478, + "step": 3376 + }, + { + "epoch": 1.8333333333333335, + "grad_norm": 8.171417716573657, + "learning_rate": 1.4622901025970406e-05, + "loss": 0.8098, + "step": 3377 + }, + { + "epoch": 1.8338762214983713, + "grad_norm": 8.935159598593566, + "learning_rate": 1.461978225755807e-05, + "loss": 0.9121, + "step": 3378 + }, + { + "epoch": 1.8344191096634095, + "grad_norm": 9.059802045574934, + "learning_rate": 1.46166629177637e-05, + "loss": 0.952, + "step": 3379 + }, + { + "epoch": 1.8349619978284473, + "grad_norm": 6.404641837192718, + "learning_rate": 1.4613543006973103e-05, + "loss": 0.7178, + "step": 3380 + }, + { + "epoch": 1.8355048859934855, + "grad_norm": 7.714640842582964, + "learning_rate": 1.4610422525572155e-05, + "loss": 0.7053, + "step": 3381 + }, + { + "epoch": 1.8360477741585233, + "grad_norm": 9.232659815202982, + "learning_rate": 1.46073014739468e-05, + "loss": 0.7707, + "step": 3382 + }, + { + "epoch": 1.8365906623235615, + "grad_norm": 7.597273066110282, + "learning_rate": 1.4604179852483056e-05, + "loss": 0.9593, + "step": 3383 + }, + { + "epoch": 1.8371335504885993, + "grad_norm": 7.246167686662127, + "learning_rate": 1.460105766156701e-05, + "loss": 0.8489, + "step": 3384 + }, + { + "epoch": 1.8376764386536375, + "grad_norm": 9.724005322694314, + "learning_rate": 1.4597934901584818e-05, + "loss": 1.4838, + "step": 3385 + }, + { + "epoch": 1.8382193268186753, + "grad_norm": 5.623806840885622, + "learning_rate": 1.459481157292271e-05, + "loss": 0.4218, + "step": 3386 + }, + { + "epoch": 1.8387622149837135, + "grad_norm": 10.505590233471066, + "learning_rate": 1.4591687675966984e-05, + "loss": 1.3634, + "step": 3387 + }, + { + "epoch": 1.8393051031487513, + "grad_norm": 7.225825527188378, + "learning_rate": 1.458856321110401e-05, + "loss": 1.0878, + "step": 3388 + }, + { + "epoch": 1.8398479913137895, + "grad_norm": 6.92346369662784, + "learning_rate": 1.4585438178720221e-05, + "loss": 0.8579, + "step": 3389 + }, + { + "epoch": 1.8403908794788273, + "grad_norm": 6.78903317106508, + "learning_rate": 1.4582312579202134e-05, + "loss": 0.9986, + "step": 3390 + }, + { + "epoch": 1.8409337676438655, + "grad_norm": 7.493115074862596, + "learning_rate": 1.4579186412936323e-05, + "loss": 0.6403, + "step": 3391 + }, + { + "epoch": 1.8414766558089033, + "grad_norm": 7.292140639462925, + "learning_rate": 1.457605968030944e-05, + "loss": 0.7705, + "step": 3392 + }, + { + "epoch": 1.8420195439739415, + "grad_norm": 8.413656799764597, + "learning_rate": 1.4572932381708198e-05, + "loss": 0.7362, + "step": 3393 + }, + { + "epoch": 1.8425624321389793, + "grad_norm": 8.092404119538978, + "learning_rate": 1.4569804517519392e-05, + "loss": 0.8419, + "step": 3394 + }, + { + "epoch": 1.8431053203040175, + "grad_norm": 8.746713556642208, + "learning_rate": 1.456667608812988e-05, + "loss": 0.8992, + "step": 3395 + }, + { + "epoch": 1.8436482084690553, + "grad_norm": 9.315011907817567, + "learning_rate": 1.4563547093926595e-05, + "loss": 1.1811, + "step": 3396 + }, + { + "epoch": 1.8441910966340935, + "grad_norm": 7.921580843874425, + "learning_rate": 1.4560417535296529e-05, + "loss": 0.9896, + "step": 3397 + }, + { + "epoch": 1.8447339847991313, + "grad_norm": 5.324625495868631, + "learning_rate": 1.4557287412626755e-05, + "loss": 0.5757, + "step": 3398 + }, + { + "epoch": 1.8452768729641695, + "grad_norm": 7.061326556489923, + "learning_rate": 1.4554156726304411e-05, + "loss": 0.7016, + "step": 3399 + }, + { + "epoch": 1.8458197611292073, + "grad_norm": 7.833004564597817, + "learning_rate": 1.4551025476716704e-05, + "loss": 0.6112, + "step": 3400 + }, + { + "epoch": 1.8463626492942455, + "grad_norm": 6.762019290951184, + "learning_rate": 1.4547893664250912e-05, + "loss": 0.6197, + "step": 3401 + }, + { + "epoch": 1.8469055374592833, + "grad_norm": 8.02135944193365, + "learning_rate": 1.4544761289294384e-05, + "loss": 0.7665, + "step": 3402 + }, + { + "epoch": 1.8474484256243215, + "grad_norm": 9.145981041987138, + "learning_rate": 1.4541628352234538e-05, + "loss": 1.4491, + "step": 3403 + }, + { + "epoch": 1.8479913137893593, + "grad_norm": 10.127072377919164, + "learning_rate": 1.4538494853458858e-05, + "loss": 1.1929, + "step": 3404 + }, + { + "epoch": 1.8485342019543975, + "grad_norm": 9.041637256167727, + "learning_rate": 1.45353607933549e-05, + "loss": 0.8471, + "step": 3405 + }, + { + "epoch": 1.8490770901194353, + "grad_norm": 7.08038124220055, + "learning_rate": 1.453222617231029e-05, + "loss": 0.5152, + "step": 3406 + }, + { + "epoch": 1.8496199782844736, + "grad_norm": 8.460108981189723, + "learning_rate": 1.4529090990712726e-05, + "loss": 0.9495, + "step": 3407 + }, + { + "epoch": 1.8501628664495113, + "grad_norm": 8.939702203617836, + "learning_rate": 1.452595524894997e-05, + "loss": 1.0959, + "step": 3408 + }, + { + "epoch": 1.8507057546145496, + "grad_norm": 7.5616663713048595, + "learning_rate": 1.4522818947409855e-05, + "loss": 1.0391, + "step": 3409 + }, + { + "epoch": 1.8512486427795873, + "grad_norm": 8.75531450227851, + "learning_rate": 1.4519682086480287e-05, + "loss": 0.7298, + "step": 3410 + }, + { + "epoch": 1.8517915309446256, + "grad_norm": 7.003753183958329, + "learning_rate": 1.4516544666549233e-05, + "loss": 0.4788, + "step": 3411 + }, + { + "epoch": 1.8523344191096633, + "grad_norm": 10.952441700747523, + "learning_rate": 1.4513406688004734e-05, + "loss": 1.0903, + "step": 3412 + }, + { + "epoch": 1.8528773072747016, + "grad_norm": 8.61536651212902, + "learning_rate": 1.4510268151234903e-05, + "loss": 0.6924, + "step": 3413 + }, + { + "epoch": 1.8534201954397393, + "grad_norm": 9.237257682023538, + "learning_rate": 1.4507129056627922e-05, + "loss": 0.8686, + "step": 3414 + }, + { + "epoch": 1.8539630836047776, + "grad_norm": 7.792081331119547, + "learning_rate": 1.4503989404572031e-05, + "loss": 0.9818, + "step": 3415 + }, + { + "epoch": 1.8545059717698154, + "grad_norm": 8.622336937881878, + "learning_rate": 1.4500849195455557e-05, + "loss": 0.9444, + "step": 3416 + }, + { + "epoch": 1.8550488599348536, + "grad_norm": 7.607437519589793, + "learning_rate": 1.4497708429666882e-05, + "loss": 0.846, + "step": 3417 + }, + { + "epoch": 1.8555917480998914, + "grad_norm": 8.358779281223866, + "learning_rate": 1.4494567107594457e-05, + "loss": 0.7125, + "step": 3418 + }, + { + "epoch": 1.8561346362649296, + "grad_norm": 6.8176708905901044, + "learning_rate": 1.449142522962681e-05, + "loss": 0.8921, + "step": 3419 + }, + { + "epoch": 1.8566775244299674, + "grad_norm": 8.931750092418099, + "learning_rate": 1.448828279615253e-05, + "loss": 0.7989, + "step": 3420 + }, + { + "epoch": 1.8572204125950056, + "grad_norm": 7.481066786036915, + "learning_rate": 1.4485139807560281e-05, + "loss": 0.6642, + "step": 3421 + }, + { + "epoch": 1.8577633007600434, + "grad_norm": 7.460924029721227, + "learning_rate": 1.4481996264238796e-05, + "loss": 0.6509, + "step": 3422 + }, + { + "epoch": 1.8583061889250816, + "grad_norm": 7.224784834236484, + "learning_rate": 1.4478852166576869e-05, + "loss": 0.7244, + "step": 3423 + }, + { + "epoch": 1.8588490770901194, + "grad_norm": 5.912102755785491, + "learning_rate": 1.4475707514963367e-05, + "loss": 0.5713, + "step": 3424 + }, + { + "epoch": 1.8593919652551576, + "grad_norm": 6.994884495015334, + "learning_rate": 1.4472562309787224e-05, + "loss": 0.4875, + "step": 3425 + }, + { + "epoch": 1.8599348534201954, + "grad_norm": 6.38845062016643, + "learning_rate": 1.4469416551437445e-05, + "loss": 0.4708, + "step": 3426 + }, + { + "epoch": 1.8604777415852336, + "grad_norm": 11.250341421225603, + "learning_rate": 1.4466270240303109e-05, + "loss": 1.4338, + "step": 3427 + }, + { + "epoch": 1.8610206297502714, + "grad_norm": 9.828853213376544, + "learning_rate": 1.4463123376773348e-05, + "loss": 0.7307, + "step": 3428 + }, + { + "epoch": 1.8615635179153096, + "grad_norm": 7.96498248613748, + "learning_rate": 1.4459975961237375e-05, + "loss": 0.6727, + "step": 3429 + }, + { + "epoch": 1.8621064060803474, + "grad_norm": 9.736354023562699, + "learning_rate": 1.4456827994084473e-05, + "loss": 0.8091, + "step": 3430 + }, + { + "epoch": 1.8626492942453856, + "grad_norm": 10.317215804231997, + "learning_rate": 1.4453679475703974e-05, + "loss": 1.0997, + "step": 3431 + }, + { + "epoch": 1.8631921824104234, + "grad_norm": 10.00211430988322, + "learning_rate": 1.4450530406485301e-05, + "loss": 1.4007, + "step": 3432 + }, + { + "epoch": 1.8637350705754616, + "grad_norm": 10.665872978467544, + "learning_rate": 1.4447380786817934e-05, + "loss": 0.9221, + "step": 3433 + }, + { + "epoch": 1.8642779587404994, + "grad_norm": 8.8489121199226, + "learning_rate": 1.4444230617091424e-05, + "loss": 0.7745, + "step": 3434 + }, + { + "epoch": 1.8648208469055376, + "grad_norm": 7.083707820329958, + "learning_rate": 1.444107989769539e-05, + "loss": 0.7152, + "step": 3435 + }, + { + "epoch": 1.8653637350705754, + "grad_norm": 7.496522090143239, + "learning_rate": 1.4437928629019515e-05, + "loss": 0.5901, + "step": 3436 + }, + { + "epoch": 1.8659066232356136, + "grad_norm": 10.576310319834935, + "learning_rate": 1.4434776811453552e-05, + "loss": 0.7752, + "step": 3437 + }, + { + "epoch": 1.8664495114006514, + "grad_norm": 8.686730112665717, + "learning_rate": 1.4431624445387327e-05, + "loss": 0.8748, + "step": 3438 + }, + { + "epoch": 1.8669923995656896, + "grad_norm": 9.435491336462407, + "learning_rate": 1.4428471531210726e-05, + "loss": 1.1954, + "step": 3439 + }, + { + "epoch": 1.8675352877307274, + "grad_norm": 9.012159097359817, + "learning_rate": 1.4425318069313712e-05, + "loss": 0.7353, + "step": 3440 + }, + { + "epoch": 1.8680781758957656, + "grad_norm": 9.617409218513528, + "learning_rate": 1.4422164060086304e-05, + "loss": 0.6611, + "step": 3441 + }, + { + "epoch": 1.8686210640608034, + "grad_norm": 9.612516235298202, + "learning_rate": 1.4419009503918598e-05, + "loss": 0.8101, + "step": 3442 + }, + { + "epoch": 1.8691639522258416, + "grad_norm": 6.542519560792006, + "learning_rate": 1.4415854401200751e-05, + "loss": 0.473, + "step": 3443 + }, + { + "epoch": 1.8697068403908794, + "grad_norm": 10.696166776054064, + "learning_rate": 1.4412698752322998e-05, + "loss": 1.0001, + "step": 3444 + }, + { + "epoch": 1.8702497285559176, + "grad_norm": 10.087499378994824, + "learning_rate": 1.4409542557675625e-05, + "loss": 0.9323, + "step": 3445 + }, + { + "epoch": 1.8707926167209554, + "grad_norm": 7.829085764895678, + "learning_rate": 1.4406385817649002e-05, + "loss": 0.8624, + "step": 3446 + }, + { + "epoch": 1.8713355048859937, + "grad_norm": 11.554825371147222, + "learning_rate": 1.4403228532633561e-05, + "loss": 0.7735, + "step": 3447 + }, + { + "epoch": 1.8718783930510314, + "grad_norm": 9.43390979006707, + "learning_rate": 1.4400070703019797e-05, + "loss": 0.9128, + "step": 3448 + }, + { + "epoch": 1.8724212812160697, + "grad_norm": 11.851717520188132, + "learning_rate": 1.4396912329198276e-05, + "loss": 1.2379, + "step": 3449 + }, + { + "epoch": 1.8729641693811074, + "grad_norm": 8.408242915866566, + "learning_rate": 1.4393753411559632e-05, + "loss": 0.7549, + "step": 3450 + }, + { + "epoch": 1.8735070575461457, + "grad_norm": 8.144856314123112, + "learning_rate": 1.4390593950494557e-05, + "loss": 0.7281, + "step": 3451 + }, + { + "epoch": 1.8740499457111834, + "grad_norm": 8.202846690293642, + "learning_rate": 1.438743394639383e-05, + "loss": 0.6773, + "step": 3452 + }, + { + "epoch": 1.8745928338762217, + "grad_norm": 9.423562458268494, + "learning_rate": 1.438427339964828e-05, + "loss": 1.1, + "step": 3453 + }, + { + "epoch": 1.8751357220412594, + "grad_norm": 7.196483568403804, + "learning_rate": 1.4381112310648806e-05, + "loss": 0.5809, + "step": 3454 + }, + { + "epoch": 1.8756786102062974, + "grad_norm": 7.26770042201738, + "learning_rate": 1.4377950679786382e-05, + "loss": 0.8053, + "step": 3455 + }, + { + "epoch": 1.8762214983713354, + "grad_norm": 9.472969211327861, + "learning_rate": 1.4374788507452039e-05, + "loss": 0.7998, + "step": 3456 + }, + { + "epoch": 1.8767643865363735, + "grad_norm": 9.514618174868874, + "learning_rate": 1.437162579403688e-05, + "loss": 0.9572, + "step": 3457 + }, + { + "epoch": 1.8773072747014115, + "grad_norm": 9.363034597703797, + "learning_rate": 1.4368462539932073e-05, + "loss": 0.8528, + "step": 3458 + }, + { + "epoch": 1.8778501628664495, + "grad_norm": 7.113386807438533, + "learning_rate": 1.4365298745528863e-05, + "loss": 0.733, + "step": 3459 + }, + { + "epoch": 1.8783930510314875, + "grad_norm": 9.09654372206602, + "learning_rate": 1.4362134411218545e-05, + "loss": 0.7779, + "step": 3460 + }, + { + "epoch": 1.8789359391965255, + "grad_norm": 7.449242636613164, + "learning_rate": 1.4358969537392489e-05, + "loss": 0.899, + "step": 3461 + }, + { + "epoch": 1.8794788273615635, + "grad_norm": 10.755480560042127, + "learning_rate": 1.4355804124442133e-05, + "loss": 1.099, + "step": 3462 + }, + { + "epoch": 1.8800217155266015, + "grad_norm": 9.091218073930861, + "learning_rate": 1.4352638172758984e-05, + "loss": 0.7489, + "step": 3463 + }, + { + "epoch": 1.8805646036916395, + "grad_norm": 9.701260384390375, + "learning_rate": 1.4349471682734603e-05, + "loss": 0.8495, + "step": 3464 + }, + { + "epoch": 1.8811074918566775, + "grad_norm": 9.008809447502522, + "learning_rate": 1.4346304654760637e-05, + "loss": 0.8984, + "step": 3465 + }, + { + "epoch": 1.8816503800217155, + "grad_norm": 8.176404536031864, + "learning_rate": 1.4343137089228783e-05, + "loss": 0.8278, + "step": 3466 + }, + { + "epoch": 1.8821932681867535, + "grad_norm": 9.197307290475337, + "learning_rate": 1.4339968986530813e-05, + "loss": 0.7142, + "step": 3467 + }, + { + "epoch": 1.8827361563517915, + "grad_norm": 5.505228455971892, + "learning_rate": 1.433680034705856e-05, + "loss": 0.4204, + "step": 3468 + }, + { + "epoch": 1.8832790445168295, + "grad_norm": 9.014964705534542, + "learning_rate": 1.433363117120393e-05, + "loss": 0.8988, + "step": 3469 + }, + { + "epoch": 1.8838219326818675, + "grad_norm": 7.290687267145188, + "learning_rate": 1.4330461459358889e-05, + "loss": 0.7683, + "step": 3470 + }, + { + "epoch": 1.8843648208469055, + "grad_norm": 10.834879009670003, + "learning_rate": 1.4327291211915473e-05, + "loss": 1.0446, + "step": 3471 + }, + { + "epoch": 1.8849077090119435, + "grad_norm": 8.240452054600638, + "learning_rate": 1.4324120429265781e-05, + "loss": 0.893, + "step": 3472 + }, + { + "epoch": 1.8854505971769815, + "grad_norm": 8.17919491677486, + "learning_rate": 1.4320949111801987e-05, + "loss": 0.9846, + "step": 3473 + }, + { + "epoch": 1.8859934853420195, + "grad_norm": 7.693481880154859, + "learning_rate": 1.431777725991632e-05, + "loss": 0.6054, + "step": 3474 + }, + { + "epoch": 1.8865363735070575, + "grad_norm": 7.929110152467141, + "learning_rate": 1.4314604874001076e-05, + "loss": 0.7299, + "step": 3475 + }, + { + "epoch": 1.8870792616720955, + "grad_norm": 9.650029930461764, + "learning_rate": 1.4311431954448629e-05, + "loss": 1.0855, + "step": 3476 + }, + { + "epoch": 1.8876221498371335, + "grad_norm": 8.487868906333745, + "learning_rate": 1.4308258501651404e-05, + "loss": 0.7117, + "step": 3477 + }, + { + "epoch": 1.8881650380021715, + "grad_norm": 7.111157851383161, + "learning_rate": 1.43050845160019e-05, + "loss": 0.6477, + "step": 3478 + }, + { + "epoch": 1.8887079261672095, + "grad_norm": 8.2908796123436, + "learning_rate": 1.4301909997892684e-05, + "loss": 0.9819, + "step": 3479 + }, + { + "epoch": 1.8892508143322475, + "grad_norm": 10.322900044501562, + "learning_rate": 1.4298734947716385e-05, + "loss": 0.9368, + "step": 3480 + }, + { + "epoch": 1.8897937024972855, + "grad_norm": 6.902927389625658, + "learning_rate": 1.4295559365865694e-05, + "loss": 0.6201, + "step": 3481 + }, + { + "epoch": 1.8903365906623235, + "grad_norm": 8.909574060437844, + "learning_rate": 1.4292383252733375e-05, + "loss": 0.9393, + "step": 3482 + }, + { + "epoch": 1.8908794788273615, + "grad_norm": 7.643883869480783, + "learning_rate": 1.4289206608712252e-05, + "loss": 0.9349, + "step": 3483 + }, + { + "epoch": 1.8914223669923995, + "grad_norm": 7.103928943926094, + "learning_rate": 1.428602943419522e-05, + "loss": 0.6511, + "step": 3484 + }, + { + "epoch": 1.8919652551574375, + "grad_norm": 8.843983286039231, + "learning_rate": 1.428285172957524e-05, + "loss": 0.6958, + "step": 3485 + }, + { + "epoch": 1.8925081433224755, + "grad_norm": 5.044738479793456, + "learning_rate": 1.427967349524533e-05, + "loss": 0.59, + "step": 3486 + }, + { + "epoch": 1.8930510314875135, + "grad_norm": 6.9998626206641195, + "learning_rate": 1.4276494731598582e-05, + "loss": 0.5973, + "step": 3487 + }, + { + "epoch": 1.8935939196525515, + "grad_norm": 8.316471360458495, + "learning_rate": 1.4273315439028151e-05, + "loss": 0.8941, + "step": 3488 + }, + { + "epoch": 1.8941368078175895, + "grad_norm": 12.713883241844957, + "learning_rate": 1.4270135617927254e-05, + "loss": 1.4921, + "step": 3489 + }, + { + "epoch": 1.8946796959826275, + "grad_norm": 10.043253133675226, + "learning_rate": 1.426695526868918e-05, + "loss": 1.0242, + "step": 3490 + }, + { + "epoch": 1.8952225841476655, + "grad_norm": 11.142761788047633, + "learning_rate": 1.4263774391707274e-05, + "loss": 1.265, + "step": 3491 + }, + { + "epoch": 1.8957654723127035, + "grad_norm": 7.12450280478247, + "learning_rate": 1.4260592987374962e-05, + "loss": 0.484, + "step": 3492 + }, + { + "epoch": 1.8963083604777415, + "grad_norm": 7.017938163026924, + "learning_rate": 1.4257411056085712e-05, + "loss": 0.6549, + "step": 3493 + }, + { + "epoch": 1.8968512486427795, + "grad_norm": 7.643680244898419, + "learning_rate": 1.4254228598233082e-05, + "loss": 0.614, + "step": 3494 + }, + { + "epoch": 1.8973941368078175, + "grad_norm": 9.484247292353862, + "learning_rate": 1.4251045614210678e-05, + "loss": 0.9667, + "step": 3495 + }, + { + "epoch": 1.8979370249728555, + "grad_norm": 10.846455465412243, + "learning_rate": 1.4247862104412175e-05, + "loss": 1.4425, + "step": 3496 + }, + { + "epoch": 1.8984799131378935, + "grad_norm": 7.211983265810251, + "learning_rate": 1.4244678069231319e-05, + "loss": 0.7508, + "step": 3497 + }, + { + "epoch": 1.8990228013029316, + "grad_norm": 8.283019454355765, + "learning_rate": 1.4241493509061912e-05, + "loss": 0.9747, + "step": 3498 + }, + { + "epoch": 1.8995656894679696, + "grad_norm": 6.678762476220025, + "learning_rate": 1.4238308424297833e-05, + "loss": 0.4702, + "step": 3499 + }, + { + "epoch": 1.9001085776330076, + "grad_norm": 8.750314729983545, + "learning_rate": 1.423512281533301e-05, + "loss": 0.7636, + "step": 3500 + }, + { + "epoch": 1.9006514657980456, + "grad_norm": 7.304249105438763, + "learning_rate": 1.4231936682561446e-05, + "loss": 0.7198, + "step": 3501 + }, + { + "epoch": 1.9011943539630836, + "grad_norm": 7.676960806676722, + "learning_rate": 1.4228750026377212e-05, + "loss": 0.8639, + "step": 3502 + }, + { + "epoch": 1.9017372421281216, + "grad_norm": 10.327339388528987, + "learning_rate": 1.4225562847174431e-05, + "loss": 0.9386, + "step": 3503 + }, + { + "epoch": 1.9022801302931596, + "grad_norm": 8.828430779713369, + "learning_rate": 1.4222375145347304e-05, + "loss": 1.0821, + "step": 3504 + }, + { + "epoch": 1.9028230184581976, + "grad_norm": 8.099300889556567, + "learning_rate": 1.421918692129009e-05, + "loss": 0.6251, + "step": 3505 + }, + { + "epoch": 1.9033659066232356, + "grad_norm": 9.197832742387652, + "learning_rate": 1.4215998175397115e-05, + "loss": 1.1794, + "step": 3506 + }, + { + "epoch": 1.9039087947882736, + "grad_norm": 9.305579433655693, + "learning_rate": 1.4212808908062763e-05, + "loss": 1.3447, + "step": 3507 + }, + { + "epoch": 1.9044516829533116, + "grad_norm": 6.897677914233707, + "learning_rate": 1.4209619119681497e-05, + "loss": 0.5178, + "step": 3508 + }, + { + "epoch": 1.9049945711183496, + "grad_norm": 9.359746343598495, + "learning_rate": 1.420642881064782e-05, + "loss": 0.985, + "step": 3509 + }, + { + "epoch": 1.9055374592833876, + "grad_norm": 5.855418324834113, + "learning_rate": 1.4203237981356331e-05, + "loss": 0.5144, + "step": 3510 + }, + { + "epoch": 1.9060803474484256, + "grad_norm": 7.408991833255532, + "learning_rate": 1.4200046632201665e-05, + "loss": 0.7358, + "step": 3511 + }, + { + "epoch": 1.9066232356134636, + "grad_norm": 9.024544122531433, + "learning_rate": 1.419685476357854e-05, + "loss": 0.778, + "step": 3512 + }, + { + "epoch": 1.9071661237785016, + "grad_norm": 7.874409137592207, + "learning_rate": 1.419366237588173e-05, + "loss": 0.8703, + "step": 3513 + }, + { + "epoch": 1.9077090119435396, + "grad_norm": 8.221027989379657, + "learning_rate": 1.4190469469506073e-05, + "loss": 0.7436, + "step": 3514 + }, + { + "epoch": 1.9082519001085776, + "grad_norm": 10.797659549408706, + "learning_rate": 1.4187276044846473e-05, + "loss": 1.0635, + "step": 3515 + }, + { + "epoch": 1.9087947882736156, + "grad_norm": 9.718910477411724, + "learning_rate": 1.4184082102297896e-05, + "loss": 0.8802, + "step": 3516 + }, + { + "epoch": 1.9093376764386536, + "grad_norm": 7.365463419520298, + "learning_rate": 1.4180887642255376e-05, + "loss": 0.5919, + "step": 3517 + }, + { + "epoch": 1.9098805646036916, + "grad_norm": 6.487376521266865, + "learning_rate": 1.4177692665114014e-05, + "loss": 0.7098, + "step": 3518 + }, + { + "epoch": 1.9104234527687296, + "grad_norm": 8.328541010595174, + "learning_rate": 1.4174497171268962e-05, + "loss": 0.6124, + "step": 3519 + }, + { + "epoch": 1.9109663409337676, + "grad_norm": 7.888322073346588, + "learning_rate": 1.4171301161115447e-05, + "loss": 0.827, + "step": 3520 + }, + { + "epoch": 1.9115092290988056, + "grad_norm": 10.407728423373431, + "learning_rate": 1.4168104635048756e-05, + "loss": 1.0511, + "step": 3521 + }, + { + "epoch": 1.9120521172638436, + "grad_norm": 8.154984984105669, + "learning_rate": 1.4164907593464239e-05, + "loss": 0.8156, + "step": 3522 + }, + { + "epoch": 1.9125950054288816, + "grad_norm": 9.142522776167024, + "learning_rate": 1.4161710036757314e-05, + "loss": 0.9046, + "step": 3523 + }, + { + "epoch": 1.9131378935939196, + "grad_norm": 8.610209396948733, + "learning_rate": 1.415851196532346e-05, + "loss": 0.5673, + "step": 3524 + }, + { + "epoch": 1.9136807817589576, + "grad_norm": 8.072698611315632, + "learning_rate": 1.415531337955822e-05, + "loss": 0.7375, + "step": 3525 + }, + { + "epoch": 1.9142236699239956, + "grad_norm": 9.878888535422037, + "learning_rate": 1.4152114279857197e-05, + "loss": 0.9677, + "step": 3526 + }, + { + "epoch": 1.9147665580890336, + "grad_norm": 8.524331697287977, + "learning_rate": 1.4148914666616062e-05, + "loss": 0.7617, + "step": 3527 + }, + { + "epoch": 1.9153094462540716, + "grad_norm": 8.04939882692512, + "learning_rate": 1.4145714540230549e-05, + "loss": 0.8767, + "step": 3528 + }, + { + "epoch": 1.9158523344191096, + "grad_norm": 10.69782500673993, + "learning_rate": 1.4142513901096453e-05, + "loss": 0.6357, + "step": 3529 + }, + { + "epoch": 1.9163952225841476, + "grad_norm": 7.279779735048647, + "learning_rate": 1.4139312749609637e-05, + "loss": 0.6474, + "step": 3530 + }, + { + "epoch": 1.9169381107491856, + "grad_norm": 7.290241222497209, + "learning_rate": 1.4136111086166024e-05, + "loss": 0.6929, + "step": 3531 + }, + { + "epoch": 1.9174809989142236, + "grad_norm": 8.3613074350218, + "learning_rate": 1.4132908911161598e-05, + "loss": 0.8199, + "step": 3532 + }, + { + "epoch": 1.9180238870792616, + "grad_norm": 9.015830183265928, + "learning_rate": 1.4129706224992413e-05, + "loss": 0.9811, + "step": 3533 + }, + { + "epoch": 1.9185667752442996, + "grad_norm": 9.621458967761185, + "learning_rate": 1.4126503028054579e-05, + "loss": 0.7498, + "step": 3534 + }, + { + "epoch": 1.9191096634093376, + "grad_norm": 11.105255757566027, + "learning_rate": 1.412329932074427e-05, + "loss": 0.7587, + "step": 3535 + }, + { + "epoch": 1.9196525515743756, + "grad_norm": 8.599145695443788, + "learning_rate": 1.4120095103457734e-05, + "loss": 0.934, + "step": 3536 + }, + { + "epoch": 1.9201954397394136, + "grad_norm": 7.727889210926642, + "learning_rate": 1.4116890376591268e-05, + "loss": 0.8305, + "step": 3537 + }, + { + "epoch": 1.9207383279044516, + "grad_norm": 12.029529213228301, + "learning_rate": 1.4113685140541242e-05, + "loss": 1.494, + "step": 3538 + }, + { + "epoch": 1.9212812160694897, + "grad_norm": 7.404838471492053, + "learning_rate": 1.411047939570408e-05, + "loss": 0.5168, + "step": 3539 + }, + { + "epoch": 1.9218241042345277, + "grad_norm": 7.540024472508733, + "learning_rate": 1.4107273142476272e-05, + "loss": 0.7162, + "step": 3540 + }, + { + "epoch": 1.9223669923995657, + "grad_norm": 7.885769945154571, + "learning_rate": 1.4104066381254378e-05, + "loss": 0.5964, + "step": 3541 + }, + { + "epoch": 1.9229098805646037, + "grad_norm": 8.287652724014645, + "learning_rate": 1.4100859112435013e-05, + "loss": 0.6175, + "step": 3542 + }, + { + "epoch": 1.9234527687296417, + "grad_norm": 8.335133777216624, + "learning_rate": 1.4097651336414857e-05, + "loss": 0.7163, + "step": 3543 + }, + { + "epoch": 1.9239956568946797, + "grad_norm": 7.765701865468595, + "learning_rate": 1.4094443053590652e-05, + "loss": 0.4534, + "step": 3544 + }, + { + "epoch": 1.9245385450597177, + "grad_norm": 7.716709219714201, + "learning_rate": 1.4091234264359206e-05, + "loss": 0.4425, + "step": 3545 + }, + { + "epoch": 1.9250814332247557, + "grad_norm": 11.979255681165771, + "learning_rate": 1.4088024969117387e-05, + "loss": 1.0249, + "step": 3546 + }, + { + "epoch": 1.9256243213897937, + "grad_norm": 9.424753754574079, + "learning_rate": 1.4084815168262123e-05, + "loss": 0.9663, + "step": 3547 + }, + { + "epoch": 1.9261672095548317, + "grad_norm": 10.101408847774994, + "learning_rate": 1.4081604862190407e-05, + "loss": 0.8136, + "step": 3548 + }, + { + "epoch": 1.9267100977198697, + "grad_norm": 11.755751498543427, + "learning_rate": 1.4078394051299298e-05, + "loss": 0.9451, + "step": 3549 + }, + { + "epoch": 1.9272529858849077, + "grad_norm": 12.251136923575924, + "learning_rate": 1.4075182735985913e-05, + "loss": 1.5716, + "step": 3550 + }, + { + "epoch": 1.9277958740499457, + "grad_norm": 10.087922707091717, + "learning_rate": 1.4071970916647432e-05, + "loss": 0.7026, + "step": 3551 + }, + { + "epoch": 1.9283387622149837, + "grad_norm": 8.507872704310849, + "learning_rate": 1.40687585936811e-05, + "loss": 0.7016, + "step": 3552 + }, + { + "epoch": 1.9288816503800217, + "grad_norm": 8.98138340766033, + "learning_rate": 1.4065545767484218e-05, + "loss": 0.6834, + "step": 3553 + }, + { + "epoch": 1.9294245385450597, + "grad_norm": 7.1085330003672045, + "learning_rate": 1.4062332438454156e-05, + "loss": 0.6813, + "step": 3554 + }, + { + "epoch": 1.9299674267100977, + "grad_norm": 13.499346110936985, + "learning_rate": 1.4059118606988345e-05, + "loss": 1.0068, + "step": 3555 + }, + { + "epoch": 1.9305103148751357, + "grad_norm": 7.698407375447584, + "learning_rate": 1.4055904273484275e-05, + "loss": 0.8418, + "step": 3556 + }, + { + "epoch": 1.9310532030401737, + "grad_norm": 8.424294783097983, + "learning_rate": 1.40526894383395e-05, + "loss": 0.8087, + "step": 3557 + }, + { + "epoch": 1.9315960912052117, + "grad_norm": 8.962780671623113, + "learning_rate": 1.4049474101951639e-05, + "loss": 0.8644, + "step": 3558 + }, + { + "epoch": 1.9321389793702497, + "grad_norm": 8.580581507909985, + "learning_rate": 1.4046258264718363e-05, + "loss": 1.0325, + "step": 3559 + }, + { + "epoch": 1.9326818675352877, + "grad_norm": 10.417878574145533, + "learning_rate": 1.4043041927037418e-05, + "loss": 0.7705, + "step": 3560 + }, + { + "epoch": 1.9332247557003257, + "grad_norm": 8.250628073154727, + "learning_rate": 1.4039825089306603e-05, + "loss": 0.794, + "step": 3561 + }, + { + "epoch": 1.9337676438653637, + "grad_norm": 8.228275233396532, + "learning_rate": 1.4036607751923783e-05, + "loss": 0.8759, + "step": 3562 + }, + { + "epoch": 1.9343105320304017, + "grad_norm": 7.760702739983814, + "learning_rate": 1.4033389915286884e-05, + "loss": 0.6458, + "step": 3563 + }, + { + "epoch": 1.9348534201954397, + "grad_norm": 6.508974748032192, + "learning_rate": 1.403017157979389e-05, + "loss": 0.6348, + "step": 3564 + }, + { + "epoch": 1.9353963083604777, + "grad_norm": 11.570824985629717, + "learning_rate": 1.4026952745842853e-05, + "loss": 1.0935, + "step": 3565 + }, + { + "epoch": 1.9359391965255157, + "grad_norm": 9.901046454739335, + "learning_rate": 1.4023733413831881e-05, + "loss": 0.7458, + "step": 3566 + }, + { + "epoch": 1.9364820846905537, + "grad_norm": 9.103951624002182, + "learning_rate": 1.402051358415915e-05, + "loss": 0.8253, + "step": 3567 + }, + { + "epoch": 1.9370249728555917, + "grad_norm": 9.293062225974824, + "learning_rate": 1.4017293257222887e-05, + "loss": 0.5801, + "step": 3568 + }, + { + "epoch": 1.9375678610206297, + "grad_norm": 7.180133735690273, + "learning_rate": 1.4014072433421398e-05, + "loss": 0.6063, + "step": 3569 + }, + { + "epoch": 1.9381107491856677, + "grad_norm": 9.028614205338394, + "learning_rate": 1.4010851113153028e-05, + "loss": 0.9412, + "step": 3570 + }, + { + "epoch": 1.9386536373507057, + "grad_norm": 9.51192022384654, + "learning_rate": 1.4007629296816202e-05, + "loss": 0.6459, + "step": 3571 + }, + { + "epoch": 1.9391965255157437, + "grad_norm": 9.883078787028563, + "learning_rate": 1.4004406984809396e-05, + "loss": 1.0324, + "step": 3572 + }, + { + "epoch": 1.9397394136807817, + "grad_norm": 7.433318187249177, + "learning_rate": 1.4001184177531154e-05, + "loss": 0.6152, + "step": 3573 + }, + { + "epoch": 1.9402823018458197, + "grad_norm": 10.165800890508075, + "learning_rate": 1.399796087538007e-05, + "loss": 0.8751, + "step": 3574 + }, + { + "epoch": 1.9408251900108577, + "grad_norm": 7.668138219396951, + "learning_rate": 1.3994737078754819e-05, + "loss": 0.43, + "step": 3575 + }, + { + "epoch": 1.9413680781758957, + "grad_norm": 8.929591410847367, + "learning_rate": 1.3991512788054115e-05, + "loss": 0.6173, + "step": 3576 + }, + { + "epoch": 1.9419109663409337, + "grad_norm": 8.069899029370944, + "learning_rate": 1.3988288003676755e-05, + "loss": 0.6782, + "step": 3577 + }, + { + "epoch": 1.9424538545059717, + "grad_norm": 7.635775018643444, + "learning_rate": 1.3985062726021574e-05, + "loss": 0.7503, + "step": 3578 + }, + { + "epoch": 1.9429967426710097, + "grad_norm": 11.649932328973133, + "learning_rate": 1.3981836955487485e-05, + "loss": 0.9154, + "step": 3579 + }, + { + "epoch": 1.9435396308360477, + "grad_norm": 9.11865472044189, + "learning_rate": 1.397861069247345e-05, + "loss": 0.8799, + "step": 3580 + }, + { + "epoch": 1.9440825190010858, + "grad_norm": 7.628563459317163, + "learning_rate": 1.3975383937378508e-05, + "loss": 0.7199, + "step": 3581 + }, + { + "epoch": 1.9446254071661238, + "grad_norm": 10.640127088045624, + "learning_rate": 1.3972156690601747e-05, + "loss": 0.7885, + "step": 3582 + }, + { + "epoch": 1.9451682953311618, + "grad_norm": 10.192406733466123, + "learning_rate": 1.3968928952542313e-05, + "loss": 0.6036, + "step": 3583 + }, + { + "epoch": 1.9457111834961998, + "grad_norm": 12.020287450972278, + "learning_rate": 1.396570072359942e-05, + "loss": 0.9477, + "step": 3584 + }, + { + "epoch": 1.9462540716612378, + "grad_norm": 8.855072111194623, + "learning_rate": 1.3962472004172343e-05, + "loss": 0.4384, + "step": 3585 + }, + { + "epoch": 1.9467969598262758, + "grad_norm": 7.173236338345466, + "learning_rate": 1.3959242794660412e-05, + "loss": 0.4969, + "step": 3586 + }, + { + "epoch": 1.9473398479913138, + "grad_norm": 6.505005692892076, + "learning_rate": 1.3956013095463024e-05, + "loss": 0.4585, + "step": 3587 + }, + { + "epoch": 1.9478827361563518, + "grad_norm": 7.041517208688192, + "learning_rate": 1.395278290697963e-05, + "loss": 0.5439, + "step": 3588 + }, + { + "epoch": 1.9484256243213898, + "grad_norm": 9.730465359550896, + "learning_rate": 1.3949552229609746e-05, + "loss": 0.8865, + "step": 3589 + }, + { + "epoch": 1.9489685124864278, + "grad_norm": 7.73678668778, + "learning_rate": 1.3946321063752948e-05, + "loss": 0.5094, + "step": 3590 + }, + { + "epoch": 1.9495114006514658, + "grad_norm": 7.095011334397053, + "learning_rate": 1.3943089409808872e-05, + "loss": 0.5224, + "step": 3591 + }, + { + "epoch": 1.9500542888165038, + "grad_norm": 8.086388468165458, + "learning_rate": 1.393985726817721e-05, + "loss": 0.7798, + "step": 3592 + }, + { + "epoch": 1.9505971769815418, + "grad_norm": 10.262604564335476, + "learning_rate": 1.3936624639257726e-05, + "loss": 1.2595, + "step": 3593 + }, + { + "epoch": 1.9511400651465798, + "grad_norm": 8.535907967530573, + "learning_rate": 1.393339152345023e-05, + "loss": 0.8205, + "step": 3594 + }, + { + "epoch": 1.9516829533116178, + "grad_norm": 7.864477340620307, + "learning_rate": 1.3930157921154601e-05, + "loss": 0.5967, + "step": 3595 + }, + { + "epoch": 1.9522258414766558, + "grad_norm": 6.739275201436143, + "learning_rate": 1.392692383277078e-05, + "loss": 0.5836, + "step": 3596 + }, + { + "epoch": 1.9527687296416938, + "grad_norm": 9.533970605538912, + "learning_rate": 1.392368925869876e-05, + "loss": 0.5761, + "step": 3597 + }, + { + "epoch": 1.9533116178067318, + "grad_norm": 11.376343280359167, + "learning_rate": 1.3920454199338598e-05, + "loss": 0.7596, + "step": 3598 + }, + { + "epoch": 1.9538545059717698, + "grad_norm": 7.280453649201749, + "learning_rate": 1.391721865509041e-05, + "loss": 0.4949, + "step": 3599 + }, + { + "epoch": 1.9543973941368078, + "grad_norm": 7.734190007799969, + "learning_rate": 1.391398262635438e-05, + "loss": 0.9177, + "step": 3600 + }, + { + "epoch": 1.9549402823018458, + "grad_norm": 13.106124045444247, + "learning_rate": 1.3910746113530738e-05, + "loss": 1.1884, + "step": 3601 + }, + { + "epoch": 1.9554831704668838, + "grad_norm": 11.868912051178873, + "learning_rate": 1.3907509117019783e-05, + "loss": 0.9245, + "step": 3602 + }, + { + "epoch": 1.9560260586319218, + "grad_norm": 10.307498071834987, + "learning_rate": 1.3904271637221876e-05, + "loss": 1.0802, + "step": 3603 + }, + { + "epoch": 1.9565689467969598, + "grad_norm": 8.308697157532139, + "learning_rate": 1.390103367453743e-05, + "loss": 0.4979, + "step": 3604 + }, + { + "epoch": 1.9571118349619978, + "grad_norm": 8.447604525372762, + "learning_rate": 1.3897795229366919e-05, + "loss": 0.8942, + "step": 3605 + }, + { + "epoch": 1.9576547231270358, + "grad_norm": 7.864013969756676, + "learning_rate": 1.3894556302110883e-05, + "loss": 0.642, + "step": 3606 + }, + { + "epoch": 1.9581976112920738, + "grad_norm": 7.903822544650853, + "learning_rate": 1.389131689316992e-05, + "loss": 0.8251, + "step": 3607 + }, + { + "epoch": 1.9587404994571118, + "grad_norm": 9.744275276467558, + "learning_rate": 1.3888077002944678e-05, + "loss": 1.0052, + "step": 3608 + }, + { + "epoch": 1.9592833876221498, + "grad_norm": 7.731164092689308, + "learning_rate": 1.3884836631835877e-05, + "loss": 0.6883, + "step": 3609 + }, + { + "epoch": 1.9598262757871878, + "grad_norm": 8.849291385162568, + "learning_rate": 1.3881595780244288e-05, + "loss": 0.9985, + "step": 3610 + }, + { + "epoch": 1.9603691639522258, + "grad_norm": 9.531929458494986, + "learning_rate": 1.3878354448570748e-05, + "loss": 0.7643, + "step": 3611 + }, + { + "epoch": 1.9609120521172638, + "grad_norm": 10.32443732773501, + "learning_rate": 1.3875112637216145e-05, + "loss": 0.848, + "step": 3612 + }, + { + "epoch": 1.9614549402823018, + "grad_norm": 7.417155617136626, + "learning_rate": 1.387187034658144e-05, + "loss": 0.5682, + "step": 3613 + }, + { + "epoch": 1.9619978284473398, + "grad_norm": 14.072831276291689, + "learning_rate": 1.386862757706764e-05, + "loss": 0.8279, + "step": 3614 + }, + { + "epoch": 1.9625407166123778, + "grad_norm": 7.764451928370161, + "learning_rate": 1.3865384329075812e-05, + "loss": 0.6687, + "step": 3615 + }, + { + "epoch": 1.9630836047774158, + "grad_norm": 8.686094567618806, + "learning_rate": 1.3862140603007095e-05, + "loss": 1.1034, + "step": 3616 + }, + { + "epoch": 1.9636264929424538, + "grad_norm": 10.753407337858437, + "learning_rate": 1.3858896399262669e-05, + "loss": 1.1692, + "step": 3617 + }, + { + "epoch": 1.9641693811074918, + "grad_norm": 9.096005648867607, + "learning_rate": 1.3855651718243786e-05, + "loss": 0.6191, + "step": 3618 + }, + { + "epoch": 1.9647122692725298, + "grad_norm": 10.387647049254962, + "learning_rate": 1.3852406560351752e-05, + "loss": 1.3573, + "step": 3619 + }, + { + "epoch": 1.9652551574375678, + "grad_norm": 7.173958455599766, + "learning_rate": 1.3849160925987936e-05, + "loss": 0.8106, + "step": 3620 + }, + { + "epoch": 1.9657980456026058, + "grad_norm": 9.434903281870886, + "learning_rate": 1.3845914815553765e-05, + "loss": 0.9581, + "step": 3621 + }, + { + "epoch": 1.9663409337676439, + "grad_norm": 7.671221453670278, + "learning_rate": 1.3842668229450717e-05, + "loss": 0.6989, + "step": 3622 + }, + { + "epoch": 1.9668838219326819, + "grad_norm": 7.360534692664803, + "learning_rate": 1.3839421168080338e-05, + "loss": 0.4912, + "step": 3623 + }, + { + "epoch": 1.9674267100977199, + "grad_norm": 6.853527063108637, + "learning_rate": 1.3836173631844231e-05, + "loss": 0.7485, + "step": 3624 + }, + { + "epoch": 1.9679695982627579, + "grad_norm": 8.619792011062271, + "learning_rate": 1.3832925621144057e-05, + "loss": 0.712, + "step": 3625 + }, + { + "epoch": 1.9685124864277959, + "grad_norm": 8.062990147577946, + "learning_rate": 1.3829677136381532e-05, + "loss": 0.7126, + "step": 3626 + }, + { + "epoch": 1.9690553745928339, + "grad_norm": 6.35400375341867, + "learning_rate": 1.3826428177958433e-05, + "loss": 0.5262, + "step": 3627 + }, + { + "epoch": 1.9695982627578719, + "grad_norm": 7.322594621519839, + "learning_rate": 1.3823178746276603e-05, + "loss": 0.6644, + "step": 3628 + }, + { + "epoch": 1.9701411509229099, + "grad_norm": 9.976906946653425, + "learning_rate": 1.3819928841737929e-05, + "loss": 0.7411, + "step": 3629 + }, + { + "epoch": 1.9706840390879479, + "grad_norm": 11.052922199516026, + "learning_rate": 1.3816678464744368e-05, + "loss": 0.9998, + "step": 3630 + }, + { + "epoch": 1.9712269272529859, + "grad_norm": 7.695988647358978, + "learning_rate": 1.381342761569793e-05, + "loss": 0.6805, + "step": 3631 + }, + { + "epoch": 1.9717698154180239, + "grad_norm": 10.387946538947885, + "learning_rate": 1.381017629500069e-05, + "loss": 0.8954, + "step": 3632 + }, + { + "epoch": 1.9723127035830619, + "grad_norm": 8.215553841008688, + "learning_rate": 1.380692450305477e-05, + "loss": 1.0488, + "step": 3633 + }, + { + "epoch": 1.9728555917480999, + "grad_norm": 8.550898170705977, + "learning_rate": 1.3803672240262364e-05, + "loss": 0.9301, + "step": 3634 + }, + { + "epoch": 1.9733984799131379, + "grad_norm": 9.978410770263547, + "learning_rate": 1.380041950702571e-05, + "loss": 0.9725, + "step": 3635 + }, + { + "epoch": 1.9739413680781759, + "grad_norm": 10.211160158432088, + "learning_rate": 1.3797166303747119e-05, + "loss": 1.1789, + "step": 3636 + }, + { + "epoch": 1.9744842562432139, + "grad_norm": 7.47091491709439, + "learning_rate": 1.3793912630828942e-05, + "loss": 1.2311, + "step": 3637 + }, + { + "epoch": 1.975027144408252, + "grad_norm": 9.966534074490724, + "learning_rate": 1.3790658488673607e-05, + "loss": 1.0057, + "step": 3638 + }, + { + "epoch": 1.97557003257329, + "grad_norm": 9.341517091004942, + "learning_rate": 1.378740387768359e-05, + "loss": 0.7708, + "step": 3639 + }, + { + "epoch": 1.976112920738328, + "grad_norm": 10.066755071795829, + "learning_rate": 1.3784148798261422e-05, + "loss": 1.3146, + "step": 3640 + }, + { + "epoch": 1.976655808903366, + "grad_norm": 8.306959980081418, + "learning_rate": 1.3780893250809705e-05, + "loss": 0.6926, + "step": 3641 + }, + { + "epoch": 1.977198697068404, + "grad_norm": 5.980548702977229, + "learning_rate": 1.377763723573108e-05, + "loss": 0.6776, + "step": 3642 + }, + { + "epoch": 1.977741585233442, + "grad_norm": 10.796897781225315, + "learning_rate": 1.3774380753428263e-05, + "loss": 1.0422, + "step": 3643 + }, + { + "epoch": 1.97828447339848, + "grad_norm": 9.650854393967004, + "learning_rate": 1.3771123804304018e-05, + "loss": 1.2566, + "step": 3644 + }, + { + "epoch": 1.978827361563518, + "grad_norm": 6.718663747121732, + "learning_rate": 1.3767866388761168e-05, + "loss": 0.5018, + "step": 3645 + }, + { + "epoch": 1.979370249728556, + "grad_norm": 8.345918705516796, + "learning_rate": 1.3764608507202604e-05, + "loss": 0.8501, + "step": 3646 + }, + { + "epoch": 1.979913137893594, + "grad_norm": 11.065999609001787, + "learning_rate": 1.3761350160031258e-05, + "loss": 1.2627, + "step": 3647 + }, + { + "epoch": 1.980456026058632, + "grad_norm": 8.567448073304528, + "learning_rate": 1.3758091347650126e-05, + "loss": 0.4477, + "step": 3648 + }, + { + "epoch": 1.98099891422367, + "grad_norm": 11.930236219703469, + "learning_rate": 1.3754832070462269e-05, + "loss": 1.5386, + "step": 3649 + }, + { + "epoch": 1.981541802388708, + "grad_norm": 7.334841489157127, + "learning_rate": 1.3751572328870797e-05, + "loss": 0.5759, + "step": 3650 + }, + { + "epoch": 1.982084690553746, + "grad_norm": 6.392865090419794, + "learning_rate": 1.3748312123278879e-05, + "loss": 0.6078, + "step": 3651 + }, + { + "epoch": 1.982627578718784, + "grad_norm": 8.598388892264492, + "learning_rate": 1.3745051454089744e-05, + "loss": 0.7682, + "step": 3652 + }, + { + "epoch": 1.983170466883822, + "grad_norm": 9.310542957266444, + "learning_rate": 1.3741790321706678e-05, + "loss": 0.9833, + "step": 3653 + }, + { + "epoch": 1.98371335504886, + "grad_norm": 8.917570196452148, + "learning_rate": 1.3738528726533021e-05, + "loss": 0.5915, + "step": 3654 + }, + { + "epoch": 1.984256243213898, + "grad_norm": 9.503202436394638, + "learning_rate": 1.3735266668972174e-05, + "loss": 0.9347, + "step": 3655 + }, + { + "epoch": 1.984799131378936, + "grad_norm": 9.187024367480415, + "learning_rate": 1.3732004149427592e-05, + "loss": 0.6669, + "step": 3656 + }, + { + "epoch": 1.985342019543974, + "grad_norm": 10.497812893691144, + "learning_rate": 1.3728741168302785e-05, + "loss": 1.0663, + "step": 3657 + }, + { + "epoch": 1.985884907709012, + "grad_norm": 9.646030688042543, + "learning_rate": 1.3725477726001332e-05, + "loss": 0.9913, + "step": 3658 + }, + { + "epoch": 1.98642779587405, + "grad_norm": 9.475431343889321, + "learning_rate": 1.3722213822926855e-05, + "loss": 0.8377, + "step": 3659 + }, + { + "epoch": 1.986970684039088, + "grad_norm": 9.449591675048769, + "learning_rate": 1.371894945948304e-05, + "loss": 0.8763, + "step": 3660 + }, + { + "epoch": 1.987513572204126, + "grad_norm": 7.3204142422253025, + "learning_rate": 1.3715684636073628e-05, + "loss": 0.7384, + "step": 3661 + }, + { + "epoch": 1.988056460369164, + "grad_norm": 10.339298014193695, + "learning_rate": 1.371241935310242e-05, + "loss": 1.0206, + "step": 3662 + }, + { + "epoch": 1.988599348534202, + "grad_norm": 7.083819102334954, + "learning_rate": 1.3709153610973266e-05, + "loss": 0.6905, + "step": 3663 + }, + { + "epoch": 1.98914223669924, + "grad_norm": 6.318629816327554, + "learning_rate": 1.3705887410090085e-05, + "loss": 0.7946, + "step": 3664 + }, + { + "epoch": 1.989685124864278, + "grad_norm": 7.109392074745231, + "learning_rate": 1.3702620750856843e-05, + "loss": 0.7125, + "step": 3665 + }, + { + "epoch": 1.990228013029316, + "grad_norm": 6.7832401368423545, + "learning_rate": 1.3699353633677565e-05, + "loss": 0.6091, + "step": 3666 + }, + { + "epoch": 1.990770901194354, + "grad_norm": 9.301471508167843, + "learning_rate": 1.3696086058956333e-05, + "loss": 0.9335, + "step": 3667 + }, + { + "epoch": 1.991313789359392, + "grad_norm": 8.167349191349723, + "learning_rate": 1.3692818027097288e-05, + "loss": 0.7148, + "step": 3668 + }, + { + "epoch": 1.99185667752443, + "grad_norm": 7.582331201930941, + "learning_rate": 1.3689549538504622e-05, + "loss": 0.5952, + "step": 3669 + }, + { + "epoch": 1.992399565689468, + "grad_norm": 8.743965167551632, + "learning_rate": 1.3686280593582588e-05, + "loss": 0.9268, + "step": 3670 + }, + { + "epoch": 1.992942453854506, + "grad_norm": 10.291738284427778, + "learning_rate": 1.3683011192735496e-05, + "loss": 0.9934, + "step": 3671 + }, + { + "epoch": 1.993485342019544, + "grad_norm": 9.808936601876956, + "learning_rate": 1.3679741336367711e-05, + "loss": 1.0506, + "step": 3672 + }, + { + "epoch": 1.994028230184582, + "grad_norm": 10.002691567142412, + "learning_rate": 1.3676471024883654e-05, + "loss": 1.1439, + "step": 3673 + }, + { + "epoch": 1.99457111834962, + "grad_norm": 7.558103827510508, + "learning_rate": 1.36732002586878e-05, + "loss": 1.0059, + "step": 3674 + }, + { + "epoch": 1.995114006514658, + "grad_norm": 7.958890420546252, + "learning_rate": 1.3669929038184684e-05, + "loss": 1.0301, + "step": 3675 + }, + { + "epoch": 1.995656894679696, + "grad_norm": 8.042682701022132, + "learning_rate": 1.3666657363778895e-05, + "loss": 0.6881, + "step": 3676 + }, + { + "epoch": 1.996199782844734, + "grad_norm": 7.54271762136155, + "learning_rate": 1.366338523587508e-05, + "loss": 0.5227, + "step": 3677 + }, + { + "epoch": 1.996742671009772, + "grad_norm": 9.060117135659972, + "learning_rate": 1.3660112654877939e-05, + "loss": 0.9847, + "step": 3678 + }, + { + "epoch": 1.99728555917481, + "grad_norm": 11.00270077841202, + "learning_rate": 1.3656839621192233e-05, + "loss": 1.0918, + "step": 3679 + }, + { + "epoch": 1.997828447339848, + "grad_norm": 9.156434005073692, + "learning_rate": 1.3653566135222774e-05, + "loss": 0.8104, + "step": 3680 + }, + { + "epoch": 1.998371335504886, + "grad_norm": 8.405413514743108, + "learning_rate": 1.3650292197374433e-05, + "loss": 0.9414, + "step": 3681 + }, + { + "epoch": 1.998914223669924, + "grad_norm": 7.9594907003419495, + "learning_rate": 1.3647017808052135e-05, + "loss": 0.9257, + "step": 3682 + }, + { + "epoch": 1.999457111834962, + "grad_norm": 9.821777823228977, + "learning_rate": 1.3643742967660859e-05, + "loss": 0.8787, + "step": 3683 + }, + { + "epoch": 2.0, + "grad_norm": 8.766507481527489, + "learning_rate": 1.3640467676605648e-05, + "loss": 0.8472, + "step": 3684 + } + ], + "logging_steps": 1.0, + "max_steps": 9210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4395608982945792.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/training_args.bin b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b16de63f342ea53d59c4d97aa8a3bd463a79531f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9801fdd09c6ba798e12b77f77221db3b5747b6e1296785789d44cf2bc492ff9e +size 7160 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/zero_to_fp32.py b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-3684/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/README.md b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/README.md new file mode 100644 index 0000000000000000000000000000000000000000..972a036203d877262d3c6673f4d81814e7409dc5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/README.md @@ -0,0 +1,202 @@ +--- +base_model: llava-hf/LLaVA-NeXT-Video-7B-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/adapter_config.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..719f6e47b29561ff93d519680ccd115b480d825e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/adapter_config.json @@ -0,0 +1,132 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "llava-hf/LLaVA-NeXT-Video-7B-hf", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 8, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [], + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "model.layers.17.self_attn.q_proj", + "model.layers.17.self_attn.k_proj", + "model.layers.13.self_attn.k_proj", + "model.layers.20.self_attn.q_proj", + "model.layers.18.self_attn.k_proj", + "29.self_attn.k_proj", + "model.layers.16.self_attn.v_proj", + "model.layers.14.self_attn.q_proj", + "29.self_attn.v_proj", + "24.self_attn.k_proj", + "model.layers.18.self_attn.v_proj", + "model.layers.5.self_attn.q_proj", + "model.layers.11.self_attn.k_proj", + "27.self_attn.v_proj", + "model.layers.16.self_attn.q_proj", + "model.layers.10.self_attn.q_proj", + "model.layers.12.self_attn.v_proj", + "model.layers.20.self_attn.v_proj", + "model.layers.0.self_attn.v_proj", + "model.layers.7.self_attn.q_proj", + "model.layers.23.self_attn.q_proj", + "model.layers.1.self_attn.q_proj", + "25.self_attn.v_proj", + "model.layers.4.self_attn.v_proj", + "model.layers.19.self_attn.q_proj", + "model.layers.10.self_attn.k_proj", + "model.layers.22.self_attn.k_proj", + "model.layers.14.self_attn.v_proj", + "model.layers.13.self_attn.v_proj", + "25.self_attn.k_proj", + "24.self_attn.v_proj", + "model.layers.21.self_attn.q_proj", + "model.layers.8.self_attn.q_proj", + "model.layers.6.self_attn.q_proj", + "model.layers.6.self_attn.k_proj", + "model.layers.1.self_attn.v_proj", + "model.layers.9.self_attn.k_proj", + "model.layers.13.self_attn.q_proj", + "o_proj", + "model.layers.16.self_attn.k_proj", + "model.layers.6.self_attn.v_proj", + "model.layers.21.self_attn.k_proj", + "model.layers.5.self_attn.k_proj", + "27.self_attn.q_proj", + "model.layers.20.self_attn.k_proj", + "model.layers.11.self_attn.v_proj", + "model.layers.22.self_attn.q_proj", + "model.layers.14.self_attn.k_proj", + "28.self_attn.v_proj", + "model.layers.22.self_attn.v_proj", + "gate_proj", + "model.layers.0.self_attn.k_proj", + "model.layers.5.self_attn.v_proj", + "30.self_attn.q_proj", + "model.layers.3.self_attn.k_proj", + "29.self_attn.q_proj", + "model.layers.19.self_attn.k_proj", + "model.layers.10.self_attn.v_proj", + "model.layers.15.self_attn.k_proj", + "model.layers.7.self_attn.v_proj", + "model.layers.9.self_attn.v_proj", + "up_proj", + "model.layers.4.self_attn.q_proj", + "28.self_attn.q_proj", + "model.layers.3.self_attn.q_proj", + "24.self_attn.q_proj", + "30.self_attn.v_proj", + "model.layers.12.self_attn.k_proj", + "model.layers.23.self_attn.k_proj", + "26.self_attn.q_proj", + "28.self_attn.k_proj", + "31.self_attn.k_proj", + "down_proj", + "model.layers.2.self_attn.q_proj", + "31.self_attn.q_proj", + "model.layers.15.self_attn.q_proj", + "model.layers.12.self_attn.q_proj", + "model.layers.0.self_attn.q_proj", + "model.layers.2.self_attn.k_proj", + "model.layers.3.self_attn.v_proj", + "model.layers.11.self_attn.q_proj", + "25.self_attn.q_proj", + "model.layers.23.self_attn.v_proj", + "30.self_attn.k_proj", + "31.self_attn.v_proj", + "model.layers.15.self_attn.v_proj", + "model.layers.8.self_attn.v_proj", + "26.self_attn.k_proj", + "model.layers.19.self_attn.v_proj", + "model.layers.2.self_attn.v_proj", + "26.self_attn.v_proj", + "model.layers.21.self_attn.v_proj", + "model.layers.17.self_attn.v_proj", + "model.layers.8.self_attn.k_proj", + "model.layers.18.self_attn.q_proj", + "model.layers.7.self_attn.k_proj", + "model.layers.4.self_attn.k_proj", + "model.layers.9.self_attn.q_proj", + "27.self_attn.k_proj", + "model.layers.1.self_attn.k_proj" + ], + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/adapter_model.safetensors b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5b12fd0c6c842ae3d70317296be37aaa08fbaa09 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3162ee7a1b8f9f928900896cbe075d9def17a36072478e02c0414319075d1eb7 +size 40043208 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..bf100c5de334b90e58e668d8ba34a5830ce416ae --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4eaf3a792568f725fc1fa8bef29e204323640fc124cbcac318d77efabebd6812 +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..a1f26206f926ea33de0b14d6ca4a14dde5d05e14 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bc45de9d0dbd6d17aaa23d4cfaf9d90caaab5586b6175d02a6696880cc7447d +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/zero_pp_rank_0_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/zero_pp_rank_0_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..86ab23fbf0edd92d3920d8e8328ef25bb61fc6e9 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/zero_pp_rank_0_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe91b67e635e24c30f3684d8e295a1a45cfc1df4a6beee8e45ac127a25edc29c +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/zero_pp_rank_1_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/zero_pp_rank_1_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..908f7b46cbf86bcc104bee134cabdc053ba6bb99 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/global_step5526/zero_pp_rank_1_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8269f1798031dc831ef87903143e78e03055189857b7a17b74f186994c185951 +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/latest b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/latest new file mode 100644 index 0000000000000000000000000000000000000000..f2454b3aa69e6e9f720ce82327b9f778c77264fa --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/latest @@ -0,0 +1 @@ +global_step5526 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/rng_state_0.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..b37f7f9bfbe06b9470433749d11d743cd3b78023 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56fcf9c09f0b9c14ba3a15873bfc2362a8af95437fa546cd2703630f9c158bb9 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/rng_state_1.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..f5a264f7d67424631867b7ca710aa64eb246a58b --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23e9de0d3da1183e04a47f936a89d7d8b00ea617879e2e534baf07e6386450ef +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/scheduler.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..2259c855f9150cf306b53673129579299274df8e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ecc2d122326a666161a455198867c5cb5c47dc9132607e11da61da51736ed6b +size 1064 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/trainer_state.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..bd0ddfe531545380b56a5b226870237dfb315c4e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/trainer_state.json @@ -0,0 +1,38715 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.0, + "eval_steps": 500, + "global_step": 5526, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005428881650380022, + "grad_norm": 2.1320506281889067, + "learning_rate": 7.220216606498195e-08, + "loss": 5.1936, + "step": 1 + }, + { + "epoch": 0.0010857763300760044, + "grad_norm": 1.8633488766761366, + "learning_rate": 1.444043321299639e-07, + "loss": 4.6995, + "step": 2 + }, + { + "epoch": 0.0016286644951140066, + "grad_norm": 2.063286098879776, + "learning_rate": 2.1660649819494586e-07, + "loss": 4.4902, + "step": 3 + }, + { + "epoch": 0.002171552660152009, + "grad_norm": 1.7123050230907495, + "learning_rate": 2.888086642599278e-07, + "loss": 4.9125, + "step": 4 + }, + { + "epoch": 0.0027144408251900108, + "grad_norm": 1.7499657276799891, + "learning_rate": 3.610108303249098e-07, + "loss": 3.7648, + "step": 5 + }, + { + "epoch": 0.003257328990228013, + "grad_norm": 2.0419045636066886, + "learning_rate": 4.332129963898917e-07, + "loss": 4.7842, + "step": 6 + }, + { + "epoch": 0.003800217155266015, + "grad_norm": 1.4960282529105982, + "learning_rate": 5.054151624548736e-07, + "loss": 4.9333, + "step": 7 + }, + { + "epoch": 0.004343105320304018, + "grad_norm": 1.640954967476452, + "learning_rate": 5.776173285198556e-07, + "loss": 3.9576, + "step": 8 + }, + { + "epoch": 0.004885993485342019, + "grad_norm": 2.0590500459610244, + "learning_rate": 6.498194945848375e-07, + "loss": 4.3833, + "step": 9 + }, + { + "epoch": 0.0054288816503800215, + "grad_norm": 1.07709353257816, + "learning_rate": 7.220216606498196e-07, + "loss": 3.7669, + "step": 10 + }, + { + "epoch": 0.005971769815418024, + "grad_norm": 1.9506235655782926, + "learning_rate": 7.942238267148016e-07, + "loss": 4.5975, + "step": 11 + }, + { + "epoch": 0.006514657980456026, + "grad_norm": 1.411983096954528, + "learning_rate": 8.664259927797834e-07, + "loss": 3.9706, + "step": 12 + }, + { + "epoch": 0.007057546145494028, + "grad_norm": 1.7110053131090028, + "learning_rate": 9.386281588447654e-07, + "loss": 4.9081, + "step": 13 + }, + { + "epoch": 0.00760043431053203, + "grad_norm": 1.3689315757199692, + "learning_rate": 1.0108303249097473e-06, + "loss": 4.2426, + "step": 14 + }, + { + "epoch": 0.008143322475570033, + "grad_norm": 1.5612091671030481, + "learning_rate": 1.0830324909747293e-06, + "loss": 4.5144, + "step": 15 + }, + { + "epoch": 0.008686210640608035, + "grad_norm": 1.3725168953865754, + "learning_rate": 1.1552346570397112e-06, + "loss": 4.9378, + "step": 16 + }, + { + "epoch": 0.009229098805646038, + "grad_norm": 2.027359913808926, + "learning_rate": 1.2274368231046932e-06, + "loss": 3.7477, + "step": 17 + }, + { + "epoch": 0.009771986970684038, + "grad_norm": 2.1695456962160224, + "learning_rate": 1.299638989169675e-06, + "loss": 5.7636, + "step": 18 + }, + { + "epoch": 0.01031487513572204, + "grad_norm": 3.2637002287680428, + "learning_rate": 1.3718411552346572e-06, + "loss": 6.1883, + "step": 19 + }, + { + "epoch": 0.010857763300760043, + "grad_norm": 2.5723616120643857, + "learning_rate": 1.4440433212996392e-06, + "loss": 4.0267, + "step": 20 + }, + { + "epoch": 0.011400651465798045, + "grad_norm": 1.4473754577877487, + "learning_rate": 1.516245487364621e-06, + "loss": 4.526, + "step": 21 + }, + { + "epoch": 0.011943539630836048, + "grad_norm": 1.6098208731761, + "learning_rate": 1.5884476534296031e-06, + "loss": 4.5163, + "step": 22 + }, + { + "epoch": 0.01248642779587405, + "grad_norm": 2.899624682573202, + "learning_rate": 1.6606498194945849e-06, + "loss": 4.8188, + "step": 23 + }, + { + "epoch": 0.013029315960912053, + "grad_norm": 2.273571202553131, + "learning_rate": 1.7328519855595669e-06, + "loss": 5.0074, + "step": 24 + }, + { + "epoch": 0.013572204125950055, + "grad_norm": 1.1877323869171315, + "learning_rate": 1.805054151624549e-06, + "loss": 4.4331, + "step": 25 + }, + { + "epoch": 0.014115092290988056, + "grad_norm": 1.9365501019385494, + "learning_rate": 1.8772563176895308e-06, + "loss": 5.0525, + "step": 26 + }, + { + "epoch": 0.014657980456026058, + "grad_norm": 1.7710485953105986, + "learning_rate": 1.949458483754513e-06, + "loss": 4.1346, + "step": 27 + }, + { + "epoch": 0.01520086862106406, + "grad_norm": 1.231848881292397, + "learning_rate": 2.0216606498194946e-06, + "loss": 4.2431, + "step": 28 + }, + { + "epoch": 0.015743756786102063, + "grad_norm": 1.6372838462475754, + "learning_rate": 2.0938628158844768e-06, + "loss": 5.4317, + "step": 29 + }, + { + "epoch": 0.016286644951140065, + "grad_norm": 1.7326568199523718, + "learning_rate": 2.1660649819494585e-06, + "loss": 4.4774, + "step": 30 + }, + { + "epoch": 0.016829533116178068, + "grad_norm": 2.1315641946923387, + "learning_rate": 2.2382671480144407e-06, + "loss": 4.1418, + "step": 31 + }, + { + "epoch": 0.01737242128121607, + "grad_norm": 1.916538366799816, + "learning_rate": 2.3104693140794225e-06, + "loss": 5.7926, + "step": 32 + }, + { + "epoch": 0.017915309446254073, + "grad_norm": 2.057531274363787, + "learning_rate": 2.3826714801444047e-06, + "loss": 4.2862, + "step": 33 + }, + { + "epoch": 0.018458197611292075, + "grad_norm": 2.6727877745045023, + "learning_rate": 2.4548736462093864e-06, + "loss": 5.1323, + "step": 34 + }, + { + "epoch": 0.019001085776330078, + "grad_norm": 1.5043525736330898, + "learning_rate": 2.527075812274368e-06, + "loss": 4.0634, + "step": 35 + }, + { + "epoch": 0.019543973941368076, + "grad_norm": 2.6324248758076862, + "learning_rate": 2.59927797833935e-06, + "loss": 5.4211, + "step": 36 + }, + { + "epoch": 0.02008686210640608, + "grad_norm": 1.6581905646911415, + "learning_rate": 2.6714801444043326e-06, + "loss": 4.9825, + "step": 37 + }, + { + "epoch": 0.02062975027144408, + "grad_norm": 1.6667970880898455, + "learning_rate": 2.7436823104693144e-06, + "loss": 4.3219, + "step": 38 + }, + { + "epoch": 0.021172638436482084, + "grad_norm": 1.7448150724822071, + "learning_rate": 2.815884476534296e-06, + "loss": 4.3912, + "step": 39 + }, + { + "epoch": 0.021715526601520086, + "grad_norm": 3.720503777726443, + "learning_rate": 2.8880866425992783e-06, + "loss": 6.0592, + "step": 40 + }, + { + "epoch": 0.02225841476655809, + "grad_norm": 1.6853001074010774, + "learning_rate": 2.96028880866426e-06, + "loss": 3.8395, + "step": 41 + }, + { + "epoch": 0.02280130293159609, + "grad_norm": 1.797029067943667, + "learning_rate": 3.032490974729242e-06, + "loss": 4.4651, + "step": 42 + }, + { + "epoch": 0.023344191096634093, + "grad_norm": 1.6298824221707349, + "learning_rate": 3.1046931407942245e-06, + "loss": 4.4507, + "step": 43 + }, + { + "epoch": 0.023887079261672096, + "grad_norm": 1.4793189209059368, + "learning_rate": 3.1768953068592062e-06, + "loss": 4.7849, + "step": 44 + }, + { + "epoch": 0.024429967426710098, + "grad_norm": 1.1741346570472608, + "learning_rate": 3.249097472924188e-06, + "loss": 3.5897, + "step": 45 + }, + { + "epoch": 0.0249728555917481, + "grad_norm": 2.1468867306169064, + "learning_rate": 3.3212996389891698e-06, + "loss": 3.8969, + "step": 46 + }, + { + "epoch": 0.025515743756786103, + "grad_norm": 1.657494000697021, + "learning_rate": 3.393501805054152e-06, + "loss": 4.3412, + "step": 47 + }, + { + "epoch": 0.026058631921824105, + "grad_norm": 1.7125864055710684, + "learning_rate": 3.4657039711191337e-06, + "loss": 3.8059, + "step": 48 + }, + { + "epoch": 0.026601520086862108, + "grad_norm": 2.9515347000925622, + "learning_rate": 3.5379061371841155e-06, + "loss": 4.1069, + "step": 49 + }, + { + "epoch": 0.02714440825190011, + "grad_norm": 3.0185188918098262, + "learning_rate": 3.610108303249098e-06, + "loss": 5.2646, + "step": 50 + }, + { + "epoch": 0.02768729641693811, + "grad_norm": 1.5546972252207274, + "learning_rate": 3.68231046931408e-06, + "loss": 3.7212, + "step": 51 + }, + { + "epoch": 0.02823018458197611, + "grad_norm": 2.6376587426282474, + "learning_rate": 3.7545126353790616e-06, + "loss": 4.2862, + "step": 52 + }, + { + "epoch": 0.028773072747014114, + "grad_norm": 2.053905428064986, + "learning_rate": 3.826714801444043e-06, + "loss": 3.8304, + "step": 53 + }, + { + "epoch": 0.029315960912052116, + "grad_norm": 2.9083183656849725, + "learning_rate": 3.898916967509026e-06, + "loss": 4.8223, + "step": 54 + }, + { + "epoch": 0.02985884907709012, + "grad_norm": 1.9919239162425852, + "learning_rate": 3.971119133574007e-06, + "loss": 3.9054, + "step": 55 + }, + { + "epoch": 0.03040173724212812, + "grad_norm": 2.4076122411223135, + "learning_rate": 4.043321299638989e-06, + "loss": 4.8568, + "step": 56 + }, + { + "epoch": 0.030944625407166124, + "grad_norm": 3.854434602830178, + "learning_rate": 4.115523465703971e-06, + "loss": 5.7771, + "step": 57 + }, + { + "epoch": 0.031487513572204126, + "grad_norm": 2.510303564095041, + "learning_rate": 4.1877256317689535e-06, + "loss": 5.5851, + "step": 58 + }, + { + "epoch": 0.03203040173724213, + "grad_norm": 2.8884448366576305, + "learning_rate": 4.259927797833936e-06, + "loss": 4.1774, + "step": 59 + }, + { + "epoch": 0.03257328990228013, + "grad_norm": 2.0893778739721, + "learning_rate": 4.332129963898917e-06, + "loss": 4.4679, + "step": 60 + }, + { + "epoch": 0.03311617806731813, + "grad_norm": 1.5816439039579229, + "learning_rate": 4.404332129963899e-06, + "loss": 4.1999, + "step": 61 + }, + { + "epoch": 0.033659066232356136, + "grad_norm": 1.9591574137365761, + "learning_rate": 4.4765342960288814e-06, + "loss": 4.1539, + "step": 62 + }, + { + "epoch": 0.03420195439739414, + "grad_norm": 3.1533909413995924, + "learning_rate": 4.548736462093864e-06, + "loss": 4.5842, + "step": 63 + }, + { + "epoch": 0.03474484256243214, + "grad_norm": 2.543715437360663, + "learning_rate": 4.620938628158845e-06, + "loss": 4.1473, + "step": 64 + }, + { + "epoch": 0.03528773072747014, + "grad_norm": 2.7608093188531955, + "learning_rate": 4.693140794223827e-06, + "loss": 4.9354, + "step": 65 + }, + { + "epoch": 0.035830618892508145, + "grad_norm": 4.192163162864289, + "learning_rate": 4.765342960288809e-06, + "loss": 5.4263, + "step": 66 + }, + { + "epoch": 0.03637350705754615, + "grad_norm": 2.809206875027472, + "learning_rate": 4.837545126353791e-06, + "loss": 4.4259, + "step": 67 + }, + { + "epoch": 0.03691639522258415, + "grad_norm": 3.9318073942863534, + "learning_rate": 4.909747292418773e-06, + "loss": 4.1483, + "step": 68 + }, + { + "epoch": 0.03745928338762215, + "grad_norm": 2.935065385226121, + "learning_rate": 4.981949458483755e-06, + "loss": 4.6854, + "step": 69 + }, + { + "epoch": 0.038002171552660155, + "grad_norm": 2.3687140876301305, + "learning_rate": 5.054151624548736e-06, + "loss": 4.5875, + "step": 70 + }, + { + "epoch": 0.03854505971769816, + "grad_norm": 2.987709377109457, + "learning_rate": 5.126353790613719e-06, + "loss": 5.5784, + "step": 71 + }, + { + "epoch": 0.03908794788273615, + "grad_norm": 2.618514747422557, + "learning_rate": 5.1985559566787e-06, + "loss": 3.9901, + "step": 72 + }, + { + "epoch": 0.039630836047774155, + "grad_norm": 3.729239515680902, + "learning_rate": 5.270758122743683e-06, + "loss": 4.6192, + "step": 73 + }, + { + "epoch": 0.04017372421281216, + "grad_norm": 2.488658936956188, + "learning_rate": 5.342960288808665e-06, + "loss": 3.9889, + "step": 74 + }, + { + "epoch": 0.04071661237785016, + "grad_norm": 2.6055640979236223, + "learning_rate": 5.415162454873647e-06, + "loss": 4.3166, + "step": 75 + }, + { + "epoch": 0.04125950054288816, + "grad_norm": 2.091905540809132, + "learning_rate": 5.487364620938629e-06, + "loss": 3.5848, + "step": 76 + }, + { + "epoch": 0.041802388707926165, + "grad_norm": 4.153140315460812, + "learning_rate": 5.559566787003611e-06, + "loss": 4.4889, + "step": 77 + }, + { + "epoch": 0.04234527687296417, + "grad_norm": 4.828033123535034, + "learning_rate": 5.631768953068592e-06, + "loss": 4.886, + "step": 78 + }, + { + "epoch": 0.04288816503800217, + "grad_norm": 3.106146715385438, + "learning_rate": 5.7039711191335744e-06, + "loss": 4.3698, + "step": 79 + }, + { + "epoch": 0.04343105320304017, + "grad_norm": 4.458698539182903, + "learning_rate": 5.776173285198557e-06, + "loss": 4.7018, + "step": 80 + }, + { + "epoch": 0.043973941368078175, + "grad_norm": 4.619902873416999, + "learning_rate": 5.848375451263538e-06, + "loss": 4.3449, + "step": 81 + }, + { + "epoch": 0.04451682953311618, + "grad_norm": 2.411841615517987, + "learning_rate": 5.92057761732852e-06, + "loss": 3.9952, + "step": 82 + }, + { + "epoch": 0.04505971769815418, + "grad_norm": 3.6594837742651705, + "learning_rate": 5.992779783393502e-06, + "loss": 4.1308, + "step": 83 + }, + { + "epoch": 0.04560260586319218, + "grad_norm": 2.3760888273966074, + "learning_rate": 6.064981949458484e-06, + "loss": 4.0315, + "step": 84 + }, + { + "epoch": 0.046145494028230184, + "grad_norm": 4.579655616004927, + "learning_rate": 6.137184115523466e-06, + "loss": 4.1051, + "step": 85 + }, + { + "epoch": 0.04668838219326819, + "grad_norm": 3.5909339610827846, + "learning_rate": 6.209386281588449e-06, + "loss": 4.3126, + "step": 86 + }, + { + "epoch": 0.04723127035830619, + "grad_norm": 3.1030422711991625, + "learning_rate": 6.28158844765343e-06, + "loss": 4.3038, + "step": 87 + }, + { + "epoch": 0.04777415852334419, + "grad_norm": 4.134211324790428, + "learning_rate": 6.3537906137184125e-06, + "loss": 4.906, + "step": 88 + }, + { + "epoch": 0.048317046688382194, + "grad_norm": 3.538946913288816, + "learning_rate": 6.425992779783395e-06, + "loss": 4.2182, + "step": 89 + }, + { + "epoch": 0.048859934853420196, + "grad_norm": 1.9828152543749473, + "learning_rate": 6.498194945848376e-06, + "loss": 3.8487, + "step": 90 + }, + { + "epoch": 0.0494028230184582, + "grad_norm": 2.127969936938475, + "learning_rate": 6.570397111913358e-06, + "loss": 3.9221, + "step": 91 + }, + { + "epoch": 0.0499457111834962, + "grad_norm": 2.740653026382279, + "learning_rate": 6.6425992779783395e-06, + "loss": 4.4661, + "step": 92 + }, + { + "epoch": 0.050488599348534204, + "grad_norm": 3.278496788216742, + "learning_rate": 6.714801444043322e-06, + "loss": 4.1995, + "step": 93 + }, + { + "epoch": 0.051031487513572206, + "grad_norm": 1.5455749978381657, + "learning_rate": 6.787003610108304e-06, + "loss": 3.3941, + "step": 94 + }, + { + "epoch": 0.05157437567861021, + "grad_norm": 2.6938438745410975, + "learning_rate": 6.859205776173285e-06, + "loss": 3.9863, + "step": 95 + }, + { + "epoch": 0.05211726384364821, + "grad_norm": 2.754193515919841, + "learning_rate": 6.9314079422382674e-06, + "loss": 4.1621, + "step": 96 + }, + { + "epoch": 0.05266015200868621, + "grad_norm": 1.787757508585929, + "learning_rate": 7.00361010830325e-06, + "loss": 3.728, + "step": 97 + }, + { + "epoch": 0.053203040173724216, + "grad_norm": 1.6756343581979376, + "learning_rate": 7.075812274368231e-06, + "loss": 3.4483, + "step": 98 + }, + { + "epoch": 0.05374592833876222, + "grad_norm": 3.0418610323064317, + "learning_rate": 7.148014440433214e-06, + "loss": 4.1131, + "step": 99 + }, + { + "epoch": 0.05428881650380022, + "grad_norm": 2.710846119482301, + "learning_rate": 7.220216606498196e-06, + "loss": 5.0449, + "step": 100 + }, + { + "epoch": 0.054831704668838216, + "grad_norm": 1.906382949957388, + "learning_rate": 7.2924187725631776e-06, + "loss": 3.3181, + "step": 101 + }, + { + "epoch": 0.05537459283387622, + "grad_norm": 2.4669878697837, + "learning_rate": 7.36462093862816e-06, + "loss": 3.7996, + "step": 102 + }, + { + "epoch": 0.05591748099891422, + "grad_norm": 2.3755842712374586, + "learning_rate": 7.436823104693142e-06, + "loss": 4.0321, + "step": 103 + }, + { + "epoch": 0.05646036916395222, + "grad_norm": 2.032793864149989, + "learning_rate": 7.509025270758123e-06, + "loss": 3.4121, + "step": 104 + }, + { + "epoch": 0.057003257328990226, + "grad_norm": 2.1561554352810686, + "learning_rate": 7.5812274368231055e-06, + "loss": 4.2585, + "step": 105 + }, + { + "epoch": 0.05754614549402823, + "grad_norm": 1.9627108654626784, + "learning_rate": 7.653429602888087e-06, + "loss": 4.3085, + "step": 106 + }, + { + "epoch": 0.05808903365906623, + "grad_norm": 1.8622753095278954, + "learning_rate": 7.72563176895307e-06, + "loss": 3.8459, + "step": 107 + }, + { + "epoch": 0.05863192182410423, + "grad_norm": 1.4438765470100654, + "learning_rate": 7.797833935018051e-06, + "loss": 3.4914, + "step": 108 + }, + { + "epoch": 0.059174809989142235, + "grad_norm": 1.6164085905580508, + "learning_rate": 7.870036101083033e-06, + "loss": 3.4958, + "step": 109 + }, + { + "epoch": 0.05971769815418024, + "grad_norm": 1.6075014519578008, + "learning_rate": 7.942238267148014e-06, + "loss": 3.202, + "step": 110 + }, + { + "epoch": 0.06026058631921824, + "grad_norm": 1.7630994896021268, + "learning_rate": 8.014440433212997e-06, + "loss": 3.0916, + "step": 111 + }, + { + "epoch": 0.06080347448425624, + "grad_norm": 1.7600666006428431, + "learning_rate": 8.086642599277978e-06, + "loss": 3.5791, + "step": 112 + }, + { + "epoch": 0.061346362649294245, + "grad_norm": 2.324243905399748, + "learning_rate": 8.158844765342961e-06, + "loss": 3.6247, + "step": 113 + }, + { + "epoch": 0.06188925081433225, + "grad_norm": 1.2715449173786617, + "learning_rate": 8.231046931407943e-06, + "loss": 3.2381, + "step": 114 + }, + { + "epoch": 0.06243213897937025, + "grad_norm": 1.9029014476619814, + "learning_rate": 8.303249097472926e-06, + "loss": 3.5269, + "step": 115 + }, + { + "epoch": 0.06297502714440825, + "grad_norm": 1.8782574049467182, + "learning_rate": 8.375451263537907e-06, + "loss": 2.8698, + "step": 116 + }, + { + "epoch": 0.06351791530944625, + "grad_norm": 1.8522383705853351, + "learning_rate": 8.447653429602888e-06, + "loss": 3.5003, + "step": 117 + }, + { + "epoch": 0.06406080347448426, + "grad_norm": 1.6480852413427443, + "learning_rate": 8.519855595667871e-06, + "loss": 3.5665, + "step": 118 + }, + { + "epoch": 0.06460369163952226, + "grad_norm": 2.1628448626122685, + "learning_rate": 8.592057761732853e-06, + "loss": 3.343, + "step": 119 + }, + { + "epoch": 0.06514657980456026, + "grad_norm": 1.568456113423694, + "learning_rate": 8.664259927797834e-06, + "loss": 2.474, + "step": 120 + }, + { + "epoch": 0.06568946796959826, + "grad_norm": 1.8054357065473743, + "learning_rate": 8.736462093862817e-06, + "loss": 2.9228, + "step": 121 + }, + { + "epoch": 0.06623235613463627, + "grad_norm": 1.9418450135728782, + "learning_rate": 8.808664259927798e-06, + "loss": 3.1995, + "step": 122 + }, + { + "epoch": 0.06677524429967427, + "grad_norm": 1.321851521267606, + "learning_rate": 8.88086642599278e-06, + "loss": 3.5515, + "step": 123 + }, + { + "epoch": 0.06731813246471227, + "grad_norm": 1.390198419578882, + "learning_rate": 8.953068592057763e-06, + "loss": 2.7741, + "step": 124 + }, + { + "epoch": 0.06786102062975027, + "grad_norm": 2.119370139523294, + "learning_rate": 9.025270758122744e-06, + "loss": 3.3427, + "step": 125 + }, + { + "epoch": 0.06840390879478828, + "grad_norm": 2.042189734142807, + "learning_rate": 9.097472924187727e-06, + "loss": 3.1414, + "step": 126 + }, + { + "epoch": 0.06894679695982628, + "grad_norm": 2.0399787373405123, + "learning_rate": 9.169675090252709e-06, + "loss": 3.0862, + "step": 127 + }, + { + "epoch": 0.06948968512486428, + "grad_norm": 2.0523727295223915, + "learning_rate": 9.24187725631769e-06, + "loss": 2.6461, + "step": 128 + }, + { + "epoch": 0.07003257328990228, + "grad_norm": 1.459190908283032, + "learning_rate": 9.314079422382673e-06, + "loss": 2.9907, + "step": 129 + }, + { + "epoch": 0.07057546145494029, + "grad_norm": 1.7403521896723462, + "learning_rate": 9.386281588447654e-06, + "loss": 3.0681, + "step": 130 + }, + { + "epoch": 0.07111834961997829, + "grad_norm": 2.2283604183714383, + "learning_rate": 9.458483754512636e-06, + "loss": 3.4619, + "step": 131 + }, + { + "epoch": 0.07166123778501629, + "grad_norm": 1.3291461832292721, + "learning_rate": 9.530685920577619e-06, + "loss": 3.096, + "step": 132 + }, + { + "epoch": 0.0722041259500543, + "grad_norm": 1.7839357457835638, + "learning_rate": 9.6028880866426e-06, + "loss": 2.7512, + "step": 133 + }, + { + "epoch": 0.0727470141150923, + "grad_norm": 1.7836465130823984, + "learning_rate": 9.675090252707581e-06, + "loss": 3.3118, + "step": 134 + }, + { + "epoch": 0.0732899022801303, + "grad_norm": 2.0407172747925135, + "learning_rate": 9.747292418772564e-06, + "loss": 2.1032, + "step": 135 + }, + { + "epoch": 0.0738327904451683, + "grad_norm": 2.335226908691354, + "learning_rate": 9.819494584837546e-06, + "loss": 2.9039, + "step": 136 + }, + { + "epoch": 0.0743756786102063, + "grad_norm": 2.567081845339778, + "learning_rate": 9.891696750902527e-06, + "loss": 3.1071, + "step": 137 + }, + { + "epoch": 0.0749185667752443, + "grad_norm": 1.9549184920601244, + "learning_rate": 9.96389891696751e-06, + "loss": 2.8253, + "step": 138 + }, + { + "epoch": 0.07546145494028231, + "grad_norm": 1.6389815727698025, + "learning_rate": 1.0036101083032491e-05, + "loss": 2.5701, + "step": 139 + }, + { + "epoch": 0.07600434310532031, + "grad_norm": 1.6558924195158018, + "learning_rate": 1.0108303249097473e-05, + "loss": 2.6428, + "step": 140 + }, + { + "epoch": 0.07654723127035831, + "grad_norm": 2.0257343427634087, + "learning_rate": 1.0180505415162456e-05, + "loss": 2.2828, + "step": 141 + }, + { + "epoch": 0.07709011943539631, + "grad_norm": 1.879991343961764, + "learning_rate": 1.0252707581227437e-05, + "loss": 3.1436, + "step": 142 + }, + { + "epoch": 0.07763300760043432, + "grad_norm": 1.8718236498893788, + "learning_rate": 1.032490974729242e-05, + "loss": 3.2434, + "step": 143 + }, + { + "epoch": 0.0781758957654723, + "grad_norm": 1.8252190287923398, + "learning_rate": 1.03971119133574e-05, + "loss": 2.4851, + "step": 144 + }, + { + "epoch": 0.07871878393051031, + "grad_norm": 2.6564413093838213, + "learning_rate": 1.0469314079422383e-05, + "loss": 2.7315, + "step": 145 + }, + { + "epoch": 0.07926167209554831, + "grad_norm": 2.054556112475997, + "learning_rate": 1.0541516245487366e-05, + "loss": 2.647, + "step": 146 + }, + { + "epoch": 0.07980456026058631, + "grad_norm": 1.072343898976271, + "learning_rate": 1.0613718411552347e-05, + "loss": 2.5868, + "step": 147 + }, + { + "epoch": 0.08034744842562432, + "grad_norm": 1.1443585178571194, + "learning_rate": 1.068592057761733e-05, + "loss": 2.3247, + "step": 148 + }, + { + "epoch": 0.08089033659066232, + "grad_norm": 2.0424670761496646, + "learning_rate": 1.0758122743682312e-05, + "loss": 1.989, + "step": 149 + }, + { + "epoch": 0.08143322475570032, + "grad_norm": 1.3715615162918033, + "learning_rate": 1.0830324909747295e-05, + "loss": 2.3989, + "step": 150 + }, + { + "epoch": 0.08197611292073832, + "grad_norm": 1.3740116291532023, + "learning_rate": 1.0902527075812274e-05, + "loss": 1.8726, + "step": 151 + }, + { + "epoch": 0.08251900108577633, + "grad_norm": 2.5824759155435655, + "learning_rate": 1.0974729241877257e-05, + "loss": 2.3019, + "step": 152 + }, + { + "epoch": 0.08306188925081433, + "grad_norm": 1.5584514490971877, + "learning_rate": 1.1046931407942239e-05, + "loss": 2.4209, + "step": 153 + }, + { + "epoch": 0.08360477741585233, + "grad_norm": 1.6184456176727229, + "learning_rate": 1.1119133574007222e-05, + "loss": 2.2377, + "step": 154 + }, + { + "epoch": 0.08414766558089033, + "grad_norm": 2.6665151568462586, + "learning_rate": 1.1191335740072201e-05, + "loss": 2.6381, + "step": 155 + }, + { + "epoch": 0.08469055374592833, + "grad_norm": 1.442929928712006, + "learning_rate": 1.1263537906137184e-05, + "loss": 2.3405, + "step": 156 + }, + { + "epoch": 0.08523344191096634, + "grad_norm": 1.906547769868643, + "learning_rate": 1.1335740072202166e-05, + "loss": 2.4391, + "step": 157 + }, + { + "epoch": 0.08577633007600434, + "grad_norm": 1.2153727696829317, + "learning_rate": 1.1407942238267149e-05, + "loss": 1.8401, + "step": 158 + }, + { + "epoch": 0.08631921824104234, + "grad_norm": 1.4810729195972283, + "learning_rate": 1.1480144404332132e-05, + "loss": 2.7248, + "step": 159 + }, + { + "epoch": 0.08686210640608034, + "grad_norm": 2.9401118999625058, + "learning_rate": 1.1552346570397113e-05, + "loss": 2.6582, + "step": 160 + }, + { + "epoch": 0.08740499457111835, + "grad_norm": 1.145106458573816, + "learning_rate": 1.1624548736462096e-05, + "loss": 2.6648, + "step": 161 + }, + { + "epoch": 0.08794788273615635, + "grad_norm": 1.9228128025198137, + "learning_rate": 1.1696750902527076e-05, + "loss": 2.3702, + "step": 162 + }, + { + "epoch": 0.08849077090119435, + "grad_norm": 1.318978856783607, + "learning_rate": 1.1768953068592059e-05, + "loss": 1.9388, + "step": 163 + }, + { + "epoch": 0.08903365906623235, + "grad_norm": 1.58104775083856, + "learning_rate": 1.184115523465704e-05, + "loss": 2.0774, + "step": 164 + }, + { + "epoch": 0.08957654723127036, + "grad_norm": 1.8206646323260607, + "learning_rate": 1.1913357400722023e-05, + "loss": 2.5221, + "step": 165 + }, + { + "epoch": 0.09011943539630836, + "grad_norm": 1.0815960351491627, + "learning_rate": 1.1985559566787005e-05, + "loss": 2.128, + "step": 166 + }, + { + "epoch": 0.09066232356134636, + "grad_norm": 1.3341486432977188, + "learning_rate": 1.2057761732851988e-05, + "loss": 1.9418, + "step": 167 + }, + { + "epoch": 0.09120521172638436, + "grad_norm": 1.3772486036170464, + "learning_rate": 1.2129963898916967e-05, + "loss": 1.84, + "step": 168 + }, + { + "epoch": 0.09174809989142237, + "grad_norm": 1.42874152479176, + "learning_rate": 1.220216606498195e-05, + "loss": 2.2172, + "step": 169 + }, + { + "epoch": 0.09229098805646037, + "grad_norm": 1.9691394708584433, + "learning_rate": 1.2274368231046932e-05, + "loss": 2.9247, + "step": 170 + }, + { + "epoch": 0.09283387622149837, + "grad_norm": 1.4773974635910447, + "learning_rate": 1.2346570397111915e-05, + "loss": 2.1028, + "step": 171 + }, + { + "epoch": 0.09337676438653637, + "grad_norm": 1.5937344901476087, + "learning_rate": 1.2418772563176898e-05, + "loss": 2.4073, + "step": 172 + }, + { + "epoch": 0.09391965255157438, + "grad_norm": 1.630783959399718, + "learning_rate": 1.2490974729241878e-05, + "loss": 2.6978, + "step": 173 + }, + { + "epoch": 0.09446254071661238, + "grad_norm": 1.44736447594164, + "learning_rate": 1.256317689530686e-05, + "loss": 1.8851, + "step": 174 + }, + { + "epoch": 0.09500542888165038, + "grad_norm": 1.3909459856510267, + "learning_rate": 1.2635379061371842e-05, + "loss": 1.576, + "step": 175 + }, + { + "epoch": 0.09554831704668838, + "grad_norm": 1.7838393677575515, + "learning_rate": 1.2707581227436825e-05, + "loss": 2.3895, + "step": 176 + }, + { + "epoch": 0.09609120521172639, + "grad_norm": 1.1294127385799255, + "learning_rate": 1.2779783393501806e-05, + "loss": 2.4855, + "step": 177 + }, + { + "epoch": 0.09663409337676439, + "grad_norm": 1.259662421149648, + "learning_rate": 1.285198555956679e-05, + "loss": 2.5469, + "step": 178 + }, + { + "epoch": 0.09717698154180239, + "grad_norm": 1.4621367635995284, + "learning_rate": 1.2924187725631769e-05, + "loss": 1.9015, + "step": 179 + }, + { + "epoch": 0.09771986970684039, + "grad_norm": 1.6960126954903645, + "learning_rate": 1.2996389891696752e-05, + "loss": 2.9545, + "step": 180 + }, + { + "epoch": 0.0982627578718784, + "grad_norm": 1.5419022503429414, + "learning_rate": 1.3068592057761733e-05, + "loss": 1.9912, + "step": 181 + }, + { + "epoch": 0.0988056460369164, + "grad_norm": 1.4073422496120207, + "learning_rate": 1.3140794223826716e-05, + "loss": 2.3353, + "step": 182 + }, + { + "epoch": 0.0993485342019544, + "grad_norm": 1.2881659254471862, + "learning_rate": 1.3212996389891696e-05, + "loss": 2.7784, + "step": 183 + }, + { + "epoch": 0.0998914223669924, + "grad_norm": 1.361969663712106, + "learning_rate": 1.3285198555956679e-05, + "loss": 2.523, + "step": 184 + }, + { + "epoch": 0.1004343105320304, + "grad_norm": 2.5841731119141316, + "learning_rate": 1.3357400722021662e-05, + "loss": 2.43, + "step": 185 + }, + { + "epoch": 0.10097719869706841, + "grad_norm": 1.3203537847943392, + "learning_rate": 1.3429602888086643e-05, + "loss": 1.9404, + "step": 186 + }, + { + "epoch": 0.10152008686210641, + "grad_norm": 1.393097901493087, + "learning_rate": 1.3501805054151626e-05, + "loss": 2.6202, + "step": 187 + }, + { + "epoch": 0.10206297502714441, + "grad_norm": 0.9414377380652101, + "learning_rate": 1.3574007220216608e-05, + "loss": 2.0879, + "step": 188 + }, + { + "epoch": 0.10260586319218241, + "grad_norm": 1.6547581806701488, + "learning_rate": 1.3646209386281591e-05, + "loss": 2.8912, + "step": 189 + }, + { + "epoch": 0.10314875135722042, + "grad_norm": 1.1236261864015666, + "learning_rate": 1.371841155234657e-05, + "loss": 2.1802, + "step": 190 + }, + { + "epoch": 0.10369163952225842, + "grad_norm": 1.1766384878462062, + "learning_rate": 1.3790613718411554e-05, + "loss": 1.9127, + "step": 191 + }, + { + "epoch": 0.10423452768729642, + "grad_norm": 2.0453724866488656, + "learning_rate": 1.3862815884476535e-05, + "loss": 1.8522, + "step": 192 + }, + { + "epoch": 0.10477741585233442, + "grad_norm": 1.550964370734908, + "learning_rate": 1.3935018050541518e-05, + "loss": 2.5852, + "step": 193 + }, + { + "epoch": 0.10532030401737243, + "grad_norm": 1.303227180334529, + "learning_rate": 1.40072202166065e-05, + "loss": 2.407, + "step": 194 + }, + { + "epoch": 0.10586319218241043, + "grad_norm": 1.2009193115283474, + "learning_rate": 1.4079422382671482e-05, + "loss": 1.8978, + "step": 195 + }, + { + "epoch": 0.10640608034744843, + "grad_norm": 1.353648963417667, + "learning_rate": 1.4151624548736462e-05, + "loss": 2.283, + "step": 196 + }, + { + "epoch": 0.10694896851248643, + "grad_norm": 1.266154745219424, + "learning_rate": 1.4223826714801445e-05, + "loss": 2.0679, + "step": 197 + }, + { + "epoch": 0.10749185667752444, + "grad_norm": 4.186447043886807, + "learning_rate": 1.4296028880866428e-05, + "loss": 1.9175, + "step": 198 + }, + { + "epoch": 0.10803474484256244, + "grad_norm": 1.147065238667786, + "learning_rate": 1.436823104693141e-05, + "loss": 2.2748, + "step": 199 + }, + { + "epoch": 0.10857763300760044, + "grad_norm": 1.2760699256137895, + "learning_rate": 1.4440433212996392e-05, + "loss": 2.3031, + "step": 200 + }, + { + "epoch": 0.10912052117263844, + "grad_norm": 0.9452152631580967, + "learning_rate": 1.4512635379061372e-05, + "loss": 1.4833, + "step": 201 + }, + { + "epoch": 0.10966340933767643, + "grad_norm": 1.3980490034032074, + "learning_rate": 1.4584837545126355e-05, + "loss": 1.5865, + "step": 202 + }, + { + "epoch": 0.11020629750271443, + "grad_norm": 1.6440838011007093, + "learning_rate": 1.4657039711191336e-05, + "loss": 1.632, + "step": 203 + }, + { + "epoch": 0.11074918566775244, + "grad_norm": 1.3969764794458068, + "learning_rate": 1.472924187725632e-05, + "loss": 1.4301, + "step": 204 + }, + { + "epoch": 0.11129207383279044, + "grad_norm": 1.223663090997016, + "learning_rate": 1.48014440433213e-05, + "loss": 2.3993, + "step": 205 + }, + { + "epoch": 0.11183496199782844, + "grad_norm": 1.183881027275754, + "learning_rate": 1.4873646209386284e-05, + "loss": 1.4987, + "step": 206 + }, + { + "epoch": 0.11237785016286644, + "grad_norm": 1.3788483357578956, + "learning_rate": 1.4945848375451264e-05, + "loss": 2.7742, + "step": 207 + }, + { + "epoch": 0.11292073832790445, + "grad_norm": 1.1260622559595288, + "learning_rate": 1.5018050541516247e-05, + "loss": 1.755, + "step": 208 + }, + { + "epoch": 0.11346362649294245, + "grad_norm": 1.8439396062381286, + "learning_rate": 1.5090252707581228e-05, + "loss": 2.9875, + "step": 209 + }, + { + "epoch": 0.11400651465798045, + "grad_norm": 1.2789617727573501, + "learning_rate": 1.5162454873646211e-05, + "loss": 2.0119, + "step": 210 + }, + { + "epoch": 0.11454940282301845, + "grad_norm": 1.3385913561767113, + "learning_rate": 1.5234657039711192e-05, + "loss": 1.7693, + "step": 211 + }, + { + "epoch": 0.11509229098805646, + "grad_norm": 1.049162702021749, + "learning_rate": 1.5306859205776174e-05, + "loss": 1.7541, + "step": 212 + }, + { + "epoch": 0.11563517915309446, + "grad_norm": 1.0797407129441892, + "learning_rate": 1.537906137184116e-05, + "loss": 2.0427, + "step": 213 + }, + { + "epoch": 0.11617806731813246, + "grad_norm": 1.3661735952693994, + "learning_rate": 1.545126353790614e-05, + "loss": 2.722, + "step": 214 + }, + { + "epoch": 0.11672095548317046, + "grad_norm": 1.7084933140661187, + "learning_rate": 1.552346570397112e-05, + "loss": 2.2125, + "step": 215 + }, + { + "epoch": 0.11726384364820847, + "grad_norm": 1.3814428141700028, + "learning_rate": 1.5595667870036102e-05, + "loss": 2.448, + "step": 216 + }, + { + "epoch": 0.11780673181324647, + "grad_norm": 1.298309196615407, + "learning_rate": 1.5667870036101084e-05, + "loss": 2.2706, + "step": 217 + }, + { + "epoch": 0.11834961997828447, + "grad_norm": 1.4553148411944858, + "learning_rate": 1.5740072202166065e-05, + "loss": 2.0087, + "step": 218 + }, + { + "epoch": 0.11889250814332247, + "grad_norm": 1.5083115470870359, + "learning_rate": 1.581227436823105e-05, + "loss": 1.7179, + "step": 219 + }, + { + "epoch": 0.11943539630836048, + "grad_norm": 1.2335179878839475, + "learning_rate": 1.5884476534296028e-05, + "loss": 2.5337, + "step": 220 + }, + { + "epoch": 0.11997828447339848, + "grad_norm": 1.1729282902419653, + "learning_rate": 1.5956678700361013e-05, + "loss": 1.9604, + "step": 221 + }, + { + "epoch": 0.12052117263843648, + "grad_norm": 1.5700347334422509, + "learning_rate": 1.6028880866425994e-05, + "loss": 2.3655, + "step": 222 + }, + { + "epoch": 0.12106406080347448, + "grad_norm": 1.435764746128967, + "learning_rate": 1.6101083032490975e-05, + "loss": 2.109, + "step": 223 + }, + { + "epoch": 0.12160694896851248, + "grad_norm": 1.0866216936301165, + "learning_rate": 1.6173285198555957e-05, + "loss": 2.3334, + "step": 224 + }, + { + "epoch": 0.12214983713355049, + "grad_norm": 1.8060789361277199, + "learning_rate": 1.624548736462094e-05, + "loss": 1.6605, + "step": 225 + }, + { + "epoch": 0.12269272529858849, + "grad_norm": 1.539101200174871, + "learning_rate": 1.6317689530685923e-05, + "loss": 2.2597, + "step": 226 + }, + { + "epoch": 0.12323561346362649, + "grad_norm": 1.7392984350243024, + "learning_rate": 1.6389891696750904e-05, + "loss": 2.5979, + "step": 227 + }, + { + "epoch": 0.1237785016286645, + "grad_norm": 1.6792982939949113, + "learning_rate": 1.6462093862815885e-05, + "loss": 2.0626, + "step": 228 + }, + { + "epoch": 0.1243213897937025, + "grad_norm": 2.067856070276011, + "learning_rate": 1.6534296028880867e-05, + "loss": 2.056, + "step": 229 + }, + { + "epoch": 0.1248642779587405, + "grad_norm": 1.2773224600362139, + "learning_rate": 1.660649819494585e-05, + "loss": 1.5852, + "step": 230 + }, + { + "epoch": 0.1254071661237785, + "grad_norm": 1.381245750855955, + "learning_rate": 1.6678700361010833e-05, + "loss": 2.0675, + "step": 231 + }, + { + "epoch": 0.1259500542888165, + "grad_norm": 1.4215755121774605, + "learning_rate": 1.6750902527075814e-05, + "loss": 2.1002, + "step": 232 + }, + { + "epoch": 0.1264929424538545, + "grad_norm": 1.4581673093175722, + "learning_rate": 1.6823104693140795e-05, + "loss": 1.9279, + "step": 233 + }, + { + "epoch": 0.1270358306188925, + "grad_norm": 1.2656102558648035, + "learning_rate": 1.6895306859205777e-05, + "loss": 1.4913, + "step": 234 + }, + { + "epoch": 0.1275787187839305, + "grad_norm": 1.3985246778048182, + "learning_rate": 1.6967509025270758e-05, + "loss": 2.1629, + "step": 235 + }, + { + "epoch": 0.1281216069489685, + "grad_norm": 1.757194392177075, + "learning_rate": 1.7039711191335743e-05, + "loss": 2.5688, + "step": 236 + }, + { + "epoch": 0.12866449511400652, + "grad_norm": 1.3766112824147358, + "learning_rate": 1.711191335740072e-05, + "loss": 2.3344, + "step": 237 + }, + { + "epoch": 0.12920738327904452, + "grad_norm": 1.2985043105700416, + "learning_rate": 1.7184115523465706e-05, + "loss": 2.0626, + "step": 238 + }, + { + "epoch": 0.12975027144408252, + "grad_norm": 1.715796674825951, + "learning_rate": 1.7256317689530687e-05, + "loss": 2.1559, + "step": 239 + }, + { + "epoch": 0.13029315960912052, + "grad_norm": 1.6374209306607368, + "learning_rate": 1.7328519855595668e-05, + "loss": 2.0801, + "step": 240 + }, + { + "epoch": 0.13083604777415853, + "grad_norm": 1.244682506367769, + "learning_rate": 1.7400722021660653e-05, + "loss": 1.6125, + "step": 241 + }, + { + "epoch": 0.13137893593919653, + "grad_norm": 1.3001838781366861, + "learning_rate": 1.7472924187725634e-05, + "loss": 2.0934, + "step": 242 + }, + { + "epoch": 0.13192182410423453, + "grad_norm": 1.1600879910417528, + "learning_rate": 1.7545126353790616e-05, + "loss": 1.7862, + "step": 243 + }, + { + "epoch": 0.13246471226927253, + "grad_norm": 1.3503312968411385, + "learning_rate": 1.7617328519855597e-05, + "loss": 2.1636, + "step": 244 + }, + { + "epoch": 0.13300760043431054, + "grad_norm": 2.1299761628528167, + "learning_rate": 1.768953068592058e-05, + "loss": 2.4826, + "step": 245 + }, + { + "epoch": 0.13355048859934854, + "grad_norm": 1.3232052544377066, + "learning_rate": 1.776173285198556e-05, + "loss": 2.0497, + "step": 246 + }, + { + "epoch": 0.13409337676438654, + "grad_norm": 1.5079892467642568, + "learning_rate": 1.7833935018050544e-05, + "loss": 2.4552, + "step": 247 + }, + { + "epoch": 0.13463626492942454, + "grad_norm": 1.716355318914646, + "learning_rate": 1.7906137184115526e-05, + "loss": 1.5309, + "step": 248 + }, + { + "epoch": 0.13517915309446255, + "grad_norm": 2.1523878955969193, + "learning_rate": 1.7978339350180507e-05, + "loss": 2.7403, + "step": 249 + }, + { + "epoch": 0.13572204125950055, + "grad_norm": 1.8169301927997485, + "learning_rate": 1.805054151624549e-05, + "loss": 2.3324, + "step": 250 + }, + { + "epoch": 0.13626492942453855, + "grad_norm": 1.4464924096148426, + "learning_rate": 1.812274368231047e-05, + "loss": 1.8399, + "step": 251 + }, + { + "epoch": 0.13680781758957655, + "grad_norm": 1.9818298103877654, + "learning_rate": 1.8194945848375454e-05, + "loss": 2.4237, + "step": 252 + }, + { + "epoch": 0.13735070575461455, + "grad_norm": 1.6371003739671923, + "learning_rate": 1.8267148014440436e-05, + "loss": 2.0477, + "step": 253 + }, + { + "epoch": 0.13789359391965256, + "grad_norm": 1.2944832148952543, + "learning_rate": 1.8339350180505417e-05, + "loss": 2.2269, + "step": 254 + }, + { + "epoch": 0.13843648208469056, + "grad_norm": 1.4284289887933213, + "learning_rate": 1.84115523465704e-05, + "loss": 2.2465, + "step": 255 + }, + { + "epoch": 0.13897937024972856, + "grad_norm": 1.4633226693151502, + "learning_rate": 1.848375451263538e-05, + "loss": 2.1582, + "step": 256 + }, + { + "epoch": 0.13952225841476656, + "grad_norm": 1.6169546473154692, + "learning_rate": 1.855595667870036e-05, + "loss": 1.9526, + "step": 257 + }, + { + "epoch": 0.14006514657980457, + "grad_norm": 1.47082651679944, + "learning_rate": 1.8628158844765346e-05, + "loss": 2.0314, + "step": 258 + }, + { + "epoch": 0.14060803474484257, + "grad_norm": 1.2279964838302115, + "learning_rate": 1.8700361010830327e-05, + "loss": 1.4345, + "step": 259 + }, + { + "epoch": 0.14115092290988057, + "grad_norm": 1.5727648286246085, + "learning_rate": 1.877256317689531e-05, + "loss": 2.7371, + "step": 260 + }, + { + "epoch": 0.14169381107491857, + "grad_norm": 1.5536143039045338, + "learning_rate": 1.884476534296029e-05, + "loss": 1.6691, + "step": 261 + }, + { + "epoch": 0.14223669923995658, + "grad_norm": 1.504953118413033, + "learning_rate": 1.891696750902527e-05, + "loss": 1.8422, + "step": 262 + }, + { + "epoch": 0.14277958740499458, + "grad_norm": 1.6092868309857633, + "learning_rate": 1.8989169675090253e-05, + "loss": 1.6446, + "step": 263 + }, + { + "epoch": 0.14332247557003258, + "grad_norm": 1.434484546738067, + "learning_rate": 1.9061371841155237e-05, + "loss": 1.5945, + "step": 264 + }, + { + "epoch": 0.14386536373507058, + "grad_norm": 1.3472204852366787, + "learning_rate": 1.913357400722022e-05, + "loss": 1.4511, + "step": 265 + }, + { + "epoch": 0.1444082519001086, + "grad_norm": 2.493842366649246, + "learning_rate": 1.92057761732852e-05, + "loss": 2.1144, + "step": 266 + }, + { + "epoch": 0.1449511400651466, + "grad_norm": 1.3138428233046642, + "learning_rate": 1.927797833935018e-05, + "loss": 1.6592, + "step": 267 + }, + { + "epoch": 0.1454940282301846, + "grad_norm": 1.7201458829291465, + "learning_rate": 1.9350180505415163e-05, + "loss": 1.6895, + "step": 268 + }, + { + "epoch": 0.1460369163952226, + "grad_norm": 1.9777013455620547, + "learning_rate": 1.9422382671480147e-05, + "loss": 2.2628, + "step": 269 + }, + { + "epoch": 0.1465798045602606, + "grad_norm": 1.8559862870685497, + "learning_rate": 1.949458483754513e-05, + "loss": 1.3374, + "step": 270 + }, + { + "epoch": 0.1471226927252986, + "grad_norm": 1.919404449298866, + "learning_rate": 1.956678700361011e-05, + "loss": 2.3805, + "step": 271 + }, + { + "epoch": 0.1476655808903366, + "grad_norm": 1.4760010543412088, + "learning_rate": 1.963898916967509e-05, + "loss": 1.9791, + "step": 272 + }, + { + "epoch": 0.1482084690553746, + "grad_norm": 1.4491842831658432, + "learning_rate": 1.9711191335740073e-05, + "loss": 1.8119, + "step": 273 + }, + { + "epoch": 0.1487513572204126, + "grad_norm": 1.4098796806726177, + "learning_rate": 1.9783393501805054e-05, + "loss": 1.7398, + "step": 274 + }, + { + "epoch": 0.1492942453854506, + "grad_norm": 1.5009369942229867, + "learning_rate": 1.985559566787004e-05, + "loss": 1.5194, + "step": 275 + }, + { + "epoch": 0.1498371335504886, + "grad_norm": 1.5073462671342261, + "learning_rate": 1.992779783393502e-05, + "loss": 1.8125, + "step": 276 + }, + { + "epoch": 0.1503800217155266, + "grad_norm": 1.9315604890805302, + "learning_rate": 2e-05, + "loss": 1.5429, + "step": 277 + }, + { + "epoch": 0.15092290988056462, + "grad_norm": 1.5059106777628657, + "learning_rate": 1.999999938159203e-05, + "loss": 1.6402, + "step": 278 + }, + { + "epoch": 0.15146579804560262, + "grad_norm": 1.7255630320192266, + "learning_rate": 1.9999997526368205e-05, + "loss": 2.2806, + "step": 279 + }, + { + "epoch": 0.15200868621064062, + "grad_norm": 1.7985407961594548, + "learning_rate": 1.999999443432874e-05, + "loss": 2.5213, + "step": 280 + }, + { + "epoch": 0.15255157437567862, + "grad_norm": 1.801471230172198, + "learning_rate": 1.999999010547403e-05, + "loss": 1.8362, + "step": 281 + }, + { + "epoch": 0.15309446254071662, + "grad_norm": 3.236881896883975, + "learning_rate": 1.999998453980461e-05, + "loss": 2.7148, + "step": 282 + }, + { + "epoch": 0.15363735070575463, + "grad_norm": 1.7343786362856708, + "learning_rate": 1.9999977737321156e-05, + "loss": 1.8195, + "step": 283 + }, + { + "epoch": 0.15418023887079263, + "grad_norm": 1.5006942215232695, + "learning_rate": 1.999996969802452e-05, + "loss": 1.8744, + "step": 284 + }, + { + "epoch": 0.15472312703583063, + "grad_norm": 4.171967924853298, + "learning_rate": 1.99999604219157e-05, + "loss": 2.4329, + "step": 285 + }, + { + "epoch": 0.15526601520086863, + "grad_norm": 1.7546486556973808, + "learning_rate": 1.9999949908995832e-05, + "loss": 1.3931, + "step": 286 + }, + { + "epoch": 0.15580890336590664, + "grad_norm": 2.0223702721102375, + "learning_rate": 1.999993815926622e-05, + "loss": 1.6704, + "step": 287 + }, + { + "epoch": 0.1563517915309446, + "grad_norm": 1.8736762651082781, + "learning_rate": 1.9999925172728324e-05, + "loss": 1.9895, + "step": 288 + }, + { + "epoch": 0.15689467969598261, + "grad_norm": 1.8878827291463698, + "learning_rate": 1.9999910949383742e-05, + "loss": 2.4939, + "step": 289 + }, + { + "epoch": 0.15743756786102062, + "grad_norm": 1.7277080078786893, + "learning_rate": 1.9999895489234234e-05, + "loss": 1.9265, + "step": 290 + }, + { + "epoch": 0.15798045602605862, + "grad_norm": 1.7402172969699101, + "learning_rate": 1.999987879228172e-05, + "loss": 1.8101, + "step": 291 + }, + { + "epoch": 0.15852334419109662, + "grad_norm": 2.077278713661093, + "learning_rate": 1.9999860858528257e-05, + "loss": 2.1763, + "step": 292 + }, + { + "epoch": 0.15906623235613462, + "grad_norm": 2.247624069618552, + "learning_rate": 1.9999841687976067e-05, + "loss": 1.9727, + "step": 293 + }, + { + "epoch": 0.15960912052117263, + "grad_norm": 1.6712153667666279, + "learning_rate": 1.999982128062752e-05, + "loss": 1.9921, + "step": 294 + }, + { + "epoch": 0.16015200868621063, + "grad_norm": 1.7126188149003763, + "learning_rate": 1.999979963648514e-05, + "loss": 1.9375, + "step": 295 + }, + { + "epoch": 0.16069489685124863, + "grad_norm": 1.949228534239786, + "learning_rate": 1.99997767555516e-05, + "loss": 2.1628, + "step": 296 + }, + { + "epoch": 0.16123778501628663, + "grad_norm": 1.8253305155157105, + "learning_rate": 1.9999752637829734e-05, + "loss": 1.8293, + "step": 297 + }, + { + "epoch": 0.16178067318132464, + "grad_norm": 1.5881557560095252, + "learning_rate": 1.9999727283322524e-05, + "loss": 1.3688, + "step": 298 + }, + { + "epoch": 0.16232356134636264, + "grad_norm": 2.4008202914116503, + "learning_rate": 1.9999700692033112e-05, + "loss": 1.5919, + "step": 299 + }, + { + "epoch": 0.16286644951140064, + "grad_norm": 1.610839420604411, + "learning_rate": 1.9999672863964778e-05, + "loss": 1.8724, + "step": 300 + }, + { + "epoch": 0.16340933767643864, + "grad_norm": 1.4663439746230646, + "learning_rate": 1.9999643799120964e-05, + "loss": 1.4628, + "step": 301 + }, + { + "epoch": 0.16395222584147665, + "grad_norm": 1.7051101810518976, + "learning_rate": 1.9999613497505272e-05, + "loss": 1.2916, + "step": 302 + }, + { + "epoch": 0.16449511400651465, + "grad_norm": 1.8967399031374104, + "learning_rate": 1.9999581959121443e-05, + "loss": 1.7559, + "step": 303 + }, + { + "epoch": 0.16503800217155265, + "grad_norm": 1.6206520761555416, + "learning_rate": 1.9999549183973382e-05, + "loss": 1.9196, + "step": 304 + }, + { + "epoch": 0.16558089033659065, + "grad_norm": 2.3545652344963544, + "learning_rate": 1.999951517206514e-05, + "loss": 1.2731, + "step": 305 + }, + { + "epoch": 0.16612377850162866, + "grad_norm": 1.7952736257095967, + "learning_rate": 1.9999479923400926e-05, + "loss": 2.0502, + "step": 306 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 1.7199420678903319, + "learning_rate": 1.99994434379851e-05, + "loss": 1.7059, + "step": 307 + }, + { + "epoch": 0.16720955483170466, + "grad_norm": 2.3318553047346, + "learning_rate": 1.9999405715822167e-05, + "loss": 1.4667, + "step": 308 + }, + { + "epoch": 0.16775244299674266, + "grad_norm": 1.618961362672766, + "learning_rate": 1.9999366756916804e-05, + "loss": 1.1866, + "step": 309 + }, + { + "epoch": 0.16829533116178066, + "grad_norm": 2.0805239071290114, + "learning_rate": 1.999932656127382e-05, + "loss": 1.7601, + "step": 310 + }, + { + "epoch": 0.16883821932681867, + "grad_norm": 1.782022796323131, + "learning_rate": 1.9999285128898193e-05, + "loss": 1.6207, + "step": 311 + }, + { + "epoch": 0.16938110749185667, + "grad_norm": 2.2438870289045028, + "learning_rate": 1.9999242459795045e-05, + "loss": 2.1329, + "step": 312 + }, + { + "epoch": 0.16992399565689467, + "grad_norm": 2.2550840823235587, + "learning_rate": 1.9999198553969652e-05, + "loss": 1.6055, + "step": 313 + }, + { + "epoch": 0.17046688382193267, + "grad_norm": 1.8012459587153988, + "learning_rate": 1.9999153411427445e-05, + "loss": 2.018, + "step": 314 + }, + { + "epoch": 0.17100977198697068, + "grad_norm": 1.8498840013582993, + "learning_rate": 1.9999107032174007e-05, + "loss": 1.8113, + "step": 315 + }, + { + "epoch": 0.17155266015200868, + "grad_norm": 2.2799557163953965, + "learning_rate": 1.9999059416215078e-05, + "loss": 1.9494, + "step": 316 + }, + { + "epoch": 0.17209554831704668, + "grad_norm": 2.383891228288174, + "learning_rate": 1.999901056355654e-05, + "loss": 1.8323, + "step": 317 + }, + { + "epoch": 0.17263843648208468, + "grad_norm": 1.9081077971876759, + "learning_rate": 1.9998960474204443e-05, + "loss": 1.5476, + "step": 318 + }, + { + "epoch": 0.1731813246471227, + "grad_norm": 1.9010783437994137, + "learning_rate": 1.999890914816498e-05, + "loss": 1.4602, + "step": 319 + }, + { + "epoch": 0.1737242128121607, + "grad_norm": 2.295573334252978, + "learning_rate": 1.9998856585444493e-05, + "loss": 1.0965, + "step": 320 + }, + { + "epoch": 0.1742671009771987, + "grad_norm": 2.1014332124066954, + "learning_rate": 1.999880278604949e-05, + "loss": 1.3726, + "step": 321 + }, + { + "epoch": 0.1748099891422367, + "grad_norm": 2.0418722317428206, + "learning_rate": 1.9998747749986625e-05, + "loss": 1.8875, + "step": 322 + }, + { + "epoch": 0.1753528773072747, + "grad_norm": 1.768897707524193, + "learning_rate": 1.99986914772627e-05, + "loss": 1.8841, + "step": 323 + }, + { + "epoch": 0.1758957654723127, + "grad_norm": 2.603335158781435, + "learning_rate": 1.9998633967884676e-05, + "loss": 1.4262, + "step": 324 + }, + { + "epoch": 0.1764386536373507, + "grad_norm": 2.1485447383197607, + "learning_rate": 1.999857522185967e-05, + "loss": 1.9283, + "step": 325 + }, + { + "epoch": 0.1769815418023887, + "grad_norm": 1.8398472624384716, + "learning_rate": 1.9998515239194945e-05, + "loss": 1.9089, + "step": 326 + }, + { + "epoch": 0.1775244299674267, + "grad_norm": 2.6988014220525374, + "learning_rate": 1.9998454019897918e-05, + "loss": 1.3636, + "step": 327 + }, + { + "epoch": 0.1780673181324647, + "grad_norm": 2.182794188412918, + "learning_rate": 1.9998391563976166e-05, + "loss": 1.7041, + "step": 328 + }, + { + "epoch": 0.1786102062975027, + "grad_norm": 1.970858895313084, + "learning_rate": 1.9998327871437405e-05, + "loss": 1.4917, + "step": 329 + }, + { + "epoch": 0.1791530944625407, + "grad_norm": 1.829655959511745, + "learning_rate": 1.9998262942289524e-05, + "loss": 1.9694, + "step": 330 + }, + { + "epoch": 0.17969598262757872, + "grad_norm": 1.9990017706312653, + "learning_rate": 1.9998196776540545e-05, + "loss": 2.067, + "step": 331 + }, + { + "epoch": 0.18023887079261672, + "grad_norm": 2.2829395740777234, + "learning_rate": 1.9998129374198655e-05, + "loss": 1.6707, + "step": 332 + }, + { + "epoch": 0.18078175895765472, + "grad_norm": 2.174263401724357, + "learning_rate": 1.9998060735272186e-05, + "loss": 1.6897, + "step": 333 + }, + { + "epoch": 0.18132464712269272, + "grad_norm": 1.8438064757899841, + "learning_rate": 1.9997990859769633e-05, + "loss": 1.4446, + "step": 334 + }, + { + "epoch": 0.18186753528773072, + "grad_norm": 2.0988396037795507, + "learning_rate": 1.9997919747699638e-05, + "loss": 1.8837, + "step": 335 + }, + { + "epoch": 0.18241042345276873, + "grad_norm": 2.582616969632433, + "learning_rate": 1.999784739907099e-05, + "loss": 2.459, + "step": 336 + }, + { + "epoch": 0.18295331161780673, + "grad_norm": 2.5198677585207707, + "learning_rate": 1.9997773813892644e-05, + "loss": 1.5117, + "step": 337 + }, + { + "epoch": 0.18349619978284473, + "grad_norm": 2.1361317951692635, + "learning_rate": 1.9997698992173697e-05, + "loss": 1.5892, + "step": 338 + }, + { + "epoch": 0.18403908794788273, + "grad_norm": 1.7284990063627486, + "learning_rate": 1.9997622933923406e-05, + "loss": 1.5151, + "step": 339 + }, + { + "epoch": 0.18458197611292074, + "grad_norm": 2.076752898704788, + "learning_rate": 1.9997545639151176e-05, + "loss": 2.3424, + "step": 340 + }, + { + "epoch": 0.18512486427795874, + "grad_norm": 2.479369276757794, + "learning_rate": 1.999746710786657e-05, + "loss": 1.5923, + "step": 341 + }, + { + "epoch": 0.18566775244299674, + "grad_norm": 1.9193986648366674, + "learning_rate": 1.9997387340079294e-05, + "loss": 1.6222, + "step": 342 + }, + { + "epoch": 0.18621064060803474, + "grad_norm": 1.8622986153150223, + "learning_rate": 1.999730633579922e-05, + "loss": 1.8402, + "step": 343 + }, + { + "epoch": 0.18675352877307275, + "grad_norm": 2.4012194427686238, + "learning_rate": 1.999722409503637e-05, + "loss": 1.9406, + "step": 344 + }, + { + "epoch": 0.18729641693811075, + "grad_norm": 2.3641518463127538, + "learning_rate": 1.9997140617800907e-05, + "loss": 2.0108, + "step": 345 + }, + { + "epoch": 0.18783930510314875, + "grad_norm": 2.406749148558192, + "learning_rate": 1.9997055904103156e-05, + "loss": 1.1323, + "step": 346 + }, + { + "epoch": 0.18838219326818675, + "grad_norm": 2.002785642517938, + "learning_rate": 1.99969699539536e-05, + "loss": 1.8745, + "step": 347 + }, + { + "epoch": 0.18892508143322476, + "grad_norm": 2.0002882431712816, + "learning_rate": 1.9996882767362874e-05, + "loss": 1.6311, + "step": 348 + }, + { + "epoch": 0.18946796959826276, + "grad_norm": 2.14418043520414, + "learning_rate": 1.9996794344341744e-05, + "loss": 1.7789, + "step": 349 + }, + { + "epoch": 0.19001085776330076, + "grad_norm": 2.0976983604919144, + "learning_rate": 1.9996704684901163e-05, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.19055374592833876, + "grad_norm": 2.017704706993719, + "learning_rate": 1.9996613789052214e-05, + "loss": 1.7974, + "step": 351 + }, + { + "epoch": 0.19109663409337677, + "grad_norm": 1.9907769296520095, + "learning_rate": 1.999652165680614e-05, + "loss": 1.9079, + "step": 352 + }, + { + "epoch": 0.19163952225841477, + "grad_norm": 2.3031451756435746, + "learning_rate": 1.999642828817433e-05, + "loss": 1.7022, + "step": 353 + }, + { + "epoch": 0.19218241042345277, + "grad_norm": 1.6722822353766713, + "learning_rate": 1.9996333683168342e-05, + "loss": 1.6655, + "step": 354 + }, + { + "epoch": 0.19272529858849077, + "grad_norm": 2.2974087830498027, + "learning_rate": 1.9996237841799874e-05, + "loss": 1.761, + "step": 355 + }, + { + "epoch": 0.19326818675352878, + "grad_norm": 1.9181254747298617, + "learning_rate": 1.9996140764080777e-05, + "loss": 1.8259, + "step": 356 + }, + { + "epoch": 0.19381107491856678, + "grad_norm": 2.07933834321742, + "learning_rate": 1.9996042450023053e-05, + "loss": 1.3783, + "step": 357 + }, + { + "epoch": 0.19435396308360478, + "grad_norm": 2.8332787295887436, + "learning_rate": 1.9995942899638875e-05, + "loss": 1.9796, + "step": 358 + }, + { + "epoch": 0.19489685124864278, + "grad_norm": 1.7924418183793758, + "learning_rate": 1.9995842112940545e-05, + "loss": 1.5063, + "step": 359 + }, + { + "epoch": 0.19543973941368079, + "grad_norm": 2.78627522830663, + "learning_rate": 1.9995740089940532e-05, + "loss": 1.891, + "step": 360 + }, + { + "epoch": 0.1959826275787188, + "grad_norm": 2.067311381839208, + "learning_rate": 1.9995636830651453e-05, + "loss": 1.9235, + "step": 361 + }, + { + "epoch": 0.1965255157437568, + "grad_norm": 2.539112412358468, + "learning_rate": 1.9995532335086078e-05, + "loss": 2.3062, + "step": 362 + }, + { + "epoch": 0.1970684039087948, + "grad_norm": 1.9040915545393904, + "learning_rate": 1.999542660325734e-05, + "loss": 1.9263, + "step": 363 + }, + { + "epoch": 0.1976112920738328, + "grad_norm": 1.7664296640025938, + "learning_rate": 1.9995319635178305e-05, + "loss": 1.5562, + "step": 364 + }, + { + "epoch": 0.1981541802388708, + "grad_norm": 1.687746217581272, + "learning_rate": 1.9995211430862206e-05, + "loss": 1.738, + "step": 365 + }, + { + "epoch": 0.1986970684039088, + "grad_norm": 1.9292930460598687, + "learning_rate": 1.9995101990322428e-05, + "loss": 1.5446, + "step": 366 + }, + { + "epoch": 0.1992399565689468, + "grad_norm": 1.9890390069401638, + "learning_rate": 1.9994991313572508e-05, + "loss": 1.9086, + "step": 367 + }, + { + "epoch": 0.1997828447339848, + "grad_norm": 1.8632430045841502, + "learning_rate": 1.999487940062613e-05, + "loss": 1.9747, + "step": 368 + }, + { + "epoch": 0.2003257328990228, + "grad_norm": 2.0677818303742237, + "learning_rate": 1.999476625149714e-05, + "loss": 1.4377, + "step": 369 + }, + { + "epoch": 0.2008686210640608, + "grad_norm": 1.7412205305358945, + "learning_rate": 1.9994651866199527e-05, + "loss": 1.9352, + "step": 370 + }, + { + "epoch": 0.2014115092290988, + "grad_norm": 1.8562175126383418, + "learning_rate": 1.9994536244747448e-05, + "loss": 1.5469, + "step": 371 + }, + { + "epoch": 0.20195439739413681, + "grad_norm": 1.8688345857586197, + "learning_rate": 1.9994419387155194e-05, + "loss": 1.8327, + "step": 372 + }, + { + "epoch": 0.20249728555917482, + "grad_norm": 2.1538407513899647, + "learning_rate": 1.9994301293437223e-05, + "loss": 1.9621, + "step": 373 + }, + { + "epoch": 0.20304017372421282, + "grad_norm": 2.4849664096044752, + "learning_rate": 1.999418196360814e-05, + "loss": 1.8179, + "step": 374 + }, + { + "epoch": 0.20358306188925082, + "grad_norm": 2.098949652063271, + "learning_rate": 1.99940613976827e-05, + "loss": 1.8719, + "step": 375 + }, + { + "epoch": 0.20412595005428882, + "grad_norm": 1.8802547433747465, + "learning_rate": 1.999393959567582e-05, + "loss": 1.5697, + "step": 376 + }, + { + "epoch": 0.20466883821932683, + "grad_norm": 2.090861017905227, + "learning_rate": 1.9993816557602567e-05, + "loss": 1.8249, + "step": 377 + }, + { + "epoch": 0.20521172638436483, + "grad_norm": 1.775481861127165, + "learning_rate": 1.999369228347815e-05, + "loss": 1.3095, + "step": 378 + }, + { + "epoch": 0.20575461454940283, + "grad_norm": 2.260330600420479, + "learning_rate": 1.999356677331794e-05, + "loss": 1.9153, + "step": 379 + }, + { + "epoch": 0.20629750271444083, + "grad_norm": 2.0769016143551484, + "learning_rate": 1.999344002713747e-05, + "loss": 1.958, + "step": 380 + }, + { + "epoch": 0.20684039087947884, + "grad_norm": 2.063584775140761, + "learning_rate": 1.9993312044952408e-05, + "loss": 1.7887, + "step": 381 + }, + { + "epoch": 0.20738327904451684, + "grad_norm": 1.9638736053910157, + "learning_rate": 1.9993182826778588e-05, + "loss": 1.3178, + "step": 382 + }, + { + "epoch": 0.20792616720955484, + "grad_norm": 1.7752329715548703, + "learning_rate": 1.9993052372631988e-05, + "loss": 1.4473, + "step": 383 + }, + { + "epoch": 0.20846905537459284, + "grad_norm": 2.1002641405528952, + "learning_rate": 1.999292068252874e-05, + "loss": 1.9698, + "step": 384 + }, + { + "epoch": 0.20901194353963085, + "grad_norm": 2.757803188261632, + "learning_rate": 1.999278775648514e-05, + "loss": 1.1517, + "step": 385 + }, + { + "epoch": 0.20955483170466885, + "grad_norm": 3.0341284449704378, + "learning_rate": 1.9992653594517624e-05, + "loss": 1.987, + "step": 386 + }, + { + "epoch": 0.21009771986970685, + "grad_norm": 2.445830333821562, + "learning_rate": 1.9992518196642786e-05, + "loss": 1.7656, + "step": 387 + }, + { + "epoch": 0.21064060803474485, + "grad_norm": 2.737783962568244, + "learning_rate": 1.9992381562877368e-05, + "loss": 2.3012, + "step": 388 + }, + { + "epoch": 0.21118349619978286, + "grad_norm": 2.0108432149096793, + "learning_rate": 1.9992243693238275e-05, + "loss": 2.1096, + "step": 389 + }, + { + "epoch": 0.21172638436482086, + "grad_norm": 2.1448465445226397, + "learning_rate": 1.9992104587742558e-05, + "loss": 1.6912, + "step": 390 + }, + { + "epoch": 0.21226927252985886, + "grad_norm": 2.842739011518123, + "learning_rate": 1.999196424640742e-05, + "loss": 1.9987, + "step": 391 + }, + { + "epoch": 0.21281216069489686, + "grad_norm": 2.026900017823922, + "learning_rate": 1.9991822669250216e-05, + "loss": 1.6251, + "step": 392 + }, + { + "epoch": 0.21335504885993486, + "grad_norm": 2.139818571931489, + "learning_rate": 1.9991679856288462e-05, + "loss": 1.4181, + "step": 393 + }, + { + "epoch": 0.21389793702497287, + "grad_norm": 2.061327407258535, + "learning_rate": 1.999153580753982e-05, + "loss": 2.0976, + "step": 394 + }, + { + "epoch": 0.21444082519001087, + "grad_norm": 1.9510646261617408, + "learning_rate": 1.9991390523022105e-05, + "loss": 1.3091, + "step": 395 + }, + { + "epoch": 0.21498371335504887, + "grad_norm": 2.0892427860245664, + "learning_rate": 1.9991244002753287e-05, + "loss": 1.3693, + "step": 396 + }, + { + "epoch": 0.21552660152008687, + "grad_norm": 1.9528708238712815, + "learning_rate": 1.9991096246751483e-05, + "loss": 1.2807, + "step": 397 + }, + { + "epoch": 0.21606948968512488, + "grad_norm": 2.0084632909809983, + "learning_rate": 1.9990947255034977e-05, + "loss": 1.7429, + "step": 398 + }, + { + "epoch": 0.21661237785016288, + "grad_norm": 2.19414527352725, + "learning_rate": 1.999079702762219e-05, + "loss": 1.5962, + "step": 399 + }, + { + "epoch": 0.21715526601520088, + "grad_norm": 2.184484203373308, + "learning_rate": 1.9990645564531702e-05, + "loss": 1.6887, + "step": 400 + }, + { + "epoch": 0.21769815418023888, + "grad_norm": 2.214473402714986, + "learning_rate": 1.9990492865782248e-05, + "loss": 1.5226, + "step": 401 + }, + { + "epoch": 0.2182410423452769, + "grad_norm": 2.1467748369867414, + "learning_rate": 1.9990338931392714e-05, + "loss": 1.5634, + "step": 402 + }, + { + "epoch": 0.21878393051031486, + "grad_norm": 2.1969353910959684, + "learning_rate": 1.999018376138214e-05, + "loss": 1.8652, + "step": 403 + }, + { + "epoch": 0.21932681867535286, + "grad_norm": 2.245031544698939, + "learning_rate": 1.9990027355769715e-05, + "loss": 2.1811, + "step": 404 + }, + { + "epoch": 0.21986970684039087, + "grad_norm": 2.381275606702572, + "learning_rate": 1.9989869714574784e-05, + "loss": 1.6356, + "step": 405 + }, + { + "epoch": 0.22041259500542887, + "grad_norm": 2.8847531540936053, + "learning_rate": 1.9989710837816846e-05, + "loss": 1.9953, + "step": 406 + }, + { + "epoch": 0.22095548317046687, + "grad_norm": 2.383074182943421, + "learning_rate": 1.9989550725515553e-05, + "loss": 1.5589, + "step": 407 + }, + { + "epoch": 0.22149837133550487, + "grad_norm": 2.3867744628533987, + "learning_rate": 1.99893893776907e-05, + "loss": 1.5351, + "step": 408 + }, + { + "epoch": 0.22204125950054288, + "grad_norm": 2.4139767232435463, + "learning_rate": 1.998922679436225e-05, + "loss": 1.5949, + "step": 409 + }, + { + "epoch": 0.22258414766558088, + "grad_norm": 3.034840719531937, + "learning_rate": 1.9989062975550313e-05, + "loss": 1.7994, + "step": 410 + }, + { + "epoch": 0.22312703583061888, + "grad_norm": 2.6245673280102544, + "learning_rate": 1.9988897921275144e-05, + "loss": 2.2593, + "step": 411 + }, + { + "epoch": 0.22366992399565688, + "grad_norm": 2.403934048820897, + "learning_rate": 1.998873163155716e-05, + "loss": 1.3179, + "step": 412 + }, + { + "epoch": 0.22421281216069489, + "grad_norm": 2.0174357446052524, + "learning_rate": 1.998856410641693e-05, + "loss": 1.8698, + "step": 413 + }, + { + "epoch": 0.2247557003257329, + "grad_norm": 2.25869172419699, + "learning_rate": 1.998839534587517e-05, + "loss": 1.596, + "step": 414 + }, + { + "epoch": 0.2252985884907709, + "grad_norm": 3.384596068816443, + "learning_rate": 1.9988225349952758e-05, + "loss": 1.5232, + "step": 415 + }, + { + "epoch": 0.2258414766558089, + "grad_norm": 2.2109577343653246, + "learning_rate": 1.9988054118670712e-05, + "loss": 1.2403, + "step": 416 + }, + { + "epoch": 0.2263843648208469, + "grad_norm": 2.1825505717325053, + "learning_rate": 1.9987881652050215e-05, + "loss": 1.7205, + "step": 417 + }, + { + "epoch": 0.2269272529858849, + "grad_norm": 2.7992033856118583, + "learning_rate": 1.99877079501126e-05, + "loss": 1.7486, + "step": 418 + }, + { + "epoch": 0.2274701411509229, + "grad_norm": 2.6351781282432793, + "learning_rate": 1.9987533012879344e-05, + "loss": 1.6693, + "step": 419 + }, + { + "epoch": 0.2280130293159609, + "grad_norm": 2.2385207734696384, + "learning_rate": 1.9987356840372088e-05, + "loss": 1.1296, + "step": 420 + }, + { + "epoch": 0.2285559174809989, + "grad_norm": 2.4921881643709107, + "learning_rate": 1.998717943261262e-05, + "loss": 1.1379, + "step": 421 + }, + { + "epoch": 0.2290988056460369, + "grad_norm": 2.842685080646979, + "learning_rate": 1.9987000789622884e-05, + "loss": 1.7963, + "step": 422 + }, + { + "epoch": 0.2296416938110749, + "grad_norm": 2.4776027353414145, + "learning_rate": 1.9986820911424972e-05, + "loss": 1.6939, + "step": 423 + }, + { + "epoch": 0.2301845819761129, + "grad_norm": 2.688785422525701, + "learning_rate": 1.9986639798041134e-05, + "loss": 1.2282, + "step": 424 + }, + { + "epoch": 0.23072747014115091, + "grad_norm": 3.4758865457335495, + "learning_rate": 1.998645744949377e-05, + "loss": 1.7664, + "step": 425 + }, + { + "epoch": 0.23127035830618892, + "grad_norm": 2.5586740557561605, + "learning_rate": 1.9986273865805432e-05, + "loss": 1.5637, + "step": 426 + }, + { + "epoch": 0.23181324647122692, + "grad_norm": 2.645306116903878, + "learning_rate": 1.9986089046998827e-05, + "loss": 1.5331, + "step": 427 + }, + { + "epoch": 0.23235613463626492, + "grad_norm": 2.748145883999443, + "learning_rate": 1.998590299309681e-05, + "loss": 1.7316, + "step": 428 + }, + { + "epoch": 0.23289902280130292, + "grad_norm": 2.4357923833746438, + "learning_rate": 1.99857157041224e-05, + "loss": 1.1479, + "step": 429 + }, + { + "epoch": 0.23344191096634093, + "grad_norm": 2.7322633978331377, + "learning_rate": 1.9985527180098755e-05, + "loss": 1.6089, + "step": 430 + }, + { + "epoch": 0.23398479913137893, + "grad_norm": 2.625531215380387, + "learning_rate": 1.9985337421049193e-05, + "loss": 1.8973, + "step": 431 + }, + { + "epoch": 0.23452768729641693, + "grad_norm": 2.1586487666504754, + "learning_rate": 1.9985146426997185e-05, + "loss": 1.664, + "step": 432 + }, + { + "epoch": 0.23507057546145493, + "grad_norm": 2.748704307086674, + "learning_rate": 1.9984954197966355e-05, + "loss": 1.677, + "step": 433 + }, + { + "epoch": 0.23561346362649294, + "grad_norm": 2.345532526959197, + "learning_rate": 1.9984760733980476e-05, + "loss": 1.5133, + "step": 434 + }, + { + "epoch": 0.23615635179153094, + "grad_norm": 2.1655585346308848, + "learning_rate": 1.9984566035063473e-05, + "loss": 1.4206, + "step": 435 + }, + { + "epoch": 0.23669923995656894, + "grad_norm": 2.697584333049271, + "learning_rate": 1.9984370101239434e-05, + "loss": 1.8131, + "step": 436 + }, + { + "epoch": 0.23724212812160694, + "grad_norm": 2.235069158786981, + "learning_rate": 1.9984172932532583e-05, + "loss": 1.7839, + "step": 437 + }, + { + "epoch": 0.23778501628664495, + "grad_norm": 2.548562657856099, + "learning_rate": 1.998397452896731e-05, + "loss": 1.4876, + "step": 438 + }, + { + "epoch": 0.23832790445168295, + "grad_norm": 2.1698874138883673, + "learning_rate": 1.9983774890568163e-05, + "loss": 1.2916, + "step": 439 + }, + { + "epoch": 0.23887079261672095, + "grad_norm": 2.3237954527681084, + "learning_rate": 1.998357401735982e-05, + "loss": 1.398, + "step": 440 + }, + { + "epoch": 0.23941368078175895, + "grad_norm": 2.448364938314344, + "learning_rate": 1.9983371909367135e-05, + "loss": 1.2663, + "step": 441 + }, + { + "epoch": 0.23995656894679696, + "grad_norm": 1.935899726785714, + "learning_rate": 1.99831685666151e-05, + "loss": 1.2448, + "step": 442 + }, + { + "epoch": 0.24049945711183496, + "grad_norm": 2.14357016947003, + "learning_rate": 1.9982963989128864e-05, + "loss": 1.4786, + "step": 443 + }, + { + "epoch": 0.24104234527687296, + "grad_norm": 2.5830135196717148, + "learning_rate": 1.998275817693373e-05, + "loss": 1.8324, + "step": 444 + }, + { + "epoch": 0.24158523344191096, + "grad_norm": 3.159289299734819, + "learning_rate": 1.9982551130055157e-05, + "loss": 1.9091, + "step": 445 + }, + { + "epoch": 0.24212812160694897, + "grad_norm": 3.2712239674501755, + "learning_rate": 1.9982342848518753e-05, + "loss": 1.7869, + "step": 446 + }, + { + "epoch": 0.24267100977198697, + "grad_norm": 3.35834841128443, + "learning_rate": 1.998213333235027e-05, + "loss": 1.2772, + "step": 447 + }, + { + "epoch": 0.24321389793702497, + "grad_norm": 2.9379666136057354, + "learning_rate": 1.998192258157563e-05, + "loss": 1.3497, + "step": 448 + }, + { + "epoch": 0.24375678610206297, + "grad_norm": 2.334119075845609, + "learning_rate": 1.9981710596220897e-05, + "loss": 1.5541, + "step": 449 + }, + { + "epoch": 0.24429967426710097, + "grad_norm": 2.781469856763384, + "learning_rate": 1.998149737631229e-05, + "loss": 1.9413, + "step": 450 + }, + { + "epoch": 0.24484256243213898, + "grad_norm": 3.0196781534130452, + "learning_rate": 1.9981282921876177e-05, + "loss": 1.3238, + "step": 451 + }, + { + "epoch": 0.24538545059717698, + "grad_norm": 3.547169321727429, + "learning_rate": 1.9981067232939086e-05, + "loss": 1.9952, + "step": 452 + }, + { + "epoch": 0.24592833876221498, + "grad_norm": 3.2411499530913535, + "learning_rate": 1.9980850309527693e-05, + "loss": 1.8244, + "step": 453 + }, + { + "epoch": 0.24647122692725298, + "grad_norm": 2.670313260104859, + "learning_rate": 1.9980632151668822e-05, + "loss": 1.607, + "step": 454 + }, + { + "epoch": 0.247014115092291, + "grad_norm": 2.8822100638306143, + "learning_rate": 1.9980412759389468e-05, + "loss": 1.8868, + "step": 455 + }, + { + "epoch": 0.247557003257329, + "grad_norm": 2.4776152417583317, + "learning_rate": 1.9980192132716748e-05, + "loss": 1.778, + "step": 456 + }, + { + "epoch": 0.248099891422367, + "grad_norm": 2.2001923672712076, + "learning_rate": 1.9979970271677967e-05, + "loss": 1.3544, + "step": 457 + }, + { + "epoch": 0.248642779587405, + "grad_norm": 2.7694932683911837, + "learning_rate": 1.9979747176300553e-05, + "loss": 1.6521, + "step": 458 + }, + { + "epoch": 0.249185667752443, + "grad_norm": 2.8464573075472845, + "learning_rate": 1.99795228466121e-05, + "loss": 1.8803, + "step": 459 + }, + { + "epoch": 0.249728555917481, + "grad_norm": 2.577989994947286, + "learning_rate": 1.9979297282640365e-05, + "loss": 1.8838, + "step": 460 + }, + { + "epoch": 0.250271444082519, + "grad_norm": 2.9859834231033164, + "learning_rate": 1.997907048441323e-05, + "loss": 1.5122, + "step": 461 + }, + { + "epoch": 0.250814332247557, + "grad_norm": 3.0857580735568098, + "learning_rate": 1.9978842451958757e-05, + "loss": 1.7789, + "step": 462 + }, + { + "epoch": 0.251357220412595, + "grad_norm": 2.523783848348806, + "learning_rate": 1.9978613185305145e-05, + "loss": 1.5815, + "step": 463 + }, + { + "epoch": 0.251900108577633, + "grad_norm": 2.699094362013039, + "learning_rate": 1.9978382684480747e-05, + "loss": 1.7448, + "step": 464 + }, + { + "epoch": 0.252442996742671, + "grad_norm": 2.736268550628698, + "learning_rate": 1.997815094951408e-05, + "loss": 1.7052, + "step": 465 + }, + { + "epoch": 0.252985884907709, + "grad_norm": 3.1150828428090014, + "learning_rate": 1.99779179804338e-05, + "loss": 1.7743, + "step": 466 + }, + { + "epoch": 0.253528773072747, + "grad_norm": 3.3513164619888482, + "learning_rate": 1.997768377726872e-05, + "loss": 2.3905, + "step": 467 + }, + { + "epoch": 0.254071661237785, + "grad_norm": 2.400886982379507, + "learning_rate": 1.9977448340047808e-05, + "loss": 1.6096, + "step": 468 + }, + { + "epoch": 0.254614549402823, + "grad_norm": 2.633533044966171, + "learning_rate": 1.9977211668800186e-05, + "loss": 1.7796, + "step": 469 + }, + { + "epoch": 0.255157437567861, + "grad_norm": 3.9297139007235042, + "learning_rate": 1.997697376355512e-05, + "loss": 1.4602, + "step": 470 + }, + { + "epoch": 0.255700325732899, + "grad_norm": 5.588260619709643, + "learning_rate": 1.9976734624342044e-05, + "loss": 2.0389, + "step": 471 + }, + { + "epoch": 0.256243213897937, + "grad_norm": 2.600723171476426, + "learning_rate": 1.9976494251190522e-05, + "loss": 1.3676, + "step": 472 + }, + { + "epoch": 0.25678610206297503, + "grad_norm": 2.61945002649116, + "learning_rate": 1.9976252644130297e-05, + "loss": 1.7902, + "step": 473 + }, + { + "epoch": 0.25732899022801303, + "grad_norm": 3.7513561790803838, + "learning_rate": 1.997600980319124e-05, + "loss": 1.5997, + "step": 474 + }, + { + "epoch": 0.25787187839305103, + "grad_norm": 2.8832359552778737, + "learning_rate": 1.9975765728403395e-05, + "loss": 1.9636, + "step": 475 + }, + { + "epoch": 0.25841476655808904, + "grad_norm": 3.0975018752600243, + "learning_rate": 1.9975520419796942e-05, + "loss": 1.1165, + "step": 476 + }, + { + "epoch": 0.25895765472312704, + "grad_norm": 2.121708502818221, + "learning_rate": 1.9975273877402227e-05, + "loss": 1.5108, + "step": 477 + }, + { + "epoch": 0.25950054288816504, + "grad_norm": 2.4073592870530116, + "learning_rate": 1.997502610124974e-05, + "loss": 1.7828, + "step": 478 + }, + { + "epoch": 0.26004343105320304, + "grad_norm": 2.6858679986632974, + "learning_rate": 1.997477709137013e-05, + "loss": 1.8483, + "step": 479 + }, + { + "epoch": 0.26058631921824105, + "grad_norm": 3.283752190131325, + "learning_rate": 1.997452684779419e-05, + "loss": 1.6105, + "step": 480 + }, + { + "epoch": 0.26112920738327905, + "grad_norm": 2.500181185675909, + "learning_rate": 1.997427537055287e-05, + "loss": 1.5475, + "step": 481 + }, + { + "epoch": 0.26167209554831705, + "grad_norm": 2.3992324550953885, + "learning_rate": 1.9974022659677278e-05, + "loss": 1.6062, + "step": 482 + }, + { + "epoch": 0.26221498371335505, + "grad_norm": 2.7692293822867837, + "learning_rate": 1.9973768715198667e-05, + "loss": 1.4995, + "step": 483 + }, + { + "epoch": 0.26275787187839306, + "grad_norm": 2.971423364277874, + "learning_rate": 1.9973513537148447e-05, + "loss": 1.7904, + "step": 484 + }, + { + "epoch": 0.26330076004343106, + "grad_norm": 2.2769736321644105, + "learning_rate": 1.9973257125558177e-05, + "loss": 1.3121, + "step": 485 + }, + { + "epoch": 0.26384364820846906, + "grad_norm": 2.5949046051899254, + "learning_rate": 1.997299948045957e-05, + "loss": 1.4555, + "step": 486 + }, + { + "epoch": 0.26438653637350706, + "grad_norm": 3.0210593683445204, + "learning_rate": 1.997274060188449e-05, + "loss": 1.8211, + "step": 487 + }, + { + "epoch": 0.26492942453854507, + "grad_norm": 3.076626224616319, + "learning_rate": 1.9972480489864962e-05, + "loss": 1.6366, + "step": 488 + }, + { + "epoch": 0.26547231270358307, + "grad_norm": 3.1369728365663536, + "learning_rate": 1.9972219144433148e-05, + "loss": 1.5027, + "step": 489 + }, + { + "epoch": 0.26601520086862107, + "grad_norm": 2.2870450349164635, + "learning_rate": 1.9971956565621383e-05, + "loss": 1.2784, + "step": 490 + }, + { + "epoch": 0.2665580890336591, + "grad_norm": 3.0130036065633776, + "learning_rate": 1.9971692753462134e-05, + "loss": 1.2083, + "step": 491 + }, + { + "epoch": 0.2671009771986971, + "grad_norm": 2.6063157323029733, + "learning_rate": 1.9971427707988034e-05, + "loss": 1.0083, + "step": 492 + }, + { + "epoch": 0.2676438653637351, + "grad_norm": 2.579371053895234, + "learning_rate": 1.997116142923186e-05, + "loss": 1.1937, + "step": 493 + }, + { + "epoch": 0.2681867535287731, + "grad_norm": 2.532537971800688, + "learning_rate": 1.9970893917226554e-05, + "loss": 1.4735, + "step": 494 + }, + { + "epoch": 0.2687296416938111, + "grad_norm": 2.3483150144294105, + "learning_rate": 1.997062517200519e-05, + "loss": 1.7269, + "step": 495 + }, + { + "epoch": 0.2692725298588491, + "grad_norm": 2.594809867192747, + "learning_rate": 1.997035519360102e-05, + "loss": 1.8283, + "step": 496 + }, + { + "epoch": 0.2698154180238871, + "grad_norm": 2.613813750609998, + "learning_rate": 1.9970083982047428e-05, + "loss": 1.2302, + "step": 497 + }, + { + "epoch": 0.2703583061889251, + "grad_norm": 2.648279162964909, + "learning_rate": 1.9969811537377956e-05, + "loss": 1.6225, + "step": 498 + }, + { + "epoch": 0.2709011943539631, + "grad_norm": 2.3790190706794325, + "learning_rate": 1.9969537859626308e-05, + "loss": 1.5172, + "step": 499 + }, + { + "epoch": 0.2714440825190011, + "grad_norm": 2.7054998578606364, + "learning_rate": 1.9969262948826326e-05, + "loss": 1.4525, + "step": 500 + }, + { + "epoch": 0.2719869706840391, + "grad_norm": 2.412151508264948, + "learning_rate": 1.9968986805012012e-05, + "loss": 1.3299, + "step": 501 + }, + { + "epoch": 0.2725298588490771, + "grad_norm": 2.4836460319285414, + "learning_rate": 1.9968709428217525e-05, + "loss": 1.5217, + "step": 502 + }, + { + "epoch": 0.2730727470141151, + "grad_norm": 3.3112944949184606, + "learning_rate": 1.9968430818477168e-05, + "loss": 2.0643, + "step": 503 + }, + { + "epoch": 0.2736156351791531, + "grad_norm": 3.266395629954733, + "learning_rate": 1.9968150975825397e-05, + "loss": 1.1719, + "step": 504 + }, + { + "epoch": 0.2741585233441911, + "grad_norm": 2.3487351103507073, + "learning_rate": 1.996786990029683e-05, + "loss": 1.6876, + "step": 505 + }, + { + "epoch": 0.2747014115092291, + "grad_norm": 3.0672442719402673, + "learning_rate": 1.9967587591926227e-05, + "loss": 1.3946, + "step": 506 + }, + { + "epoch": 0.2752442996742671, + "grad_norm": 2.84320621483769, + "learning_rate": 1.99673040507485e-05, + "loss": 1.3259, + "step": 507 + }, + { + "epoch": 0.2757871878393051, + "grad_norm": 2.5078449617010707, + "learning_rate": 1.9967019276798728e-05, + "loss": 0.8478, + "step": 508 + }, + { + "epoch": 0.2763300760043431, + "grad_norm": 3.137784699454796, + "learning_rate": 1.9966733270112126e-05, + "loss": 1.1688, + "step": 509 + }, + { + "epoch": 0.2768729641693811, + "grad_norm": 3.7277826491955017, + "learning_rate": 1.996644603072407e-05, + "loss": 1.1091, + "step": 510 + }, + { + "epoch": 0.2774158523344191, + "grad_norm": 2.5276781604415635, + "learning_rate": 1.996615755867008e-05, + "loss": 1.1299, + "step": 511 + }, + { + "epoch": 0.2779587404994571, + "grad_norm": 3.1012700661738744, + "learning_rate": 1.996586785398584e-05, + "loss": 1.7218, + "step": 512 + }, + { + "epoch": 0.2785016286644951, + "grad_norm": 2.285166712515903, + "learning_rate": 1.9965576916707182e-05, + "loss": 1.2868, + "step": 513 + }, + { + "epoch": 0.27904451682953313, + "grad_norm": 2.33097906349044, + "learning_rate": 1.9965284746870088e-05, + "loss": 0.9887, + "step": 514 + }, + { + "epoch": 0.27958740499457113, + "grad_norm": 2.6473787082237927, + "learning_rate": 1.9964991344510697e-05, + "loss": 1.8543, + "step": 515 + }, + { + "epoch": 0.28013029315960913, + "grad_norm": 2.4628160599533366, + "learning_rate": 1.996469670966529e-05, + "loss": 1.2263, + "step": 516 + }, + { + "epoch": 0.28067318132464714, + "grad_norm": 3.2897847068350905, + "learning_rate": 1.9964400842370314e-05, + "loss": 1.6338, + "step": 517 + }, + { + "epoch": 0.28121606948968514, + "grad_norm": 2.4439319341540324, + "learning_rate": 1.9964103742662363e-05, + "loss": 1.0836, + "step": 518 + }, + { + "epoch": 0.28175895765472314, + "grad_norm": 2.3221991020412003, + "learning_rate": 1.996380541057818e-05, + "loss": 1.2331, + "step": 519 + }, + { + "epoch": 0.28230184581976114, + "grad_norm": 2.9571040634251564, + "learning_rate": 1.9963505846154662e-05, + "loss": 1.3066, + "step": 520 + }, + { + "epoch": 0.28284473398479915, + "grad_norm": 3.7512706020225624, + "learning_rate": 1.996320504942886e-05, + "loss": 1.7482, + "step": 521 + }, + { + "epoch": 0.28338762214983715, + "grad_norm": 2.4620109793388267, + "learning_rate": 1.9962903020437983e-05, + "loss": 1.5334, + "step": 522 + }, + { + "epoch": 0.28393051031487515, + "grad_norm": 3.030374272795485, + "learning_rate": 1.9962599759219383e-05, + "loss": 1.8957, + "step": 523 + }, + { + "epoch": 0.28447339847991315, + "grad_norm": 2.452389821491403, + "learning_rate": 1.9962295265810563e-05, + "loss": 1.5438, + "step": 524 + }, + { + "epoch": 0.28501628664495116, + "grad_norm": 2.713028369466205, + "learning_rate": 1.996198954024919e-05, + "loss": 1.4272, + "step": 525 + }, + { + "epoch": 0.28555917480998916, + "grad_norm": 3.810321275175567, + "learning_rate": 1.996168258257307e-05, + "loss": 1.9028, + "step": 526 + }, + { + "epoch": 0.28610206297502716, + "grad_norm": 2.7774100977441236, + "learning_rate": 1.9961374392820173e-05, + "loss": 1.5644, + "step": 527 + }, + { + "epoch": 0.28664495114006516, + "grad_norm": 2.6798712089104186, + "learning_rate": 1.9961064971028616e-05, + "loss": 1.188, + "step": 528 + }, + { + "epoch": 0.28718783930510317, + "grad_norm": 2.889465990486677, + "learning_rate": 1.9960754317236666e-05, + "loss": 1.5393, + "step": 529 + }, + { + "epoch": 0.28773072747014117, + "grad_norm": 3.465251366831076, + "learning_rate": 1.996044243148275e-05, + "loss": 2.1899, + "step": 530 + }, + { + "epoch": 0.28827361563517917, + "grad_norm": 2.645941940974219, + "learning_rate": 1.9960129313805437e-05, + "loss": 1.3691, + "step": 531 + }, + { + "epoch": 0.2888165038002172, + "grad_norm": 2.3914199977194293, + "learning_rate": 1.9959814964243455e-05, + "loss": 1.5219, + "step": 532 + }, + { + "epoch": 0.2893593919652552, + "grad_norm": 2.3023810529281343, + "learning_rate": 1.995949938283569e-05, + "loss": 1.5147, + "step": 533 + }, + { + "epoch": 0.2899022801302932, + "grad_norm": 2.7362205671791155, + "learning_rate": 1.9959182569621164e-05, + "loss": 1.7571, + "step": 534 + }, + { + "epoch": 0.2904451682953312, + "grad_norm": 3.971162331076012, + "learning_rate": 1.9958864524639066e-05, + "loss": 1.3425, + "step": 535 + }, + { + "epoch": 0.2909880564603692, + "grad_norm": 3.4144928239616514, + "learning_rate": 1.9958545247928727e-05, + "loss": 1.6962, + "step": 536 + }, + { + "epoch": 0.2915309446254072, + "grad_norm": 3.5063126675319043, + "learning_rate": 1.9958224739529647e-05, + "loss": 1.6406, + "step": 537 + }, + { + "epoch": 0.2920738327904452, + "grad_norm": 2.9013783116047547, + "learning_rate": 1.995790299948146e-05, + "loss": 1.6376, + "step": 538 + }, + { + "epoch": 0.2926167209554832, + "grad_norm": 2.4827691033904693, + "learning_rate": 1.9957580027823957e-05, + "loss": 1.8672, + "step": 539 + }, + { + "epoch": 0.2931596091205212, + "grad_norm": 3.1245563354940242, + "learning_rate": 1.9957255824597087e-05, + "loss": 1.885, + "step": 540 + }, + { + "epoch": 0.2937024972855592, + "grad_norm": 2.7941860466759896, + "learning_rate": 1.9956930389840945e-05, + "loss": 1.0903, + "step": 541 + }, + { + "epoch": 0.2942453854505972, + "grad_norm": 3.1348104146717772, + "learning_rate": 1.9956603723595784e-05, + "loss": 2.1446, + "step": 542 + }, + { + "epoch": 0.2947882736156352, + "grad_norm": 2.4248222663066747, + "learning_rate": 1.995627582590201e-05, + "loss": 1.3503, + "step": 543 + }, + { + "epoch": 0.2953311617806732, + "grad_norm": 3.1538245864476337, + "learning_rate": 1.995594669680017e-05, + "loss": 1.3275, + "step": 544 + }, + { + "epoch": 0.2958740499457112, + "grad_norm": 3.436905240480997, + "learning_rate": 1.9955616336330976e-05, + "loss": 1.7249, + "step": 545 + }, + { + "epoch": 0.2964169381107492, + "grad_norm": 3.9145352507266393, + "learning_rate": 1.9955284744535287e-05, + "loss": 2.0089, + "step": 546 + }, + { + "epoch": 0.2969598262757872, + "grad_norm": 2.8600509325831185, + "learning_rate": 1.9954951921454113e-05, + "loss": 1.4527, + "step": 547 + }, + { + "epoch": 0.2975027144408252, + "grad_norm": 2.865760632888347, + "learning_rate": 1.995461786712862e-05, + "loss": 1.4553, + "step": 548 + }, + { + "epoch": 0.2980456026058632, + "grad_norm": 3.3559882259900706, + "learning_rate": 1.9954282581600127e-05, + "loss": 2.0456, + "step": 549 + }, + { + "epoch": 0.2985884907709012, + "grad_norm": 3.3424656149019008, + "learning_rate": 1.9953946064910098e-05, + "loss": 1.8253, + "step": 550 + }, + { + "epoch": 0.2991313789359392, + "grad_norm": 2.569119318410074, + "learning_rate": 1.9953608317100153e-05, + "loss": 1.3623, + "step": 551 + }, + { + "epoch": 0.2996742671009772, + "grad_norm": 2.8918953822102424, + "learning_rate": 1.995326933821207e-05, + "loss": 1.7521, + "step": 552 + }, + { + "epoch": 0.3002171552660152, + "grad_norm": 2.7627900710192246, + "learning_rate": 1.995292912828777e-05, + "loss": 1.775, + "step": 553 + }, + { + "epoch": 0.3007600434310532, + "grad_norm": 2.8053609300694804, + "learning_rate": 1.9952587687369334e-05, + "loss": 1.7536, + "step": 554 + }, + { + "epoch": 0.30130293159609123, + "grad_norm": 2.9775715301146803, + "learning_rate": 1.995224501549899e-05, + "loss": 1.6715, + "step": 555 + }, + { + "epoch": 0.30184581976112923, + "grad_norm": 3.146995410263436, + "learning_rate": 1.9951901112719123e-05, + "loss": 1.1032, + "step": 556 + }, + { + "epoch": 0.30238870792616723, + "grad_norm": 3.219197817112143, + "learning_rate": 1.9951555979072266e-05, + "loss": 1.6326, + "step": 557 + }, + { + "epoch": 0.30293159609120524, + "grad_norm": 2.676508071644292, + "learning_rate": 1.99512096146011e-05, + "loss": 1.4836, + "step": 558 + }, + { + "epoch": 0.30347448425624324, + "grad_norm": 3.8806900857620374, + "learning_rate": 1.9950862019348474e-05, + "loss": 1.7794, + "step": 559 + }, + { + "epoch": 0.30401737242128124, + "grad_norm": 2.547409032322543, + "learning_rate": 1.995051319335737e-05, + "loss": 1.3263, + "step": 560 + }, + { + "epoch": 0.30456026058631924, + "grad_norm": 3.025189850713409, + "learning_rate": 1.995016313667094e-05, + "loss": 1.2409, + "step": 561 + }, + { + "epoch": 0.30510314875135724, + "grad_norm": 3.3644665856402614, + "learning_rate": 1.9949811849332476e-05, + "loss": 1.2988, + "step": 562 + }, + { + "epoch": 0.30564603691639525, + "grad_norm": 3.4126932100522755, + "learning_rate": 1.9949459331385422e-05, + "loss": 1.6126, + "step": 563 + }, + { + "epoch": 0.30618892508143325, + "grad_norm": 3.3751971677759416, + "learning_rate": 1.994910558287338e-05, + "loss": 1.6243, + "step": 564 + }, + { + "epoch": 0.30673181324647125, + "grad_norm": 3.0948952637866105, + "learning_rate": 1.9948750603840102e-05, + "loss": 1.6553, + "step": 565 + }, + { + "epoch": 0.30727470141150925, + "grad_norm": 3.1116929004314224, + "learning_rate": 1.9948394394329494e-05, + "loss": 1.0466, + "step": 566 + }, + { + "epoch": 0.30781758957654726, + "grad_norm": 2.649244034262683, + "learning_rate": 1.9948036954385613e-05, + "loss": 1.2914, + "step": 567 + }, + { + "epoch": 0.30836047774158526, + "grad_norm": 3.9338023936000965, + "learning_rate": 1.9947678284052667e-05, + "loss": 1.7532, + "step": 568 + }, + { + "epoch": 0.30890336590662326, + "grad_norm": 2.887133065759567, + "learning_rate": 1.9947318383375017e-05, + "loss": 1.8001, + "step": 569 + }, + { + "epoch": 0.30944625407166126, + "grad_norm": 3.6996119456889915, + "learning_rate": 1.9946957252397173e-05, + "loss": 2.4852, + "step": 570 + }, + { + "epoch": 0.30998914223669927, + "grad_norm": 2.829855772438557, + "learning_rate": 1.9946594891163808e-05, + "loss": 1.6048, + "step": 571 + }, + { + "epoch": 0.31053203040173727, + "grad_norm": 4.288772356209683, + "learning_rate": 1.9946231299719732e-05, + "loss": 1.4841, + "step": 572 + }, + { + "epoch": 0.31107491856677527, + "grad_norm": 2.37958986026152, + "learning_rate": 1.9945866478109914e-05, + "loss": 1.1797, + "step": 573 + }, + { + "epoch": 0.3116178067318133, + "grad_norm": 2.9977970906442932, + "learning_rate": 1.9945500426379483e-05, + "loss": 1.224, + "step": 574 + }, + { + "epoch": 0.3121606948968513, + "grad_norm": 3.406596416686285, + "learning_rate": 1.9945133144573705e-05, + "loss": 1.4793, + "step": 575 + }, + { + "epoch": 0.3127035830618892, + "grad_norm": 3.3772239188722244, + "learning_rate": 1.994476463273801e-05, + "loss": 1.3696, + "step": 576 + }, + { + "epoch": 0.3132464712269272, + "grad_norm": 2.729299768057245, + "learning_rate": 1.9944394890917977e-05, + "loss": 1.448, + "step": 577 + }, + { + "epoch": 0.31378935939196523, + "grad_norm": 2.6526406591248297, + "learning_rate": 1.9944023919159335e-05, + "loss": 1.6905, + "step": 578 + }, + { + "epoch": 0.31433224755700323, + "grad_norm": 2.9512501195242944, + "learning_rate": 1.9943651717507965e-05, + "loss": 1.6277, + "step": 579 + }, + { + "epoch": 0.31487513572204123, + "grad_norm": 2.8292191327941723, + "learning_rate": 1.9943278286009903e-05, + "loss": 1.0532, + "step": 580 + }, + { + "epoch": 0.31541802388707924, + "grad_norm": 3.2105870294745436, + "learning_rate": 1.9942903624711335e-05, + "loss": 1.1823, + "step": 581 + }, + { + "epoch": 0.31596091205211724, + "grad_norm": 3.224137115744835, + "learning_rate": 1.9942527733658602e-05, + "loss": 1.5409, + "step": 582 + }, + { + "epoch": 0.31650380021715524, + "grad_norm": 3.126134123334164, + "learning_rate": 1.9942150612898194e-05, + "loss": 1.2423, + "step": 583 + }, + { + "epoch": 0.31704668838219324, + "grad_norm": 2.709954974374804, + "learning_rate": 1.994177226247675e-05, + "loss": 1.3419, + "step": 584 + }, + { + "epoch": 0.31758957654723124, + "grad_norm": 2.9030367366777927, + "learning_rate": 1.9941392682441066e-05, + "loss": 1.427, + "step": 585 + }, + { + "epoch": 0.31813246471226925, + "grad_norm": 3.272875981776567, + "learning_rate": 1.9941011872838092e-05, + "loss": 2.0196, + "step": 586 + }, + { + "epoch": 0.31867535287730725, + "grad_norm": 3.3480185179867, + "learning_rate": 1.994062983371493e-05, + "loss": 1.6038, + "step": 587 + }, + { + "epoch": 0.31921824104234525, + "grad_norm": 2.801738772846361, + "learning_rate": 1.9940246565118822e-05, + "loss": 1.7505, + "step": 588 + }, + { + "epoch": 0.31976112920738325, + "grad_norm": 2.743550967049156, + "learning_rate": 1.993986206709718e-05, + "loss": 1.3914, + "step": 589 + }, + { + "epoch": 0.32030401737242126, + "grad_norm": 2.481845489278486, + "learning_rate": 1.9939476339697555e-05, + "loss": 1.6927, + "step": 590 + }, + { + "epoch": 0.32084690553745926, + "grad_norm": 3.145019330878407, + "learning_rate": 1.993908938296765e-05, + "loss": 1.42, + "step": 591 + }, + { + "epoch": 0.32138979370249726, + "grad_norm": 3.326427662456394, + "learning_rate": 1.9938701196955335e-05, + "loss": 1.2516, + "step": 592 + }, + { + "epoch": 0.32193268186753526, + "grad_norm": 3.195815443331326, + "learning_rate": 1.9938311781708616e-05, + "loss": 2.1428, + "step": 593 + }, + { + "epoch": 0.32247557003257327, + "grad_norm": 2.461395361556941, + "learning_rate": 1.9937921137275657e-05, + "loss": 1.4448, + "step": 594 + }, + { + "epoch": 0.32301845819761127, + "grad_norm": 3.1793351214549794, + "learning_rate": 1.993752926370477e-05, + "loss": 1.4609, + "step": 595 + }, + { + "epoch": 0.32356134636264927, + "grad_norm": 2.783909288864463, + "learning_rate": 1.9937136161044427e-05, + "loss": 1.3355, + "step": 596 + }, + { + "epoch": 0.3241042345276873, + "grad_norm": 2.8156929696256734, + "learning_rate": 1.9936741829343247e-05, + "loss": 2.101, + "step": 597 + }, + { + "epoch": 0.3246471226927253, + "grad_norm": 3.793352093788154, + "learning_rate": 1.993634626865e-05, + "loss": 2.19, + "step": 598 + }, + { + "epoch": 0.3251900108577633, + "grad_norm": 2.6680863350639545, + "learning_rate": 1.993594947901361e-05, + "loss": 1.5199, + "step": 599 + }, + { + "epoch": 0.3257328990228013, + "grad_norm": 2.6498586813134297, + "learning_rate": 1.9935551460483155e-05, + "loss": 0.9282, + "step": 600 + }, + { + "epoch": 0.3262757871878393, + "grad_norm": 3.443063689484479, + "learning_rate": 1.993515221310786e-05, + "loss": 1.9107, + "step": 601 + }, + { + "epoch": 0.3268186753528773, + "grad_norm": 2.45686855619251, + "learning_rate": 1.9934751736937103e-05, + "loss": 1.2929, + "step": 602 + }, + { + "epoch": 0.3273615635179153, + "grad_norm": 3.3974155688994077, + "learning_rate": 1.9934350032020417e-05, + "loss": 1.546, + "step": 603 + }, + { + "epoch": 0.3279044516829533, + "grad_norm": 2.3692191456624783, + "learning_rate": 1.993394709840749e-05, + "loss": 1.2239, + "step": 604 + }, + { + "epoch": 0.3284473398479913, + "grad_norm": 4.044352925058249, + "learning_rate": 1.993354293614815e-05, + "loss": 1.801, + "step": 605 + }, + { + "epoch": 0.3289902280130293, + "grad_norm": 4.051192057617293, + "learning_rate": 1.993313754529239e-05, + "loss": 1.8043, + "step": 606 + }, + { + "epoch": 0.3295331161780673, + "grad_norm": 3.3680281079594634, + "learning_rate": 1.9932730925890344e-05, + "loss": 1.4915, + "step": 607 + }, + { + "epoch": 0.3300760043431053, + "grad_norm": 2.993677842102555, + "learning_rate": 1.9932323077992312e-05, + "loss": 1.3457, + "step": 608 + }, + { + "epoch": 0.3306188925081433, + "grad_norm": 3.325666927669253, + "learning_rate": 1.9931914001648726e-05, + "loss": 1.4221, + "step": 609 + }, + { + "epoch": 0.3311617806731813, + "grad_norm": 2.459382097232459, + "learning_rate": 1.993150369691019e-05, + "loss": 1.2551, + "step": 610 + }, + { + "epoch": 0.3317046688382193, + "grad_norm": 5.281638549560053, + "learning_rate": 1.993109216382745e-05, + "loss": 1.142, + "step": 611 + }, + { + "epoch": 0.3322475570032573, + "grad_norm": 2.6931033883715374, + "learning_rate": 1.99306794024514e-05, + "loss": 1.2573, + "step": 612 + }, + { + "epoch": 0.3327904451682953, + "grad_norm": 2.8066393604642714, + "learning_rate": 1.9930265412833097e-05, + "loss": 0.9847, + "step": 613 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 3.682172765988149, + "learning_rate": 1.992985019502374e-05, + "loss": 1.4482, + "step": 614 + }, + { + "epoch": 0.3338762214983713, + "grad_norm": 3.7190248655227873, + "learning_rate": 1.9929433749074684e-05, + "loss": 1.3682, + "step": 615 + }, + { + "epoch": 0.3344191096634093, + "grad_norm": 2.8594798198555527, + "learning_rate": 1.9929016075037438e-05, + "loss": 1.2396, + "step": 616 + }, + { + "epoch": 0.3349619978284473, + "grad_norm": 3.5439348644341035, + "learning_rate": 1.992859717296366e-05, + "loss": 1.5843, + "step": 617 + }, + { + "epoch": 0.3355048859934853, + "grad_norm": 2.3447575030783283, + "learning_rate": 1.992817704290516e-05, + "loss": 1.2025, + "step": 618 + }, + { + "epoch": 0.3360477741585233, + "grad_norm": 3.027039186876229, + "learning_rate": 1.99277556849139e-05, + "loss": 1.7133, + "step": 619 + }, + { + "epoch": 0.33659066232356133, + "grad_norm": 2.9868315770124343, + "learning_rate": 1.9927333099041992e-05, + "loss": 1.2309, + "step": 620 + }, + { + "epoch": 0.33713355048859933, + "grad_norm": 3.215889472723888, + "learning_rate": 1.9926909285341705e-05, + "loss": 1.2217, + "step": 621 + }, + { + "epoch": 0.33767643865363733, + "grad_norm": 3.293119071439456, + "learning_rate": 1.992648424386546e-05, + "loss": 1.7828, + "step": 622 + }, + { + "epoch": 0.33821932681867534, + "grad_norm": 2.7023629491691623, + "learning_rate": 1.992605797466582e-05, + "loss": 1.3745, + "step": 623 + }, + { + "epoch": 0.33876221498371334, + "grad_norm": 4.234534944070885, + "learning_rate": 1.9925630477795514e-05, + "loss": 1.3415, + "step": 624 + }, + { + "epoch": 0.33930510314875134, + "grad_norm": 3.6994706259321815, + "learning_rate": 1.9925201753307406e-05, + "loss": 1.4234, + "step": 625 + }, + { + "epoch": 0.33984799131378934, + "grad_norm": 3.156992028368943, + "learning_rate": 1.992477180125453e-05, + "loss": 1.4845, + "step": 626 + }, + { + "epoch": 0.34039087947882735, + "grad_norm": 3.1286393814611397, + "learning_rate": 1.9924340621690058e-05, + "loss": 0.9685, + "step": 627 + }, + { + "epoch": 0.34093376764386535, + "grad_norm": 3.336970766050665, + "learning_rate": 1.9923908214667323e-05, + "loss": 1.3268, + "step": 628 + }, + { + "epoch": 0.34147665580890335, + "grad_norm": 4.290496076646803, + "learning_rate": 1.99234745802398e-05, + "loss": 1.2785, + "step": 629 + }, + { + "epoch": 0.34201954397394135, + "grad_norm": 2.7491777959300023, + "learning_rate": 1.9923039718461127e-05, + "loss": 1.3798, + "step": 630 + }, + { + "epoch": 0.34256243213897936, + "grad_norm": 5.312529540414799, + "learning_rate": 1.992260362938509e-05, + "loss": 2.2512, + "step": 631 + }, + { + "epoch": 0.34310532030401736, + "grad_norm": 3.9557409230221676, + "learning_rate": 1.9922166313065618e-05, + "loss": 1.849, + "step": 632 + }, + { + "epoch": 0.34364820846905536, + "grad_norm": 3.7327574664165786, + "learning_rate": 1.9921727769556805e-05, + "loss": 1.4078, + "step": 633 + }, + { + "epoch": 0.34419109663409336, + "grad_norm": 3.7644055867642594, + "learning_rate": 1.992128799891289e-05, + "loss": 1.714, + "step": 634 + }, + { + "epoch": 0.34473398479913137, + "grad_norm": 3.923952524278675, + "learning_rate": 1.9920847001188258e-05, + "loss": 1.7613, + "step": 635 + }, + { + "epoch": 0.34527687296416937, + "grad_norm": 3.1523646374054155, + "learning_rate": 1.992040477643746e-05, + "loss": 1.2563, + "step": 636 + }, + { + "epoch": 0.34581976112920737, + "grad_norm": 3.6157151039083444, + "learning_rate": 1.991996132471519e-05, + "loss": 1.1795, + "step": 637 + }, + { + "epoch": 0.3463626492942454, + "grad_norm": 4.222007576307266, + "learning_rate": 1.991951664607629e-05, + "loss": 1.6605, + "step": 638 + }, + { + "epoch": 0.3469055374592834, + "grad_norm": 3.1555785318795544, + "learning_rate": 1.9919070740575764e-05, + "loss": 1.526, + "step": 639 + }, + { + "epoch": 0.3474484256243214, + "grad_norm": 3.2645269107818677, + "learning_rate": 1.991862360826876e-05, + "loss": 1.3182, + "step": 640 + }, + { + "epoch": 0.3479913137893594, + "grad_norm": 3.53044955128402, + "learning_rate": 1.991817524921058e-05, + "loss": 1.5583, + "step": 641 + }, + { + "epoch": 0.3485342019543974, + "grad_norm": 3.24710260787145, + "learning_rate": 1.9917725663456678e-05, + "loss": 0.9878, + "step": 642 + }, + { + "epoch": 0.3490770901194354, + "grad_norm": 3.9154742975791654, + "learning_rate": 1.991727485106266e-05, + "loss": 1.6034, + "step": 643 + }, + { + "epoch": 0.3496199782844734, + "grad_norm": 3.8199965805178318, + "learning_rate": 1.9916822812084282e-05, + "loss": 1.3768, + "step": 644 + }, + { + "epoch": 0.3501628664495114, + "grad_norm": 3.198793670499035, + "learning_rate": 1.9916369546577455e-05, + "loss": 1.4308, + "step": 645 + }, + { + "epoch": 0.3507057546145494, + "grad_norm": 2.908177540585374, + "learning_rate": 1.9915915054598237e-05, + "loss": 1.3964, + "step": 646 + }, + { + "epoch": 0.3512486427795874, + "grad_norm": 2.920889003436479, + "learning_rate": 1.9915459336202844e-05, + "loss": 1.056, + "step": 647 + }, + { + "epoch": 0.3517915309446254, + "grad_norm": 3.7046400158340864, + "learning_rate": 1.991500239144763e-05, + "loss": 1.9052, + "step": 648 + }, + { + "epoch": 0.3523344191096634, + "grad_norm": 4.412988121206581, + "learning_rate": 1.9914544220389124e-05, + "loss": 1.944, + "step": 649 + }, + { + "epoch": 0.3528773072747014, + "grad_norm": 3.87124697251994, + "learning_rate": 1.9914084823083988e-05, + "loss": 1.4951, + "step": 650 + }, + { + "epoch": 0.3534201954397394, + "grad_norm": 3.8734663848108584, + "learning_rate": 1.9913624199589037e-05, + "loss": 1.5462, + "step": 651 + }, + { + "epoch": 0.3539630836047774, + "grad_norm": 2.7283359997443126, + "learning_rate": 1.9913162349961248e-05, + "loss": 0.9188, + "step": 652 + }, + { + "epoch": 0.3545059717698154, + "grad_norm": 2.885333878853398, + "learning_rate": 1.991269927425774e-05, + "loss": 1.2292, + "step": 653 + }, + { + "epoch": 0.3550488599348534, + "grad_norm": 3.2425772460053257, + "learning_rate": 1.9912234972535788e-05, + "loss": 1.2863, + "step": 654 + }, + { + "epoch": 0.3555917480998914, + "grad_norm": 3.060966260943164, + "learning_rate": 1.991176944485281e-05, + "loss": 1.5056, + "step": 655 + }, + { + "epoch": 0.3561346362649294, + "grad_norm": 4.533553635387312, + "learning_rate": 1.99113026912664e-05, + "loss": 1.8229, + "step": 656 + }, + { + "epoch": 0.3566775244299674, + "grad_norm": 2.8582150809054045, + "learning_rate": 1.9910834711834267e-05, + "loss": 1.3233, + "step": 657 + }, + { + "epoch": 0.3572204125950054, + "grad_norm": 4.930680768753159, + "learning_rate": 1.9910365506614308e-05, + "loss": 1.5997, + "step": 658 + }, + { + "epoch": 0.3577633007600434, + "grad_norm": 3.0733335483268083, + "learning_rate": 1.9909895075664545e-05, + "loss": 0.9206, + "step": 659 + }, + { + "epoch": 0.3583061889250814, + "grad_norm": 3.4659704922712686, + "learning_rate": 1.990942341904317e-05, + "loss": 1.2752, + "step": 660 + }, + { + "epoch": 0.35884907709011943, + "grad_norm": 3.493850558129149, + "learning_rate": 1.9908950536808508e-05, + "loss": 1.588, + "step": 661 + }, + { + "epoch": 0.35939196525515743, + "grad_norm": 3.61256229272583, + "learning_rate": 1.9908476429019056e-05, + "loss": 1.469, + "step": 662 + }, + { + "epoch": 0.35993485342019543, + "grad_norm": 3.3430037514089803, + "learning_rate": 1.9908001095733445e-05, + "loss": 1.3038, + "step": 663 + }, + { + "epoch": 0.36047774158523344, + "grad_norm": 3.8918915484374557, + "learning_rate": 1.9907524537010467e-05, + "loss": 1.3683, + "step": 664 + }, + { + "epoch": 0.36102062975027144, + "grad_norm": 3.6032272045245053, + "learning_rate": 1.9907046752909064e-05, + "loss": 1.8694, + "step": 665 + }, + { + "epoch": 0.36156351791530944, + "grad_norm": 4.546268377661869, + "learning_rate": 1.9906567743488326e-05, + "loss": 1.2871, + "step": 666 + }, + { + "epoch": 0.36210640608034744, + "grad_norm": 3.5283169402860777, + "learning_rate": 1.9906087508807504e-05, + "loss": 1.5334, + "step": 667 + }, + { + "epoch": 0.36264929424538545, + "grad_norm": 2.7984191262784504, + "learning_rate": 1.9905606048925993e-05, + "loss": 1.1924, + "step": 668 + }, + { + "epoch": 0.36319218241042345, + "grad_norm": 3.54703835950394, + "learning_rate": 1.9905123363903335e-05, + "loss": 1.5972, + "step": 669 + }, + { + "epoch": 0.36373507057546145, + "grad_norm": 3.0126938064772855, + "learning_rate": 1.9904639453799236e-05, + "loss": 1.2294, + "step": 670 + }, + { + "epoch": 0.36427795874049945, + "grad_norm": 3.16358553989926, + "learning_rate": 1.990415431867354e-05, + "loss": 1.3564, + "step": 671 + }, + { + "epoch": 0.36482084690553745, + "grad_norm": 4.118910539945259, + "learning_rate": 1.990366795858626e-05, + "loss": 1.6838, + "step": 672 + }, + { + "epoch": 0.36536373507057546, + "grad_norm": 3.8466533902335502, + "learning_rate": 1.9903180373597534e-05, + "loss": 1.7986, + "step": 673 + }, + { + "epoch": 0.36590662323561346, + "grad_norm": 3.6195384409794684, + "learning_rate": 1.990269156376768e-05, + "loss": 1.6113, + "step": 674 + }, + { + "epoch": 0.36644951140065146, + "grad_norm": 3.509724379422402, + "learning_rate": 1.9902201529157152e-05, + "loss": 1.5496, + "step": 675 + }, + { + "epoch": 0.36699239956568946, + "grad_norm": 3.579737463592409, + "learning_rate": 1.9901710269826554e-05, + "loss": 1.4856, + "step": 676 + }, + { + "epoch": 0.36753528773072747, + "grad_norm": 3.780431786449365, + "learning_rate": 1.9901217785836655e-05, + "loss": 1.8519, + "step": 677 + }, + { + "epoch": 0.36807817589576547, + "grad_norm": 4.0659352627131735, + "learning_rate": 1.9900724077248354e-05, + "loss": 1.4382, + "step": 678 + }, + { + "epoch": 0.36862106406080347, + "grad_norm": 2.9524656376446274, + "learning_rate": 1.9900229144122723e-05, + "loss": 1.4541, + "step": 679 + }, + { + "epoch": 0.3691639522258415, + "grad_norm": 2.388269642575342, + "learning_rate": 1.989973298652097e-05, + "loss": 0.9567, + "step": 680 + }, + { + "epoch": 0.3697068403908795, + "grad_norm": 2.8322419431529453, + "learning_rate": 1.9899235604504467e-05, + "loss": 1.1187, + "step": 681 + }, + { + "epoch": 0.3702497285559175, + "grad_norm": 3.0469435898757613, + "learning_rate": 1.9898736998134726e-05, + "loss": 1.5571, + "step": 682 + }, + { + "epoch": 0.3707926167209555, + "grad_norm": 3.623982705749655, + "learning_rate": 1.9898237167473416e-05, + "loss": 1.2047, + "step": 683 + }, + { + "epoch": 0.3713355048859935, + "grad_norm": 3.669362267695381, + "learning_rate": 1.9897736112582357e-05, + "loss": 1.7747, + "step": 684 + }, + { + "epoch": 0.3718783930510315, + "grad_norm": 2.9109325710576353, + "learning_rate": 1.989723383352352e-05, + "loss": 1.0044, + "step": 685 + }, + { + "epoch": 0.3724212812160695, + "grad_norm": 3.4478434843760146, + "learning_rate": 1.9896730330359032e-05, + "loss": 1.1245, + "step": 686 + }, + { + "epoch": 0.3729641693811075, + "grad_norm": 3.2957514959337275, + "learning_rate": 1.989622560315116e-05, + "loss": 1.3243, + "step": 687 + }, + { + "epoch": 0.3735070575461455, + "grad_norm": 3.5500423086355988, + "learning_rate": 1.989571965196234e-05, + "loss": 2.0478, + "step": 688 + }, + { + "epoch": 0.3740499457111835, + "grad_norm": 2.621987663839103, + "learning_rate": 1.9895212476855136e-05, + "loss": 1.3135, + "step": 689 + }, + { + "epoch": 0.3745928338762215, + "grad_norm": 4.055390555970487, + "learning_rate": 1.989470407789228e-05, + "loss": 1.5165, + "step": 690 + }, + { + "epoch": 0.3751357220412595, + "grad_norm": 2.94204110816024, + "learning_rate": 1.989419445513666e-05, + "loss": 1.3458, + "step": 691 + }, + { + "epoch": 0.3756786102062975, + "grad_norm": 3.5714132672245977, + "learning_rate": 1.98936836086513e-05, + "loss": 1.3034, + "step": 692 + }, + { + "epoch": 0.3762214983713355, + "grad_norm": 3.659137782783938, + "learning_rate": 1.9893171538499382e-05, + "loss": 1.6203, + "step": 693 + }, + { + "epoch": 0.3767643865363735, + "grad_norm": 5.116097300755018, + "learning_rate": 1.9892658244744236e-05, + "loss": 2.1071, + "step": 694 + }, + { + "epoch": 0.3773072747014115, + "grad_norm": 2.8449662018307005, + "learning_rate": 1.9892143727449357e-05, + "loss": 1.0477, + "step": 695 + }, + { + "epoch": 0.3778501628664495, + "grad_norm": 4.021030963638319, + "learning_rate": 1.989162798667838e-05, + "loss": 1.9528, + "step": 696 + }, + { + "epoch": 0.3783930510314875, + "grad_norm": 3.377368666687089, + "learning_rate": 1.989111102249508e-05, + "loss": 1.2481, + "step": 697 + }, + { + "epoch": 0.3789359391965255, + "grad_norm": 3.403268816169458, + "learning_rate": 1.9890592834963406e-05, + "loss": 1.1864, + "step": 698 + }, + { + "epoch": 0.3794788273615635, + "grad_norm": 3.082879601892987, + "learning_rate": 1.9890073424147453e-05, + "loss": 1.7365, + "step": 699 + }, + { + "epoch": 0.3800217155266015, + "grad_norm": 4.122610250215441, + "learning_rate": 1.988955279011145e-05, + "loss": 1.4139, + "step": 700 + }, + { + "epoch": 0.3805646036916395, + "grad_norm": 3.298087120576001, + "learning_rate": 1.98890309329198e-05, + "loss": 1.5904, + "step": 701 + }, + { + "epoch": 0.3811074918566775, + "grad_norm": 3.1183850173888703, + "learning_rate": 1.9888507852637043e-05, + "loss": 1.2146, + "step": 702 + }, + { + "epoch": 0.38165038002171553, + "grad_norm": 3.128042971411966, + "learning_rate": 1.9887983549327873e-05, + "loss": 1.2751, + "step": 703 + }, + { + "epoch": 0.38219326818675353, + "grad_norm": 3.362555156234201, + "learning_rate": 1.988745802305714e-05, + "loss": 1.1363, + "step": 704 + }, + { + "epoch": 0.38273615635179153, + "grad_norm": 3.29574344438116, + "learning_rate": 1.988693127388984e-05, + "loss": 1.227, + "step": 705 + }, + { + "epoch": 0.38327904451682954, + "grad_norm": 4.2418853227437205, + "learning_rate": 1.9886403301891123e-05, + "loss": 1.7091, + "step": 706 + }, + { + "epoch": 0.38382193268186754, + "grad_norm": 4.165071133964158, + "learning_rate": 1.9885874107126287e-05, + "loss": 1.9403, + "step": 707 + }, + { + "epoch": 0.38436482084690554, + "grad_norm": 4.639609386348944, + "learning_rate": 1.9885343689660787e-05, + "loss": 1.221, + "step": 708 + }, + { + "epoch": 0.38490770901194354, + "grad_norm": 3.0746775894794034, + "learning_rate": 1.9884812049560226e-05, + "loss": 1.4958, + "step": 709 + }, + { + "epoch": 0.38545059717698155, + "grad_norm": 4.822804859325608, + "learning_rate": 1.9884279186890357e-05, + "loss": 1.6928, + "step": 710 + }, + { + "epoch": 0.38599348534201955, + "grad_norm": 4.299811491105524, + "learning_rate": 1.9883745101717084e-05, + "loss": 1.3075, + "step": 711 + }, + { + "epoch": 0.38653637350705755, + "grad_norm": 3.882170994492157, + "learning_rate": 1.9883209794106464e-05, + "loss": 1.9712, + "step": 712 + }, + { + "epoch": 0.38707926167209555, + "grad_norm": 4.164019029255246, + "learning_rate": 1.9882673264124705e-05, + "loss": 1.6109, + "step": 713 + }, + { + "epoch": 0.38762214983713356, + "grad_norm": 3.691226317297577, + "learning_rate": 1.9882135511838167e-05, + "loss": 1.3225, + "step": 714 + }, + { + "epoch": 0.38816503800217156, + "grad_norm": 5.777282505586183, + "learning_rate": 1.988159653731336e-05, + "loss": 1.7945, + "step": 715 + }, + { + "epoch": 0.38870792616720956, + "grad_norm": 3.3926512133724165, + "learning_rate": 1.9881056340616944e-05, + "loss": 1.5797, + "step": 716 + }, + { + "epoch": 0.38925081433224756, + "grad_norm": 3.7756965718668467, + "learning_rate": 1.988051492181573e-05, + "loss": 1.5515, + "step": 717 + }, + { + "epoch": 0.38979370249728557, + "grad_norm": 3.5509658648264613, + "learning_rate": 1.987997228097668e-05, + "loss": 1.2004, + "step": 718 + }, + { + "epoch": 0.39033659066232357, + "grad_norm": 4.682493032945398, + "learning_rate": 1.987942841816692e-05, + "loss": 1.2957, + "step": 719 + }, + { + "epoch": 0.39087947882736157, + "grad_norm": 2.923319971884201, + "learning_rate": 1.9878883333453704e-05, + "loss": 0.753, + "step": 720 + }, + { + "epoch": 0.3914223669923996, + "grad_norm": 2.8706465146507845, + "learning_rate": 1.987833702690445e-05, + "loss": 1.1057, + "step": 721 + }, + { + "epoch": 0.3919652551574376, + "grad_norm": 3.4516078574834057, + "learning_rate": 1.987778949858673e-05, + "loss": 1.1275, + "step": 722 + }, + { + "epoch": 0.3925081433224756, + "grad_norm": 3.3400553957950567, + "learning_rate": 1.9877240748568263e-05, + "loss": 1.1538, + "step": 723 + }, + { + "epoch": 0.3930510314875136, + "grad_norm": 3.155540544638446, + "learning_rate": 1.987669077691692e-05, + "loss": 1.2486, + "step": 724 + }, + { + "epoch": 0.3935939196525516, + "grad_norm": 3.1450276486107054, + "learning_rate": 1.987613958370072e-05, + "loss": 1.265, + "step": 725 + }, + { + "epoch": 0.3941368078175896, + "grad_norm": 3.0493327058864406, + "learning_rate": 1.9875587168987834e-05, + "loss": 1.2097, + "step": 726 + }, + { + "epoch": 0.3946796959826276, + "grad_norm": 3.716259137065701, + "learning_rate": 1.987503353284659e-05, + "loss": 1.5386, + "step": 727 + }, + { + "epoch": 0.3952225841476656, + "grad_norm": 3.441725300410296, + "learning_rate": 1.9874478675345458e-05, + "loss": 1.3936, + "step": 728 + }, + { + "epoch": 0.3957654723127036, + "grad_norm": 3.773353353497139, + "learning_rate": 1.9873922596553067e-05, + "loss": 1.731, + "step": 729 + }, + { + "epoch": 0.3963083604777416, + "grad_norm": 3.527645597828566, + "learning_rate": 1.987336529653819e-05, + "loss": 1.708, + "step": 730 + }, + { + "epoch": 0.3968512486427796, + "grad_norm": 3.8440953147603643, + "learning_rate": 1.9872806775369762e-05, + "loss": 1.6102, + "step": 731 + }, + { + "epoch": 0.3973941368078176, + "grad_norm": 3.4036622029265966, + "learning_rate": 1.9872247033116855e-05, + "loss": 0.9298, + "step": 732 + }, + { + "epoch": 0.3979370249728556, + "grad_norm": 4.034027632823603, + "learning_rate": 1.98716860698487e-05, + "loss": 1.78, + "step": 733 + }, + { + "epoch": 0.3984799131378936, + "grad_norm": 4.10524690369694, + "learning_rate": 1.987112388563468e-05, + "loss": 1.2354, + "step": 734 + }, + { + "epoch": 0.3990228013029316, + "grad_norm": 4.059289261823474, + "learning_rate": 1.9870560480544325e-05, + "loss": 1.804, + "step": 735 + }, + { + "epoch": 0.3995656894679696, + "grad_norm": 3.2544839104292667, + "learning_rate": 1.986999585464732e-05, + "loss": 1.1228, + "step": 736 + }, + { + "epoch": 0.4001085776330076, + "grad_norm": 3.741117309135591, + "learning_rate": 1.9869430008013496e-05, + "loss": 1.329, + "step": 737 + }, + { + "epoch": 0.4006514657980456, + "grad_norm": 4.446330579981585, + "learning_rate": 1.9868862940712838e-05, + "loss": 1.6506, + "step": 738 + }, + { + "epoch": 0.4011943539630836, + "grad_norm": 3.139296317304318, + "learning_rate": 1.9868294652815483e-05, + "loss": 1.144, + "step": 739 + }, + { + "epoch": 0.4017372421281216, + "grad_norm": 4.025270331554587, + "learning_rate": 1.986772514439172e-05, + "loss": 1.336, + "step": 740 + }, + { + "epoch": 0.4022801302931596, + "grad_norm": 3.4545518889756557, + "learning_rate": 1.986715441551198e-05, + "loss": 1.6006, + "step": 741 + }, + { + "epoch": 0.4028230184581976, + "grad_norm": 3.748598610621174, + "learning_rate": 1.986658246624686e-05, + "loss": 1.4812, + "step": 742 + }, + { + "epoch": 0.4033659066232356, + "grad_norm": 5.126058311631635, + "learning_rate": 1.9866009296667093e-05, + "loss": 2.0129, + "step": 743 + }, + { + "epoch": 0.40390879478827363, + "grad_norm": 8.145692299678405, + "learning_rate": 1.9865434906843574e-05, + "loss": 1.8321, + "step": 744 + }, + { + "epoch": 0.40445168295331163, + "grad_norm": 4.23918362592579, + "learning_rate": 1.9864859296847343e-05, + "loss": 1.5765, + "step": 745 + }, + { + "epoch": 0.40499457111834963, + "grad_norm": 4.85145039296263, + "learning_rate": 1.986428246674959e-05, + "loss": 1.6486, + "step": 746 + }, + { + "epoch": 0.40553745928338764, + "grad_norm": 3.7681072947070415, + "learning_rate": 1.986370441662166e-05, + "loss": 1.4466, + "step": 747 + }, + { + "epoch": 0.40608034744842564, + "grad_norm": 4.206136090938125, + "learning_rate": 1.986312514653505e-05, + "loss": 1.6717, + "step": 748 + }, + { + "epoch": 0.40662323561346364, + "grad_norm": 3.7336223866799965, + "learning_rate": 1.9862544656561403e-05, + "loss": 0.9599, + "step": 749 + }, + { + "epoch": 0.40716612377850164, + "grad_norm": 4.037759008136602, + "learning_rate": 1.986196294677251e-05, + "loss": 1.9387, + "step": 750 + }, + { + "epoch": 0.40770901194353965, + "grad_norm": 3.8832525632606156, + "learning_rate": 1.9861380017240324e-05, + "loss": 2.0184, + "step": 751 + }, + { + "epoch": 0.40825190010857765, + "grad_norm": 3.9521376069923546, + "learning_rate": 1.986079586803694e-05, + "loss": 1.0926, + "step": 752 + }, + { + "epoch": 0.40879478827361565, + "grad_norm": 5.128973903987411, + "learning_rate": 1.986021049923461e-05, + "loss": 1.5943, + "step": 753 + }, + { + "epoch": 0.40933767643865365, + "grad_norm": 4.5372892013581865, + "learning_rate": 1.9859623910905728e-05, + "loss": 1.643, + "step": 754 + }, + { + "epoch": 0.40988056460369166, + "grad_norm": 4.2137978897196415, + "learning_rate": 1.985903610312285e-05, + "loss": 1.7237, + "step": 755 + }, + { + "epoch": 0.41042345276872966, + "grad_norm": 2.713265124864733, + "learning_rate": 1.985844707595867e-05, + "loss": 1.0521, + "step": 756 + }, + { + "epoch": 0.41096634093376766, + "grad_norm": 3.362872283003281, + "learning_rate": 1.9857856829486045e-05, + "loss": 1.2879, + "step": 757 + }, + { + "epoch": 0.41150922909880566, + "grad_norm": 4.022235417044952, + "learning_rate": 1.9857265363777975e-05, + "loss": 1.398, + "step": 758 + }, + { + "epoch": 0.41205211726384366, + "grad_norm": 4.083280210123792, + "learning_rate": 1.9856672678907616e-05, + "loss": 1.6652, + "step": 759 + }, + { + "epoch": 0.41259500542888167, + "grad_norm": 4.469771898345589, + "learning_rate": 1.985607877494827e-05, + "loss": 1.6242, + "step": 760 + }, + { + "epoch": 0.41313789359391967, + "grad_norm": 3.0490808076449016, + "learning_rate": 1.9855483651973396e-05, + "loss": 1.0783, + "step": 761 + }, + { + "epoch": 0.41368078175895767, + "grad_norm": 3.488230005104396, + "learning_rate": 1.9854887310056593e-05, + "loss": 1.3953, + "step": 762 + }, + { + "epoch": 0.4142236699239957, + "grad_norm": 3.5610333355465653, + "learning_rate": 1.9854289749271624e-05, + "loss": 1.3692, + "step": 763 + }, + { + "epoch": 0.4147665580890337, + "grad_norm": 3.5572639096784533, + "learning_rate": 1.9853690969692393e-05, + "loss": 1.646, + "step": 764 + }, + { + "epoch": 0.4153094462540717, + "grad_norm": 3.811741925703646, + "learning_rate": 1.9853090971392953e-05, + "loss": 1.6637, + "step": 765 + }, + { + "epoch": 0.4158523344191097, + "grad_norm": 2.912609699248709, + "learning_rate": 1.9852489754447526e-05, + "loss": 0.9086, + "step": 766 + }, + { + "epoch": 0.4163952225841477, + "grad_norm": 3.7608762982374646, + "learning_rate": 1.985188731893046e-05, + "loss": 1.5131, + "step": 767 + }, + { + "epoch": 0.4169381107491857, + "grad_norm": 3.776000457813264, + "learning_rate": 1.985128366491627e-05, + "loss": 1.5929, + "step": 768 + }, + { + "epoch": 0.4174809989142237, + "grad_norm": 4.129873624081247, + "learning_rate": 1.9850678792479613e-05, + "loss": 1.5461, + "step": 769 + }, + { + "epoch": 0.4180238870792617, + "grad_norm": 3.2233425858963645, + "learning_rate": 1.9850072701695306e-05, + "loss": 0.9614, + "step": 770 + }, + { + "epoch": 0.4185667752442997, + "grad_norm": 3.8215876727885054, + "learning_rate": 1.984946539263831e-05, + "loss": 1.2549, + "step": 771 + }, + { + "epoch": 0.4191096634093377, + "grad_norm": 3.0062072101032125, + "learning_rate": 1.9848856865383732e-05, + "loss": 0.9796, + "step": 772 + }, + { + "epoch": 0.4196525515743757, + "grad_norm": 2.937513473620578, + "learning_rate": 1.984824712000684e-05, + "loss": 1.0204, + "step": 773 + }, + { + "epoch": 0.4201954397394137, + "grad_norm": 2.949279428268976, + "learning_rate": 1.984763615658305e-05, + "loss": 1.332, + "step": 774 + }, + { + "epoch": 0.4207383279044517, + "grad_norm": 3.455520053658832, + "learning_rate": 1.9847023975187925e-05, + "loss": 1.3786, + "step": 775 + }, + { + "epoch": 0.4212812160694897, + "grad_norm": 3.8352233275650733, + "learning_rate": 1.9846410575897183e-05, + "loss": 1.4897, + "step": 776 + }, + { + "epoch": 0.4218241042345277, + "grad_norm": 3.262791090623995, + "learning_rate": 1.984579595878669e-05, + "loss": 1.5097, + "step": 777 + }, + { + "epoch": 0.4223669923995657, + "grad_norm": 3.603809745585289, + "learning_rate": 1.9845180123932456e-05, + "loss": 1.8419, + "step": 778 + }, + { + "epoch": 0.4229098805646037, + "grad_norm": 3.9648489260279196, + "learning_rate": 1.9844563071410656e-05, + "loss": 1.5159, + "step": 779 + }, + { + "epoch": 0.4234527687296417, + "grad_norm": 4.576987136894564, + "learning_rate": 1.9843944801297605e-05, + "loss": 1.6866, + "step": 780 + }, + { + "epoch": 0.4239956568946797, + "grad_norm": 3.4164331578089406, + "learning_rate": 1.9843325313669774e-05, + "loss": 1.5533, + "step": 781 + }, + { + "epoch": 0.4245385450597177, + "grad_norm": 3.2121441901059367, + "learning_rate": 1.9842704608603774e-05, + "loss": 0.8879, + "step": 782 + }, + { + "epoch": 0.4250814332247557, + "grad_norm": 3.6213877563860537, + "learning_rate": 1.9842082686176388e-05, + "loss": 1.3098, + "step": 783 + }, + { + "epoch": 0.4256243213897937, + "grad_norm": 3.719993086277978, + "learning_rate": 1.9841459546464527e-05, + "loss": 1.331, + "step": 784 + }, + { + "epoch": 0.4261672095548317, + "grad_norm": 2.7952357248876516, + "learning_rate": 1.9840835189545266e-05, + "loss": 1.1298, + "step": 785 + }, + { + "epoch": 0.42671009771986973, + "grad_norm": 3.450597586563723, + "learning_rate": 1.9840209615495822e-05, + "loss": 1.0023, + "step": 786 + }, + { + "epoch": 0.42725298588490773, + "grad_norm": 4.338480682869939, + "learning_rate": 1.983958282439357e-05, + "loss": 1.7044, + "step": 787 + }, + { + "epoch": 0.42779587404994573, + "grad_norm": 3.4831324579282765, + "learning_rate": 1.983895481631603e-05, + "loss": 1.4948, + "step": 788 + }, + { + "epoch": 0.42833876221498374, + "grad_norm": 6.382810905837672, + "learning_rate": 1.9838325591340885e-05, + "loss": 2.3008, + "step": 789 + }, + { + "epoch": 0.42888165038002174, + "grad_norm": 3.063848408601439, + "learning_rate": 1.9837695149545945e-05, + "loss": 1.2045, + "step": 790 + }, + { + "epoch": 0.42942453854505974, + "grad_norm": 4.80361465821882, + "learning_rate": 1.9837063491009193e-05, + "loss": 1.2552, + "step": 791 + }, + { + "epoch": 0.42996742671009774, + "grad_norm": 4.16884813133452, + "learning_rate": 1.9836430615808745e-05, + "loss": 1.6618, + "step": 792 + }, + { + "epoch": 0.43051031487513575, + "grad_norm": 4.215747683493222, + "learning_rate": 1.9835796524022886e-05, + "loss": 1.2571, + "step": 793 + }, + { + "epoch": 0.43105320304017375, + "grad_norm": 4.894853560761769, + "learning_rate": 1.9835161215730038e-05, + "loss": 0.793, + "step": 794 + }, + { + "epoch": 0.43159609120521175, + "grad_norm": 3.264854272979086, + "learning_rate": 1.983452469100877e-05, + "loss": 1.0721, + "step": 795 + }, + { + "epoch": 0.43213897937024975, + "grad_norm": 3.2346237132359335, + "learning_rate": 1.9833886949937823e-05, + "loss": 1.2396, + "step": 796 + }, + { + "epoch": 0.43268186753528776, + "grad_norm": 3.7265266370894023, + "learning_rate": 1.9833247992596058e-05, + "loss": 1.8845, + "step": 797 + }, + { + "epoch": 0.43322475570032576, + "grad_norm": 3.0655010730082126, + "learning_rate": 1.9832607819062513e-05, + "loss": 1.556, + "step": 798 + }, + { + "epoch": 0.43376764386536376, + "grad_norm": 3.209417077142042, + "learning_rate": 1.983196642941636e-05, + "loss": 1.2183, + "step": 799 + }, + { + "epoch": 0.43431053203040176, + "grad_norm": 4.384697511143271, + "learning_rate": 1.9831323823736933e-05, + "loss": 2.0889, + "step": 800 + }, + { + "epoch": 0.43485342019543977, + "grad_norm": 3.2773759507081084, + "learning_rate": 1.9830680002103703e-05, + "loss": 0.919, + "step": 801 + }, + { + "epoch": 0.43539630836047777, + "grad_norm": 4.1436406822638805, + "learning_rate": 1.9830034964596304e-05, + "loss": 1.7759, + "step": 802 + }, + { + "epoch": 0.43593919652551577, + "grad_norm": 5.715436881791243, + "learning_rate": 1.9829388711294512e-05, + "loss": 1.4277, + "step": 803 + }, + { + "epoch": 0.4364820846905538, + "grad_norm": 3.166985124043999, + "learning_rate": 1.982874124227826e-05, + "loss": 1.56, + "step": 804 + }, + { + "epoch": 0.4370249728555918, + "grad_norm": 4.237249076660815, + "learning_rate": 1.9828092557627626e-05, + "loss": 1.2366, + "step": 805 + }, + { + "epoch": 0.4375678610206297, + "grad_norm": 3.073398749866342, + "learning_rate": 1.982744265742284e-05, + "loss": 1.1475, + "step": 806 + }, + { + "epoch": 0.4381107491856677, + "grad_norm": 3.356777818996544, + "learning_rate": 1.9826791541744285e-05, + "loss": 1.2666, + "step": 807 + }, + { + "epoch": 0.4386536373507057, + "grad_norm": 3.171383453112648, + "learning_rate": 1.982613921067249e-05, + "loss": 1.1407, + "step": 808 + }, + { + "epoch": 0.43919652551574373, + "grad_norm": 3.334493176451811, + "learning_rate": 1.9825485664288138e-05, + "loss": 1.3309, + "step": 809 + }, + { + "epoch": 0.43973941368078173, + "grad_norm": 3.199045603289457, + "learning_rate": 1.982483090267206e-05, + "loss": 1.1634, + "step": 810 + }, + { + "epoch": 0.44028230184581973, + "grad_norm": 2.9970299057571856, + "learning_rate": 1.9824174925905235e-05, + "loss": 0.9642, + "step": 811 + }, + { + "epoch": 0.44082519001085774, + "grad_norm": 3.9851247078132994, + "learning_rate": 1.98235177340688e-05, + "loss": 1.4926, + "step": 812 + }, + { + "epoch": 0.44136807817589574, + "grad_norm": 3.586604601363014, + "learning_rate": 1.9822859327244034e-05, + "loss": 1.3105, + "step": 813 + }, + { + "epoch": 0.44191096634093374, + "grad_norm": 4.198634753759357, + "learning_rate": 1.9822199705512372e-05, + "loss": 1.4437, + "step": 814 + }, + { + "epoch": 0.44245385450597174, + "grad_norm": 3.6499375531563096, + "learning_rate": 1.9821538868955394e-05, + "loss": 1.2063, + "step": 815 + }, + { + "epoch": 0.44299674267100975, + "grad_norm": 4.3973905860052716, + "learning_rate": 1.9820876817654836e-05, + "loss": 1.5002, + "step": 816 + }, + { + "epoch": 0.44353963083604775, + "grad_norm": 2.7871960857121287, + "learning_rate": 1.9820213551692585e-05, + "loss": 0.6589, + "step": 817 + }, + { + "epoch": 0.44408251900108575, + "grad_norm": 3.4126665543063184, + "learning_rate": 1.981954907115067e-05, + "loss": 1.1651, + "step": 818 + }, + { + "epoch": 0.44462540716612375, + "grad_norm": 4.229651941034759, + "learning_rate": 1.981888337611127e-05, + "loss": 1.6444, + "step": 819 + }, + { + "epoch": 0.44516829533116176, + "grad_norm": 3.4135993440364056, + "learning_rate": 1.981821646665673e-05, + "loss": 1.2614, + "step": 820 + }, + { + "epoch": 0.44571118349619976, + "grad_norm": 4.274078711060124, + "learning_rate": 1.9817548342869527e-05, + "loss": 1.66, + "step": 821 + }, + { + "epoch": 0.44625407166123776, + "grad_norm": 2.9406863812447903, + "learning_rate": 1.98168790048323e-05, + "loss": 1.0412, + "step": 822 + }, + { + "epoch": 0.44679695982627576, + "grad_norm": 4.300279758003301, + "learning_rate": 1.981620845262783e-05, + "loss": 1.3689, + "step": 823 + }, + { + "epoch": 0.44733984799131377, + "grad_norm": 3.130525530503507, + "learning_rate": 1.9815536686339056e-05, + "loss": 1.1771, + "step": 824 + }, + { + "epoch": 0.44788273615635177, + "grad_norm": 3.6606030046920597, + "learning_rate": 1.981486370604906e-05, + "loss": 1.0323, + "step": 825 + }, + { + "epoch": 0.44842562432138977, + "grad_norm": 3.495865966877518, + "learning_rate": 1.981418951184108e-05, + "loss": 1.4493, + "step": 826 + }, + { + "epoch": 0.4489685124864278, + "grad_norm": 3.621229661878418, + "learning_rate": 1.9813514103798498e-05, + "loss": 1.5392, + "step": 827 + }, + { + "epoch": 0.4495114006514658, + "grad_norm": 3.6798995423063485, + "learning_rate": 1.9812837482004853e-05, + "loss": 1.0593, + "step": 828 + }, + { + "epoch": 0.4500542888165038, + "grad_norm": 3.4514040085362954, + "learning_rate": 1.9812159646543824e-05, + "loss": 0.9262, + "step": 829 + }, + { + "epoch": 0.4505971769815418, + "grad_norm": 3.195454430712678, + "learning_rate": 1.9811480597499257e-05, + "loss": 1.1469, + "step": 830 + }, + { + "epoch": 0.4511400651465798, + "grad_norm": 4.590874916307442, + "learning_rate": 1.981080033495513e-05, + "loss": 1.6534, + "step": 831 + }, + { + "epoch": 0.4516829533116178, + "grad_norm": 3.668702667937113, + "learning_rate": 1.981011885899558e-05, + "loss": 1.0597, + "step": 832 + }, + { + "epoch": 0.4522258414766558, + "grad_norm": 2.7669639542025046, + "learning_rate": 1.98094361697049e-05, + "loss": 1.0387, + "step": 833 + }, + { + "epoch": 0.4527687296416938, + "grad_norm": 4.734510249253564, + "learning_rate": 1.9808752267167515e-05, + "loss": 1.6757, + "step": 834 + }, + { + "epoch": 0.4533116178067318, + "grad_norm": 9.006078348758516, + "learning_rate": 1.9808067151468018e-05, + "loss": 2.1559, + "step": 835 + }, + { + "epoch": 0.4538545059717698, + "grad_norm": 4.169764802563779, + "learning_rate": 1.9807380822691146e-05, + "loss": 1.2557, + "step": 836 + }, + { + "epoch": 0.4543973941368078, + "grad_norm": 3.970541817152741, + "learning_rate": 1.9806693280921783e-05, + "loss": 1.5586, + "step": 837 + }, + { + "epoch": 0.4549402823018458, + "grad_norm": 5.042448582457387, + "learning_rate": 1.9806004526244966e-05, + "loss": 1.3191, + "step": 838 + }, + { + "epoch": 0.4554831704668838, + "grad_norm": 3.7467475303782236, + "learning_rate": 1.9805314558745876e-05, + "loss": 1.3318, + "step": 839 + }, + { + "epoch": 0.4560260586319218, + "grad_norm": 4.214820131429537, + "learning_rate": 1.980462337850986e-05, + "loss": 1.3958, + "step": 840 + }, + { + "epoch": 0.4565689467969598, + "grad_norm": 3.9655409992603965, + "learning_rate": 1.980393098562239e-05, + "loss": 0.9978, + "step": 841 + }, + { + "epoch": 0.4571118349619978, + "grad_norm": 5.522638140748038, + "learning_rate": 1.980323738016912e-05, + "loss": 1.4151, + "step": 842 + }, + { + "epoch": 0.4576547231270358, + "grad_norm": 3.84359618797871, + "learning_rate": 1.9802542562235822e-05, + "loss": 1.3567, + "step": 843 + }, + { + "epoch": 0.4581976112920738, + "grad_norm": 5.115138943098397, + "learning_rate": 1.9801846531908437e-05, + "loss": 1.6437, + "step": 844 + }, + { + "epoch": 0.4587404994571118, + "grad_norm": 3.811847477591368, + "learning_rate": 1.9801149289273054e-05, + "loss": 1.2335, + "step": 845 + }, + { + "epoch": 0.4592833876221498, + "grad_norm": 4.4009848681180115, + "learning_rate": 1.98004508344159e-05, + "loss": 1.2147, + "step": 846 + }, + { + "epoch": 0.4598262757871878, + "grad_norm": 3.850991850744113, + "learning_rate": 1.9799751167423376e-05, + "loss": 1.4865, + "step": 847 + }, + { + "epoch": 0.4603691639522258, + "grad_norm": 3.1303042065802917, + "learning_rate": 1.9799050288382004e-05, + "loss": 1.4926, + "step": 848 + }, + { + "epoch": 0.4609120521172638, + "grad_norm": 4.225513214699549, + "learning_rate": 1.9798348197378475e-05, + "loss": 1.2044, + "step": 849 + }, + { + "epoch": 0.46145494028230183, + "grad_norm": 4.139979880818486, + "learning_rate": 1.9797644894499625e-05, + "loss": 1.5682, + "step": 850 + }, + { + "epoch": 0.46199782844733983, + "grad_norm": 4.414209735978709, + "learning_rate": 1.979694037983244e-05, + "loss": 1.9033, + "step": 851 + }, + { + "epoch": 0.46254071661237783, + "grad_norm": 5.079031167256292, + "learning_rate": 1.9796234653464057e-05, + "loss": 1.8745, + "step": 852 + }, + { + "epoch": 0.46308360477741584, + "grad_norm": 6.59971413476838, + "learning_rate": 1.9795527715481755e-05, + "loss": 1.2449, + "step": 853 + }, + { + "epoch": 0.46362649294245384, + "grad_norm": 3.863740127637264, + "learning_rate": 1.9794819565972973e-05, + "loss": 1.3912, + "step": 854 + }, + { + "epoch": 0.46416938110749184, + "grad_norm": 3.8578186204961327, + "learning_rate": 1.9794110205025302e-05, + "loss": 1.3487, + "step": 855 + }, + { + "epoch": 0.46471226927252984, + "grad_norm": 4.19751015099396, + "learning_rate": 1.9793399632726466e-05, + "loss": 1.129, + "step": 856 + }, + { + "epoch": 0.46525515743756785, + "grad_norm": 4.261483190159906, + "learning_rate": 1.979268784916436e-05, + "loss": 1.7298, + "step": 857 + }, + { + "epoch": 0.46579804560260585, + "grad_norm": 4.1774094914047435, + "learning_rate": 1.9791974854427008e-05, + "loss": 0.8716, + "step": 858 + }, + { + "epoch": 0.46634093376764385, + "grad_norm": 3.845723101011648, + "learning_rate": 1.9791260648602603e-05, + "loss": 0.9086, + "step": 859 + }, + { + "epoch": 0.46688382193268185, + "grad_norm": 3.4922208632687393, + "learning_rate": 1.9790545231779476e-05, + "loss": 1.3688, + "step": 860 + }, + { + "epoch": 0.46742671009771986, + "grad_norm": 2.976761133296452, + "learning_rate": 1.978982860404611e-05, + "loss": 0.7865, + "step": 861 + }, + { + "epoch": 0.46796959826275786, + "grad_norm": 3.934765402497862, + "learning_rate": 1.978911076549114e-05, + "loss": 1.2993, + "step": 862 + }, + { + "epoch": 0.46851248642779586, + "grad_norm": 3.029198772099951, + "learning_rate": 1.978839171620335e-05, + "loss": 1.1737, + "step": 863 + }, + { + "epoch": 0.46905537459283386, + "grad_norm": 3.5637545602445946, + "learning_rate": 1.978767145627167e-05, + "loss": 1.1932, + "step": 864 + }, + { + "epoch": 0.46959826275787186, + "grad_norm": 3.131289093388952, + "learning_rate": 1.9786949985785187e-05, + "loss": 0.69, + "step": 865 + }, + { + "epoch": 0.47014115092290987, + "grad_norm": 3.522468470152659, + "learning_rate": 1.978622730483313e-05, + "loss": 1.0277, + "step": 866 + }, + { + "epoch": 0.47068403908794787, + "grad_norm": 4.046439598363532, + "learning_rate": 1.9785503413504883e-05, + "loss": 1.2917, + "step": 867 + }, + { + "epoch": 0.47122692725298587, + "grad_norm": 5.55259287782346, + "learning_rate": 1.978477831188998e-05, + "loss": 1.866, + "step": 868 + }, + { + "epoch": 0.4717698154180239, + "grad_norm": 5.874766516228883, + "learning_rate": 1.97840520000781e-05, + "loss": 1.6221, + "step": 869 + }, + { + "epoch": 0.4723127035830619, + "grad_norm": 5.530080877073305, + "learning_rate": 1.9783324478159074e-05, + "loss": 1.7617, + "step": 870 + }, + { + "epoch": 0.4728555917480999, + "grad_norm": 4.996268624326209, + "learning_rate": 1.9782595746222886e-05, + "loss": 1.5794, + "step": 871 + }, + { + "epoch": 0.4733984799131379, + "grad_norm": 3.934049315300115, + "learning_rate": 1.9781865804359663e-05, + "loss": 1.2603, + "step": 872 + }, + { + "epoch": 0.4739413680781759, + "grad_norm": 4.534039775579426, + "learning_rate": 1.978113465265969e-05, + "loss": 1.2766, + "step": 873 + }, + { + "epoch": 0.4744842562432139, + "grad_norm": 4.281791787646871, + "learning_rate": 1.9780402291213393e-05, + "loss": 1.6424, + "step": 874 + }, + { + "epoch": 0.4750271444082519, + "grad_norm": 4.443571747878234, + "learning_rate": 1.977966872011135e-05, + "loss": 1.514, + "step": 875 + }, + { + "epoch": 0.4755700325732899, + "grad_norm": 4.432653652742011, + "learning_rate": 1.9778933939444298e-05, + "loss": 1.2405, + "step": 876 + }, + { + "epoch": 0.4761129207383279, + "grad_norm": 3.408851551186198, + "learning_rate": 1.9778197949303107e-05, + "loss": 0.9597, + "step": 877 + }, + { + "epoch": 0.4766558089033659, + "grad_norm": 3.13258933410851, + "learning_rate": 1.9777460749778812e-05, + "loss": 0.8015, + "step": 878 + }, + { + "epoch": 0.4771986970684039, + "grad_norm": 3.535707436018844, + "learning_rate": 1.977672234096259e-05, + "loss": 1.3375, + "step": 879 + }, + { + "epoch": 0.4777415852334419, + "grad_norm": 5.210333639552661, + "learning_rate": 1.9775982722945764e-05, + "loss": 1.5039, + "step": 880 + }, + { + "epoch": 0.4782844733984799, + "grad_norm": 4.14115837125392, + "learning_rate": 1.9775241895819818e-05, + "loss": 1.0267, + "step": 881 + }, + { + "epoch": 0.4788273615635179, + "grad_norm": 4.390843049257208, + "learning_rate": 1.977449985967637e-05, + "loss": 1.6639, + "step": 882 + }, + { + "epoch": 0.4793702497285559, + "grad_norm": 3.5906251649150405, + "learning_rate": 1.9773756614607205e-05, + "loss": 1.2382, + "step": 883 + }, + { + "epoch": 0.4799131378935939, + "grad_norm": 4.78880370244169, + "learning_rate": 1.9773012160704242e-05, + "loss": 1.9436, + "step": 884 + }, + { + "epoch": 0.4804560260586319, + "grad_norm": 4.024679917575935, + "learning_rate": 1.977226649805956e-05, + "loss": 1.7126, + "step": 885 + }, + { + "epoch": 0.4809989142236699, + "grad_norm": 4.789782875875197, + "learning_rate": 1.9771519626765384e-05, + "loss": 1.4639, + "step": 886 + }, + { + "epoch": 0.4815418023887079, + "grad_norm": 4.188325588019175, + "learning_rate": 1.9770771546914088e-05, + "loss": 0.9698, + "step": 887 + }, + { + "epoch": 0.4820846905537459, + "grad_norm": 4.491436069698424, + "learning_rate": 1.9770022258598192e-05, + "loss": 1.5504, + "step": 888 + }, + { + "epoch": 0.4826275787187839, + "grad_norm": 4.889270563723842, + "learning_rate": 1.9769271761910373e-05, + "loss": 1.6052, + "step": 889 + }, + { + "epoch": 0.4831704668838219, + "grad_norm": 3.7615089990618493, + "learning_rate": 1.9768520056943454e-05, + "loss": 1.2445, + "step": 890 + }, + { + "epoch": 0.4837133550488599, + "grad_norm": 4.153654659347469, + "learning_rate": 1.9767767143790403e-05, + "loss": 1.4304, + "step": 891 + }, + { + "epoch": 0.48425624321389793, + "grad_norm": 4.040032964648755, + "learning_rate": 1.9767013022544346e-05, + "loss": 1.2906, + "step": 892 + }, + { + "epoch": 0.48479913137893593, + "grad_norm": 4.6559235863058275, + "learning_rate": 1.976625769329855e-05, + "loss": 1.5214, + "step": 893 + }, + { + "epoch": 0.48534201954397393, + "grad_norm": 4.224494142526306, + "learning_rate": 1.976550115614644e-05, + "loss": 1.2299, + "step": 894 + }, + { + "epoch": 0.48588490770901194, + "grad_norm": 5.225621473809677, + "learning_rate": 1.9764743411181585e-05, + "loss": 1.3489, + "step": 895 + }, + { + "epoch": 0.48642779587404994, + "grad_norm": 3.298440710545404, + "learning_rate": 1.97639844584977e-05, + "loss": 1.2107, + "step": 896 + }, + { + "epoch": 0.48697068403908794, + "grad_norm": 4.8373671778516485, + "learning_rate": 1.9763224298188652e-05, + "loss": 1.7748, + "step": 897 + }, + { + "epoch": 0.48751357220412594, + "grad_norm": 3.9332967347674224, + "learning_rate": 1.9762462930348465e-05, + "loss": 1.1036, + "step": 898 + }, + { + "epoch": 0.48805646036916395, + "grad_norm": 3.555513906222738, + "learning_rate": 1.9761700355071306e-05, + "loss": 1.5988, + "step": 899 + }, + { + "epoch": 0.48859934853420195, + "grad_norm": 4.115001448738943, + "learning_rate": 1.9760936572451487e-05, + "loss": 0.9789, + "step": 900 + }, + { + "epoch": 0.48914223669923995, + "grad_norm": 3.955269044937872, + "learning_rate": 1.9760171582583476e-05, + "loss": 1.3833, + "step": 901 + }, + { + "epoch": 0.48968512486427795, + "grad_norm": 2.718909991804504, + "learning_rate": 1.975940538556189e-05, + "loss": 1.1708, + "step": 902 + }, + { + "epoch": 0.49022801302931596, + "grad_norm": 4.234679090051759, + "learning_rate": 1.975863798148149e-05, + "loss": 1.3002, + "step": 903 + }, + { + "epoch": 0.49077090119435396, + "grad_norm": 4.162638017444828, + "learning_rate": 1.975786937043719e-05, + "loss": 1.246, + "step": 904 + }, + { + "epoch": 0.49131378935939196, + "grad_norm": 4.466027820294199, + "learning_rate": 1.975709955252406e-05, + "loss": 1.7345, + "step": 905 + }, + { + "epoch": 0.49185667752442996, + "grad_norm": 4.120456111509664, + "learning_rate": 1.97563285278373e-05, + "loss": 1.1339, + "step": 906 + }, + { + "epoch": 0.49239956568946797, + "grad_norm": 3.7932999801551346, + "learning_rate": 1.9755556296472286e-05, + "loss": 1.241, + "step": 907 + }, + { + "epoch": 0.49294245385450597, + "grad_norm": 3.3311865780622107, + "learning_rate": 1.9754782858524515e-05, + "loss": 1.2329, + "step": 908 + }, + { + "epoch": 0.49348534201954397, + "grad_norm": 4.21189840914184, + "learning_rate": 1.9754008214089654e-05, + "loss": 1.089, + "step": 909 + }, + { + "epoch": 0.494028230184582, + "grad_norm": 4.882048591403942, + "learning_rate": 1.9753232363263513e-05, + "loss": 1.3579, + "step": 910 + }, + { + "epoch": 0.49457111834962, + "grad_norm": 3.8495412115497833, + "learning_rate": 1.9752455306142052e-05, + "loss": 0.876, + "step": 911 + }, + { + "epoch": 0.495114006514658, + "grad_norm": 5.35378911928701, + "learning_rate": 1.975167704282137e-05, + "loss": 1.5025, + "step": 912 + }, + { + "epoch": 0.495656894679696, + "grad_norm": 3.4776107222093176, + "learning_rate": 1.9750897573397733e-05, + "loss": 1.1576, + "step": 913 + }, + { + "epoch": 0.496199782844734, + "grad_norm": 4.3680755226003525, + "learning_rate": 1.9750116897967545e-05, + "loss": 1.2519, + "step": 914 + }, + { + "epoch": 0.496742671009772, + "grad_norm": 4.154395835162648, + "learning_rate": 1.9749335016627355e-05, + "loss": 1.3397, + "step": 915 + }, + { + "epoch": 0.49728555917481, + "grad_norm": 5.465210053229008, + "learning_rate": 1.9748551929473873e-05, + "loss": 1.7555, + "step": 916 + }, + { + "epoch": 0.497828447339848, + "grad_norm": 4.391089608848125, + "learning_rate": 1.9747767636603953e-05, + "loss": 0.9685, + "step": 917 + }, + { + "epoch": 0.498371335504886, + "grad_norm": 4.215127655197256, + "learning_rate": 1.9746982138114597e-05, + "loss": 1.3889, + "step": 918 + }, + { + "epoch": 0.498914223669924, + "grad_norm": 4.559291789845839, + "learning_rate": 1.9746195434102956e-05, + "loss": 1.2411, + "step": 919 + }, + { + "epoch": 0.499457111834962, + "grad_norm": 4.226267506705079, + "learning_rate": 1.9745407524666327e-05, + "loss": 0.9517, + "step": 920 + }, + { + "epoch": 0.5, + "grad_norm": 4.289151336000736, + "learning_rate": 1.9744618409902164e-05, + "loss": 0.9123, + "step": 921 + }, + { + "epoch": 0.500542888165038, + "grad_norm": 5.356043714827742, + "learning_rate": 1.9743828089908067e-05, + "loss": 1.4039, + "step": 922 + }, + { + "epoch": 0.501085776330076, + "grad_norm": 4.867388642606108, + "learning_rate": 1.9743036564781785e-05, + "loss": 1.877, + "step": 923 + }, + { + "epoch": 0.501628664495114, + "grad_norm": 4.034273796102648, + "learning_rate": 1.9742243834621207e-05, + "loss": 1.6151, + "step": 924 + }, + { + "epoch": 0.502171552660152, + "grad_norm": 4.67131234305618, + "learning_rate": 1.9741449899524393e-05, + "loss": 1.6415, + "step": 925 + }, + { + "epoch": 0.50271444082519, + "grad_norm": 4.609211834325983, + "learning_rate": 1.9740654759589524e-05, + "loss": 1.2702, + "step": 926 + }, + { + "epoch": 0.503257328990228, + "grad_norm": 3.8693712873725064, + "learning_rate": 1.973985841491495e-05, + "loss": 1.0518, + "step": 927 + }, + { + "epoch": 0.503800217155266, + "grad_norm": 4.224951197105749, + "learning_rate": 1.9739060865599163e-05, + "loss": 1.1367, + "step": 928 + }, + { + "epoch": 0.504343105320304, + "grad_norm": 4.851656117293343, + "learning_rate": 1.973826211174081e-05, + "loss": 1.2753, + "step": 929 + }, + { + "epoch": 0.504885993485342, + "grad_norm": 4.255261482914816, + "learning_rate": 1.9737462153438675e-05, + "loss": 1.1723, + "step": 930 + }, + { + "epoch": 0.50542888165038, + "grad_norm": 3.8038614228711616, + "learning_rate": 1.9736660990791705e-05, + "loss": 1.1919, + "step": 931 + }, + { + "epoch": 0.505971769815418, + "grad_norm": 3.934227455395002, + "learning_rate": 1.9735858623898984e-05, + "loss": 1.1311, + "step": 932 + }, + { + "epoch": 0.506514657980456, + "grad_norm": 6.087081162565673, + "learning_rate": 1.9735055052859752e-05, + "loss": 1.4226, + "step": 933 + }, + { + "epoch": 0.507057546145494, + "grad_norm": 3.901680463400142, + "learning_rate": 1.9734250277773398e-05, + "loss": 0.8336, + "step": 934 + }, + { + "epoch": 0.507600434310532, + "grad_norm": 4.29804205692996, + "learning_rate": 1.9733444298739454e-05, + "loss": 1.5767, + "step": 935 + }, + { + "epoch": 0.50814332247557, + "grad_norm": 4.114842645062852, + "learning_rate": 1.9732637115857606e-05, + "loss": 1.3353, + "step": 936 + }, + { + "epoch": 0.508686210640608, + "grad_norm": 3.8525228530533706, + "learning_rate": 1.9731828729227686e-05, + "loss": 0.8208, + "step": 937 + }, + { + "epoch": 0.509229098805646, + "grad_norm": 4.592070280997817, + "learning_rate": 1.973101913894968e-05, + "loss": 1.3908, + "step": 938 + }, + { + "epoch": 0.509771986970684, + "grad_norm": 3.5842386277465628, + "learning_rate": 1.9730208345123718e-05, + "loss": 1.1996, + "step": 939 + }, + { + "epoch": 0.510314875135722, + "grad_norm": 6.053647281932304, + "learning_rate": 1.9729396347850082e-05, + "loss": 1.565, + "step": 940 + }, + { + "epoch": 0.51085776330076, + "grad_norm": 4.515664489508703, + "learning_rate": 1.9728583147229196e-05, + "loss": 1.4823, + "step": 941 + }, + { + "epoch": 0.511400651465798, + "grad_norm": 4.7752750595730475, + "learning_rate": 1.9727768743361644e-05, + "loss": 1.3867, + "step": 942 + }, + { + "epoch": 0.511943539630836, + "grad_norm": 3.983262894008015, + "learning_rate": 1.972695313634815e-05, + "loss": 1.1327, + "step": 943 + }, + { + "epoch": 0.512486427795874, + "grad_norm": 3.8510372454489326, + "learning_rate": 1.9726136326289586e-05, + "loss": 1.0233, + "step": 944 + }, + { + "epoch": 0.5130293159609121, + "grad_norm": 3.7881258264140634, + "learning_rate": 1.9725318313286985e-05, + "loss": 1.0239, + "step": 945 + }, + { + "epoch": 0.5135722041259501, + "grad_norm": 5.532602386318144, + "learning_rate": 1.9724499097441513e-05, + "loss": 1.8325, + "step": 946 + }, + { + "epoch": 0.5141150922909881, + "grad_norm": 5.978249809716352, + "learning_rate": 1.9723678678854493e-05, + "loss": 1.4899, + "step": 947 + }, + { + "epoch": 0.5146579804560261, + "grad_norm": 4.164217442125778, + "learning_rate": 1.9722857057627398e-05, + "loss": 1.5428, + "step": 948 + }, + { + "epoch": 0.5152008686210641, + "grad_norm": 3.315839086509451, + "learning_rate": 1.9722034233861848e-05, + "loss": 1.161, + "step": 949 + }, + { + "epoch": 0.5157437567861021, + "grad_norm": 4.362448119896709, + "learning_rate": 1.9721210207659608e-05, + "loss": 1.3402, + "step": 950 + }, + { + "epoch": 0.5162866449511401, + "grad_norm": 4.196652705893848, + "learning_rate": 1.9720384979122594e-05, + "loss": 1.4782, + "step": 951 + }, + { + "epoch": 0.5168295331161781, + "grad_norm": 4.093486911730782, + "learning_rate": 1.9719558548352876e-05, + "loss": 0.8508, + "step": 952 + }, + { + "epoch": 0.5173724212812161, + "grad_norm": 3.623412868803193, + "learning_rate": 1.9718730915452664e-05, + "loss": 1.4979, + "step": 953 + }, + { + "epoch": 0.5179153094462541, + "grad_norm": 3.4964911024452845, + "learning_rate": 1.9717902080524324e-05, + "loss": 1.0495, + "step": 954 + }, + { + "epoch": 0.5184581976112921, + "grad_norm": 4.021867363378223, + "learning_rate": 1.9717072043670367e-05, + "loss": 1.5284, + "step": 955 + }, + { + "epoch": 0.5190010857763301, + "grad_norm": 4.383918034859983, + "learning_rate": 1.9716240804993454e-05, + "loss": 1.3092, + "step": 956 + }, + { + "epoch": 0.5195439739413681, + "grad_norm": 4.61200677033352, + "learning_rate": 1.971540836459639e-05, + "loss": 1.8809, + "step": 957 + }, + { + "epoch": 0.5200868621064061, + "grad_norm": 4.490685725956694, + "learning_rate": 1.9714574722582142e-05, + "loss": 1.4004, + "step": 958 + }, + { + "epoch": 0.5206297502714441, + "grad_norm": 3.4814367857792736, + "learning_rate": 1.9713739879053802e-05, + "loss": 0.7918, + "step": 959 + }, + { + "epoch": 0.5211726384364821, + "grad_norm": 4.751597566480749, + "learning_rate": 1.9712903834114635e-05, + "loss": 1.6595, + "step": 960 + }, + { + "epoch": 0.5217155266015201, + "grad_norm": 4.535333205317335, + "learning_rate": 1.9712066587868042e-05, + "loss": 1.5676, + "step": 961 + }, + { + "epoch": 0.5222584147665581, + "grad_norm": 4.285001993552699, + "learning_rate": 1.9711228140417577e-05, + "loss": 1.0156, + "step": 962 + }, + { + "epoch": 0.5228013029315961, + "grad_norm": 4.61361421766516, + "learning_rate": 1.9710388491866934e-05, + "loss": 0.9552, + "step": 963 + }, + { + "epoch": 0.5233441910966341, + "grad_norm": 5.760482951437034, + "learning_rate": 1.9709547642319968e-05, + "loss": 2.2573, + "step": 964 + }, + { + "epoch": 0.5238870792616721, + "grad_norm": 4.38974180509626, + "learning_rate": 1.9708705591880674e-05, + "loss": 0.9939, + "step": 965 + }, + { + "epoch": 0.5244299674267101, + "grad_norm": 4.856793718321527, + "learning_rate": 1.97078623406532e-05, + "loss": 1.6161, + "step": 966 + }, + { + "epoch": 0.5249728555917481, + "grad_norm": 4.164767885626303, + "learning_rate": 1.9707017888741838e-05, + "loss": 1.4327, + "step": 967 + }, + { + "epoch": 0.5255157437567861, + "grad_norm": 5.163022804287371, + "learning_rate": 1.970617223625104e-05, + "loss": 1.2063, + "step": 968 + }, + { + "epoch": 0.5260586319218241, + "grad_norm": 4.826390195584734, + "learning_rate": 1.9705325383285384e-05, + "loss": 1.609, + "step": 969 + }, + { + "epoch": 0.5266015200868621, + "grad_norm": 6.3570394668292405, + "learning_rate": 1.9704477329949617e-05, + "loss": 1.5845, + "step": 970 + }, + { + "epoch": 0.5271444082519001, + "grad_norm": 5.176208926963624, + "learning_rate": 1.9703628076348628e-05, + "loss": 1.5792, + "step": 971 + }, + { + "epoch": 0.5276872964169381, + "grad_norm": 4.968725075311641, + "learning_rate": 1.9702777622587452e-05, + "loss": 1.774, + "step": 972 + }, + { + "epoch": 0.5282301845819761, + "grad_norm": 3.733580423042607, + "learning_rate": 1.9701925968771277e-05, + "loss": 0.8424, + "step": 973 + }, + { + "epoch": 0.5287730727470141, + "grad_norm": 4.330993358163065, + "learning_rate": 1.9701073115005437e-05, + "loss": 1.6198, + "step": 974 + }, + { + "epoch": 0.5293159609120521, + "grad_norm": 4.64032700965104, + "learning_rate": 1.9700219061395408e-05, + "loss": 1.2958, + "step": 975 + }, + { + "epoch": 0.5298588490770901, + "grad_norm": 4.08538215893539, + "learning_rate": 1.969936380804683e-05, + "loss": 1.5371, + "step": 976 + }, + { + "epoch": 0.5304017372421281, + "grad_norm": 3.7672102191405012, + "learning_rate": 1.9698507355065478e-05, + "loss": 1.4831, + "step": 977 + }, + { + "epoch": 0.5309446254071661, + "grad_norm": 4.237599031471779, + "learning_rate": 1.969764970255728e-05, + "loss": 1.4561, + "step": 978 + }, + { + "epoch": 0.5314875135722041, + "grad_norm": 4.296489590459447, + "learning_rate": 1.9696790850628308e-05, + "loss": 1.3813, + "step": 979 + }, + { + "epoch": 0.5320304017372421, + "grad_norm": 4.037479773042692, + "learning_rate": 1.9695930799384787e-05, + "loss": 1.103, + "step": 980 + }, + { + "epoch": 0.5325732899022801, + "grad_norm": 3.388491089813824, + "learning_rate": 1.9695069548933097e-05, + "loss": 1.0416, + "step": 981 + }, + { + "epoch": 0.5331161780673181, + "grad_norm": 3.6808593992848873, + "learning_rate": 1.969420709937975e-05, + "loss": 1.3158, + "step": 982 + }, + { + "epoch": 0.5336590662323561, + "grad_norm": 3.574434479685402, + "learning_rate": 1.969334345083142e-05, + "loss": 0.9542, + "step": 983 + }, + { + "epoch": 0.5342019543973942, + "grad_norm": 4.205886917864195, + "learning_rate": 1.9692478603394926e-05, + "loss": 1.1542, + "step": 984 + }, + { + "epoch": 0.5347448425624322, + "grad_norm": 3.89660120144171, + "learning_rate": 1.9691612557177225e-05, + "loss": 1.2374, + "step": 985 + }, + { + "epoch": 0.5352877307274702, + "grad_norm": 5.508260629607227, + "learning_rate": 1.9690745312285443e-05, + "loss": 2.0836, + "step": 986 + }, + { + "epoch": 0.5358306188925082, + "grad_norm": 5.482641917586033, + "learning_rate": 1.968987686882683e-05, + "loss": 1.5608, + "step": 987 + }, + { + "epoch": 0.5363735070575462, + "grad_norm": 5.4640893579957845, + "learning_rate": 1.9689007226908807e-05, + "loss": 1.3632, + "step": 988 + }, + { + "epoch": 0.5369163952225842, + "grad_norm": 4.163793092253212, + "learning_rate": 1.9688136386638926e-05, + "loss": 1.0509, + "step": 989 + }, + { + "epoch": 0.5374592833876222, + "grad_norm": 7.431185891169344, + "learning_rate": 1.96872643481249e-05, + "loss": 2.2147, + "step": 990 + }, + { + "epoch": 0.5380021715526602, + "grad_norm": 4.243102861037138, + "learning_rate": 1.9686391111474574e-05, + "loss": 0.9978, + "step": 991 + }, + { + "epoch": 0.5385450597176982, + "grad_norm": 4.250454886057412, + "learning_rate": 1.968551667679596e-05, + "loss": 1.5265, + "step": 992 + }, + { + "epoch": 0.5390879478827362, + "grad_norm": 5.989466882905416, + "learning_rate": 1.9684641044197207e-05, + "loss": 1.3344, + "step": 993 + }, + { + "epoch": 0.5396308360477742, + "grad_norm": 6.003524822087355, + "learning_rate": 1.9683764213786617e-05, + "loss": 2.0785, + "step": 994 + }, + { + "epoch": 0.5401737242128122, + "grad_norm": 4.535858561115482, + "learning_rate": 1.9682886185672633e-05, + "loss": 1.7654, + "step": 995 + }, + { + "epoch": 0.5407166123778502, + "grad_norm": 3.304287633610362, + "learning_rate": 1.9682006959963854e-05, + "loss": 0.9151, + "step": 996 + }, + { + "epoch": 0.5412595005428882, + "grad_norm": 3.6154194334954797, + "learning_rate": 1.9681126536769022e-05, + "loss": 1.1871, + "step": 997 + }, + { + "epoch": 0.5418023887079262, + "grad_norm": 5.596808777173772, + "learning_rate": 1.968024491619703e-05, + "loss": 1.1773, + "step": 998 + }, + { + "epoch": 0.5423452768729642, + "grad_norm": 4.486728043355676, + "learning_rate": 1.9679362098356923e-05, + "loss": 1.6495, + "step": 999 + }, + { + "epoch": 0.5428881650380022, + "grad_norm": 5.254525660355539, + "learning_rate": 1.9678478083357882e-05, + "loss": 1.7632, + "step": 1000 + }, + { + "epoch": 0.5434310532030402, + "grad_norm": 4.084222439189464, + "learning_rate": 1.9677592871309248e-05, + "loss": 1.4108, + "step": 1001 + }, + { + "epoch": 0.5439739413680782, + "grad_norm": 3.7647794378542248, + "learning_rate": 1.9676706462320504e-05, + "loss": 1.0893, + "step": 1002 + }, + { + "epoch": 0.5445168295331162, + "grad_norm": 3.658128786211824, + "learning_rate": 1.967581885650128e-05, + "loss": 1.0626, + "step": 1003 + }, + { + "epoch": 0.5450597176981542, + "grad_norm": 4.111525087755858, + "learning_rate": 1.967493005396136e-05, + "loss": 1.1915, + "step": 1004 + }, + { + "epoch": 0.5456026058631922, + "grad_norm": 4.130761699023899, + "learning_rate": 1.967404005481067e-05, + "loss": 1.1007, + "step": 1005 + }, + { + "epoch": 0.5461454940282302, + "grad_norm": 4.610164776763781, + "learning_rate": 1.9673148859159292e-05, + "loss": 1.3626, + "step": 1006 + }, + { + "epoch": 0.5466883821932682, + "grad_norm": 3.3516729554956686, + "learning_rate": 1.9672256467117445e-05, + "loss": 1.1138, + "step": 1007 + }, + { + "epoch": 0.5472312703583062, + "grad_norm": 5.185808250719794, + "learning_rate": 1.9671362878795502e-05, + "loss": 1.5639, + "step": 1008 + }, + { + "epoch": 0.5477741585233442, + "grad_norm": 3.9549787550673323, + "learning_rate": 1.9670468094303983e-05, + "loss": 1.0094, + "step": 1009 + }, + { + "epoch": 0.5483170466883822, + "grad_norm": 5.025417088271204, + "learning_rate": 1.966957211375356e-05, + "loss": 1.9741, + "step": 1010 + }, + { + "epoch": 0.5488599348534202, + "grad_norm": 5.1560673391175, + "learning_rate": 1.9668674937255044e-05, + "loss": 0.9653, + "step": 1011 + }, + { + "epoch": 0.5494028230184582, + "grad_norm": 4.843540796107876, + "learning_rate": 1.9667776564919404e-05, + "loss": 1.4745, + "step": 1012 + }, + { + "epoch": 0.5499457111834962, + "grad_norm": 3.8379573247832193, + "learning_rate": 1.966687699685775e-05, + "loss": 1.537, + "step": 1013 + }, + { + "epoch": 0.5504885993485342, + "grad_norm": 4.291804816024517, + "learning_rate": 1.9665976233181342e-05, + "loss": 1.6859, + "step": 1014 + }, + { + "epoch": 0.5510314875135722, + "grad_norm": 4.827630960232388, + "learning_rate": 1.966507427400159e-05, + "loss": 1.3205, + "step": 1015 + }, + { + "epoch": 0.5515743756786102, + "grad_norm": 4.251540690047392, + "learning_rate": 1.9664171119430044e-05, + "loss": 0.8213, + "step": 1016 + }, + { + "epoch": 0.5521172638436482, + "grad_norm": 3.7693342275615476, + "learning_rate": 1.9663266769578414e-05, + "loss": 1.0349, + "step": 1017 + }, + { + "epoch": 0.5526601520086862, + "grad_norm": 4.075472936638236, + "learning_rate": 1.966236122455855e-05, + "loss": 1.0853, + "step": 1018 + }, + { + "epoch": 0.5532030401737242, + "grad_norm": 4.771167333111926, + "learning_rate": 1.9661454484482448e-05, + "loss": 1.2135, + "step": 1019 + }, + { + "epoch": 0.5537459283387622, + "grad_norm": 3.834854303591261, + "learning_rate": 1.9660546549462262e-05, + "loss": 1.0972, + "step": 1020 + }, + { + "epoch": 0.5542888165038002, + "grad_norm": 4.712560042125338, + "learning_rate": 1.9659637419610278e-05, + "loss": 0.8779, + "step": 1021 + }, + { + "epoch": 0.5548317046688382, + "grad_norm": 3.8174143787104566, + "learning_rate": 1.9658727095038942e-05, + "loss": 0.7395, + "step": 1022 + }, + { + "epoch": 0.5553745928338762, + "grad_norm": 5.970646625648977, + "learning_rate": 1.965781557586085e-05, + "loss": 1.4233, + "step": 1023 + }, + { + "epoch": 0.5559174809989142, + "grad_norm": 3.872798187838392, + "learning_rate": 1.9656902862188732e-05, + "loss": 0.8147, + "step": 1024 + }, + { + "epoch": 0.5564603691639523, + "grad_norm": 5.632609234869673, + "learning_rate": 1.9655988954135473e-05, + "loss": 1.1581, + "step": 1025 + }, + { + "epoch": 0.5570032573289903, + "grad_norm": 4.141188191911578, + "learning_rate": 1.9655073851814117e-05, + "loss": 1.1707, + "step": 1026 + }, + { + "epoch": 0.5575461454940283, + "grad_norm": 5.7835323867879564, + "learning_rate": 1.9654157555337837e-05, + "loss": 1.2383, + "step": 1027 + }, + { + "epoch": 0.5580890336590663, + "grad_norm": 6.410603262158535, + "learning_rate": 1.9653240064819965e-05, + "loss": 2.3461, + "step": 1028 + }, + { + "epoch": 0.5586319218241043, + "grad_norm": 4.661944777700264, + "learning_rate": 1.9652321380373974e-05, + "loss": 1.1718, + "step": 1029 + }, + { + "epoch": 0.5591748099891423, + "grad_norm": 4.224334614804915, + "learning_rate": 1.9651401502113497e-05, + "loss": 1.0937, + "step": 1030 + }, + { + "epoch": 0.5597176981541803, + "grad_norm": 4.292561821167698, + "learning_rate": 1.9650480430152295e-05, + "loss": 0.8306, + "step": 1031 + }, + { + "epoch": 0.5602605863192183, + "grad_norm": 4.515070808084738, + "learning_rate": 1.9649558164604293e-05, + "loss": 1.4042, + "step": 1032 + }, + { + "epoch": 0.5608034744842563, + "grad_norm": 4.906478484057087, + "learning_rate": 1.964863470558356e-05, + "loss": 1.3396, + "step": 1033 + }, + { + "epoch": 0.5613463626492943, + "grad_norm": 3.9530548146664306, + "learning_rate": 1.9647710053204307e-05, + "loss": 0.917, + "step": 1034 + }, + { + "epoch": 0.5618892508143323, + "grad_norm": 4.640160859146004, + "learning_rate": 1.96467842075809e-05, + "loss": 1.6625, + "step": 1035 + }, + { + "epoch": 0.5624321389793703, + "grad_norm": 4.921887921875984, + "learning_rate": 1.964585716882785e-05, + "loss": 1.5133, + "step": 1036 + }, + { + "epoch": 0.5629750271444083, + "grad_norm": 4.667679101301949, + "learning_rate": 1.964492893705981e-05, + "loss": 0.9713, + "step": 1037 + }, + { + "epoch": 0.5635179153094463, + "grad_norm": 6.0457527064758825, + "learning_rate": 1.9643999512391586e-05, + "loss": 1.1101, + "step": 1038 + }, + { + "epoch": 0.5640608034744843, + "grad_norm": 5.188403972218186, + "learning_rate": 1.964306889493813e-05, + "loss": 1.8276, + "step": 1039 + }, + { + "epoch": 0.5646036916395223, + "grad_norm": 4.449370010668479, + "learning_rate": 1.9642137084814548e-05, + "loss": 1.0797, + "step": 1040 + }, + { + "epoch": 0.5651465798045603, + "grad_norm": 5.97660491495462, + "learning_rate": 1.9641204082136085e-05, + "loss": 1.913, + "step": 1041 + }, + { + "epoch": 0.5656894679695983, + "grad_norm": 5.611282387277076, + "learning_rate": 1.9640269887018135e-05, + "loss": 2.4593, + "step": 1042 + }, + { + "epoch": 0.5662323561346363, + "grad_norm": 3.6226559029227996, + "learning_rate": 1.9639334499576237e-05, + "loss": 0.7939, + "step": 1043 + }, + { + "epoch": 0.5667752442996743, + "grad_norm": 4.701569628633213, + "learning_rate": 1.963839791992609e-05, + "loss": 1.4092, + "step": 1044 + }, + { + "epoch": 0.5673181324647123, + "grad_norm": 5.185222096624812, + "learning_rate": 1.9637460148183525e-05, + "loss": 1.3851, + "step": 1045 + }, + { + "epoch": 0.5678610206297503, + "grad_norm": 4.094772946058271, + "learning_rate": 1.963652118446453e-05, + "loss": 1.1872, + "step": 1046 + }, + { + "epoch": 0.5684039087947883, + "grad_norm": 4.653748340189311, + "learning_rate": 1.9635581028885233e-05, + "loss": 1.1546, + "step": 1047 + }, + { + "epoch": 0.5689467969598263, + "grad_norm": 3.9426831069951844, + "learning_rate": 1.9634639681561924e-05, + "loss": 0.9324, + "step": 1048 + }, + { + "epoch": 0.5694896851248643, + "grad_norm": 4.949304407204856, + "learning_rate": 1.9633697142611017e-05, + "loss": 1.2372, + "step": 1049 + }, + { + "epoch": 0.5700325732899023, + "grad_norm": 5.364689744345368, + "learning_rate": 1.9632753412149096e-05, + "loss": 1.5104, + "step": 1050 + }, + { + "epoch": 0.5705754614549403, + "grad_norm": 5.307969223857254, + "learning_rate": 1.9631808490292884e-05, + "loss": 1.212, + "step": 1051 + }, + { + "epoch": 0.5711183496199783, + "grad_norm": 5.197310781163775, + "learning_rate": 1.963086237715924e-05, + "loss": 1.0175, + "step": 1052 + }, + { + "epoch": 0.5716612377850163, + "grad_norm": 4.095170540464123, + "learning_rate": 1.9629915072865194e-05, + "loss": 1.2711, + "step": 1053 + }, + { + "epoch": 0.5722041259500543, + "grad_norm": 4.816474093418903, + "learning_rate": 1.9628966577527902e-05, + "loss": 0.7274, + "step": 1054 + }, + { + "epoch": 0.5727470141150923, + "grad_norm": 4.000604469041167, + "learning_rate": 1.962801689126468e-05, + "loss": 1.2913, + "step": 1055 + }, + { + "epoch": 0.5732899022801303, + "grad_norm": 4.31668064977018, + "learning_rate": 1.962706601419298e-05, + "loss": 1.2598, + "step": 1056 + }, + { + "epoch": 0.5738327904451683, + "grad_norm": 5.033983318277113, + "learning_rate": 1.9626113946430414e-05, + "loss": 1.5149, + "step": 1057 + }, + { + "epoch": 0.5743756786102063, + "grad_norm": 4.51738649586986, + "learning_rate": 1.9625160688094733e-05, + "loss": 1.2389, + "step": 1058 + }, + { + "epoch": 0.5749185667752443, + "grad_norm": 3.6899120907028853, + "learning_rate": 1.9624206239303837e-05, + "loss": 1.1632, + "step": 1059 + }, + { + "epoch": 0.5754614549402823, + "grad_norm": 4.423539955787714, + "learning_rate": 1.9623250600175775e-05, + "loss": 0.9526, + "step": 1060 + }, + { + "epoch": 0.5760043431053203, + "grad_norm": 5.062986942342047, + "learning_rate": 1.962229377082874e-05, + "loss": 1.6768, + "step": 1061 + }, + { + "epoch": 0.5765472312703583, + "grad_norm": 5.044415439938055, + "learning_rate": 1.962133575138108e-05, + "loss": 1.3808, + "step": 1062 + }, + { + "epoch": 0.5770901194353963, + "grad_norm": 3.1405283495786587, + "learning_rate": 1.962037654195128e-05, + "loss": 0.7588, + "step": 1063 + }, + { + "epoch": 0.5776330076004343, + "grad_norm": 5.449873922495929, + "learning_rate": 1.9619416142657974e-05, + "loss": 1.1544, + "step": 1064 + }, + { + "epoch": 0.5781758957654723, + "grad_norm": 4.268527750000251, + "learning_rate": 1.961845455361995e-05, + "loss": 1.1941, + "step": 1065 + }, + { + "epoch": 0.5787187839305103, + "grad_norm": 4.6641878464575095, + "learning_rate": 1.9617491774956137e-05, + "loss": 1.7162, + "step": 1066 + }, + { + "epoch": 0.5792616720955484, + "grad_norm": 4.583539478633983, + "learning_rate": 1.9616527806785612e-05, + "loss": 1.1415, + "step": 1067 + }, + { + "epoch": 0.5798045602605864, + "grad_norm": 4.325093393868848, + "learning_rate": 1.96155626492276e-05, + "loss": 1.3284, + "step": 1068 + }, + { + "epoch": 0.5803474484256244, + "grad_norm": 4.180109104666535, + "learning_rate": 1.9614596302401478e-05, + "loss": 1.136, + "step": 1069 + }, + { + "epoch": 0.5808903365906624, + "grad_norm": 4.108700840520527, + "learning_rate": 1.9613628766426762e-05, + "loss": 1.6142, + "step": 1070 + }, + { + "epoch": 0.5814332247557004, + "grad_norm": 5.1931558422083555, + "learning_rate": 1.9612660041423116e-05, + "loss": 0.8479, + "step": 1071 + }, + { + "epoch": 0.5819761129207384, + "grad_norm": 6.008601497593343, + "learning_rate": 1.9611690127510358e-05, + "loss": 1.4743, + "step": 1072 + }, + { + "epoch": 0.5825190010857764, + "grad_norm": 4.531622764164067, + "learning_rate": 1.9610719024808444e-05, + "loss": 1.5755, + "step": 1073 + }, + { + "epoch": 0.5830618892508144, + "grad_norm": 6.18647874774261, + "learning_rate": 1.9609746733437486e-05, + "loss": 1.7369, + "step": 1074 + }, + { + "epoch": 0.5836047774158524, + "grad_norm": 5.5880249590216025, + "learning_rate": 1.9608773253517738e-05, + "loss": 1.4402, + "step": 1075 + }, + { + "epoch": 0.5841476655808904, + "grad_norm": 4.798692544278573, + "learning_rate": 1.9607798585169595e-05, + "loss": 1.2257, + "step": 1076 + }, + { + "epoch": 0.5846905537459284, + "grad_norm": 5.635786862796798, + "learning_rate": 1.960682272851361e-05, + "loss": 1.183, + "step": 1077 + }, + { + "epoch": 0.5852334419109664, + "grad_norm": 3.3979344506068037, + "learning_rate": 1.9605845683670484e-05, + "loss": 0.6463, + "step": 1078 + }, + { + "epoch": 0.5857763300760044, + "grad_norm": 4.5152680576801405, + "learning_rate": 1.960486745076105e-05, + "loss": 0.8476, + "step": 1079 + }, + { + "epoch": 0.5863192182410424, + "grad_norm": 4.4606140059612835, + "learning_rate": 1.9603888029906305e-05, + "loss": 1.185, + "step": 1080 + }, + { + "epoch": 0.5868621064060804, + "grad_norm": 4.229219175605004, + "learning_rate": 1.960290742122738e-05, + "loss": 1.3672, + "step": 1081 + }, + { + "epoch": 0.5874049945711184, + "grad_norm": 4.964172517040214, + "learning_rate": 1.960192562484556e-05, + "loss": 1.0561, + "step": 1082 + }, + { + "epoch": 0.5879478827361564, + "grad_norm": 4.8348336861653785, + "learning_rate": 1.960094264088228e-05, + "loss": 1.1005, + "step": 1083 + }, + { + "epoch": 0.5884907709011944, + "grad_norm": 4.3153540955607514, + "learning_rate": 1.959995846945911e-05, + "loss": 1.6513, + "step": 1084 + }, + { + "epoch": 0.5890336590662324, + "grad_norm": 4.733352525701168, + "learning_rate": 1.9598973110697773e-05, + "loss": 1.6162, + "step": 1085 + }, + { + "epoch": 0.5895765472312704, + "grad_norm": 5.402815147148315, + "learning_rate": 1.959798656472015e-05, + "loss": 1.4178, + "step": 1086 + }, + { + "epoch": 0.5901194353963084, + "grad_norm": 3.8276911935691147, + "learning_rate": 1.9596998831648247e-05, + "loss": 1.1554, + "step": 1087 + }, + { + "epoch": 0.5906623235613464, + "grad_norm": 4.499779496087119, + "learning_rate": 1.9596009911604232e-05, + "loss": 0.9786, + "step": 1088 + }, + { + "epoch": 0.5912052117263844, + "grad_norm": 4.3705352636049595, + "learning_rate": 1.959501980471042e-05, + "loss": 1.0827, + "step": 1089 + }, + { + "epoch": 0.5917480998914224, + "grad_norm": 4.9770508472387025, + "learning_rate": 1.9594028511089264e-05, + "loss": 1.2512, + "step": 1090 + }, + { + "epoch": 0.5922909880564604, + "grad_norm": 4.23068807802663, + "learning_rate": 1.9593036030863376e-05, + "loss": 0.7528, + "step": 1091 + }, + { + "epoch": 0.5928338762214984, + "grad_norm": 4.321323716192747, + "learning_rate": 1.9592042364155496e-05, + "loss": 1.0773, + "step": 1092 + }, + { + "epoch": 0.5933767643865364, + "grad_norm": 3.380036112638087, + "learning_rate": 1.9591047511088535e-05, + "loss": 0.7238, + "step": 1093 + }, + { + "epoch": 0.5939196525515744, + "grad_norm": 3.9722637658815447, + "learning_rate": 1.959005147178553e-05, + "loss": 0.8132, + "step": 1094 + }, + { + "epoch": 0.5944625407166124, + "grad_norm": 4.354488865716506, + "learning_rate": 1.9589054246369673e-05, + "loss": 1.1214, + "step": 1095 + }, + { + "epoch": 0.5950054288816504, + "grad_norm": 4.337355429292092, + "learning_rate": 1.9588055834964307e-05, + "loss": 0.9039, + "step": 1096 + }, + { + "epoch": 0.5955483170466884, + "grad_norm": 5.864047590425705, + "learning_rate": 1.9587056237692912e-05, + "loss": 1.6648, + "step": 1097 + }, + { + "epoch": 0.5960912052117264, + "grad_norm": 4.904173131706714, + "learning_rate": 1.9586055454679123e-05, + "loss": 1.1465, + "step": 1098 + }, + { + "epoch": 0.5966340933767644, + "grad_norm": 4.411877930208604, + "learning_rate": 1.9585053486046717e-05, + "loss": 1.4579, + "step": 1099 + }, + { + "epoch": 0.5971769815418024, + "grad_norm": 4.985171448881133, + "learning_rate": 1.9584050331919616e-05, + "loss": 1.5497, + "step": 1100 + }, + { + "epoch": 0.5977198697068404, + "grad_norm": 4.950532301400581, + "learning_rate": 1.9583045992421902e-05, + "loss": 1.3811, + "step": 1101 + }, + { + "epoch": 0.5982627578718784, + "grad_norm": 4.154119544907876, + "learning_rate": 1.9582040467677782e-05, + "loss": 1.186, + "step": 1102 + }, + { + "epoch": 0.5988056460369164, + "grad_norm": 4.069018604381756, + "learning_rate": 1.9581033757811628e-05, + "loss": 0.9312, + "step": 1103 + }, + { + "epoch": 0.5993485342019544, + "grad_norm": 5.33070083129983, + "learning_rate": 1.9580025862947948e-05, + "loss": 1.4274, + "step": 1104 + }, + { + "epoch": 0.5998914223669924, + "grad_norm": 4.855606632567906, + "learning_rate": 1.95790167832114e-05, + "loss": 1.0875, + "step": 1105 + }, + { + "epoch": 0.6004343105320304, + "grad_norm": 5.30182267546013, + "learning_rate": 1.957800651872679e-05, + "loss": 1.6501, + "step": 1106 + }, + { + "epoch": 0.6009771986970684, + "grad_norm": 4.433903391255917, + "learning_rate": 1.957699506961907e-05, + "loss": 1.3897, + "step": 1107 + }, + { + "epoch": 0.6015200868621065, + "grad_norm": 4.687653782723915, + "learning_rate": 1.9575982436013335e-05, + "loss": 0.8484, + "step": 1108 + }, + { + "epoch": 0.6020629750271445, + "grad_norm": 5.650426140105992, + "learning_rate": 1.957496861803483e-05, + "loss": 1.923, + "step": 1109 + }, + { + "epoch": 0.6026058631921825, + "grad_norm": 4.024650772512973, + "learning_rate": 1.957395361580895e-05, + "loss": 1.2713, + "step": 1110 + }, + { + "epoch": 0.6031487513572205, + "grad_norm": 4.343921133786845, + "learning_rate": 1.9572937429461223e-05, + "loss": 1.3152, + "step": 1111 + }, + { + "epoch": 0.6036916395222585, + "grad_norm": 4.1467176317930265, + "learning_rate": 1.957192005911734e-05, + "loss": 1.3325, + "step": 1112 + }, + { + "epoch": 0.6042345276872965, + "grad_norm": 4.305465221711973, + "learning_rate": 1.9570901504903128e-05, + "loss": 0.9237, + "step": 1113 + }, + { + "epoch": 0.6047774158523345, + "grad_norm": 4.101045495900469, + "learning_rate": 1.9569881766944564e-05, + "loss": 1.3932, + "step": 1114 + }, + { + "epoch": 0.6053203040173725, + "grad_norm": 4.0920748344339835, + "learning_rate": 1.956886084536777e-05, + "loss": 1.0067, + "step": 1115 + }, + { + "epoch": 0.6058631921824105, + "grad_norm": 4.844431281019285, + "learning_rate": 1.956783874029902e-05, + "loss": 1.2726, + "step": 1116 + }, + { + "epoch": 0.6064060803474485, + "grad_norm": 4.329792546128024, + "learning_rate": 1.9566815451864723e-05, + "loss": 1.1144, + "step": 1117 + }, + { + "epoch": 0.6069489685124865, + "grad_norm": 5.430322578642178, + "learning_rate": 1.9565790980191447e-05, + "loss": 1.3307, + "step": 1118 + }, + { + "epoch": 0.6074918566775245, + "grad_norm": 3.672369600344093, + "learning_rate": 1.9564765325405895e-05, + "loss": 0.8514, + "step": 1119 + }, + { + "epoch": 0.6080347448425625, + "grad_norm": 3.6433282223345334, + "learning_rate": 1.9563738487634924e-05, + "loss": 1.1972, + "step": 1120 + }, + { + "epoch": 0.6085776330076005, + "grad_norm": 4.23249445053715, + "learning_rate": 1.956271046700553e-05, + "loss": 1.0638, + "step": 1121 + }, + { + "epoch": 0.6091205211726385, + "grad_norm": 3.9162807435580134, + "learning_rate": 1.956168126364487e-05, + "loss": 1.121, + "step": 1122 + }, + { + "epoch": 0.6096634093376765, + "grad_norm": 3.3651337591577812, + "learning_rate": 1.956065087768023e-05, + "loss": 0.9604, + "step": 1123 + }, + { + "epoch": 0.6102062975027145, + "grad_norm": 5.5905307606778685, + "learning_rate": 1.955961930923905e-05, + "loss": 1.4509, + "step": 1124 + }, + { + "epoch": 0.6107491856677525, + "grad_norm": 3.6120233107886404, + "learning_rate": 1.955858655844892e-05, + "loss": 1.0106, + "step": 1125 + }, + { + "epoch": 0.6112920738327905, + "grad_norm": 3.6242967727445468, + "learning_rate": 1.9557552625437574e-05, + "loss": 1.0768, + "step": 1126 + }, + { + "epoch": 0.6118349619978285, + "grad_norm": 4.856763858492221, + "learning_rate": 1.9556517510332883e-05, + "loss": 1.6372, + "step": 1127 + }, + { + "epoch": 0.6123778501628665, + "grad_norm": 4.648427809898363, + "learning_rate": 1.9555481213262873e-05, + "loss": 1.1622, + "step": 1128 + }, + { + "epoch": 0.6129207383279045, + "grad_norm": 4.228089396781667, + "learning_rate": 1.9554443734355723e-05, + "loss": 1.067, + "step": 1129 + }, + { + "epoch": 0.6134636264929425, + "grad_norm": 4.069786144676624, + "learning_rate": 1.9553405073739743e-05, + "loss": 1.1427, + "step": 1130 + }, + { + "epoch": 0.6140065146579805, + "grad_norm": 3.112182391585594, + "learning_rate": 1.9552365231543395e-05, + "loss": 0.7862, + "step": 1131 + }, + { + "epoch": 0.6145494028230185, + "grad_norm": 4.783518734299315, + "learning_rate": 1.955132420789529e-05, + "loss": 1.0366, + "step": 1132 + }, + { + "epoch": 0.6150922909880565, + "grad_norm": 4.905128048667836, + "learning_rate": 1.9550282002924187e-05, + "loss": 1.2566, + "step": 1133 + }, + { + "epoch": 0.6156351791530945, + "grad_norm": 4.643678450407587, + "learning_rate": 1.9549238616758987e-05, + "loss": 1.2749, + "step": 1134 + }, + { + "epoch": 0.6161780673181325, + "grad_norm": 4.18363664236292, + "learning_rate": 1.954819404952873e-05, + "loss": 1.1523, + "step": 1135 + }, + { + "epoch": 0.6167209554831705, + "grad_norm": 3.9776065520060335, + "learning_rate": 1.9547148301362623e-05, + "loss": 0.8672, + "step": 1136 + }, + { + "epoch": 0.6172638436482085, + "grad_norm": 5.417170242762362, + "learning_rate": 1.9546101372389994e-05, + "loss": 1.6434, + "step": 1137 + }, + { + "epoch": 0.6178067318132465, + "grad_norm": 3.3325133642660068, + "learning_rate": 1.9545053262740335e-05, + "loss": 1.1447, + "step": 1138 + }, + { + "epoch": 0.6183496199782845, + "grad_norm": 5.4727757064949545, + "learning_rate": 1.9544003972543273e-05, + "loss": 1.5013, + "step": 1139 + }, + { + "epoch": 0.6188925081433225, + "grad_norm": 4.619714236951128, + "learning_rate": 1.954295350192859e-05, + "loss": 0.9758, + "step": 1140 + }, + { + "epoch": 0.6194353963083605, + "grad_norm": 5.189787937143953, + "learning_rate": 1.954190185102621e-05, + "loss": 1.2771, + "step": 1141 + }, + { + "epoch": 0.6199782844733985, + "grad_norm": 6.571114312185869, + "learning_rate": 1.9540849019966198e-05, + "loss": 1.2805, + "step": 1142 + }, + { + "epoch": 0.6205211726384365, + "grad_norm": 5.6343515388196455, + "learning_rate": 1.9539795008878774e-05, + "loss": 1.4061, + "step": 1143 + }, + { + "epoch": 0.6210640608034745, + "grad_norm": 5.582702517092888, + "learning_rate": 1.9538739817894302e-05, + "loss": 1.613, + "step": 1144 + }, + { + "epoch": 0.6216069489685125, + "grad_norm": 6.194616996988272, + "learning_rate": 1.9537683447143287e-05, + "loss": 1.114, + "step": 1145 + }, + { + "epoch": 0.6221498371335505, + "grad_norm": 6.0395907363668515, + "learning_rate": 1.9536625896756377e-05, + "loss": 1.5233, + "step": 1146 + }, + { + "epoch": 0.6226927252985885, + "grad_norm": 4.672511820840566, + "learning_rate": 1.9535567166864382e-05, + "loss": 1.1491, + "step": 1147 + }, + { + "epoch": 0.6232356134636265, + "grad_norm": 4.954133576744497, + "learning_rate": 1.9534507257598244e-05, + "loss": 1.1974, + "step": 1148 + }, + { + "epoch": 0.6237785016286646, + "grad_norm": 4.447205012745297, + "learning_rate": 1.953344616908905e-05, + "loss": 1.019, + "step": 1149 + }, + { + "epoch": 0.6243213897937026, + "grad_norm": 4.284087771325503, + "learning_rate": 1.9532383901468038e-05, + "loss": 1.0814, + "step": 1150 + }, + { + "epoch": 0.6248642779587406, + "grad_norm": 3.576067836847051, + "learning_rate": 1.9531320454866595e-05, + "loss": 1.1025, + "step": 1151 + }, + { + "epoch": 0.6254071661237784, + "grad_norm": 3.6831292979487777, + "learning_rate": 1.9530255829416246e-05, + "loss": 0.967, + "step": 1152 + }, + { + "epoch": 0.6259500542888164, + "grad_norm": 3.842496728015195, + "learning_rate": 1.952919002524867e-05, + "loss": 1.0903, + "step": 1153 + }, + { + "epoch": 0.6264929424538545, + "grad_norm": 4.077633480381933, + "learning_rate": 1.952812304249568e-05, + "loss": 0.9227, + "step": 1154 + }, + { + "epoch": 0.6270358306188925, + "grad_norm": 4.157004955121598, + "learning_rate": 1.952705488128925e-05, + "loss": 0.9427, + "step": 1155 + }, + { + "epoch": 0.6275787187839305, + "grad_norm": 5.267614104282471, + "learning_rate": 1.952598554176149e-05, + "loss": 1.9875, + "step": 1156 + }, + { + "epoch": 0.6281216069489685, + "grad_norm": 3.8938350314376735, + "learning_rate": 1.952491502404465e-05, + "loss": 1.1662, + "step": 1157 + }, + { + "epoch": 0.6286644951140065, + "grad_norm": 4.851409026807682, + "learning_rate": 1.9523843328271144e-05, + "loss": 1.9559, + "step": 1158 + }, + { + "epoch": 0.6292073832790445, + "grad_norm": 4.4104613077358845, + "learning_rate": 1.9522770454573513e-05, + "loss": 0.8011, + "step": 1159 + }, + { + "epoch": 0.6297502714440825, + "grad_norm": 3.6883335641947856, + "learning_rate": 1.952169640308446e-05, + "loss": 1.1838, + "step": 1160 + }, + { + "epoch": 0.6302931596091205, + "grad_norm": 3.728203039459734, + "learning_rate": 1.9520621173936818e-05, + "loss": 1.1423, + "step": 1161 + }, + { + "epoch": 0.6308360477741585, + "grad_norm": 3.6272612276564296, + "learning_rate": 1.9519544767263574e-05, + "loss": 0.8376, + "step": 1162 + }, + { + "epoch": 0.6313789359391965, + "grad_norm": 3.928276333701112, + "learning_rate": 1.951846718319786e-05, + "loss": 1.0749, + "step": 1163 + }, + { + "epoch": 0.6319218241042345, + "grad_norm": 5.941949883397345, + "learning_rate": 1.951738842187296e-05, + "loss": 1.8535, + "step": 1164 + }, + { + "epoch": 0.6324647122692725, + "grad_norm": 3.4741200080360644, + "learning_rate": 1.951630848342229e-05, + "loss": 0.7537, + "step": 1165 + }, + { + "epoch": 0.6330076004343105, + "grad_norm": 5.392480994365815, + "learning_rate": 1.9515227367979416e-05, + "loss": 0.8776, + "step": 1166 + }, + { + "epoch": 0.6335504885993485, + "grad_norm": 5.203169892005258, + "learning_rate": 1.951414507567806e-05, + "loss": 1.2678, + "step": 1167 + }, + { + "epoch": 0.6340933767643865, + "grad_norm": 7.499226723318992, + "learning_rate": 1.9513061606652076e-05, + "loss": 1.3295, + "step": 1168 + }, + { + "epoch": 0.6346362649294245, + "grad_norm": 4.355172848230307, + "learning_rate": 1.9511976961035474e-05, + "loss": 0.9188, + "step": 1169 + }, + { + "epoch": 0.6351791530944625, + "grad_norm": 4.983220943925928, + "learning_rate": 1.9510891138962398e-05, + "loss": 1.1951, + "step": 1170 + }, + { + "epoch": 0.6357220412595005, + "grad_norm": 5.3098934312394395, + "learning_rate": 1.950980414056715e-05, + "loss": 0.855, + "step": 1171 + }, + { + "epoch": 0.6362649294245385, + "grad_norm": 5.125683116898323, + "learning_rate": 1.950871596598417e-05, + "loss": 1.1371, + "step": 1172 + }, + { + "epoch": 0.6368078175895765, + "grad_norm": 4.3010814264850215, + "learning_rate": 1.950762661534804e-05, + "loss": 1.7401, + "step": 1173 + }, + { + "epoch": 0.6373507057546145, + "grad_norm": 5.860625964437982, + "learning_rate": 1.95065360887935e-05, + "loss": 1.3557, + "step": 1174 + }, + { + "epoch": 0.6378935939196525, + "grad_norm": 6.3356570754666155, + "learning_rate": 1.9505444386455426e-05, + "loss": 1.2281, + "step": 1175 + }, + { + "epoch": 0.6384364820846905, + "grad_norm": 4.788366086551355, + "learning_rate": 1.9504351508468842e-05, + "loss": 1.3458, + "step": 1176 + }, + { + "epoch": 0.6389793702497285, + "grad_norm": 4.5553889359557065, + "learning_rate": 1.9503257454968914e-05, + "loss": 0.803, + "step": 1177 + }, + { + "epoch": 0.6395222584147665, + "grad_norm": 4.9063187966995345, + "learning_rate": 1.950216222609096e-05, + "loss": 0.8306, + "step": 1178 + }, + { + "epoch": 0.6400651465798045, + "grad_norm": 4.775138114897886, + "learning_rate": 1.9501065821970435e-05, + "loss": 0.9783, + "step": 1179 + }, + { + "epoch": 0.6406080347448425, + "grad_norm": 4.53672284240203, + "learning_rate": 1.9499968242742948e-05, + "loss": 1.3076, + "step": 1180 + }, + { + "epoch": 0.6411509229098805, + "grad_norm": 4.539860050199487, + "learning_rate": 1.949886948854425e-05, + "loss": 1.4212, + "step": 1181 + }, + { + "epoch": 0.6416938110749185, + "grad_norm": 4.251462432005476, + "learning_rate": 1.9497769559510232e-05, + "loss": 1.0274, + "step": 1182 + }, + { + "epoch": 0.6422366992399565, + "grad_norm": 4.49791289564495, + "learning_rate": 1.9496668455776938e-05, + "loss": 1.7531, + "step": 1183 + }, + { + "epoch": 0.6427795874049945, + "grad_norm": 4.996740483321485, + "learning_rate": 1.9495566177480555e-05, + "loss": 1.201, + "step": 1184 + }, + { + "epoch": 0.6433224755700325, + "grad_norm": 5.347352236818569, + "learning_rate": 1.9494462724757413e-05, + "loss": 1.5723, + "step": 1185 + }, + { + "epoch": 0.6438653637350705, + "grad_norm": 4.2242558589374015, + "learning_rate": 1.9493358097743988e-05, + "loss": 1.8014, + "step": 1186 + }, + { + "epoch": 0.6444082519001085, + "grad_norm": 5.042556201755935, + "learning_rate": 1.9492252296576906e-05, + "loss": 1.5466, + "step": 1187 + }, + { + "epoch": 0.6449511400651465, + "grad_norm": 4.284314238035697, + "learning_rate": 1.949114532139293e-05, + "loss": 0.9515, + "step": 1188 + }, + { + "epoch": 0.6454940282301845, + "grad_norm": 4.699826119048575, + "learning_rate": 1.9490037172328974e-05, + "loss": 1.1677, + "step": 1189 + }, + { + "epoch": 0.6460369163952225, + "grad_norm": 6.035247135554916, + "learning_rate": 1.9488927849522095e-05, + "loss": 1.957, + "step": 1190 + }, + { + "epoch": 0.6465798045602605, + "grad_norm": 4.042004235065528, + "learning_rate": 1.94878173531095e-05, + "loss": 1.2182, + "step": 1191 + }, + { + "epoch": 0.6471226927252985, + "grad_norm": 4.114859597350168, + "learning_rate": 1.948670568322853e-05, + "loss": 1.1787, + "step": 1192 + }, + { + "epoch": 0.6476655808903365, + "grad_norm": 4.380177461001504, + "learning_rate": 1.9485592840016682e-05, + "loss": 1.3912, + "step": 1193 + }, + { + "epoch": 0.6482084690553745, + "grad_norm": 5.160710982722015, + "learning_rate": 1.948447882361159e-05, + "loss": 1.2642, + "step": 1194 + }, + { + "epoch": 0.6487513572204126, + "grad_norm": 3.950160510485583, + "learning_rate": 1.9483363634151046e-05, + "loss": 1.1246, + "step": 1195 + }, + { + "epoch": 0.6492942453854506, + "grad_norm": 4.587449618376974, + "learning_rate": 1.9482247271772974e-05, + "loss": 1.3275, + "step": 1196 + }, + { + "epoch": 0.6498371335504886, + "grad_norm": 4.919831513668322, + "learning_rate": 1.9481129736615445e-05, + "loss": 1.4007, + "step": 1197 + }, + { + "epoch": 0.6503800217155266, + "grad_norm": 4.298954423192263, + "learning_rate": 1.948001102881668e-05, + "loss": 0.8076, + "step": 1198 + }, + { + "epoch": 0.6509229098805646, + "grad_norm": 5.525772014179547, + "learning_rate": 1.9478891148515043e-05, + "loss": 1.4558, + "step": 1199 + }, + { + "epoch": 0.6514657980456026, + "grad_norm": 5.473549669240299, + "learning_rate": 1.947777009584904e-05, + "loss": 1.0976, + "step": 1200 + }, + { + "epoch": 0.6520086862106406, + "grad_norm": 5.360010284326786, + "learning_rate": 1.947664787095733e-05, + "loss": 1.0462, + "step": 1201 + }, + { + "epoch": 0.6525515743756786, + "grad_norm": 4.696691081444868, + "learning_rate": 1.9475524473978705e-05, + "loss": 0.933, + "step": 1202 + }, + { + "epoch": 0.6530944625407166, + "grad_norm": 6.0503072087093965, + "learning_rate": 1.947439990505211e-05, + "loss": 0.9628, + "step": 1203 + }, + { + "epoch": 0.6536373507057546, + "grad_norm": 5.476272854511399, + "learning_rate": 1.9473274164316637e-05, + "loss": 0.9716, + "step": 1204 + }, + { + "epoch": 0.6541802388707926, + "grad_norm": 3.910448715873912, + "learning_rate": 1.9472147251911517e-05, + "loss": 1.0732, + "step": 1205 + }, + { + "epoch": 0.6547231270358306, + "grad_norm": 4.596114022138067, + "learning_rate": 1.9471019167976126e-05, + "loss": 0.9399, + "step": 1206 + }, + { + "epoch": 0.6552660152008686, + "grad_norm": 6.764840749864346, + "learning_rate": 1.946988991264999e-05, + "loss": 2.0529, + "step": 1207 + }, + { + "epoch": 0.6558089033659066, + "grad_norm": 5.217328230116913, + "learning_rate": 1.9468759486072778e-05, + "loss": 1.0477, + "step": 1208 + }, + { + "epoch": 0.6563517915309446, + "grad_norm": 4.03429219354135, + "learning_rate": 1.9467627888384303e-05, + "loss": 1.1103, + "step": 1209 + }, + { + "epoch": 0.6568946796959826, + "grad_norm": 4.301388894222403, + "learning_rate": 1.946649511972452e-05, + "loss": 0.8515, + "step": 1210 + }, + { + "epoch": 0.6574375678610206, + "grad_norm": 5.285446092583105, + "learning_rate": 1.9465361180233536e-05, + "loss": 1.2881, + "step": 1211 + }, + { + "epoch": 0.6579804560260586, + "grad_norm": 5.879897026479874, + "learning_rate": 1.9464226070051593e-05, + "loss": 0.9362, + "step": 1212 + }, + { + "epoch": 0.6585233441910966, + "grad_norm": 5.335257600571856, + "learning_rate": 1.9463089789319083e-05, + "loss": 1.4372, + "step": 1213 + }, + { + "epoch": 0.6590662323561346, + "grad_norm": 5.272070125431448, + "learning_rate": 1.9461952338176552e-05, + "loss": 1.2063, + "step": 1214 + }, + { + "epoch": 0.6596091205211726, + "grad_norm": 5.188765715525815, + "learning_rate": 1.946081371676467e-05, + "loss": 0.8912, + "step": 1215 + }, + { + "epoch": 0.6601520086862106, + "grad_norm": 4.747539421286818, + "learning_rate": 1.9459673925224275e-05, + "loss": 1.0381, + "step": 1216 + }, + { + "epoch": 0.6606948968512486, + "grad_norm": 4.486254331013965, + "learning_rate": 1.945853296369633e-05, + "loss": 1.2584, + "step": 1217 + }, + { + "epoch": 0.6612377850162866, + "grad_norm": 4.730012415000151, + "learning_rate": 1.945739083232195e-05, + "loss": 1.1061, + "step": 1218 + }, + { + "epoch": 0.6617806731813246, + "grad_norm": 5.713566521874704, + "learning_rate": 1.9456247531242405e-05, + "loss": 1.4628, + "step": 1219 + }, + { + "epoch": 0.6623235613463626, + "grad_norm": 6.08134803588137, + "learning_rate": 1.9455103060599093e-05, + "loss": 1.6743, + "step": 1220 + }, + { + "epoch": 0.6628664495114006, + "grad_norm": 4.778151760566848, + "learning_rate": 1.9453957420533562e-05, + "loss": 0.8454, + "step": 1221 + }, + { + "epoch": 0.6634093376764386, + "grad_norm": 5.007171333863363, + "learning_rate": 1.945281061118751e-05, + "loss": 1.079, + "step": 1222 + }, + { + "epoch": 0.6639522258414766, + "grad_norm": 4.034543591270571, + "learning_rate": 1.945166263270278e-05, + "loss": 0.8378, + "step": 1223 + }, + { + "epoch": 0.6644951140065146, + "grad_norm": 4.191899722379318, + "learning_rate": 1.9450513485221352e-05, + "loss": 1.3602, + "step": 1224 + }, + { + "epoch": 0.6650380021715526, + "grad_norm": 6.349672663190246, + "learning_rate": 1.944936316888535e-05, + "loss": 1.861, + "step": 1225 + }, + { + "epoch": 0.6655808903365906, + "grad_norm": 5.991894404952759, + "learning_rate": 1.9448211683837055e-05, + "loss": 1.0931, + "step": 1226 + }, + { + "epoch": 0.6661237785016286, + "grad_norm": 4.405247018509644, + "learning_rate": 1.9447059030218876e-05, + "loss": 1.2239, + "step": 1227 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 4.810361623786047, + "learning_rate": 1.9445905208173387e-05, + "loss": 1.1468, + "step": 1228 + }, + { + "epoch": 0.6672095548317046, + "grad_norm": 4.6967843893557015, + "learning_rate": 1.944475021784328e-05, + "loss": 1.5304, + "step": 1229 + }, + { + "epoch": 0.6677524429967426, + "grad_norm": 5.325911382516086, + "learning_rate": 1.9443594059371417e-05, + "loss": 1.2218, + "step": 1230 + }, + { + "epoch": 0.6682953311617806, + "grad_norm": 4.416437363913103, + "learning_rate": 1.9442436732900788e-05, + "loss": 1.223, + "step": 1231 + }, + { + "epoch": 0.6688382193268186, + "grad_norm": 3.720560087272861, + "learning_rate": 1.9441278238574537e-05, + "loss": 0.8155, + "step": 1232 + }, + { + "epoch": 0.6693811074918566, + "grad_norm": 5.273159297494141, + "learning_rate": 1.9440118576535947e-05, + "loss": 1.1958, + "step": 1233 + }, + { + "epoch": 0.6699239956568946, + "grad_norm": 5.684867015184436, + "learning_rate": 1.9438957746928443e-05, + "loss": 1.2384, + "step": 1234 + }, + { + "epoch": 0.6704668838219326, + "grad_norm": 4.952541482610055, + "learning_rate": 1.9437795749895604e-05, + "loss": 1.337, + "step": 1235 + }, + { + "epoch": 0.6710097719869706, + "grad_norm": 4.788863413585636, + "learning_rate": 1.9436632585581145e-05, + "loss": 0.5182, + "step": 1236 + }, + { + "epoch": 0.6715526601520087, + "grad_norm": 5.723878805619219, + "learning_rate": 1.9435468254128925e-05, + "loss": 1.1444, + "step": 1237 + }, + { + "epoch": 0.6720955483170467, + "grad_norm": 4.3068839361839, + "learning_rate": 1.9434302755682958e-05, + "loss": 1.1286, + "step": 1238 + }, + { + "epoch": 0.6726384364820847, + "grad_norm": 3.7588872381668477, + "learning_rate": 1.9433136090387384e-05, + "loss": 0.9718, + "step": 1239 + }, + { + "epoch": 0.6731813246471227, + "grad_norm": 6.802207677110681, + "learning_rate": 1.9431968258386508e-05, + "loss": 1.2389, + "step": 1240 + }, + { + "epoch": 0.6737242128121607, + "grad_norm": 3.6180158875124735, + "learning_rate": 1.9430799259824766e-05, + "loss": 0.644, + "step": 1241 + }, + { + "epoch": 0.6742671009771987, + "grad_norm": 5.559822163300458, + "learning_rate": 1.9429629094846742e-05, + "loss": 1.1157, + "step": 1242 + }, + { + "epoch": 0.6748099891422367, + "grad_norm": 4.084355750224745, + "learning_rate": 1.942845776359716e-05, + "loss": 0.91, + "step": 1243 + }, + { + "epoch": 0.6753528773072747, + "grad_norm": 3.9402358983627583, + "learning_rate": 1.9427285266220895e-05, + "loss": 0.9409, + "step": 1244 + }, + { + "epoch": 0.6758957654723127, + "grad_norm": 5.093172430302549, + "learning_rate": 1.9426111602862968e-05, + "loss": 1.5905, + "step": 1245 + }, + { + "epoch": 0.6764386536373507, + "grad_norm": 4.02536040473524, + "learning_rate": 1.942493677366853e-05, + "loss": 0.9277, + "step": 1246 + }, + { + "epoch": 0.6769815418023887, + "grad_norm": 4.741533426757492, + "learning_rate": 1.942376077878289e-05, + "loss": 1.1027, + "step": 1247 + }, + { + "epoch": 0.6775244299674267, + "grad_norm": 5.365058237804207, + "learning_rate": 1.9422583618351503e-05, + "loss": 1.072, + "step": 1248 + }, + { + "epoch": 0.6780673181324647, + "grad_norm": 4.479339123691231, + "learning_rate": 1.9421405292519956e-05, + "loss": 1.474, + "step": 1249 + }, + { + "epoch": 0.6786102062975027, + "grad_norm": 7.79086989645888, + "learning_rate": 1.942022580143398e-05, + "loss": 1.6973, + "step": 1250 + }, + { + "epoch": 0.6791530944625407, + "grad_norm": 5.621467578044424, + "learning_rate": 1.9419045145239474e-05, + "loss": 1.5432, + "step": 1251 + }, + { + "epoch": 0.6796959826275787, + "grad_norm": 5.041038447258229, + "learning_rate": 1.9417863324082444e-05, + "loss": 1.2977, + "step": 1252 + }, + { + "epoch": 0.6802388707926167, + "grad_norm": 4.717189339301973, + "learning_rate": 1.941668033810907e-05, + "loss": 0.924, + "step": 1253 + }, + { + "epoch": 0.6807817589576547, + "grad_norm": 5.086833595497465, + "learning_rate": 1.9415496187465667e-05, + "loss": 1.5545, + "step": 1254 + }, + { + "epoch": 0.6813246471226927, + "grad_norm": 4.790403842984273, + "learning_rate": 1.941431087229869e-05, + "loss": 1.2545, + "step": 1255 + }, + { + "epoch": 0.6818675352877307, + "grad_norm": 4.470785518257313, + "learning_rate": 1.9413124392754735e-05, + "loss": 1.2036, + "step": 1256 + }, + { + "epoch": 0.6824104234527687, + "grad_norm": 3.820710537112263, + "learning_rate": 1.9411936748980557e-05, + "loss": 1.0856, + "step": 1257 + }, + { + "epoch": 0.6829533116178067, + "grad_norm": 4.453720544384727, + "learning_rate": 1.9410747941123036e-05, + "loss": 1.1645, + "step": 1258 + }, + { + "epoch": 0.6834961997828447, + "grad_norm": 4.692503519890958, + "learning_rate": 1.9409557969329218e-05, + "loss": 1.1174, + "step": 1259 + }, + { + "epoch": 0.6840390879478827, + "grad_norm": 5.666971506629844, + "learning_rate": 1.940836683374627e-05, + "loss": 1.4904, + "step": 1260 + }, + { + "epoch": 0.6845819761129207, + "grad_norm": 5.468580615919992, + "learning_rate": 1.940717453452152e-05, + "loss": 1.2834, + "step": 1261 + }, + { + "epoch": 0.6851248642779587, + "grad_norm": 6.42319990350808, + "learning_rate": 1.9405981071802425e-05, + "loss": 1.1144, + "step": 1262 + }, + { + "epoch": 0.6856677524429967, + "grad_norm": 6.078674980286373, + "learning_rate": 1.9404786445736607e-05, + "loss": 1.4214, + "step": 1263 + }, + { + "epoch": 0.6862106406080347, + "grad_norm": 4.465631588283012, + "learning_rate": 1.9403590656471806e-05, + "loss": 0.7813, + "step": 1264 + }, + { + "epoch": 0.6867535287730727, + "grad_norm": 5.220582351132811, + "learning_rate": 1.940239370415593e-05, + "loss": 1.2112, + "step": 1265 + }, + { + "epoch": 0.6872964169381107, + "grad_norm": 5.1374760487841895, + "learning_rate": 1.9401195588937014e-05, + "loss": 1.2402, + "step": 1266 + }, + { + "epoch": 0.6878393051031487, + "grad_norm": 5.106543779494583, + "learning_rate": 1.9399996310963243e-05, + "loss": 1.1846, + "step": 1267 + }, + { + "epoch": 0.6883821932681867, + "grad_norm": 3.9521299273320363, + "learning_rate": 1.939879587038295e-05, + "loss": 0.8975, + "step": 1268 + }, + { + "epoch": 0.6889250814332247, + "grad_norm": 5.398700470714175, + "learning_rate": 1.9397594267344604e-05, + "loss": 1.4686, + "step": 1269 + }, + { + "epoch": 0.6894679695982627, + "grad_norm": 5.8287973073589265, + "learning_rate": 1.939639150199682e-05, + "loss": 1.011, + "step": 1270 + }, + { + "epoch": 0.6900108577633007, + "grad_norm": 4.812101240482426, + "learning_rate": 1.9395187574488358e-05, + "loss": 0.948, + "step": 1271 + }, + { + "epoch": 0.6905537459283387, + "grad_norm": 5.828731649028842, + "learning_rate": 1.939398248496813e-05, + "loss": 1.2233, + "step": 1272 + }, + { + "epoch": 0.6910966340933767, + "grad_norm": 4.890080173319348, + "learning_rate": 1.9392776233585167e-05, + "loss": 0.8859, + "step": 1273 + }, + { + "epoch": 0.6916395222584147, + "grad_norm": 4.920508250004042, + "learning_rate": 1.9391568820488674e-05, + "loss": 0.73, + "step": 1274 + }, + { + "epoch": 0.6921824104234527, + "grad_norm": 5.635401553567158, + "learning_rate": 1.9390360245827983e-05, + "loss": 1.1432, + "step": 1275 + }, + { + "epoch": 0.6927252985884907, + "grad_norm": 4.588336191443314, + "learning_rate": 1.9389150509752566e-05, + "loss": 1.3326, + "step": 1276 + }, + { + "epoch": 0.6932681867535287, + "grad_norm": 4.631694372660094, + "learning_rate": 1.9387939612412056e-05, + "loss": 1.3762, + "step": 1277 + }, + { + "epoch": 0.6938110749185668, + "grad_norm": 5.585985863722577, + "learning_rate": 1.938672755395621e-05, + "loss": 1.3606, + "step": 1278 + }, + { + "epoch": 0.6943539630836048, + "grad_norm": 4.764783952643876, + "learning_rate": 1.938551433453494e-05, + "loss": 0.9858, + "step": 1279 + }, + { + "epoch": 0.6948968512486428, + "grad_norm": 4.234252918062867, + "learning_rate": 1.9384299954298297e-05, + "loss": 0.7702, + "step": 1280 + }, + { + "epoch": 0.6954397394136808, + "grad_norm": 4.36117548276201, + "learning_rate": 1.938308441339648e-05, + "loss": 1.1846, + "step": 1281 + }, + { + "epoch": 0.6959826275787188, + "grad_norm": 4.311957696864392, + "learning_rate": 1.938186771197983e-05, + "loss": 0.9934, + "step": 1282 + }, + { + "epoch": 0.6965255157437568, + "grad_norm": 5.117807004985933, + "learning_rate": 1.9380649850198824e-05, + "loss": 1.4474, + "step": 1283 + }, + { + "epoch": 0.6970684039087948, + "grad_norm": 5.695329527120276, + "learning_rate": 1.93794308282041e-05, + "loss": 1.3001, + "step": 1284 + }, + { + "epoch": 0.6976112920738328, + "grad_norm": 4.305244126285458, + "learning_rate": 1.937821064614642e-05, + "loss": 1.1421, + "step": 1285 + }, + { + "epoch": 0.6981541802388708, + "grad_norm": 4.574797635236639, + "learning_rate": 1.93769893041767e-05, + "loss": 1.1104, + "step": 1286 + }, + { + "epoch": 0.6986970684039088, + "grad_norm": 5.051193103711606, + "learning_rate": 1.9375766802446002e-05, + "loss": 1.2858, + "step": 1287 + }, + { + "epoch": 0.6992399565689468, + "grad_norm": 5.002235181102662, + "learning_rate": 1.9374543141105518e-05, + "loss": 1.2931, + "step": 1288 + }, + { + "epoch": 0.6997828447339848, + "grad_norm": 5.259386594746194, + "learning_rate": 1.93733183203066e-05, + "loss": 1.0432, + "step": 1289 + }, + { + "epoch": 0.7003257328990228, + "grad_norm": 3.7218449672815117, + "learning_rate": 1.9372092340200736e-05, + "loss": 0.9913, + "step": 1290 + }, + { + "epoch": 0.7008686210640608, + "grad_norm": 4.45109753373239, + "learning_rate": 1.937086520093955e-05, + "loss": 1.2012, + "step": 1291 + }, + { + "epoch": 0.7014115092290988, + "grad_norm": 4.73529586440727, + "learning_rate": 1.9369636902674823e-05, + "loss": 1.3995, + "step": 1292 + }, + { + "epoch": 0.7019543973941368, + "grad_norm": 3.845443933095641, + "learning_rate": 1.936840744555847e-05, + "loss": 1.0971, + "step": 1293 + }, + { + "epoch": 0.7024972855591748, + "grad_norm": 6.67046211091574, + "learning_rate": 1.9367176829742553e-05, + "loss": 1.5273, + "step": 1294 + }, + { + "epoch": 0.7030401737242128, + "grad_norm": 4.454761961182613, + "learning_rate": 1.9365945055379275e-05, + "loss": 1.1282, + "step": 1295 + }, + { + "epoch": 0.7035830618892508, + "grad_norm": 6.552953195473038, + "learning_rate": 1.936471212262099e-05, + "loss": 2.171, + "step": 1296 + }, + { + "epoch": 0.7041259500542888, + "grad_norm": 4.894073730847946, + "learning_rate": 1.9363478031620182e-05, + "loss": 1.2303, + "step": 1297 + }, + { + "epoch": 0.7046688382193268, + "grad_norm": 3.863202427039748, + "learning_rate": 1.936224278252949e-05, + "loss": 0.8515, + "step": 1298 + }, + { + "epoch": 0.7052117263843648, + "grad_norm": 5.503562663348476, + "learning_rate": 1.9361006375501685e-05, + "loss": 1.1834, + "step": 1299 + }, + { + "epoch": 0.7057546145494028, + "grad_norm": 5.318422186143503, + "learning_rate": 1.9359768810689697e-05, + "loss": 1.3378, + "step": 1300 + }, + { + "epoch": 0.7062975027144408, + "grad_norm": 4.46194766163203, + "learning_rate": 1.9358530088246582e-05, + "loss": 1.3318, + "step": 1301 + }, + { + "epoch": 0.7068403908794788, + "grad_norm": 5.6623901137247765, + "learning_rate": 1.9357290208325552e-05, + "loss": 1.3512, + "step": 1302 + }, + { + "epoch": 0.7073832790445168, + "grad_norm": 5.805010217787694, + "learning_rate": 1.9356049171079957e-05, + "loss": 1.2383, + "step": 1303 + }, + { + "epoch": 0.7079261672095548, + "grad_norm": 4.847713425075124, + "learning_rate": 1.9354806976663286e-05, + "loss": 0.7329, + "step": 1304 + }, + { + "epoch": 0.7084690553745928, + "grad_norm": 4.207200307039602, + "learning_rate": 1.935356362522918e-05, + "loss": 1.0119, + "step": 1305 + }, + { + "epoch": 0.7090119435396308, + "grad_norm": 5.76731204400203, + "learning_rate": 1.9352319116931417e-05, + "loss": 1.7383, + "step": 1306 + }, + { + "epoch": 0.7095548317046688, + "grad_norm": 4.989804434976344, + "learning_rate": 1.935107345192392e-05, + "loss": 1.1958, + "step": 1307 + }, + { + "epoch": 0.7100977198697068, + "grad_norm": 4.6099552434159135, + "learning_rate": 1.9349826630360757e-05, + "loss": 0.9933, + "step": 1308 + }, + { + "epoch": 0.7106406080347448, + "grad_norm": 4.693361205108708, + "learning_rate": 1.9348578652396136e-05, + "loss": 1.3553, + "step": 1309 + }, + { + "epoch": 0.7111834961997828, + "grad_norm": 5.473441077550542, + "learning_rate": 1.9347329518184406e-05, + "loss": 1.2202, + "step": 1310 + }, + { + "epoch": 0.7117263843648208, + "grad_norm": 5.885376904454794, + "learning_rate": 1.9346079227880062e-05, + "loss": 1.3698, + "step": 1311 + }, + { + "epoch": 0.7122692725298588, + "grad_norm": 4.115405735194196, + "learning_rate": 1.9344827781637744e-05, + "loss": 1.3361, + "step": 1312 + }, + { + "epoch": 0.7128121606948968, + "grad_norm": 6.745730370060005, + "learning_rate": 1.9343575179612236e-05, + "loss": 1.6411, + "step": 1313 + }, + { + "epoch": 0.7133550488599348, + "grad_norm": 4.363445628534919, + "learning_rate": 1.9342321421958455e-05, + "loss": 0.792, + "step": 1314 + }, + { + "epoch": 0.7138979370249728, + "grad_norm": 6.167581448099132, + "learning_rate": 1.9341066508831472e-05, + "loss": 1.4264, + "step": 1315 + }, + { + "epoch": 0.7144408251900108, + "grad_norm": 4.973571230650737, + "learning_rate": 1.9339810440386495e-05, + "loss": 0.9769, + "step": 1316 + }, + { + "epoch": 0.7149837133550488, + "grad_norm": 5.2637238951774945, + "learning_rate": 1.933855321677888e-05, + "loss": 1.2478, + "step": 1317 + }, + { + "epoch": 0.7155266015200868, + "grad_norm": 5.2296234862723505, + "learning_rate": 1.9337294838164118e-05, + "loss": 1.3999, + "step": 1318 + }, + { + "epoch": 0.7160694896851248, + "grad_norm": 4.720322767524389, + "learning_rate": 1.9336035304697848e-05, + "loss": 1.3561, + "step": 1319 + }, + { + "epoch": 0.7166123778501629, + "grad_norm": 4.2513159313917015, + "learning_rate": 1.9334774616535854e-05, + "loss": 1.176, + "step": 1320 + }, + { + "epoch": 0.7171552660152009, + "grad_norm": 4.5267349508423225, + "learning_rate": 1.9333512773834057e-05, + "loss": 1.1211, + "step": 1321 + }, + { + "epoch": 0.7176981541802389, + "grad_norm": 5.384909301403834, + "learning_rate": 1.9332249776748523e-05, + "loss": 1.3904, + "step": 1322 + }, + { + "epoch": 0.7182410423452769, + "grad_norm": 4.67839864320759, + "learning_rate": 1.9330985625435468e-05, + "loss": 0.8608, + "step": 1323 + }, + { + "epoch": 0.7187839305103149, + "grad_norm": 5.193826637693982, + "learning_rate": 1.9329720320051233e-05, + "loss": 1.3172, + "step": 1324 + }, + { + "epoch": 0.7193268186753529, + "grad_norm": 4.8605293911870096, + "learning_rate": 1.9328453860752324e-05, + "loss": 1.3134, + "step": 1325 + }, + { + "epoch": 0.7198697068403909, + "grad_norm": 5.685265101122876, + "learning_rate": 1.9327186247695377e-05, + "loss": 1.2036, + "step": 1326 + }, + { + "epoch": 0.7204125950054289, + "grad_norm": 5.159715270999642, + "learning_rate": 1.9325917481037164e-05, + "loss": 1.4898, + "step": 1327 + }, + { + "epoch": 0.7209554831704669, + "grad_norm": 5.793178164657722, + "learning_rate": 1.9324647560934613e-05, + "loss": 1.3909, + "step": 1328 + }, + { + "epoch": 0.7214983713355049, + "grad_norm": 3.7442056178189267, + "learning_rate": 1.9323376487544795e-05, + "loss": 0.774, + "step": 1329 + }, + { + "epoch": 0.7220412595005429, + "grad_norm": 4.962252385233631, + "learning_rate": 1.9322104261024912e-05, + "loss": 1.3528, + "step": 1330 + }, + { + "epoch": 0.7225841476655809, + "grad_norm": 5.4534463408495, + "learning_rate": 1.9320830881532316e-05, + "loss": 1.6754, + "step": 1331 + }, + { + "epoch": 0.7231270358306189, + "grad_norm": 6.301262380555263, + "learning_rate": 1.93195563492245e-05, + "loss": 1.5583, + "step": 1332 + }, + { + "epoch": 0.7236699239956569, + "grad_norm": 3.970618104957748, + "learning_rate": 1.9318280664259103e-05, + "loss": 0.8964, + "step": 1333 + }, + { + "epoch": 0.7242128121606949, + "grad_norm": 3.490105469488415, + "learning_rate": 1.9317003826793904e-05, + "loss": 0.8098, + "step": 1334 + }, + { + "epoch": 0.7247557003257329, + "grad_norm": 5.240119278714349, + "learning_rate": 1.9315725836986822e-05, + "loss": 1.1688, + "step": 1335 + }, + { + "epoch": 0.7252985884907709, + "grad_norm": 4.550130688769963, + "learning_rate": 1.931444669499592e-05, + "loss": 1.2569, + "step": 1336 + }, + { + "epoch": 0.7258414766558089, + "grad_norm": 5.058353172472431, + "learning_rate": 1.9313166400979404e-05, + "loss": 1.3471, + "step": 1337 + }, + { + "epoch": 0.7263843648208469, + "grad_norm": 4.528967594836655, + "learning_rate": 1.931188495509563e-05, + "loss": 1.1831, + "step": 1338 + }, + { + "epoch": 0.7269272529858849, + "grad_norm": 4.493564615513058, + "learning_rate": 1.931060235750308e-05, + "loss": 1.4889, + "step": 1339 + }, + { + "epoch": 0.7274701411509229, + "grad_norm": 5.176344629525538, + "learning_rate": 1.9309318608360392e-05, + "loss": 1.6704, + "step": 1340 + }, + { + "epoch": 0.7280130293159609, + "grad_norm": 6.738195156936204, + "learning_rate": 1.930803370782634e-05, + "loss": 1.2147, + "step": 1341 + }, + { + "epoch": 0.7285559174809989, + "grad_norm": 4.248439922934747, + "learning_rate": 1.9306747656059847e-05, + "loss": 1.1943, + "step": 1342 + }, + { + "epoch": 0.7290988056460369, + "grad_norm": 5.175389864390461, + "learning_rate": 1.930546045321997e-05, + "loss": 1.4587, + "step": 1343 + }, + { + "epoch": 0.7296416938110749, + "grad_norm": 5.19934365393201, + "learning_rate": 1.9304172099465914e-05, + "loss": 0.9433, + "step": 1344 + }, + { + "epoch": 0.7301845819761129, + "grad_norm": 6.18681550522946, + "learning_rate": 1.9302882594957025e-05, + "loss": 1.5576, + "step": 1345 + }, + { + "epoch": 0.7307274701411509, + "grad_norm": 4.311510847746277, + "learning_rate": 1.930159193985279e-05, + "loss": 1.327, + "step": 1346 + }, + { + "epoch": 0.7312703583061889, + "grad_norm": 4.044345899710192, + "learning_rate": 1.9300300134312838e-05, + "loss": 1.2106, + "step": 1347 + }, + { + "epoch": 0.7318132464712269, + "grad_norm": 4.181894752624144, + "learning_rate": 1.929900717849694e-05, + "loss": 0.8703, + "step": 1348 + }, + { + "epoch": 0.7323561346362649, + "grad_norm": 5.37566626429507, + "learning_rate": 1.929771307256502e-05, + "loss": 1.446, + "step": 1349 + }, + { + "epoch": 0.7328990228013029, + "grad_norm": 5.37247658500195, + "learning_rate": 1.9296417816677123e-05, + "loss": 0.9239, + "step": 1350 + }, + { + "epoch": 0.7334419109663409, + "grad_norm": 5.749510968066806, + "learning_rate": 1.929512141099346e-05, + "loss": 1.4799, + "step": 1351 + }, + { + "epoch": 0.7339847991313789, + "grad_norm": 5.079635654383841, + "learning_rate": 1.929382385567436e-05, + "loss": 1.2571, + "step": 1352 + }, + { + "epoch": 0.7345276872964169, + "grad_norm": 4.923400923431618, + "learning_rate": 1.929252515088032e-05, + "loss": 1.1967, + "step": 1353 + }, + { + "epoch": 0.7350705754614549, + "grad_norm": 4.397561354835364, + "learning_rate": 1.9291225296771957e-05, + "loss": 0.9088, + "step": 1354 + }, + { + "epoch": 0.7356134636264929, + "grad_norm": 4.13744065040371, + "learning_rate": 1.9289924293510037e-05, + "loss": 1.023, + "step": 1355 + }, + { + "epoch": 0.7361563517915309, + "grad_norm": 4.266553618249588, + "learning_rate": 1.9288622141255477e-05, + "loss": 1.019, + "step": 1356 + }, + { + "epoch": 0.7366992399565689, + "grad_norm": 4.52588883698767, + "learning_rate": 1.928731884016933e-05, + "loss": 1.0935, + "step": 1357 + }, + { + "epoch": 0.7372421281216069, + "grad_norm": 5.46382933116543, + "learning_rate": 1.9286014390412786e-05, + "loss": 1.2833, + "step": 1358 + }, + { + "epoch": 0.737785016286645, + "grad_norm": 4.88651312061201, + "learning_rate": 1.928470879214718e-05, + "loss": 1.6064, + "step": 1359 + }, + { + "epoch": 0.738327904451683, + "grad_norm": 5.403882256899124, + "learning_rate": 1.9283402045533995e-05, + "loss": 1.1537, + "step": 1360 + }, + { + "epoch": 0.738870792616721, + "grad_norm": 6.381532932652649, + "learning_rate": 1.928209415073485e-05, + "loss": 1.527, + "step": 1361 + }, + { + "epoch": 0.739413680781759, + "grad_norm": 4.795244725313847, + "learning_rate": 1.9280785107911505e-05, + "loss": 1.4696, + "step": 1362 + }, + { + "epoch": 0.739956568946797, + "grad_norm": 5.8579061053451955, + "learning_rate": 1.9279474917225866e-05, + "loss": 1.5566, + "step": 1363 + }, + { + "epoch": 0.740499457111835, + "grad_norm": 4.5018596923036185, + "learning_rate": 1.927816357883998e-05, + "loss": 0.9842, + "step": 1364 + }, + { + "epoch": 0.741042345276873, + "grad_norm": 3.596935443077127, + "learning_rate": 1.927685109291604e-05, + "loss": 0.632, + "step": 1365 + }, + { + "epoch": 0.741585233441911, + "grad_norm": 4.535193817325199, + "learning_rate": 1.9275537459616364e-05, + "loss": 1.332, + "step": 1366 + }, + { + "epoch": 0.742128121606949, + "grad_norm": 6.678844775309246, + "learning_rate": 1.9274222679103437e-05, + "loss": 1.6826, + "step": 1367 + }, + { + "epoch": 0.742671009771987, + "grad_norm": 4.444554760823683, + "learning_rate": 1.927290675153987e-05, + "loss": 0.9597, + "step": 1368 + }, + { + "epoch": 0.743213897937025, + "grad_norm": 5.19917061501543, + "learning_rate": 1.927158967708841e-05, + "loss": 1.4333, + "step": 1369 + }, + { + "epoch": 0.743756786102063, + "grad_norm": 4.742331177279521, + "learning_rate": 1.927027145591197e-05, + "loss": 1.1659, + "step": 1370 + }, + { + "epoch": 0.744299674267101, + "grad_norm": 3.713496752478956, + "learning_rate": 1.926895208817358e-05, + "loss": 0.8688, + "step": 1371 + }, + { + "epoch": 0.744842562432139, + "grad_norm": 4.424173231823696, + "learning_rate": 1.9267631574036417e-05, + "loss": 1.0751, + "step": 1372 + }, + { + "epoch": 0.745385450597177, + "grad_norm": 4.7028219143924055, + "learning_rate": 1.9266309913663815e-05, + "loss": 1.3478, + "step": 1373 + }, + { + "epoch": 0.745928338762215, + "grad_norm": 5.498103807374834, + "learning_rate": 1.9264987107219237e-05, + "loss": 1.168, + "step": 1374 + }, + { + "epoch": 0.746471226927253, + "grad_norm": 4.326699254091958, + "learning_rate": 1.9263663154866285e-05, + "loss": 1.0713, + "step": 1375 + }, + { + "epoch": 0.747014115092291, + "grad_norm": 5.58697502498226, + "learning_rate": 1.926233805676871e-05, + "loss": 1.2474, + "step": 1376 + }, + { + "epoch": 0.747557003257329, + "grad_norm": 6.475832093356038, + "learning_rate": 1.92610118130904e-05, + "loss": 1.2805, + "step": 1377 + }, + { + "epoch": 0.748099891422367, + "grad_norm": 5.194166453296571, + "learning_rate": 1.925968442399539e-05, + "loss": 0.9736, + "step": 1378 + }, + { + "epoch": 0.748642779587405, + "grad_norm": 3.867607548943643, + "learning_rate": 1.9258355889647855e-05, + "loss": 1.2325, + "step": 1379 + }, + { + "epoch": 0.749185667752443, + "grad_norm": 5.602424635230674, + "learning_rate": 1.925702621021211e-05, + "loss": 1.0153, + "step": 1380 + }, + { + "epoch": 0.749728555917481, + "grad_norm": 5.011820196133583, + "learning_rate": 1.9255695385852604e-05, + "loss": 1.1823, + "step": 1381 + }, + { + "epoch": 0.750271444082519, + "grad_norm": 5.238515717778492, + "learning_rate": 1.9254363416733944e-05, + "loss": 1.4279, + "step": 1382 + }, + { + "epoch": 0.750814332247557, + "grad_norm": 5.600089025591521, + "learning_rate": 1.925303030302087e-05, + "loss": 0.9212, + "step": 1383 + }, + { + "epoch": 0.751357220412595, + "grad_norm": 4.318444011047328, + "learning_rate": 1.9251696044878255e-05, + "loss": 0.9623, + "step": 1384 + }, + { + "epoch": 0.751900108577633, + "grad_norm": 5.471464995723797, + "learning_rate": 1.925036064247113e-05, + "loss": 1.5446, + "step": 1385 + }, + { + "epoch": 0.752442996742671, + "grad_norm": 5.153911794581426, + "learning_rate": 1.9249024095964663e-05, + "loss": 1.7223, + "step": 1386 + }, + { + "epoch": 0.752985884907709, + "grad_norm": 3.7202924563691537, + "learning_rate": 1.924768640552415e-05, + "loss": 1.1469, + "step": 1387 + }, + { + "epoch": 0.753528773072747, + "grad_norm": 4.612135079041308, + "learning_rate": 1.9246347571315043e-05, + "loss": 0.9572, + "step": 1388 + }, + { + "epoch": 0.754071661237785, + "grad_norm": 4.618898432554816, + "learning_rate": 1.9245007593502937e-05, + "loss": 1.0443, + "step": 1389 + }, + { + "epoch": 0.754614549402823, + "grad_norm": 5.2787203720857345, + "learning_rate": 1.9243666472253554e-05, + "loss": 1.2034, + "step": 1390 + }, + { + "epoch": 0.755157437567861, + "grad_norm": 4.510719578546688, + "learning_rate": 1.9242324207732766e-05, + "loss": 0.8055, + "step": 1391 + }, + { + "epoch": 0.755700325732899, + "grad_norm": 6.256234902874701, + "learning_rate": 1.9240980800106596e-05, + "loss": 1.5985, + "step": 1392 + }, + { + "epoch": 0.756243213897937, + "grad_norm": 9.01213198061367, + "learning_rate": 1.923963624954119e-05, + "loss": 1.7921, + "step": 1393 + }, + { + "epoch": 0.756786102062975, + "grad_norm": 5.149498439254725, + "learning_rate": 1.923829055620285e-05, + "loss": 1.2406, + "step": 1394 + }, + { + "epoch": 0.757328990228013, + "grad_norm": 5.083183084254609, + "learning_rate": 1.9236943720258007e-05, + "loss": 0.9887, + "step": 1395 + }, + { + "epoch": 0.757871878393051, + "grad_norm": 4.5118989088500685, + "learning_rate": 1.9235595741873247e-05, + "loss": 1.3528, + "step": 1396 + }, + { + "epoch": 0.758414766558089, + "grad_norm": 4.583236692880461, + "learning_rate": 1.923424662121528e-05, + "loss": 1.0185, + "step": 1397 + }, + { + "epoch": 0.758957654723127, + "grad_norm": 5.3602373686113625, + "learning_rate": 1.9232896358450976e-05, + "loss": 1.1827, + "step": 1398 + }, + { + "epoch": 0.759500542888165, + "grad_norm": 4.510149132944334, + "learning_rate": 1.9231544953747336e-05, + "loss": 0.9981, + "step": 1399 + }, + { + "epoch": 0.760043431053203, + "grad_norm": 5.008678356958532, + "learning_rate": 1.9230192407271506e-05, + "loss": 1.4957, + "step": 1400 + }, + { + "epoch": 0.760586319218241, + "grad_norm": 4.6938647576746995, + "learning_rate": 1.9228838719190765e-05, + "loss": 1.361, + "step": 1401 + }, + { + "epoch": 0.761129207383279, + "grad_norm": 5.263462060803471, + "learning_rate": 1.9227483889672544e-05, + "loss": 1.1716, + "step": 1402 + }, + { + "epoch": 0.761672095548317, + "grad_norm": 4.737410394333335, + "learning_rate": 1.9226127918884407e-05, + "loss": 1.3924, + "step": 1403 + }, + { + "epoch": 0.762214983713355, + "grad_norm": 5.117476933198257, + "learning_rate": 1.9224770806994066e-05, + "loss": 1.1215, + "step": 1404 + }, + { + "epoch": 0.7627578718783931, + "grad_norm": 4.525492833460315, + "learning_rate": 1.922341255416937e-05, + "loss": 1.0626, + "step": 1405 + }, + { + "epoch": 0.7633007600434311, + "grad_norm": 4.928417801176309, + "learning_rate": 1.9222053160578312e-05, + "loss": 0.9576, + "step": 1406 + }, + { + "epoch": 0.7638436482084691, + "grad_norm": 4.44076437857558, + "learning_rate": 1.9220692626389018e-05, + "loss": 0.9186, + "step": 1407 + }, + { + "epoch": 0.7643865363735071, + "grad_norm": 4.779705057681976, + "learning_rate": 1.9219330951769763e-05, + "loss": 1.3392, + "step": 1408 + }, + { + "epoch": 0.7649294245385451, + "grad_norm": 5.37856562129718, + "learning_rate": 1.9217968136888965e-05, + "loss": 1.043, + "step": 1409 + }, + { + "epoch": 0.7654723127035831, + "grad_norm": 5.596447438931628, + "learning_rate": 1.9216604181915178e-05, + "loss": 1.3223, + "step": 1410 + }, + { + "epoch": 0.7660152008686211, + "grad_norm": 5.267245970848837, + "learning_rate": 1.9215239087017093e-05, + "loss": 1.0484, + "step": 1411 + }, + { + "epoch": 0.7665580890336591, + "grad_norm": 6.124227707475327, + "learning_rate": 1.9213872852363552e-05, + "loss": 1.5361, + "step": 1412 + }, + { + "epoch": 0.7671009771986971, + "grad_norm": 5.632804630433347, + "learning_rate": 1.9212505478123532e-05, + "loss": 1.2227, + "step": 1413 + }, + { + "epoch": 0.7676438653637351, + "grad_norm": 4.903635376409972, + "learning_rate": 1.9211136964466152e-05, + "loss": 0.9045, + "step": 1414 + }, + { + "epoch": 0.7681867535287731, + "grad_norm": 4.995962297168909, + "learning_rate": 1.9209767311560673e-05, + "loss": 1.2364, + "step": 1415 + }, + { + "epoch": 0.7687296416938111, + "grad_norm": 4.76685077713632, + "learning_rate": 1.9208396519576494e-05, + "loss": 1.4849, + "step": 1416 + }, + { + "epoch": 0.7692725298588491, + "grad_norm": 5.66289239913894, + "learning_rate": 1.9207024588683158e-05, + "loss": 1.389, + "step": 1417 + }, + { + "epoch": 0.7698154180238871, + "grad_norm": 4.044195406366437, + "learning_rate": 1.920565151905035e-05, + "loss": 0.5736, + "step": 1418 + }, + { + "epoch": 0.7703583061889251, + "grad_norm": 5.610527750585898, + "learning_rate": 1.9204277310847887e-05, + "loss": 1.5147, + "step": 1419 + }, + { + "epoch": 0.7709011943539631, + "grad_norm": 3.8629108074125424, + "learning_rate": 1.9202901964245734e-05, + "loss": 0.9184, + "step": 1420 + }, + { + "epoch": 0.7714440825190011, + "grad_norm": 4.582445031278247, + "learning_rate": 1.9201525479414e-05, + "loss": 1.071, + "step": 1421 + }, + { + "epoch": 0.7719869706840391, + "grad_norm": 5.016846104390101, + "learning_rate": 1.9200147856522933e-05, + "loss": 1.3673, + "step": 1422 + }, + { + "epoch": 0.7725298588490771, + "grad_norm": 4.798189213061551, + "learning_rate": 1.9198769095742914e-05, + "loss": 1.3483, + "step": 1423 + }, + { + "epoch": 0.7730727470141151, + "grad_norm": 5.871902023790772, + "learning_rate": 1.9197389197244473e-05, + "loss": 1.7625, + "step": 1424 + }, + { + "epoch": 0.7736156351791531, + "grad_norm": 5.895934775040147, + "learning_rate": 1.9196008161198277e-05, + "loss": 0.999, + "step": 1425 + }, + { + "epoch": 0.7741585233441911, + "grad_norm": 5.0199402476408155, + "learning_rate": 1.9194625987775138e-05, + "loss": 1.3251, + "step": 1426 + }, + { + "epoch": 0.7747014115092291, + "grad_norm": 6.650397672217608, + "learning_rate": 1.9193242677146e-05, + "loss": 1.2162, + "step": 1427 + }, + { + "epoch": 0.7752442996742671, + "grad_norm": 5.515267491505962, + "learning_rate": 1.9191858229481958e-05, + "loss": 1.3849, + "step": 1428 + }, + { + "epoch": 0.7757871878393051, + "grad_norm": 5.694611687374825, + "learning_rate": 1.9190472644954236e-05, + "loss": 1.0831, + "step": 1429 + }, + { + "epoch": 0.7763300760043431, + "grad_norm": 4.548114219835821, + "learning_rate": 1.9189085923734215e-05, + "loss": 1.2549, + "step": 1430 + }, + { + "epoch": 0.7768729641693811, + "grad_norm": 6.119997613777156, + "learning_rate": 1.9187698065993398e-05, + "loss": 1.6137, + "step": 1431 + }, + { + "epoch": 0.7774158523344191, + "grad_norm": 4.574150272616086, + "learning_rate": 1.9186309071903445e-05, + "loss": 1.3015, + "step": 1432 + }, + { + "epoch": 0.7779587404994571, + "grad_norm": 4.77854731853541, + "learning_rate": 1.9184918941636142e-05, + "loss": 0.6973, + "step": 1433 + }, + { + "epoch": 0.7785016286644951, + "grad_norm": 6.926122760031406, + "learning_rate": 1.9183527675363425e-05, + "loss": 1.4034, + "step": 1434 + }, + { + "epoch": 0.7790445168295331, + "grad_norm": 4.6748890605309645, + "learning_rate": 1.9182135273257372e-05, + "loss": 0.9854, + "step": 1435 + }, + { + "epoch": 0.7795874049945711, + "grad_norm": 5.072338861625223, + "learning_rate": 1.9180741735490194e-05, + "loss": 1.1604, + "step": 1436 + }, + { + "epoch": 0.7801302931596091, + "grad_norm": 5.995183838581222, + "learning_rate": 1.9179347062234245e-05, + "loss": 1.3913, + "step": 1437 + }, + { + "epoch": 0.7806731813246471, + "grad_norm": 5.494006248546126, + "learning_rate": 1.917795125366202e-05, + "loss": 1.2541, + "step": 1438 + }, + { + "epoch": 0.7812160694896851, + "grad_norm": 3.8223879163574694, + "learning_rate": 1.917655430994616e-05, + "loss": 0.7292, + "step": 1439 + }, + { + "epoch": 0.7817589576547231, + "grad_norm": 4.1476284953657405, + "learning_rate": 1.9175156231259434e-05, + "loss": 0.608, + "step": 1440 + }, + { + "epoch": 0.7823018458197611, + "grad_norm": 5.5144637244676495, + "learning_rate": 1.9173757017774764e-05, + "loss": 1.2674, + "step": 1441 + }, + { + "epoch": 0.7828447339847991, + "grad_norm": 5.4660654663594945, + "learning_rate": 1.9172356669665206e-05, + "loss": 1.3043, + "step": 1442 + }, + { + "epoch": 0.7833876221498371, + "grad_norm": 3.526312966998694, + "learning_rate": 1.9170955187103957e-05, + "loss": 0.8721, + "step": 1443 + }, + { + "epoch": 0.7839305103148752, + "grad_norm": 4.395429767339641, + "learning_rate": 1.9169552570264355e-05, + "loss": 0.985, + "step": 1444 + }, + { + "epoch": 0.7844733984799132, + "grad_norm": 4.608635000415997, + "learning_rate": 1.9168148819319874e-05, + "loss": 1.3492, + "step": 1445 + }, + { + "epoch": 0.7850162866449512, + "grad_norm": 5.729965497961937, + "learning_rate": 1.9166743934444137e-05, + "loss": 1.3405, + "step": 1446 + }, + { + "epoch": 0.7855591748099892, + "grad_norm": 6.60908414105679, + "learning_rate": 1.91653379158109e-05, + "loss": 1.5271, + "step": 1447 + }, + { + "epoch": 0.7861020629750272, + "grad_norm": 4.91386728282805, + "learning_rate": 1.916393076359406e-05, + "loss": 1.475, + "step": 1448 + }, + { + "epoch": 0.7866449511400652, + "grad_norm": 5.286158051966196, + "learning_rate": 1.916252247796766e-05, + "loss": 1.5235, + "step": 1449 + }, + { + "epoch": 0.7871878393051032, + "grad_norm": 4.699459928019414, + "learning_rate": 1.916111305910588e-05, + "loss": 1.151, + "step": 1450 + }, + { + "epoch": 0.7877307274701412, + "grad_norm": 5.125659016648778, + "learning_rate": 1.915970250718303e-05, + "loss": 1.2952, + "step": 1451 + }, + { + "epoch": 0.7882736156351792, + "grad_norm": 4.976170683421487, + "learning_rate": 1.915829082237358e-05, + "loss": 1.3291, + "step": 1452 + }, + { + "epoch": 0.7888165038002172, + "grad_norm": 5.237853683905863, + "learning_rate": 1.9156878004852123e-05, + "loss": 1.4775, + "step": 1453 + }, + { + "epoch": 0.7893593919652552, + "grad_norm": 4.914941394388547, + "learning_rate": 1.9155464054793404e-05, + "loss": 1.2151, + "step": 1454 + }, + { + "epoch": 0.7899022801302932, + "grad_norm": 5.050785426148085, + "learning_rate": 1.9154048972372293e-05, + "loss": 1.163, + "step": 1455 + }, + { + "epoch": 0.7904451682953312, + "grad_norm": 4.2016259300832255, + "learning_rate": 1.915263275776382e-05, + "loss": 0.9601, + "step": 1456 + }, + { + "epoch": 0.7909880564603692, + "grad_norm": 5.48804064939896, + "learning_rate": 1.915121541114314e-05, + "loss": 1.3026, + "step": 1457 + }, + { + "epoch": 0.7915309446254072, + "grad_norm": 5.230495684608947, + "learning_rate": 1.9149796932685552e-05, + "loss": 1.1923, + "step": 1458 + }, + { + "epoch": 0.7920738327904452, + "grad_norm": 4.788655104859546, + "learning_rate": 1.91483773225665e-05, + "loss": 1.2437, + "step": 1459 + }, + { + "epoch": 0.7926167209554832, + "grad_norm": 4.506216689801701, + "learning_rate": 1.9146956580961556e-05, + "loss": 0.9364, + "step": 1460 + }, + { + "epoch": 0.7931596091205212, + "grad_norm": 4.2335852812311865, + "learning_rate": 1.9145534708046446e-05, + "loss": 0.7104, + "step": 1461 + }, + { + "epoch": 0.7937024972855592, + "grad_norm": 6.204688311211956, + "learning_rate": 1.914411170399703e-05, + "loss": 1.0825, + "step": 1462 + }, + { + "epoch": 0.7942453854505972, + "grad_norm": 3.445489329210515, + "learning_rate": 1.91426875689893e-05, + "loss": 0.9921, + "step": 1463 + }, + { + "epoch": 0.7947882736156352, + "grad_norm": 5.204416925095863, + "learning_rate": 1.9141262303199403e-05, + "loss": 1.3043, + "step": 1464 + }, + { + "epoch": 0.7953311617806732, + "grad_norm": 6.3486214559668985, + "learning_rate": 1.9139835906803612e-05, + "loss": 1.3193, + "step": 1465 + }, + { + "epoch": 0.7958740499457112, + "grad_norm": 5.610159156463615, + "learning_rate": 1.913840837997835e-05, + "loss": 1.0455, + "step": 1466 + }, + { + "epoch": 0.7964169381107492, + "grad_norm": 4.625080334899242, + "learning_rate": 1.913697972290018e-05, + "loss": 0.7981, + "step": 1467 + }, + { + "epoch": 0.7969598262757872, + "grad_norm": 5.993005746484773, + "learning_rate": 1.9135549935745792e-05, + "loss": 1.1674, + "step": 1468 + }, + { + "epoch": 0.7975027144408252, + "grad_norm": 5.084136125969368, + "learning_rate": 1.913411901869203e-05, + "loss": 1.0362, + "step": 1469 + }, + { + "epoch": 0.7980456026058632, + "grad_norm": 5.701736254232889, + "learning_rate": 1.913268697191587e-05, + "loss": 1.2159, + "step": 1470 + }, + { + "epoch": 0.7985884907709012, + "grad_norm": 4.715190322082246, + "learning_rate": 1.9131253795594428e-05, + "loss": 0.9848, + "step": 1471 + }, + { + "epoch": 0.7991313789359392, + "grad_norm": 5.375233653105075, + "learning_rate": 1.9129819489904964e-05, + "loss": 1.0476, + "step": 1472 + }, + { + "epoch": 0.7996742671009772, + "grad_norm": 6.352674291337691, + "learning_rate": 1.9128384055024874e-05, + "loss": 1.2362, + "step": 1473 + }, + { + "epoch": 0.8002171552660152, + "grad_norm": 5.46716259791096, + "learning_rate": 1.91269474911317e-05, + "loss": 1.2949, + "step": 1474 + }, + { + "epoch": 0.8007600434310532, + "grad_norm": 5.3150808603597826, + "learning_rate": 1.912550979840311e-05, + "loss": 1.1587, + "step": 1475 + }, + { + "epoch": 0.8013029315960912, + "grad_norm": 4.46814887402293, + "learning_rate": 1.9124070977016926e-05, + "loss": 0.9649, + "step": 1476 + }, + { + "epoch": 0.8018458197611292, + "grad_norm": 4.916267658604107, + "learning_rate": 1.9122631027151103e-05, + "loss": 1.2117, + "step": 1477 + }, + { + "epoch": 0.8023887079261672, + "grad_norm": 5.506988713852874, + "learning_rate": 1.9121189948983733e-05, + "loss": 1.3387, + "step": 1478 + }, + { + "epoch": 0.8029315960912052, + "grad_norm": 4.7704152930487895, + "learning_rate": 1.911974774269305e-05, + "loss": 1.3379, + "step": 1479 + }, + { + "epoch": 0.8034744842562432, + "grad_norm": 5.667769720352476, + "learning_rate": 1.9118304408457435e-05, + "loss": 1.0552, + "step": 1480 + }, + { + "epoch": 0.8040173724212812, + "grad_norm": 4.9471232370904925, + "learning_rate": 1.91168599464554e-05, + "loss": 1.0013, + "step": 1481 + }, + { + "epoch": 0.8045602605863192, + "grad_norm": 4.856623754785127, + "learning_rate": 1.9115414356865594e-05, + "loss": 0.9001, + "step": 1482 + }, + { + "epoch": 0.8051031487513572, + "grad_norm": 4.004976425614515, + "learning_rate": 1.9113967639866815e-05, + "loss": 0.8114, + "step": 1483 + }, + { + "epoch": 0.8056460369163952, + "grad_norm": 5.983237168687411, + "learning_rate": 1.911251979563799e-05, + "loss": 1.1678, + "step": 1484 + }, + { + "epoch": 0.8061889250814332, + "grad_norm": 4.808701495369871, + "learning_rate": 1.9111070824358196e-05, + "loss": 0.9181, + "step": 1485 + }, + { + "epoch": 0.8067318132464713, + "grad_norm": 5.080557830592386, + "learning_rate": 1.910962072620664e-05, + "loss": 1.0751, + "step": 1486 + }, + { + "epoch": 0.8072747014115093, + "grad_norm": 5.730412968009966, + "learning_rate": 1.9108169501362674e-05, + "loss": 1.2727, + "step": 1487 + }, + { + "epoch": 0.8078175895765473, + "grad_norm": 6.47838124684283, + "learning_rate": 1.9106717150005785e-05, + "loss": 1.6491, + "step": 1488 + }, + { + "epoch": 0.8083604777415853, + "grad_norm": 5.311235841870027, + "learning_rate": 1.910526367231561e-05, + "loss": 0.8382, + "step": 1489 + }, + { + "epoch": 0.8089033659066233, + "grad_norm": 5.43072146168114, + "learning_rate": 1.9103809068471914e-05, + "loss": 1.3026, + "step": 1490 + }, + { + "epoch": 0.8094462540716613, + "grad_norm": 4.46156225654375, + "learning_rate": 1.9102353338654597e-05, + "loss": 0.8071, + "step": 1491 + }, + { + "epoch": 0.8099891422366993, + "grad_norm": 5.297520452299967, + "learning_rate": 1.9100896483043714e-05, + "loss": 1.0625, + "step": 1492 + }, + { + "epoch": 0.8105320304017373, + "grad_norm": 6.736735698357962, + "learning_rate": 1.909943850181945e-05, + "loss": 1.6497, + "step": 1493 + }, + { + "epoch": 0.8110749185667753, + "grad_norm": 4.8675910913183955, + "learning_rate": 1.9097979395162132e-05, + "loss": 1.0822, + "step": 1494 + }, + { + "epoch": 0.8116178067318133, + "grad_norm": 4.5489525578867305, + "learning_rate": 1.909651916325222e-05, + "loss": 0.7908, + "step": 1495 + }, + { + "epoch": 0.8121606948968513, + "grad_norm": 6.002121574887706, + "learning_rate": 1.909505780627032e-05, + "loss": 0.5178, + "step": 1496 + }, + { + "epoch": 0.8127035830618893, + "grad_norm": 4.34564271493731, + "learning_rate": 1.9093595324397175e-05, + "loss": 0.9818, + "step": 1497 + }, + { + "epoch": 0.8132464712269273, + "grad_norm": 5.0382562486474525, + "learning_rate": 1.9092131717813668e-05, + "loss": 1.0997, + "step": 1498 + }, + { + "epoch": 0.8137893593919653, + "grad_norm": 6.8289191133749, + "learning_rate": 1.909066698670082e-05, + "loss": 1.0046, + "step": 1499 + }, + { + "epoch": 0.8143322475570033, + "grad_norm": 6.6968746410007585, + "learning_rate": 1.908920113123979e-05, + "loss": 1.5191, + "step": 1500 + }, + { + "epoch": 0.8148751357220413, + "grad_norm": 5.548259576405749, + "learning_rate": 1.9087734151611877e-05, + "loss": 1.3272, + "step": 1501 + }, + { + "epoch": 0.8154180238870793, + "grad_norm": 5.684602861026239, + "learning_rate": 1.9086266047998522e-05, + "loss": 1.0698, + "step": 1502 + }, + { + "epoch": 0.8159609120521173, + "grad_norm": 6.832627276636624, + "learning_rate": 1.90847968205813e-05, + "loss": 1.5796, + "step": 1503 + }, + { + "epoch": 0.8165038002171553, + "grad_norm": 5.605724679057613, + "learning_rate": 1.908332646954193e-05, + "loss": 1.1709, + "step": 1504 + }, + { + "epoch": 0.8170466883821933, + "grad_norm": 4.607345782065147, + "learning_rate": 1.908185499506226e-05, + "loss": 0.8686, + "step": 1505 + }, + { + "epoch": 0.8175895765472313, + "grad_norm": 4.687252800873217, + "learning_rate": 1.9080382397324296e-05, + "loss": 0.8881, + "step": 1506 + }, + { + "epoch": 0.8181324647122693, + "grad_norm": 6.357714385620256, + "learning_rate": 1.907890867651016e-05, + "loss": 1.1948, + "step": 1507 + }, + { + "epoch": 0.8186753528773073, + "grad_norm": 4.8608334992124425, + "learning_rate": 1.9077433832802135e-05, + "loss": 1.0311, + "step": 1508 + }, + { + "epoch": 0.8192182410423453, + "grad_norm": 5.948864135911491, + "learning_rate": 1.9075957866382623e-05, + "loss": 1.1314, + "step": 1509 + }, + { + "epoch": 0.8197611292073833, + "grad_norm": 4.807850657807276, + "learning_rate": 1.9074480777434178e-05, + "loss": 0.9478, + "step": 1510 + }, + { + "epoch": 0.8203040173724213, + "grad_norm": 4.26222909817726, + "learning_rate": 1.9073002566139486e-05, + "loss": 0.8541, + "step": 1511 + }, + { + "epoch": 0.8208469055374593, + "grad_norm": 6.6130092616391005, + "learning_rate": 1.9071523232681382e-05, + "loss": 1.2754, + "step": 1512 + }, + { + "epoch": 0.8213897937024973, + "grad_norm": 5.255155673043404, + "learning_rate": 1.907004277724282e-05, + "loss": 1.0813, + "step": 1513 + }, + { + "epoch": 0.8219326818675353, + "grad_norm": 5.079098182105948, + "learning_rate": 1.9068561200006917e-05, + "loss": 1.0016, + "step": 1514 + }, + { + "epoch": 0.8224755700325733, + "grad_norm": 5.491172879434626, + "learning_rate": 1.906707850115691e-05, + "loss": 1.2884, + "step": 1515 + }, + { + "epoch": 0.8230184581976113, + "grad_norm": 6.718962166599785, + "learning_rate": 1.9065594680876182e-05, + "loss": 1.4973, + "step": 1516 + }, + { + "epoch": 0.8235613463626493, + "grad_norm": 5.348428262646105, + "learning_rate": 1.9064109739348257e-05, + "loss": 1.1113, + "step": 1517 + }, + { + "epoch": 0.8241042345276873, + "grad_norm": 4.604059650726469, + "learning_rate": 1.906262367675679e-05, + "loss": 1.1614, + "step": 1518 + }, + { + "epoch": 0.8246471226927253, + "grad_norm": 5.895860662978225, + "learning_rate": 1.9061136493285586e-05, + "loss": 1.1532, + "step": 1519 + }, + { + "epoch": 0.8251900108577633, + "grad_norm": 5.621847434524929, + "learning_rate": 1.905964818911858e-05, + "loss": 0.9277, + "step": 1520 + }, + { + "epoch": 0.8257328990228013, + "grad_norm": 4.585955492662189, + "learning_rate": 1.9058158764439844e-05, + "loss": 0.7988, + "step": 1521 + }, + { + "epoch": 0.8262757871878393, + "grad_norm": 5.784308925988881, + "learning_rate": 1.9056668219433595e-05, + "loss": 1.6078, + "step": 1522 + }, + { + "epoch": 0.8268186753528773, + "grad_norm": 5.621216253388429, + "learning_rate": 1.905517655428419e-05, + "loss": 1.389, + "step": 1523 + }, + { + "epoch": 0.8273615635179153, + "grad_norm": 5.514208208061458, + "learning_rate": 1.9053683769176115e-05, + "loss": 0.7612, + "step": 1524 + }, + { + "epoch": 0.8279044516829533, + "grad_norm": 4.6572185133859065, + "learning_rate": 1.9052189864294002e-05, + "loss": 0.744, + "step": 1525 + }, + { + "epoch": 0.8284473398479913, + "grad_norm": 5.239632270223703, + "learning_rate": 1.905069483982262e-05, + "loss": 0.939, + "step": 1526 + }, + { + "epoch": 0.8289902280130294, + "grad_norm": 3.7037650901217454, + "learning_rate": 1.9049198695946876e-05, + "loss": 1.0177, + "step": 1527 + }, + { + "epoch": 0.8295331161780674, + "grad_norm": 6.974869726679236, + "learning_rate": 1.9047701432851813e-05, + "loss": 1.2722, + "step": 1528 + }, + { + "epoch": 0.8300760043431054, + "grad_norm": 6.642878483620589, + "learning_rate": 1.904620305072262e-05, + "loss": 1.5369, + "step": 1529 + }, + { + "epoch": 0.8306188925081434, + "grad_norm": 7.99082645392899, + "learning_rate": 1.9044703549744616e-05, + "loss": 1.2245, + "step": 1530 + }, + { + "epoch": 0.8311617806731814, + "grad_norm": 6.5593948883008135, + "learning_rate": 1.904320293010326e-05, + "loss": 1.307, + "step": 1531 + }, + { + "epoch": 0.8317046688382194, + "grad_norm": 4.930764759519961, + "learning_rate": 1.9041701191984155e-05, + "loss": 0.9564, + "step": 1532 + }, + { + "epoch": 0.8322475570032574, + "grad_norm": 3.9621684331427773, + "learning_rate": 1.9040198335573033e-05, + "loss": 0.8153, + "step": 1533 + }, + { + "epoch": 0.8327904451682954, + "grad_norm": 4.859084711241092, + "learning_rate": 1.9038694361055774e-05, + "loss": 1.0967, + "step": 1534 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.9476278172327595, + "learning_rate": 1.903718926861839e-05, + "loss": 1.4743, + "step": 1535 + }, + { + "epoch": 0.8338762214983714, + "grad_norm": 4.312795088472635, + "learning_rate": 1.903568305844704e-05, + "loss": 0.8249, + "step": 1536 + }, + { + "epoch": 0.8344191096634094, + "grad_norm": 6.638027572281857, + "learning_rate": 1.9034175730728e-05, + "loss": 1.6353, + "step": 1537 + }, + { + "epoch": 0.8349619978284474, + "grad_norm": 5.709548912096025, + "learning_rate": 1.9032667285647714e-05, + "loss": 1.8512, + "step": 1538 + }, + { + "epoch": 0.8355048859934854, + "grad_norm": 5.098275426179771, + "learning_rate": 1.9031157723392738e-05, + "loss": 1.0348, + "step": 1539 + }, + { + "epoch": 0.8360477741585234, + "grad_norm": 5.05236269448673, + "learning_rate": 1.9029647044149783e-05, + "loss": 1.108, + "step": 1540 + }, + { + "epoch": 0.8365906623235614, + "grad_norm": 4.674708875536442, + "learning_rate": 1.9028135248105692e-05, + "loss": 1.0453, + "step": 1541 + }, + { + "epoch": 0.8371335504885994, + "grad_norm": 4.717321956389267, + "learning_rate": 1.902662233544744e-05, + "loss": 0.9042, + "step": 1542 + }, + { + "epoch": 0.8376764386536374, + "grad_norm": 6.217040367936212, + "learning_rate": 1.9025108306362158e-05, + "loss": 1.0762, + "step": 1543 + }, + { + "epoch": 0.8382193268186754, + "grad_norm": 5.338744288323046, + "learning_rate": 1.9023593161037094e-05, + "loss": 1.1631, + "step": 1544 + }, + { + "epoch": 0.8387622149837134, + "grad_norm": 4.8224874086687874, + "learning_rate": 1.9022076899659643e-05, + "loss": 1.3907, + "step": 1545 + }, + { + "epoch": 0.8393051031487514, + "grad_norm": 4.699800413179793, + "learning_rate": 1.9020559522417345e-05, + "loss": 0.7682, + "step": 1546 + }, + { + "epoch": 0.8398479913137894, + "grad_norm": 4.271439794022252, + "learning_rate": 1.9019041029497866e-05, + "loss": 0.8475, + "step": 1547 + }, + { + "epoch": 0.8403908794788274, + "grad_norm": 4.240562925498168, + "learning_rate": 1.9017521421089022e-05, + "loss": 0.8201, + "step": 1548 + }, + { + "epoch": 0.8409337676438654, + "grad_norm": 5.24718786534657, + "learning_rate": 1.9016000697378755e-05, + "loss": 1.0728, + "step": 1549 + }, + { + "epoch": 0.8414766558089034, + "grad_norm": 6.492220902503762, + "learning_rate": 1.9014478858555156e-05, + "loss": 0.9432, + "step": 1550 + }, + { + "epoch": 0.8420195439739414, + "grad_norm": 4.917811923935393, + "learning_rate": 1.9012955904806438e-05, + "loss": 1.1672, + "step": 1551 + }, + { + "epoch": 0.8425624321389794, + "grad_norm": 5.66709669173795, + "learning_rate": 1.9011431836320976e-05, + "loss": 1.5058, + "step": 1552 + }, + { + "epoch": 0.8431053203040174, + "grad_norm": 5.658634152413846, + "learning_rate": 1.9009906653287258e-05, + "loss": 1.3653, + "step": 1553 + }, + { + "epoch": 0.8436482084690554, + "grad_norm": 4.956250367213818, + "learning_rate": 1.9008380355893925e-05, + "loss": 0.8309, + "step": 1554 + }, + { + "epoch": 0.8441910966340934, + "grad_norm": 4.329497016180362, + "learning_rate": 1.9006852944329753e-05, + "loss": 1.1141, + "step": 1555 + }, + { + "epoch": 0.8447339847991314, + "grad_norm": 5.717864943326053, + "learning_rate": 1.9005324418783658e-05, + "loss": 1.3274, + "step": 1556 + }, + { + "epoch": 0.8452768729641694, + "grad_norm": 5.657300743895006, + "learning_rate": 1.900379477944468e-05, + "loss": 1.2432, + "step": 1557 + }, + { + "epoch": 0.8458197611292074, + "grad_norm": 5.325761071371062, + "learning_rate": 1.900226402650202e-05, + "loss": 1.3428, + "step": 1558 + }, + { + "epoch": 0.8463626492942454, + "grad_norm": 5.207304704126981, + "learning_rate": 1.9000732160144996e-05, + "loss": 1.455, + "step": 1559 + }, + { + "epoch": 0.8469055374592834, + "grad_norm": 5.248656371435267, + "learning_rate": 1.8999199180563074e-05, + "loss": 0.7851, + "step": 1560 + }, + { + "epoch": 0.8474484256243214, + "grad_norm": 6.92723307445887, + "learning_rate": 1.899766508794585e-05, + "loss": 1.5236, + "step": 1561 + }, + { + "epoch": 0.8479913137893594, + "grad_norm": 4.3434986726191935, + "learning_rate": 1.899612988248307e-05, + "loss": 0.9117, + "step": 1562 + }, + { + "epoch": 0.8485342019543974, + "grad_norm": 5.815800316925401, + "learning_rate": 1.8994593564364612e-05, + "loss": 1.0097, + "step": 1563 + }, + { + "epoch": 0.8490770901194354, + "grad_norm": 5.5494362406379905, + "learning_rate": 1.8993056133780484e-05, + "loss": 1.1829, + "step": 1564 + }, + { + "epoch": 0.8496199782844734, + "grad_norm": 4.323494913553237, + "learning_rate": 1.899151759092084e-05, + "loss": 0.9359, + "step": 1565 + }, + { + "epoch": 0.8501628664495114, + "grad_norm": 4.455911135073797, + "learning_rate": 1.898997793597597e-05, + "loss": 0.9354, + "step": 1566 + }, + { + "epoch": 0.8507057546145494, + "grad_norm": 6.684553981467422, + "learning_rate": 1.8988437169136302e-05, + "loss": 1.3693, + "step": 1567 + }, + { + "epoch": 0.8512486427795874, + "grad_norm": 6.341757024960091, + "learning_rate": 1.89868952905924e-05, + "loss": 0.9441, + "step": 1568 + }, + { + "epoch": 0.8517915309446255, + "grad_norm": 6.017605683279906, + "learning_rate": 1.8985352300534965e-05, + "loss": 1.2747, + "step": 1569 + }, + { + "epoch": 0.8523344191096635, + "grad_norm": 5.501152081837135, + "learning_rate": 1.8983808199154835e-05, + "loss": 0.8414, + "step": 1570 + }, + { + "epoch": 0.8528773072747015, + "grad_norm": 5.075164131477861, + "learning_rate": 1.8982262986642993e-05, + "loss": 1.124, + "step": 1571 + }, + { + "epoch": 0.8534201954397395, + "grad_norm": 5.7750932469486065, + "learning_rate": 1.8980716663190545e-05, + "loss": 1.3968, + "step": 1572 + }, + { + "epoch": 0.8539630836047775, + "grad_norm": 6.967872140444835, + "learning_rate": 1.897916922898875e-05, + "loss": 1.2395, + "step": 1573 + }, + { + "epoch": 0.8545059717698155, + "grad_norm": 5.037804725894728, + "learning_rate": 1.8977620684228994e-05, + "loss": 0.9741, + "step": 1574 + }, + { + "epoch": 0.8550488599348535, + "grad_norm": 4.586000206489282, + "learning_rate": 1.8976071029102802e-05, + "loss": 0.6757, + "step": 1575 + }, + { + "epoch": 0.8555917480998915, + "grad_norm": 4.686417605581938, + "learning_rate": 1.897452026380184e-05, + "loss": 0.8382, + "step": 1576 + }, + { + "epoch": 0.8561346362649295, + "grad_norm": 5.133601528061789, + "learning_rate": 1.8972968388517908e-05, + "loss": 1.2999, + "step": 1577 + }, + { + "epoch": 0.8566775244299675, + "grad_norm": 6.170796021401621, + "learning_rate": 1.8971415403442942e-05, + "loss": 1.1513, + "step": 1578 + }, + { + "epoch": 0.8572204125950055, + "grad_norm": 5.229050623246509, + "learning_rate": 1.8969861308769025e-05, + "loss": 0.9371, + "step": 1579 + }, + { + "epoch": 0.8577633007600435, + "grad_norm": 5.081229057143735, + "learning_rate": 1.8968306104688365e-05, + "loss": 0.9398, + "step": 1580 + }, + { + "epoch": 0.8583061889250815, + "grad_norm": 4.641876540865143, + "learning_rate": 1.896674979139331e-05, + "loss": 1.1401, + "step": 1581 + }, + { + "epoch": 0.8588490770901195, + "grad_norm": 4.149586485324207, + "learning_rate": 1.8965192369076356e-05, + "loss": 0.6768, + "step": 1582 + }, + { + "epoch": 0.8593919652551575, + "grad_norm": 6.422772444923944, + "learning_rate": 1.8963633837930114e-05, + "loss": 1.392, + "step": 1583 + }, + { + "epoch": 0.8599348534201955, + "grad_norm": 4.860635992388028, + "learning_rate": 1.8962074198147357e-05, + "loss": 0.9983, + "step": 1584 + }, + { + "epoch": 0.8604777415852335, + "grad_norm": 4.216877618358879, + "learning_rate": 1.8960513449920982e-05, + "loss": 0.6958, + "step": 1585 + }, + { + "epoch": 0.8610206297502715, + "grad_norm": 5.012340070254638, + "learning_rate": 1.8958951593444017e-05, + "loss": 0.9859, + "step": 1586 + }, + { + "epoch": 0.8615635179153095, + "grad_norm": 5.467321056155664, + "learning_rate": 1.8957388628909644e-05, + "loss": 1.3782, + "step": 1587 + }, + { + "epoch": 0.8621064060803475, + "grad_norm": 5.59589685839463, + "learning_rate": 1.8955824556511168e-05, + "loss": 0.8297, + "step": 1588 + }, + { + "epoch": 0.8626492942453855, + "grad_norm": 6.519624998298618, + "learning_rate": 1.895425937644204e-05, + "loss": 1.7264, + "step": 1589 + }, + { + "epoch": 0.8631921824104235, + "grad_norm": 6.6951270307269795, + "learning_rate": 1.8952693088895837e-05, + "loss": 1.2919, + "step": 1590 + }, + { + "epoch": 0.8637350705754615, + "grad_norm": 4.348741825803296, + "learning_rate": 1.895112569406629e-05, + "loss": 1.0062, + "step": 1591 + }, + { + "epoch": 0.8642779587404995, + "grad_norm": 5.0213162819553565, + "learning_rate": 1.8949557192147243e-05, + "loss": 1.2381, + "step": 1592 + }, + { + "epoch": 0.8648208469055375, + "grad_norm": 6.1637021381056885, + "learning_rate": 1.8947987583332705e-05, + "loss": 1.2561, + "step": 1593 + }, + { + "epoch": 0.8653637350705755, + "grad_norm": 6.638257635344733, + "learning_rate": 1.89464168678168e-05, + "loss": 0.9861, + "step": 1594 + }, + { + "epoch": 0.8659066232356135, + "grad_norm": 4.861237127098975, + "learning_rate": 1.89448450457938e-05, + "loss": 0.9407, + "step": 1595 + }, + { + "epoch": 0.8664495114006515, + "grad_norm": 6.4248430682281565, + "learning_rate": 1.894327211745811e-05, + "loss": 1.3079, + "step": 1596 + }, + { + "epoch": 0.8669923995656895, + "grad_norm": 5.712502208347676, + "learning_rate": 1.8941698083004265e-05, + "loss": 1.4422, + "step": 1597 + }, + { + "epoch": 0.8675352877307275, + "grad_norm": 5.634308307822415, + "learning_rate": 1.8940122942626957e-05, + "loss": 0.7869, + "step": 1598 + }, + { + "epoch": 0.8680781758957655, + "grad_norm": 7.839535783401759, + "learning_rate": 1.893854669652099e-05, + "loss": 1.3397, + "step": 1599 + }, + { + "epoch": 0.8686210640608035, + "grad_norm": 5.156814072355032, + "learning_rate": 1.8936969344881323e-05, + "loss": 1.2541, + "step": 1600 + }, + { + "epoch": 0.8691639522258415, + "grad_norm": 4.912293322601371, + "learning_rate": 1.8935390887903044e-05, + "loss": 0.8418, + "step": 1601 + }, + { + "epoch": 0.8697068403908795, + "grad_norm": 5.436641391280079, + "learning_rate": 1.8933811325781382e-05, + "loss": 1.4456, + "step": 1602 + }, + { + "epoch": 0.8702497285559175, + "grad_norm": 5.221596044730403, + "learning_rate": 1.8932230658711696e-05, + "loss": 0.9626, + "step": 1603 + }, + { + "epoch": 0.8707926167209555, + "grad_norm": 6.103653494209826, + "learning_rate": 1.8930648886889482e-05, + "loss": 1.2338, + "step": 1604 + }, + { + "epoch": 0.8713355048859935, + "grad_norm": 7.003639368124227, + "learning_rate": 1.8929066010510383e-05, + "loss": 1.2216, + "step": 1605 + }, + { + "epoch": 0.8718783930510315, + "grad_norm": 4.911905632269173, + "learning_rate": 1.8927482029770168e-05, + "loss": 0.9049, + "step": 1606 + }, + { + "epoch": 0.8724212812160695, + "grad_norm": 6.079178298711795, + "learning_rate": 1.8925896944864748e-05, + "loss": 1.6408, + "step": 1607 + }, + { + "epoch": 0.8729641693811075, + "grad_norm": 6.045568168649525, + "learning_rate": 1.892431075599017e-05, + "loss": 1.084, + "step": 1608 + }, + { + "epoch": 0.8735070575461455, + "grad_norm": 4.827633470123435, + "learning_rate": 1.892272346334261e-05, + "loss": 0.8121, + "step": 1609 + }, + { + "epoch": 0.8740499457111836, + "grad_norm": 5.314535878915851, + "learning_rate": 1.8921135067118396e-05, + "loss": 1.2657, + "step": 1610 + }, + { + "epoch": 0.8745928338762216, + "grad_norm": 4.786606110240068, + "learning_rate": 1.8919545567513976e-05, + "loss": 1.0534, + "step": 1611 + }, + { + "epoch": 0.8751357220412594, + "grad_norm": 4.075700855026623, + "learning_rate": 1.8917954964725948e-05, + "loss": 0.7757, + "step": 1612 + }, + { + "epoch": 0.8756786102062974, + "grad_norm": 6.286980562376595, + "learning_rate": 1.8916363258951033e-05, + "loss": 0.9782, + "step": 1613 + }, + { + "epoch": 0.8762214983713354, + "grad_norm": 6.745303990081763, + "learning_rate": 1.8914770450386102e-05, + "loss": 0.9488, + "step": 1614 + }, + { + "epoch": 0.8767643865363735, + "grad_norm": 6.2663007227218275, + "learning_rate": 1.8913176539228152e-05, + "loss": 1.9529, + "step": 1615 + }, + { + "epoch": 0.8773072747014115, + "grad_norm": 4.793153881320079, + "learning_rate": 1.8911581525674324e-05, + "loss": 1.0242, + "step": 1616 + }, + { + "epoch": 0.8778501628664495, + "grad_norm": 4.6844946413894, + "learning_rate": 1.890998540992189e-05, + "loss": 0.8565, + "step": 1617 + }, + { + "epoch": 0.8783930510314875, + "grad_norm": 3.993737542716379, + "learning_rate": 1.8908388192168256e-05, + "loss": 0.6107, + "step": 1618 + }, + { + "epoch": 0.8789359391965255, + "grad_norm": 6.099718972513945, + "learning_rate": 1.8906789872610977e-05, + "loss": 0.8043, + "step": 1619 + }, + { + "epoch": 0.8794788273615635, + "grad_norm": 7.113418049135383, + "learning_rate": 1.8905190451447726e-05, + "loss": 1.241, + "step": 1620 + }, + { + "epoch": 0.8800217155266015, + "grad_norm": 4.439254236407197, + "learning_rate": 1.8903589928876337e-05, + "loss": 1.0627, + "step": 1621 + }, + { + "epoch": 0.8805646036916395, + "grad_norm": 5.773948255627926, + "learning_rate": 1.8901988305094746e-05, + "loss": 1.3241, + "step": 1622 + }, + { + "epoch": 0.8811074918566775, + "grad_norm": 6.231870498005159, + "learning_rate": 1.890038558030106e-05, + "loss": 1.5241, + "step": 1623 + }, + { + "epoch": 0.8816503800217155, + "grad_norm": 5.888612961801449, + "learning_rate": 1.8898781754693495e-05, + "loss": 1.109, + "step": 1624 + }, + { + "epoch": 0.8821932681867535, + "grad_norm": 4.880160750019388, + "learning_rate": 1.8897176828470424e-05, + "loss": 1.0124, + "step": 1625 + }, + { + "epoch": 0.8827361563517915, + "grad_norm": 6.365487465081988, + "learning_rate": 1.889557080183034e-05, + "loss": 0.8999, + "step": 1626 + }, + { + "epoch": 0.8832790445168295, + "grad_norm": 4.461166628483842, + "learning_rate": 1.8893963674971883e-05, + "loss": 0.8762, + "step": 1627 + }, + { + "epoch": 0.8838219326818675, + "grad_norm": 5.675449614336963, + "learning_rate": 1.8892355448093825e-05, + "loss": 1.049, + "step": 1628 + }, + { + "epoch": 0.8843648208469055, + "grad_norm": 4.320721965444944, + "learning_rate": 1.8890746121395072e-05, + "loss": 1.1291, + "step": 1629 + }, + { + "epoch": 0.8849077090119435, + "grad_norm": 5.155835351166163, + "learning_rate": 1.8889135695074668e-05, + "loss": 1.1035, + "step": 1630 + }, + { + "epoch": 0.8854505971769815, + "grad_norm": 6.506050711845877, + "learning_rate": 1.8887524169331794e-05, + "loss": 2.1954, + "step": 1631 + }, + { + "epoch": 0.8859934853420195, + "grad_norm": 5.514989348863087, + "learning_rate": 1.8885911544365766e-05, + "loss": 1.0237, + "step": 1632 + }, + { + "epoch": 0.8865363735070575, + "grad_norm": 5.123449587076818, + "learning_rate": 1.8884297820376038e-05, + "loss": 0.9908, + "step": 1633 + }, + { + "epoch": 0.8870792616720955, + "grad_norm": 5.421671567850239, + "learning_rate": 1.8882682997562197e-05, + "loss": 1.7734, + "step": 1634 + }, + { + "epoch": 0.8876221498371335, + "grad_norm": 7.1012436118787186, + "learning_rate": 1.8881067076123963e-05, + "loss": 1.5488, + "step": 1635 + }, + { + "epoch": 0.8881650380021715, + "grad_norm": 5.328780061143714, + "learning_rate": 1.88794500562612e-05, + "loss": 1.3832, + "step": 1636 + }, + { + "epoch": 0.8887079261672095, + "grad_norm": 5.1503624666971595, + "learning_rate": 1.88778319381739e-05, + "loss": 0.8524, + "step": 1637 + }, + { + "epoch": 0.8892508143322475, + "grad_norm": 5.131553278195334, + "learning_rate": 1.88762127220622e-05, + "loss": 0.9769, + "step": 1638 + }, + { + "epoch": 0.8897937024972855, + "grad_norm": 3.1519686008526135, + "learning_rate": 1.8874592408126365e-05, + "loss": 0.472, + "step": 1639 + }, + { + "epoch": 0.8903365906623235, + "grad_norm": 4.164586140054025, + "learning_rate": 1.8872970996566794e-05, + "loss": 0.67, + "step": 1640 + }, + { + "epoch": 0.8908794788273615, + "grad_norm": 7.488659246842809, + "learning_rate": 1.8871348487584028e-05, + "loss": 1.3141, + "step": 1641 + }, + { + "epoch": 0.8914223669923995, + "grad_norm": 5.911298710361154, + "learning_rate": 1.8869724881378743e-05, + "loss": 1.084, + "step": 1642 + }, + { + "epoch": 0.8919652551574375, + "grad_norm": 5.977793625957034, + "learning_rate": 1.886810017815175e-05, + "loss": 0.9273, + "step": 1643 + }, + { + "epoch": 0.8925081433224755, + "grad_norm": 5.2773273690434515, + "learning_rate": 1.8866474378103993e-05, + "loss": 1.2594, + "step": 1644 + }, + { + "epoch": 0.8930510314875135, + "grad_norm": 8.049487951903771, + "learning_rate": 1.8864847481436554e-05, + "loss": 1.0104, + "step": 1645 + }, + { + "epoch": 0.8935939196525515, + "grad_norm": 5.428394606766542, + "learning_rate": 1.886321948835065e-05, + "loss": 0.827, + "step": 1646 + }, + { + "epoch": 0.8941368078175895, + "grad_norm": 5.687133062399138, + "learning_rate": 1.8861590399047635e-05, + "loss": 1.2759, + "step": 1647 + }, + { + "epoch": 0.8946796959826275, + "grad_norm": 5.037591954761082, + "learning_rate": 1.885996021372899e-05, + "loss": 0.8928, + "step": 1648 + }, + { + "epoch": 0.8952225841476655, + "grad_norm": 4.911533421640023, + "learning_rate": 1.8858328932596352e-05, + "loss": 0.705, + "step": 1649 + }, + { + "epoch": 0.8957654723127035, + "grad_norm": 6.054608860222289, + "learning_rate": 1.885669655585147e-05, + "loss": 1.296, + "step": 1650 + }, + { + "epoch": 0.8963083604777415, + "grad_norm": 6.221099773279408, + "learning_rate": 1.8855063083696244e-05, + "loss": 1.5253, + "step": 1651 + }, + { + "epoch": 0.8968512486427795, + "grad_norm": 5.0718329796444035, + "learning_rate": 1.8853428516332702e-05, + "loss": 1.1683, + "step": 1652 + }, + { + "epoch": 0.8973941368078175, + "grad_norm": 5.721985007773526, + "learning_rate": 1.8851792853963015e-05, + "loss": 1.1635, + "step": 1653 + }, + { + "epoch": 0.8979370249728555, + "grad_norm": 7.224045168649458, + "learning_rate": 1.8850156096789473e-05, + "loss": 1.0042, + "step": 1654 + }, + { + "epoch": 0.8984799131378935, + "grad_norm": 4.938883702947416, + "learning_rate": 1.8848518245014526e-05, + "loss": 0.7712, + "step": 1655 + }, + { + "epoch": 0.8990228013029316, + "grad_norm": 5.9261619850877, + "learning_rate": 1.8846879298840735e-05, + "loss": 1.4387, + "step": 1656 + }, + { + "epoch": 0.8995656894679696, + "grad_norm": 5.181216185736725, + "learning_rate": 1.8845239258470817e-05, + "loss": 0.9389, + "step": 1657 + }, + { + "epoch": 0.9001085776330076, + "grad_norm": 5.108249236166185, + "learning_rate": 1.8843598124107608e-05, + "loss": 1.0034, + "step": 1658 + }, + { + "epoch": 0.9006514657980456, + "grad_norm": 4.839799276438571, + "learning_rate": 1.8841955895954088e-05, + "loss": 0.8524, + "step": 1659 + }, + { + "epoch": 0.9011943539630836, + "grad_norm": 5.598018889964705, + "learning_rate": 1.8840312574213372e-05, + "loss": 1.0939, + "step": 1660 + }, + { + "epoch": 0.9017372421281216, + "grad_norm": 5.2496226470868725, + "learning_rate": 1.8838668159088707e-05, + "loss": 1.0812, + "step": 1661 + }, + { + "epoch": 0.9022801302931596, + "grad_norm": 5.2298567821168, + "learning_rate": 1.8837022650783477e-05, + "loss": 0.953, + "step": 1662 + }, + { + "epoch": 0.9028230184581976, + "grad_norm": 4.471282412199049, + "learning_rate": 1.88353760495012e-05, + "loss": 1.2652, + "step": 1663 + }, + { + "epoch": 0.9033659066232356, + "grad_norm": 5.892082884595572, + "learning_rate": 1.8833728355445534e-05, + "loss": 1.1165, + "step": 1664 + }, + { + "epoch": 0.9039087947882736, + "grad_norm": 5.49174731306936, + "learning_rate": 1.8832079568820268e-05, + "loss": 0.9262, + "step": 1665 + }, + { + "epoch": 0.9044516829533116, + "grad_norm": 5.55211658654164, + "learning_rate": 1.883042968982932e-05, + "loss": 1.3112, + "step": 1666 + }, + { + "epoch": 0.9049945711183496, + "grad_norm": 5.597392580982551, + "learning_rate": 1.8828778718676757e-05, + "loss": 0.8511, + "step": 1667 + }, + { + "epoch": 0.9055374592833876, + "grad_norm": 6.183127994440369, + "learning_rate": 1.8827126655566773e-05, + "loss": 1.2784, + "step": 1668 + }, + { + "epoch": 0.9060803474484256, + "grad_norm": 6.991796130593696, + "learning_rate": 1.882547350070369e-05, + "loss": 1.1703, + "step": 1669 + }, + { + "epoch": 0.9066232356134636, + "grad_norm": 5.7794859741652305, + "learning_rate": 1.8823819254291986e-05, + "loss": 1.0974, + "step": 1670 + }, + { + "epoch": 0.9071661237785016, + "grad_norm": 5.388338879165358, + "learning_rate": 1.8822163916536245e-05, + "loss": 1.2087, + "step": 1671 + }, + { + "epoch": 0.9077090119435396, + "grad_norm": 5.072420136637763, + "learning_rate": 1.8820507487641218e-05, + "loss": 0.9244, + "step": 1672 + }, + { + "epoch": 0.9082519001085776, + "grad_norm": 5.9571835605705115, + "learning_rate": 1.8818849967811762e-05, + "loss": 1.3299, + "step": 1673 + }, + { + "epoch": 0.9087947882736156, + "grad_norm": 5.706573186001289, + "learning_rate": 1.8817191357252892e-05, + "loss": 1.1059, + "step": 1674 + }, + { + "epoch": 0.9093376764386536, + "grad_norm": 5.4984127833712435, + "learning_rate": 1.8815531656169737e-05, + "loss": 1.1428, + "step": 1675 + }, + { + "epoch": 0.9098805646036916, + "grad_norm": 4.824306323799433, + "learning_rate": 1.8813870864767582e-05, + "loss": 0.7313, + "step": 1676 + }, + { + "epoch": 0.9104234527687296, + "grad_norm": 6.180183255961245, + "learning_rate": 1.8812208983251828e-05, + "loss": 1.0729, + "step": 1677 + }, + { + "epoch": 0.9109663409337676, + "grad_norm": 6.586598617269493, + "learning_rate": 1.8810546011828024e-05, + "loss": 1.5871, + "step": 1678 + }, + { + "epoch": 0.9115092290988056, + "grad_norm": 5.5611281281711875, + "learning_rate": 1.8808881950701845e-05, + "loss": 0.8133, + "step": 1679 + }, + { + "epoch": 0.9120521172638436, + "grad_norm": 5.214863992160379, + "learning_rate": 1.8807216800079108e-05, + "loss": 1.0437, + "step": 1680 + }, + { + "epoch": 0.9125950054288816, + "grad_norm": 6.631045200232836, + "learning_rate": 1.8805550560165763e-05, + "loss": 1.446, + "step": 1681 + }, + { + "epoch": 0.9131378935939196, + "grad_norm": 6.140822200691373, + "learning_rate": 1.8803883231167887e-05, + "loss": 1.8672, + "step": 1682 + }, + { + "epoch": 0.9136807817589576, + "grad_norm": 4.5606387908479995, + "learning_rate": 1.8802214813291708e-05, + "loss": 1.0662, + "step": 1683 + }, + { + "epoch": 0.9142236699239956, + "grad_norm": 4.166651706834997, + "learning_rate": 1.8800545306743567e-05, + "loss": 0.749, + "step": 1684 + }, + { + "epoch": 0.9147665580890336, + "grad_norm": 4.674238905131916, + "learning_rate": 1.8798874711729957e-05, + "loss": 1.1106, + "step": 1685 + }, + { + "epoch": 0.9153094462540716, + "grad_norm": 5.224600466137879, + "learning_rate": 1.8797203028457497e-05, + "loss": 0.9652, + "step": 1686 + }, + { + "epoch": 0.9158523344191096, + "grad_norm": 5.87706899393542, + "learning_rate": 1.8795530257132947e-05, + "loss": 1.2146, + "step": 1687 + }, + { + "epoch": 0.9163952225841476, + "grad_norm": 7.587853042260599, + "learning_rate": 1.87938563979632e-05, + "loss": 1.9051, + "step": 1688 + }, + { + "epoch": 0.9169381107491856, + "grad_norm": 4.988334377581133, + "learning_rate": 1.8792181451155275e-05, + "loss": 0.8502, + "step": 1689 + }, + { + "epoch": 0.9174809989142236, + "grad_norm": 5.454638378956608, + "learning_rate": 1.8790505416916338e-05, + "loss": 1.0493, + "step": 1690 + }, + { + "epoch": 0.9180238870792616, + "grad_norm": 5.95672873256444, + "learning_rate": 1.878882829545368e-05, + "loss": 1.1296, + "step": 1691 + }, + { + "epoch": 0.9185667752442996, + "grad_norm": 6.921816321230723, + "learning_rate": 1.8787150086974734e-05, + "loss": 1.6388, + "step": 1692 + }, + { + "epoch": 0.9191096634093376, + "grad_norm": 5.423998305707574, + "learning_rate": 1.878547079168706e-05, + "loss": 1.4865, + "step": 1693 + }, + { + "epoch": 0.9196525515743756, + "grad_norm": 4.479238264802495, + "learning_rate": 1.878379040979835e-05, + "loss": 1.0257, + "step": 1694 + }, + { + "epoch": 0.9201954397394136, + "grad_norm": 6.325821606507086, + "learning_rate": 1.8782108941516446e-05, + "loss": 1.1156, + "step": 1695 + }, + { + "epoch": 0.9207383279044516, + "grad_norm": 6.220712298120886, + "learning_rate": 1.8780426387049315e-05, + "loss": 1.375, + "step": 1696 + }, + { + "epoch": 0.9212812160694897, + "grad_norm": 5.813906943099404, + "learning_rate": 1.877874274660505e-05, + "loss": 1.1869, + "step": 1697 + }, + { + "epoch": 0.9218241042345277, + "grad_norm": 5.0160254446208965, + "learning_rate": 1.8777058020391893e-05, + "loss": 1.0443, + "step": 1698 + }, + { + "epoch": 0.9223669923995657, + "grad_norm": 4.273981514908416, + "learning_rate": 1.877537220861821e-05, + "loss": 1.0015, + "step": 1699 + }, + { + "epoch": 0.9229098805646037, + "grad_norm": 4.908796260576538, + "learning_rate": 1.8773685311492513e-05, + "loss": 1.0203, + "step": 1700 + }, + { + "epoch": 0.9234527687296417, + "grad_norm": 5.566794452728763, + "learning_rate": 1.8771997329223425e-05, + "loss": 1.4637, + "step": 1701 + }, + { + "epoch": 0.9239956568946797, + "grad_norm": 4.946565856604512, + "learning_rate": 1.8770308262019733e-05, + "loss": 0.8093, + "step": 1702 + }, + { + "epoch": 0.9245385450597177, + "grad_norm": 4.767566465605265, + "learning_rate": 1.8768618110090334e-05, + "loss": 0.5938, + "step": 1703 + }, + { + "epoch": 0.9250814332247557, + "grad_norm": 6.919371263621627, + "learning_rate": 1.8766926873644272e-05, + "loss": 0.9857, + "step": 1704 + }, + { + "epoch": 0.9256243213897937, + "grad_norm": 3.5297488420709944, + "learning_rate": 1.876523455289072e-05, + "loss": 0.8744, + "step": 1705 + }, + { + "epoch": 0.9261672095548317, + "grad_norm": 4.910386600622229, + "learning_rate": 1.8763541148038994e-05, + "loss": 0.9491, + "step": 1706 + }, + { + "epoch": 0.9267100977198697, + "grad_norm": 5.550892558363115, + "learning_rate": 1.876184665929853e-05, + "loss": 1.2793, + "step": 1707 + }, + { + "epoch": 0.9272529858849077, + "grad_norm": 4.470721855980947, + "learning_rate": 1.8760151086878905e-05, + "loss": 0.9706, + "step": 1708 + }, + { + "epoch": 0.9277958740499457, + "grad_norm": 7.1893201073258926, + "learning_rate": 1.8758454430989833e-05, + "loss": 1.2473, + "step": 1709 + }, + { + "epoch": 0.9283387622149837, + "grad_norm": 5.933304434057961, + "learning_rate": 1.875675669184116e-05, + "loss": 1.4917, + "step": 1710 + }, + { + "epoch": 0.9288816503800217, + "grad_norm": 5.386119916834561, + "learning_rate": 1.8755057869642857e-05, + "loss": 0.9417, + "step": 1711 + }, + { + "epoch": 0.9294245385450597, + "grad_norm": 6.121437468126703, + "learning_rate": 1.875335796460505e-05, + "loss": 1.203, + "step": 1712 + }, + { + "epoch": 0.9299674267100977, + "grad_norm": 5.34957970840911, + "learning_rate": 1.8751656976937974e-05, + "loss": 1.2557, + "step": 1713 + }, + { + "epoch": 0.9305103148751357, + "grad_norm": 7.404283054108375, + "learning_rate": 1.8749954906852023e-05, + "loss": 1.449, + "step": 1714 + }, + { + "epoch": 0.9310532030401737, + "grad_norm": 6.5511839319022585, + "learning_rate": 1.8748251754557696e-05, + "loss": 1.5548, + "step": 1715 + }, + { + "epoch": 0.9315960912052117, + "grad_norm": 5.0254557279034815, + "learning_rate": 1.8746547520265654e-05, + "loss": 0.9777, + "step": 1716 + }, + { + "epoch": 0.9321389793702497, + "grad_norm": 6.385252198006408, + "learning_rate": 1.874484220418667e-05, + "loss": 1.254, + "step": 1717 + }, + { + "epoch": 0.9326818675352877, + "grad_norm": 6.872281478717196, + "learning_rate": 1.874313580653167e-05, + "loss": 1.278, + "step": 1718 + }, + { + "epoch": 0.9332247557003257, + "grad_norm": 8.525842687889295, + "learning_rate": 1.8741428327511696e-05, + "loss": 1.3712, + "step": 1719 + }, + { + "epoch": 0.9337676438653637, + "grad_norm": 7.97420765579221, + "learning_rate": 1.8739719767337933e-05, + "loss": 1.7493, + "step": 1720 + }, + { + "epoch": 0.9343105320304017, + "grad_norm": 6.107109526814546, + "learning_rate": 1.8738010126221705e-05, + "loss": 0.9861, + "step": 1721 + }, + { + "epoch": 0.9348534201954397, + "grad_norm": 6.4666849584489645, + "learning_rate": 1.8736299404374453e-05, + "loss": 0.8041, + "step": 1722 + }, + { + "epoch": 0.9353963083604777, + "grad_norm": 7.489590627219774, + "learning_rate": 1.873458760200777e-05, + "loss": 1.2325, + "step": 1723 + }, + { + "epoch": 0.9359391965255157, + "grad_norm": 7.020182643385788, + "learning_rate": 1.8732874719333373e-05, + "loss": 1.1698, + "step": 1724 + }, + { + "epoch": 0.9364820846905537, + "grad_norm": 6.436449840675483, + "learning_rate": 1.873116075656311e-05, + "loss": 1.084, + "step": 1725 + }, + { + "epoch": 0.9370249728555917, + "grad_norm": 7.27100690341834, + "learning_rate": 1.872944571390897e-05, + "loss": 0.9793, + "step": 1726 + }, + { + "epoch": 0.9375678610206297, + "grad_norm": 6.337617737677888, + "learning_rate": 1.872772959158307e-05, + "loss": 1.0558, + "step": 1727 + }, + { + "epoch": 0.9381107491856677, + "grad_norm": 5.356493873446841, + "learning_rate": 1.8726012389797667e-05, + "loss": 1.1031, + "step": 1728 + }, + { + "epoch": 0.9386536373507057, + "grad_norm": 6.302965546816682, + "learning_rate": 1.8724294108765142e-05, + "loss": 1.1401, + "step": 1729 + }, + { + "epoch": 0.9391965255157437, + "grad_norm": 6.116650787967334, + "learning_rate": 1.872257474869802e-05, + "loss": 1.0848, + "step": 1730 + }, + { + "epoch": 0.9397394136807817, + "grad_norm": 4.3874746967683, + "learning_rate": 1.8720854309808948e-05, + "loss": 0.7463, + "step": 1731 + }, + { + "epoch": 0.9402823018458197, + "grad_norm": 5.00673534942558, + "learning_rate": 1.871913279231072e-05, + "loss": 1.2365, + "step": 1732 + }, + { + "epoch": 0.9408251900108577, + "grad_norm": 5.222845312743783, + "learning_rate": 1.871741019641625e-05, + "loss": 1.062, + "step": 1733 + }, + { + "epoch": 0.9413680781758957, + "grad_norm": 4.624403224726164, + "learning_rate": 1.871568652233859e-05, + "loss": 1.0155, + "step": 1734 + }, + { + "epoch": 0.9419109663409337, + "grad_norm": 5.386412889478532, + "learning_rate": 1.8713961770290936e-05, + "loss": 1.4316, + "step": 1735 + }, + { + "epoch": 0.9424538545059717, + "grad_norm": 6.0899860043030705, + "learning_rate": 1.87122359404866e-05, + "loss": 1.1372, + "step": 1736 + }, + { + "epoch": 0.9429967426710097, + "grad_norm": 5.306362358042622, + "learning_rate": 1.8710509033139037e-05, + "loss": 0.9353, + "step": 1737 + }, + { + "epoch": 0.9435396308360477, + "grad_norm": 6.479420699028343, + "learning_rate": 1.8708781048461832e-05, + "loss": 1.2435, + "step": 1738 + }, + { + "epoch": 0.9440825190010858, + "grad_norm": 6.205746500302448, + "learning_rate": 1.8707051986668712e-05, + "loss": 0.7872, + "step": 1739 + }, + { + "epoch": 0.9446254071661238, + "grad_norm": 4.157411034856641, + "learning_rate": 1.8705321847973523e-05, + "loss": 0.8004, + "step": 1740 + }, + { + "epoch": 0.9451682953311618, + "grad_norm": 5.5814599600578285, + "learning_rate": 1.8703590632590254e-05, + "loss": 0.7384, + "step": 1741 + }, + { + "epoch": 0.9457111834961998, + "grad_norm": 6.040759722497843, + "learning_rate": 1.8701858340733023e-05, + "loss": 1.5941, + "step": 1742 + }, + { + "epoch": 0.9462540716612378, + "grad_norm": 4.592040464655725, + "learning_rate": 1.8700124972616085e-05, + "loss": 1.0662, + "step": 1743 + }, + { + "epoch": 0.9467969598262758, + "grad_norm": 6.077319934267432, + "learning_rate": 1.8698390528453823e-05, + "loss": 1.3726, + "step": 1744 + }, + { + "epoch": 0.9473398479913138, + "grad_norm": 4.255180894541495, + "learning_rate": 1.869665500846076e-05, + "loss": 0.7001, + "step": 1745 + }, + { + "epoch": 0.9478827361563518, + "grad_norm": 6.390593673413393, + "learning_rate": 1.869491841285154e-05, + "loss": 1.2268, + "step": 1746 + }, + { + "epoch": 0.9484256243213898, + "grad_norm": 5.347324129941674, + "learning_rate": 1.8693180741840957e-05, + "loss": 1.2193, + "step": 1747 + }, + { + "epoch": 0.9489685124864278, + "grad_norm": 6.407261723862006, + "learning_rate": 1.8691441995643927e-05, + "loss": 0.868, + "step": 1748 + }, + { + "epoch": 0.9495114006514658, + "grad_norm": 4.400520878629861, + "learning_rate": 1.8689702174475496e-05, + "loss": 0.4679, + "step": 1749 + }, + { + "epoch": 0.9500542888165038, + "grad_norm": 5.6161548427455505, + "learning_rate": 1.8687961278550852e-05, + "loss": 0.8842, + "step": 1750 + }, + { + "epoch": 0.9505971769815418, + "grad_norm": 4.7163538240048, + "learning_rate": 1.8686219308085306e-05, + "loss": 0.9972, + "step": 1751 + }, + { + "epoch": 0.9511400651465798, + "grad_norm": 4.845854095585142, + "learning_rate": 1.8684476263294318e-05, + "loss": 1.1046, + "step": 1752 + }, + { + "epoch": 0.9516829533116178, + "grad_norm": 3.878984502983999, + "learning_rate": 1.8682732144393463e-05, + "loss": 0.7873, + "step": 1753 + }, + { + "epoch": 0.9522258414766558, + "grad_norm": 5.736472248757515, + "learning_rate": 1.8680986951598458e-05, + "loss": 1.2046, + "step": 1754 + }, + { + "epoch": 0.9527687296416938, + "grad_norm": 5.12988223940441, + "learning_rate": 1.867924068512515e-05, + "loss": 0.6293, + "step": 1755 + }, + { + "epoch": 0.9533116178067318, + "grad_norm": 7.822467968073818, + "learning_rate": 1.867749334518952e-05, + "loss": 1.3625, + "step": 1756 + }, + { + "epoch": 0.9538545059717698, + "grad_norm": 5.488388947029871, + "learning_rate": 1.8675744932007687e-05, + "loss": 1.078, + "step": 1757 + }, + { + "epoch": 0.9543973941368078, + "grad_norm": 4.631848438676548, + "learning_rate": 1.8673995445795894e-05, + "loss": 0.69, + "step": 1758 + }, + { + "epoch": 0.9549402823018458, + "grad_norm": 7.078171019993124, + "learning_rate": 1.8672244886770516e-05, + "loss": 1.1036, + "step": 1759 + }, + { + "epoch": 0.9554831704668838, + "grad_norm": 5.9259972832349215, + "learning_rate": 1.8670493255148073e-05, + "loss": 0.7919, + "step": 1760 + }, + { + "epoch": 0.9560260586319218, + "grad_norm": 5.846343833877514, + "learning_rate": 1.8668740551145205e-05, + "loss": 1.1653, + "step": 1761 + }, + { + "epoch": 0.9565689467969598, + "grad_norm": 5.559082125902064, + "learning_rate": 1.8666986774978685e-05, + "loss": 1.4214, + "step": 1762 + }, + { + "epoch": 0.9571118349619978, + "grad_norm": 6.511207981127819, + "learning_rate": 1.8665231926865433e-05, + "loss": 1.3552, + "step": 1763 + }, + { + "epoch": 0.9576547231270358, + "grad_norm": 5.728999474717542, + "learning_rate": 1.8663476007022482e-05, + "loss": 0.7861, + "step": 1764 + }, + { + "epoch": 0.9581976112920738, + "grad_norm": 6.152118564149309, + "learning_rate": 1.8661719015667016e-05, + "loss": 0.8679, + "step": 1765 + }, + { + "epoch": 0.9587404994571118, + "grad_norm": 6.77121947972117, + "learning_rate": 1.8659960953016334e-05, + "loss": 0.8103, + "step": 1766 + }, + { + "epoch": 0.9592833876221498, + "grad_norm": 7.822575173268641, + "learning_rate": 1.865820181928788e-05, + "loss": 1.2558, + "step": 1767 + }, + { + "epoch": 0.9598262757871878, + "grad_norm": 6.389073592671253, + "learning_rate": 1.8656441614699225e-05, + "loss": 1.0314, + "step": 1768 + }, + { + "epoch": 0.9603691639522258, + "grad_norm": 6.141689972496849, + "learning_rate": 1.8654680339468076e-05, + "loss": 1.2452, + "step": 1769 + }, + { + "epoch": 0.9609120521172638, + "grad_norm": 6.934092850250001, + "learning_rate": 1.8652917993812267e-05, + "loss": 1.1186, + "step": 1770 + }, + { + "epoch": 0.9614549402823018, + "grad_norm": 6.129320663785202, + "learning_rate": 1.865115457794977e-05, + "loss": 0.9076, + "step": 1771 + }, + { + "epoch": 0.9619978284473398, + "grad_norm": 7.008261896063745, + "learning_rate": 1.8649390092098693e-05, + "loss": 1.7152, + "step": 1772 + }, + { + "epoch": 0.9625407166123778, + "grad_norm": 6.36833090372477, + "learning_rate": 1.8647624536477255e-05, + "loss": 0.8087, + "step": 1773 + }, + { + "epoch": 0.9630836047774158, + "grad_norm": 4.7595903392838865, + "learning_rate": 1.8645857911303838e-05, + "loss": 0.6374, + "step": 1774 + }, + { + "epoch": 0.9636264929424538, + "grad_norm": 6.410325139984648, + "learning_rate": 1.8644090216796934e-05, + "loss": 1.3611, + "step": 1775 + }, + { + "epoch": 0.9641693811074918, + "grad_norm": 5.900212705071564, + "learning_rate": 1.8642321453175177e-05, + "loss": 0.891, + "step": 1776 + }, + { + "epoch": 0.9647122692725298, + "grad_norm": 6.90562183471938, + "learning_rate": 1.8640551620657326e-05, + "loss": 0.7899, + "step": 1777 + }, + { + "epoch": 0.9652551574375678, + "grad_norm": 6.476066490933977, + "learning_rate": 1.8638780719462278e-05, + "loss": 0.9614, + "step": 1778 + }, + { + "epoch": 0.9657980456026058, + "grad_norm": 6.659586852509603, + "learning_rate": 1.8637008749809065e-05, + "loss": 1.2419, + "step": 1779 + }, + { + "epoch": 0.9663409337676439, + "grad_norm": 6.587754652016677, + "learning_rate": 1.8635235711916847e-05, + "loss": 1.3664, + "step": 1780 + }, + { + "epoch": 0.9668838219326819, + "grad_norm": 6.558578755813245, + "learning_rate": 1.863346160600491e-05, + "loss": 1.1249, + "step": 1781 + }, + { + "epoch": 0.9674267100977199, + "grad_norm": 4.791301716076818, + "learning_rate": 1.8631686432292685e-05, + "loss": 0.6836, + "step": 1782 + }, + { + "epoch": 0.9679695982627579, + "grad_norm": 5.120778292537024, + "learning_rate": 1.862991019099972e-05, + "loss": 0.592, + "step": 1783 + }, + { + "epoch": 0.9685124864277959, + "grad_norm": 8.250755784410368, + "learning_rate": 1.8628132882345713e-05, + "loss": 1.2579, + "step": 1784 + }, + { + "epoch": 0.9690553745928339, + "grad_norm": 5.811263926579046, + "learning_rate": 1.862635450655048e-05, + "loss": 1.2935, + "step": 1785 + }, + { + "epoch": 0.9695982627578719, + "grad_norm": 6.042040633945999, + "learning_rate": 1.862457506383397e-05, + "loss": 0.8784, + "step": 1786 + }, + { + "epoch": 0.9701411509229099, + "grad_norm": 7.115011060082062, + "learning_rate": 1.8622794554416272e-05, + "loss": 1.0717, + "step": 1787 + }, + { + "epoch": 0.9706840390879479, + "grad_norm": 7.520775839533288, + "learning_rate": 1.8621012978517604e-05, + "loss": 1.3468, + "step": 1788 + }, + { + "epoch": 0.9712269272529859, + "grad_norm": 6.949921961354188, + "learning_rate": 1.8619230336358306e-05, + "loss": 1.5609, + "step": 1789 + }, + { + "epoch": 0.9717698154180239, + "grad_norm": 6.396218455762339, + "learning_rate": 1.8617446628158866e-05, + "loss": 1.1807, + "step": 1790 + }, + { + "epoch": 0.9723127035830619, + "grad_norm": 6.314794279037124, + "learning_rate": 1.861566185413989e-05, + "loss": 0.9255, + "step": 1791 + }, + { + "epoch": 0.9728555917480999, + "grad_norm": 5.393752442926871, + "learning_rate": 1.8613876014522128e-05, + "loss": 0.7926, + "step": 1792 + }, + { + "epoch": 0.9733984799131379, + "grad_norm": 5.908313693688933, + "learning_rate": 1.8612089109526453e-05, + "loss": 0.7984, + "step": 1793 + }, + { + "epoch": 0.9739413680781759, + "grad_norm": 4.6952509401002125, + "learning_rate": 1.8610301139373867e-05, + "loss": 1.0344, + "step": 1794 + }, + { + "epoch": 0.9744842562432139, + "grad_norm": 6.81835335105175, + "learning_rate": 1.8608512104285517e-05, + "loss": 1.1532, + "step": 1795 + }, + { + "epoch": 0.9750271444082519, + "grad_norm": 4.159411114912345, + "learning_rate": 1.860672200448267e-05, + "loss": 1.0665, + "step": 1796 + }, + { + "epoch": 0.9755700325732899, + "grad_norm": 4.911404446825004, + "learning_rate": 1.8604930840186726e-05, + "loss": 1.1784, + "step": 1797 + }, + { + "epoch": 0.9761129207383279, + "grad_norm": 5.399453855489055, + "learning_rate": 1.860313861161922e-05, + "loss": 0.8215, + "step": 1798 + }, + { + "epoch": 0.9766558089033659, + "grad_norm": 4.797137239148531, + "learning_rate": 1.860134531900182e-05, + "loss": 0.6405, + "step": 1799 + }, + { + "epoch": 0.9771986970684039, + "grad_norm": 3.937076316066272, + "learning_rate": 1.859955096255633e-05, + "loss": 0.5912, + "step": 1800 + }, + { + "epoch": 0.9777415852334419, + "grad_norm": 5.569110777302129, + "learning_rate": 1.859775554250466e-05, + "loss": 0.8089, + "step": 1801 + }, + { + "epoch": 0.9782844733984799, + "grad_norm": 6.364258901894282, + "learning_rate": 1.859595905906889e-05, + "loss": 1.5181, + "step": 1802 + }, + { + "epoch": 0.9788273615635179, + "grad_norm": 4.995224609073908, + "learning_rate": 1.85941615124712e-05, + "loss": 0.8573, + "step": 1803 + }, + { + "epoch": 0.9793702497285559, + "grad_norm": 5.041992740138782, + "learning_rate": 1.8592362902933918e-05, + "loss": 0.8982, + "step": 1804 + }, + { + "epoch": 0.9799131378935939, + "grad_norm": 5.6054574490953275, + "learning_rate": 1.8590563230679496e-05, + "loss": 0.9285, + "step": 1805 + }, + { + "epoch": 0.9804560260586319, + "grad_norm": 5.269503319178264, + "learning_rate": 1.8588762495930526e-05, + "loss": 1.0963, + "step": 1806 + }, + { + "epoch": 0.9809989142236699, + "grad_norm": 6.184787731593613, + "learning_rate": 1.8586960698909718e-05, + "loss": 1.4005, + "step": 1807 + }, + { + "epoch": 0.9815418023887079, + "grad_norm": 8.50824591479003, + "learning_rate": 1.858515783983993e-05, + "loss": 1.9607, + "step": 1808 + }, + { + "epoch": 0.9820846905537459, + "grad_norm": 5.399464732589782, + "learning_rate": 1.8583353918944134e-05, + "loss": 0.951, + "step": 1809 + }, + { + "epoch": 0.9826275787187839, + "grad_norm": 6.577634699737779, + "learning_rate": 1.8581548936445447e-05, + "loss": 1.5139, + "step": 1810 + }, + { + "epoch": 0.9831704668838219, + "grad_norm": 10.183489163585788, + "learning_rate": 1.8579742892567107e-05, + "loss": 2.0539, + "step": 1811 + }, + { + "epoch": 0.9837133550488599, + "grad_norm": 6.578605968943458, + "learning_rate": 1.8577935787532494e-05, + "loss": 1.1454, + "step": 1812 + }, + { + "epoch": 0.9842562432138979, + "grad_norm": 4.923285404215955, + "learning_rate": 1.8576127621565113e-05, + "loss": 0.8965, + "step": 1813 + }, + { + "epoch": 0.9847991313789359, + "grad_norm": 5.912445932400008, + "learning_rate": 1.85743183948886e-05, + "loss": 1.1311, + "step": 1814 + }, + { + "epoch": 0.9853420195439739, + "grad_norm": 4.5183231221646105, + "learning_rate": 1.8572508107726725e-05, + "loss": 0.7511, + "step": 1815 + }, + { + "epoch": 0.9858849077090119, + "grad_norm": 5.2218543332097935, + "learning_rate": 1.8570696760303378e-05, + "loss": 1.0063, + "step": 1816 + }, + { + "epoch": 0.9864277958740499, + "grad_norm": 7.215748803707923, + "learning_rate": 1.85688843528426e-05, + "loss": 1.3674, + "step": 1817 + }, + { + "epoch": 0.9869706840390879, + "grad_norm": 6.362354074709609, + "learning_rate": 1.8567070885568547e-05, + "loss": 1.1319, + "step": 1818 + }, + { + "epoch": 0.987513572204126, + "grad_norm": 4.680755706450828, + "learning_rate": 1.8565256358705513e-05, + "loss": 0.6613, + "step": 1819 + }, + { + "epoch": 0.988056460369164, + "grad_norm": 4.80241945107321, + "learning_rate": 1.8563440772477922e-05, + "loss": 0.8619, + "step": 1820 + }, + { + "epoch": 0.988599348534202, + "grad_norm": 5.858709563059296, + "learning_rate": 1.856162412711033e-05, + "loss": 0.9432, + "step": 1821 + }, + { + "epoch": 0.98914223669924, + "grad_norm": 6.478119299588854, + "learning_rate": 1.855980642282742e-05, + "loss": 1.1866, + "step": 1822 + }, + { + "epoch": 0.989685124864278, + "grad_norm": 7.081924554034699, + "learning_rate": 1.8557987659854006e-05, + "loss": 1.3657, + "step": 1823 + }, + { + "epoch": 0.990228013029316, + "grad_norm": 5.404410220733361, + "learning_rate": 1.855616783841504e-05, + "loss": 1.1051, + "step": 1824 + }, + { + "epoch": 0.990770901194354, + "grad_norm": 6.742207962837618, + "learning_rate": 1.8554346958735602e-05, + "loss": 1.4667, + "step": 1825 + }, + { + "epoch": 0.991313789359392, + "grad_norm": 6.164178667211635, + "learning_rate": 1.8552525021040895e-05, + "loss": 0.7861, + "step": 1826 + }, + { + "epoch": 0.99185667752443, + "grad_norm": 6.078017237490637, + "learning_rate": 1.8550702025556265e-05, + "loss": 1.3294, + "step": 1827 + }, + { + "epoch": 0.992399565689468, + "grad_norm": 4.7372949568909455, + "learning_rate": 1.8548877972507182e-05, + "loss": 0.9779, + "step": 1828 + }, + { + "epoch": 0.992942453854506, + "grad_norm": 5.25971277777658, + "learning_rate": 1.8547052862119247e-05, + "loss": 1.049, + "step": 1829 + }, + { + "epoch": 0.993485342019544, + "grad_norm": 4.950709970125391, + "learning_rate": 1.854522669461819e-05, + "loss": 0.6665, + "step": 1830 + }, + { + "epoch": 0.994028230184582, + "grad_norm": 5.856428194694733, + "learning_rate": 1.8543399470229876e-05, + "loss": 1.0954, + "step": 1831 + }, + { + "epoch": 0.99457111834962, + "grad_norm": 4.895286613591861, + "learning_rate": 1.85415711891803e-05, + "loss": 0.7069, + "step": 1832 + }, + { + "epoch": 0.995114006514658, + "grad_norm": 5.249228502887773, + "learning_rate": 1.8539741851695586e-05, + "loss": 1.0655, + "step": 1833 + }, + { + "epoch": 0.995656894679696, + "grad_norm": 4.855694799243912, + "learning_rate": 1.8537911458001988e-05, + "loss": 1.1872, + "step": 1834 + }, + { + "epoch": 0.996199782844734, + "grad_norm": 5.180821237913711, + "learning_rate": 1.8536080008325896e-05, + "loss": 0.8281, + "step": 1835 + }, + { + "epoch": 0.996742671009772, + "grad_norm": 5.073254295771654, + "learning_rate": 1.8534247502893823e-05, + "loss": 0.7002, + "step": 1836 + }, + { + "epoch": 0.99728555917481, + "grad_norm": 5.4022604097623255, + "learning_rate": 1.8532413941932416e-05, + "loss": 1.1793, + "step": 1837 + }, + { + "epoch": 0.997828447339848, + "grad_norm": 4.981761324197401, + "learning_rate": 1.8530579325668455e-05, + "loss": 0.8331, + "step": 1838 + }, + { + "epoch": 0.998371335504886, + "grad_norm": 7.325363733761658, + "learning_rate": 1.852874365432885e-05, + "loss": 1.3071, + "step": 1839 + }, + { + "epoch": 0.998914223669924, + "grad_norm": 7.5668251202705505, + "learning_rate": 1.852690692814063e-05, + "loss": 1.4955, + "step": 1840 + }, + { + "epoch": 0.999457111834962, + "grad_norm": 6.706936559053067, + "learning_rate": 1.8525069147330978e-05, + "loss": 1.0286, + "step": 1841 + }, + { + "epoch": 1.0, + "grad_norm": 6.338369283394405, + "learning_rate": 1.8523230312127183e-05, + "loss": 1.0509, + "step": 1842 + }, + { + "epoch": 1.000542888165038, + "grad_norm": 6.119502670416093, + "learning_rate": 1.8521390422756683e-05, + "loss": 1.0805, + "step": 1843 + }, + { + "epoch": 1.001085776330076, + "grad_norm": 5.074420967885639, + "learning_rate": 1.851954947944703e-05, + "loss": 1.0931, + "step": 1844 + }, + { + "epoch": 1.001628664495114, + "grad_norm": 8.194067497007946, + "learning_rate": 1.851770748242592e-05, + "loss": 2.0732, + "step": 1845 + }, + { + "epoch": 1.002171552660152, + "grad_norm": 4.389522929363591, + "learning_rate": 1.8515864431921177e-05, + "loss": 0.8684, + "step": 1846 + }, + { + "epoch": 1.00271444082519, + "grad_norm": 5.720373954211297, + "learning_rate": 1.8514020328160748e-05, + "loss": 1.0223, + "step": 1847 + }, + { + "epoch": 1.003257328990228, + "grad_norm": 5.163678744409467, + "learning_rate": 1.8512175171372713e-05, + "loss": 0.9611, + "step": 1848 + }, + { + "epoch": 1.003800217155266, + "grad_norm": 5.786056891749039, + "learning_rate": 1.8510328961785286e-05, + "loss": 1.3473, + "step": 1849 + }, + { + "epoch": 1.004343105320304, + "grad_norm": 5.9194250471944985, + "learning_rate": 1.850848169962681e-05, + "loss": 0.8218, + "step": 1850 + }, + { + "epoch": 1.004885993485342, + "grad_norm": 5.890241796399565, + "learning_rate": 1.850663338512576e-05, + "loss": 0.842, + "step": 1851 + }, + { + "epoch": 1.00542888165038, + "grad_norm": 6.939341519595904, + "learning_rate": 1.8504784018510732e-05, + "loss": 0.9263, + "step": 1852 + }, + { + "epoch": 1.005971769815418, + "grad_norm": 4.592014594714093, + "learning_rate": 1.850293360001046e-05, + "loss": 1.015, + "step": 1853 + }, + { + "epoch": 1.006514657980456, + "grad_norm": 4.575145990552055, + "learning_rate": 1.8501082129853816e-05, + "loss": 1.0505, + "step": 1854 + }, + { + "epoch": 1.007057546145494, + "grad_norm": 5.999218616951103, + "learning_rate": 1.849922960826978e-05, + "loss": 1.0945, + "step": 1855 + }, + { + "epoch": 1.007600434310532, + "grad_norm": 5.459227552994586, + "learning_rate": 1.8497376035487483e-05, + "loss": 1.0167, + "step": 1856 + }, + { + "epoch": 1.00814332247557, + "grad_norm": 4.785837605715454, + "learning_rate": 1.8495521411736173e-05, + "loss": 1.0559, + "step": 1857 + }, + { + "epoch": 1.008686210640608, + "grad_norm": 5.813312187108195, + "learning_rate": 1.8493665737245236e-05, + "loss": 1.3477, + "step": 1858 + }, + { + "epoch": 1.009229098805646, + "grad_norm": 6.512945450475729, + "learning_rate": 1.8491809012244182e-05, + "loss": 1.4416, + "step": 1859 + }, + { + "epoch": 1.009771986970684, + "grad_norm": 5.599962385405486, + "learning_rate": 1.8489951236962658e-05, + "loss": 0.8961, + "step": 1860 + }, + { + "epoch": 1.010314875135722, + "grad_norm": 4.863989016796961, + "learning_rate": 1.848809241163043e-05, + "loss": 0.639, + "step": 1861 + }, + { + "epoch": 1.01085776330076, + "grad_norm": 5.937471136238718, + "learning_rate": 1.848623253647741e-05, + "loss": 1.0548, + "step": 1862 + }, + { + "epoch": 1.011400651465798, + "grad_norm": 4.75172125273428, + "learning_rate": 1.8484371611733625e-05, + "loss": 0.6493, + "step": 1863 + }, + { + "epoch": 1.011943539630836, + "grad_norm": 5.867263430143688, + "learning_rate": 1.848250963762923e-05, + "loss": 0.8569, + "step": 1864 + }, + { + "epoch": 1.012486427795874, + "grad_norm": 5.932410925593062, + "learning_rate": 1.848064661439453e-05, + "loss": 0.8351, + "step": 1865 + }, + { + "epoch": 1.013029315960912, + "grad_norm": 5.8320858990674, + "learning_rate": 1.847878254225994e-05, + "loss": 0.8916, + "step": 1866 + }, + { + "epoch": 1.01357220412595, + "grad_norm": 5.836721791741835, + "learning_rate": 1.847691742145601e-05, + "loss": 1.1742, + "step": 1867 + }, + { + "epoch": 1.014115092290988, + "grad_norm": 6.806668722890977, + "learning_rate": 1.8475051252213423e-05, + "loss": 1.0714, + "step": 1868 + }, + { + "epoch": 1.014657980456026, + "grad_norm": 7.026347641082525, + "learning_rate": 1.8473184034762992e-05, + "loss": 1.4101, + "step": 1869 + }, + { + "epoch": 1.015200868621064, + "grad_norm": 5.9532414283206005, + "learning_rate": 1.8471315769335657e-05, + "loss": 0.7855, + "step": 1870 + }, + { + "epoch": 1.015743756786102, + "grad_norm": 5.7578706206451855, + "learning_rate": 1.846944645616248e-05, + "loss": 0.7644, + "step": 1871 + }, + { + "epoch": 1.01628664495114, + "grad_norm": 6.340541706204924, + "learning_rate": 1.846757609547467e-05, + "loss": 1.1607, + "step": 1872 + }, + { + "epoch": 1.016829533116178, + "grad_norm": 4.942991579941027, + "learning_rate": 1.8465704687503558e-05, + "loss": 0.7578, + "step": 1873 + }, + { + "epoch": 1.017372421281216, + "grad_norm": 6.42121491667965, + "learning_rate": 1.846383223248059e-05, + "loss": 0.9924, + "step": 1874 + }, + { + "epoch": 1.017915309446254, + "grad_norm": 6.252498990648505, + "learning_rate": 1.8461958730637368e-05, + "loss": 0.811, + "step": 1875 + }, + { + "epoch": 1.018458197611292, + "grad_norm": 5.941336187529878, + "learning_rate": 1.84600841822056e-05, + "loss": 0.772, + "step": 1876 + }, + { + "epoch": 1.01900108577633, + "grad_norm": 5.323895762793836, + "learning_rate": 1.845820858741714e-05, + "loss": 0.8812, + "step": 1877 + }, + { + "epoch": 1.019543973941368, + "grad_norm": 7.2574721612646105, + "learning_rate": 1.845633194650396e-05, + "loss": 1.2759, + "step": 1878 + }, + { + "epoch": 1.020086862106406, + "grad_norm": 4.307176114235153, + "learning_rate": 1.8454454259698165e-05, + "loss": 0.569, + "step": 1879 + }, + { + "epoch": 1.020629750271444, + "grad_norm": 6.504543473682945, + "learning_rate": 1.8452575527231997e-05, + "loss": 0.9355, + "step": 1880 + }, + { + "epoch": 1.021172638436482, + "grad_norm": 5.822065969327671, + "learning_rate": 1.8450695749337816e-05, + "loss": 1.0287, + "step": 1881 + }, + { + "epoch": 1.02171552660152, + "grad_norm": 8.90668319984316, + "learning_rate": 1.8448814926248112e-05, + "loss": 1.5348, + "step": 1882 + }, + { + "epoch": 1.022258414766558, + "grad_norm": 5.233217719773852, + "learning_rate": 1.844693305819552e-05, + "loss": 0.564, + "step": 1883 + }, + { + "epoch": 1.022801302931596, + "grad_norm": 5.806961114054563, + "learning_rate": 1.844505014541278e-05, + "loss": 0.9349, + "step": 1884 + }, + { + "epoch": 1.023344191096634, + "grad_norm": 6.239514346807307, + "learning_rate": 1.8443166188132777e-05, + "loss": 0.9677, + "step": 1885 + }, + { + "epoch": 1.023887079261672, + "grad_norm": 5.596402826374599, + "learning_rate": 1.8441281186588528e-05, + "loss": 0.695, + "step": 1886 + }, + { + "epoch": 1.02442996742671, + "grad_norm": 6.531429104200087, + "learning_rate": 1.8439395141013165e-05, + "loss": 1.1543, + "step": 1887 + }, + { + "epoch": 1.024972855591748, + "grad_norm": 5.129461327017003, + "learning_rate": 1.843750805163996e-05, + "loss": 1.2624, + "step": 1888 + }, + { + "epoch": 1.0255157437567861, + "grad_norm": 6.862993011679228, + "learning_rate": 1.8435619918702318e-05, + "loss": 1.2673, + "step": 1889 + }, + { + "epoch": 1.0260586319218241, + "grad_norm": 5.134671966877655, + "learning_rate": 1.8433730742433755e-05, + "loss": 0.6032, + "step": 1890 + }, + { + "epoch": 1.0266015200868621, + "grad_norm": 6.9199078267443195, + "learning_rate": 1.8431840523067932e-05, + "loss": 1.1255, + "step": 1891 + }, + { + "epoch": 1.0271444082519001, + "grad_norm": 5.808796921918892, + "learning_rate": 1.8429949260838635e-05, + "loss": 0.8488, + "step": 1892 + }, + { + "epoch": 1.0276872964169381, + "grad_norm": 7.102540931353532, + "learning_rate": 1.842805695597978e-05, + "loss": 1.0634, + "step": 1893 + }, + { + "epoch": 1.0282301845819761, + "grad_norm": 9.932212718124791, + "learning_rate": 1.8426163608725403e-05, + "loss": 1.3909, + "step": 1894 + }, + { + "epoch": 1.0287730727470141, + "grad_norm": 6.999752698080143, + "learning_rate": 1.8424269219309686e-05, + "loss": 1.1106, + "step": 1895 + }, + { + "epoch": 1.0293159609120521, + "grad_norm": 6.569873028691542, + "learning_rate": 1.842237378796693e-05, + "loss": 0.9455, + "step": 1896 + }, + { + "epoch": 1.0298588490770901, + "grad_norm": 6.418746090533204, + "learning_rate": 1.8420477314931554e-05, + "loss": 0.934, + "step": 1897 + }, + { + "epoch": 1.0304017372421281, + "grad_norm": 7.7235490740904815, + "learning_rate": 1.8418579800438125e-05, + "loss": 1.1582, + "step": 1898 + }, + { + "epoch": 1.0309446254071661, + "grad_norm": 5.4350436368749895, + "learning_rate": 1.841668124472133e-05, + "loss": 0.7726, + "step": 1899 + }, + { + "epoch": 1.0314875135722041, + "grad_norm": 5.34390520812136, + "learning_rate": 1.8414781648015983e-05, + "loss": 0.7497, + "step": 1900 + }, + { + "epoch": 1.0320304017372421, + "grad_norm": 6.563688266956639, + "learning_rate": 1.841288101055703e-05, + "loss": 1.3352, + "step": 1901 + }, + { + "epoch": 1.0325732899022801, + "grad_norm": 7.710283890330315, + "learning_rate": 1.841097933257955e-05, + "loss": 0.8804, + "step": 1902 + }, + { + "epoch": 1.0331161780673181, + "grad_norm": 5.31500431944041, + "learning_rate": 1.840907661431874e-05, + "loss": 0.8667, + "step": 1903 + }, + { + "epoch": 1.0336590662323561, + "grad_norm": 7.476982099026401, + "learning_rate": 1.840717285600993e-05, + "loss": 0.8317, + "step": 1904 + }, + { + "epoch": 1.0342019543973942, + "grad_norm": 5.1563433597221415, + "learning_rate": 1.840526805788858e-05, + "loss": 1.0108, + "step": 1905 + }, + { + "epoch": 1.0347448425624322, + "grad_norm": 6.411832164370409, + "learning_rate": 1.8403362220190284e-05, + "loss": 0.7778, + "step": 1906 + }, + { + "epoch": 1.0352877307274702, + "grad_norm": 6.688312152010462, + "learning_rate": 1.8401455343150757e-05, + "loss": 1.2438, + "step": 1907 + }, + { + "epoch": 1.0358306188925082, + "grad_norm": 6.412659783246144, + "learning_rate": 1.839954742700584e-05, + "loss": 0.8191, + "step": 1908 + }, + { + "epoch": 1.0363735070575462, + "grad_norm": 6.399618830323376, + "learning_rate": 1.839763847199151e-05, + "loss": 1.4485, + "step": 1909 + }, + { + "epoch": 1.0369163952225842, + "grad_norm": 8.041718562623554, + "learning_rate": 1.8395728478343873e-05, + "loss": 1.0043, + "step": 1910 + }, + { + "epoch": 1.0374592833876222, + "grad_norm": 6.4118854777282905, + "learning_rate": 1.8393817446299152e-05, + "loss": 0.9183, + "step": 1911 + }, + { + "epoch": 1.0380021715526602, + "grad_norm": 6.400834366436549, + "learning_rate": 1.8391905376093717e-05, + "loss": 1.1517, + "step": 1912 + }, + { + "epoch": 1.0385450597176982, + "grad_norm": 6.160364431172616, + "learning_rate": 1.8389992267964046e-05, + "loss": 0.5949, + "step": 1913 + }, + { + "epoch": 1.0390879478827362, + "grad_norm": 7.298564934931577, + "learning_rate": 1.8388078122146763e-05, + "loss": 1.0234, + "step": 1914 + }, + { + "epoch": 1.0396308360477742, + "grad_norm": 6.132796013684569, + "learning_rate": 1.838616293887861e-05, + "loss": 0.9417, + "step": 1915 + }, + { + "epoch": 1.0401737242128122, + "grad_norm": 8.083369310536504, + "learning_rate": 1.8384246718396458e-05, + "loss": 1.2617, + "step": 1916 + }, + { + "epoch": 1.0407166123778502, + "grad_norm": 5.492413210001038, + "learning_rate": 1.8382329460937306e-05, + "loss": 0.7906, + "step": 1917 + }, + { + "epoch": 1.0412595005428882, + "grad_norm": 5.936786319330656, + "learning_rate": 1.838041116673829e-05, + "loss": 0.8254, + "step": 1918 + }, + { + "epoch": 1.0418023887079262, + "grad_norm": 6.637068024744901, + "learning_rate": 1.8378491836036666e-05, + "loss": 1.1779, + "step": 1919 + }, + { + "epoch": 1.0423452768729642, + "grad_norm": 5.481354066687985, + "learning_rate": 1.8376571469069814e-05, + "loss": 0.9586, + "step": 1920 + }, + { + "epoch": 1.0428881650380022, + "grad_norm": 7.669719057080363, + "learning_rate": 1.8374650066075257e-05, + "loss": 1.2058, + "step": 1921 + }, + { + "epoch": 1.0434310532030402, + "grad_norm": 7.387160499244698, + "learning_rate": 1.8372727627290627e-05, + "loss": 1.3861, + "step": 1922 + }, + { + "epoch": 1.0439739413680782, + "grad_norm": 4.975886150156898, + "learning_rate": 1.8370804152953704e-05, + "loss": 0.5849, + "step": 1923 + }, + { + "epoch": 1.0445168295331162, + "grad_norm": 5.413290989228805, + "learning_rate": 1.8368879643302383e-05, + "loss": 0.8425, + "step": 1924 + }, + { + "epoch": 1.0450597176981542, + "grad_norm": 6.564098001301886, + "learning_rate": 1.836695409857469e-05, + "loss": 0.9658, + "step": 1925 + }, + { + "epoch": 1.0456026058631922, + "grad_norm": 8.050721283259312, + "learning_rate": 1.8365027519008774e-05, + "loss": 1.684, + "step": 1926 + }, + { + "epoch": 1.0461454940282302, + "grad_norm": 5.409117554986433, + "learning_rate": 1.836309990484293e-05, + "loss": 0.8247, + "step": 1927 + }, + { + "epoch": 1.0466883821932682, + "grad_norm": 5.108392155413988, + "learning_rate": 1.8361171256315555e-05, + "loss": 0.8244, + "step": 1928 + }, + { + "epoch": 1.0472312703583062, + "grad_norm": 6.806129539780182, + "learning_rate": 1.8359241573665194e-05, + "loss": 1.0678, + "step": 1929 + }, + { + "epoch": 1.0477741585233442, + "grad_norm": 4.43019983774416, + "learning_rate": 1.8357310857130514e-05, + "loss": 0.6213, + "step": 1930 + }, + { + "epoch": 1.0483170466883822, + "grad_norm": 6.194875830854703, + "learning_rate": 1.835537910695031e-05, + "loss": 0.8774, + "step": 1931 + }, + { + "epoch": 1.0488599348534202, + "grad_norm": 4.540451693265499, + "learning_rate": 1.8353446323363496e-05, + "loss": 0.8508, + "step": 1932 + }, + { + "epoch": 1.0494028230184582, + "grad_norm": 5.9491142341914225, + "learning_rate": 1.8351512506609133e-05, + "loss": 1.0139, + "step": 1933 + }, + { + "epoch": 1.0499457111834962, + "grad_norm": 6.436571178648067, + "learning_rate": 1.834957765692639e-05, + "loss": 0.9272, + "step": 1934 + }, + { + "epoch": 1.0504885993485342, + "grad_norm": 5.6538429161280535, + "learning_rate": 1.8347641774554573e-05, + "loss": 0.7834, + "step": 1935 + }, + { + "epoch": 1.0510314875135722, + "grad_norm": 6.481163094620501, + "learning_rate": 1.8345704859733123e-05, + "loss": 1.0916, + "step": 1936 + }, + { + "epoch": 1.0515743756786102, + "grad_norm": 6.165149195118496, + "learning_rate": 1.8343766912701588e-05, + "loss": 1.1054, + "step": 1937 + }, + { + "epoch": 1.0521172638436482, + "grad_norm": 7.778083794032891, + "learning_rate": 1.834182793369967e-05, + "loss": 1.297, + "step": 1938 + }, + { + "epoch": 1.0526601520086862, + "grad_norm": 8.871857523729659, + "learning_rate": 1.8339887922967176e-05, + "loss": 1.5749, + "step": 1939 + }, + { + "epoch": 1.0532030401737242, + "grad_norm": 6.563413115451867, + "learning_rate": 1.8337946880744047e-05, + "loss": 1.0271, + "step": 1940 + }, + { + "epoch": 1.0537459283387622, + "grad_norm": 6.5304770661428275, + "learning_rate": 1.833600480727036e-05, + "loss": 0.9236, + "step": 1941 + }, + { + "epoch": 1.0542888165038002, + "grad_norm": 6.802479270907137, + "learning_rate": 1.8334061702786317e-05, + "loss": 0.9076, + "step": 1942 + }, + { + "epoch": 1.0548317046688382, + "grad_norm": 5.92560525391674, + "learning_rate": 1.833211756753224e-05, + "loss": 0.9601, + "step": 1943 + }, + { + "epoch": 1.0553745928338762, + "grad_norm": 7.436926375817581, + "learning_rate": 1.8330172401748584e-05, + "loss": 1.1758, + "step": 1944 + }, + { + "epoch": 1.0559174809989142, + "grad_norm": 6.369742715679364, + "learning_rate": 1.8328226205675927e-05, + "loss": 0.8424, + "step": 1945 + }, + { + "epoch": 1.0564603691639523, + "grad_norm": 7.60632781901004, + "learning_rate": 1.8326278979554976e-05, + "loss": 1.7815, + "step": 1946 + }, + { + "epoch": 1.0570032573289903, + "grad_norm": 8.499242837541805, + "learning_rate": 1.8324330723626578e-05, + "loss": 1.1341, + "step": 1947 + }, + { + "epoch": 1.0575461454940283, + "grad_norm": 6.469082998862602, + "learning_rate": 1.8322381438131686e-05, + "loss": 0.7798, + "step": 1948 + }, + { + "epoch": 1.0580890336590663, + "grad_norm": 5.923622010018089, + "learning_rate": 1.8320431123311388e-05, + "loss": 0.7971, + "step": 1949 + }, + { + "epoch": 1.0586319218241043, + "grad_norm": 6.617488371294151, + "learning_rate": 1.8318479779406914e-05, + "loss": 1.1551, + "step": 1950 + }, + { + "epoch": 1.0591748099891423, + "grad_norm": 6.813827391520987, + "learning_rate": 1.8316527406659604e-05, + "loss": 1.0234, + "step": 1951 + }, + { + "epoch": 1.0597176981541803, + "grad_norm": 7.580867549521637, + "learning_rate": 1.831457400531093e-05, + "loss": 0.7896, + "step": 1952 + }, + { + "epoch": 1.0602605863192183, + "grad_norm": 5.975929968595557, + "learning_rate": 1.8312619575602486e-05, + "loss": 1.1248, + "step": 1953 + }, + { + "epoch": 1.0608034744842563, + "grad_norm": 6.2878616306091955, + "learning_rate": 1.831066411777601e-05, + "loss": 1.0833, + "step": 1954 + }, + { + "epoch": 1.0613463626492943, + "grad_norm": 6.284963948133933, + "learning_rate": 1.8308707632073345e-05, + "loss": 1.0773, + "step": 1955 + }, + { + "epoch": 1.0618892508143323, + "grad_norm": 4.009278799804313, + "learning_rate": 1.830675011873648e-05, + "loss": 0.4782, + "step": 1956 + }, + { + "epoch": 1.0624321389793703, + "grad_norm": 6.648653088879531, + "learning_rate": 1.8304791578007524e-05, + "loss": 0.9761, + "step": 1957 + }, + { + "epoch": 1.0629750271444083, + "grad_norm": 6.867797713608168, + "learning_rate": 1.830283201012871e-05, + "loss": 0.9315, + "step": 1958 + }, + { + "epoch": 1.0635179153094463, + "grad_norm": 5.989268423819992, + "learning_rate": 1.83008714153424e-05, + "loss": 0.8312, + "step": 1959 + }, + { + "epoch": 1.0640608034744843, + "grad_norm": 6.207518262073172, + "learning_rate": 1.8298909793891083e-05, + "loss": 1.1003, + "step": 1960 + }, + { + "epoch": 1.0646036916395223, + "grad_norm": 5.5435710943195815, + "learning_rate": 1.8296947146017373e-05, + "loss": 0.8012, + "step": 1961 + }, + { + "epoch": 1.0651465798045603, + "grad_norm": 8.000995164961113, + "learning_rate": 1.829498347196402e-05, + "loss": 0.9864, + "step": 1962 + }, + { + "epoch": 1.0656894679695983, + "grad_norm": 6.483090321755992, + "learning_rate": 1.829301877197389e-05, + "loss": 1.2823, + "step": 1963 + }, + { + "epoch": 1.0662323561346363, + "grad_norm": 7.147142303433649, + "learning_rate": 1.8291053046289985e-05, + "loss": 0.9185, + "step": 1964 + }, + { + "epoch": 1.0667752442996743, + "grad_norm": 6.000816189695904, + "learning_rate": 1.828908629515542e-05, + "loss": 1.0284, + "step": 1965 + }, + { + "epoch": 1.0673181324647123, + "grad_norm": 6.188793151434744, + "learning_rate": 1.8287118518813453e-05, + "loss": 0.7041, + "step": 1966 + }, + { + "epoch": 1.0678610206297503, + "grad_norm": 7.353722352349769, + "learning_rate": 1.828514971750746e-05, + "loss": 1.0707, + "step": 1967 + }, + { + "epoch": 1.0684039087947883, + "grad_norm": 7.553378116857428, + "learning_rate": 1.8283179891480944e-05, + "loss": 1.1426, + "step": 1968 + }, + { + "epoch": 1.0689467969598263, + "grad_norm": 7.134594743515001, + "learning_rate": 1.828120904097754e-05, + "loss": 0.9583, + "step": 1969 + }, + { + "epoch": 1.0694896851248643, + "grad_norm": 5.680617024729498, + "learning_rate": 1.8279237166241004e-05, + "loss": 0.9632, + "step": 1970 + }, + { + "epoch": 1.0700325732899023, + "grad_norm": 5.932910663063727, + "learning_rate": 1.8277264267515218e-05, + "loss": 0.9984, + "step": 1971 + }, + { + "epoch": 1.0705754614549403, + "grad_norm": 7.833159555663921, + "learning_rate": 1.8275290345044198e-05, + "loss": 0.8038, + "step": 1972 + }, + { + "epoch": 1.0711183496199783, + "grad_norm": 6.429988603781813, + "learning_rate": 1.8273315399072076e-05, + "loss": 0.9125, + "step": 1973 + }, + { + "epoch": 1.0716612377850163, + "grad_norm": 7.955595430340166, + "learning_rate": 1.827133942984312e-05, + "loss": 0.8697, + "step": 1974 + }, + { + "epoch": 1.0722041259500543, + "grad_norm": 5.983514211221058, + "learning_rate": 1.826936243760172e-05, + "loss": 0.8155, + "step": 1975 + }, + { + "epoch": 1.0727470141150923, + "grad_norm": 6.19913176090901, + "learning_rate": 1.8267384422592398e-05, + "loss": 1.3207, + "step": 1976 + }, + { + "epoch": 1.0732899022801303, + "grad_norm": 6.467400855860047, + "learning_rate": 1.8265405385059792e-05, + "loss": 1.1279, + "step": 1977 + }, + { + "epoch": 1.0738327904451683, + "grad_norm": 6.515150741776156, + "learning_rate": 1.8263425325248675e-05, + "loss": 1.372, + "step": 1978 + }, + { + "epoch": 1.0743756786102063, + "grad_norm": 6.181497786496646, + "learning_rate": 1.8261444243403945e-05, + "loss": 0.8571, + "step": 1979 + }, + { + "epoch": 1.0749185667752443, + "grad_norm": 7.637816753034266, + "learning_rate": 1.8259462139770624e-05, + "loss": 1.4861, + "step": 1980 + }, + { + "epoch": 1.0754614549402823, + "grad_norm": 7.0094523855293485, + "learning_rate": 1.825747901459386e-05, + "loss": 1.0113, + "step": 1981 + }, + { + "epoch": 1.0760043431053203, + "grad_norm": 7.6230550634556655, + "learning_rate": 1.8255494868118933e-05, + "loss": 1.3818, + "step": 1982 + }, + { + "epoch": 1.0765472312703583, + "grad_norm": 6.946643864799888, + "learning_rate": 1.8253509700591242e-05, + "loss": 1.1384, + "step": 1983 + }, + { + "epoch": 1.0770901194353963, + "grad_norm": 7.288235795183878, + "learning_rate": 1.825152351225632e-05, + "loss": 1.1288, + "step": 1984 + }, + { + "epoch": 1.0776330076004343, + "grad_norm": 8.872287367521855, + "learning_rate": 1.8249536303359816e-05, + "loss": 1.6016, + "step": 1985 + }, + { + "epoch": 1.0781758957654723, + "grad_norm": 5.75562973378046, + "learning_rate": 1.8247548074147515e-05, + "loss": 1.1803, + "step": 1986 + }, + { + "epoch": 1.0787187839305103, + "grad_norm": 6.377752945158632, + "learning_rate": 1.824555882486532e-05, + "loss": 0.6991, + "step": 1987 + }, + { + "epoch": 1.0792616720955484, + "grad_norm": 7.178965937765225, + "learning_rate": 1.8243568555759274e-05, + "loss": 0.988, + "step": 1988 + }, + { + "epoch": 1.0798045602605864, + "grad_norm": 6.0869015160234525, + "learning_rate": 1.824157726707553e-05, + "loss": 0.8587, + "step": 1989 + }, + { + "epoch": 1.0803474484256244, + "grad_norm": 5.9113747723637395, + "learning_rate": 1.823958495906037e-05, + "loss": 0.8815, + "step": 1990 + }, + { + "epoch": 1.0808903365906624, + "grad_norm": 7.84172234151508, + "learning_rate": 1.8237591631960218e-05, + "loss": 1.2621, + "step": 1991 + }, + { + "epoch": 1.0814332247557004, + "grad_norm": 7.188505727546616, + "learning_rate": 1.8235597286021597e-05, + "loss": 0.5991, + "step": 1992 + }, + { + "epoch": 1.0819761129207384, + "grad_norm": 6.358900329219305, + "learning_rate": 1.823360192149118e-05, + "loss": 1.3325, + "step": 1993 + }, + { + "epoch": 1.0825190010857764, + "grad_norm": 6.545296131842583, + "learning_rate": 1.8231605538615756e-05, + "loss": 1.1091, + "step": 1994 + }, + { + "epoch": 1.0830618892508144, + "grad_norm": 6.708966367161533, + "learning_rate": 1.8229608137642238e-05, + "loss": 1.1352, + "step": 1995 + }, + { + "epoch": 1.0836047774158524, + "grad_norm": 6.651776068492304, + "learning_rate": 1.822760971881767e-05, + "loss": 0.6818, + "step": 1996 + }, + { + "epoch": 1.0841476655808904, + "grad_norm": 4.745518059898734, + "learning_rate": 1.8225610282389222e-05, + "loss": 0.9403, + "step": 1997 + }, + { + "epoch": 1.0846905537459284, + "grad_norm": 6.526942367542791, + "learning_rate": 1.8223609828604184e-05, + "loss": 1.0543, + "step": 1998 + }, + { + "epoch": 1.0852334419109664, + "grad_norm": 5.331318824902818, + "learning_rate": 1.8221608357709973e-05, + "loss": 0.6626, + "step": 1999 + }, + { + "epoch": 1.0857763300760044, + "grad_norm": 6.640324946305343, + "learning_rate": 1.8219605869954134e-05, + "loss": 1.0166, + "step": 2000 + }, + { + "epoch": 1.0863192182410424, + "grad_norm": 5.684199608008352, + "learning_rate": 1.8217602365584352e-05, + "loss": 0.6796, + "step": 2001 + }, + { + "epoch": 1.0868621064060804, + "grad_norm": 5.429738239817468, + "learning_rate": 1.8215597844848403e-05, + "loss": 0.6783, + "step": 2002 + }, + { + "epoch": 1.0874049945711184, + "grad_norm": 6.1832795338939235, + "learning_rate": 1.821359230799422e-05, + "loss": 1.4244, + "step": 2003 + }, + { + "epoch": 1.0879478827361564, + "grad_norm": 7.118410329483546, + "learning_rate": 1.8211585755269852e-05, + "loss": 1.0274, + "step": 2004 + }, + { + "epoch": 1.0884907709011944, + "grad_norm": 5.5061218005078265, + "learning_rate": 1.820957818692347e-05, + "loss": 0.7143, + "step": 2005 + }, + { + "epoch": 1.0890336590662324, + "grad_norm": 5.518412565715972, + "learning_rate": 1.8207569603203373e-05, + "loss": 1.3158, + "step": 2006 + }, + { + "epoch": 1.0895765472312704, + "grad_norm": 5.47000403426222, + "learning_rate": 1.820556000435799e-05, + "loss": 1.0694, + "step": 2007 + }, + { + "epoch": 1.0901194353963084, + "grad_norm": 7.469961154104622, + "learning_rate": 1.820354939063586e-05, + "loss": 1.0702, + "step": 2008 + }, + { + "epoch": 1.0906623235613464, + "grad_norm": 5.889961089255285, + "learning_rate": 1.8201537762285674e-05, + "loss": 1.0598, + "step": 2009 + }, + { + "epoch": 1.0912052117263844, + "grad_norm": 6.659595389259375, + "learning_rate": 1.8199525119556226e-05, + "loss": 0.8456, + "step": 2010 + }, + { + "epoch": 1.0917480998914224, + "grad_norm": 6.328157257432415, + "learning_rate": 1.8197511462696443e-05, + "loss": 0.9763, + "step": 2011 + }, + { + "epoch": 1.0922909880564604, + "grad_norm": 5.916734142619132, + "learning_rate": 1.8195496791955373e-05, + "loss": 1.1194, + "step": 2012 + }, + { + "epoch": 1.0928338762214984, + "grad_norm": 5.904087624332602, + "learning_rate": 1.8193481107582203e-05, + "loss": 0.756, + "step": 2013 + }, + { + "epoch": 1.0933767643865364, + "grad_norm": 6.284745154848233, + "learning_rate": 1.8191464409826227e-05, + "loss": 0.7931, + "step": 2014 + }, + { + "epoch": 1.0939196525515744, + "grad_norm": 7.838306243079945, + "learning_rate": 1.8189446698936878e-05, + "loss": 1.3635, + "step": 2015 + }, + { + "epoch": 1.0944625407166124, + "grad_norm": 6.3405758352546995, + "learning_rate": 1.818742797516371e-05, + "loss": 0.5365, + "step": 2016 + }, + { + "epoch": 1.0950054288816504, + "grad_norm": 8.315910084823138, + "learning_rate": 1.8185408238756405e-05, + "loss": 1.6592, + "step": 2017 + }, + { + "epoch": 1.0955483170466884, + "grad_norm": 5.205053142209431, + "learning_rate": 1.8183387489964762e-05, + "loss": 0.6416, + "step": 2018 + }, + { + "epoch": 1.0960912052117264, + "grad_norm": 7.018373673519811, + "learning_rate": 1.8181365729038706e-05, + "loss": 1.0814, + "step": 2019 + }, + { + "epoch": 1.0966340933767644, + "grad_norm": 5.705860306884454, + "learning_rate": 1.8179342956228307e-05, + "loss": 0.5806, + "step": 2020 + }, + { + "epoch": 1.0971769815418024, + "grad_norm": 7.077662290720525, + "learning_rate": 1.8177319171783728e-05, + "loss": 0.8969, + "step": 2021 + }, + { + "epoch": 1.0977198697068404, + "grad_norm": 6.540901861387866, + "learning_rate": 1.8175294375955284e-05, + "loss": 0.7982, + "step": 2022 + }, + { + "epoch": 1.0982627578718784, + "grad_norm": 7.283213833491601, + "learning_rate": 1.81732685689934e-05, + "loss": 1.2659, + "step": 2023 + }, + { + "epoch": 1.0988056460369164, + "grad_norm": 7.016589270114781, + "learning_rate": 1.8171241751148633e-05, + "loss": 1.0011, + "step": 2024 + }, + { + "epoch": 1.0993485342019544, + "grad_norm": 5.724817250167612, + "learning_rate": 1.8169213922671666e-05, + "loss": 0.8967, + "step": 2025 + }, + { + "epoch": 1.0998914223669924, + "grad_norm": 7.431207902407107, + "learning_rate": 1.81671850838133e-05, + "loss": 0.9445, + "step": 2026 + }, + { + "epoch": 1.1004343105320304, + "grad_norm": 6.9152342645144955, + "learning_rate": 1.816515523482447e-05, + "loss": 0.8187, + "step": 2027 + }, + { + "epoch": 1.1009771986970684, + "grad_norm": 7.045028624610342, + "learning_rate": 1.816312437595622e-05, + "loss": 0.7323, + "step": 2028 + }, + { + "epoch": 1.1015200868621065, + "grad_norm": 6.63265573559803, + "learning_rate": 1.816109250745974e-05, + "loss": 0.6689, + "step": 2029 + }, + { + "epoch": 1.1020629750271445, + "grad_norm": 8.605675554379312, + "learning_rate": 1.8159059629586333e-05, + "loss": 1.3733, + "step": 2030 + }, + { + "epoch": 1.1026058631921825, + "grad_norm": 7.481808301387036, + "learning_rate": 1.8157025742587426e-05, + "loss": 0.9767, + "step": 2031 + }, + { + "epoch": 1.1031487513572205, + "grad_norm": 7.113214270917786, + "learning_rate": 1.8154990846714575e-05, + "loss": 0.6365, + "step": 2032 + }, + { + "epoch": 1.1036916395222585, + "grad_norm": 7.672152203903941, + "learning_rate": 1.8152954942219462e-05, + "loss": 1.3314, + "step": 2033 + }, + { + "epoch": 1.1042345276872965, + "grad_norm": 7.533257156973554, + "learning_rate": 1.8150918029353885e-05, + "loss": 0.967, + "step": 2034 + }, + { + "epoch": 1.1047774158523345, + "grad_norm": 6.469759228237502, + "learning_rate": 1.8148880108369775e-05, + "loss": 1.348, + "step": 2035 + }, + { + "epoch": 1.1053203040173725, + "grad_norm": 6.620692148711997, + "learning_rate": 1.8146841179519186e-05, + "loss": 0.7538, + "step": 2036 + }, + { + "epoch": 1.1058631921824105, + "grad_norm": 6.974705446587608, + "learning_rate": 1.8144801243054297e-05, + "loss": 1.3664, + "step": 2037 + }, + { + "epoch": 1.1064060803474485, + "grad_norm": 8.047700738112871, + "learning_rate": 1.8142760299227408e-05, + "loss": 1.0466, + "step": 2038 + }, + { + "epoch": 1.1069489685124865, + "grad_norm": 5.469412128245213, + "learning_rate": 1.814071834829095e-05, + "loss": 0.5464, + "step": 2039 + }, + { + "epoch": 1.1074918566775245, + "grad_norm": 5.732997692484992, + "learning_rate": 1.813867539049747e-05, + "loss": 0.7307, + "step": 2040 + }, + { + "epoch": 1.1080347448425625, + "grad_norm": 8.89935072196041, + "learning_rate": 1.8136631426099646e-05, + "loss": 1.8231, + "step": 2041 + }, + { + "epoch": 1.1085776330076005, + "grad_norm": 7.0036806828899465, + "learning_rate": 1.813458645535028e-05, + "loss": 1.063, + "step": 2042 + }, + { + "epoch": 1.1091205211726385, + "grad_norm": 6.013891951698058, + "learning_rate": 1.8132540478502297e-05, + "loss": 0.9836, + "step": 2043 + }, + { + "epoch": 1.1096634093376765, + "grad_norm": 4.234811309290956, + "learning_rate": 1.813049349580875e-05, + "loss": 0.3708, + "step": 2044 + }, + { + "epoch": 1.1102062975027145, + "grad_norm": 6.259459841046158, + "learning_rate": 1.8128445507522806e-05, + "loss": 0.8463, + "step": 2045 + }, + { + "epoch": 1.1107491856677525, + "grad_norm": 7.9568342619649775, + "learning_rate": 1.8126396513897764e-05, + "loss": 1.0563, + "step": 2046 + }, + { + "epoch": 1.1112920738327905, + "grad_norm": 6.937848428371049, + "learning_rate": 1.8124346515187056e-05, + "loss": 0.6359, + "step": 2047 + }, + { + "epoch": 1.1118349619978285, + "grad_norm": 7.672632972051955, + "learning_rate": 1.8122295511644218e-05, + "loss": 0.657, + "step": 2048 + }, + { + "epoch": 1.1123778501628665, + "grad_norm": 6.450214026534184, + "learning_rate": 1.8120243503522924e-05, + "loss": 0.7279, + "step": 2049 + }, + { + "epoch": 1.1129207383279045, + "grad_norm": 6.628216836060006, + "learning_rate": 1.8118190491076978e-05, + "loss": 0.5804, + "step": 2050 + }, + { + "epoch": 1.1134636264929425, + "grad_norm": 6.698074316145226, + "learning_rate": 1.8116136474560288e-05, + "loss": 1.011, + "step": 2051 + }, + { + "epoch": 1.1140065146579805, + "grad_norm": 5.4126983291571795, + "learning_rate": 1.8114081454226905e-05, + "loss": 0.8564, + "step": 2052 + }, + { + "epoch": 1.1145494028230185, + "grad_norm": 8.596797841329312, + "learning_rate": 1.8112025430331e-05, + "loss": 1.1924, + "step": 2053 + }, + { + "epoch": 1.1150922909880565, + "grad_norm": 7.416912053391948, + "learning_rate": 1.8109968403126856e-05, + "loss": 0.9894, + "step": 2054 + }, + { + "epoch": 1.1156351791530945, + "grad_norm": 7.434308789484601, + "learning_rate": 1.8107910372868898e-05, + "loss": 1.1544, + "step": 2055 + }, + { + "epoch": 1.1161780673181325, + "grad_norm": 8.325689788099726, + "learning_rate": 1.8105851339811663e-05, + "loss": 0.9994, + "step": 2056 + }, + { + "epoch": 1.1167209554831705, + "grad_norm": 8.384964960696278, + "learning_rate": 1.8103791304209813e-05, + "loss": 1.1768, + "step": 2057 + }, + { + "epoch": 1.1172638436482085, + "grad_norm": 7.044535285708613, + "learning_rate": 1.810173026631814e-05, + "loss": 1.1801, + "step": 2058 + }, + { + "epoch": 1.1178067318132465, + "grad_norm": 6.665810359686943, + "learning_rate": 1.8099668226391552e-05, + "loss": 0.7865, + "step": 2059 + }, + { + "epoch": 1.1183496199782845, + "grad_norm": 7.395697719004186, + "learning_rate": 1.8097605184685093e-05, + "loss": 1.1357, + "step": 2060 + }, + { + "epoch": 1.1188925081433225, + "grad_norm": 7.1495559268215185, + "learning_rate": 1.809554114145392e-05, + "loss": 1.1507, + "step": 2061 + }, + { + "epoch": 1.1194353963083605, + "grad_norm": 7.877063869804663, + "learning_rate": 1.8093476096953315e-05, + "loss": 1.634, + "step": 2062 + }, + { + "epoch": 1.1199782844733985, + "grad_norm": 6.528060830825324, + "learning_rate": 1.809141005143869e-05, + "loss": 0.697, + "step": 2063 + }, + { + "epoch": 1.1205211726384365, + "grad_norm": 6.5949576624011, + "learning_rate": 1.808934300516557e-05, + "loss": 1.1506, + "step": 2064 + }, + { + "epoch": 1.1210640608034745, + "grad_norm": 6.787348272435354, + "learning_rate": 1.8087274958389612e-05, + "loss": 0.9986, + "step": 2065 + }, + { + "epoch": 1.1216069489685125, + "grad_norm": 6.310955809912108, + "learning_rate": 1.8085205911366602e-05, + "loss": 0.89, + "step": 2066 + }, + { + "epoch": 1.1221498371335505, + "grad_norm": 6.8487503881448895, + "learning_rate": 1.8083135864352442e-05, + "loss": 1.1061, + "step": 2067 + }, + { + "epoch": 1.1226927252985885, + "grad_norm": 5.329494375939551, + "learning_rate": 1.808106481760315e-05, + "loss": 1.0011, + "step": 2068 + }, + { + "epoch": 1.1232356134636265, + "grad_norm": 6.286005313228098, + "learning_rate": 1.8078992771374886e-05, + "loss": 0.8399, + "step": 2069 + }, + { + "epoch": 1.1237785016286646, + "grad_norm": 6.324896503266866, + "learning_rate": 1.8076919725923917e-05, + "loss": 1.2716, + "step": 2070 + }, + { + "epoch": 1.1243213897937026, + "grad_norm": 5.663105266452551, + "learning_rate": 1.8074845681506644e-05, + "loss": 0.9943, + "step": 2071 + }, + { + "epoch": 1.1248642779587406, + "grad_norm": 5.175572461557621, + "learning_rate": 1.807277063837959e-05, + "loss": 0.6319, + "step": 2072 + }, + { + "epoch": 1.1254071661237786, + "grad_norm": 7.297487212372251, + "learning_rate": 1.8070694596799397e-05, + "loss": 0.7533, + "step": 2073 + }, + { + "epoch": 1.1259500542888166, + "grad_norm": 9.046578233072974, + "learning_rate": 1.806861755702283e-05, + "loss": 0.9991, + "step": 2074 + }, + { + "epoch": 1.1264929424538546, + "grad_norm": 4.516511292735119, + "learning_rate": 1.8066539519306786e-05, + "loss": 0.6393, + "step": 2075 + }, + { + "epoch": 1.1270358306188926, + "grad_norm": 5.16623460840236, + "learning_rate": 1.8064460483908283e-05, + "loss": 0.7083, + "step": 2076 + }, + { + "epoch": 1.1275787187839306, + "grad_norm": 10.7323945119629, + "learning_rate": 1.8062380451084445e-05, + "loss": 2.4193, + "step": 2077 + }, + { + "epoch": 1.1281216069489686, + "grad_norm": 6.438876475868993, + "learning_rate": 1.8060299421092554e-05, + "loss": 0.8468, + "step": 2078 + }, + { + "epoch": 1.1286644951140066, + "grad_norm": 8.437791617086905, + "learning_rate": 1.8058217394189976e-05, + "loss": 1.4895, + "step": 2079 + }, + { + "epoch": 1.1292073832790446, + "grad_norm": 5.237528459736948, + "learning_rate": 1.805613437063423e-05, + "loss": 0.643, + "step": 2080 + }, + { + "epoch": 1.1297502714440826, + "grad_norm": 6.010151293762264, + "learning_rate": 1.8054050350682947e-05, + "loss": 0.9001, + "step": 2081 + }, + { + "epoch": 1.1302931596091206, + "grad_norm": 4.816849804237718, + "learning_rate": 1.805196533459388e-05, + "loss": 0.6349, + "step": 2082 + }, + { + "epoch": 1.1308360477741586, + "grad_norm": 6.872496495387926, + "learning_rate": 1.8049879322624906e-05, + "loss": 0.8808, + "step": 2083 + }, + { + "epoch": 1.1313789359391966, + "grad_norm": 5.86067647581003, + "learning_rate": 1.804779231503403e-05, + "loss": 0.6811, + "step": 2084 + }, + { + "epoch": 1.1319218241042346, + "grad_norm": 6.073217718554074, + "learning_rate": 1.8045704312079376e-05, + "loss": 1.0086, + "step": 2085 + }, + { + "epoch": 1.1324647122692726, + "grad_norm": 7.519215180241719, + "learning_rate": 1.804361531401918e-05, + "loss": 0.8815, + "step": 2086 + }, + { + "epoch": 1.1330076004343106, + "grad_norm": 6.49012361077388, + "learning_rate": 1.8041525321111835e-05, + "loss": 1.0129, + "step": 2087 + }, + { + "epoch": 1.1335504885993486, + "grad_norm": 5.938773374319115, + "learning_rate": 1.8039434333615814e-05, + "loss": 0.8485, + "step": 2088 + }, + { + "epoch": 1.1340933767643866, + "grad_norm": 5.607636698742016, + "learning_rate": 1.8037342351789743e-05, + "loss": 0.6209, + "step": 2089 + }, + { + "epoch": 1.1346362649294246, + "grad_norm": 5.522935966563286, + "learning_rate": 1.803524937589236e-05, + "loss": 0.6576, + "step": 2090 + }, + { + "epoch": 1.1351791530944626, + "grad_norm": 8.130974168982206, + "learning_rate": 1.8033155406182533e-05, + "loss": 1.0676, + "step": 2091 + }, + { + "epoch": 1.1357220412595006, + "grad_norm": 6.027276275319545, + "learning_rate": 1.803106044291924e-05, + "loss": 0.9853, + "step": 2092 + }, + { + "epoch": 1.1362649294245386, + "grad_norm": 4.645985748184023, + "learning_rate": 1.8028964486361586e-05, + "loss": 0.7505, + "step": 2093 + }, + { + "epoch": 1.1368078175895766, + "grad_norm": 6.820894579082682, + "learning_rate": 1.8026867536768816e-05, + "loss": 0.9358, + "step": 2094 + }, + { + "epoch": 1.1373507057546146, + "grad_norm": 3.9530727606720872, + "learning_rate": 1.802476959440027e-05, + "loss": 0.7736, + "step": 2095 + }, + { + "epoch": 1.1378935939196526, + "grad_norm": 7.650275740883737, + "learning_rate": 1.8022670659515432e-05, + "loss": 1.4832, + "step": 2096 + }, + { + "epoch": 1.1384364820846906, + "grad_norm": 6.94341164115752, + "learning_rate": 1.80205707323739e-05, + "loss": 1.0336, + "step": 2097 + }, + { + "epoch": 1.1389793702497286, + "grad_norm": 5.608528053150328, + "learning_rate": 1.8018469813235403e-05, + "loss": 0.829, + "step": 2098 + }, + { + "epoch": 1.1395222584147666, + "grad_norm": 5.2191336182398995, + "learning_rate": 1.8016367902359776e-05, + "loss": 0.9708, + "step": 2099 + }, + { + "epoch": 1.1400651465798046, + "grad_norm": 6.773346810380977, + "learning_rate": 1.801426500000699e-05, + "loss": 0.6755, + "step": 2100 + }, + { + "epoch": 1.1406080347448426, + "grad_norm": 7.292992073718992, + "learning_rate": 1.8012161106437137e-05, + "loss": 0.8842, + "step": 2101 + }, + { + "epoch": 1.1411509229098806, + "grad_norm": 6.699434744454233, + "learning_rate": 1.8010056221910427e-05, + "loss": 1.0385, + "step": 2102 + }, + { + "epoch": 1.1416938110749186, + "grad_norm": 6.565299107509569, + "learning_rate": 1.8007950346687198e-05, + "loss": 0.6557, + "step": 2103 + }, + { + "epoch": 1.1422366992399566, + "grad_norm": 7.2741514068627025, + "learning_rate": 1.800584348102791e-05, + "loss": 1.0347, + "step": 2104 + }, + { + "epoch": 1.1427795874049946, + "grad_norm": 6.467241522225306, + "learning_rate": 1.800373562519314e-05, + "loss": 0.8026, + "step": 2105 + }, + { + "epoch": 1.1433224755700326, + "grad_norm": 6.109318081193457, + "learning_rate": 1.800162677944359e-05, + "loss": 0.9182, + "step": 2106 + }, + { + "epoch": 1.1438653637350706, + "grad_norm": 7.136347555199431, + "learning_rate": 1.7999516944040087e-05, + "loss": 1.2743, + "step": 2107 + }, + { + "epoch": 1.1444082519001086, + "grad_norm": 7.791284332349926, + "learning_rate": 1.7997406119243582e-05, + "loss": 0.9738, + "step": 2108 + }, + { + "epoch": 1.1449511400651466, + "grad_norm": 7.315957477694666, + "learning_rate": 1.7995294305315137e-05, + "loss": 1.0674, + "step": 2109 + }, + { + "epoch": 1.1454940282301846, + "grad_norm": 6.305984776426275, + "learning_rate": 1.7993181502515957e-05, + "loss": 0.8204, + "step": 2110 + }, + { + "epoch": 1.1460369163952226, + "grad_norm": 5.824114865699788, + "learning_rate": 1.7991067711107345e-05, + "loss": 0.7953, + "step": 2111 + }, + { + "epoch": 1.1465798045602607, + "grad_norm": 7.4284096250312786, + "learning_rate": 1.798895293135074e-05, + "loss": 0.7349, + "step": 2112 + }, + { + "epoch": 1.1471226927252987, + "grad_norm": 6.415949338549894, + "learning_rate": 1.798683716350771e-05, + "loss": 0.8346, + "step": 2113 + }, + { + "epoch": 1.1476655808903367, + "grad_norm": 6.360883520037482, + "learning_rate": 1.7984720407839925e-05, + "loss": 0.7966, + "step": 2114 + }, + { + "epoch": 1.1482084690553747, + "grad_norm": 6.703196393264975, + "learning_rate": 1.7982602664609198e-05, + "loss": 1.3341, + "step": 2115 + }, + { + "epoch": 1.1487513572204127, + "grad_norm": 8.115384707506347, + "learning_rate": 1.798048393407745e-05, + "loss": 1.0284, + "step": 2116 + }, + { + "epoch": 1.1492942453854507, + "grad_norm": 8.519166176194911, + "learning_rate": 1.797836421650673e-05, + "loss": 1.4075, + "step": 2117 + }, + { + "epoch": 1.1498371335504887, + "grad_norm": 7.120345871314324, + "learning_rate": 1.7976243512159207e-05, + "loss": 1.1933, + "step": 2118 + }, + { + "epoch": 1.1503800217155267, + "grad_norm": 5.887743000051609, + "learning_rate": 1.7974121821297178e-05, + "loss": 0.9143, + "step": 2119 + }, + { + "epoch": 1.1509229098805647, + "grad_norm": 6.264158167802545, + "learning_rate": 1.797199914418305e-05, + "loss": 0.805, + "step": 2120 + }, + { + "epoch": 1.1514657980456027, + "grad_norm": 7.747345696360016, + "learning_rate": 1.7969875481079363e-05, + "loss": 1.1374, + "step": 2121 + }, + { + "epoch": 1.1520086862106407, + "grad_norm": 5.932381639003466, + "learning_rate": 1.7967750832248774e-05, + "loss": 0.7801, + "step": 2122 + }, + { + "epoch": 1.1525515743756787, + "grad_norm": 5.646019326027894, + "learning_rate": 1.7965625197954064e-05, + "loss": 0.5053, + "step": 2123 + }, + { + "epoch": 1.1530944625407167, + "grad_norm": 6.085604242778934, + "learning_rate": 1.7963498578458135e-05, + "loss": 0.6711, + "step": 2124 + }, + { + "epoch": 1.1536373507057547, + "grad_norm": 6.390132028289734, + "learning_rate": 1.7961370974024008e-05, + "loss": 0.9897, + "step": 2125 + }, + { + "epoch": 1.1541802388707927, + "grad_norm": 5.052558465006266, + "learning_rate": 1.795924238491483e-05, + "loss": 0.8742, + "step": 2126 + }, + { + "epoch": 1.1547231270358307, + "grad_norm": 5.86713723099216, + "learning_rate": 1.7957112811393867e-05, + "loss": 0.965, + "step": 2127 + }, + { + "epoch": 1.1552660152008687, + "grad_norm": 6.092033075473227, + "learning_rate": 1.7954982253724515e-05, + "loss": 1.104, + "step": 2128 + }, + { + "epoch": 1.1558089033659067, + "grad_norm": 7.051795930424219, + "learning_rate": 1.7952850712170278e-05, + "loss": 1.0918, + "step": 2129 + }, + { + "epoch": 1.1563517915309447, + "grad_norm": 5.613114568974867, + "learning_rate": 1.7950718186994788e-05, + "loss": 0.6698, + "step": 2130 + }, + { + "epoch": 1.1568946796959827, + "grad_norm": 5.533439816776461, + "learning_rate": 1.7948584678461802e-05, + "loss": 1.4395, + "step": 2131 + }, + { + "epoch": 1.1574375678610207, + "grad_norm": 7.73758559868109, + "learning_rate": 1.7946450186835195e-05, + "loss": 1.0806, + "step": 2132 + }, + { + "epoch": 1.1579804560260587, + "grad_norm": 5.982378084297589, + "learning_rate": 1.7944314712378966e-05, + "loss": 1.0428, + "step": 2133 + }, + { + "epoch": 1.1585233441910967, + "grad_norm": 5.604421383798244, + "learning_rate": 1.7942178255357227e-05, + "loss": 0.9411, + "step": 2134 + }, + { + "epoch": 1.1590662323561347, + "grad_norm": 6.5637069585312435, + "learning_rate": 1.794004081603423e-05, + "loss": 0.8285, + "step": 2135 + }, + { + "epoch": 1.1596091205211727, + "grad_norm": 5.567535422744373, + "learning_rate": 1.7937902394674326e-05, + "loss": 1.0185, + "step": 2136 + }, + { + "epoch": 1.1601520086862107, + "grad_norm": 7.898198443147329, + "learning_rate": 1.7935762991542004e-05, + "loss": 0.8878, + "step": 2137 + }, + { + "epoch": 1.1606948968512487, + "grad_norm": 5.12407458777131, + "learning_rate": 1.7933622606901865e-05, + "loss": 0.9174, + "step": 2138 + }, + { + "epoch": 1.1612377850162867, + "grad_norm": 4.480165604953252, + "learning_rate": 1.7931481241018642e-05, + "loss": 0.9424, + "step": 2139 + }, + { + "epoch": 1.1617806731813247, + "grad_norm": 7.6708161194386495, + "learning_rate": 1.7929338894157173e-05, + "loss": 0.9875, + "step": 2140 + }, + { + "epoch": 1.1623235613463627, + "grad_norm": 9.749916559657342, + "learning_rate": 1.7927195566582435e-05, + "loss": 1.7168, + "step": 2141 + }, + { + "epoch": 1.1628664495114007, + "grad_norm": 5.50355689991458, + "learning_rate": 1.7925051258559516e-05, + "loss": 0.7957, + "step": 2142 + }, + { + "epoch": 1.1634093376764387, + "grad_norm": 5.910752363178181, + "learning_rate": 1.7922905970353627e-05, + "loss": 1.0525, + "step": 2143 + }, + { + "epoch": 1.1639522258414767, + "grad_norm": 6.523454927461212, + "learning_rate": 1.7920759702230098e-05, + "loss": 0.9938, + "step": 2144 + }, + { + "epoch": 1.1644951140065147, + "grad_norm": 6.963057538162332, + "learning_rate": 1.7918612454454387e-05, + "loss": 0.8677, + "step": 2145 + }, + { + "epoch": 1.1650380021715527, + "grad_norm": 6.55011455398606, + "learning_rate": 1.7916464227292067e-05, + "loss": 0.7268, + "step": 2146 + }, + { + "epoch": 1.1655808903365907, + "grad_norm": 6.47957363507043, + "learning_rate": 1.7914315021008836e-05, + "loss": 0.7449, + "step": 2147 + }, + { + "epoch": 1.1661237785016287, + "grad_norm": 6.574809242111077, + "learning_rate": 1.7912164835870506e-05, + "loss": 0.6999, + "step": 2148 + }, + { + "epoch": 1.1666666666666667, + "grad_norm": 8.307297080129953, + "learning_rate": 1.791001367214302e-05, + "loss": 1.0402, + "step": 2149 + }, + { + "epoch": 1.1672095548317047, + "grad_norm": 8.01469917888599, + "learning_rate": 1.7907861530092438e-05, + "loss": 1.1024, + "step": 2150 + }, + { + "epoch": 1.1677524429967427, + "grad_norm": 7.336381076899337, + "learning_rate": 1.7905708409984938e-05, + "loss": 0.9112, + "step": 2151 + }, + { + "epoch": 1.1682953311617807, + "grad_norm": 7.872105686379811, + "learning_rate": 1.7903554312086823e-05, + "loss": 0.9002, + "step": 2152 + }, + { + "epoch": 1.1688382193268188, + "grad_norm": 5.842761146463113, + "learning_rate": 1.7901399236664514e-05, + "loss": 0.9457, + "step": 2153 + }, + { + "epoch": 1.1693811074918568, + "grad_norm": 5.8100183378475005, + "learning_rate": 1.7899243183984552e-05, + "loss": 0.9675, + "step": 2154 + }, + { + "epoch": 1.1699239956568948, + "grad_norm": 5.208689262592845, + "learning_rate": 1.7897086154313604e-05, + "loss": 0.7562, + "step": 2155 + }, + { + "epoch": 1.1704668838219328, + "grad_norm": 7.695958654055186, + "learning_rate": 1.7894928147918458e-05, + "loss": 0.9995, + "step": 2156 + }, + { + "epoch": 1.1710097719869708, + "grad_norm": 5.8108200610880525, + "learning_rate": 1.789276916506601e-05, + "loss": 0.8179, + "step": 2157 + }, + { + "epoch": 1.1715526601520088, + "grad_norm": 5.776078861908261, + "learning_rate": 1.7890609206023298e-05, + "loss": 0.6612, + "step": 2158 + }, + { + "epoch": 1.1720955483170468, + "grad_norm": 6.458218841014727, + "learning_rate": 1.7888448271057463e-05, + "loss": 0.8419, + "step": 2159 + }, + { + "epoch": 1.1726384364820848, + "grad_norm": 7.236506499419343, + "learning_rate": 1.7886286360435772e-05, + "loss": 0.6336, + "step": 2160 + }, + { + "epoch": 1.1731813246471228, + "grad_norm": 7.20552304926134, + "learning_rate": 1.7884123474425614e-05, + "loss": 0.9157, + "step": 2161 + }, + { + "epoch": 1.1737242128121608, + "grad_norm": 8.090329675463652, + "learning_rate": 1.78819596132945e-05, + "loss": 1.6894, + "step": 2162 + }, + { + "epoch": 1.1742671009771988, + "grad_norm": 7.500510561892478, + "learning_rate": 1.787979477731006e-05, + "loss": 0.9469, + "step": 2163 + }, + { + "epoch": 1.1748099891422368, + "grad_norm": 7.6126459328076415, + "learning_rate": 1.7877628966740044e-05, + "loss": 0.9179, + "step": 2164 + }, + { + "epoch": 1.1753528773072748, + "grad_norm": 7.065751172246334, + "learning_rate": 1.787546218185232e-05, + "loss": 0.9649, + "step": 2165 + }, + { + "epoch": 1.1758957654723128, + "grad_norm": 6.923487399525626, + "learning_rate": 1.787329442291488e-05, + "loss": 1.0791, + "step": 2166 + }, + { + "epoch": 1.1764386536373508, + "grad_norm": 7.614292041669914, + "learning_rate": 1.7871125690195843e-05, + "loss": 1.2367, + "step": 2167 + }, + { + "epoch": 1.1769815418023888, + "grad_norm": 8.560468932577576, + "learning_rate": 1.7868955983963434e-05, + "loss": 1.0876, + "step": 2168 + }, + { + "epoch": 1.1775244299674268, + "grad_norm": 6.209836424501219, + "learning_rate": 1.786678530448601e-05, + "loss": 0.6251, + "step": 2169 + }, + { + "epoch": 1.1780673181324648, + "grad_norm": 5.926645140708837, + "learning_rate": 1.7864613652032035e-05, + "loss": 0.5939, + "step": 2170 + }, + { + "epoch": 1.1786102062975028, + "grad_norm": 5.460593849347875, + "learning_rate": 1.7862441026870114e-05, + "loss": 0.8769, + "step": 2171 + }, + { + "epoch": 1.1791530944625408, + "grad_norm": 8.063368903057139, + "learning_rate": 1.7860267429268954e-05, + "loss": 1.1492, + "step": 2172 + }, + { + "epoch": 1.1796959826275788, + "grad_norm": 6.246398264149539, + "learning_rate": 1.7858092859497392e-05, + "loss": 0.9146, + "step": 2173 + }, + { + "epoch": 1.1802388707926168, + "grad_norm": 9.483282452267977, + "learning_rate": 1.7855917317824383e-05, + "loss": 1.3714, + "step": 2174 + }, + { + "epoch": 1.1807817589576548, + "grad_norm": 7.377757446359213, + "learning_rate": 1.7853740804519e-05, + "loss": 0.7914, + "step": 2175 + }, + { + "epoch": 1.1813246471226928, + "grad_norm": 6.555194448443568, + "learning_rate": 1.7851563319850435e-05, + "loss": 1.0025, + "step": 2176 + }, + { + "epoch": 1.1818675352877308, + "grad_norm": 5.968820734002013, + "learning_rate": 1.7849384864088005e-05, + "loss": 1.0449, + "step": 2177 + }, + { + "epoch": 1.1824104234527688, + "grad_norm": 8.050731398373367, + "learning_rate": 1.7847205437501145e-05, + "loss": 1.2944, + "step": 2178 + }, + { + "epoch": 1.1829533116178068, + "grad_norm": 6.479093872062473, + "learning_rate": 1.784502504035941e-05, + "loss": 0.8312, + "step": 2179 + }, + { + "epoch": 1.1834961997828448, + "grad_norm": 9.645172625032446, + "learning_rate": 1.7842843672932473e-05, + "loss": 1.4877, + "step": 2180 + }, + { + "epoch": 1.1840390879478828, + "grad_norm": 7.811721210097083, + "learning_rate": 1.7840661335490133e-05, + "loss": 1.0861, + "step": 2181 + }, + { + "epoch": 1.1845819761129208, + "grad_norm": 5.939133632197409, + "learning_rate": 1.7838478028302303e-05, + "loss": 0.7359, + "step": 2182 + }, + { + "epoch": 1.1851248642779588, + "grad_norm": 7.121597254982603, + "learning_rate": 1.7836293751639017e-05, + "loss": 0.626, + "step": 2183 + }, + { + "epoch": 1.1856677524429968, + "grad_norm": 7.225734005859657, + "learning_rate": 1.783410850577043e-05, + "loss": 1.0651, + "step": 2184 + }, + { + "epoch": 1.1862106406080348, + "grad_norm": 6.3733995711762965, + "learning_rate": 1.783192229096682e-05, + "loss": 1.0399, + "step": 2185 + }, + { + "epoch": 1.1867535287730728, + "grad_norm": 7.215840642904109, + "learning_rate": 1.7829735107498576e-05, + "loss": 1.0958, + "step": 2186 + }, + { + "epoch": 1.1872964169381108, + "grad_norm": 7.2349637750528135, + "learning_rate": 1.7827546955636216e-05, + "loss": 0.8005, + "step": 2187 + }, + { + "epoch": 1.1878393051031488, + "grad_norm": 7.592927799049675, + "learning_rate": 1.7825357835650376e-05, + "loss": 1.0981, + "step": 2188 + }, + { + "epoch": 1.1883821932681868, + "grad_norm": 8.010034188435748, + "learning_rate": 1.7823167747811805e-05, + "loss": 1.3088, + "step": 2189 + }, + { + "epoch": 1.1889250814332248, + "grad_norm": 5.674359355487917, + "learning_rate": 1.7820976692391377e-05, + "loss": 0.8725, + "step": 2190 + }, + { + "epoch": 1.1894679695982628, + "grad_norm": 5.236415159647034, + "learning_rate": 1.781878466966009e-05, + "loss": 0.7162, + "step": 2191 + }, + { + "epoch": 1.1900108577633008, + "grad_norm": 5.677268305787393, + "learning_rate": 1.781659167988905e-05, + "loss": 0.7101, + "step": 2192 + }, + { + "epoch": 1.1905537459283388, + "grad_norm": 6.370730170332413, + "learning_rate": 1.7814397723349496e-05, + "loss": 0.9087, + "step": 2193 + }, + { + "epoch": 1.1910966340933768, + "grad_norm": 6.045624314512159, + "learning_rate": 1.7812202800312776e-05, + "loss": 1.1714, + "step": 2194 + }, + { + "epoch": 1.1916395222584149, + "grad_norm": 10.798495850925244, + "learning_rate": 1.7810006911050366e-05, + "loss": 1.1997, + "step": 2195 + }, + { + "epoch": 1.1921824104234529, + "grad_norm": 5.89010386817935, + "learning_rate": 1.780781005583385e-05, + "loss": 1.0702, + "step": 2196 + }, + { + "epoch": 1.1927252985884909, + "grad_norm": 6.09421937705502, + "learning_rate": 1.7805612234934946e-05, + "loss": 0.825, + "step": 2197 + }, + { + "epoch": 1.1932681867535289, + "grad_norm": 5.582107046729241, + "learning_rate": 1.780341344862548e-05, + "loss": 0.8548, + "step": 2198 + }, + { + "epoch": 1.1938110749185669, + "grad_norm": 7.598179521038406, + "learning_rate": 1.78012136971774e-05, + "loss": 1.0015, + "step": 2199 + }, + { + "epoch": 1.1943539630836049, + "grad_norm": 7.4068666684121585, + "learning_rate": 1.7799012980862777e-05, + "loss": 1.0645, + "step": 2200 + }, + { + "epoch": 1.1948968512486429, + "grad_norm": 6.087565282375561, + "learning_rate": 1.7796811299953796e-05, + "loss": 0.9169, + "step": 2201 + }, + { + "epoch": 1.1954397394136809, + "grad_norm": 6.440325481735819, + "learning_rate": 1.7794608654722772e-05, + "loss": 1.0707, + "step": 2202 + }, + { + "epoch": 1.1959826275787189, + "grad_norm": 7.562903613167714, + "learning_rate": 1.7792405045442125e-05, + "loss": 1.0086, + "step": 2203 + }, + { + "epoch": 1.1965255157437569, + "grad_norm": 6.2377346284888, + "learning_rate": 1.77902004723844e-05, + "loss": 0.6111, + "step": 2204 + }, + { + "epoch": 1.1970684039087949, + "grad_norm": 5.978299124706355, + "learning_rate": 1.7787994935822268e-05, + "loss": 0.5782, + "step": 2205 + }, + { + "epoch": 1.1976112920738329, + "grad_norm": 5.9766366336106715, + "learning_rate": 1.778578843602851e-05, + "loss": 0.8409, + "step": 2206 + }, + { + "epoch": 1.1981541802388709, + "grad_norm": 7.700010929432555, + "learning_rate": 1.7783580973276027e-05, + "loss": 1.091, + "step": 2207 + }, + { + "epoch": 1.1986970684039089, + "grad_norm": 8.043903647823342, + "learning_rate": 1.778137254783785e-05, + "loss": 1.2894, + "step": 2208 + }, + { + "epoch": 1.1992399565689469, + "grad_norm": 8.709529168595635, + "learning_rate": 1.777916315998711e-05, + "loss": 1.5601, + "step": 2209 + }, + { + "epoch": 1.1997828447339849, + "grad_norm": 6.234520197542074, + "learning_rate": 1.7776952809997073e-05, + "loss": 1.0414, + "step": 2210 + }, + { + "epoch": 1.200325732899023, + "grad_norm": 5.8108372705767, + "learning_rate": 1.7774741498141116e-05, + "loss": 0.8811, + "step": 2211 + }, + { + "epoch": 1.200868621064061, + "grad_norm": 6.423388338643525, + "learning_rate": 1.7772529224692744e-05, + "loss": 0.8371, + "step": 2212 + }, + { + "epoch": 1.201411509229099, + "grad_norm": 7.740996825838731, + "learning_rate": 1.7770315989925565e-05, + "loss": 0.9993, + "step": 2213 + }, + { + "epoch": 1.201954397394137, + "grad_norm": 6.33257469290407, + "learning_rate": 1.7768101794113323e-05, + "loss": 0.8563, + "step": 2214 + }, + { + "epoch": 1.202497285559175, + "grad_norm": 8.126954888789392, + "learning_rate": 1.776588663752987e-05, + "loss": 0.9479, + "step": 2215 + }, + { + "epoch": 1.203040173724213, + "grad_norm": 7.781450888823224, + "learning_rate": 1.7763670520449178e-05, + "loss": 1.0559, + "step": 2216 + }, + { + "epoch": 1.203583061889251, + "grad_norm": 6.615763983931515, + "learning_rate": 1.7761453443145348e-05, + "loss": 0.7463, + "step": 2217 + }, + { + "epoch": 1.204125950054289, + "grad_norm": 10.193394757123674, + "learning_rate": 1.7759235405892584e-05, + "loss": 1.3597, + "step": 2218 + }, + { + "epoch": 1.204668838219327, + "grad_norm": 7.365896137130193, + "learning_rate": 1.7757016408965217e-05, + "loss": 0.5822, + "step": 2219 + }, + { + "epoch": 1.205211726384365, + "grad_norm": 8.530190774406401, + "learning_rate": 1.77547964526377e-05, + "loss": 1.0075, + "step": 2220 + }, + { + "epoch": 1.205754614549403, + "grad_norm": 7.143571088845206, + "learning_rate": 1.7752575537184597e-05, + "loss": 0.9395, + "step": 2221 + }, + { + "epoch": 1.206297502714441, + "grad_norm": 7.790375269151535, + "learning_rate": 1.7750353662880595e-05, + "loss": 1.0556, + "step": 2222 + }, + { + "epoch": 1.206840390879479, + "grad_norm": 6.5033471136457255, + "learning_rate": 1.77481308300005e-05, + "loss": 0.7464, + "step": 2223 + }, + { + "epoch": 1.207383279044517, + "grad_norm": 7.06026002468489, + "learning_rate": 1.774590703881924e-05, + "loss": 0.6315, + "step": 2224 + }, + { + "epoch": 1.207926167209555, + "grad_norm": 8.773109114641178, + "learning_rate": 1.774368228961185e-05, + "loss": 0.9884, + "step": 2225 + }, + { + "epoch": 1.208469055374593, + "grad_norm": 7.343364313617542, + "learning_rate": 1.774145658265349e-05, + "loss": 1.1542, + "step": 2226 + }, + { + "epoch": 1.209011943539631, + "grad_norm": 5.953950803365371, + "learning_rate": 1.773922991821944e-05, + "loss": 0.8485, + "step": 2227 + }, + { + "epoch": 1.209554831704669, + "grad_norm": 8.905484324028157, + "learning_rate": 1.7737002296585105e-05, + "loss": 1.0171, + "step": 2228 + }, + { + "epoch": 1.210097719869707, + "grad_norm": 7.158929117808072, + "learning_rate": 1.7734773718025992e-05, + "loss": 0.6134, + "step": 2229 + }, + { + "epoch": 1.210640608034745, + "grad_norm": 8.342029051568577, + "learning_rate": 1.7732544182817737e-05, + "loss": 0.9775, + "step": 2230 + }, + { + "epoch": 1.211183496199783, + "grad_norm": 5.587278578636139, + "learning_rate": 1.7730313691236098e-05, + "loss": 0.9395, + "step": 2231 + }, + { + "epoch": 1.211726384364821, + "grad_norm": 6.1948883535007475, + "learning_rate": 1.7728082243556936e-05, + "loss": 0.9977, + "step": 2232 + }, + { + "epoch": 1.212269272529859, + "grad_norm": 6.095377323439518, + "learning_rate": 1.7725849840056248e-05, + "loss": 0.9141, + "step": 2233 + }, + { + "epoch": 1.212812160694897, + "grad_norm": 8.71522968216411, + "learning_rate": 1.7723616481010137e-05, + "loss": 1.4356, + "step": 2234 + }, + { + "epoch": 1.213355048859935, + "grad_norm": 5.416863970522961, + "learning_rate": 1.7721382166694834e-05, + "loss": 0.6594, + "step": 2235 + }, + { + "epoch": 1.213897937024973, + "grad_norm": 8.046753742934493, + "learning_rate": 1.7719146897386674e-05, + "loss": 1.1187, + "step": 2236 + }, + { + "epoch": 1.214440825190011, + "grad_norm": 6.625446441699582, + "learning_rate": 1.7716910673362123e-05, + "loss": 0.8731, + "step": 2237 + }, + { + "epoch": 1.214983713355049, + "grad_norm": 6.47539429072121, + "learning_rate": 1.7714673494897767e-05, + "loss": 0.9473, + "step": 2238 + }, + { + "epoch": 1.215526601520087, + "grad_norm": 7.458434209050833, + "learning_rate": 1.7712435362270292e-05, + "loss": 0.7844, + "step": 2239 + }, + { + "epoch": 1.216069489685125, + "grad_norm": 6.935847404573923, + "learning_rate": 1.7710196275756524e-05, + "loss": 0.8431, + "step": 2240 + }, + { + "epoch": 1.216612377850163, + "grad_norm": 8.015017144687471, + "learning_rate": 1.770795623563339e-05, + "loss": 1.2329, + "step": 2241 + }, + { + "epoch": 1.217155266015201, + "grad_norm": 6.914919733505818, + "learning_rate": 1.7705715242177944e-05, + "loss": 1.1335, + "step": 2242 + }, + { + "epoch": 1.217698154180239, + "grad_norm": 5.900061304870694, + "learning_rate": 1.7703473295667354e-05, + "loss": 0.7453, + "step": 2243 + }, + { + "epoch": 1.218241042345277, + "grad_norm": 4.959614765477869, + "learning_rate": 1.770123039637891e-05, + "loss": 0.6093, + "step": 2244 + }, + { + "epoch": 1.2187839305103148, + "grad_norm": 8.038465432918807, + "learning_rate": 1.769898654459002e-05, + "loss": 1.0029, + "step": 2245 + }, + { + "epoch": 1.219326818675353, + "grad_norm": 8.08635156220498, + "learning_rate": 1.7696741740578204e-05, + "loss": 1.0839, + "step": 2246 + }, + { + "epoch": 1.2198697068403908, + "grad_norm": 7.985335036181301, + "learning_rate": 1.7694495984621097e-05, + "loss": 1.262, + "step": 2247 + }, + { + "epoch": 1.220412595005429, + "grad_norm": 8.31037377117703, + "learning_rate": 1.769224927699647e-05, + "loss": 1.0265, + "step": 2248 + }, + { + "epoch": 1.2209554831704668, + "grad_norm": 6.241752286447289, + "learning_rate": 1.769000161798219e-05, + "loss": 0.6179, + "step": 2249 + }, + { + "epoch": 1.221498371335505, + "grad_norm": 6.767956890107702, + "learning_rate": 1.7687753007856253e-05, + "loss": 0.9897, + "step": 2250 + }, + { + "epoch": 1.2220412595005428, + "grad_norm": 7.119621854687504, + "learning_rate": 1.7685503446896772e-05, + "loss": 1.0529, + "step": 2251 + }, + { + "epoch": 1.222584147665581, + "grad_norm": 6.37786350129091, + "learning_rate": 1.7683252935381976e-05, + "loss": 0.8575, + "step": 2252 + }, + { + "epoch": 1.2231270358306188, + "grad_norm": 9.280258739154695, + "learning_rate": 1.768100147359021e-05, + "loss": 0.959, + "step": 2253 + }, + { + "epoch": 1.223669923995657, + "grad_norm": 7.172898793975323, + "learning_rate": 1.7678749061799942e-05, + "loss": 1.3745, + "step": 2254 + }, + { + "epoch": 1.2242128121606948, + "grad_norm": 5.3297508420409825, + "learning_rate": 1.7676495700289753e-05, + "loss": 0.6697, + "step": 2255 + }, + { + "epoch": 1.224755700325733, + "grad_norm": 6.518884308754971, + "learning_rate": 1.767424138933834e-05, + "loss": 0.9854, + "step": 2256 + }, + { + "epoch": 1.2252985884907708, + "grad_norm": 7.939498291049385, + "learning_rate": 1.767198612922452e-05, + "loss": 0.8918, + "step": 2257 + }, + { + "epoch": 1.225841476655809, + "grad_norm": 6.309602610605377, + "learning_rate": 1.7669729920227226e-05, + "loss": 0.9939, + "step": 2258 + }, + { + "epoch": 1.2263843648208468, + "grad_norm": 7.462197154770194, + "learning_rate": 1.7667472762625518e-05, + "loss": 1.1299, + "step": 2259 + }, + { + "epoch": 1.226927252985885, + "grad_norm": 6.700031938849371, + "learning_rate": 1.7665214656698555e-05, + "loss": 0.8655, + "step": 2260 + }, + { + "epoch": 1.2274701411509228, + "grad_norm": 6.72124660471967, + "learning_rate": 1.7662955602725627e-05, + "loss": 0.7383, + "step": 2261 + }, + { + "epoch": 1.228013029315961, + "grad_norm": 6.071348581517431, + "learning_rate": 1.7660695600986137e-05, + "loss": 0.8571, + "step": 2262 + }, + { + "epoch": 1.2285559174809988, + "grad_norm": 5.914184858865564, + "learning_rate": 1.7658434651759605e-05, + "loss": 0.7494, + "step": 2263 + }, + { + "epoch": 1.229098805646037, + "grad_norm": 12.685614201985125, + "learning_rate": 1.7656172755325675e-05, + "loss": 1.608, + "step": 2264 + }, + { + "epoch": 1.2296416938110748, + "grad_norm": 7.001237249554534, + "learning_rate": 1.765390991196409e-05, + "loss": 1.0437, + "step": 2265 + }, + { + "epoch": 1.230184581976113, + "grad_norm": 8.232094967642231, + "learning_rate": 1.765164612195473e-05, + "loss": 0.8574, + "step": 2266 + }, + { + "epoch": 1.2307274701411508, + "grad_norm": 6.367083948330669, + "learning_rate": 1.7649381385577582e-05, + "loss": 0.7106, + "step": 2267 + }, + { + "epoch": 1.231270358306189, + "grad_norm": 6.2062819181758195, + "learning_rate": 1.7647115703112756e-05, + "loss": 0.6838, + "step": 2268 + }, + { + "epoch": 1.2318132464712268, + "grad_norm": 6.896096048797596, + "learning_rate": 1.7644849074840472e-05, + "loss": 0.9595, + "step": 2269 + }, + { + "epoch": 1.232356134636265, + "grad_norm": 5.915145487902616, + "learning_rate": 1.7642581501041067e-05, + "loss": 0.5741, + "step": 2270 + }, + { + "epoch": 1.2328990228013028, + "grad_norm": 5.966871119524657, + "learning_rate": 1.7640312981995004e-05, + "loss": 0.8686, + "step": 2271 + }, + { + "epoch": 1.233441910966341, + "grad_norm": 6.835680775363237, + "learning_rate": 1.7638043517982854e-05, + "loss": 0.9532, + "step": 2272 + }, + { + "epoch": 1.2339847991313788, + "grad_norm": 8.228029492129801, + "learning_rate": 1.7635773109285306e-05, + "loss": 1.0834, + "step": 2273 + }, + { + "epoch": 1.234527687296417, + "grad_norm": 8.272927259623488, + "learning_rate": 1.7633501756183174e-05, + "loss": 1.2591, + "step": 2274 + }, + { + "epoch": 1.2350705754614548, + "grad_norm": 6.755748928056399, + "learning_rate": 1.7631229458957377e-05, + "loss": 0.997, + "step": 2275 + }, + { + "epoch": 1.235613463626493, + "grad_norm": 7.335027842642818, + "learning_rate": 1.762895621788896e-05, + "loss": 0.8695, + "step": 2276 + }, + { + "epoch": 1.2361563517915308, + "grad_norm": 5.237971036796901, + "learning_rate": 1.7626682033259077e-05, + "loss": 0.6846, + "step": 2277 + }, + { + "epoch": 1.236699239956569, + "grad_norm": 6.333455514649705, + "learning_rate": 1.7624406905349004e-05, + "loss": 0.7532, + "step": 2278 + }, + { + "epoch": 1.2372421281216068, + "grad_norm": 6.255556726847127, + "learning_rate": 1.7622130834440136e-05, + "loss": 0.7801, + "step": 2279 + }, + { + "epoch": 1.237785016286645, + "grad_norm": 5.602832960919912, + "learning_rate": 1.761985382081398e-05, + "loss": 0.5105, + "step": 2280 + }, + { + "epoch": 1.2383279044516828, + "grad_norm": 5.814619917758832, + "learning_rate": 1.7617575864752155e-05, + "loss": 0.9034, + "step": 2281 + }, + { + "epoch": 1.238870792616721, + "grad_norm": 6.743402665368447, + "learning_rate": 1.7615296966536407e-05, + "loss": 0.7713, + "step": 2282 + }, + { + "epoch": 1.2394136807817588, + "grad_norm": 7.176469994889228, + "learning_rate": 1.7613017126448597e-05, + "loss": 0.957, + "step": 2283 + }, + { + "epoch": 1.239956568946797, + "grad_norm": 8.5427155590311, + "learning_rate": 1.7610736344770693e-05, + "loss": 0.8789, + "step": 2284 + }, + { + "epoch": 1.2404994571118348, + "grad_norm": 6.674205629131359, + "learning_rate": 1.7608454621784786e-05, + "loss": 1.0239, + "step": 2285 + }, + { + "epoch": 1.241042345276873, + "grad_norm": 7.687894188099568, + "learning_rate": 1.7606171957773088e-05, + "loss": 1.3164, + "step": 2286 + }, + { + "epoch": 1.2415852334419109, + "grad_norm": 6.331613648212601, + "learning_rate": 1.760388835301792e-05, + "loss": 0.8047, + "step": 2287 + }, + { + "epoch": 1.242128121606949, + "grad_norm": 7.0776252458357725, + "learning_rate": 1.760160380780172e-05, + "loss": 1.1893, + "step": 2288 + }, + { + "epoch": 1.2426710097719869, + "grad_norm": 5.434346705496786, + "learning_rate": 1.7599318322407047e-05, + "loss": 0.763, + "step": 2289 + }, + { + "epoch": 1.243213897937025, + "grad_norm": 6.0271106768536775, + "learning_rate": 1.759703189711657e-05, + "loss": 1.1129, + "step": 2290 + }, + { + "epoch": 1.2437567861020629, + "grad_norm": 7.286292908201965, + "learning_rate": 1.7594744532213088e-05, + "loss": 1.1136, + "step": 2291 + }, + { + "epoch": 1.244299674267101, + "grad_norm": 6.685086643433707, + "learning_rate": 1.7592456227979488e-05, + "loss": 0.724, + "step": 2292 + }, + { + "epoch": 1.2448425624321389, + "grad_norm": 7.184174802374185, + "learning_rate": 1.7590166984698807e-05, + "loss": 1.1471, + "step": 2293 + }, + { + "epoch": 1.245385450597177, + "grad_norm": 7.787563610771557, + "learning_rate": 1.7587876802654176e-05, + "loss": 1.01, + "step": 2294 + }, + { + "epoch": 1.2459283387622149, + "grad_norm": 7.036440767647583, + "learning_rate": 1.7585585682128848e-05, + "loss": 0.8612, + "step": 2295 + }, + { + "epoch": 1.246471226927253, + "grad_norm": 6.6372154433448705, + "learning_rate": 1.7583293623406192e-05, + "loss": 1.3394, + "step": 2296 + }, + { + "epoch": 1.2470141150922909, + "grad_norm": 7.707986168491473, + "learning_rate": 1.7581000626769697e-05, + "loss": 0.8688, + "step": 2297 + }, + { + "epoch": 1.247557003257329, + "grad_norm": 7.862359368033938, + "learning_rate": 1.757870669250296e-05, + "loss": 1.1999, + "step": 2298 + }, + { + "epoch": 1.2480998914223669, + "grad_norm": 7.942280078814929, + "learning_rate": 1.75764118208897e-05, + "loss": 1.241, + "step": 2299 + }, + { + "epoch": 1.248642779587405, + "grad_norm": 7.258825765708901, + "learning_rate": 1.7574116012213753e-05, + "loss": 0.8336, + "step": 2300 + }, + { + "epoch": 1.2491856677524429, + "grad_norm": 7.0149158036864065, + "learning_rate": 1.7571819266759066e-05, + "loss": 0.8637, + "step": 2301 + }, + { + "epoch": 1.249728555917481, + "grad_norm": 7.069351546856077, + "learning_rate": 1.7569521584809703e-05, + "loss": 0.8146, + "step": 2302 + }, + { + "epoch": 1.250271444082519, + "grad_norm": 4.6424312647112425, + "learning_rate": 1.756722296664985e-05, + "loss": 0.5056, + "step": 2303 + }, + { + "epoch": 1.2508143322475571, + "grad_norm": 7.857412677197525, + "learning_rate": 1.7564923412563797e-05, + "loss": 1.3072, + "step": 2304 + }, + { + "epoch": 1.251357220412595, + "grad_norm": 6.081845076053873, + "learning_rate": 1.7562622922835962e-05, + "loss": 1.0212, + "step": 2305 + }, + { + "epoch": 1.2519001085776331, + "grad_norm": 5.244697868381976, + "learning_rate": 1.7560321497750867e-05, + "loss": 0.9677, + "step": 2306 + }, + { + "epoch": 1.252442996742671, + "grad_norm": 11.027915321544773, + "learning_rate": 1.7558019137593162e-05, + "loss": 0.8984, + "step": 2307 + }, + { + "epoch": 1.2529858849077091, + "grad_norm": 5.955816485468965, + "learning_rate": 1.7555715842647603e-05, + "loss": 0.6282, + "step": 2308 + }, + { + "epoch": 1.253528773072747, + "grad_norm": 8.925991341226974, + "learning_rate": 1.755341161319907e-05, + "loss": 1.2121, + "step": 2309 + }, + { + "epoch": 1.2540716612377851, + "grad_norm": 7.947406679428915, + "learning_rate": 1.7551106449532547e-05, + "loss": 1.0545, + "step": 2310 + }, + { + "epoch": 1.254614549402823, + "grad_norm": 7.329950391674132, + "learning_rate": 1.7548800351933147e-05, + "loss": 1.0585, + "step": 2311 + }, + { + "epoch": 1.2551574375678611, + "grad_norm": 5.0931347466751955, + "learning_rate": 1.7546493320686084e-05, + "loss": 0.5533, + "step": 2312 + }, + { + "epoch": 1.255700325732899, + "grad_norm": 7.125279667319334, + "learning_rate": 1.7544185356076703e-05, + "loss": 1.1011, + "step": 2313 + }, + { + "epoch": 1.2562432138979371, + "grad_norm": 8.060697991979964, + "learning_rate": 1.7541876458390453e-05, + "loss": 0.8732, + "step": 2314 + }, + { + "epoch": 1.256786102062975, + "grad_norm": 9.732797197135078, + "learning_rate": 1.75395666279129e-05, + "loss": 1.5663, + "step": 2315 + }, + { + "epoch": 1.2573289902280131, + "grad_norm": 5.3673655570709435, + "learning_rate": 1.7537255864929733e-05, + "loss": 0.7057, + "step": 2316 + }, + { + "epoch": 1.257871878393051, + "grad_norm": 5.997561584131186, + "learning_rate": 1.7534944169726752e-05, + "loss": 0.8526, + "step": 2317 + }, + { + "epoch": 1.2584147665580891, + "grad_norm": 7.236766116786611, + "learning_rate": 1.753263154258986e-05, + "loss": 0.8921, + "step": 2318 + }, + { + "epoch": 1.258957654723127, + "grad_norm": 7.663917187467161, + "learning_rate": 1.7530317983805096e-05, + "loss": 0.9309, + "step": 2319 + }, + { + "epoch": 1.2595005428881652, + "grad_norm": 5.704193869412544, + "learning_rate": 1.75280034936586e-05, + "loss": 0.6291, + "step": 2320 + }, + { + "epoch": 1.260043431053203, + "grad_norm": 7.365455665485428, + "learning_rate": 1.7525688072436636e-05, + "loss": 0.9946, + "step": 2321 + }, + { + "epoch": 1.2605863192182412, + "grad_norm": 4.8267670121648205, + "learning_rate": 1.752337172042558e-05, + "loss": 0.8254, + "step": 2322 + }, + { + "epoch": 1.261129207383279, + "grad_norm": 7.231116660749922, + "learning_rate": 1.752105443791192e-05, + "loss": 1.3835, + "step": 2323 + }, + { + "epoch": 1.2616720955483172, + "grad_norm": 6.02849395913157, + "learning_rate": 1.7518736225182253e-05, + "loss": 0.6229, + "step": 2324 + }, + { + "epoch": 1.262214983713355, + "grad_norm": 6.458823304206971, + "learning_rate": 1.7516417082523314e-05, + "loss": 0.7707, + "step": 2325 + }, + { + "epoch": 1.2627578718783932, + "grad_norm": 7.163598438660781, + "learning_rate": 1.7514097010221927e-05, + "loss": 1.0007, + "step": 2326 + }, + { + "epoch": 1.263300760043431, + "grad_norm": 6.607394766457111, + "learning_rate": 1.7511776008565043e-05, + "loss": 1.0289, + "step": 2327 + }, + { + "epoch": 1.2638436482084692, + "grad_norm": 8.20658895752514, + "learning_rate": 1.7509454077839736e-05, + "loss": 0.7553, + "step": 2328 + }, + { + "epoch": 1.264386536373507, + "grad_norm": 8.963409997263136, + "learning_rate": 1.750713121833318e-05, + "loss": 0.9558, + "step": 2329 + }, + { + "epoch": 1.2649294245385452, + "grad_norm": 6.463632425962391, + "learning_rate": 1.7504807430332668e-05, + "loss": 0.6553, + "step": 2330 + }, + { + "epoch": 1.265472312703583, + "grad_norm": 6.936173778617146, + "learning_rate": 1.7502482714125612e-05, + "loss": 0.7861, + "step": 2331 + }, + { + "epoch": 1.2660152008686212, + "grad_norm": 6.975529559460763, + "learning_rate": 1.7500157069999536e-05, + "loss": 0.665, + "step": 2332 + }, + { + "epoch": 1.266558089033659, + "grad_norm": 6.768480922885786, + "learning_rate": 1.7497830498242082e-05, + "loss": 1.0163, + "step": 2333 + }, + { + "epoch": 1.2671009771986972, + "grad_norm": 6.357421024100016, + "learning_rate": 1.7495502999141004e-05, + "loss": 0.5897, + "step": 2334 + }, + { + "epoch": 1.267643865363735, + "grad_norm": 7.82539792030223, + "learning_rate": 1.7493174572984168e-05, + "loss": 1.0509, + "step": 2335 + }, + { + "epoch": 1.2681867535287732, + "grad_norm": 7.6523935004894215, + "learning_rate": 1.7490845220059554e-05, + "loss": 0.7603, + "step": 2336 + }, + { + "epoch": 1.268729641693811, + "grad_norm": 6.677360283026087, + "learning_rate": 1.7488514940655267e-05, + "loss": 1.1797, + "step": 2337 + }, + { + "epoch": 1.2692725298588492, + "grad_norm": 7.837606738097467, + "learning_rate": 1.7486183735059517e-05, + "loss": 1.1995, + "step": 2338 + }, + { + "epoch": 1.269815418023887, + "grad_norm": 8.913928805014676, + "learning_rate": 1.7483851603560634e-05, + "loss": 1.6007, + "step": 2339 + }, + { + "epoch": 1.2703583061889252, + "grad_norm": 8.467363918664116, + "learning_rate": 1.7481518546447058e-05, + "loss": 0.9515, + "step": 2340 + }, + { + "epoch": 1.270901194353963, + "grad_norm": 6.758882762498143, + "learning_rate": 1.747918456400734e-05, + "loss": 0.7935, + "step": 2341 + }, + { + "epoch": 1.2714440825190012, + "grad_norm": 5.455764070482765, + "learning_rate": 1.7476849656530158e-05, + "loss": 0.4623, + "step": 2342 + }, + { + "epoch": 1.271986970684039, + "grad_norm": 8.417334291043444, + "learning_rate": 1.7474513824304294e-05, + "loss": 1.1563, + "step": 2343 + }, + { + "epoch": 1.2725298588490772, + "grad_norm": 5.61084844590966, + "learning_rate": 1.7472177067618646e-05, + "loss": 0.8566, + "step": 2344 + }, + { + "epoch": 1.273072747014115, + "grad_norm": 7.726478476188633, + "learning_rate": 1.746983938676223e-05, + "loss": 1.0568, + "step": 2345 + }, + { + "epoch": 1.2736156351791532, + "grad_norm": 8.882013629978706, + "learning_rate": 1.7467500782024177e-05, + "loss": 1.0921, + "step": 2346 + }, + { + "epoch": 1.274158523344191, + "grad_norm": 6.976574380776494, + "learning_rate": 1.746516125369372e-05, + "loss": 1.1222, + "step": 2347 + }, + { + "epoch": 1.2747014115092292, + "grad_norm": 7.641347465211616, + "learning_rate": 1.7462820802060224e-05, + "loss": 1.0672, + "step": 2348 + }, + { + "epoch": 1.275244299674267, + "grad_norm": 6.473243122755322, + "learning_rate": 1.7460479427413156e-05, + "loss": 0.9019, + "step": 2349 + }, + { + "epoch": 1.2757871878393052, + "grad_norm": 6.494958002593054, + "learning_rate": 1.7458137130042105e-05, + "loss": 0.7862, + "step": 2350 + }, + { + "epoch": 1.276330076004343, + "grad_norm": 6.54459648554074, + "learning_rate": 1.7455793910236764e-05, + "loss": 0.9572, + "step": 2351 + }, + { + "epoch": 1.2768729641693812, + "grad_norm": 6.249218254368162, + "learning_rate": 1.7453449768286952e-05, + "loss": 0.9415, + "step": 2352 + }, + { + "epoch": 1.277415852334419, + "grad_norm": 6.335956299458767, + "learning_rate": 1.7451104704482592e-05, + "loss": 0.7647, + "step": 2353 + }, + { + "epoch": 1.2779587404994572, + "grad_norm": 6.296362619304004, + "learning_rate": 1.744875871911373e-05, + "loss": 0.6293, + "step": 2354 + }, + { + "epoch": 1.278501628664495, + "grad_norm": 6.629983181103369, + "learning_rate": 1.7446411812470516e-05, + "loss": 0.7816, + "step": 2355 + }, + { + "epoch": 1.2790445168295332, + "grad_norm": 6.921460108918273, + "learning_rate": 1.744406398484322e-05, + "loss": 0.6373, + "step": 2356 + }, + { + "epoch": 1.279587404994571, + "grad_norm": 9.4421099603369, + "learning_rate": 1.7441715236522228e-05, + "loss": 1.5291, + "step": 2357 + }, + { + "epoch": 1.2801302931596092, + "grad_norm": 7.838644313955947, + "learning_rate": 1.7439365567798035e-05, + "loss": 0.7934, + "step": 2358 + }, + { + "epoch": 1.280673181324647, + "grad_norm": 5.7659414273416, + "learning_rate": 1.743701497896125e-05, + "loss": 0.8358, + "step": 2359 + }, + { + "epoch": 1.2812160694896852, + "grad_norm": 4.731198780322981, + "learning_rate": 1.7434663470302602e-05, + "loss": 0.7357, + "step": 2360 + }, + { + "epoch": 1.281758957654723, + "grad_norm": 8.993227122901354, + "learning_rate": 1.7432311042112926e-05, + "loss": 1.0679, + "step": 2361 + }, + { + "epoch": 1.2823018458197613, + "grad_norm": 7.193753575449645, + "learning_rate": 1.7429957694683175e-05, + "loss": 0.9257, + "step": 2362 + }, + { + "epoch": 1.282844733984799, + "grad_norm": 7.1027186942584635, + "learning_rate": 1.7427603428304416e-05, + "loss": 0.64, + "step": 2363 + }, + { + "epoch": 1.2833876221498373, + "grad_norm": 9.441739719572269, + "learning_rate": 1.7425248243267824e-05, + "loss": 1.3236, + "step": 2364 + }, + { + "epoch": 1.283930510314875, + "grad_norm": 6.478965109994284, + "learning_rate": 1.7422892139864696e-05, + "loss": 0.7726, + "step": 2365 + }, + { + "epoch": 1.2844733984799133, + "grad_norm": 6.372951649068815, + "learning_rate": 1.742053511838644e-05, + "loss": 0.7785, + "step": 2366 + }, + { + "epoch": 1.285016286644951, + "grad_norm": 10.052356612127076, + "learning_rate": 1.7418177179124574e-05, + "loss": 1.7877, + "step": 2367 + }, + { + "epoch": 1.2855591748099893, + "grad_norm": 5.24633801429042, + "learning_rate": 1.741581832237073e-05, + "loss": 0.4195, + "step": 2368 + }, + { + "epoch": 1.286102062975027, + "grad_norm": 7.958362658884473, + "learning_rate": 1.7413458548416656e-05, + "loss": 1.1782, + "step": 2369 + }, + { + "epoch": 1.2866449511400653, + "grad_norm": 7.923031881378813, + "learning_rate": 1.7411097857554216e-05, + "loss": 1.0861, + "step": 2370 + }, + { + "epoch": 1.287187839305103, + "grad_norm": 6.53099588563486, + "learning_rate": 1.7408736250075378e-05, + "loss": 0.8717, + "step": 2371 + }, + { + "epoch": 1.2877307274701413, + "grad_norm": 7.364911313605863, + "learning_rate": 1.740637372627224e-05, + "loss": 0.9341, + "step": 2372 + }, + { + "epoch": 1.288273615635179, + "grad_norm": 8.083622859612968, + "learning_rate": 1.740401028643699e-05, + "loss": 1.1654, + "step": 2373 + }, + { + "epoch": 1.2888165038002173, + "grad_norm": 6.561315990238234, + "learning_rate": 1.7401645930861944e-05, + "loss": 0.8088, + "step": 2374 + }, + { + "epoch": 1.289359391965255, + "grad_norm": 6.926193447339149, + "learning_rate": 1.739928065983954e-05, + "loss": 0.8335, + "step": 2375 + }, + { + "epoch": 1.2899022801302933, + "grad_norm": 7.876730774519387, + "learning_rate": 1.7396914473662306e-05, + "loss": 0.9969, + "step": 2376 + }, + { + "epoch": 1.290445168295331, + "grad_norm": 6.513645878383447, + "learning_rate": 1.73945473726229e-05, + "loss": 0.9302, + "step": 2377 + }, + { + "epoch": 1.2909880564603693, + "grad_norm": 6.441562626276137, + "learning_rate": 1.73921793570141e-05, + "loss": 1.0318, + "step": 2378 + }, + { + "epoch": 1.291530944625407, + "grad_norm": 5.226443422910506, + "learning_rate": 1.7389810427128766e-05, + "loss": 0.6142, + "step": 2379 + }, + { + "epoch": 1.2920738327904453, + "grad_norm": 7.436131751002477, + "learning_rate": 1.7387440583259906e-05, + "loss": 0.9907, + "step": 2380 + }, + { + "epoch": 1.292616720955483, + "grad_norm": 6.335395136223753, + "learning_rate": 1.7385069825700615e-05, + "loss": 0.8431, + "step": 2381 + }, + { + "epoch": 1.2931596091205213, + "grad_norm": 8.93013417703781, + "learning_rate": 1.738269815474412e-05, + "loss": 1.3162, + "step": 2382 + }, + { + "epoch": 1.293702497285559, + "grad_norm": 6.616000665685319, + "learning_rate": 1.7380325570683754e-05, + "loss": 0.81, + "step": 2383 + }, + { + "epoch": 1.2942453854505973, + "grad_norm": 9.126678309827314, + "learning_rate": 1.737795207381296e-05, + "loss": 1.4976, + "step": 2384 + }, + { + "epoch": 1.294788273615635, + "grad_norm": 9.726273416685123, + "learning_rate": 1.737557766442529e-05, + "loss": 1.0284, + "step": 2385 + }, + { + "epoch": 1.2953311617806733, + "grad_norm": 6.207571191669611, + "learning_rate": 1.737320234281442e-05, + "loss": 0.9409, + "step": 2386 + }, + { + "epoch": 1.295874049945711, + "grad_norm": 6.150632184961992, + "learning_rate": 1.7370826109274136e-05, + "loss": 0.7558, + "step": 2387 + }, + { + "epoch": 1.2964169381107493, + "grad_norm": 6.555861715157727, + "learning_rate": 1.736844896409833e-05, + "loss": 1.4217, + "step": 2388 + }, + { + "epoch": 1.296959826275787, + "grad_norm": 7.291607947091272, + "learning_rate": 1.7366070907581015e-05, + "loss": 0.9969, + "step": 2389 + }, + { + "epoch": 1.2975027144408253, + "grad_norm": 7.003873752430508, + "learning_rate": 1.7363691940016307e-05, + "loss": 0.9569, + "step": 2390 + }, + { + "epoch": 1.298045602605863, + "grad_norm": 6.355975398274633, + "learning_rate": 1.7361312061698444e-05, + "loss": 0.9244, + "step": 2391 + }, + { + "epoch": 1.2985884907709013, + "grad_norm": 5.093818720596728, + "learning_rate": 1.7358931272921773e-05, + "loss": 0.5581, + "step": 2392 + }, + { + "epoch": 1.299131378935939, + "grad_norm": 4.532047411652702, + "learning_rate": 1.7356549573980753e-05, + "loss": 0.5153, + "step": 2393 + }, + { + "epoch": 1.2996742671009773, + "grad_norm": 7.378796483754393, + "learning_rate": 1.735416696516996e-05, + "loss": 0.9631, + "step": 2394 + }, + { + "epoch": 1.3002171552660151, + "grad_norm": 6.186232096125507, + "learning_rate": 1.7351783446784075e-05, + "loss": 0.7876, + "step": 2395 + }, + { + "epoch": 1.3007600434310533, + "grad_norm": 7.851387526494335, + "learning_rate": 1.7349399019117897e-05, + "loss": 0.8983, + "step": 2396 + }, + { + "epoch": 1.3013029315960911, + "grad_norm": 7.275506442499215, + "learning_rate": 1.7347013682466335e-05, + "loss": 0.8334, + "step": 2397 + }, + { + "epoch": 1.3018458197611293, + "grad_norm": 6.264072188984264, + "learning_rate": 1.7344627437124407e-05, + "loss": 0.6585, + "step": 2398 + }, + { + "epoch": 1.3023887079261671, + "grad_norm": 7.537073303434296, + "learning_rate": 1.734224028338726e-05, + "loss": 1.1033, + "step": 2399 + }, + { + "epoch": 1.3029315960912053, + "grad_norm": 6.850281208830108, + "learning_rate": 1.7339852221550126e-05, + "loss": 0.8133, + "step": 2400 + }, + { + "epoch": 1.3034744842562431, + "grad_norm": 8.089656198012262, + "learning_rate": 1.7337463251908374e-05, + "loss": 1.1342, + "step": 2401 + }, + { + "epoch": 1.3040173724212814, + "grad_norm": 6.2919884156176105, + "learning_rate": 1.733507337475747e-05, + "loss": 0.6533, + "step": 2402 + }, + { + "epoch": 1.3045602605863191, + "grad_norm": 5.114953151143873, + "learning_rate": 1.7332682590393008e-05, + "loss": 0.7451, + "step": 2403 + }, + { + "epoch": 1.3051031487513574, + "grad_norm": 6.18470588828327, + "learning_rate": 1.733029089911067e-05, + "loss": 1.1299, + "step": 2404 + }, + { + "epoch": 1.3056460369163951, + "grad_norm": 5.950129534147113, + "learning_rate": 1.7327898301206273e-05, + "loss": 0.7329, + "step": 2405 + }, + { + "epoch": 1.3061889250814334, + "grad_norm": 6.293661255114301, + "learning_rate": 1.7325504796975732e-05, + "loss": 0.7535, + "step": 2406 + }, + { + "epoch": 1.3067318132464711, + "grad_norm": 7.682262802579054, + "learning_rate": 1.732311038671509e-05, + "loss": 1.2399, + "step": 2407 + }, + { + "epoch": 1.3072747014115094, + "grad_norm": 7.314299514485556, + "learning_rate": 1.7320715070720478e-05, + "loss": 0.9298, + "step": 2408 + }, + { + "epoch": 1.3078175895765471, + "grad_norm": 8.921126356976833, + "learning_rate": 1.7318318849288158e-05, + "loss": 1.4254, + "step": 2409 + }, + { + "epoch": 1.3083604777415854, + "grad_norm": 7.060288459365394, + "learning_rate": 1.7315921722714503e-05, + "loss": 0.9428, + "step": 2410 + }, + { + "epoch": 1.3089033659066232, + "grad_norm": 8.568251970472561, + "learning_rate": 1.7313523691295988e-05, + "loss": 0.9011, + "step": 2411 + }, + { + "epoch": 1.3094462540716614, + "grad_norm": 6.049363733159396, + "learning_rate": 1.7311124755329206e-05, + "loss": 1.0718, + "step": 2412 + }, + { + "epoch": 1.3099891422366992, + "grad_norm": 8.287809242433424, + "learning_rate": 1.7308724915110864e-05, + "loss": 0.9137, + "step": 2413 + }, + { + "epoch": 1.3105320304017374, + "grad_norm": 8.436233795197621, + "learning_rate": 1.7306324170937774e-05, + "loss": 1.2427, + "step": 2414 + }, + { + "epoch": 1.3110749185667752, + "grad_norm": 7.035710404708695, + "learning_rate": 1.7303922523106863e-05, + "loss": 0.8206, + "step": 2415 + }, + { + "epoch": 1.3116178067318134, + "grad_norm": 5.9728457186472275, + "learning_rate": 1.730151997191518e-05, + "loss": 0.7307, + "step": 2416 + }, + { + "epoch": 1.3121606948968512, + "grad_norm": 6.296494925379207, + "learning_rate": 1.7299116517659865e-05, + "loss": 0.9808, + "step": 2417 + }, + { + "epoch": 1.3127035830618892, + "grad_norm": 7.672949954931739, + "learning_rate": 1.7296712160638192e-05, + "loss": 1.0289, + "step": 2418 + }, + { + "epoch": 1.3132464712269272, + "grad_norm": 5.958844838187463, + "learning_rate": 1.7294306901147525e-05, + "loss": 0.9657, + "step": 2419 + }, + { + "epoch": 1.3137893593919652, + "grad_norm": 6.5874736925476745, + "learning_rate": 1.7291900739485356e-05, + "loss": 0.8701, + "step": 2420 + }, + { + "epoch": 1.3143322475570032, + "grad_norm": 8.273467490572564, + "learning_rate": 1.7289493675949282e-05, + "loss": 0.9353, + "step": 2421 + }, + { + "epoch": 1.3148751357220412, + "grad_norm": 8.115143215098843, + "learning_rate": 1.7287085710837013e-05, + "loss": 0.6241, + "step": 2422 + }, + { + "epoch": 1.3154180238870792, + "grad_norm": 7.004703854660018, + "learning_rate": 1.7284676844446368e-05, + "loss": 0.801, + "step": 2423 + }, + { + "epoch": 1.3159609120521172, + "grad_norm": 6.039976470078682, + "learning_rate": 1.728226707707528e-05, + "loss": 1.085, + "step": 2424 + }, + { + "epoch": 1.3165038002171552, + "grad_norm": 5.318299059101813, + "learning_rate": 1.72798564090218e-05, + "loss": 0.4945, + "step": 2425 + }, + { + "epoch": 1.3170466883821932, + "grad_norm": 6.095578475774351, + "learning_rate": 1.727744484058407e-05, + "loss": 0.7696, + "step": 2426 + }, + { + "epoch": 1.3175895765472312, + "grad_norm": 6.4886080927006695, + "learning_rate": 1.7275032372060368e-05, + "loss": 0.5364, + "step": 2427 + }, + { + "epoch": 1.3181324647122692, + "grad_norm": 8.681669485105683, + "learning_rate": 1.7272619003749066e-05, + "loss": 0.9132, + "step": 2428 + }, + { + "epoch": 1.3186753528773072, + "grad_norm": 6.819144122945338, + "learning_rate": 1.7270204735948653e-05, + "loss": 0.7867, + "step": 2429 + }, + { + "epoch": 1.3192182410423452, + "grad_norm": 6.65338033222052, + "learning_rate": 1.7267789568957734e-05, + "loss": 0.7236, + "step": 2430 + }, + { + "epoch": 1.3197611292073832, + "grad_norm": 6.622862530496735, + "learning_rate": 1.7265373503075014e-05, + "loss": 0.9038, + "step": 2431 + }, + { + "epoch": 1.3203040173724212, + "grad_norm": 6.284216416307323, + "learning_rate": 1.7262956538599323e-05, + "loss": 0.677, + "step": 2432 + }, + { + "epoch": 1.3208469055374592, + "grad_norm": 9.150803536423062, + "learning_rate": 1.7260538675829593e-05, + "loss": 1.7953, + "step": 2433 + }, + { + "epoch": 1.3213897937024972, + "grad_norm": 9.236250071710021, + "learning_rate": 1.7258119915064867e-05, + "loss": 0.8568, + "step": 2434 + }, + { + "epoch": 1.3219326818675352, + "grad_norm": 7.061272817814159, + "learning_rate": 1.72557002566043e-05, + "loss": 0.9355, + "step": 2435 + }, + { + "epoch": 1.3224755700325732, + "grad_norm": 6.786187626307571, + "learning_rate": 1.7253279700747164e-05, + "loss": 0.7707, + "step": 2436 + }, + { + "epoch": 1.3230184581976112, + "grad_norm": 5.592273858747665, + "learning_rate": 1.725085824779283e-05, + "loss": 0.5932, + "step": 2437 + }, + { + "epoch": 1.3235613463626492, + "grad_norm": 8.137496643840409, + "learning_rate": 1.72484358980408e-05, + "loss": 1.0244, + "step": 2438 + }, + { + "epoch": 1.3241042345276872, + "grad_norm": 5.981122138792478, + "learning_rate": 1.724601265179066e-05, + "loss": 0.9807, + "step": 2439 + }, + { + "epoch": 1.3246471226927252, + "grad_norm": 6.912027532729551, + "learning_rate": 1.7243588509342127e-05, + "loss": 0.8673, + "step": 2440 + }, + { + "epoch": 1.3251900108577632, + "grad_norm": 5.836398731450734, + "learning_rate": 1.7241163470995024e-05, + "loss": 0.5738, + "step": 2441 + }, + { + "epoch": 1.3257328990228012, + "grad_norm": 5.0689326588185795, + "learning_rate": 1.7238737537049283e-05, + "loss": 0.4449, + "step": 2442 + }, + { + "epoch": 1.3262757871878392, + "grad_norm": 6.628337387569613, + "learning_rate": 1.7236310707804943e-05, + "loss": 0.7975, + "step": 2443 + }, + { + "epoch": 1.3268186753528772, + "grad_norm": 9.359654484357105, + "learning_rate": 1.7233882983562168e-05, + "loss": 0.8996, + "step": 2444 + }, + { + "epoch": 1.3273615635179152, + "grad_norm": 5.979606493030731, + "learning_rate": 1.723145436462121e-05, + "loss": 0.732, + "step": 2445 + }, + { + "epoch": 1.3279044516829532, + "grad_norm": 6.393853922089353, + "learning_rate": 1.7229024851282453e-05, + "loss": 0.9713, + "step": 2446 + }, + { + "epoch": 1.3284473398479912, + "grad_norm": 9.783386290725945, + "learning_rate": 1.722659444384638e-05, + "loss": 1.2991, + "step": 2447 + }, + { + "epoch": 1.3289902280130292, + "grad_norm": 8.471557152983001, + "learning_rate": 1.722416314261359e-05, + "loss": 0.8517, + "step": 2448 + }, + { + "epoch": 1.3295331161780672, + "grad_norm": 7.20996107804807, + "learning_rate": 1.7221730947884793e-05, + "loss": 1.0892, + "step": 2449 + }, + { + "epoch": 1.3300760043431052, + "grad_norm": 9.97079093211024, + "learning_rate": 1.7219297859960796e-05, + "loss": 0.8866, + "step": 2450 + }, + { + "epoch": 1.3306188925081432, + "grad_norm": 5.88899239355285, + "learning_rate": 1.7216863879142536e-05, + "loss": 0.7286, + "step": 2451 + }, + { + "epoch": 1.3311617806731812, + "grad_norm": 8.317968838525326, + "learning_rate": 1.7214429005731054e-05, + "loss": 0.9374, + "step": 2452 + }, + { + "epoch": 1.3317046688382193, + "grad_norm": 8.278980033154179, + "learning_rate": 1.721199324002749e-05, + "loss": 1.2977, + "step": 2453 + }, + { + "epoch": 1.3322475570032573, + "grad_norm": 7.50073001311901, + "learning_rate": 1.7209556582333106e-05, + "loss": 0.9479, + "step": 2454 + }, + { + "epoch": 1.3327904451682953, + "grad_norm": 9.34140520001984, + "learning_rate": 1.720711903294928e-05, + "loss": 1.1834, + "step": 2455 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 6.155163014555799, + "learning_rate": 1.720468059217748e-05, + "loss": 1.0674, + "step": 2456 + }, + { + "epoch": 1.3338762214983713, + "grad_norm": 8.207152122618293, + "learning_rate": 1.7202241260319305e-05, + "loss": 1.1623, + "step": 2457 + }, + { + "epoch": 1.3344191096634093, + "grad_norm": 8.477127177193351, + "learning_rate": 1.719980103767645e-05, + "loss": 1.0812, + "step": 2458 + }, + { + "epoch": 1.3349619978284473, + "grad_norm": 9.352009366934986, + "learning_rate": 1.7197359924550726e-05, + "loss": 0.9593, + "step": 2459 + }, + { + "epoch": 1.3355048859934853, + "grad_norm": 6.377220189803284, + "learning_rate": 1.719491792124406e-05, + "loss": 0.8569, + "step": 2460 + }, + { + "epoch": 1.3360477741585233, + "grad_norm": 10.372722210417326, + "learning_rate": 1.7192475028058475e-05, + "loss": 0.8778, + "step": 2461 + }, + { + "epoch": 1.3365906623235613, + "grad_norm": 6.943692981514271, + "learning_rate": 1.7190031245296118e-05, + "loss": 0.8143, + "step": 2462 + }, + { + "epoch": 1.3371335504885993, + "grad_norm": 8.53588364120296, + "learning_rate": 1.7187586573259237e-05, + "loss": 1.2608, + "step": 2463 + }, + { + "epoch": 1.3376764386536373, + "grad_norm": 6.85497949710551, + "learning_rate": 1.7185141012250195e-05, + "loss": 0.621, + "step": 2464 + }, + { + "epoch": 1.3382193268186753, + "grad_norm": 9.575890345813113, + "learning_rate": 1.7182694562571458e-05, + "loss": 1.5686, + "step": 2465 + }, + { + "epoch": 1.3387622149837133, + "grad_norm": 8.637844905126164, + "learning_rate": 1.718024722452561e-05, + "loss": 0.8287, + "step": 2466 + }, + { + "epoch": 1.3393051031487513, + "grad_norm": 7.464168837468054, + "learning_rate": 1.7177798998415344e-05, + "loss": 0.9837, + "step": 2467 + }, + { + "epoch": 1.3398479913137893, + "grad_norm": 9.27927006463804, + "learning_rate": 1.7175349884543458e-05, + "loss": 1.3479, + "step": 2468 + }, + { + "epoch": 1.3403908794788273, + "grad_norm": 6.9150180643503765, + "learning_rate": 1.717289988321286e-05, + "loss": 0.7731, + "step": 2469 + }, + { + "epoch": 1.3409337676438653, + "grad_norm": 8.264286643294954, + "learning_rate": 1.7170448994726574e-05, + "loss": 1.1252, + "step": 2470 + }, + { + "epoch": 1.3414766558089033, + "grad_norm": 6.1769106952218715, + "learning_rate": 1.7167997219387728e-05, + "loss": 0.6077, + "step": 2471 + }, + { + "epoch": 1.3420195439739413, + "grad_norm": 6.00046554758586, + "learning_rate": 1.7165544557499563e-05, + "loss": 0.9807, + "step": 2472 + }, + { + "epoch": 1.3425624321389793, + "grad_norm": 5.684026968717126, + "learning_rate": 1.7163091009365427e-05, + "loss": 0.8755, + "step": 2473 + }, + { + "epoch": 1.3431053203040173, + "grad_norm": 8.910694398982464, + "learning_rate": 1.7160636575288777e-05, + "loss": 1.1781, + "step": 2474 + }, + { + "epoch": 1.3436482084690553, + "grad_norm": 6.090059050660724, + "learning_rate": 1.7158181255573184e-05, + "loss": 0.8757, + "step": 2475 + }, + { + "epoch": 1.3441910966340933, + "grad_norm": 5.3693699493967575, + "learning_rate": 1.7155725050522325e-05, + "loss": 0.4111, + "step": 2476 + }, + { + "epoch": 1.3447339847991313, + "grad_norm": 6.4883376045754675, + "learning_rate": 1.7153267960439987e-05, + "loss": 1.1667, + "step": 2477 + }, + { + "epoch": 1.3452768729641693, + "grad_norm": 6.9649527868325585, + "learning_rate": 1.7150809985630065e-05, + "loss": 1.2642, + "step": 2478 + }, + { + "epoch": 1.3458197611292073, + "grad_norm": 7.458967124353787, + "learning_rate": 1.714835112639657e-05, + "loss": 1.1198, + "step": 2479 + }, + { + "epoch": 1.3463626492942453, + "grad_norm": 7.4820540779231015, + "learning_rate": 1.7145891383043613e-05, + "loss": 1.069, + "step": 2480 + }, + { + "epoch": 1.3469055374592833, + "grad_norm": 6.8442083181868165, + "learning_rate": 1.7143430755875422e-05, + "loss": 1.1187, + "step": 2481 + }, + { + "epoch": 1.3474484256243213, + "grad_norm": 8.389465670773259, + "learning_rate": 1.7140969245196332e-05, + "loss": 1.216, + "step": 2482 + }, + { + "epoch": 1.3479913137893593, + "grad_norm": 8.099481746156716, + "learning_rate": 1.713850685131078e-05, + "loss": 1.1365, + "step": 2483 + }, + { + "epoch": 1.3485342019543973, + "grad_norm": 7.005684437567773, + "learning_rate": 1.713604357452333e-05, + "loss": 0.7665, + "step": 2484 + }, + { + "epoch": 1.3490770901194353, + "grad_norm": 6.032904961736961, + "learning_rate": 1.7133579415138634e-05, + "loss": 0.7527, + "step": 2485 + }, + { + "epoch": 1.3496199782844733, + "grad_norm": 7.644359802639278, + "learning_rate": 1.7131114373461466e-05, + "loss": 0.7293, + "step": 2486 + }, + { + "epoch": 1.3501628664495113, + "grad_norm": 9.681466854129468, + "learning_rate": 1.7128648449796706e-05, + "loss": 1.0718, + "step": 2487 + }, + { + "epoch": 1.3507057546145493, + "grad_norm": 7.703820713192501, + "learning_rate": 1.7126181644449348e-05, + "loss": 1.0049, + "step": 2488 + }, + { + "epoch": 1.3512486427795873, + "grad_norm": 8.313943023532504, + "learning_rate": 1.7123713957724482e-05, + "loss": 1.1647, + "step": 2489 + }, + { + "epoch": 1.3517915309446253, + "grad_norm": 6.869318843676937, + "learning_rate": 1.7121245389927326e-05, + "loss": 0.7056, + "step": 2490 + }, + { + "epoch": 1.3523344191096633, + "grad_norm": 8.354366894281007, + "learning_rate": 1.7118775941363186e-05, + "loss": 1.4107, + "step": 2491 + }, + { + "epoch": 1.3528773072747013, + "grad_norm": 7.594996874862703, + "learning_rate": 1.7116305612337493e-05, + "loss": 1.3947, + "step": 2492 + }, + { + "epoch": 1.3534201954397393, + "grad_norm": 5.912640331169303, + "learning_rate": 1.7113834403155782e-05, + "loss": 0.678, + "step": 2493 + }, + { + "epoch": 1.3539630836047774, + "grad_norm": 8.603089873660597, + "learning_rate": 1.7111362314123693e-05, + "loss": 1.1846, + "step": 2494 + }, + { + "epoch": 1.3545059717698154, + "grad_norm": 5.899587064063632, + "learning_rate": 1.710888934554698e-05, + "loss": 0.7241, + "step": 2495 + }, + { + "epoch": 1.3550488599348534, + "grad_norm": 5.1366064982711315, + "learning_rate": 1.7106415497731502e-05, + "loss": 0.6714, + "step": 2496 + }, + { + "epoch": 1.3555917480998914, + "grad_norm": 6.55484035818926, + "learning_rate": 1.710394077098323e-05, + "loss": 1.0163, + "step": 2497 + }, + { + "epoch": 1.3561346362649294, + "grad_norm": 7.060748735507699, + "learning_rate": 1.710146516560824e-05, + "loss": 0.8407, + "step": 2498 + }, + { + "epoch": 1.3566775244299674, + "grad_norm": 7.488417378829399, + "learning_rate": 1.709898868191272e-05, + "loss": 1.0936, + "step": 2499 + }, + { + "epoch": 1.3572204125950054, + "grad_norm": 5.616560879457259, + "learning_rate": 1.7096511320202965e-05, + "loss": 0.6962, + "step": 2500 + }, + { + "epoch": 1.3577633007600434, + "grad_norm": 6.5457002772857775, + "learning_rate": 1.7094033080785384e-05, + "loss": 0.6338, + "step": 2501 + }, + { + "epoch": 1.3583061889250814, + "grad_norm": 5.788382651374094, + "learning_rate": 1.709155396396648e-05, + "loss": 0.9251, + "step": 2502 + }, + { + "epoch": 1.3588490770901194, + "grad_norm": 5.94278314769549, + "learning_rate": 1.7089073970052883e-05, + "loss": 0.5738, + "step": 2503 + }, + { + "epoch": 1.3593919652551574, + "grad_norm": 8.987615230603502, + "learning_rate": 1.7086593099351318e-05, + "loss": 1.1561, + "step": 2504 + }, + { + "epoch": 1.3599348534201954, + "grad_norm": 6.25509019167934, + "learning_rate": 1.7084111352168627e-05, + "loss": 0.5584, + "step": 2505 + }, + { + "epoch": 1.3604777415852334, + "grad_norm": 6.42626784615777, + "learning_rate": 1.708162872881175e-05, + "loss": 0.6385, + "step": 2506 + }, + { + "epoch": 1.3610206297502714, + "grad_norm": 7.342680004921898, + "learning_rate": 1.707914522958775e-05, + "loss": 1.0729, + "step": 2507 + }, + { + "epoch": 1.3615635179153094, + "grad_norm": 6.858864236878158, + "learning_rate": 1.707666085480378e-05, + "loss": 0.8641, + "step": 2508 + }, + { + "epoch": 1.3621064060803474, + "grad_norm": 6.176457712037886, + "learning_rate": 1.707417560476712e-05, + "loss": 0.9472, + "step": 2509 + }, + { + "epoch": 1.3626492942453854, + "grad_norm": 7.346954574146698, + "learning_rate": 1.7071689479785145e-05, + "loss": 1.0253, + "step": 2510 + }, + { + "epoch": 1.3631921824104234, + "grad_norm": 4.9331066104603085, + "learning_rate": 1.7069202480165344e-05, + "loss": 0.7307, + "step": 2511 + }, + { + "epoch": 1.3637350705754614, + "grad_norm": 8.316711829847868, + "learning_rate": 1.7066714606215316e-05, + "loss": 0.8078, + "step": 2512 + }, + { + "epoch": 1.3642779587404994, + "grad_norm": 8.079444940270628, + "learning_rate": 1.706422585824276e-05, + "loss": 1.2922, + "step": 2513 + }, + { + "epoch": 1.3648208469055374, + "grad_norm": 7.503741374806181, + "learning_rate": 1.7061736236555494e-05, + "loss": 0.7943, + "step": 2514 + }, + { + "epoch": 1.3653637350705754, + "grad_norm": 6.4023293864360635, + "learning_rate": 1.7059245741461435e-05, + "loss": 1.0111, + "step": 2515 + }, + { + "epoch": 1.3659066232356134, + "grad_norm": 5.7325920409358595, + "learning_rate": 1.705675437326861e-05, + "loss": 1.1201, + "step": 2516 + }, + { + "epoch": 1.3664495114006514, + "grad_norm": 8.452660578831315, + "learning_rate": 1.705426213228516e-05, + "loss": 0.8818, + "step": 2517 + }, + { + "epoch": 1.3669923995656894, + "grad_norm": 7.8301031574049755, + "learning_rate": 1.7051769018819327e-05, + "loss": 0.8803, + "step": 2518 + }, + { + "epoch": 1.3675352877307274, + "grad_norm": 10.77827779179992, + "learning_rate": 1.704927503317946e-05, + "loss": 1.6667, + "step": 2519 + }, + { + "epoch": 1.3680781758957654, + "grad_norm": 4.579120603442138, + "learning_rate": 1.7046780175674023e-05, + "loss": 0.6101, + "step": 2520 + }, + { + "epoch": 1.3686210640608034, + "grad_norm": 7.443733277623912, + "learning_rate": 1.7044284446611584e-05, + "loss": 0.8174, + "step": 2521 + }, + { + "epoch": 1.3691639522258414, + "grad_norm": 10.860855146774309, + "learning_rate": 1.7041787846300817e-05, + "loss": 1.2849, + "step": 2522 + }, + { + "epoch": 1.3697068403908794, + "grad_norm": 9.311283837684355, + "learning_rate": 1.7039290375050503e-05, + "loss": 1.4923, + "step": 2523 + }, + { + "epoch": 1.3702497285559174, + "grad_norm": 7.171226901795553, + "learning_rate": 1.703679203316954e-05, + "loss": 0.9493, + "step": 2524 + }, + { + "epoch": 1.3707926167209554, + "grad_norm": 8.658053683953712, + "learning_rate": 1.7034292820966923e-05, + "loss": 1.1366, + "step": 2525 + }, + { + "epoch": 1.3713355048859934, + "grad_norm": 6.400723985129287, + "learning_rate": 1.703179273875176e-05, + "loss": 1.0484, + "step": 2526 + }, + { + "epoch": 1.3718783930510314, + "grad_norm": 6.2215780206112274, + "learning_rate": 1.702929178683326e-05, + "loss": 1.1058, + "step": 2527 + }, + { + "epoch": 1.3724212812160694, + "grad_norm": 7.548758658264992, + "learning_rate": 1.7026789965520755e-05, + "loss": 1.094, + "step": 2528 + }, + { + "epoch": 1.3729641693811074, + "grad_norm": 7.250992484079327, + "learning_rate": 1.7024287275123664e-05, + "loss": 0.8424, + "step": 2529 + }, + { + "epoch": 1.3735070575461454, + "grad_norm": 6.664109303085317, + "learning_rate": 1.702178371595153e-05, + "loss": 1.0498, + "step": 2530 + }, + { + "epoch": 1.3740499457111834, + "grad_norm": 6.379180216401721, + "learning_rate": 1.7019279288313994e-05, + "loss": 0.9038, + "step": 2531 + }, + { + "epoch": 1.3745928338762214, + "grad_norm": 7.524913368583668, + "learning_rate": 1.7016773992520806e-05, + "loss": 0.8787, + "step": 2532 + }, + { + "epoch": 1.3751357220412594, + "grad_norm": 5.722778911929788, + "learning_rate": 1.701426782888183e-05, + "loss": 0.8199, + "step": 2533 + }, + { + "epoch": 1.3756786102062974, + "grad_norm": 6.26381521799836, + "learning_rate": 1.701176079770703e-05, + "loss": 0.6989, + "step": 2534 + }, + { + "epoch": 1.3762214983713354, + "grad_norm": 7.2300935887481765, + "learning_rate": 1.700925289930648e-05, + "loss": 1.094, + "step": 2535 + }, + { + "epoch": 1.3767643865363735, + "grad_norm": 7.385614391862118, + "learning_rate": 1.7006744133990358e-05, + "loss": 0.7753, + "step": 2536 + }, + { + "epoch": 1.3773072747014115, + "grad_norm": 5.52011787904346, + "learning_rate": 1.7004234502068952e-05, + "loss": 0.9232, + "step": 2537 + }, + { + "epoch": 1.3778501628664495, + "grad_norm": 6.523028453950784, + "learning_rate": 1.7001724003852668e-05, + "loss": 1.1181, + "step": 2538 + }, + { + "epoch": 1.3783930510314875, + "grad_norm": 6.334090281128193, + "learning_rate": 1.6999212639651995e-05, + "loss": 0.9384, + "step": 2539 + }, + { + "epoch": 1.3789359391965255, + "grad_norm": 7.3687722051890265, + "learning_rate": 1.6996700409777548e-05, + "loss": 0.9678, + "step": 2540 + }, + { + "epoch": 1.3794788273615635, + "grad_norm": 9.49113126498042, + "learning_rate": 1.6994187314540042e-05, + "loss": 0.7875, + "step": 2541 + }, + { + "epoch": 1.3800217155266015, + "grad_norm": 4.885613246464195, + "learning_rate": 1.69916733542503e-05, + "loss": 0.4488, + "step": 2542 + }, + { + "epoch": 1.3805646036916395, + "grad_norm": 7.362521236541499, + "learning_rate": 1.6989158529219262e-05, + "loss": 0.9808, + "step": 2543 + }, + { + "epoch": 1.3811074918566775, + "grad_norm": 8.6488909724118, + "learning_rate": 1.6986642839757953e-05, + "loss": 1.0599, + "step": 2544 + }, + { + "epoch": 1.3816503800217155, + "grad_norm": 6.054541159551389, + "learning_rate": 1.698412628617752e-05, + "loss": 0.8252, + "step": 2545 + }, + { + "epoch": 1.3821932681867535, + "grad_norm": 7.77360294326381, + "learning_rate": 1.698160886878922e-05, + "loss": 1.0718, + "step": 2546 + }, + { + "epoch": 1.3827361563517915, + "grad_norm": 9.493714306510151, + "learning_rate": 1.697909058790441e-05, + "loss": 1.2784, + "step": 2547 + }, + { + "epoch": 1.3832790445168295, + "grad_norm": 7.7307988823000455, + "learning_rate": 1.6976571443834555e-05, + "loss": 1.0308, + "step": 2548 + }, + { + "epoch": 1.3838219326818675, + "grad_norm": 7.149534093427325, + "learning_rate": 1.697405143689122e-05, + "loss": 0.9702, + "step": 2549 + }, + { + "epoch": 1.3843648208469055, + "grad_norm": 6.16619254042592, + "learning_rate": 1.6971530567386087e-05, + "loss": 0.6922, + "step": 2550 + }, + { + "epoch": 1.3849077090119435, + "grad_norm": 6.555255328428672, + "learning_rate": 1.6969008835630947e-05, + "loss": 0.8771, + "step": 2551 + }, + { + "epoch": 1.3854505971769815, + "grad_norm": 7.652524555180243, + "learning_rate": 1.6966486241937685e-05, + "loss": 1.0635, + "step": 2552 + }, + { + "epoch": 1.3859934853420195, + "grad_norm": 9.675328775672968, + "learning_rate": 1.69639627866183e-05, + "loss": 1.1616, + "step": 2553 + }, + { + "epoch": 1.3865363735070575, + "grad_norm": 6.5667057730439256, + "learning_rate": 1.69614384699849e-05, + "loss": 0.7181, + "step": 2554 + }, + { + "epoch": 1.3870792616720955, + "grad_norm": 7.374476144001139, + "learning_rate": 1.6958913292349698e-05, + "loss": 1.123, + "step": 2555 + }, + { + "epoch": 1.3876221498371335, + "grad_norm": 8.873033944832025, + "learning_rate": 1.6956387254025005e-05, + "loss": 1.2651, + "step": 2556 + }, + { + "epoch": 1.3881650380021715, + "grad_norm": 6.877978468812787, + "learning_rate": 1.695386035532325e-05, + "loss": 1.0039, + "step": 2557 + }, + { + "epoch": 1.3887079261672095, + "grad_norm": 4.746919583421645, + "learning_rate": 1.6951332596556966e-05, + "loss": 0.7222, + "step": 2558 + }, + { + "epoch": 1.3892508143322475, + "grad_norm": 5.538693020619686, + "learning_rate": 1.6948803978038787e-05, + "loss": 0.8463, + "step": 2559 + }, + { + "epoch": 1.3897937024972855, + "grad_norm": 6.636448326582328, + "learning_rate": 1.6946274500081455e-05, + "loss": 0.7594, + "step": 2560 + }, + { + "epoch": 1.3903365906623235, + "grad_norm": 8.74658695496399, + "learning_rate": 1.6943744162997825e-05, + "loss": 1.224, + "step": 2561 + }, + { + "epoch": 1.3908794788273615, + "grad_norm": 7.305739264312674, + "learning_rate": 1.694121296710085e-05, + "loss": 1.0755, + "step": 2562 + }, + { + "epoch": 1.3914223669923995, + "grad_norm": 4.6475508332537165, + "learning_rate": 1.693868091270359e-05, + "loss": 0.6127, + "step": 2563 + }, + { + "epoch": 1.3919652551574375, + "grad_norm": 8.484116710765257, + "learning_rate": 1.6936148000119218e-05, + "loss": 1.029, + "step": 2564 + }, + { + "epoch": 1.3925081433224755, + "grad_norm": 6.532276109161297, + "learning_rate": 1.6933614229661008e-05, + "loss": 1.0517, + "step": 2565 + }, + { + "epoch": 1.3930510314875135, + "grad_norm": 5.43723324019315, + "learning_rate": 1.6931079601642336e-05, + "loss": 0.6915, + "step": 2566 + }, + { + "epoch": 1.3935939196525515, + "grad_norm": 7.562306350008547, + "learning_rate": 1.69285441163767e-05, + "loss": 0.5726, + "step": 2567 + }, + { + "epoch": 1.3941368078175895, + "grad_norm": 8.500602954507755, + "learning_rate": 1.6926007774177678e-05, + "loss": 1.0297, + "step": 2568 + }, + { + "epoch": 1.3946796959826275, + "grad_norm": 6.432151204429844, + "learning_rate": 1.6923470575358977e-05, + "loss": 1.0016, + "step": 2569 + }, + { + "epoch": 1.3952225841476655, + "grad_norm": 9.552257208599151, + "learning_rate": 1.6920932520234402e-05, + "loss": 1.0814, + "step": 2570 + }, + { + "epoch": 1.3957654723127035, + "grad_norm": 8.075918082294125, + "learning_rate": 1.6918393609117863e-05, + "loss": 1.1699, + "step": 2571 + }, + { + "epoch": 1.3963083604777415, + "grad_norm": 6.296577538864887, + "learning_rate": 1.6915853842323373e-05, + "loss": 1.0211, + "step": 2572 + }, + { + "epoch": 1.3968512486427795, + "grad_norm": 7.785873884786678, + "learning_rate": 1.691331322016506e-05, + "loss": 0.9225, + "step": 2573 + }, + { + "epoch": 1.3973941368078175, + "grad_norm": 6.768130802749829, + "learning_rate": 1.691077174295715e-05, + "loss": 0.8092, + "step": 2574 + }, + { + "epoch": 1.3979370249728555, + "grad_norm": 5.566346022945414, + "learning_rate": 1.6908229411013977e-05, + "loss": 0.8312, + "step": 2575 + }, + { + "epoch": 1.3984799131378935, + "grad_norm": 7.2860384645139815, + "learning_rate": 1.6905686224649978e-05, + "loss": 0.9357, + "step": 2576 + }, + { + "epoch": 1.3990228013029316, + "grad_norm": 7.072468614016554, + "learning_rate": 1.6903142184179704e-05, + "loss": 0.7362, + "step": 2577 + }, + { + "epoch": 1.3995656894679696, + "grad_norm": 6.421225685489366, + "learning_rate": 1.6900597289917803e-05, + "loss": 0.6828, + "step": 2578 + }, + { + "epoch": 1.4001085776330076, + "grad_norm": 8.167489592155093, + "learning_rate": 1.689805154217903e-05, + "loss": 1.1237, + "step": 2579 + }, + { + "epoch": 1.4006514657980456, + "grad_norm": 11.676833397587332, + "learning_rate": 1.6895504941278246e-05, + "loss": 0.9532, + "step": 2580 + }, + { + "epoch": 1.4011943539630836, + "grad_norm": 6.04293054204543, + "learning_rate": 1.689295748753042e-05, + "loss": 0.6541, + "step": 2581 + }, + { + "epoch": 1.4017372421281216, + "grad_norm": 8.432023560966908, + "learning_rate": 1.6890409181250632e-05, + "loss": 1.0167, + "step": 2582 + }, + { + "epoch": 1.4022801302931596, + "grad_norm": 7.494983940960417, + "learning_rate": 1.688786002275405e-05, + "loss": 0.8913, + "step": 2583 + }, + { + "epoch": 1.4028230184581976, + "grad_norm": 6.475912046078606, + "learning_rate": 1.6885310012355964e-05, + "loss": 0.8599, + "step": 2584 + }, + { + "epoch": 1.4033659066232356, + "grad_norm": 5.145369710998311, + "learning_rate": 1.6882759150371765e-05, + "loss": 0.8079, + "step": 2585 + }, + { + "epoch": 1.4039087947882736, + "grad_norm": 6.1855027756804155, + "learning_rate": 1.688020743711694e-05, + "loss": 0.8046, + "step": 2586 + }, + { + "epoch": 1.4044516829533116, + "grad_norm": 8.409570707749053, + "learning_rate": 1.68776548729071e-05, + "loss": 0.75, + "step": 2587 + }, + { + "epoch": 1.4049945711183496, + "grad_norm": 6.150666283054011, + "learning_rate": 1.687510145805794e-05, + "loss": 0.8339, + "step": 2588 + }, + { + "epoch": 1.4055374592833876, + "grad_norm": 5.4082205122977625, + "learning_rate": 1.6872547192885272e-05, + "loss": 0.7878, + "step": 2589 + }, + { + "epoch": 1.4060803474484256, + "grad_norm": 7.011727228130495, + "learning_rate": 1.686999207770502e-05, + "loss": 0.6415, + "step": 2590 + }, + { + "epoch": 1.4066232356134636, + "grad_norm": 7.384524419917869, + "learning_rate": 1.6867436112833193e-05, + "loss": 0.8489, + "step": 2591 + }, + { + "epoch": 1.4071661237785016, + "grad_norm": 6.846659366372164, + "learning_rate": 1.6864879298585925e-05, + "loss": 1.128, + "step": 2592 + }, + { + "epoch": 1.4077090119435396, + "grad_norm": 6.142948047083128, + "learning_rate": 1.6862321635279444e-05, + "loss": 0.7886, + "step": 2593 + }, + { + "epoch": 1.4082519001085776, + "grad_norm": 10.23557881641297, + "learning_rate": 1.6859763123230086e-05, + "loss": 0.8323, + "step": 2594 + }, + { + "epoch": 1.4087947882736156, + "grad_norm": 6.995135696669204, + "learning_rate": 1.6857203762754294e-05, + "loss": 1.0283, + "step": 2595 + }, + { + "epoch": 1.4093376764386536, + "grad_norm": 7.709014920487576, + "learning_rate": 1.685464355416861e-05, + "loss": 0.8586, + "step": 2596 + }, + { + "epoch": 1.4098805646036916, + "grad_norm": 8.153794053960306, + "learning_rate": 1.6852082497789684e-05, + "loss": 1.1797, + "step": 2597 + }, + { + "epoch": 1.4104234527687296, + "grad_norm": 7.721286185320884, + "learning_rate": 1.684952059393428e-05, + "loss": 0.9196, + "step": 2598 + }, + { + "epoch": 1.4109663409337676, + "grad_norm": 8.024349951964256, + "learning_rate": 1.684695784291925e-05, + "loss": 0.7769, + "step": 2599 + }, + { + "epoch": 1.4115092290988056, + "grad_norm": 8.135395328882872, + "learning_rate": 1.684439424506156e-05, + "loss": 1.1642, + "step": 2600 + }, + { + "epoch": 1.4120521172638436, + "grad_norm": 6.057801451822921, + "learning_rate": 1.684182980067828e-05, + "loss": 0.7352, + "step": 2601 + }, + { + "epoch": 1.4125950054288816, + "grad_norm": 7.808755828672067, + "learning_rate": 1.683926451008659e-05, + "loss": 0.9788, + "step": 2602 + }, + { + "epoch": 1.4131378935939196, + "grad_norm": 7.426450515083944, + "learning_rate": 1.6836698373603765e-05, + "loss": 0.7696, + "step": 2603 + }, + { + "epoch": 1.4136807817589576, + "grad_norm": 9.382792907740685, + "learning_rate": 1.6834131391547187e-05, + "loss": 1.0872, + "step": 2604 + }, + { + "epoch": 1.4142236699239956, + "grad_norm": 6.171700145892427, + "learning_rate": 1.6831563564234347e-05, + "loss": 0.7233, + "step": 2605 + }, + { + "epoch": 1.4147665580890336, + "grad_norm": 11.270930151197227, + "learning_rate": 1.682899489198284e-05, + "loss": 0.9739, + "step": 2606 + }, + { + "epoch": 1.4153094462540716, + "grad_norm": 9.050544540368543, + "learning_rate": 1.6826425375110357e-05, + "loss": 0.9436, + "step": 2607 + }, + { + "epoch": 1.4158523344191096, + "grad_norm": 6.969930349987723, + "learning_rate": 1.6823855013934705e-05, + "loss": 0.7632, + "step": 2608 + }, + { + "epoch": 1.4163952225841476, + "grad_norm": 5.791480821515424, + "learning_rate": 1.682128380877379e-05, + "loss": 0.8202, + "step": 2609 + }, + { + "epoch": 1.4169381107491856, + "grad_norm": 9.905757175497309, + "learning_rate": 1.6818711759945623e-05, + "loss": 1.2161, + "step": 2610 + }, + { + "epoch": 1.4174809989142236, + "grad_norm": 6.082684402096193, + "learning_rate": 1.6816138867768318e-05, + "loss": 0.7245, + "step": 2611 + }, + { + "epoch": 1.4180238870792616, + "grad_norm": 10.15499139340414, + "learning_rate": 1.6813565132560092e-05, + "loss": 1.1514, + "step": 2612 + }, + { + "epoch": 1.4185667752442996, + "grad_norm": 6.978513243855861, + "learning_rate": 1.6810990554639276e-05, + "loss": 0.5331, + "step": 2613 + }, + { + "epoch": 1.4191096634093376, + "grad_norm": 8.037739371135425, + "learning_rate": 1.6808415134324288e-05, + "loss": 0.7089, + "step": 2614 + }, + { + "epoch": 1.4196525515743756, + "grad_norm": 6.925791024229076, + "learning_rate": 1.6805838871933664e-05, + "loss": 0.8981, + "step": 2615 + }, + { + "epoch": 1.4201954397394136, + "grad_norm": 6.352165521713976, + "learning_rate": 1.6803261767786048e-05, + "loss": 0.5644, + "step": 2616 + }, + { + "epoch": 1.4207383279044516, + "grad_norm": 10.267497627922179, + "learning_rate": 1.680068382220017e-05, + "loss": 0.9607, + "step": 2617 + }, + { + "epoch": 1.4212812160694897, + "grad_norm": 7.696468630335302, + "learning_rate": 1.679810503549488e-05, + "loss": 0.729, + "step": 2618 + }, + { + "epoch": 1.4218241042345277, + "grad_norm": 6.868981955235214, + "learning_rate": 1.679552540798912e-05, + "loss": 0.9152, + "step": 2619 + }, + { + "epoch": 1.4223669923995657, + "grad_norm": 8.909790381587262, + "learning_rate": 1.6792944940001952e-05, + "loss": 0.9335, + "step": 2620 + }, + { + "epoch": 1.4229098805646037, + "grad_norm": 9.62415635421133, + "learning_rate": 1.6790363631852524e-05, + "loss": 1.1012, + "step": 2621 + }, + { + "epoch": 1.4234527687296417, + "grad_norm": 7.307385111665146, + "learning_rate": 1.67877814838601e-05, + "loss": 0.6963, + "step": 2622 + }, + { + "epoch": 1.4239956568946797, + "grad_norm": 9.018642559605246, + "learning_rate": 1.678519849634405e-05, + "loss": 1.0612, + "step": 2623 + }, + { + "epoch": 1.4245385450597177, + "grad_norm": 6.929244675286548, + "learning_rate": 1.6782614669623827e-05, + "loss": 0.9462, + "step": 2624 + }, + { + "epoch": 1.4250814332247557, + "grad_norm": 8.577419731292203, + "learning_rate": 1.6780030004019016e-05, + "loss": 1.112, + "step": 2625 + }, + { + "epoch": 1.4256243213897937, + "grad_norm": 7.125694710248634, + "learning_rate": 1.677744449984929e-05, + "loss": 0.851, + "step": 2626 + }, + { + "epoch": 1.4261672095548317, + "grad_norm": 9.488635353984538, + "learning_rate": 1.6774858157434425e-05, + "loss": 1.3064, + "step": 2627 + }, + { + "epoch": 1.4267100977198697, + "grad_norm": 7.500971938576216, + "learning_rate": 1.6772270977094307e-05, + "loss": 1.3675, + "step": 2628 + }, + { + "epoch": 1.4272529858849077, + "grad_norm": 6.979855088071626, + "learning_rate": 1.676968295914892e-05, + "loss": 0.8896, + "step": 2629 + }, + { + "epoch": 1.4277958740499457, + "grad_norm": 7.795843207695432, + "learning_rate": 1.6767094103918357e-05, + "loss": 0.7758, + "step": 2630 + }, + { + "epoch": 1.4283387622149837, + "grad_norm": 6.131884468994572, + "learning_rate": 1.6764504411722806e-05, + "loss": 0.7744, + "step": 2631 + }, + { + "epoch": 1.4288816503800217, + "grad_norm": 10.33693974618067, + "learning_rate": 1.676191388288257e-05, + "loss": 1.5364, + "step": 2632 + }, + { + "epoch": 1.4294245385450597, + "grad_norm": 10.005426546410897, + "learning_rate": 1.6759322517718048e-05, + "loss": 0.9713, + "step": 2633 + }, + { + "epoch": 1.4299674267100977, + "grad_norm": 9.480236119835816, + "learning_rate": 1.6756730316549745e-05, + "loss": 1.2573, + "step": 2634 + }, + { + "epoch": 1.4305103148751357, + "grad_norm": 10.874101086577593, + "learning_rate": 1.675413727969827e-05, + "loss": 1.0308, + "step": 2635 + }, + { + "epoch": 1.4310532030401737, + "grad_norm": 5.34984499382639, + "learning_rate": 1.675154340748433e-05, + "loss": 0.6013, + "step": 2636 + }, + { + "epoch": 1.4315960912052117, + "grad_norm": 7.464386702227026, + "learning_rate": 1.674894870022874e-05, + "loss": 0.9918, + "step": 2637 + }, + { + "epoch": 1.4321389793702497, + "grad_norm": 8.280480153626241, + "learning_rate": 1.674635315825242e-05, + "loss": 0.8745, + "step": 2638 + }, + { + "epoch": 1.4326818675352877, + "grad_norm": 6.169575731939732, + "learning_rate": 1.6743756781876385e-05, + "loss": 0.6984, + "step": 2639 + }, + { + "epoch": 1.4332247557003257, + "grad_norm": 5.390920772379431, + "learning_rate": 1.6741159571421768e-05, + "loss": 0.8886, + "step": 2640 + }, + { + "epoch": 1.4337676438653637, + "grad_norm": 7.003336357340687, + "learning_rate": 1.6738561527209792e-05, + "loss": 1.0335, + "step": 2641 + }, + { + "epoch": 1.4343105320304017, + "grad_norm": 7.722166856340297, + "learning_rate": 1.6735962649561784e-05, + "loss": 0.9804, + "step": 2642 + }, + { + "epoch": 1.4348534201954397, + "grad_norm": 7.328242752900802, + "learning_rate": 1.673336293879918e-05, + "loss": 0.8675, + "step": 2643 + }, + { + "epoch": 1.4353963083604777, + "grad_norm": 7.344577484351048, + "learning_rate": 1.6730762395243515e-05, + "loss": 1.2518, + "step": 2644 + }, + { + "epoch": 1.4359391965255157, + "grad_norm": 7.560893408391802, + "learning_rate": 1.6728161019216433e-05, + "loss": 1.0196, + "step": 2645 + }, + { + "epoch": 1.4364820846905537, + "grad_norm": 7.509821569775923, + "learning_rate": 1.6725558811039674e-05, + "loss": 1.1112, + "step": 2646 + }, + { + "epoch": 1.4370249728555917, + "grad_norm": 7.672412677248201, + "learning_rate": 1.672295577103508e-05, + "loss": 0.9075, + "step": 2647 + }, + { + "epoch": 1.4375678610206297, + "grad_norm": 8.562389123203614, + "learning_rate": 1.67203518995246e-05, + "loss": 1.2123, + "step": 2648 + }, + { + "epoch": 1.4381107491856677, + "grad_norm": 8.172008773566091, + "learning_rate": 1.671774719683029e-05, + "loss": 0.8064, + "step": 2649 + }, + { + "epoch": 1.4386536373507057, + "grad_norm": 7.5145061670971405, + "learning_rate": 1.6715141663274297e-05, + "loss": 1.3513, + "step": 2650 + }, + { + "epoch": 1.4391965255157437, + "grad_norm": 5.675981880534199, + "learning_rate": 1.6712535299178883e-05, + "loss": 0.5029, + "step": 2651 + }, + { + "epoch": 1.4397394136807817, + "grad_norm": 8.195158801834715, + "learning_rate": 1.6709928104866403e-05, + "loss": 0.9414, + "step": 2652 + }, + { + "epoch": 1.4402823018458197, + "grad_norm": 7.374510239535514, + "learning_rate": 1.6707320080659322e-05, + "loss": 0.8103, + "step": 2653 + }, + { + "epoch": 1.4408251900108577, + "grad_norm": 5.703232221762084, + "learning_rate": 1.6704711226880204e-05, + "loss": 1.1646, + "step": 2654 + }, + { + "epoch": 1.4413680781758957, + "grad_norm": 6.981749095787225, + "learning_rate": 1.6702101543851714e-05, + "loss": 0.7174, + "step": 2655 + }, + { + "epoch": 1.4419109663409337, + "grad_norm": 6.456866833246808, + "learning_rate": 1.6699491031896625e-05, + "loss": 0.9795, + "step": 2656 + }, + { + "epoch": 1.4424538545059717, + "grad_norm": 10.193228844116039, + "learning_rate": 1.6696879691337807e-05, + "loss": 1.1288, + "step": 2657 + }, + { + "epoch": 1.4429967426710097, + "grad_norm": 9.920534268729181, + "learning_rate": 1.6694267522498237e-05, + "loss": 0.8379, + "step": 2658 + }, + { + "epoch": 1.4435396308360477, + "grad_norm": 6.153776792682404, + "learning_rate": 1.669165452570099e-05, + "loss": 0.8818, + "step": 2659 + }, + { + "epoch": 1.4440825190010858, + "grad_norm": 5.650871346064418, + "learning_rate": 1.6689040701269245e-05, + "loss": 0.732, + "step": 2660 + }, + { + "epoch": 1.4446254071661238, + "grad_norm": 7.829649689315363, + "learning_rate": 1.668642604952629e-05, + "loss": 0.8119, + "step": 2661 + }, + { + "epoch": 1.4451682953311618, + "grad_norm": 5.959350361834891, + "learning_rate": 1.6683810570795498e-05, + "loss": 0.6829, + "step": 2662 + }, + { + "epoch": 1.4457111834961998, + "grad_norm": 8.76687690585514, + "learning_rate": 1.6681194265400365e-05, + "loss": 1.4888, + "step": 2663 + }, + { + "epoch": 1.4462540716612378, + "grad_norm": 6.129763847493397, + "learning_rate": 1.6678577133664476e-05, + "loss": 0.8372, + "step": 2664 + }, + { + "epoch": 1.4467969598262758, + "grad_norm": 6.529235674005995, + "learning_rate": 1.6675959175911527e-05, + "loss": 0.9655, + "step": 2665 + }, + { + "epoch": 1.4473398479913138, + "grad_norm": 7.741547370588863, + "learning_rate": 1.6673340392465304e-05, + "loss": 1.0772, + "step": 2666 + }, + { + "epoch": 1.4478827361563518, + "grad_norm": 6.674042204787818, + "learning_rate": 1.6670720783649706e-05, + "loss": 0.9681, + "step": 2667 + }, + { + "epoch": 1.4484256243213898, + "grad_norm": 7.920578917847642, + "learning_rate": 1.666810034978873e-05, + "loss": 1.1068, + "step": 2668 + }, + { + "epoch": 1.4489685124864278, + "grad_norm": 7.249023283604718, + "learning_rate": 1.6665479091206476e-05, + "loss": 1.1025, + "step": 2669 + }, + { + "epoch": 1.4495114006514658, + "grad_norm": 6.407638163624444, + "learning_rate": 1.6662857008227145e-05, + "loss": 0.8236, + "step": 2670 + }, + { + "epoch": 1.4500542888165038, + "grad_norm": 7.289336395940614, + "learning_rate": 1.6660234101175036e-05, + "loss": 0.9386, + "step": 2671 + }, + { + "epoch": 1.4505971769815418, + "grad_norm": 6.973959594036107, + "learning_rate": 1.665761037037456e-05, + "loss": 0.8435, + "step": 2672 + }, + { + "epoch": 1.4511400651465798, + "grad_norm": 6.971928597160326, + "learning_rate": 1.665498581615023e-05, + "loss": 0.9065, + "step": 2673 + }, + { + "epoch": 1.4516829533116178, + "grad_norm": 6.1608850694063735, + "learning_rate": 1.665236043882664e-05, + "loss": 0.9289, + "step": 2674 + }, + { + "epoch": 1.4522258414766558, + "grad_norm": 6.62873967155635, + "learning_rate": 1.6649734238728512e-05, + "loss": 0.7797, + "step": 2675 + }, + { + "epoch": 1.4527687296416938, + "grad_norm": 7.089586063125632, + "learning_rate": 1.6647107216180655e-05, + "loss": 0.6306, + "step": 2676 + }, + { + "epoch": 1.4533116178067318, + "grad_norm": 10.986990416571043, + "learning_rate": 1.6644479371507985e-05, + "loss": 0.8764, + "step": 2677 + }, + { + "epoch": 1.4538545059717698, + "grad_norm": 5.865852703001031, + "learning_rate": 1.664185070503551e-05, + "loss": 0.9106, + "step": 2678 + }, + { + "epoch": 1.4543973941368078, + "grad_norm": 7.303323841843, + "learning_rate": 1.663922121708836e-05, + "loss": 0.8188, + "step": 2679 + }, + { + "epoch": 1.4549402823018458, + "grad_norm": 5.715280365147904, + "learning_rate": 1.663659090799175e-05, + "loss": 1.0573, + "step": 2680 + }, + { + "epoch": 1.4554831704668838, + "grad_norm": 7.084313061900368, + "learning_rate": 1.6633959778070992e-05, + "loss": 1.0749, + "step": 2681 + }, + { + "epoch": 1.4560260586319218, + "grad_norm": 7.902311581251679, + "learning_rate": 1.6631327827651524e-05, + "loss": 0.8677, + "step": 2682 + }, + { + "epoch": 1.4565689467969598, + "grad_norm": 8.1594384957164, + "learning_rate": 1.6628695057058855e-05, + "loss": 0.9999, + "step": 2683 + }, + { + "epoch": 1.4571118349619978, + "grad_norm": 5.580141751242927, + "learning_rate": 1.6626061466618623e-05, + "loss": 0.6387, + "step": 2684 + }, + { + "epoch": 1.4576547231270358, + "grad_norm": 8.145754034329466, + "learning_rate": 1.6623427056656544e-05, + "loss": 0.8984, + "step": 2685 + }, + { + "epoch": 1.4581976112920738, + "grad_norm": 6.453637820501138, + "learning_rate": 1.6620791827498454e-05, + "loss": 0.6922, + "step": 2686 + }, + { + "epoch": 1.4587404994571118, + "grad_norm": 7.3511834666776075, + "learning_rate": 1.6618155779470275e-05, + "loss": 0.8462, + "step": 2687 + }, + { + "epoch": 1.4592833876221498, + "grad_norm": 7.962874909840381, + "learning_rate": 1.6615518912898043e-05, + "loss": 0.9639, + "step": 2688 + }, + { + "epoch": 1.4598262757871878, + "grad_norm": 6.975602073560851, + "learning_rate": 1.6612881228107886e-05, + "loss": 0.9364, + "step": 2689 + }, + { + "epoch": 1.4603691639522258, + "grad_norm": 6.916712660881488, + "learning_rate": 1.6610242725426044e-05, + "loss": 0.8033, + "step": 2690 + }, + { + "epoch": 1.4609120521172638, + "grad_norm": 6.268208953592866, + "learning_rate": 1.6607603405178842e-05, + "loss": 0.7542, + "step": 2691 + }, + { + "epoch": 1.4614549402823018, + "grad_norm": 9.047138051843492, + "learning_rate": 1.660496326769272e-05, + "loss": 1.5956, + "step": 2692 + }, + { + "epoch": 1.4619978284473398, + "grad_norm": 7.228556514923675, + "learning_rate": 1.6602322313294216e-05, + "loss": 0.8461, + "step": 2693 + }, + { + "epoch": 1.4625407166123778, + "grad_norm": 6.195288408674038, + "learning_rate": 1.659968054230997e-05, + "loss": 0.9851, + "step": 2694 + }, + { + "epoch": 1.4630836047774158, + "grad_norm": 7.413116367868534, + "learning_rate": 1.6597037955066713e-05, + "loss": 0.9693, + "step": 2695 + }, + { + "epoch": 1.4636264929424538, + "grad_norm": 7.014031390984585, + "learning_rate": 1.6594394551891288e-05, + "loss": 1.0539, + "step": 2696 + }, + { + "epoch": 1.4641693811074918, + "grad_norm": 6.8196013668965225, + "learning_rate": 1.6591750333110634e-05, + "loss": 0.712, + "step": 2697 + }, + { + "epoch": 1.4647122692725298, + "grad_norm": 8.160690900976675, + "learning_rate": 1.658910529905179e-05, + "loss": 0.6551, + "step": 2698 + }, + { + "epoch": 1.4652551574375678, + "grad_norm": 5.174705288640052, + "learning_rate": 1.6586459450041906e-05, + "loss": 0.8285, + "step": 2699 + }, + { + "epoch": 1.4657980456026058, + "grad_norm": 7.029482771999071, + "learning_rate": 1.6583812786408216e-05, + "loss": 0.923, + "step": 2700 + }, + { + "epoch": 1.4663409337676439, + "grad_norm": 7.8877107275923946, + "learning_rate": 1.658116530847807e-05, + "loss": 1.1915, + "step": 2701 + }, + { + "epoch": 1.4668838219326819, + "grad_norm": 9.269305925663183, + "learning_rate": 1.657851701657891e-05, + "loss": 1.098, + "step": 2702 + }, + { + "epoch": 1.4674267100977199, + "grad_norm": 10.478546362437942, + "learning_rate": 1.657586791103828e-05, + "loss": 1.2416, + "step": 2703 + }, + { + "epoch": 1.4679695982627579, + "grad_norm": 8.207786471433751, + "learning_rate": 1.6573217992183826e-05, + "loss": 0.8658, + "step": 2704 + }, + { + "epoch": 1.4685124864277959, + "grad_norm": 7.862118410069405, + "learning_rate": 1.6570567260343294e-05, + "loss": 0.8714, + "step": 2705 + }, + { + "epoch": 1.4690553745928339, + "grad_norm": 7.744957075186954, + "learning_rate": 1.6567915715844534e-05, + "loss": 0.9374, + "step": 2706 + }, + { + "epoch": 1.4695982627578719, + "grad_norm": 6.97424297060321, + "learning_rate": 1.6565263359015488e-05, + "loss": 0.7183, + "step": 2707 + }, + { + "epoch": 1.4701411509229099, + "grad_norm": 7.349900880491086, + "learning_rate": 1.6562610190184206e-05, + "loss": 0.6066, + "step": 2708 + }, + { + "epoch": 1.4706840390879479, + "grad_norm": 6.982564531951699, + "learning_rate": 1.655995620967884e-05, + "loss": 1.2218, + "step": 2709 + }, + { + "epoch": 1.4712269272529859, + "grad_norm": 7.572476514419708, + "learning_rate": 1.6557301417827632e-05, + "loss": 0.89, + "step": 2710 + }, + { + "epoch": 1.4717698154180239, + "grad_norm": 8.654751087453503, + "learning_rate": 1.6554645814958932e-05, + "loss": 0.7188, + "step": 2711 + }, + { + "epoch": 1.4723127035830619, + "grad_norm": 6.282711052010393, + "learning_rate": 1.6551989401401196e-05, + "loss": 0.7586, + "step": 2712 + }, + { + "epoch": 1.4728555917480999, + "grad_norm": 8.732262593936317, + "learning_rate": 1.6549332177482966e-05, + "loss": 1.0178, + "step": 2713 + }, + { + "epoch": 1.4733984799131379, + "grad_norm": 6.989470691687101, + "learning_rate": 1.6546674143532895e-05, + "loss": 0.6981, + "step": 2714 + }, + { + "epoch": 1.4739413680781759, + "grad_norm": 7.179372740296565, + "learning_rate": 1.6544015299879734e-05, + "loss": 0.7617, + "step": 2715 + }, + { + "epoch": 1.4744842562432139, + "grad_norm": 9.117294639273645, + "learning_rate": 1.6541355646852327e-05, + "loss": 0.8822, + "step": 2716 + }, + { + "epoch": 1.475027144408252, + "grad_norm": 7.875727317127435, + "learning_rate": 1.653869518477963e-05, + "loss": 0.9149, + "step": 2717 + }, + { + "epoch": 1.47557003257329, + "grad_norm": 7.926910105501724, + "learning_rate": 1.6536033913990687e-05, + "loss": 0.8797, + "step": 2718 + }, + { + "epoch": 1.476112920738328, + "grad_norm": 5.14878192977047, + "learning_rate": 1.6533371834814657e-05, + "loss": 0.4491, + "step": 2719 + }, + { + "epoch": 1.476655808903366, + "grad_norm": 5.327596982546826, + "learning_rate": 1.6530708947580785e-05, + "loss": 0.4335, + "step": 2720 + }, + { + "epoch": 1.477198697068404, + "grad_norm": 9.158336314582998, + "learning_rate": 1.6528045252618423e-05, + "loss": 1.1208, + "step": 2721 + }, + { + "epoch": 1.477741585233442, + "grad_norm": 10.630652314970348, + "learning_rate": 1.6525380750257022e-05, + "loss": 1.097, + "step": 2722 + }, + { + "epoch": 1.47828447339848, + "grad_norm": 8.97220182564547, + "learning_rate": 1.652271544082613e-05, + "loss": 0.7838, + "step": 2723 + }, + { + "epoch": 1.478827361563518, + "grad_norm": 10.609251261516084, + "learning_rate": 1.652004932465539e-05, + "loss": 0.8861, + "step": 2724 + }, + { + "epoch": 1.479370249728556, + "grad_norm": 7.303801635012744, + "learning_rate": 1.6517382402074563e-05, + "loss": 0.5437, + "step": 2725 + }, + { + "epoch": 1.479913137893594, + "grad_norm": 11.770584674671793, + "learning_rate": 1.651471467341349e-05, + "loss": 1.4947, + "step": 2726 + }, + { + "epoch": 1.480456026058632, + "grad_norm": 6.824299484966605, + "learning_rate": 1.6512046139002128e-05, + "loss": 0.7144, + "step": 2727 + }, + { + "epoch": 1.48099891422367, + "grad_norm": 7.595224046166877, + "learning_rate": 1.650937679917052e-05, + "loss": 1.0259, + "step": 2728 + }, + { + "epoch": 1.481541802388708, + "grad_norm": 8.979875171863954, + "learning_rate": 1.6506706654248813e-05, + "loss": 1.0951, + "step": 2729 + }, + { + "epoch": 1.482084690553746, + "grad_norm": 7.700209331252355, + "learning_rate": 1.650403570456726e-05, + "loss": 1.0228, + "step": 2730 + }, + { + "epoch": 1.482627578718784, + "grad_norm": 9.090566362046857, + "learning_rate": 1.65013639504562e-05, + "loss": 1.0199, + "step": 2731 + }, + { + "epoch": 1.483170466883822, + "grad_norm": 6.644791741009562, + "learning_rate": 1.6498691392246088e-05, + "loss": 0.5478, + "step": 2732 + }, + { + "epoch": 1.48371335504886, + "grad_norm": 7.067718277939746, + "learning_rate": 1.6496018030267467e-05, + "loss": 0.864, + "step": 2733 + }, + { + "epoch": 1.484256243213898, + "grad_norm": 8.017737461642762, + "learning_rate": 1.6493343864850984e-05, + "loss": 1.1688, + "step": 2734 + }, + { + "epoch": 1.484799131378936, + "grad_norm": 6.479930584435578, + "learning_rate": 1.6490668896327382e-05, + "loss": 0.7227, + "step": 2735 + }, + { + "epoch": 1.485342019543974, + "grad_norm": 8.392289588009671, + "learning_rate": 1.6487993125027504e-05, + "loss": 0.7213, + "step": 2736 + }, + { + "epoch": 1.485884907709012, + "grad_norm": 7.817891248616158, + "learning_rate": 1.64853165512823e-05, + "loss": 0.7521, + "step": 2737 + }, + { + "epoch": 1.48642779587405, + "grad_norm": 8.824885297850575, + "learning_rate": 1.6482639175422804e-05, + "loss": 1.016, + "step": 2738 + }, + { + "epoch": 1.486970684039088, + "grad_norm": 11.48655873033237, + "learning_rate": 1.6479960997780165e-05, + "loss": 0.9814, + "step": 2739 + }, + { + "epoch": 1.487513572204126, + "grad_norm": 10.88221555712033, + "learning_rate": 1.6477282018685628e-05, + "loss": 1.464, + "step": 2740 + }, + { + "epoch": 1.488056460369164, + "grad_norm": 8.701247869952772, + "learning_rate": 1.6474602238470524e-05, + "loss": 0.918, + "step": 2741 + }, + { + "epoch": 1.488599348534202, + "grad_norm": 6.4327269926453505, + "learning_rate": 1.6471921657466294e-05, + "loss": 0.7949, + "step": 2742 + }, + { + "epoch": 1.48914223669924, + "grad_norm": 7.592315420975657, + "learning_rate": 1.6469240276004477e-05, + "loss": 0.9485, + "step": 2743 + }, + { + "epoch": 1.489685124864278, + "grad_norm": 7.682698616258038, + "learning_rate": 1.6466558094416717e-05, + "loss": 0.8439, + "step": 2744 + }, + { + "epoch": 1.490228013029316, + "grad_norm": 6.127626205001148, + "learning_rate": 1.6463875113034743e-05, + "loss": 0.8652, + "step": 2745 + }, + { + "epoch": 1.490770901194354, + "grad_norm": 9.65248236696942, + "learning_rate": 1.6461191332190397e-05, + "loss": 1.6278, + "step": 2746 + }, + { + "epoch": 1.491313789359392, + "grad_norm": 8.46309710838681, + "learning_rate": 1.6458506752215603e-05, + "loss": 1.5697, + "step": 2747 + }, + { + "epoch": 1.49185667752443, + "grad_norm": 6.578253971678976, + "learning_rate": 1.6455821373442407e-05, + "loss": 0.8836, + "step": 2748 + }, + { + "epoch": 1.492399565689468, + "grad_norm": 5.927853369454149, + "learning_rate": 1.645313519620293e-05, + "loss": 0.7182, + "step": 2749 + }, + { + "epoch": 1.492942453854506, + "grad_norm": 7.725891960915116, + "learning_rate": 1.645044822082941e-05, + "loss": 1.1794, + "step": 2750 + }, + { + "epoch": 1.493485342019544, + "grad_norm": 8.975220195142995, + "learning_rate": 1.644776044765417e-05, + "loss": 0.8909, + "step": 2751 + }, + { + "epoch": 1.494028230184582, + "grad_norm": 7.34643792387778, + "learning_rate": 1.6445071877009643e-05, + "loss": 1.0047, + "step": 2752 + }, + { + "epoch": 1.49457111834962, + "grad_norm": 7.561064198594243, + "learning_rate": 1.6442382509228355e-05, + "loss": 0.7533, + "step": 2753 + }, + { + "epoch": 1.495114006514658, + "grad_norm": 9.327085390245353, + "learning_rate": 1.6439692344642933e-05, + "loss": 0.8048, + "step": 2754 + }, + { + "epoch": 1.495656894679696, + "grad_norm": 8.662401438640156, + "learning_rate": 1.6437001383586095e-05, + "loss": 1.2513, + "step": 2755 + }, + { + "epoch": 1.496199782844734, + "grad_norm": 7.683989985812991, + "learning_rate": 1.6434309626390667e-05, + "loss": 0.8005, + "step": 2756 + }, + { + "epoch": 1.496742671009772, + "grad_norm": 6.295413547214335, + "learning_rate": 1.6431617073389574e-05, + "loss": 0.8778, + "step": 2757 + }, + { + "epoch": 1.49728555917481, + "grad_norm": 10.390849259618786, + "learning_rate": 1.6428923724915825e-05, + "loss": 1.6234, + "step": 2758 + }, + { + "epoch": 1.497828447339848, + "grad_norm": 6.815996692584369, + "learning_rate": 1.6426229581302545e-05, + "loss": 0.8142, + "step": 2759 + }, + { + "epoch": 1.498371335504886, + "grad_norm": 7.203388917066917, + "learning_rate": 1.642353464288295e-05, + "loss": 1.1273, + "step": 2760 + }, + { + "epoch": 1.498914223669924, + "grad_norm": 6.45381693158517, + "learning_rate": 1.6420838909990356e-05, + "loss": 0.7989, + "step": 2761 + }, + { + "epoch": 1.499457111834962, + "grad_norm": 7.135095705156932, + "learning_rate": 1.6418142382958167e-05, + "loss": 0.6753, + "step": 2762 + }, + { + "epoch": 1.5, + "grad_norm": 8.263302638252904, + "learning_rate": 1.64154450621199e-05, + "loss": 1.0357, + "step": 2763 + }, + { + "epoch": 1.500542888165038, + "grad_norm": 8.19537793615059, + "learning_rate": 1.6412746947809165e-05, + "loss": 0.8135, + "step": 2764 + }, + { + "epoch": 1.501085776330076, + "grad_norm": 5.084112201229359, + "learning_rate": 1.6410048040359665e-05, + "loss": 0.3466, + "step": 2765 + }, + { + "epoch": 1.501628664495114, + "grad_norm": 6.621661501175659, + "learning_rate": 1.6407348340105208e-05, + "loss": 1.1468, + "step": 2766 + }, + { + "epoch": 1.502171552660152, + "grad_norm": 7.5339066223692495, + "learning_rate": 1.6404647847379696e-05, + "loss": 0.8808, + "step": 2767 + }, + { + "epoch": 1.50271444082519, + "grad_norm": 6.501647846893153, + "learning_rate": 1.6401946562517134e-05, + "loss": 0.7589, + "step": 2768 + }, + { + "epoch": 1.503257328990228, + "grad_norm": 8.29356168583173, + "learning_rate": 1.6399244485851614e-05, + "loss": 1.3969, + "step": 2769 + }, + { + "epoch": 1.503800217155266, + "grad_norm": 6.518146966456392, + "learning_rate": 1.6396541617717337e-05, + "loss": 0.7951, + "step": 2770 + }, + { + "epoch": 1.504343105320304, + "grad_norm": 6.51134736256057, + "learning_rate": 1.63938379584486e-05, + "loss": 0.9073, + "step": 2771 + }, + { + "epoch": 1.504885993485342, + "grad_norm": 6.682459044312825, + "learning_rate": 1.6391133508379797e-05, + "loss": 0.733, + "step": 2772 + }, + { + "epoch": 1.50542888165038, + "grad_norm": 6.390687464891663, + "learning_rate": 1.638842826784541e-05, + "loss": 0.6934, + "step": 2773 + }, + { + "epoch": 1.505971769815418, + "grad_norm": 8.130197298030478, + "learning_rate": 1.6385722237180038e-05, + "loss": 0.8567, + "step": 2774 + }, + { + "epoch": 1.506514657980456, + "grad_norm": 7.959636836831885, + "learning_rate": 1.6383015416718356e-05, + "loss": 0.9926, + "step": 2775 + }, + { + "epoch": 1.507057546145494, + "grad_norm": 7.732366564125716, + "learning_rate": 1.638030780679516e-05, + "loss": 1.2228, + "step": 2776 + }, + { + "epoch": 1.507600434310532, + "grad_norm": 4.633651796670195, + "learning_rate": 1.6377599407745324e-05, + "loss": 0.4633, + "step": 2777 + }, + { + "epoch": 1.50814332247557, + "grad_norm": 7.016748827947926, + "learning_rate": 1.6374890219903828e-05, + "loss": 0.9593, + "step": 2778 + }, + { + "epoch": 1.508686210640608, + "grad_norm": 8.428067394567327, + "learning_rate": 1.637218024360575e-05, + "loss": 1.0493, + "step": 2779 + }, + { + "epoch": 1.509229098805646, + "grad_norm": 8.297567044395613, + "learning_rate": 1.6369469479186266e-05, + "loss": 1.2335, + "step": 2780 + }, + { + "epoch": 1.509771986970684, + "grad_norm": 5.8755950768313205, + "learning_rate": 1.6366757926980643e-05, + "loss": 0.6556, + "step": 2781 + }, + { + "epoch": 1.510314875135722, + "grad_norm": 6.837923725489964, + "learning_rate": 1.6364045587324254e-05, + "loss": 0.7031, + "step": 2782 + }, + { + "epoch": 1.51085776330076, + "grad_norm": 6.718777411432131, + "learning_rate": 1.6361332460552565e-05, + "loss": 0.8982, + "step": 2783 + }, + { + "epoch": 1.511400651465798, + "grad_norm": 8.788667184000067, + "learning_rate": 1.6358618547001137e-05, + "loss": 0.8548, + "step": 2784 + }, + { + "epoch": 1.511943539630836, + "grad_norm": 6.559897969720367, + "learning_rate": 1.635590384700563e-05, + "loss": 0.9627, + "step": 2785 + }, + { + "epoch": 1.512486427795874, + "grad_norm": 6.538968751920423, + "learning_rate": 1.635318836090181e-05, + "loss": 1.0731, + "step": 2786 + }, + { + "epoch": 1.513029315960912, + "grad_norm": 8.149646454407613, + "learning_rate": 1.6350472089025523e-05, + "loss": 0.9898, + "step": 2787 + }, + { + "epoch": 1.51357220412595, + "grad_norm": 8.544281989077668, + "learning_rate": 1.6347755031712734e-05, + "loss": 1.1125, + "step": 2788 + }, + { + "epoch": 1.514115092290988, + "grad_norm": 6.930736082842491, + "learning_rate": 1.634503718929948e-05, + "loss": 0.7617, + "step": 2789 + }, + { + "epoch": 1.514657980456026, + "grad_norm": 8.527931500643358, + "learning_rate": 1.6342318562121916e-05, + "loss": 1.2179, + "step": 2790 + }, + { + "epoch": 1.515200868621064, + "grad_norm": 7.509610271006853, + "learning_rate": 1.6339599150516283e-05, + "loss": 0.908, + "step": 2791 + }, + { + "epoch": 1.515743756786102, + "grad_norm": 6.0902820346277045, + "learning_rate": 1.6336878954818926e-05, + "loss": 0.829, + "step": 2792 + }, + { + "epoch": 1.51628664495114, + "grad_norm": 7.3274444131512935, + "learning_rate": 1.6334157975366278e-05, + "loss": 0.9667, + "step": 2793 + }, + { + "epoch": 1.516829533116178, + "grad_norm": 7.753641450951333, + "learning_rate": 1.633143621249488e-05, + "loss": 0.7961, + "step": 2794 + }, + { + "epoch": 1.517372421281216, + "grad_norm": 14.786266283108317, + "learning_rate": 1.6328713666541357e-05, + "loss": 1.4119, + "step": 2795 + }, + { + "epoch": 1.517915309446254, + "grad_norm": 6.1388168060810875, + "learning_rate": 1.632599033784244e-05, + "loss": 0.7726, + "step": 2796 + }, + { + "epoch": 1.518458197611292, + "grad_norm": 9.134190897524304, + "learning_rate": 1.632326622673496e-05, + "loss": 1.208, + "step": 2797 + }, + { + "epoch": 1.51900108577633, + "grad_norm": 6.898277731717493, + "learning_rate": 1.632054133355583e-05, + "loss": 0.6929, + "step": 2798 + }, + { + "epoch": 1.519543973941368, + "grad_norm": 6.90513711625334, + "learning_rate": 1.631781565864208e-05, + "loss": 0.7279, + "step": 2799 + }, + { + "epoch": 1.520086862106406, + "grad_norm": 8.332173251791465, + "learning_rate": 1.6315089202330817e-05, + "loss": 0.987, + "step": 2800 + }, + { + "epoch": 1.520629750271444, + "grad_norm": 6.783078863499027, + "learning_rate": 1.631236196495926e-05, + "loss": 0.6793, + "step": 2801 + }, + { + "epoch": 1.521172638436482, + "grad_norm": 7.841204229725534, + "learning_rate": 1.6309633946864712e-05, + "loss": 0.832, + "step": 2802 + }, + { + "epoch": 1.52171552660152, + "grad_norm": 6.4931904342190405, + "learning_rate": 1.630690514838458e-05, + "loss": 0.6662, + "step": 2803 + }, + { + "epoch": 1.522258414766558, + "grad_norm": 8.50187086749661, + "learning_rate": 1.6304175569856368e-05, + "loss": 0.8359, + "step": 2804 + }, + { + "epoch": 1.522801302931596, + "grad_norm": 6.196174007418195, + "learning_rate": 1.6301445211617676e-05, + "loss": 0.8437, + "step": 2805 + }, + { + "epoch": 1.523344191096634, + "grad_norm": 8.232047070633076, + "learning_rate": 1.6298714074006196e-05, + "loss": 0.9134, + "step": 2806 + }, + { + "epoch": 1.523887079261672, + "grad_norm": 6.594428796101092, + "learning_rate": 1.629598215735972e-05, + "loss": 0.6703, + "step": 2807 + }, + { + "epoch": 1.52442996742671, + "grad_norm": 7.634546429927066, + "learning_rate": 1.629324946201614e-05, + "loss": 0.8208, + "step": 2808 + }, + { + "epoch": 1.524972855591748, + "grad_norm": 8.225245989940067, + "learning_rate": 1.6290515988313432e-05, + "loss": 1.0002, + "step": 2809 + }, + { + "epoch": 1.5255157437567861, + "grad_norm": 8.334512060802671, + "learning_rate": 1.628778173658968e-05, + "loss": 0.8883, + "step": 2810 + }, + { + "epoch": 1.5260586319218241, + "grad_norm": 7.733244019009635, + "learning_rate": 1.6285046707183068e-05, + "loss": 0.7155, + "step": 2811 + }, + { + "epoch": 1.5266015200868621, + "grad_norm": 9.607081900344873, + "learning_rate": 1.628231090043186e-05, + "loss": 0.8321, + "step": 2812 + }, + { + "epoch": 1.5271444082519001, + "grad_norm": 8.390046905361096, + "learning_rate": 1.6279574316674426e-05, + "loss": 0.8818, + "step": 2813 + }, + { + "epoch": 1.5276872964169381, + "grad_norm": 7.608191285481891, + "learning_rate": 1.6276836956249235e-05, + "loss": 0.7382, + "step": 2814 + }, + { + "epoch": 1.5282301845819761, + "grad_norm": 8.998269112751723, + "learning_rate": 1.6274098819494844e-05, + "loss": 0.9285, + "step": 2815 + }, + { + "epoch": 1.5287730727470141, + "grad_norm": 5.877268783539307, + "learning_rate": 1.627135990674991e-05, + "loss": 0.8434, + "step": 2816 + }, + { + "epoch": 1.5293159609120521, + "grad_norm": 6.956730702284698, + "learning_rate": 1.6268620218353188e-05, + "loss": 0.5463, + "step": 2817 + }, + { + "epoch": 1.5298588490770901, + "grad_norm": 6.516430093047641, + "learning_rate": 1.626587975464353e-05, + "loss": 0.4886, + "step": 2818 + }, + { + "epoch": 1.5304017372421281, + "grad_norm": 6.751333342823053, + "learning_rate": 1.626313851595987e-05, + "loss": 0.6279, + "step": 2819 + }, + { + "epoch": 1.5309446254071661, + "grad_norm": 8.120808986494415, + "learning_rate": 1.6260396502641264e-05, + "loss": 0.8192, + "step": 2820 + }, + { + "epoch": 1.5314875135722041, + "grad_norm": 7.570228095805488, + "learning_rate": 1.6257653715026837e-05, + "loss": 0.5813, + "step": 2821 + }, + { + "epoch": 1.5320304017372421, + "grad_norm": 8.436330907807918, + "learning_rate": 1.625491015345583e-05, + "loss": 0.9115, + "step": 2822 + }, + { + "epoch": 1.5325732899022801, + "grad_norm": 8.384048301051982, + "learning_rate": 1.6252165818267564e-05, + "loss": 1.0131, + "step": 2823 + }, + { + "epoch": 1.5331161780673181, + "grad_norm": 6.665532827172479, + "learning_rate": 1.6249420709801462e-05, + "loss": 0.5454, + "step": 2824 + }, + { + "epoch": 1.5336590662323561, + "grad_norm": 7.096056294730765, + "learning_rate": 1.624667482839705e-05, + "loss": 1.043, + "step": 2825 + }, + { + "epoch": 1.5342019543973942, + "grad_norm": 5.384161272714602, + "learning_rate": 1.6243928174393935e-05, + "loss": 0.8345, + "step": 2826 + }, + { + "epoch": 1.5347448425624322, + "grad_norm": 8.398527018753486, + "learning_rate": 1.6241180748131834e-05, + "loss": 0.6911, + "step": 2827 + }, + { + "epoch": 1.5352877307274702, + "grad_norm": 6.880924484125559, + "learning_rate": 1.6238432549950552e-05, + "loss": 0.6039, + "step": 2828 + }, + { + "epoch": 1.5358306188925082, + "grad_norm": 9.23669884568822, + "learning_rate": 1.623568358018999e-05, + "loss": 1.1145, + "step": 2829 + }, + { + "epoch": 1.5363735070575462, + "grad_norm": 9.838376476241027, + "learning_rate": 1.6232933839190146e-05, + "loss": 1.0435, + "step": 2830 + }, + { + "epoch": 1.5369163952225842, + "grad_norm": 8.64475205930121, + "learning_rate": 1.6230183327291108e-05, + "loss": 1.0886, + "step": 2831 + }, + { + "epoch": 1.5374592833876222, + "grad_norm": 8.446847258487077, + "learning_rate": 1.6227432044833072e-05, + "loss": 0.8922, + "step": 2832 + }, + { + "epoch": 1.5380021715526602, + "grad_norm": 7.373724851347831, + "learning_rate": 1.622467999215631e-05, + "loss": 0.7823, + "step": 2833 + }, + { + "epoch": 1.5385450597176982, + "grad_norm": 9.661226436708493, + "learning_rate": 1.622192716960121e-05, + "loss": 1.2651, + "step": 2834 + }, + { + "epoch": 1.5390879478827362, + "grad_norm": 9.316115050127708, + "learning_rate": 1.6219173577508237e-05, + "loss": 0.897, + "step": 2835 + }, + { + "epoch": 1.5396308360477742, + "grad_norm": 12.011440258586251, + "learning_rate": 1.621641921621797e-05, + "loss": 0.868, + "step": 2836 + }, + { + "epoch": 1.5401737242128122, + "grad_norm": 7.737580132404622, + "learning_rate": 1.6213664086071058e-05, + "loss": 1.0727, + "step": 2837 + }, + { + "epoch": 1.5407166123778502, + "grad_norm": 6.696935187657473, + "learning_rate": 1.6210908187408275e-05, + "loss": 0.7404, + "step": 2838 + }, + { + "epoch": 1.5412595005428882, + "grad_norm": 6.460846604904407, + "learning_rate": 1.6208151520570465e-05, + "loss": 0.5892, + "step": 2839 + }, + { + "epoch": 1.5418023887079262, + "grad_norm": 8.244871763303658, + "learning_rate": 1.6205394085898586e-05, + "loss": 1.0692, + "step": 2840 + }, + { + "epoch": 1.5423452768729642, + "grad_norm": 6.140143281113326, + "learning_rate": 1.620263588373367e-05, + "loss": 0.7648, + "step": 2841 + }, + { + "epoch": 1.5428881650380022, + "grad_norm": 6.3369705303866235, + "learning_rate": 1.619987691441687e-05, + "loss": 0.6665, + "step": 2842 + }, + { + "epoch": 1.5434310532030402, + "grad_norm": 8.864356355873657, + "learning_rate": 1.6197117178289405e-05, + "loss": 0.9855, + "step": 2843 + }, + { + "epoch": 1.5439739413680782, + "grad_norm": 8.487808912288623, + "learning_rate": 1.6194356675692614e-05, + "loss": 1.3813, + "step": 2844 + }, + { + "epoch": 1.5445168295331162, + "grad_norm": 6.5782618837581435, + "learning_rate": 1.619159540696792e-05, + "loss": 1.1941, + "step": 2845 + }, + { + "epoch": 1.5450597176981542, + "grad_norm": 8.668580644259926, + "learning_rate": 1.6188833372456833e-05, + "loss": 1.1044, + "step": 2846 + }, + { + "epoch": 1.5456026058631922, + "grad_norm": 8.677506532765584, + "learning_rate": 1.6186070572500972e-05, + "loss": 1.3984, + "step": 2847 + }, + { + "epoch": 1.5461454940282302, + "grad_norm": 6.822422879140709, + "learning_rate": 1.6183307007442046e-05, + "loss": 0.6903, + "step": 2848 + }, + { + "epoch": 1.5466883821932682, + "grad_norm": 8.480828664812217, + "learning_rate": 1.6180542677621852e-05, + "loss": 1.2466, + "step": 2849 + }, + { + "epoch": 1.5472312703583062, + "grad_norm": 7.804300661419004, + "learning_rate": 1.617777758338229e-05, + "loss": 0.8833, + "step": 2850 + }, + { + "epoch": 1.5477741585233442, + "grad_norm": 6.418809144902363, + "learning_rate": 1.617501172506535e-05, + "loss": 0.9174, + "step": 2851 + }, + { + "epoch": 1.5483170466883822, + "grad_norm": 6.627083812958508, + "learning_rate": 1.617224510301312e-05, + "loss": 0.7867, + "step": 2852 + }, + { + "epoch": 1.5488599348534202, + "grad_norm": 7.792114190076133, + "learning_rate": 1.616947771756778e-05, + "loss": 0.6242, + "step": 2853 + }, + { + "epoch": 1.5494028230184582, + "grad_norm": 6.965108246318713, + "learning_rate": 1.6166709569071598e-05, + "loss": 0.5844, + "step": 2854 + }, + { + "epoch": 1.5499457111834962, + "grad_norm": 6.9966373703206655, + "learning_rate": 1.616394065786695e-05, + "loss": 1.1688, + "step": 2855 + }, + { + "epoch": 1.5504885993485342, + "grad_norm": 8.466603441198636, + "learning_rate": 1.6161170984296298e-05, + "loss": 0.9506, + "step": 2856 + }, + { + "epoch": 1.5510314875135722, + "grad_norm": 6.361127259105106, + "learning_rate": 1.61584005487022e-05, + "loss": 0.5573, + "step": 2857 + }, + { + "epoch": 1.5515743756786102, + "grad_norm": 6.137955961198825, + "learning_rate": 1.6155629351427306e-05, + "loss": 1.0474, + "step": 2858 + }, + { + "epoch": 1.5521172638436482, + "grad_norm": 6.262211837195136, + "learning_rate": 1.6152857392814367e-05, + "loss": 0.6786, + "step": 2859 + }, + { + "epoch": 1.5526601520086862, + "grad_norm": 9.3980797837583, + "learning_rate": 1.6150084673206214e-05, + "loss": 1.2893, + "step": 2860 + }, + { + "epoch": 1.5532030401737242, + "grad_norm": 9.600537200283373, + "learning_rate": 1.614731119294579e-05, + "loss": 0.8668, + "step": 2861 + }, + { + "epoch": 1.5537459283387622, + "grad_norm": 8.65304941049683, + "learning_rate": 1.614453695237612e-05, + "loss": 1.0377, + "step": 2862 + }, + { + "epoch": 1.5542888165038002, + "grad_norm": 6.075747510405264, + "learning_rate": 1.6141761951840327e-05, + "loss": 0.8203, + "step": 2863 + }, + { + "epoch": 1.5548317046688382, + "grad_norm": 6.578770531031146, + "learning_rate": 1.6138986191681626e-05, + "loss": 0.6808, + "step": 2864 + }, + { + "epoch": 1.5553745928338762, + "grad_norm": 9.206515322081769, + "learning_rate": 1.6136209672243332e-05, + "loss": 0.9111, + "step": 2865 + }, + { + "epoch": 1.5559174809989142, + "grad_norm": 8.829723041453656, + "learning_rate": 1.613343239386884e-05, + "loss": 1.7392, + "step": 2866 + }, + { + "epoch": 1.5564603691639523, + "grad_norm": 9.489367814866748, + "learning_rate": 1.613065435690166e-05, + "loss": 0.9186, + "step": 2867 + }, + { + "epoch": 1.5570032573289903, + "grad_norm": 7.7196736231335175, + "learning_rate": 1.6127875561685376e-05, + "loss": 0.769, + "step": 2868 + }, + { + "epoch": 1.5575461454940283, + "grad_norm": 7.668511006453606, + "learning_rate": 1.6125096008563677e-05, + "loss": 0.7572, + "step": 2869 + }, + { + "epoch": 1.5580890336590663, + "grad_norm": 8.31551289254719, + "learning_rate": 1.6122315697880343e-05, + "loss": 1.0657, + "step": 2870 + }, + { + "epoch": 1.5586319218241043, + "grad_norm": 11.828690633927348, + "learning_rate": 1.6119534629979244e-05, + "loss": 1.0095, + "step": 2871 + }, + { + "epoch": 1.5591748099891423, + "grad_norm": 7.137823181609677, + "learning_rate": 1.611675280520435e-05, + "loss": 0.7246, + "step": 2872 + }, + { + "epoch": 1.5597176981541803, + "grad_norm": 5.051313327297567, + "learning_rate": 1.611397022389972e-05, + "loss": 0.5905, + "step": 2873 + }, + { + "epoch": 1.5602605863192183, + "grad_norm": 8.165805991671068, + "learning_rate": 1.6111186886409504e-05, + "loss": 1.0678, + "step": 2874 + }, + { + "epoch": 1.5608034744842563, + "grad_norm": 11.597649604311457, + "learning_rate": 1.6108402793077957e-05, + "loss": 1.4836, + "step": 2875 + }, + { + "epoch": 1.5613463626492943, + "grad_norm": 7.953203184026488, + "learning_rate": 1.610561794424942e-05, + "loss": 0.9216, + "step": 2876 + }, + { + "epoch": 1.5618892508143323, + "grad_norm": 7.397930684647979, + "learning_rate": 1.6102832340268322e-05, + "loss": 0.6688, + "step": 2877 + }, + { + "epoch": 1.5624321389793703, + "grad_norm": 8.999084750858366, + "learning_rate": 1.6100045981479195e-05, + "loss": 1.254, + "step": 2878 + }, + { + "epoch": 1.5629750271444083, + "grad_norm": 7.445594920489225, + "learning_rate": 1.6097258868226658e-05, + "loss": 0.6406, + "step": 2879 + }, + { + "epoch": 1.5635179153094463, + "grad_norm": 10.469355169581702, + "learning_rate": 1.609447100085543e-05, + "loss": 0.7928, + "step": 2880 + }, + { + "epoch": 1.5640608034744843, + "grad_norm": 8.853634937458189, + "learning_rate": 1.6091682379710313e-05, + "loss": 1.0788, + "step": 2881 + }, + { + "epoch": 1.5646036916395223, + "grad_norm": 7.909620213759866, + "learning_rate": 1.6088893005136206e-05, + "loss": 1.011, + "step": 2882 + }, + { + "epoch": 1.5651465798045603, + "grad_norm": 10.379252728708625, + "learning_rate": 1.6086102877478117e-05, + "loss": 0.8675, + "step": 2883 + }, + { + "epoch": 1.5656894679695983, + "grad_norm": 8.143888785034273, + "learning_rate": 1.6083311997081116e-05, + "loss": 0.8935, + "step": 2884 + }, + { + "epoch": 1.5662323561346363, + "grad_norm": 8.577520797942114, + "learning_rate": 1.6080520364290396e-05, + "loss": 0.6763, + "step": 2885 + }, + { + "epoch": 1.5667752442996743, + "grad_norm": 9.226094142154105, + "learning_rate": 1.6077727979451228e-05, + "loss": 0.6679, + "step": 2886 + }, + { + "epoch": 1.5673181324647123, + "grad_norm": 11.197536431412058, + "learning_rate": 1.607493484290897e-05, + "loss": 1.0525, + "step": 2887 + }, + { + "epoch": 1.5678610206297503, + "grad_norm": 8.43012225245798, + "learning_rate": 1.6072140955009093e-05, + "loss": 1.0317, + "step": 2888 + }, + { + "epoch": 1.5684039087947883, + "grad_norm": 6.69358812320454, + "learning_rate": 1.606934631609715e-05, + "loss": 0.7703, + "step": 2889 + }, + { + "epoch": 1.5689467969598263, + "grad_norm": 9.36327291463899, + "learning_rate": 1.6066550926518776e-05, + "loss": 1.0844, + "step": 2890 + }, + { + "epoch": 1.5694896851248643, + "grad_norm": 8.40518653624583, + "learning_rate": 1.6063754786619716e-05, + "loss": 0.6863, + "step": 2891 + }, + { + "epoch": 1.5700325732899023, + "grad_norm": 9.552486022227878, + "learning_rate": 1.60609578967458e-05, + "loss": 1.0342, + "step": 2892 + }, + { + "epoch": 1.5705754614549403, + "grad_norm": 6.76445711513327, + "learning_rate": 1.6058160257242953e-05, + "loss": 0.8777, + "step": 2893 + }, + { + "epoch": 1.5711183496199783, + "grad_norm": 9.040602694092291, + "learning_rate": 1.6055361868457188e-05, + "loss": 0.9634, + "step": 2894 + }, + { + "epoch": 1.5716612377850163, + "grad_norm": 7.00278619777266, + "learning_rate": 1.6052562730734614e-05, + "loss": 0.8995, + "step": 2895 + }, + { + "epoch": 1.5722041259500543, + "grad_norm": 8.533327622490077, + "learning_rate": 1.604976284442144e-05, + "loss": 1.0896, + "step": 2896 + }, + { + "epoch": 1.5727470141150923, + "grad_norm": 9.895218231842426, + "learning_rate": 1.6046962209863953e-05, + "loss": 0.9309, + "step": 2897 + }, + { + "epoch": 1.5732899022801303, + "grad_norm": 5.945823958044379, + "learning_rate": 1.604416082740854e-05, + "loss": 0.6512, + "step": 2898 + }, + { + "epoch": 1.5738327904451683, + "grad_norm": 10.822429137471252, + "learning_rate": 1.6041358697401687e-05, + "loss": 1.0744, + "step": 2899 + }, + { + "epoch": 1.5743756786102063, + "grad_norm": 7.728147258511819, + "learning_rate": 1.603855582018996e-05, + "loss": 0.9566, + "step": 2900 + }, + { + "epoch": 1.5749185667752443, + "grad_norm": 8.643748761880973, + "learning_rate": 1.603575219612003e-05, + "loss": 0.9188, + "step": 2901 + }, + { + "epoch": 1.5754614549402823, + "grad_norm": 9.357572097357403, + "learning_rate": 1.603294782553864e-05, + "loss": 0.9045, + "step": 2902 + }, + { + "epoch": 1.5760043431053203, + "grad_norm": 6.76207345266047, + "learning_rate": 1.6030142708792653e-05, + "loss": 0.7658, + "step": 2903 + }, + { + "epoch": 1.5765472312703583, + "grad_norm": 6.47603683733189, + "learning_rate": 1.6027336846229005e-05, + "loss": 0.5406, + "step": 2904 + }, + { + "epoch": 1.5770901194353963, + "grad_norm": 6.030033648190992, + "learning_rate": 1.602453023819473e-05, + "loss": 0.5141, + "step": 2905 + }, + { + "epoch": 1.5776330076004343, + "grad_norm": 8.344821553681937, + "learning_rate": 1.6021722885036954e-05, + "loss": 1.255, + "step": 2906 + }, + { + "epoch": 1.5781758957654723, + "grad_norm": 7.2392259468937885, + "learning_rate": 1.601891478710289e-05, + "loss": 0.8358, + "step": 2907 + }, + { + "epoch": 1.5787187839305103, + "grad_norm": 8.011046574978268, + "learning_rate": 1.6016105944739856e-05, + "loss": 1.0316, + "step": 2908 + }, + { + "epoch": 1.5792616720955484, + "grad_norm": 6.609140115647256, + "learning_rate": 1.601329635829525e-05, + "loss": 0.7324, + "step": 2909 + }, + { + "epoch": 1.5798045602605864, + "grad_norm": 6.672864395642082, + "learning_rate": 1.6010486028116568e-05, + "loss": 0.6386, + "step": 2910 + }, + { + "epoch": 1.5803474484256244, + "grad_norm": 7.270447246601273, + "learning_rate": 1.600767495455139e-05, + "loss": 0.6699, + "step": 2911 + }, + { + "epoch": 1.5808903365906624, + "grad_norm": 9.7186017734922, + "learning_rate": 1.6004863137947405e-05, + "loss": 0.9604, + "step": 2912 + }, + { + "epoch": 1.5814332247557004, + "grad_norm": 7.18305646407285, + "learning_rate": 1.6002050578652374e-05, + "loss": 0.7286, + "step": 2913 + }, + { + "epoch": 1.5819761129207384, + "grad_norm": 8.171408196684215, + "learning_rate": 1.5999237277014162e-05, + "loss": 0.9121, + "step": 2914 + }, + { + "epoch": 1.5825190010857764, + "grad_norm": 7.710673421134947, + "learning_rate": 1.599642323338072e-05, + "loss": 0.7287, + "step": 2915 + }, + { + "epoch": 1.5830618892508144, + "grad_norm": 8.264934044661224, + "learning_rate": 1.5993608448100095e-05, + "loss": 0.6537, + "step": 2916 + }, + { + "epoch": 1.5836047774158524, + "grad_norm": 7.874914961028911, + "learning_rate": 1.599079292152043e-05, + "loss": 0.6511, + "step": 2917 + }, + { + "epoch": 1.5841476655808904, + "grad_norm": 7.399378371785218, + "learning_rate": 1.5987976653989945e-05, + "loss": 0.6236, + "step": 2918 + }, + { + "epoch": 1.5846905537459284, + "grad_norm": 7.837764322959507, + "learning_rate": 1.5985159645856966e-05, + "loss": 0.8184, + "step": 2919 + }, + { + "epoch": 1.5852334419109664, + "grad_norm": 8.64272103945922, + "learning_rate": 1.5982341897469903e-05, + "loss": 0.9182, + "step": 2920 + }, + { + "epoch": 1.5857763300760044, + "grad_norm": 9.994158036669248, + "learning_rate": 1.5979523409177254e-05, + "loss": 1.3466, + "step": 2921 + }, + { + "epoch": 1.5863192182410424, + "grad_norm": 6.671268632879107, + "learning_rate": 1.5976704181327626e-05, + "loss": 0.7506, + "step": 2922 + }, + { + "epoch": 1.5868621064060804, + "grad_norm": 8.161895609948308, + "learning_rate": 1.59738842142697e-05, + "loss": 0.9683, + "step": 2923 + }, + { + "epoch": 1.5874049945711184, + "grad_norm": 7.236619510875819, + "learning_rate": 1.597106350835225e-05, + "loss": 0.798, + "step": 2924 + }, + { + "epoch": 1.5879478827361564, + "grad_norm": 10.263209688443808, + "learning_rate": 1.5968242063924152e-05, + "loss": 1.0799, + "step": 2925 + }, + { + "epoch": 1.5884907709011944, + "grad_norm": 9.433308260958457, + "learning_rate": 1.596541988133436e-05, + "loss": 0.8755, + "step": 2926 + }, + { + "epoch": 1.5890336590662324, + "grad_norm": 9.48995441422385, + "learning_rate": 1.5962596960931927e-05, + "loss": 1.4712, + "step": 2927 + }, + { + "epoch": 1.5895765472312704, + "grad_norm": 7.317297215230638, + "learning_rate": 1.5959773303066005e-05, + "loss": 1.0421, + "step": 2928 + }, + { + "epoch": 1.5901194353963084, + "grad_norm": 7.491426794362114, + "learning_rate": 1.595694890808582e-05, + "loss": 0.9468, + "step": 2929 + }, + { + "epoch": 1.5906623235613464, + "grad_norm": 7.320440852081977, + "learning_rate": 1.5954123776340702e-05, + "loss": 0.7459, + "step": 2930 + }, + { + "epoch": 1.5912052117263844, + "grad_norm": 8.334158307527314, + "learning_rate": 1.5951297908180062e-05, + "loss": 1.3182, + "step": 2931 + }, + { + "epoch": 1.5917480998914224, + "grad_norm": 7.02613629768153, + "learning_rate": 1.5948471303953418e-05, + "loss": 0.919, + "step": 2932 + }, + { + "epoch": 1.5922909880564604, + "grad_norm": 6.904443487786306, + "learning_rate": 1.594564396401036e-05, + "loss": 0.5737, + "step": 2933 + }, + { + "epoch": 1.5928338762214984, + "grad_norm": 8.041576521446402, + "learning_rate": 1.594281588870058e-05, + "loss": 0.7191, + "step": 2934 + }, + { + "epoch": 1.5933767643865364, + "grad_norm": 8.557472225459021, + "learning_rate": 1.5939987078373856e-05, + "loss": 0.8936, + "step": 2935 + }, + { + "epoch": 1.5939196525515744, + "grad_norm": 10.051674589973098, + "learning_rate": 1.5937157533380065e-05, + "loss": 1.1777, + "step": 2936 + }, + { + "epoch": 1.5944625407166124, + "grad_norm": 7.1580018473739235, + "learning_rate": 1.5934327254069167e-05, + "loss": 0.8628, + "step": 2937 + }, + { + "epoch": 1.5950054288816504, + "grad_norm": 8.689338133459966, + "learning_rate": 1.593149624079122e-05, + "loss": 1.1877, + "step": 2938 + }, + { + "epoch": 1.5955483170466884, + "grad_norm": 8.40904043481009, + "learning_rate": 1.5928664493896364e-05, + "loss": 0.9815, + "step": 2939 + }, + { + "epoch": 1.5960912052117264, + "grad_norm": 7.778305354024487, + "learning_rate": 1.5925832013734832e-05, + "loss": 0.8282, + "step": 2940 + }, + { + "epoch": 1.5966340933767644, + "grad_norm": 7.074281482828989, + "learning_rate": 1.5922998800656956e-05, + "loss": 0.8347, + "step": 2941 + }, + { + "epoch": 1.5971769815418024, + "grad_norm": 9.188840060350078, + "learning_rate": 1.5920164855013145e-05, + "loss": 0.8566, + "step": 2942 + }, + { + "epoch": 1.5977198697068404, + "grad_norm": 7.021116417649334, + "learning_rate": 1.591733017715391e-05, + "loss": 0.8367, + "step": 2943 + }, + { + "epoch": 1.5982627578718784, + "grad_norm": 8.039041768789302, + "learning_rate": 1.5914494767429846e-05, + "loss": 0.9889, + "step": 2944 + }, + { + "epoch": 1.5988056460369164, + "grad_norm": 6.252701385804125, + "learning_rate": 1.5911658626191645e-05, + "loss": 1.0478, + "step": 2945 + }, + { + "epoch": 1.5993485342019544, + "grad_norm": 8.223409826369243, + "learning_rate": 1.5908821753790083e-05, + "loss": 1.2458, + "step": 2946 + }, + { + "epoch": 1.5998914223669924, + "grad_norm": 8.972690403999096, + "learning_rate": 1.590598415057603e-05, + "loss": 0.8471, + "step": 2947 + }, + { + "epoch": 1.6004343105320304, + "grad_norm": 8.163032020074416, + "learning_rate": 1.5903145816900445e-05, + "loss": 0.8213, + "step": 2948 + }, + { + "epoch": 1.6009771986970684, + "grad_norm": 5.667868413023322, + "learning_rate": 1.5900306753114375e-05, + "loss": 0.8426, + "step": 2949 + }, + { + "epoch": 1.6015200868621065, + "grad_norm": 6.960744978365833, + "learning_rate": 1.5897466959568967e-05, + "loss": 0.7224, + "step": 2950 + }, + { + "epoch": 1.6020629750271445, + "grad_norm": 7.648210267430006, + "learning_rate": 1.589462643661544e-05, + "loss": 0.7451, + "step": 2951 + }, + { + "epoch": 1.6026058631921825, + "grad_norm": 5.9568596208122875, + "learning_rate": 1.5891785184605123e-05, + "loss": 0.5252, + "step": 2952 + }, + { + "epoch": 1.6031487513572205, + "grad_norm": 6.41676218105273, + "learning_rate": 1.5888943203889427e-05, + "loss": 0.6502, + "step": 2953 + }, + { + "epoch": 1.6036916395222585, + "grad_norm": 7.704871607753619, + "learning_rate": 1.5886100494819846e-05, + "loss": 1.1762, + "step": 2954 + }, + { + "epoch": 1.6042345276872965, + "grad_norm": 7.656219757403813, + "learning_rate": 1.5883257057747975e-05, + "loss": 0.8076, + "step": 2955 + }, + { + "epoch": 1.6047774158523345, + "grad_norm": 6.897030921112048, + "learning_rate": 1.58804128930255e-05, + "loss": 0.883, + "step": 2956 + }, + { + "epoch": 1.6053203040173725, + "grad_norm": 6.718724904599124, + "learning_rate": 1.5877568001004182e-05, + "loss": 0.8533, + "step": 2957 + }, + { + "epoch": 1.6058631921824105, + "grad_norm": 8.458861423777194, + "learning_rate": 1.5874722382035887e-05, + "loss": 1.2017, + "step": 2958 + }, + { + "epoch": 1.6064060803474485, + "grad_norm": 9.16622725543938, + "learning_rate": 1.5871876036472565e-05, + "loss": 1.0074, + "step": 2959 + }, + { + "epoch": 1.6069489685124865, + "grad_norm": 7.053786350768386, + "learning_rate": 1.5869028964666254e-05, + "loss": 0.8193, + "step": 2960 + }, + { + "epoch": 1.6074918566775245, + "grad_norm": 7.670976634760575, + "learning_rate": 1.5866181166969088e-05, + "loss": 0.8331, + "step": 2961 + }, + { + "epoch": 1.6080347448425625, + "grad_norm": 9.942186867470006, + "learning_rate": 1.586333264373329e-05, + "loss": 0.8672, + "step": 2962 + }, + { + "epoch": 1.6085776330076005, + "grad_norm": 7.744463789862724, + "learning_rate": 1.586048339531116e-05, + "loss": 0.8859, + "step": 2963 + }, + { + "epoch": 1.6091205211726385, + "grad_norm": 7.181553947846623, + "learning_rate": 1.5857633422055104e-05, + "loss": 1.0376, + "step": 2964 + }, + { + "epoch": 1.6096634093376765, + "grad_norm": 7.184586780377298, + "learning_rate": 1.5854782724317616e-05, + "loss": 0.694, + "step": 2965 + }, + { + "epoch": 1.6102062975027145, + "grad_norm": 6.9154974167720855, + "learning_rate": 1.5851931302451262e-05, + "loss": 0.8573, + "step": 2966 + }, + { + "epoch": 1.6107491856677525, + "grad_norm": 5.056944401416877, + "learning_rate": 1.5849079156808726e-05, + "loss": 0.4972, + "step": 2967 + }, + { + "epoch": 1.6112920738327905, + "grad_norm": 6.98635366168859, + "learning_rate": 1.584622628774275e-05, + "loss": 0.9002, + "step": 2968 + }, + { + "epoch": 1.6118349619978285, + "grad_norm": 9.22418255248469, + "learning_rate": 1.5843372695606196e-05, + "loss": 1.4143, + "step": 2969 + }, + { + "epoch": 1.6123778501628665, + "grad_norm": 9.49766654897752, + "learning_rate": 1.584051838075199e-05, + "loss": 1.0145, + "step": 2970 + }, + { + "epoch": 1.6129207383279045, + "grad_norm": 9.681135461557052, + "learning_rate": 1.5837663343533166e-05, + "loss": 0.8841, + "step": 2971 + }, + { + "epoch": 1.6134636264929425, + "grad_norm": 6.00661489729409, + "learning_rate": 1.583480758430283e-05, + "loss": 0.677, + "step": 2972 + }, + { + "epoch": 1.6140065146579805, + "grad_norm": 7.843885336692957, + "learning_rate": 1.5831951103414194e-05, + "loss": 0.4729, + "step": 2973 + }, + { + "epoch": 1.6145494028230185, + "grad_norm": 9.391971866087287, + "learning_rate": 1.5829093901220557e-05, + "loss": 1.0833, + "step": 2974 + }, + { + "epoch": 1.6150922909880565, + "grad_norm": 7.202123614778415, + "learning_rate": 1.582623597807529e-05, + "loss": 0.9755, + "step": 2975 + }, + { + "epoch": 1.6156351791530945, + "grad_norm": 8.050976157508988, + "learning_rate": 1.5823377334331875e-05, + "loss": 0.8981, + "step": 2976 + }, + { + "epoch": 1.6161780673181325, + "grad_norm": 6.409986462897389, + "learning_rate": 1.5820517970343867e-05, + "loss": 0.3994, + "step": 2977 + }, + { + "epoch": 1.6167209554831705, + "grad_norm": 6.001133896600598, + "learning_rate": 1.581765788646492e-05, + "loss": 0.4701, + "step": 2978 + }, + { + "epoch": 1.6172638436482085, + "grad_norm": 12.858193971885504, + "learning_rate": 1.581479708304878e-05, + "loss": 1.4327, + "step": 2979 + }, + { + "epoch": 1.6178067318132465, + "grad_norm": 9.559044098763156, + "learning_rate": 1.5811935560449262e-05, + "loss": 0.9058, + "step": 2980 + }, + { + "epoch": 1.6183496199782845, + "grad_norm": 8.041859101089647, + "learning_rate": 1.5809073319020293e-05, + "loss": 0.6251, + "step": 2981 + }, + { + "epoch": 1.6188925081433225, + "grad_norm": 7.413714357196275, + "learning_rate": 1.580621035911588e-05, + "loss": 0.7887, + "step": 2982 + }, + { + "epoch": 1.6194353963083605, + "grad_norm": 11.941805543277479, + "learning_rate": 1.5803346681090113e-05, + "loss": 1.2547, + "step": 2983 + }, + { + "epoch": 1.6199782844733985, + "grad_norm": 5.728003242952451, + "learning_rate": 1.580048228529718e-05, + "loss": 0.4941, + "step": 2984 + }, + { + "epoch": 1.6205211726384365, + "grad_norm": 8.67398306592389, + "learning_rate": 1.5797617172091354e-05, + "loss": 0.7758, + "step": 2985 + }, + { + "epoch": 1.6210640608034745, + "grad_norm": 6.198163559825324, + "learning_rate": 1.5794751341826996e-05, + "loss": 0.4601, + "step": 2986 + }, + { + "epoch": 1.6216069489685125, + "grad_norm": 8.257469675141675, + "learning_rate": 1.5791884794858557e-05, + "loss": 0.8679, + "step": 2987 + }, + { + "epoch": 1.6221498371335505, + "grad_norm": 12.953868836673164, + "learning_rate": 1.5789017531540575e-05, + "loss": 1.572, + "step": 2988 + }, + { + "epoch": 1.6226927252985885, + "grad_norm": 9.232767459675694, + "learning_rate": 1.5786149552227682e-05, + "loss": 0.7511, + "step": 2989 + }, + { + "epoch": 1.6232356134636265, + "grad_norm": 7.465499623540017, + "learning_rate": 1.5783280857274586e-05, + "loss": 0.5754, + "step": 2990 + }, + { + "epoch": 1.6237785016286646, + "grad_norm": 5.087952971939767, + "learning_rate": 1.5780411447036097e-05, + "loss": 0.532, + "step": 2991 + }, + { + "epoch": 1.6243213897937026, + "grad_norm": 11.064786965172743, + "learning_rate": 1.577754132186711e-05, + "loss": 0.7195, + "step": 2992 + }, + { + "epoch": 1.6248642779587406, + "grad_norm": 10.268274657097892, + "learning_rate": 1.57746704821226e-05, + "loss": 1.1258, + "step": 2993 + }, + { + "epoch": 1.6254071661237783, + "grad_norm": 7.650078703788778, + "learning_rate": 1.5771798928157645e-05, + "loss": 0.5949, + "step": 2994 + }, + { + "epoch": 1.6259500542888166, + "grad_norm": 9.330939436907839, + "learning_rate": 1.5768926660327396e-05, + "loss": 0.9243, + "step": 2995 + }, + { + "epoch": 1.6264929424538543, + "grad_norm": 11.566902464205237, + "learning_rate": 1.576605367898711e-05, + "loss": 1.5588, + "step": 2996 + }, + { + "epoch": 1.6270358306188926, + "grad_norm": 9.91827217636312, + "learning_rate": 1.576317998449211e-05, + "loss": 1.1418, + "step": 2997 + }, + { + "epoch": 1.6275787187839303, + "grad_norm": 5.842910545051031, + "learning_rate": 1.5760305577197824e-05, + "loss": 0.5064, + "step": 2998 + }, + { + "epoch": 1.6281216069489686, + "grad_norm": 7.20637691224012, + "learning_rate": 1.5757430457459765e-05, + "loss": 0.8129, + "step": 2999 + }, + { + "epoch": 1.6286644951140063, + "grad_norm": 7.5338669003432965, + "learning_rate": 1.5754554625633535e-05, + "loss": 0.7244, + "step": 3000 + }, + { + "epoch": 1.6292073832790446, + "grad_norm": 7.489311206471124, + "learning_rate": 1.5751678082074813e-05, + "loss": 1.1409, + "step": 3001 + }, + { + "epoch": 1.6297502714440824, + "grad_norm": 7.26299222123934, + "learning_rate": 1.574880082713938e-05, + "loss": 0.7061, + "step": 3002 + }, + { + "epoch": 1.6302931596091206, + "grad_norm": 6.292246912680897, + "learning_rate": 1.5745922861183095e-05, + "loss": 0.6383, + "step": 3003 + }, + { + "epoch": 1.6308360477741584, + "grad_norm": 7.483086031387996, + "learning_rate": 1.574304418456192e-05, + "loss": 0.8194, + "step": 3004 + }, + { + "epoch": 1.6313789359391966, + "grad_norm": 8.062451557552306, + "learning_rate": 1.5740164797631882e-05, + "loss": 0.9474, + "step": 3005 + }, + { + "epoch": 1.6319218241042344, + "grad_norm": 7.62904254588456, + "learning_rate": 1.5737284700749116e-05, + "loss": 1.2324, + "step": 3006 + }, + { + "epoch": 1.6324647122692726, + "grad_norm": 8.533361938490982, + "learning_rate": 1.573440389426983e-05, + "loss": 0.6182, + "step": 3007 + }, + { + "epoch": 1.6330076004343104, + "grad_norm": 7.0451736765954855, + "learning_rate": 1.5731522378550337e-05, + "loss": 0.7464, + "step": 3008 + }, + { + "epoch": 1.6335504885993486, + "grad_norm": 8.121336222871497, + "learning_rate": 1.572864015394702e-05, + "loss": 1.1087, + "step": 3009 + }, + { + "epoch": 1.6340933767643864, + "grad_norm": 7.2112540180194955, + "learning_rate": 1.5725757220816356e-05, + "loss": 0.7756, + "step": 3010 + }, + { + "epoch": 1.6346362649294246, + "grad_norm": 8.092755806422884, + "learning_rate": 1.5722873579514915e-05, + "loss": 1.1184, + "step": 3011 + }, + { + "epoch": 1.6351791530944624, + "grad_norm": 7.485157971844339, + "learning_rate": 1.5719989230399347e-05, + "loss": 0.6839, + "step": 3012 + }, + { + "epoch": 1.6357220412595006, + "grad_norm": 8.516980380341119, + "learning_rate": 1.5717104173826397e-05, + "loss": 0.948, + "step": 3013 + }, + { + "epoch": 1.6362649294245384, + "grad_norm": 7.602547351124631, + "learning_rate": 1.571421841015289e-05, + "loss": 0.8479, + "step": 3014 + }, + { + "epoch": 1.6368078175895766, + "grad_norm": 7.0470960880026245, + "learning_rate": 1.5711331939735744e-05, + "loss": 0.9014, + "step": 3015 + }, + { + "epoch": 1.6373507057546144, + "grad_norm": 6.760810209806174, + "learning_rate": 1.570844476293196e-05, + "loss": 0.9316, + "step": 3016 + }, + { + "epoch": 1.6378935939196526, + "grad_norm": 8.02053578699558, + "learning_rate": 1.570555688009863e-05, + "loss": 0.7837, + "step": 3017 + }, + { + "epoch": 1.6384364820846904, + "grad_norm": 4.624578610614632, + "learning_rate": 1.5702668291592936e-05, + "loss": 0.4706, + "step": 3018 + }, + { + "epoch": 1.6389793702497286, + "grad_norm": 8.011712265192347, + "learning_rate": 1.569977899777213e-05, + "loss": 1.0338, + "step": 3019 + }, + { + "epoch": 1.6395222584147664, + "grad_norm": 7.773921575637797, + "learning_rate": 1.569688899899358e-05, + "loss": 0.7512, + "step": 3020 + }, + { + "epoch": 1.6400651465798046, + "grad_norm": 12.19433508686472, + "learning_rate": 1.569399829561472e-05, + "loss": 1.3502, + "step": 3021 + }, + { + "epoch": 1.6406080347448424, + "grad_norm": 11.319107865931453, + "learning_rate": 1.569110688799307e-05, + "loss": 0.9996, + "step": 3022 + }, + { + "epoch": 1.6411509229098806, + "grad_norm": 7.1911299711947585, + "learning_rate": 1.5688214776486255e-05, + "loss": 0.9037, + "step": 3023 + }, + { + "epoch": 1.6416938110749184, + "grad_norm": 8.326843836043436, + "learning_rate": 1.5685321961451968e-05, + "loss": 1.1179, + "step": 3024 + }, + { + "epoch": 1.6422366992399566, + "grad_norm": 8.257534801333424, + "learning_rate": 1.5682428443248002e-05, + "loss": 0.7197, + "step": 3025 + }, + { + "epoch": 1.6427795874049944, + "grad_norm": 6.669062512754087, + "learning_rate": 1.567953422223223e-05, + "loss": 0.6292, + "step": 3026 + }, + { + "epoch": 1.6433224755700326, + "grad_norm": 7.99477616152165, + "learning_rate": 1.567663929876261e-05, + "loss": 0.926, + "step": 3027 + }, + { + "epoch": 1.6438653637350704, + "grad_norm": 5.3717184947082615, + "learning_rate": 1.56737436731972e-05, + "loss": 0.5378, + "step": 3028 + }, + { + "epoch": 1.6444082519001086, + "grad_norm": 6.522826762176589, + "learning_rate": 1.5670847345894125e-05, + "loss": 0.7393, + "step": 3029 + }, + { + "epoch": 1.6449511400651464, + "grad_norm": 9.503371738808127, + "learning_rate": 1.5667950317211612e-05, + "loss": 1.2727, + "step": 3030 + }, + { + "epoch": 1.6454940282301846, + "grad_norm": 8.208077816222845, + "learning_rate": 1.5665052587507974e-05, + "loss": 0.7032, + "step": 3031 + }, + { + "epoch": 1.6460369163952224, + "grad_norm": 9.101101018529372, + "learning_rate": 1.56621541571416e-05, + "loss": 0.6983, + "step": 3032 + }, + { + "epoch": 1.6465798045602607, + "grad_norm": 7.254188682380924, + "learning_rate": 1.565925502647098e-05, + "loss": 0.7196, + "step": 3033 + }, + { + "epoch": 1.6471226927252984, + "grad_norm": 7.837018641144827, + "learning_rate": 1.5656355195854676e-05, + "loss": 1.2035, + "step": 3034 + }, + { + "epoch": 1.6476655808903367, + "grad_norm": 8.949191823273495, + "learning_rate": 1.5653454665651344e-05, + "loss": 1.0169, + "step": 3035 + }, + { + "epoch": 1.6482084690553744, + "grad_norm": 8.166317934132502, + "learning_rate": 1.5650553436219732e-05, + "loss": 0.672, + "step": 3036 + }, + { + "epoch": 1.6487513572204127, + "grad_norm": 6.443983188038021, + "learning_rate": 1.564765150791866e-05, + "loss": 0.5552, + "step": 3037 + }, + { + "epoch": 1.6492942453854504, + "grad_norm": 10.403060348162544, + "learning_rate": 1.5644748881107057e-05, + "loss": 1.0709, + "step": 3038 + }, + { + "epoch": 1.6498371335504887, + "grad_norm": 8.319421686790951, + "learning_rate": 1.564184555614391e-05, + "loss": 0.7923, + "step": 3039 + }, + { + "epoch": 1.6503800217155264, + "grad_norm": 9.65474819234395, + "learning_rate": 1.5638941533388318e-05, + "loss": 0.8519, + "step": 3040 + }, + { + "epoch": 1.6509229098805647, + "grad_norm": 8.635959327913778, + "learning_rate": 1.5636036813199445e-05, + "loss": 0.8447, + "step": 3041 + }, + { + "epoch": 1.6514657980456025, + "grad_norm": 7.674155625456285, + "learning_rate": 1.563313139593656e-05, + "loss": 0.7205, + "step": 3042 + }, + { + "epoch": 1.6520086862106407, + "grad_norm": 7.249478120616782, + "learning_rate": 1.5630225281959003e-05, + "loss": 0.6632, + "step": 3043 + }, + { + "epoch": 1.6525515743756785, + "grad_norm": 8.90384172612564, + "learning_rate": 1.5627318471626208e-05, + "loss": 0.8243, + "step": 3044 + }, + { + "epoch": 1.6530944625407167, + "grad_norm": 8.274462585020828, + "learning_rate": 1.5624410965297703e-05, + "loss": 0.7381, + "step": 3045 + }, + { + "epoch": 1.6536373507057545, + "grad_norm": 7.266215459792981, + "learning_rate": 1.562150276333308e-05, + "loss": 0.4989, + "step": 3046 + }, + { + "epoch": 1.6541802388707927, + "grad_norm": 4.816150056433193, + "learning_rate": 1.5618593866092036e-05, + "loss": 0.5382, + "step": 3047 + }, + { + "epoch": 1.6547231270358305, + "grad_norm": 6.835417235425346, + "learning_rate": 1.561568427393435e-05, + "loss": 1.044, + "step": 3048 + }, + { + "epoch": 1.6552660152008687, + "grad_norm": 7.2539344971377435, + "learning_rate": 1.5612773987219885e-05, + "loss": 0.5901, + "step": 3049 + }, + { + "epoch": 1.6558089033659065, + "grad_norm": 11.003646663954257, + "learning_rate": 1.5609863006308586e-05, + "loss": 0.9407, + "step": 3050 + }, + { + "epoch": 1.6563517915309447, + "grad_norm": 8.268114443948154, + "learning_rate": 1.560695133156049e-05, + "loss": 1.0164, + "step": 3051 + }, + { + "epoch": 1.6568946796959825, + "grad_norm": 7.824779774533612, + "learning_rate": 1.5604038963335716e-05, + "loss": 0.7111, + "step": 3052 + }, + { + "epoch": 1.6574375678610207, + "grad_norm": 8.542948361290183, + "learning_rate": 1.560112590199447e-05, + "loss": 0.6501, + "step": 3053 + }, + { + "epoch": 1.6579804560260585, + "grad_norm": 7.85824626614946, + "learning_rate": 1.5598212147897047e-05, + "loss": 0.752, + "step": 3054 + }, + { + "epoch": 1.6585233441910967, + "grad_norm": 6.588621135611242, + "learning_rate": 1.559529770140382e-05, + "loss": 0.507, + "step": 3055 + }, + { + "epoch": 1.6590662323561345, + "grad_norm": 6.418764525989286, + "learning_rate": 1.559238256287526e-05, + "loss": 0.5947, + "step": 3056 + }, + { + "epoch": 1.6596091205211727, + "grad_norm": 7.681071515192806, + "learning_rate": 1.5589466732671913e-05, + "loss": 0.8854, + "step": 3057 + }, + { + "epoch": 1.6601520086862105, + "grad_norm": 5.918230957225734, + "learning_rate": 1.558655021115441e-05, + "loss": 0.5546, + "step": 3058 + }, + { + "epoch": 1.6606948968512487, + "grad_norm": 10.669803193569846, + "learning_rate": 1.5583632998683475e-05, + "loss": 1.1548, + "step": 3059 + }, + { + "epoch": 1.6612377850162865, + "grad_norm": 9.271579004183549, + "learning_rate": 1.558071509561991e-05, + "loss": 0.8049, + "step": 3060 + }, + { + "epoch": 1.6617806731813247, + "grad_norm": 8.337085337830533, + "learning_rate": 1.557779650232461e-05, + "loss": 0.954, + "step": 3061 + }, + { + "epoch": 1.6623235613463625, + "grad_norm": 11.145075141347627, + "learning_rate": 1.5574877219158543e-05, + "loss": 1.1592, + "step": 3062 + }, + { + "epoch": 1.6628664495114007, + "grad_norm": 13.953049521423079, + "learning_rate": 1.557195724648278e-05, + "loss": 1.5699, + "step": 3063 + }, + { + "epoch": 1.6634093376764385, + "grad_norm": 8.654584332416588, + "learning_rate": 1.5569036584658466e-05, + "loss": 1.0185, + "step": 3064 + }, + { + "epoch": 1.6639522258414767, + "grad_norm": 7.520260829640502, + "learning_rate": 1.556611523404683e-05, + "loss": 0.7898, + "step": 3065 + }, + { + "epoch": 1.6644951140065145, + "grad_norm": 6.7373805901465875, + "learning_rate": 1.5563193195009188e-05, + "loss": 0.8392, + "step": 3066 + }, + { + "epoch": 1.6650380021715527, + "grad_norm": 6.732598633253056, + "learning_rate": 1.556027046790695e-05, + "loss": 0.6672, + "step": 3067 + }, + { + "epoch": 1.6655808903365905, + "grad_norm": 9.217725100287737, + "learning_rate": 1.55573470531016e-05, + "loss": 0.9874, + "step": 3068 + }, + { + "epoch": 1.6661237785016287, + "grad_norm": 6.963043776606882, + "learning_rate": 1.5554422950954706e-05, + "loss": 0.812, + "step": 3069 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 6.449074540999346, + "learning_rate": 1.5551498161827928e-05, + "loss": 0.5035, + "step": 3070 + }, + { + "epoch": 1.6672095548317047, + "grad_norm": 7.793693453975787, + "learning_rate": 1.554857268608301e-05, + "loss": 0.7974, + "step": 3071 + }, + { + "epoch": 1.6677524429967425, + "grad_norm": 10.124707886106192, + "learning_rate": 1.554564652408178e-05, + "loss": 1.4946, + "step": 3072 + }, + { + "epoch": 1.6682953311617807, + "grad_norm": 5.481775138355068, + "learning_rate": 1.5542719676186153e-05, + "loss": 0.502, + "step": 3073 + }, + { + "epoch": 1.6688382193268185, + "grad_norm": 6.055974601989459, + "learning_rate": 1.5539792142758114e-05, + "loss": 0.7463, + "step": 3074 + }, + { + "epoch": 1.6693811074918568, + "grad_norm": 7.452121005302326, + "learning_rate": 1.5536863924159762e-05, + "loss": 0.7282, + "step": 3075 + }, + { + "epoch": 1.6699239956568945, + "grad_norm": 8.991453230181806, + "learning_rate": 1.5533935020753252e-05, + "loss": 0.9327, + "step": 3076 + }, + { + "epoch": 1.6704668838219328, + "grad_norm": 8.138196256677595, + "learning_rate": 1.5531005432900838e-05, + "loss": 0.658, + "step": 3077 + }, + { + "epoch": 1.6710097719869705, + "grad_norm": 6.674313540160648, + "learning_rate": 1.552807516096486e-05, + "loss": 0.569, + "step": 3078 + }, + { + "epoch": 1.6715526601520088, + "grad_norm": 7.2303438904479105, + "learning_rate": 1.5525144205307734e-05, + "loss": 0.5186, + "step": 3079 + }, + { + "epoch": 1.6720955483170465, + "grad_norm": 6.872945037718312, + "learning_rate": 1.5522212566291966e-05, + "loss": 0.7282, + "step": 3080 + }, + { + "epoch": 1.6726384364820848, + "grad_norm": 7.913543579080126, + "learning_rate": 1.5519280244280146e-05, + "loss": 0.7015, + "step": 3081 + }, + { + "epoch": 1.6731813246471225, + "grad_norm": 6.144062932600708, + "learning_rate": 1.551634723963495e-05, + "loss": 0.7829, + "step": 3082 + }, + { + "epoch": 1.6737242128121608, + "grad_norm": 7.111349951384809, + "learning_rate": 1.5513413552719137e-05, + "loss": 0.6904, + "step": 3083 + }, + { + "epoch": 1.6742671009771986, + "grad_norm": 10.453360973997349, + "learning_rate": 1.5510479183895545e-05, + "loss": 1.4584, + "step": 3084 + }, + { + "epoch": 1.6748099891422368, + "grad_norm": 13.976555611119428, + "learning_rate": 1.5507544133527108e-05, + "loss": 1.542, + "step": 3085 + }, + { + "epoch": 1.6753528773072746, + "grad_norm": 5.894489690173581, + "learning_rate": 1.5504608401976835e-05, + "loss": 0.6792, + "step": 3086 + }, + { + "epoch": 1.6758957654723128, + "grad_norm": 8.937162315762803, + "learning_rate": 1.550167198960782e-05, + "loss": 0.8243, + "step": 3087 + }, + { + "epoch": 1.6764386536373506, + "grad_norm": 9.787723535687237, + "learning_rate": 1.549873489678325e-05, + "loss": 1.0071, + "step": 3088 + }, + { + "epoch": 1.6769815418023888, + "grad_norm": 6.0371827348978755, + "learning_rate": 1.5495797123866378e-05, + "loss": 0.6407, + "step": 3089 + }, + { + "epoch": 1.6775244299674266, + "grad_norm": 5.997673025227573, + "learning_rate": 1.5492858671220563e-05, + "loss": 0.7129, + "step": 3090 + }, + { + "epoch": 1.6780673181324648, + "grad_norm": 11.08813931574187, + "learning_rate": 1.5489919539209232e-05, + "loss": 0.8435, + "step": 3091 + }, + { + "epoch": 1.6786102062975026, + "grad_norm": 8.730951161091724, + "learning_rate": 1.54869797281959e-05, + "loss": 0.9159, + "step": 3092 + }, + { + "epoch": 1.6791530944625408, + "grad_norm": 7.293119927853279, + "learning_rate": 1.5484039238544173e-05, + "loss": 1.1001, + "step": 3093 + }, + { + "epoch": 1.6796959826275786, + "grad_norm": 7.091810627268002, + "learning_rate": 1.5481098070617734e-05, + "loss": 0.747, + "step": 3094 + }, + { + "epoch": 1.6802388707926168, + "grad_norm": 9.079124871383598, + "learning_rate": 1.547815622478035e-05, + "loss": 1.2292, + "step": 3095 + }, + { + "epoch": 1.6807817589576546, + "grad_norm": 7.238161650182196, + "learning_rate": 1.5475213701395867e-05, + "loss": 0.6657, + "step": 3096 + }, + { + "epoch": 1.6813246471226928, + "grad_norm": 9.385885763398004, + "learning_rate": 1.5472270500828236e-05, + "loss": 1.305, + "step": 3097 + }, + { + "epoch": 1.6818675352877306, + "grad_norm": 8.98902696621929, + "learning_rate": 1.5469326623441463e-05, + "loss": 0.8641, + "step": 3098 + }, + { + "epoch": 1.6824104234527688, + "grad_norm": 6.519773416264212, + "learning_rate": 1.5466382069599656e-05, + "loss": 0.4567, + "step": 3099 + }, + { + "epoch": 1.6829533116178066, + "grad_norm": 6.775917495491388, + "learning_rate": 1.5463436839667007e-05, + "loss": 0.6874, + "step": 3100 + }, + { + "epoch": 1.6834961997828448, + "grad_norm": 7.82569199258335, + "learning_rate": 1.5460490934007776e-05, + "loss": 0.819, + "step": 3101 + }, + { + "epoch": 1.6840390879478826, + "grad_norm": 6.885216490352139, + "learning_rate": 1.5457544352986326e-05, + "loss": 0.7891, + "step": 3102 + }, + { + "epoch": 1.6845819761129208, + "grad_norm": 6.733247427994312, + "learning_rate": 1.5454597096967093e-05, + "loss": 0.6664, + "step": 3103 + }, + { + "epoch": 1.6851248642779586, + "grad_norm": 6.96769902934235, + "learning_rate": 1.5451649166314598e-05, + "loss": 0.6578, + "step": 3104 + }, + { + "epoch": 1.6856677524429968, + "grad_norm": 6.941713803652074, + "learning_rate": 1.5448700561393444e-05, + "loss": 0.5855, + "step": 3105 + }, + { + "epoch": 1.6862106406080346, + "grad_norm": 8.203921934361416, + "learning_rate": 1.5445751282568324e-05, + "loss": 0.9843, + "step": 3106 + }, + { + "epoch": 1.6867535287730728, + "grad_norm": 8.822160140984163, + "learning_rate": 1.5442801330204004e-05, + "loss": 1.1003, + "step": 3107 + }, + { + "epoch": 1.6872964169381106, + "grad_norm": 8.330851912949297, + "learning_rate": 1.5439850704665338e-05, + "loss": 1.0992, + "step": 3108 + }, + { + "epoch": 1.6878393051031488, + "grad_norm": 6.552171568470756, + "learning_rate": 1.543689940631727e-05, + "loss": 0.5434, + "step": 3109 + }, + { + "epoch": 1.6883821932681866, + "grad_norm": 5.772049074992556, + "learning_rate": 1.5433947435524822e-05, + "loss": 0.7515, + "step": 3110 + }, + { + "epoch": 1.6889250814332248, + "grad_norm": 9.575440058135785, + "learning_rate": 1.543099479265309e-05, + "loss": 0.9691, + "step": 3111 + }, + { + "epoch": 1.6894679695982626, + "grad_norm": 6.008162314259628, + "learning_rate": 1.5428041478067263e-05, + "loss": 0.6137, + "step": 3112 + }, + { + "epoch": 1.6900108577633008, + "grad_norm": 8.993171311918207, + "learning_rate": 1.542508749213262e-05, + "loss": 0.7608, + "step": 3113 + }, + { + "epoch": 1.6905537459283386, + "grad_norm": 8.042583538721042, + "learning_rate": 1.542213283521451e-05, + "loss": 0.8188, + "step": 3114 + }, + { + "epoch": 1.6910966340933768, + "grad_norm": 9.290959640283983, + "learning_rate": 1.541917750767837e-05, + "loss": 1.1737, + "step": 3115 + }, + { + "epoch": 1.6916395222584146, + "grad_norm": 6.3883435583657535, + "learning_rate": 1.5416221509889718e-05, + "loss": 0.4376, + "step": 3116 + }, + { + "epoch": 1.6921824104234529, + "grad_norm": 8.943225081508276, + "learning_rate": 1.541326484221416e-05, + "loss": 0.9983, + "step": 3117 + }, + { + "epoch": 1.6927252985884906, + "grad_norm": 7.583355336491315, + "learning_rate": 1.541030750501737e-05, + "loss": 0.6078, + "step": 3118 + }, + { + "epoch": 1.6932681867535289, + "grad_norm": 7.318011543277784, + "learning_rate": 1.5407349498665133e-05, + "loss": 0.745, + "step": 3119 + }, + { + "epoch": 1.6938110749185666, + "grad_norm": 6.212465291912028, + "learning_rate": 1.5404390823523287e-05, + "loss": 0.4739, + "step": 3120 + }, + { + "epoch": 1.6943539630836049, + "grad_norm": 8.223585888775384, + "learning_rate": 1.5401431479957775e-05, + "loss": 1.1819, + "step": 3121 + }, + { + "epoch": 1.6948968512486426, + "grad_norm": 7.957232407413796, + "learning_rate": 1.5398471468334605e-05, + "loss": 0.7071, + "step": 3122 + }, + { + "epoch": 1.6954397394136809, + "grad_norm": 5.52597775233848, + "learning_rate": 1.5395510789019884e-05, + "loss": 0.6171, + "step": 3123 + }, + { + "epoch": 1.6959826275787186, + "grad_norm": 8.015216509333372, + "learning_rate": 1.5392549442379785e-05, + "loss": 0.9051, + "step": 3124 + }, + { + "epoch": 1.6965255157437569, + "grad_norm": 8.330470177699917, + "learning_rate": 1.538958742878058e-05, + "loss": 0.5037, + "step": 3125 + }, + { + "epoch": 1.6970684039087947, + "grad_norm": 6.1063862074204245, + "learning_rate": 1.538662474858861e-05, + "loss": 0.6509, + "step": 3126 + }, + { + "epoch": 1.6976112920738329, + "grad_norm": 8.11378323370098, + "learning_rate": 1.5383661402170308e-05, + "loss": 1.0356, + "step": 3127 + }, + { + "epoch": 1.6981541802388707, + "grad_norm": 5.995828293111682, + "learning_rate": 1.5380697389892185e-05, + "loss": 0.5297, + "step": 3128 + }, + { + "epoch": 1.6986970684039089, + "grad_norm": 6.962108385144619, + "learning_rate": 1.537773271212083e-05, + "loss": 0.8711, + "step": 3129 + }, + { + "epoch": 1.6992399565689467, + "grad_norm": 8.69611973474616, + "learning_rate": 1.5374767369222922e-05, + "loss": 0.911, + "step": 3130 + }, + { + "epoch": 1.6997828447339849, + "grad_norm": 11.047832664292853, + "learning_rate": 1.5371801361565223e-05, + "loss": 0.8499, + "step": 3131 + }, + { + "epoch": 1.7003257328990227, + "grad_norm": 10.490928296052365, + "learning_rate": 1.5368834689514568e-05, + "loss": 0.7617, + "step": 3132 + }, + { + "epoch": 1.700868621064061, + "grad_norm": 8.723324926351989, + "learning_rate": 1.536586735343788e-05, + "loss": 0.7158, + "step": 3133 + }, + { + "epoch": 1.7014115092290987, + "grad_norm": 7.780646168464086, + "learning_rate": 1.536289935370217e-05, + "loss": 0.6452, + "step": 3134 + }, + { + "epoch": 1.701954397394137, + "grad_norm": 8.037748179528911, + "learning_rate": 1.5359930690674518e-05, + "loss": 0.8368, + "step": 3135 + }, + { + "epoch": 1.7024972855591747, + "grad_norm": 7.539822669872962, + "learning_rate": 1.5356961364722096e-05, + "loss": 0.7224, + "step": 3136 + }, + { + "epoch": 1.703040173724213, + "grad_norm": 10.929660828744602, + "learning_rate": 1.5353991376212155e-05, + "loss": 1.2459, + "step": 3137 + }, + { + "epoch": 1.7035830618892507, + "grad_norm": 11.118929306326729, + "learning_rate": 1.5351020725512028e-05, + "loss": 1.2769, + "step": 3138 + }, + { + "epoch": 1.704125950054289, + "grad_norm": 8.820097629927263, + "learning_rate": 1.534804941298913e-05, + "loss": 1.0722, + "step": 3139 + }, + { + "epoch": 1.7046688382193267, + "grad_norm": 8.79991079654506, + "learning_rate": 1.5345077439010956e-05, + "loss": 0.7596, + "step": 3140 + }, + { + "epoch": 1.705211726384365, + "grad_norm": 7.5009896900263096, + "learning_rate": 1.5342104803945087e-05, + "loss": 0.6457, + "step": 3141 + }, + { + "epoch": 1.7057546145494027, + "grad_norm": 9.190121352339093, + "learning_rate": 1.533913150815918e-05, + "loss": 0.7152, + "step": 3142 + }, + { + "epoch": 1.706297502714441, + "grad_norm": 7.209026897095485, + "learning_rate": 1.5336157552020977e-05, + "loss": 0.534, + "step": 3143 + }, + { + "epoch": 1.7068403908794787, + "grad_norm": 8.752247270964865, + "learning_rate": 1.5333182935898306e-05, + "loss": 1.0953, + "step": 3144 + }, + { + "epoch": 1.707383279044517, + "grad_norm": 8.714563086288697, + "learning_rate": 1.5330207660159068e-05, + "loss": 0.8016, + "step": 3145 + }, + { + "epoch": 1.7079261672095547, + "grad_norm": 9.540320185420228, + "learning_rate": 1.5327231725171255e-05, + "loss": 1.2679, + "step": 3146 + }, + { + "epoch": 1.708469055374593, + "grad_norm": 8.091263210809633, + "learning_rate": 1.532425513130293e-05, + "loss": 0.7521, + "step": 3147 + }, + { + "epoch": 1.7090119435396307, + "grad_norm": 10.062104841108379, + "learning_rate": 1.5321277878922246e-05, + "loss": 1.0696, + "step": 3148 + }, + { + "epoch": 1.709554831704669, + "grad_norm": 8.184716628635337, + "learning_rate": 1.531829996839743e-05, + "loss": 0.643, + "step": 3149 + }, + { + "epoch": 1.7100977198697067, + "grad_norm": 11.130045309860368, + "learning_rate": 1.53153214000968e-05, + "loss": 1.2565, + "step": 3150 + }, + { + "epoch": 1.710640608034745, + "grad_norm": 8.856470164658607, + "learning_rate": 1.5312342174388746e-05, + "loss": 0.8376, + "step": 3151 + }, + { + "epoch": 1.7111834961997827, + "grad_norm": 11.958863336707164, + "learning_rate": 1.5309362291641747e-05, + "loss": 0.978, + "step": 3152 + }, + { + "epoch": 1.711726384364821, + "grad_norm": 10.369392130776264, + "learning_rate": 1.5306381752224357e-05, + "loss": 1.1822, + "step": 3153 + }, + { + "epoch": 1.7122692725298587, + "grad_norm": 7.346196971786586, + "learning_rate": 1.5303400556505213e-05, + "loss": 0.9284, + "step": 3154 + }, + { + "epoch": 1.712812160694897, + "grad_norm": 6.652184209108613, + "learning_rate": 1.5300418704853042e-05, + "loss": 0.5787, + "step": 3155 + }, + { + "epoch": 1.7133550488599347, + "grad_norm": 9.745597378440513, + "learning_rate": 1.5297436197636634e-05, + "loss": 1.0194, + "step": 3156 + }, + { + "epoch": 1.713897937024973, + "grad_norm": 6.17200933356452, + "learning_rate": 1.5294453035224874e-05, + "loss": 0.6315, + "step": 3157 + }, + { + "epoch": 1.7144408251900107, + "grad_norm": 8.373044157038668, + "learning_rate": 1.5291469217986724e-05, + "loss": 0.9127, + "step": 3158 + }, + { + "epoch": 1.714983713355049, + "grad_norm": 9.811623309823384, + "learning_rate": 1.5288484746291227e-05, + "loss": 0.7791, + "step": 3159 + }, + { + "epoch": 1.7155266015200867, + "grad_norm": 9.264521529810194, + "learning_rate": 1.5285499620507513e-05, + "loss": 0.6764, + "step": 3160 + }, + { + "epoch": 1.716069489685125, + "grad_norm": 8.688562646594697, + "learning_rate": 1.5282513841004777e-05, + "loss": 1.0202, + "step": 3161 + }, + { + "epoch": 1.7166123778501627, + "grad_norm": 8.522452395448054, + "learning_rate": 1.527952740815231e-05, + "loss": 1.0043, + "step": 3162 + }, + { + "epoch": 1.717155266015201, + "grad_norm": 6.330008909603884, + "learning_rate": 1.527654032231948e-05, + "loss": 0.6651, + "step": 3163 + }, + { + "epoch": 1.7176981541802387, + "grad_norm": 9.314495623946685, + "learning_rate": 1.5273552583875736e-05, + "loss": 0.8312, + "step": 3164 + }, + { + "epoch": 1.718241042345277, + "grad_norm": 9.157480690732486, + "learning_rate": 1.52705641931906e-05, + "loss": 1.0017, + "step": 3165 + }, + { + "epoch": 1.7187839305103148, + "grad_norm": 8.300251884517943, + "learning_rate": 1.5267575150633687e-05, + "loss": 0.9929, + "step": 3166 + }, + { + "epoch": 1.719326818675353, + "grad_norm": 8.328644452020212, + "learning_rate": 1.5264585456574684e-05, + "loss": 0.8105, + "step": 3167 + }, + { + "epoch": 1.7198697068403908, + "grad_norm": 9.499775808882656, + "learning_rate": 1.526159511138336e-05, + "loss": 0.8593, + "step": 3168 + }, + { + "epoch": 1.720412595005429, + "grad_norm": 7.684886832360556, + "learning_rate": 1.5258604115429567e-05, + "loss": 0.5919, + "step": 3169 + }, + { + "epoch": 1.7209554831704668, + "grad_norm": 8.666869336691208, + "learning_rate": 1.5255612469083239e-05, + "loss": 0.6537, + "step": 3170 + }, + { + "epoch": 1.721498371335505, + "grad_norm": 11.871617471990923, + "learning_rate": 1.5252620172714378e-05, + "loss": 1.3826, + "step": 3171 + }, + { + "epoch": 1.7220412595005428, + "grad_norm": 6.088995555337644, + "learning_rate": 1.5249627226693089e-05, + "loss": 0.5085, + "step": 3172 + }, + { + "epoch": 1.722584147665581, + "grad_norm": 9.255760547122362, + "learning_rate": 1.5246633631389536e-05, + "loss": 0.8823, + "step": 3173 + }, + { + "epoch": 1.7231270358306188, + "grad_norm": 9.91965748862537, + "learning_rate": 1.5243639387173974e-05, + "loss": 0.8752, + "step": 3174 + }, + { + "epoch": 1.723669923995657, + "grad_norm": 6.397561552165108, + "learning_rate": 1.5240644494416734e-05, + "loss": 0.6673, + "step": 3175 + }, + { + "epoch": 1.7242128121606948, + "grad_norm": 7.066817871358342, + "learning_rate": 1.523764895348823e-05, + "loss": 1.041, + "step": 3176 + }, + { + "epoch": 1.724755700325733, + "grad_norm": 8.282830238008364, + "learning_rate": 1.5234652764758959e-05, + "loss": 0.9521, + "step": 3177 + }, + { + "epoch": 1.7252985884907708, + "grad_norm": 7.611547937947443, + "learning_rate": 1.523165592859949e-05, + "loss": 0.7182, + "step": 3178 + }, + { + "epoch": 1.725841476655809, + "grad_norm": 8.068698205768472, + "learning_rate": 1.5228658445380475e-05, + "loss": 0.9078, + "step": 3179 + }, + { + "epoch": 1.7263843648208468, + "grad_norm": 8.731772184014808, + "learning_rate": 1.5225660315472652e-05, + "loss": 1.2148, + "step": 3180 + }, + { + "epoch": 1.726927252985885, + "grad_norm": 7.785982141134894, + "learning_rate": 1.5222661539246832e-05, + "loss": 0.8653, + "step": 3181 + }, + { + "epoch": 1.7274701411509228, + "grad_norm": 7.6267653051373285, + "learning_rate": 1.521966211707391e-05, + "loss": 0.7454, + "step": 3182 + }, + { + "epoch": 1.728013029315961, + "grad_norm": 7.440711922406893, + "learning_rate": 1.521666204932486e-05, + "loss": 0.6576, + "step": 3183 + }, + { + "epoch": 1.7285559174809988, + "grad_norm": 7.69035147644333, + "learning_rate": 1.521366133637073e-05, + "loss": 0.8269, + "step": 3184 + }, + { + "epoch": 1.729098805646037, + "grad_norm": 8.387844627258719, + "learning_rate": 1.5210659978582662e-05, + "loss": 1.1049, + "step": 3185 + }, + { + "epoch": 1.7296416938110748, + "grad_norm": 9.08085381726355, + "learning_rate": 1.5207657976331862e-05, + "loss": 0.97, + "step": 3186 + }, + { + "epoch": 1.730184581976113, + "grad_norm": 7.621613071814458, + "learning_rate": 1.5204655329989617e-05, + "loss": 0.7113, + "step": 3187 + }, + { + "epoch": 1.7307274701411508, + "grad_norm": 6.743838306403499, + "learning_rate": 1.5201652039927313e-05, + "loss": 0.7307, + "step": 3188 + }, + { + "epoch": 1.731270358306189, + "grad_norm": 9.602658786805504, + "learning_rate": 1.5198648106516392e-05, + "loss": 1.0515, + "step": 3189 + }, + { + "epoch": 1.7318132464712268, + "grad_norm": 8.153765987898234, + "learning_rate": 1.5195643530128387e-05, + "loss": 1.0879, + "step": 3190 + }, + { + "epoch": 1.732356134636265, + "grad_norm": 9.745244927474419, + "learning_rate": 1.519263831113491e-05, + "loss": 0.9542, + "step": 3191 + }, + { + "epoch": 1.7328990228013028, + "grad_norm": 8.908425729214914, + "learning_rate": 1.5189632449907654e-05, + "loss": 0.8936, + "step": 3192 + }, + { + "epoch": 1.733441910966341, + "grad_norm": 7.409065356665995, + "learning_rate": 1.5186625946818382e-05, + "loss": 0.7645, + "step": 3193 + }, + { + "epoch": 1.7339847991313788, + "grad_norm": 7.958784484679066, + "learning_rate": 1.5183618802238949e-05, + "loss": 0.8512, + "step": 3194 + }, + { + "epoch": 1.734527687296417, + "grad_norm": 5.808082449761269, + "learning_rate": 1.5180611016541278e-05, + "loss": 0.7241, + "step": 3195 + }, + { + "epoch": 1.7350705754614548, + "grad_norm": 8.88976614296984, + "learning_rate": 1.5177602590097382e-05, + "loss": 1.0506, + "step": 3196 + }, + { + "epoch": 1.735613463626493, + "grad_norm": 8.197925091983134, + "learning_rate": 1.5174593523279346e-05, + "loss": 0.7959, + "step": 3197 + }, + { + "epoch": 1.7361563517915308, + "grad_norm": 10.2752921216035, + "learning_rate": 1.5171583816459334e-05, + "loss": 1.2859, + "step": 3198 + }, + { + "epoch": 1.736699239956569, + "grad_norm": 9.43498245507836, + "learning_rate": 1.5168573470009596e-05, + "loss": 1.0459, + "step": 3199 + }, + { + "epoch": 1.7372421281216068, + "grad_norm": 9.033995794822129, + "learning_rate": 1.516556248430245e-05, + "loss": 1.0003, + "step": 3200 + }, + { + "epoch": 1.737785016286645, + "grad_norm": 9.577836032406582, + "learning_rate": 1.5162550859710306e-05, + "loss": 1.2394, + "step": 3201 + }, + { + "epoch": 1.7383279044516828, + "grad_norm": 5.2063230201693775, + "learning_rate": 1.5159538596605642e-05, + "loss": 0.3625, + "step": 3202 + }, + { + "epoch": 1.738870792616721, + "grad_norm": 7.741298457702819, + "learning_rate": 1.5156525695361022e-05, + "loss": 0.6744, + "step": 3203 + }, + { + "epoch": 1.7394136807817588, + "grad_norm": 8.718382461609222, + "learning_rate": 1.5153512156349086e-05, + "loss": 1.0423, + "step": 3204 + }, + { + "epoch": 1.739956568946797, + "grad_norm": 6.706656901306365, + "learning_rate": 1.5150497979942556e-05, + "loss": 0.5121, + "step": 3205 + }, + { + "epoch": 1.7404994571118348, + "grad_norm": 9.934397785927217, + "learning_rate": 1.5147483166514225e-05, + "loss": 1.5269, + "step": 3206 + }, + { + "epoch": 1.741042345276873, + "grad_norm": 8.069224944733923, + "learning_rate": 1.514446771643697e-05, + "loss": 0.7801, + "step": 3207 + }, + { + "epoch": 1.7415852334419109, + "grad_norm": 9.193161021474104, + "learning_rate": 1.514145163008375e-05, + "loss": 1.2144, + "step": 3208 + }, + { + "epoch": 1.742128121606949, + "grad_norm": 6.5647281089901, + "learning_rate": 1.5138434907827599e-05, + "loss": 0.6692, + "step": 3209 + }, + { + "epoch": 1.7426710097719869, + "grad_norm": 9.387942500806554, + "learning_rate": 1.513541755004163e-05, + "loss": 1.1655, + "step": 3210 + }, + { + "epoch": 1.743213897937025, + "grad_norm": 6.936398185140981, + "learning_rate": 1.5132399557099031e-05, + "loss": 0.7761, + "step": 3211 + }, + { + "epoch": 1.7437567861020629, + "grad_norm": 8.699280554503527, + "learning_rate": 1.5129380929373076e-05, + "loss": 1.2891, + "step": 3212 + }, + { + "epoch": 1.744299674267101, + "grad_norm": 11.932192326740894, + "learning_rate": 1.5126361667237116e-05, + "loss": 0.9868, + "step": 3213 + }, + { + "epoch": 1.7448425624321389, + "grad_norm": 8.456694582402795, + "learning_rate": 1.512334177106457e-05, + "loss": 1.199, + "step": 3214 + }, + { + "epoch": 1.745385450597177, + "grad_norm": 8.284979161281417, + "learning_rate": 1.512032124122895e-05, + "loss": 0.658, + "step": 3215 + }, + { + "epoch": 1.7459283387622149, + "grad_norm": 8.490755188379255, + "learning_rate": 1.5117300078103841e-05, + "loss": 0.6974, + "step": 3216 + }, + { + "epoch": 1.746471226927253, + "grad_norm": 14.621603534181931, + "learning_rate": 1.5114278282062898e-05, + "loss": 1.4386, + "step": 3217 + }, + { + "epoch": 1.7470141150922909, + "grad_norm": 8.35853855400652, + "learning_rate": 1.511125585347987e-05, + "loss": 0.7991, + "step": 3218 + }, + { + "epoch": 1.747557003257329, + "grad_norm": 10.14192231306983, + "learning_rate": 1.5108232792728567e-05, + "loss": 0.9942, + "step": 3219 + }, + { + "epoch": 1.7480998914223669, + "grad_norm": 6.003113555943506, + "learning_rate": 1.5105209100182893e-05, + "loss": 0.5614, + "step": 3220 + }, + { + "epoch": 1.748642779587405, + "grad_norm": 7.420823995167257, + "learning_rate": 1.5102184776216824e-05, + "loss": 0.6918, + "step": 3221 + }, + { + "epoch": 1.7491856677524429, + "grad_norm": 7.3724702610256, + "learning_rate": 1.5099159821204406e-05, + "loss": 0.815, + "step": 3222 + }, + { + "epoch": 1.749728555917481, + "grad_norm": 8.539056209113738, + "learning_rate": 1.509613423551978e-05, + "loss": 1.0859, + "step": 3223 + }, + { + "epoch": 1.750271444082519, + "grad_norm": 10.518635534148203, + "learning_rate": 1.5093108019537147e-05, + "loss": 1.1873, + "step": 3224 + }, + { + "epoch": 1.7508143322475571, + "grad_norm": 8.049010262043616, + "learning_rate": 1.5090081173630795e-05, + "loss": 0.971, + "step": 3225 + }, + { + "epoch": 1.751357220412595, + "grad_norm": 7.874724964543097, + "learning_rate": 1.5087053698175092e-05, + "loss": 1.06, + "step": 3226 + }, + { + "epoch": 1.7519001085776331, + "grad_norm": 8.018813791724389, + "learning_rate": 1.508402559354448e-05, + "loss": 0.7192, + "step": 3227 + }, + { + "epoch": 1.752442996742671, + "grad_norm": 11.748757549078345, + "learning_rate": 1.508099686011348e-05, + "loss": 1.4476, + "step": 3228 + }, + { + "epoch": 1.7529858849077091, + "grad_norm": 7.885344513181394, + "learning_rate": 1.507796749825669e-05, + "loss": 0.7493, + "step": 3229 + }, + { + "epoch": 1.753528773072747, + "grad_norm": 6.808213775155858, + "learning_rate": 1.507493750834879e-05, + "loss": 0.6646, + "step": 3230 + }, + { + "epoch": 1.7540716612377851, + "grad_norm": 6.304489454276083, + "learning_rate": 1.5071906890764527e-05, + "loss": 0.7375, + "step": 3231 + }, + { + "epoch": 1.754614549402823, + "grad_norm": 8.217636823407403, + "learning_rate": 1.5068875645878739e-05, + "loss": 1.0069, + "step": 3232 + }, + { + "epoch": 1.7551574375678611, + "grad_norm": 10.706707510999644, + "learning_rate": 1.5065843774066329e-05, + "loss": 0.689, + "step": 3233 + }, + { + "epoch": 1.755700325732899, + "grad_norm": 7.600175930285094, + "learning_rate": 1.5062811275702291e-05, + "loss": 0.8645, + "step": 3234 + }, + { + "epoch": 1.7562432138979371, + "grad_norm": 8.463001355349205, + "learning_rate": 1.5059778151161684e-05, + "loss": 0.7185, + "step": 3235 + }, + { + "epoch": 1.756786102062975, + "grad_norm": 5.834156873881953, + "learning_rate": 1.5056744400819651e-05, + "loss": 0.6486, + "step": 3236 + }, + { + "epoch": 1.7573289902280131, + "grad_norm": 7.152665898997485, + "learning_rate": 1.5053710025051411e-05, + "loss": 0.9116, + "step": 3237 + }, + { + "epoch": 1.757871878393051, + "grad_norm": 8.037482464894001, + "learning_rate": 1.5050675024232262e-05, + "loss": 0.8231, + "step": 3238 + }, + { + "epoch": 1.7584147665580891, + "grad_norm": 10.416359782719013, + "learning_rate": 1.5047639398737573e-05, + "loss": 1.2555, + "step": 3239 + }, + { + "epoch": 1.758957654723127, + "grad_norm": 6.789336798229921, + "learning_rate": 1.50446031489428e-05, + "loss": 0.7295, + "step": 3240 + }, + { + "epoch": 1.7595005428881652, + "grad_norm": 9.927450801451293, + "learning_rate": 1.5041566275223472e-05, + "loss": 1.5417, + "step": 3241 + }, + { + "epoch": 1.760043431053203, + "grad_norm": 7.040644062299268, + "learning_rate": 1.5038528777955188e-05, + "loss": 0.4495, + "step": 3242 + }, + { + "epoch": 1.7605863192182412, + "grad_norm": 8.848733486454524, + "learning_rate": 1.5035490657513638e-05, + "loss": 1.0123, + "step": 3243 + }, + { + "epoch": 1.761129207383279, + "grad_norm": 8.109736914130819, + "learning_rate": 1.503245191427458e-05, + "loss": 0.8984, + "step": 3244 + }, + { + "epoch": 1.7616720955483172, + "grad_norm": 7.210785707597705, + "learning_rate": 1.5029412548613845e-05, + "loss": 0.6947, + "step": 3245 + }, + { + "epoch": 1.762214983713355, + "grad_norm": 10.368251730052483, + "learning_rate": 1.502637256090735e-05, + "loss": 0.8413, + "step": 3246 + }, + { + "epoch": 1.7627578718783932, + "grad_norm": 6.770955437649876, + "learning_rate": 1.5023331951531086e-05, + "loss": 0.6386, + "step": 3247 + }, + { + "epoch": 1.763300760043431, + "grad_norm": 10.818384833705863, + "learning_rate": 1.5020290720861122e-05, + "loss": 1.2544, + "step": 3248 + }, + { + "epoch": 1.7638436482084692, + "grad_norm": 7.9104385792951195, + "learning_rate": 1.50172488692736e-05, + "loss": 0.6538, + "step": 3249 + }, + { + "epoch": 1.764386536373507, + "grad_norm": 9.565689306609565, + "learning_rate": 1.5014206397144742e-05, + "loss": 1.1232, + "step": 3250 + }, + { + "epoch": 1.7649294245385452, + "grad_norm": 11.651444088862487, + "learning_rate": 1.5011163304850844e-05, + "loss": 1.2054, + "step": 3251 + }, + { + "epoch": 1.765472312703583, + "grad_norm": 10.326063989708704, + "learning_rate": 1.500811959276828e-05, + "loss": 0.9441, + "step": 3252 + }, + { + "epoch": 1.7660152008686212, + "grad_norm": 8.0529225996154, + "learning_rate": 1.5005075261273506e-05, + "loss": 0.9431, + "step": 3253 + }, + { + "epoch": 1.766558089033659, + "grad_norm": 7.391183931496241, + "learning_rate": 1.500203031074305e-05, + "loss": 0.8619, + "step": 3254 + }, + { + "epoch": 1.7671009771986972, + "grad_norm": 6.892445188811786, + "learning_rate": 1.4998984741553508e-05, + "loss": 0.7039, + "step": 3255 + }, + { + "epoch": 1.767643865363735, + "grad_norm": 6.541771639151763, + "learning_rate": 1.4995938554081568e-05, + "loss": 0.6359, + "step": 3256 + }, + { + "epoch": 1.7681867535287732, + "grad_norm": 7.690176361123399, + "learning_rate": 1.4992891748703985e-05, + "loss": 0.8998, + "step": 3257 + }, + { + "epoch": 1.768729641693811, + "grad_norm": 6.497301879748706, + "learning_rate": 1.498984432579759e-05, + "loss": 0.5743, + "step": 3258 + }, + { + "epoch": 1.7692725298588492, + "grad_norm": 6.656723951695686, + "learning_rate": 1.4986796285739298e-05, + "loss": 0.7002, + "step": 3259 + }, + { + "epoch": 1.769815418023887, + "grad_norm": 7.9129397206513685, + "learning_rate": 1.4983747628906095e-05, + "loss": 0.9239, + "step": 3260 + }, + { + "epoch": 1.7703583061889252, + "grad_norm": 4.544544755639526, + "learning_rate": 1.4980698355675043e-05, + "loss": 0.3431, + "step": 3261 + }, + { + "epoch": 1.770901194353963, + "grad_norm": 8.243743590575315, + "learning_rate": 1.4977648466423278e-05, + "loss": 1.1386, + "step": 3262 + }, + { + "epoch": 1.7714440825190012, + "grad_norm": 9.404041237838763, + "learning_rate": 1.4974597961528021e-05, + "loss": 0.9087, + "step": 3263 + }, + { + "epoch": 1.771986970684039, + "grad_norm": 8.534670509056685, + "learning_rate": 1.4971546841366556e-05, + "loss": 0.8769, + "step": 3264 + }, + { + "epoch": 1.7725298588490772, + "grad_norm": 9.442139570892468, + "learning_rate": 1.4968495106316254e-05, + "loss": 0.6942, + "step": 3265 + }, + { + "epoch": 1.773072747014115, + "grad_norm": 8.449365139837813, + "learning_rate": 1.496544275675456e-05, + "loss": 1.0743, + "step": 3266 + }, + { + "epoch": 1.7736156351791532, + "grad_norm": 9.461545898435599, + "learning_rate": 1.4962389793058994e-05, + "loss": 0.7473, + "step": 3267 + }, + { + "epoch": 1.774158523344191, + "grad_norm": 11.871672085414737, + "learning_rate": 1.4959336215607147e-05, + "loss": 1.5433, + "step": 3268 + }, + { + "epoch": 1.7747014115092292, + "grad_norm": 6.417686676489462, + "learning_rate": 1.4956282024776693e-05, + "loss": 0.5006, + "step": 3269 + }, + { + "epoch": 1.775244299674267, + "grad_norm": 10.49161190534651, + "learning_rate": 1.4953227220945381e-05, + "loss": 0.9918, + "step": 3270 + }, + { + "epoch": 1.7757871878393052, + "grad_norm": 9.298743913365943, + "learning_rate": 1.4950171804491031e-05, + "loss": 1.3149, + "step": 3271 + }, + { + "epoch": 1.776330076004343, + "grad_norm": 7.775473503246929, + "learning_rate": 1.4947115775791541e-05, + "loss": 0.6873, + "step": 3272 + }, + { + "epoch": 1.7768729641693812, + "grad_norm": 6.75591368497042, + "learning_rate": 1.4944059135224891e-05, + "loss": 0.7799, + "step": 3273 + }, + { + "epoch": 1.777415852334419, + "grad_norm": 8.234180005996857, + "learning_rate": 1.4941001883169124e-05, + "loss": 1.0393, + "step": 3274 + }, + { + "epoch": 1.7779587404994572, + "grad_norm": 8.419005181659063, + "learning_rate": 1.4937944020002371e-05, + "loss": 0.6798, + "step": 3275 + }, + { + "epoch": 1.778501628664495, + "grad_norm": 7.0982600255490205, + "learning_rate": 1.4934885546102833e-05, + "loss": 0.4936, + "step": 3276 + }, + { + "epoch": 1.7790445168295332, + "grad_norm": 8.404626943584326, + "learning_rate": 1.4931826461848785e-05, + "loss": 1.0507, + "step": 3277 + }, + { + "epoch": 1.779587404994571, + "grad_norm": 5.861882830901275, + "learning_rate": 1.4928766767618576e-05, + "loss": 0.6785, + "step": 3278 + }, + { + "epoch": 1.7801302931596092, + "grad_norm": 8.114055102204162, + "learning_rate": 1.4925706463790642e-05, + "loss": 0.765, + "step": 3279 + }, + { + "epoch": 1.780673181324647, + "grad_norm": 10.033269670592238, + "learning_rate": 1.492264555074348e-05, + "loss": 0.9163, + "step": 3280 + }, + { + "epoch": 1.7812160694896852, + "grad_norm": 7.040755628619449, + "learning_rate": 1.4919584028855671e-05, + "loss": 0.5962, + "step": 3281 + }, + { + "epoch": 1.781758957654723, + "grad_norm": 12.093987616559675, + "learning_rate": 1.4916521898505872e-05, + "loss": 0.8999, + "step": 3282 + }, + { + "epoch": 1.7823018458197613, + "grad_norm": 9.400200287574112, + "learning_rate": 1.4913459160072805e-05, + "loss": 1.0203, + "step": 3283 + }, + { + "epoch": 1.782844733984799, + "grad_norm": 8.605701710844945, + "learning_rate": 1.4910395813935279e-05, + "loss": 0.7874, + "step": 3284 + }, + { + "epoch": 1.7833876221498373, + "grad_norm": 10.279849472300688, + "learning_rate": 1.4907331860472174e-05, + "loss": 1.4306, + "step": 3285 + }, + { + "epoch": 1.783930510314875, + "grad_norm": 10.582881297295359, + "learning_rate": 1.4904267300062443e-05, + "loss": 0.6536, + "step": 3286 + }, + { + "epoch": 1.7844733984799133, + "grad_norm": 6.960885594125151, + "learning_rate": 1.4901202133085115e-05, + "loss": 0.7256, + "step": 3287 + }, + { + "epoch": 1.785016286644951, + "grad_norm": 9.088589913670967, + "learning_rate": 1.4898136359919298e-05, + "loss": 0.9299, + "step": 3288 + }, + { + "epoch": 1.7855591748099893, + "grad_norm": 8.922355254191798, + "learning_rate": 1.4895069980944168e-05, + "loss": 0.8159, + "step": 3289 + }, + { + "epoch": 1.786102062975027, + "grad_norm": 9.459478678151763, + "learning_rate": 1.4892002996538982e-05, + "loss": 1.132, + "step": 3290 + }, + { + "epoch": 1.7866449511400653, + "grad_norm": 9.746458814865415, + "learning_rate": 1.4888935407083065e-05, + "loss": 0.9874, + "step": 3291 + }, + { + "epoch": 1.787187839305103, + "grad_norm": 8.13065950431162, + "learning_rate": 1.4885867212955827e-05, + "loss": 0.8013, + "step": 3292 + }, + { + "epoch": 1.7877307274701413, + "grad_norm": 8.469962208509646, + "learning_rate": 1.4882798414536749e-05, + "loss": 1.0408, + "step": 3293 + }, + { + "epoch": 1.788273615635179, + "grad_norm": 9.018264603768202, + "learning_rate": 1.4879729012205379e-05, + "loss": 0.8324, + "step": 3294 + }, + { + "epoch": 1.7888165038002173, + "grad_norm": 8.934092671110381, + "learning_rate": 1.4876659006341347e-05, + "loss": 0.8842, + "step": 3295 + }, + { + "epoch": 1.789359391965255, + "grad_norm": 7.612464888658943, + "learning_rate": 1.4873588397324356e-05, + "loss": 0.9356, + "step": 3296 + }, + { + "epoch": 1.7899022801302933, + "grad_norm": 7.303452266512586, + "learning_rate": 1.4870517185534184e-05, + "loss": 1.0291, + "step": 3297 + }, + { + "epoch": 1.790445168295331, + "grad_norm": 9.805172223903986, + "learning_rate": 1.4867445371350687e-05, + "loss": 1.0243, + "step": 3298 + }, + { + "epoch": 1.7909880564603693, + "grad_norm": 7.742224062288237, + "learning_rate": 1.4864372955153785e-05, + "loss": 0.7173, + "step": 3299 + }, + { + "epoch": 1.791530944625407, + "grad_norm": 9.174714938761847, + "learning_rate": 1.4861299937323485e-05, + "loss": 0.9721, + "step": 3300 + }, + { + "epoch": 1.7920738327904453, + "grad_norm": 8.116980821679304, + "learning_rate": 1.4858226318239864e-05, + "loss": 0.7776, + "step": 3301 + }, + { + "epoch": 1.792616720955483, + "grad_norm": 5.885386861657401, + "learning_rate": 1.4855152098283068e-05, + "loss": 0.3833, + "step": 3302 + }, + { + "epoch": 1.7931596091205213, + "grad_norm": 6.804177263707199, + "learning_rate": 1.4852077277833319e-05, + "loss": 0.7972, + "step": 3303 + }, + { + "epoch": 1.793702497285559, + "grad_norm": 8.874804659745028, + "learning_rate": 1.4849001857270922e-05, + "loss": 1.063, + "step": 3304 + }, + { + "epoch": 1.7942453854505973, + "grad_norm": 10.231899453600786, + "learning_rate": 1.4845925836976247e-05, + "loss": 1.6617, + "step": 3305 + }, + { + "epoch": 1.794788273615635, + "grad_norm": 9.067079935451718, + "learning_rate": 1.4842849217329742e-05, + "loss": 1.1062, + "step": 3306 + }, + { + "epoch": 1.7953311617806733, + "grad_norm": 9.5261315220339, + "learning_rate": 1.4839771998711925e-05, + "loss": 1.5991, + "step": 3307 + }, + { + "epoch": 1.795874049945711, + "grad_norm": 9.53442504374803, + "learning_rate": 1.4836694181503397e-05, + "loss": 0.9267, + "step": 3308 + }, + { + "epoch": 1.7964169381107493, + "grad_norm": 8.231497735162955, + "learning_rate": 1.4833615766084821e-05, + "loss": 0.9802, + "step": 3309 + }, + { + "epoch": 1.796959826275787, + "grad_norm": 7.2171054452276815, + "learning_rate": 1.4830536752836944e-05, + "loss": 0.6116, + "step": 3310 + }, + { + "epoch": 1.7975027144408253, + "grad_norm": 8.412637332813022, + "learning_rate": 1.4827457142140583e-05, + "loss": 0.7482, + "step": 3311 + }, + { + "epoch": 1.798045602605863, + "grad_norm": 6.24239912297862, + "learning_rate": 1.482437693437663e-05, + "loss": 0.4333, + "step": 3312 + }, + { + "epoch": 1.7985884907709013, + "grad_norm": 9.706212881828902, + "learning_rate": 1.4821296129926049e-05, + "loss": 1.3862, + "step": 3313 + }, + { + "epoch": 1.799131378935939, + "grad_norm": 8.275124350006633, + "learning_rate": 1.4818214729169878e-05, + "loss": 0.812, + "step": 3314 + }, + { + "epoch": 1.7996742671009773, + "grad_norm": 7.0810692617964275, + "learning_rate": 1.4815132732489228e-05, + "loss": 0.8919, + "step": 3315 + }, + { + "epoch": 1.8002171552660151, + "grad_norm": 7.225811411306636, + "learning_rate": 1.4812050140265288e-05, + "loss": 0.7512, + "step": 3316 + }, + { + "epoch": 1.8007600434310533, + "grad_norm": 6.458874811050634, + "learning_rate": 1.4808966952879318e-05, + "loss": 0.6914, + "step": 3317 + }, + { + "epoch": 1.8013029315960911, + "grad_norm": 6.537680447766501, + "learning_rate": 1.4805883170712653e-05, + "loss": 0.5511, + "step": 3318 + }, + { + "epoch": 1.8018458197611293, + "grad_norm": 8.023777979063889, + "learning_rate": 1.4802798794146695e-05, + "loss": 0.6814, + "step": 3319 + }, + { + "epoch": 1.8023887079261671, + "grad_norm": 13.062126505855263, + "learning_rate": 1.4799713823562932e-05, + "loss": 1.0526, + "step": 3320 + }, + { + "epoch": 1.8029315960912053, + "grad_norm": 8.173027381817493, + "learning_rate": 1.479662825934291e-05, + "loss": 0.8939, + "step": 3321 + }, + { + "epoch": 1.8034744842562431, + "grad_norm": 7.465360921787222, + "learning_rate": 1.4793542101868261e-05, + "loss": 0.5965, + "step": 3322 + }, + { + "epoch": 1.8040173724212814, + "grad_norm": 6.5098081549027835, + "learning_rate": 1.4790455351520684e-05, + "loss": 0.5963, + "step": 3323 + }, + { + "epoch": 1.8045602605863191, + "grad_norm": 11.071700718231854, + "learning_rate": 1.4787368008681956e-05, + "loss": 0.9728, + "step": 3324 + }, + { + "epoch": 1.8051031487513574, + "grad_norm": 9.251168051579464, + "learning_rate": 1.4784280073733924e-05, + "loss": 0.8816, + "step": 3325 + }, + { + "epoch": 1.8056460369163951, + "grad_norm": 7.876824650963893, + "learning_rate": 1.4781191547058505e-05, + "loss": 1.0365, + "step": 3326 + }, + { + "epoch": 1.8061889250814334, + "grad_norm": 7.984254709859244, + "learning_rate": 1.4778102429037696e-05, + "loss": 0.7252, + "step": 3327 + }, + { + "epoch": 1.8067318132464711, + "grad_norm": 8.162664097932598, + "learning_rate": 1.4775012720053563e-05, + "loss": 0.8297, + "step": 3328 + }, + { + "epoch": 1.8072747014115094, + "grad_norm": 6.631733087479191, + "learning_rate": 1.4771922420488245e-05, + "loss": 0.6962, + "step": 3329 + }, + { + "epoch": 1.8078175895765471, + "grad_norm": 8.80250228426254, + "learning_rate": 1.476883153072396e-05, + "loss": 1.0576, + "step": 3330 + }, + { + "epoch": 1.8083604777415854, + "grad_norm": 10.637063152950418, + "learning_rate": 1.476574005114299e-05, + "loss": 0.8866, + "step": 3331 + }, + { + "epoch": 1.8089033659066232, + "grad_norm": 9.58332512337787, + "learning_rate": 1.4762647982127696e-05, + "loss": 1.6148, + "step": 3332 + }, + { + "epoch": 1.8094462540716614, + "grad_norm": 7.826567769007031, + "learning_rate": 1.4759555324060508e-05, + "loss": 0.9256, + "step": 3333 + }, + { + "epoch": 1.8099891422366992, + "grad_norm": 9.445307614493151, + "learning_rate": 1.4756462077323931e-05, + "loss": 1.1389, + "step": 3334 + }, + { + "epoch": 1.8105320304017374, + "grad_norm": 7.504505604214132, + "learning_rate": 1.4753368242300543e-05, + "loss": 1.1333, + "step": 3335 + }, + { + "epoch": 1.8110749185667752, + "grad_norm": 6.996150668976886, + "learning_rate": 1.4750273819372996e-05, + "loss": 0.6721, + "step": 3336 + }, + { + "epoch": 1.8116178067318134, + "grad_norm": 7.593185686378431, + "learning_rate": 1.4747178808924012e-05, + "loss": 0.6856, + "step": 3337 + }, + { + "epoch": 1.8121606948968512, + "grad_norm": 7.277948846492979, + "learning_rate": 1.474408321133639e-05, + "loss": 0.5487, + "step": 3338 + }, + { + "epoch": 1.8127035830618894, + "grad_norm": 9.500801329454344, + "learning_rate": 1.474098702699299e-05, + "loss": 1.063, + "step": 3339 + }, + { + "epoch": 1.8132464712269272, + "grad_norm": 7.732920511283457, + "learning_rate": 1.4737890256276763e-05, + "loss": 0.8875, + "step": 3340 + }, + { + "epoch": 1.8137893593919654, + "grad_norm": 9.56546309643246, + "learning_rate": 1.4734792899570716e-05, + "loss": 0.7724, + "step": 3341 + }, + { + "epoch": 1.8143322475570032, + "grad_norm": 7.735046028758256, + "learning_rate": 1.4731694957257938e-05, + "loss": 0.8498, + "step": 3342 + }, + { + "epoch": 1.8148751357220414, + "grad_norm": 9.077413763001829, + "learning_rate": 1.4728596429721587e-05, + "loss": 1.1604, + "step": 3343 + }, + { + "epoch": 1.8154180238870792, + "grad_norm": 8.33530084886159, + "learning_rate": 1.4725497317344892e-05, + "loss": 1.0126, + "step": 3344 + }, + { + "epoch": 1.8159609120521174, + "grad_norm": 11.938749223535789, + "learning_rate": 1.4722397620511158e-05, + "loss": 0.9538, + "step": 3345 + }, + { + "epoch": 1.8165038002171552, + "grad_norm": 4.946988727080576, + "learning_rate": 1.471929733960376e-05, + "loss": 0.3007, + "step": 3346 + }, + { + "epoch": 1.8170466883821934, + "grad_norm": 6.934881579248824, + "learning_rate": 1.4716196475006145e-05, + "loss": 0.5286, + "step": 3347 + }, + { + "epoch": 1.8175895765472312, + "grad_norm": 12.504257989709076, + "learning_rate": 1.4713095027101833e-05, + "loss": 0.9948, + "step": 3348 + }, + { + "epoch": 1.8181324647122694, + "grad_norm": 6.299717892798866, + "learning_rate": 1.4709992996274416e-05, + "loss": 0.6503, + "step": 3349 + }, + { + "epoch": 1.8186753528773072, + "grad_norm": 8.48347618555589, + "learning_rate": 1.4706890382907562e-05, + "loss": 0.7195, + "step": 3350 + }, + { + "epoch": 1.8192182410423454, + "grad_norm": 7.992565784639017, + "learning_rate": 1.4703787187385002e-05, + "loss": 0.7349, + "step": 3351 + }, + { + "epoch": 1.8197611292073832, + "grad_norm": 7.567197193038979, + "learning_rate": 1.4700683410090546e-05, + "loss": 0.4719, + "step": 3352 + }, + { + "epoch": 1.8203040173724214, + "grad_norm": 7.276389893333727, + "learning_rate": 1.4697579051408073e-05, + "loss": 0.93, + "step": 3353 + }, + { + "epoch": 1.8208469055374592, + "grad_norm": 8.13251629417801, + "learning_rate": 1.4694474111721537e-05, + "loss": 0.9865, + "step": 3354 + }, + { + "epoch": 1.8213897937024974, + "grad_norm": 6.788010338042944, + "learning_rate": 1.469136859141496e-05, + "loss": 0.5671, + "step": 3355 + }, + { + "epoch": 1.8219326818675352, + "grad_norm": 6.83194614525194, + "learning_rate": 1.4688262490872438e-05, + "loss": 0.5486, + "step": 3356 + }, + { + "epoch": 1.8224755700325734, + "grad_norm": 12.201391760355081, + "learning_rate": 1.4685155810478143e-05, + "loss": 0.9178, + "step": 3357 + }, + { + "epoch": 1.8230184581976112, + "grad_norm": 7.955977866907339, + "learning_rate": 1.4682048550616306e-05, + "loss": 0.9671, + "step": 3358 + }, + { + "epoch": 1.8235613463626494, + "grad_norm": 9.699881747283202, + "learning_rate": 1.4678940711671245e-05, + "loss": 0.7699, + "step": 3359 + }, + { + "epoch": 1.8241042345276872, + "grad_norm": 7.703855587772515, + "learning_rate": 1.4675832294027339e-05, + "loss": 0.6471, + "step": 3360 + }, + { + "epoch": 1.8246471226927254, + "grad_norm": 11.230217201340418, + "learning_rate": 1.4672723298069043e-05, + "loss": 1.1256, + "step": 3361 + }, + { + "epoch": 1.8251900108577632, + "grad_norm": 11.964156332533788, + "learning_rate": 1.4669613724180883e-05, + "loss": 1.1829, + "step": 3362 + }, + { + "epoch": 1.8257328990228014, + "grad_norm": 8.221701868139569, + "learning_rate": 1.4666503572747452e-05, + "loss": 0.6004, + "step": 3363 + }, + { + "epoch": 1.8262757871878392, + "grad_norm": 6.179895131215982, + "learning_rate": 1.4663392844153426e-05, + "loss": 0.6923, + "step": 3364 + }, + { + "epoch": 1.8268186753528775, + "grad_norm": 8.333959237389427, + "learning_rate": 1.4660281538783538e-05, + "loss": 0.8785, + "step": 3365 + }, + { + "epoch": 1.8273615635179152, + "grad_norm": 9.742872979143373, + "learning_rate": 1.4657169657022603e-05, + "loss": 1.0129, + "step": 3366 + }, + { + "epoch": 1.8279044516829535, + "grad_norm": 8.019435507225802, + "learning_rate": 1.46540571992555e-05, + "loss": 1.0615, + "step": 3367 + }, + { + "epoch": 1.8284473398479912, + "grad_norm": 5.872559756186675, + "learning_rate": 1.4650944165867187e-05, + "loss": 0.6391, + "step": 3368 + }, + { + "epoch": 1.8289902280130295, + "grad_norm": 7.878073300625923, + "learning_rate": 1.4647830557242685e-05, + "loss": 1.0317, + "step": 3369 + }, + { + "epoch": 1.8295331161780672, + "grad_norm": 10.290400619980408, + "learning_rate": 1.4644716373767096e-05, + "loss": 1.3108, + "step": 3370 + }, + { + "epoch": 1.8300760043431055, + "grad_norm": 7.655417729944032, + "learning_rate": 1.4641601615825582e-05, + "loss": 0.8434, + "step": 3371 + }, + { + "epoch": 1.8306188925081432, + "grad_norm": 8.894875963607664, + "learning_rate": 1.4638486283803384e-05, + "loss": 1.4973, + "step": 3372 + }, + { + "epoch": 1.8311617806731815, + "grad_norm": 7.70346849256546, + "learning_rate": 1.4635370378085809e-05, + "loss": 0.9202, + "step": 3373 + }, + { + "epoch": 1.8317046688382193, + "grad_norm": 7.428334777443758, + "learning_rate": 1.4632253899058233e-05, + "loss": 1.005, + "step": 3374 + }, + { + "epoch": 1.8322475570032575, + "grad_norm": 6.905061230055496, + "learning_rate": 1.4629136847106118e-05, + "loss": 0.7045, + "step": 3375 + }, + { + "epoch": 1.8327904451682953, + "grad_norm": 9.092059147160219, + "learning_rate": 1.4626019222614977e-05, + "loss": 1.1478, + "step": 3376 + }, + { + "epoch": 1.8333333333333335, + "grad_norm": 8.171417716573657, + "learning_rate": 1.4622901025970406e-05, + "loss": 0.8098, + "step": 3377 + }, + { + "epoch": 1.8338762214983713, + "grad_norm": 8.935159598593566, + "learning_rate": 1.461978225755807e-05, + "loss": 0.9121, + "step": 3378 + }, + { + "epoch": 1.8344191096634095, + "grad_norm": 9.059802045574934, + "learning_rate": 1.46166629177637e-05, + "loss": 0.952, + "step": 3379 + }, + { + "epoch": 1.8349619978284473, + "grad_norm": 6.404641837192718, + "learning_rate": 1.4613543006973103e-05, + "loss": 0.7178, + "step": 3380 + }, + { + "epoch": 1.8355048859934855, + "grad_norm": 7.714640842582964, + "learning_rate": 1.4610422525572155e-05, + "loss": 0.7053, + "step": 3381 + }, + { + "epoch": 1.8360477741585233, + "grad_norm": 9.232659815202982, + "learning_rate": 1.46073014739468e-05, + "loss": 0.7707, + "step": 3382 + }, + { + "epoch": 1.8365906623235615, + "grad_norm": 7.597273066110282, + "learning_rate": 1.4604179852483056e-05, + "loss": 0.9593, + "step": 3383 + }, + { + "epoch": 1.8371335504885993, + "grad_norm": 7.246167686662127, + "learning_rate": 1.460105766156701e-05, + "loss": 0.8489, + "step": 3384 + }, + { + "epoch": 1.8376764386536375, + "grad_norm": 9.724005322694314, + "learning_rate": 1.4597934901584818e-05, + "loss": 1.4838, + "step": 3385 + }, + { + "epoch": 1.8382193268186753, + "grad_norm": 5.623806840885622, + "learning_rate": 1.459481157292271e-05, + "loss": 0.4218, + "step": 3386 + }, + { + "epoch": 1.8387622149837135, + "grad_norm": 10.505590233471066, + "learning_rate": 1.4591687675966984e-05, + "loss": 1.3634, + "step": 3387 + }, + { + "epoch": 1.8393051031487513, + "grad_norm": 7.225825527188378, + "learning_rate": 1.458856321110401e-05, + "loss": 1.0878, + "step": 3388 + }, + { + "epoch": 1.8398479913137895, + "grad_norm": 6.92346369662784, + "learning_rate": 1.4585438178720221e-05, + "loss": 0.8579, + "step": 3389 + }, + { + "epoch": 1.8403908794788273, + "grad_norm": 6.78903317106508, + "learning_rate": 1.4582312579202134e-05, + "loss": 0.9986, + "step": 3390 + }, + { + "epoch": 1.8409337676438655, + "grad_norm": 7.493115074862596, + "learning_rate": 1.4579186412936323e-05, + "loss": 0.6403, + "step": 3391 + }, + { + "epoch": 1.8414766558089033, + "grad_norm": 7.292140639462925, + "learning_rate": 1.457605968030944e-05, + "loss": 0.7705, + "step": 3392 + }, + { + "epoch": 1.8420195439739415, + "grad_norm": 8.413656799764597, + "learning_rate": 1.4572932381708198e-05, + "loss": 0.7362, + "step": 3393 + }, + { + "epoch": 1.8425624321389793, + "grad_norm": 8.092404119538978, + "learning_rate": 1.4569804517519392e-05, + "loss": 0.8419, + "step": 3394 + }, + { + "epoch": 1.8431053203040175, + "grad_norm": 8.746713556642208, + "learning_rate": 1.456667608812988e-05, + "loss": 0.8992, + "step": 3395 + }, + { + "epoch": 1.8436482084690553, + "grad_norm": 9.315011907817567, + "learning_rate": 1.4563547093926595e-05, + "loss": 1.1811, + "step": 3396 + }, + { + "epoch": 1.8441910966340935, + "grad_norm": 7.921580843874425, + "learning_rate": 1.4560417535296529e-05, + "loss": 0.9896, + "step": 3397 + }, + { + "epoch": 1.8447339847991313, + "grad_norm": 5.324625495868631, + "learning_rate": 1.4557287412626755e-05, + "loss": 0.5757, + "step": 3398 + }, + { + "epoch": 1.8452768729641695, + "grad_norm": 7.061326556489923, + "learning_rate": 1.4554156726304411e-05, + "loss": 0.7016, + "step": 3399 + }, + { + "epoch": 1.8458197611292073, + "grad_norm": 7.833004564597817, + "learning_rate": 1.4551025476716704e-05, + "loss": 0.6112, + "step": 3400 + }, + { + "epoch": 1.8463626492942455, + "grad_norm": 6.762019290951184, + "learning_rate": 1.4547893664250912e-05, + "loss": 0.6197, + "step": 3401 + }, + { + "epoch": 1.8469055374592833, + "grad_norm": 8.02135944193365, + "learning_rate": 1.4544761289294384e-05, + "loss": 0.7665, + "step": 3402 + }, + { + "epoch": 1.8474484256243215, + "grad_norm": 9.145981041987138, + "learning_rate": 1.4541628352234538e-05, + "loss": 1.4491, + "step": 3403 + }, + { + "epoch": 1.8479913137893593, + "grad_norm": 10.127072377919164, + "learning_rate": 1.4538494853458858e-05, + "loss": 1.1929, + "step": 3404 + }, + { + "epoch": 1.8485342019543975, + "grad_norm": 9.041637256167727, + "learning_rate": 1.45353607933549e-05, + "loss": 0.8471, + "step": 3405 + }, + { + "epoch": 1.8490770901194353, + "grad_norm": 7.08038124220055, + "learning_rate": 1.453222617231029e-05, + "loss": 0.5152, + "step": 3406 + }, + { + "epoch": 1.8496199782844736, + "grad_norm": 8.460108981189723, + "learning_rate": 1.4529090990712726e-05, + "loss": 0.9495, + "step": 3407 + }, + { + "epoch": 1.8501628664495113, + "grad_norm": 8.939702203617836, + "learning_rate": 1.452595524894997e-05, + "loss": 1.0959, + "step": 3408 + }, + { + "epoch": 1.8507057546145496, + "grad_norm": 7.5616663713048595, + "learning_rate": 1.4522818947409855e-05, + "loss": 1.0391, + "step": 3409 + }, + { + "epoch": 1.8512486427795873, + "grad_norm": 8.75531450227851, + "learning_rate": 1.4519682086480287e-05, + "loss": 0.7298, + "step": 3410 + }, + { + "epoch": 1.8517915309446256, + "grad_norm": 7.003753183958329, + "learning_rate": 1.4516544666549233e-05, + "loss": 0.4788, + "step": 3411 + }, + { + "epoch": 1.8523344191096633, + "grad_norm": 10.952441700747523, + "learning_rate": 1.4513406688004734e-05, + "loss": 1.0903, + "step": 3412 + }, + { + "epoch": 1.8528773072747016, + "grad_norm": 8.61536651212902, + "learning_rate": 1.4510268151234903e-05, + "loss": 0.6924, + "step": 3413 + }, + { + "epoch": 1.8534201954397393, + "grad_norm": 9.237257682023538, + "learning_rate": 1.4507129056627922e-05, + "loss": 0.8686, + "step": 3414 + }, + { + "epoch": 1.8539630836047776, + "grad_norm": 7.792081331119547, + "learning_rate": 1.4503989404572031e-05, + "loss": 0.9818, + "step": 3415 + }, + { + "epoch": 1.8545059717698154, + "grad_norm": 8.622336937881878, + "learning_rate": 1.4500849195455557e-05, + "loss": 0.9444, + "step": 3416 + }, + { + "epoch": 1.8550488599348536, + "grad_norm": 7.607437519589793, + "learning_rate": 1.4497708429666882e-05, + "loss": 0.846, + "step": 3417 + }, + { + "epoch": 1.8555917480998914, + "grad_norm": 8.358779281223866, + "learning_rate": 1.4494567107594457e-05, + "loss": 0.7125, + "step": 3418 + }, + { + "epoch": 1.8561346362649296, + "grad_norm": 6.8176708905901044, + "learning_rate": 1.449142522962681e-05, + "loss": 0.8921, + "step": 3419 + }, + { + "epoch": 1.8566775244299674, + "grad_norm": 8.931750092418099, + "learning_rate": 1.448828279615253e-05, + "loss": 0.7989, + "step": 3420 + }, + { + "epoch": 1.8572204125950056, + "grad_norm": 7.481066786036915, + "learning_rate": 1.4485139807560281e-05, + "loss": 0.6642, + "step": 3421 + }, + { + "epoch": 1.8577633007600434, + "grad_norm": 7.460924029721227, + "learning_rate": 1.4481996264238796e-05, + "loss": 0.6509, + "step": 3422 + }, + { + "epoch": 1.8583061889250816, + "grad_norm": 7.224784834236484, + "learning_rate": 1.4478852166576869e-05, + "loss": 0.7244, + "step": 3423 + }, + { + "epoch": 1.8588490770901194, + "grad_norm": 5.912102755785491, + "learning_rate": 1.4475707514963367e-05, + "loss": 0.5713, + "step": 3424 + }, + { + "epoch": 1.8593919652551576, + "grad_norm": 6.994884495015334, + "learning_rate": 1.4472562309787224e-05, + "loss": 0.4875, + "step": 3425 + }, + { + "epoch": 1.8599348534201954, + "grad_norm": 6.38845062016643, + "learning_rate": 1.4469416551437445e-05, + "loss": 0.4708, + "step": 3426 + }, + { + "epoch": 1.8604777415852336, + "grad_norm": 11.250341421225603, + "learning_rate": 1.4466270240303109e-05, + "loss": 1.4338, + "step": 3427 + }, + { + "epoch": 1.8610206297502714, + "grad_norm": 9.828853213376544, + "learning_rate": 1.4463123376773348e-05, + "loss": 0.7307, + "step": 3428 + }, + { + "epoch": 1.8615635179153096, + "grad_norm": 7.96498248613748, + "learning_rate": 1.4459975961237375e-05, + "loss": 0.6727, + "step": 3429 + }, + { + "epoch": 1.8621064060803474, + "grad_norm": 9.736354023562699, + "learning_rate": 1.4456827994084473e-05, + "loss": 0.8091, + "step": 3430 + }, + { + "epoch": 1.8626492942453856, + "grad_norm": 10.317215804231997, + "learning_rate": 1.4453679475703974e-05, + "loss": 1.0997, + "step": 3431 + }, + { + "epoch": 1.8631921824104234, + "grad_norm": 10.00211430988322, + "learning_rate": 1.4450530406485301e-05, + "loss": 1.4007, + "step": 3432 + }, + { + "epoch": 1.8637350705754616, + "grad_norm": 10.665872978467544, + "learning_rate": 1.4447380786817934e-05, + "loss": 0.9221, + "step": 3433 + }, + { + "epoch": 1.8642779587404994, + "grad_norm": 8.8489121199226, + "learning_rate": 1.4444230617091424e-05, + "loss": 0.7745, + "step": 3434 + }, + { + "epoch": 1.8648208469055376, + "grad_norm": 7.083707820329958, + "learning_rate": 1.444107989769539e-05, + "loss": 0.7152, + "step": 3435 + }, + { + "epoch": 1.8653637350705754, + "grad_norm": 7.496522090143239, + "learning_rate": 1.4437928629019515e-05, + "loss": 0.5901, + "step": 3436 + }, + { + "epoch": 1.8659066232356136, + "grad_norm": 10.576310319834935, + "learning_rate": 1.4434776811453552e-05, + "loss": 0.7752, + "step": 3437 + }, + { + "epoch": 1.8664495114006514, + "grad_norm": 8.686730112665717, + "learning_rate": 1.4431624445387327e-05, + "loss": 0.8748, + "step": 3438 + }, + { + "epoch": 1.8669923995656896, + "grad_norm": 9.435491336462407, + "learning_rate": 1.4428471531210726e-05, + "loss": 1.1954, + "step": 3439 + }, + { + "epoch": 1.8675352877307274, + "grad_norm": 9.012159097359817, + "learning_rate": 1.4425318069313712e-05, + "loss": 0.7353, + "step": 3440 + }, + { + "epoch": 1.8680781758957656, + "grad_norm": 9.617409218513528, + "learning_rate": 1.4422164060086304e-05, + "loss": 0.6611, + "step": 3441 + }, + { + "epoch": 1.8686210640608034, + "grad_norm": 9.612516235298202, + "learning_rate": 1.4419009503918598e-05, + "loss": 0.8101, + "step": 3442 + }, + { + "epoch": 1.8691639522258416, + "grad_norm": 6.542519560792006, + "learning_rate": 1.4415854401200751e-05, + "loss": 0.473, + "step": 3443 + }, + { + "epoch": 1.8697068403908794, + "grad_norm": 10.696166776054064, + "learning_rate": 1.4412698752322998e-05, + "loss": 1.0001, + "step": 3444 + }, + { + "epoch": 1.8702497285559176, + "grad_norm": 10.087499378994824, + "learning_rate": 1.4409542557675625e-05, + "loss": 0.9323, + "step": 3445 + }, + { + "epoch": 1.8707926167209554, + "grad_norm": 7.829085764895678, + "learning_rate": 1.4406385817649002e-05, + "loss": 0.8624, + "step": 3446 + }, + { + "epoch": 1.8713355048859937, + "grad_norm": 11.554825371147222, + "learning_rate": 1.4403228532633561e-05, + "loss": 0.7735, + "step": 3447 + }, + { + "epoch": 1.8718783930510314, + "grad_norm": 9.43390979006707, + "learning_rate": 1.4400070703019797e-05, + "loss": 0.9128, + "step": 3448 + }, + { + "epoch": 1.8724212812160697, + "grad_norm": 11.851717520188132, + "learning_rate": 1.4396912329198276e-05, + "loss": 1.2379, + "step": 3449 + }, + { + "epoch": 1.8729641693811074, + "grad_norm": 8.408242915866566, + "learning_rate": 1.4393753411559632e-05, + "loss": 0.7549, + "step": 3450 + }, + { + "epoch": 1.8735070575461457, + "grad_norm": 8.144856314123112, + "learning_rate": 1.4390593950494557e-05, + "loss": 0.7281, + "step": 3451 + }, + { + "epoch": 1.8740499457111834, + "grad_norm": 8.202846690293642, + "learning_rate": 1.438743394639383e-05, + "loss": 0.6773, + "step": 3452 + }, + { + "epoch": 1.8745928338762217, + "grad_norm": 9.423562458268494, + "learning_rate": 1.438427339964828e-05, + "loss": 1.1, + "step": 3453 + }, + { + "epoch": 1.8751357220412594, + "grad_norm": 7.196483568403804, + "learning_rate": 1.4381112310648806e-05, + "loss": 0.5809, + "step": 3454 + }, + { + "epoch": 1.8756786102062974, + "grad_norm": 7.26770042201738, + "learning_rate": 1.4377950679786382e-05, + "loss": 0.8053, + "step": 3455 + }, + { + "epoch": 1.8762214983713354, + "grad_norm": 9.472969211327861, + "learning_rate": 1.4374788507452039e-05, + "loss": 0.7998, + "step": 3456 + }, + { + "epoch": 1.8767643865363735, + "grad_norm": 9.514618174868874, + "learning_rate": 1.437162579403688e-05, + "loss": 0.9572, + "step": 3457 + }, + { + "epoch": 1.8773072747014115, + "grad_norm": 9.363034597703797, + "learning_rate": 1.4368462539932073e-05, + "loss": 0.8528, + "step": 3458 + }, + { + "epoch": 1.8778501628664495, + "grad_norm": 7.113386807438533, + "learning_rate": 1.4365298745528863e-05, + "loss": 0.733, + "step": 3459 + }, + { + "epoch": 1.8783930510314875, + "grad_norm": 9.09654372206602, + "learning_rate": 1.4362134411218545e-05, + "loss": 0.7779, + "step": 3460 + }, + { + "epoch": 1.8789359391965255, + "grad_norm": 7.449242636613164, + "learning_rate": 1.4358969537392489e-05, + "loss": 0.899, + "step": 3461 + }, + { + "epoch": 1.8794788273615635, + "grad_norm": 10.755480560042127, + "learning_rate": 1.4355804124442133e-05, + "loss": 1.099, + "step": 3462 + }, + { + "epoch": 1.8800217155266015, + "grad_norm": 9.091218073930861, + "learning_rate": 1.4352638172758984e-05, + "loss": 0.7489, + "step": 3463 + }, + { + "epoch": 1.8805646036916395, + "grad_norm": 9.701260384390375, + "learning_rate": 1.4349471682734603e-05, + "loss": 0.8495, + "step": 3464 + }, + { + "epoch": 1.8811074918566775, + "grad_norm": 9.008809447502522, + "learning_rate": 1.4346304654760637e-05, + "loss": 0.8984, + "step": 3465 + }, + { + "epoch": 1.8816503800217155, + "grad_norm": 8.176404536031864, + "learning_rate": 1.4343137089228783e-05, + "loss": 0.8278, + "step": 3466 + }, + { + "epoch": 1.8821932681867535, + "grad_norm": 9.197307290475337, + "learning_rate": 1.4339968986530813e-05, + "loss": 0.7142, + "step": 3467 + }, + { + "epoch": 1.8827361563517915, + "grad_norm": 5.505228455971892, + "learning_rate": 1.433680034705856e-05, + "loss": 0.4204, + "step": 3468 + }, + { + "epoch": 1.8832790445168295, + "grad_norm": 9.014964705534542, + "learning_rate": 1.433363117120393e-05, + "loss": 0.8988, + "step": 3469 + }, + { + "epoch": 1.8838219326818675, + "grad_norm": 7.290687267145188, + "learning_rate": 1.4330461459358889e-05, + "loss": 0.7683, + "step": 3470 + }, + { + "epoch": 1.8843648208469055, + "grad_norm": 10.834879009670003, + "learning_rate": 1.4327291211915473e-05, + "loss": 1.0446, + "step": 3471 + }, + { + "epoch": 1.8849077090119435, + "grad_norm": 8.240452054600638, + "learning_rate": 1.4324120429265781e-05, + "loss": 0.893, + "step": 3472 + }, + { + "epoch": 1.8854505971769815, + "grad_norm": 8.17919491677486, + "learning_rate": 1.4320949111801987e-05, + "loss": 0.9846, + "step": 3473 + }, + { + "epoch": 1.8859934853420195, + "grad_norm": 7.693481880154859, + "learning_rate": 1.431777725991632e-05, + "loss": 0.6054, + "step": 3474 + }, + { + "epoch": 1.8865363735070575, + "grad_norm": 7.929110152467141, + "learning_rate": 1.4314604874001076e-05, + "loss": 0.7299, + "step": 3475 + }, + { + "epoch": 1.8870792616720955, + "grad_norm": 9.650029930461764, + "learning_rate": 1.4311431954448629e-05, + "loss": 1.0855, + "step": 3476 + }, + { + "epoch": 1.8876221498371335, + "grad_norm": 8.487868906333745, + "learning_rate": 1.4308258501651404e-05, + "loss": 0.7117, + "step": 3477 + }, + { + "epoch": 1.8881650380021715, + "grad_norm": 7.111157851383161, + "learning_rate": 1.43050845160019e-05, + "loss": 0.6477, + "step": 3478 + }, + { + "epoch": 1.8887079261672095, + "grad_norm": 8.2908796123436, + "learning_rate": 1.4301909997892684e-05, + "loss": 0.9819, + "step": 3479 + }, + { + "epoch": 1.8892508143322475, + "grad_norm": 10.322900044501562, + "learning_rate": 1.4298734947716385e-05, + "loss": 0.9368, + "step": 3480 + }, + { + "epoch": 1.8897937024972855, + "grad_norm": 6.902927389625658, + "learning_rate": 1.4295559365865694e-05, + "loss": 0.6201, + "step": 3481 + }, + { + "epoch": 1.8903365906623235, + "grad_norm": 8.909574060437844, + "learning_rate": 1.4292383252733375e-05, + "loss": 0.9393, + "step": 3482 + }, + { + "epoch": 1.8908794788273615, + "grad_norm": 7.643883869480783, + "learning_rate": 1.4289206608712252e-05, + "loss": 0.9349, + "step": 3483 + }, + { + "epoch": 1.8914223669923995, + "grad_norm": 7.103928943926094, + "learning_rate": 1.428602943419522e-05, + "loss": 0.6511, + "step": 3484 + }, + { + "epoch": 1.8919652551574375, + "grad_norm": 8.843983286039231, + "learning_rate": 1.428285172957524e-05, + "loss": 0.6958, + "step": 3485 + }, + { + "epoch": 1.8925081433224755, + "grad_norm": 5.044738479793456, + "learning_rate": 1.427967349524533e-05, + "loss": 0.59, + "step": 3486 + }, + { + "epoch": 1.8930510314875135, + "grad_norm": 6.9998626206641195, + "learning_rate": 1.4276494731598582e-05, + "loss": 0.5973, + "step": 3487 + }, + { + "epoch": 1.8935939196525515, + "grad_norm": 8.316471360458495, + "learning_rate": 1.4273315439028151e-05, + "loss": 0.8941, + "step": 3488 + }, + { + "epoch": 1.8941368078175895, + "grad_norm": 12.713883241844957, + "learning_rate": 1.4270135617927254e-05, + "loss": 1.4921, + "step": 3489 + }, + { + "epoch": 1.8946796959826275, + "grad_norm": 10.043253133675226, + "learning_rate": 1.426695526868918e-05, + "loss": 1.0242, + "step": 3490 + }, + { + "epoch": 1.8952225841476655, + "grad_norm": 11.142761788047633, + "learning_rate": 1.4263774391707274e-05, + "loss": 1.265, + "step": 3491 + }, + { + "epoch": 1.8957654723127035, + "grad_norm": 7.12450280478247, + "learning_rate": 1.4260592987374962e-05, + "loss": 0.484, + "step": 3492 + }, + { + "epoch": 1.8963083604777415, + "grad_norm": 7.017938163026924, + "learning_rate": 1.4257411056085712e-05, + "loss": 0.6549, + "step": 3493 + }, + { + "epoch": 1.8968512486427795, + "grad_norm": 7.643680244898419, + "learning_rate": 1.4254228598233082e-05, + "loss": 0.614, + "step": 3494 + }, + { + "epoch": 1.8973941368078175, + "grad_norm": 9.484247292353862, + "learning_rate": 1.4251045614210678e-05, + "loss": 0.9667, + "step": 3495 + }, + { + "epoch": 1.8979370249728555, + "grad_norm": 10.846455465412243, + "learning_rate": 1.4247862104412175e-05, + "loss": 1.4425, + "step": 3496 + }, + { + "epoch": 1.8984799131378935, + "grad_norm": 7.211983265810251, + "learning_rate": 1.4244678069231319e-05, + "loss": 0.7508, + "step": 3497 + }, + { + "epoch": 1.8990228013029316, + "grad_norm": 8.283019454355765, + "learning_rate": 1.4241493509061912e-05, + "loss": 0.9747, + "step": 3498 + }, + { + "epoch": 1.8995656894679696, + "grad_norm": 6.678762476220025, + "learning_rate": 1.4238308424297833e-05, + "loss": 0.4702, + "step": 3499 + }, + { + "epoch": 1.9001085776330076, + "grad_norm": 8.750314729983545, + "learning_rate": 1.423512281533301e-05, + "loss": 0.7636, + "step": 3500 + }, + { + "epoch": 1.9006514657980456, + "grad_norm": 7.304249105438763, + "learning_rate": 1.4231936682561446e-05, + "loss": 0.7198, + "step": 3501 + }, + { + "epoch": 1.9011943539630836, + "grad_norm": 7.676960806676722, + "learning_rate": 1.4228750026377212e-05, + "loss": 0.8639, + "step": 3502 + }, + { + "epoch": 1.9017372421281216, + "grad_norm": 10.327339388528987, + "learning_rate": 1.4225562847174431e-05, + "loss": 0.9386, + "step": 3503 + }, + { + "epoch": 1.9022801302931596, + "grad_norm": 8.828430779713369, + "learning_rate": 1.4222375145347304e-05, + "loss": 1.0821, + "step": 3504 + }, + { + "epoch": 1.9028230184581976, + "grad_norm": 8.099300889556567, + "learning_rate": 1.421918692129009e-05, + "loss": 0.6251, + "step": 3505 + }, + { + "epoch": 1.9033659066232356, + "grad_norm": 9.197832742387652, + "learning_rate": 1.4215998175397115e-05, + "loss": 1.1794, + "step": 3506 + }, + { + "epoch": 1.9039087947882736, + "grad_norm": 9.305579433655693, + "learning_rate": 1.4212808908062763e-05, + "loss": 1.3447, + "step": 3507 + }, + { + "epoch": 1.9044516829533116, + "grad_norm": 6.897677914233707, + "learning_rate": 1.4209619119681497e-05, + "loss": 0.5178, + "step": 3508 + }, + { + "epoch": 1.9049945711183496, + "grad_norm": 9.359746343598495, + "learning_rate": 1.420642881064782e-05, + "loss": 0.985, + "step": 3509 + }, + { + "epoch": 1.9055374592833876, + "grad_norm": 5.855418324834113, + "learning_rate": 1.4203237981356331e-05, + "loss": 0.5144, + "step": 3510 + }, + { + "epoch": 1.9060803474484256, + "grad_norm": 7.408991833255532, + "learning_rate": 1.4200046632201665e-05, + "loss": 0.7358, + "step": 3511 + }, + { + "epoch": 1.9066232356134636, + "grad_norm": 9.024544122531433, + "learning_rate": 1.419685476357854e-05, + "loss": 0.778, + "step": 3512 + }, + { + "epoch": 1.9071661237785016, + "grad_norm": 7.874409137592207, + "learning_rate": 1.419366237588173e-05, + "loss": 0.8703, + "step": 3513 + }, + { + "epoch": 1.9077090119435396, + "grad_norm": 8.221027989379657, + "learning_rate": 1.4190469469506073e-05, + "loss": 0.7436, + "step": 3514 + }, + { + "epoch": 1.9082519001085776, + "grad_norm": 10.797659549408706, + "learning_rate": 1.4187276044846473e-05, + "loss": 1.0635, + "step": 3515 + }, + { + "epoch": 1.9087947882736156, + "grad_norm": 9.718910477411724, + "learning_rate": 1.4184082102297896e-05, + "loss": 0.8802, + "step": 3516 + }, + { + "epoch": 1.9093376764386536, + "grad_norm": 7.365463419520298, + "learning_rate": 1.4180887642255376e-05, + "loss": 0.5919, + "step": 3517 + }, + { + "epoch": 1.9098805646036916, + "grad_norm": 6.487376521266865, + "learning_rate": 1.4177692665114014e-05, + "loss": 0.7098, + "step": 3518 + }, + { + "epoch": 1.9104234527687296, + "grad_norm": 8.328541010595174, + "learning_rate": 1.4174497171268962e-05, + "loss": 0.6124, + "step": 3519 + }, + { + "epoch": 1.9109663409337676, + "grad_norm": 7.888322073346588, + "learning_rate": 1.4171301161115447e-05, + "loss": 0.827, + "step": 3520 + }, + { + "epoch": 1.9115092290988056, + "grad_norm": 10.407728423373431, + "learning_rate": 1.4168104635048756e-05, + "loss": 1.0511, + "step": 3521 + }, + { + "epoch": 1.9120521172638436, + "grad_norm": 8.154984984105669, + "learning_rate": 1.4164907593464239e-05, + "loss": 0.8156, + "step": 3522 + }, + { + "epoch": 1.9125950054288816, + "grad_norm": 9.142522776167024, + "learning_rate": 1.4161710036757314e-05, + "loss": 0.9046, + "step": 3523 + }, + { + "epoch": 1.9131378935939196, + "grad_norm": 8.610209396948733, + "learning_rate": 1.415851196532346e-05, + "loss": 0.5673, + "step": 3524 + }, + { + "epoch": 1.9136807817589576, + "grad_norm": 8.072698611315632, + "learning_rate": 1.415531337955822e-05, + "loss": 0.7375, + "step": 3525 + }, + { + "epoch": 1.9142236699239956, + "grad_norm": 9.878888535422037, + "learning_rate": 1.4152114279857197e-05, + "loss": 0.9677, + "step": 3526 + }, + { + "epoch": 1.9147665580890336, + "grad_norm": 8.524331697287977, + "learning_rate": 1.4148914666616062e-05, + "loss": 0.7617, + "step": 3527 + }, + { + "epoch": 1.9153094462540716, + "grad_norm": 8.04939882692512, + "learning_rate": 1.4145714540230549e-05, + "loss": 0.8767, + "step": 3528 + }, + { + "epoch": 1.9158523344191096, + "grad_norm": 10.69782500673993, + "learning_rate": 1.4142513901096453e-05, + "loss": 0.6357, + "step": 3529 + }, + { + "epoch": 1.9163952225841476, + "grad_norm": 7.279779735048647, + "learning_rate": 1.4139312749609637e-05, + "loss": 0.6474, + "step": 3530 + }, + { + "epoch": 1.9169381107491856, + "grad_norm": 7.290241222497209, + "learning_rate": 1.4136111086166024e-05, + "loss": 0.6929, + "step": 3531 + }, + { + "epoch": 1.9174809989142236, + "grad_norm": 8.3613074350218, + "learning_rate": 1.4132908911161598e-05, + "loss": 0.8199, + "step": 3532 + }, + { + "epoch": 1.9180238870792616, + "grad_norm": 9.015830183265928, + "learning_rate": 1.4129706224992413e-05, + "loss": 0.9811, + "step": 3533 + }, + { + "epoch": 1.9185667752442996, + "grad_norm": 9.621458967761185, + "learning_rate": 1.4126503028054579e-05, + "loss": 0.7498, + "step": 3534 + }, + { + "epoch": 1.9191096634093376, + "grad_norm": 11.105255757566027, + "learning_rate": 1.412329932074427e-05, + "loss": 0.7587, + "step": 3535 + }, + { + "epoch": 1.9196525515743756, + "grad_norm": 8.599145695443788, + "learning_rate": 1.4120095103457734e-05, + "loss": 0.934, + "step": 3536 + }, + { + "epoch": 1.9201954397394136, + "grad_norm": 7.727889210926642, + "learning_rate": 1.4116890376591268e-05, + "loss": 0.8305, + "step": 3537 + }, + { + "epoch": 1.9207383279044516, + "grad_norm": 12.029529213228301, + "learning_rate": 1.4113685140541242e-05, + "loss": 1.494, + "step": 3538 + }, + { + "epoch": 1.9212812160694897, + "grad_norm": 7.404838471492053, + "learning_rate": 1.411047939570408e-05, + "loss": 0.5168, + "step": 3539 + }, + { + "epoch": 1.9218241042345277, + "grad_norm": 7.540024472508733, + "learning_rate": 1.4107273142476272e-05, + "loss": 0.7162, + "step": 3540 + }, + { + "epoch": 1.9223669923995657, + "grad_norm": 7.885769945154571, + "learning_rate": 1.4104066381254378e-05, + "loss": 0.5964, + "step": 3541 + }, + { + "epoch": 1.9229098805646037, + "grad_norm": 8.287652724014645, + "learning_rate": 1.4100859112435013e-05, + "loss": 0.6175, + "step": 3542 + }, + { + "epoch": 1.9234527687296417, + "grad_norm": 8.335133777216624, + "learning_rate": 1.4097651336414857e-05, + "loss": 0.7163, + "step": 3543 + }, + { + "epoch": 1.9239956568946797, + "grad_norm": 7.765701865468595, + "learning_rate": 1.4094443053590652e-05, + "loss": 0.4534, + "step": 3544 + }, + { + "epoch": 1.9245385450597177, + "grad_norm": 7.716709219714201, + "learning_rate": 1.4091234264359206e-05, + "loss": 0.4425, + "step": 3545 + }, + { + "epoch": 1.9250814332247557, + "grad_norm": 11.979255681165771, + "learning_rate": 1.4088024969117387e-05, + "loss": 1.0249, + "step": 3546 + }, + { + "epoch": 1.9256243213897937, + "grad_norm": 9.424753754574079, + "learning_rate": 1.4084815168262123e-05, + "loss": 0.9663, + "step": 3547 + }, + { + "epoch": 1.9261672095548317, + "grad_norm": 10.101408847774994, + "learning_rate": 1.4081604862190407e-05, + "loss": 0.8136, + "step": 3548 + }, + { + "epoch": 1.9267100977198697, + "grad_norm": 11.755751498543427, + "learning_rate": 1.4078394051299298e-05, + "loss": 0.9451, + "step": 3549 + }, + { + "epoch": 1.9272529858849077, + "grad_norm": 12.251136923575924, + "learning_rate": 1.4075182735985913e-05, + "loss": 1.5716, + "step": 3550 + }, + { + "epoch": 1.9277958740499457, + "grad_norm": 10.087922707091717, + "learning_rate": 1.4071970916647432e-05, + "loss": 0.7026, + "step": 3551 + }, + { + "epoch": 1.9283387622149837, + "grad_norm": 8.507872704310849, + "learning_rate": 1.40687585936811e-05, + "loss": 0.7016, + "step": 3552 + }, + { + "epoch": 1.9288816503800217, + "grad_norm": 8.98138340766033, + "learning_rate": 1.4065545767484218e-05, + "loss": 0.6834, + "step": 3553 + }, + { + "epoch": 1.9294245385450597, + "grad_norm": 7.1085330003672045, + "learning_rate": 1.4062332438454156e-05, + "loss": 0.6813, + "step": 3554 + }, + { + "epoch": 1.9299674267100977, + "grad_norm": 13.499346110936985, + "learning_rate": 1.4059118606988345e-05, + "loss": 1.0068, + "step": 3555 + }, + { + "epoch": 1.9305103148751357, + "grad_norm": 7.698407375447584, + "learning_rate": 1.4055904273484275e-05, + "loss": 0.8418, + "step": 3556 + }, + { + "epoch": 1.9310532030401737, + "grad_norm": 8.424294783097983, + "learning_rate": 1.40526894383395e-05, + "loss": 0.8087, + "step": 3557 + }, + { + "epoch": 1.9315960912052117, + "grad_norm": 8.962780671623113, + "learning_rate": 1.4049474101951639e-05, + "loss": 0.8644, + "step": 3558 + }, + { + "epoch": 1.9321389793702497, + "grad_norm": 8.580581507909985, + "learning_rate": 1.4046258264718363e-05, + "loss": 1.0325, + "step": 3559 + }, + { + "epoch": 1.9326818675352877, + "grad_norm": 10.417878574145533, + "learning_rate": 1.4043041927037418e-05, + "loss": 0.7705, + "step": 3560 + }, + { + "epoch": 1.9332247557003257, + "grad_norm": 8.250628073154727, + "learning_rate": 1.4039825089306603e-05, + "loss": 0.794, + "step": 3561 + }, + { + "epoch": 1.9337676438653637, + "grad_norm": 8.228275233396532, + "learning_rate": 1.4036607751923783e-05, + "loss": 0.8759, + "step": 3562 + }, + { + "epoch": 1.9343105320304017, + "grad_norm": 7.760702739983814, + "learning_rate": 1.4033389915286884e-05, + "loss": 0.6458, + "step": 3563 + }, + { + "epoch": 1.9348534201954397, + "grad_norm": 6.508974748032192, + "learning_rate": 1.403017157979389e-05, + "loss": 0.6348, + "step": 3564 + }, + { + "epoch": 1.9353963083604777, + "grad_norm": 11.570824985629717, + "learning_rate": 1.4026952745842853e-05, + "loss": 1.0935, + "step": 3565 + }, + { + "epoch": 1.9359391965255157, + "grad_norm": 9.901046454739335, + "learning_rate": 1.4023733413831881e-05, + "loss": 0.7458, + "step": 3566 + }, + { + "epoch": 1.9364820846905537, + "grad_norm": 9.103951624002182, + "learning_rate": 1.402051358415915e-05, + "loss": 0.8253, + "step": 3567 + }, + { + "epoch": 1.9370249728555917, + "grad_norm": 9.293062225974824, + "learning_rate": 1.4017293257222887e-05, + "loss": 0.5801, + "step": 3568 + }, + { + "epoch": 1.9375678610206297, + "grad_norm": 7.180133735690273, + "learning_rate": 1.4014072433421398e-05, + "loss": 0.6063, + "step": 3569 + }, + { + "epoch": 1.9381107491856677, + "grad_norm": 9.028614205338394, + "learning_rate": 1.4010851113153028e-05, + "loss": 0.9412, + "step": 3570 + }, + { + "epoch": 1.9386536373507057, + "grad_norm": 9.51192022384654, + "learning_rate": 1.4007629296816202e-05, + "loss": 0.6459, + "step": 3571 + }, + { + "epoch": 1.9391965255157437, + "grad_norm": 9.883078787028563, + "learning_rate": 1.4004406984809396e-05, + "loss": 1.0324, + "step": 3572 + }, + { + "epoch": 1.9397394136807817, + "grad_norm": 7.433318187249177, + "learning_rate": 1.4001184177531154e-05, + "loss": 0.6152, + "step": 3573 + }, + { + "epoch": 1.9402823018458197, + "grad_norm": 10.165800890508075, + "learning_rate": 1.399796087538007e-05, + "loss": 0.8751, + "step": 3574 + }, + { + "epoch": 1.9408251900108577, + "grad_norm": 7.668138219396951, + "learning_rate": 1.3994737078754819e-05, + "loss": 0.43, + "step": 3575 + }, + { + "epoch": 1.9413680781758957, + "grad_norm": 8.929591410847367, + "learning_rate": 1.3991512788054115e-05, + "loss": 0.6173, + "step": 3576 + }, + { + "epoch": 1.9419109663409337, + "grad_norm": 8.069899029370944, + "learning_rate": 1.3988288003676755e-05, + "loss": 0.6782, + "step": 3577 + }, + { + "epoch": 1.9424538545059717, + "grad_norm": 7.635775018643444, + "learning_rate": 1.3985062726021574e-05, + "loss": 0.7503, + "step": 3578 + }, + { + "epoch": 1.9429967426710097, + "grad_norm": 11.649932328973133, + "learning_rate": 1.3981836955487485e-05, + "loss": 0.9154, + "step": 3579 + }, + { + "epoch": 1.9435396308360477, + "grad_norm": 9.11865472044189, + "learning_rate": 1.397861069247345e-05, + "loss": 0.8799, + "step": 3580 + }, + { + "epoch": 1.9440825190010858, + "grad_norm": 7.628563459317163, + "learning_rate": 1.3975383937378508e-05, + "loss": 0.7199, + "step": 3581 + }, + { + "epoch": 1.9446254071661238, + "grad_norm": 10.640127088045624, + "learning_rate": 1.3972156690601747e-05, + "loss": 0.7885, + "step": 3582 + }, + { + "epoch": 1.9451682953311618, + "grad_norm": 10.192406733466123, + "learning_rate": 1.3968928952542313e-05, + "loss": 0.6036, + "step": 3583 + }, + { + "epoch": 1.9457111834961998, + "grad_norm": 12.020287450972278, + "learning_rate": 1.396570072359942e-05, + "loss": 0.9477, + "step": 3584 + }, + { + "epoch": 1.9462540716612378, + "grad_norm": 8.855072111194623, + "learning_rate": 1.3962472004172343e-05, + "loss": 0.4384, + "step": 3585 + }, + { + "epoch": 1.9467969598262758, + "grad_norm": 7.173236338345466, + "learning_rate": 1.3959242794660412e-05, + "loss": 0.4969, + "step": 3586 + }, + { + "epoch": 1.9473398479913138, + "grad_norm": 6.505005692892076, + "learning_rate": 1.3956013095463024e-05, + "loss": 0.4585, + "step": 3587 + }, + { + "epoch": 1.9478827361563518, + "grad_norm": 7.041517208688192, + "learning_rate": 1.395278290697963e-05, + "loss": 0.5439, + "step": 3588 + }, + { + "epoch": 1.9484256243213898, + "grad_norm": 9.730465359550896, + "learning_rate": 1.3949552229609746e-05, + "loss": 0.8865, + "step": 3589 + }, + { + "epoch": 1.9489685124864278, + "grad_norm": 7.73678668778, + "learning_rate": 1.3946321063752948e-05, + "loss": 0.5094, + "step": 3590 + }, + { + "epoch": 1.9495114006514658, + "grad_norm": 7.095011334397053, + "learning_rate": 1.3943089409808872e-05, + "loss": 0.5224, + "step": 3591 + }, + { + "epoch": 1.9500542888165038, + "grad_norm": 8.086388468165458, + "learning_rate": 1.393985726817721e-05, + "loss": 0.7798, + "step": 3592 + }, + { + "epoch": 1.9505971769815418, + "grad_norm": 10.262604564335476, + "learning_rate": 1.3936624639257726e-05, + "loss": 1.2595, + "step": 3593 + }, + { + "epoch": 1.9511400651465798, + "grad_norm": 8.535907967530573, + "learning_rate": 1.393339152345023e-05, + "loss": 0.8205, + "step": 3594 + }, + { + "epoch": 1.9516829533116178, + "grad_norm": 7.864477340620307, + "learning_rate": 1.3930157921154601e-05, + "loss": 0.5967, + "step": 3595 + }, + { + "epoch": 1.9522258414766558, + "grad_norm": 6.739275201436143, + "learning_rate": 1.392692383277078e-05, + "loss": 0.5836, + "step": 3596 + }, + { + "epoch": 1.9527687296416938, + "grad_norm": 9.533970605538912, + "learning_rate": 1.392368925869876e-05, + "loss": 0.5761, + "step": 3597 + }, + { + "epoch": 1.9533116178067318, + "grad_norm": 11.376343280359167, + "learning_rate": 1.3920454199338598e-05, + "loss": 0.7596, + "step": 3598 + }, + { + "epoch": 1.9538545059717698, + "grad_norm": 7.280453649201749, + "learning_rate": 1.391721865509041e-05, + "loss": 0.4949, + "step": 3599 + }, + { + "epoch": 1.9543973941368078, + "grad_norm": 7.734190007799969, + "learning_rate": 1.391398262635438e-05, + "loss": 0.9177, + "step": 3600 + }, + { + "epoch": 1.9549402823018458, + "grad_norm": 13.106124045444247, + "learning_rate": 1.3910746113530738e-05, + "loss": 1.1884, + "step": 3601 + }, + { + "epoch": 1.9554831704668838, + "grad_norm": 11.868912051178873, + "learning_rate": 1.3907509117019783e-05, + "loss": 0.9245, + "step": 3602 + }, + { + "epoch": 1.9560260586319218, + "grad_norm": 10.307498071834987, + "learning_rate": 1.3904271637221876e-05, + "loss": 1.0802, + "step": 3603 + }, + { + "epoch": 1.9565689467969598, + "grad_norm": 8.308697157532139, + "learning_rate": 1.390103367453743e-05, + "loss": 0.4979, + "step": 3604 + }, + { + "epoch": 1.9571118349619978, + "grad_norm": 8.447604525372762, + "learning_rate": 1.3897795229366919e-05, + "loss": 0.8942, + "step": 3605 + }, + { + "epoch": 1.9576547231270358, + "grad_norm": 7.864013969756676, + "learning_rate": 1.3894556302110883e-05, + "loss": 0.642, + "step": 3606 + }, + { + "epoch": 1.9581976112920738, + "grad_norm": 7.903822544650853, + "learning_rate": 1.389131689316992e-05, + "loss": 0.8251, + "step": 3607 + }, + { + "epoch": 1.9587404994571118, + "grad_norm": 9.744275276467558, + "learning_rate": 1.3888077002944678e-05, + "loss": 1.0052, + "step": 3608 + }, + { + "epoch": 1.9592833876221498, + "grad_norm": 7.731164092689308, + "learning_rate": 1.3884836631835877e-05, + "loss": 0.6883, + "step": 3609 + }, + { + "epoch": 1.9598262757871878, + "grad_norm": 8.849291385162568, + "learning_rate": 1.3881595780244288e-05, + "loss": 0.9985, + "step": 3610 + }, + { + "epoch": 1.9603691639522258, + "grad_norm": 9.531929458494986, + "learning_rate": 1.3878354448570748e-05, + "loss": 0.7643, + "step": 3611 + }, + { + "epoch": 1.9609120521172638, + "grad_norm": 10.32443732773501, + "learning_rate": 1.3875112637216145e-05, + "loss": 0.848, + "step": 3612 + }, + { + "epoch": 1.9614549402823018, + "grad_norm": 7.417155617136626, + "learning_rate": 1.387187034658144e-05, + "loss": 0.5682, + "step": 3613 + }, + { + "epoch": 1.9619978284473398, + "grad_norm": 14.072831276291689, + "learning_rate": 1.386862757706764e-05, + "loss": 0.8279, + "step": 3614 + }, + { + "epoch": 1.9625407166123778, + "grad_norm": 7.764451928370161, + "learning_rate": 1.3865384329075812e-05, + "loss": 0.6687, + "step": 3615 + }, + { + "epoch": 1.9630836047774158, + "grad_norm": 8.686094567618806, + "learning_rate": 1.3862140603007095e-05, + "loss": 1.1034, + "step": 3616 + }, + { + "epoch": 1.9636264929424538, + "grad_norm": 10.753407337858437, + "learning_rate": 1.3858896399262669e-05, + "loss": 1.1692, + "step": 3617 + }, + { + "epoch": 1.9641693811074918, + "grad_norm": 9.096005648867607, + "learning_rate": 1.3855651718243786e-05, + "loss": 0.6191, + "step": 3618 + }, + { + "epoch": 1.9647122692725298, + "grad_norm": 10.387647049254962, + "learning_rate": 1.3852406560351752e-05, + "loss": 1.3573, + "step": 3619 + }, + { + "epoch": 1.9652551574375678, + "grad_norm": 7.173958455599766, + "learning_rate": 1.3849160925987936e-05, + "loss": 0.8106, + "step": 3620 + }, + { + "epoch": 1.9657980456026058, + "grad_norm": 9.434903281870886, + "learning_rate": 1.3845914815553765e-05, + "loss": 0.9581, + "step": 3621 + }, + { + "epoch": 1.9663409337676439, + "grad_norm": 7.671221453670278, + "learning_rate": 1.3842668229450717e-05, + "loss": 0.6989, + "step": 3622 + }, + { + "epoch": 1.9668838219326819, + "grad_norm": 7.360534692664803, + "learning_rate": 1.3839421168080338e-05, + "loss": 0.4912, + "step": 3623 + }, + { + "epoch": 1.9674267100977199, + "grad_norm": 6.853527063108637, + "learning_rate": 1.3836173631844231e-05, + "loss": 0.7485, + "step": 3624 + }, + { + "epoch": 1.9679695982627579, + "grad_norm": 8.619792011062271, + "learning_rate": 1.3832925621144057e-05, + "loss": 0.712, + "step": 3625 + }, + { + "epoch": 1.9685124864277959, + "grad_norm": 8.062990147577946, + "learning_rate": 1.3829677136381532e-05, + "loss": 0.7126, + "step": 3626 + }, + { + "epoch": 1.9690553745928339, + "grad_norm": 6.35400375341867, + "learning_rate": 1.3826428177958433e-05, + "loss": 0.5262, + "step": 3627 + }, + { + "epoch": 1.9695982627578719, + "grad_norm": 7.322594621519839, + "learning_rate": 1.3823178746276603e-05, + "loss": 0.6644, + "step": 3628 + }, + { + "epoch": 1.9701411509229099, + "grad_norm": 9.976906946653425, + "learning_rate": 1.3819928841737929e-05, + "loss": 0.7411, + "step": 3629 + }, + { + "epoch": 1.9706840390879479, + "grad_norm": 11.052922199516026, + "learning_rate": 1.3816678464744368e-05, + "loss": 0.9998, + "step": 3630 + }, + { + "epoch": 1.9712269272529859, + "grad_norm": 7.695988647358978, + "learning_rate": 1.381342761569793e-05, + "loss": 0.6805, + "step": 3631 + }, + { + "epoch": 1.9717698154180239, + "grad_norm": 10.387946538947885, + "learning_rate": 1.381017629500069e-05, + "loss": 0.8954, + "step": 3632 + }, + { + "epoch": 1.9723127035830619, + "grad_norm": 8.215553841008688, + "learning_rate": 1.380692450305477e-05, + "loss": 1.0488, + "step": 3633 + }, + { + "epoch": 1.9728555917480999, + "grad_norm": 8.550898170705977, + "learning_rate": 1.3803672240262364e-05, + "loss": 0.9301, + "step": 3634 + }, + { + "epoch": 1.9733984799131379, + "grad_norm": 9.978410770263547, + "learning_rate": 1.380041950702571e-05, + "loss": 0.9725, + "step": 3635 + }, + { + "epoch": 1.9739413680781759, + "grad_norm": 10.211160158432088, + "learning_rate": 1.3797166303747119e-05, + "loss": 1.1789, + "step": 3636 + }, + { + "epoch": 1.9744842562432139, + "grad_norm": 7.47091491709439, + "learning_rate": 1.3793912630828942e-05, + "loss": 1.2311, + "step": 3637 + }, + { + "epoch": 1.975027144408252, + "grad_norm": 9.966534074490724, + "learning_rate": 1.3790658488673607e-05, + "loss": 1.0057, + "step": 3638 + }, + { + "epoch": 1.97557003257329, + "grad_norm": 9.341517091004942, + "learning_rate": 1.378740387768359e-05, + "loss": 0.7708, + "step": 3639 + }, + { + "epoch": 1.976112920738328, + "grad_norm": 10.066755071795829, + "learning_rate": 1.3784148798261422e-05, + "loss": 1.3146, + "step": 3640 + }, + { + "epoch": 1.976655808903366, + "grad_norm": 8.306959980081418, + "learning_rate": 1.3780893250809705e-05, + "loss": 0.6926, + "step": 3641 + }, + { + "epoch": 1.977198697068404, + "grad_norm": 5.980548702977229, + "learning_rate": 1.377763723573108e-05, + "loss": 0.6776, + "step": 3642 + }, + { + "epoch": 1.977741585233442, + "grad_norm": 10.796897781225315, + "learning_rate": 1.3774380753428263e-05, + "loss": 1.0422, + "step": 3643 + }, + { + "epoch": 1.97828447339848, + "grad_norm": 9.650854393967004, + "learning_rate": 1.3771123804304018e-05, + "loss": 1.2566, + "step": 3644 + }, + { + "epoch": 1.978827361563518, + "grad_norm": 6.718663747121732, + "learning_rate": 1.3767866388761168e-05, + "loss": 0.5018, + "step": 3645 + }, + { + "epoch": 1.979370249728556, + "grad_norm": 8.345918705516796, + "learning_rate": 1.3764608507202604e-05, + "loss": 0.8501, + "step": 3646 + }, + { + "epoch": 1.979913137893594, + "grad_norm": 11.065999609001787, + "learning_rate": 1.3761350160031258e-05, + "loss": 1.2627, + "step": 3647 + }, + { + "epoch": 1.980456026058632, + "grad_norm": 8.567448073304528, + "learning_rate": 1.3758091347650126e-05, + "loss": 0.4477, + "step": 3648 + }, + { + "epoch": 1.98099891422367, + "grad_norm": 11.930236219703469, + "learning_rate": 1.3754832070462269e-05, + "loss": 1.5386, + "step": 3649 + }, + { + "epoch": 1.981541802388708, + "grad_norm": 7.334841489157127, + "learning_rate": 1.3751572328870797e-05, + "loss": 0.5759, + "step": 3650 + }, + { + "epoch": 1.982084690553746, + "grad_norm": 6.392865090419794, + "learning_rate": 1.3748312123278879e-05, + "loss": 0.6078, + "step": 3651 + }, + { + "epoch": 1.982627578718784, + "grad_norm": 8.598388892264492, + "learning_rate": 1.3745051454089744e-05, + "loss": 0.7682, + "step": 3652 + }, + { + "epoch": 1.983170466883822, + "grad_norm": 9.310542957266444, + "learning_rate": 1.3741790321706678e-05, + "loss": 0.9833, + "step": 3653 + }, + { + "epoch": 1.98371335504886, + "grad_norm": 8.917570196452148, + "learning_rate": 1.3738528726533021e-05, + "loss": 0.5915, + "step": 3654 + }, + { + "epoch": 1.984256243213898, + "grad_norm": 9.503202436394638, + "learning_rate": 1.3735266668972174e-05, + "loss": 0.9347, + "step": 3655 + }, + { + "epoch": 1.984799131378936, + "grad_norm": 9.187024367480415, + "learning_rate": 1.3732004149427592e-05, + "loss": 0.6669, + "step": 3656 + }, + { + "epoch": 1.985342019543974, + "grad_norm": 10.497812893691144, + "learning_rate": 1.3728741168302785e-05, + "loss": 1.0663, + "step": 3657 + }, + { + "epoch": 1.985884907709012, + "grad_norm": 9.646030688042543, + "learning_rate": 1.3725477726001332e-05, + "loss": 0.9913, + "step": 3658 + }, + { + "epoch": 1.98642779587405, + "grad_norm": 9.475431343889321, + "learning_rate": 1.3722213822926855e-05, + "loss": 0.8377, + "step": 3659 + }, + { + "epoch": 1.986970684039088, + "grad_norm": 9.449591675048769, + "learning_rate": 1.371894945948304e-05, + "loss": 0.8763, + "step": 3660 + }, + { + "epoch": 1.987513572204126, + "grad_norm": 7.3204142422253025, + "learning_rate": 1.3715684636073628e-05, + "loss": 0.7384, + "step": 3661 + }, + { + "epoch": 1.988056460369164, + "grad_norm": 10.339298014193695, + "learning_rate": 1.371241935310242e-05, + "loss": 1.0206, + "step": 3662 + }, + { + "epoch": 1.988599348534202, + "grad_norm": 7.083819102334954, + "learning_rate": 1.3709153610973266e-05, + "loss": 0.6905, + "step": 3663 + }, + { + "epoch": 1.98914223669924, + "grad_norm": 6.318629816327554, + "learning_rate": 1.3705887410090085e-05, + "loss": 0.7946, + "step": 3664 + }, + { + "epoch": 1.989685124864278, + "grad_norm": 7.109392074745231, + "learning_rate": 1.3702620750856843e-05, + "loss": 0.7125, + "step": 3665 + }, + { + "epoch": 1.990228013029316, + "grad_norm": 6.7832401368423545, + "learning_rate": 1.3699353633677565e-05, + "loss": 0.6091, + "step": 3666 + }, + { + "epoch": 1.990770901194354, + "grad_norm": 9.301471508167843, + "learning_rate": 1.3696086058956333e-05, + "loss": 0.9335, + "step": 3667 + }, + { + "epoch": 1.991313789359392, + "grad_norm": 8.167349191349723, + "learning_rate": 1.3692818027097288e-05, + "loss": 0.7148, + "step": 3668 + }, + { + "epoch": 1.99185667752443, + "grad_norm": 7.582331201930941, + "learning_rate": 1.3689549538504622e-05, + "loss": 0.5952, + "step": 3669 + }, + { + "epoch": 1.992399565689468, + "grad_norm": 8.743965167551632, + "learning_rate": 1.3686280593582588e-05, + "loss": 0.9268, + "step": 3670 + }, + { + "epoch": 1.992942453854506, + "grad_norm": 10.291738284427778, + "learning_rate": 1.3683011192735496e-05, + "loss": 0.9934, + "step": 3671 + }, + { + "epoch": 1.993485342019544, + "grad_norm": 9.808936601876956, + "learning_rate": 1.3679741336367711e-05, + "loss": 1.0506, + "step": 3672 + }, + { + "epoch": 1.994028230184582, + "grad_norm": 10.002691567142412, + "learning_rate": 1.3676471024883654e-05, + "loss": 1.1439, + "step": 3673 + }, + { + "epoch": 1.99457111834962, + "grad_norm": 7.558103827510508, + "learning_rate": 1.36732002586878e-05, + "loss": 1.0059, + "step": 3674 + }, + { + "epoch": 1.995114006514658, + "grad_norm": 7.958890420546252, + "learning_rate": 1.3669929038184684e-05, + "loss": 1.0301, + "step": 3675 + }, + { + "epoch": 1.995656894679696, + "grad_norm": 8.042682701022132, + "learning_rate": 1.3666657363778895e-05, + "loss": 0.6881, + "step": 3676 + }, + { + "epoch": 1.996199782844734, + "grad_norm": 7.54271762136155, + "learning_rate": 1.366338523587508e-05, + "loss": 0.5227, + "step": 3677 + }, + { + "epoch": 1.996742671009772, + "grad_norm": 9.060117135659972, + "learning_rate": 1.3660112654877939e-05, + "loss": 0.9847, + "step": 3678 + }, + { + "epoch": 1.99728555917481, + "grad_norm": 11.00270077841202, + "learning_rate": 1.3656839621192233e-05, + "loss": 1.0918, + "step": 3679 + }, + { + "epoch": 1.997828447339848, + "grad_norm": 9.156434005073692, + "learning_rate": 1.3653566135222774e-05, + "loss": 0.8104, + "step": 3680 + }, + { + "epoch": 1.998371335504886, + "grad_norm": 8.405413514743108, + "learning_rate": 1.3650292197374433e-05, + "loss": 0.9414, + "step": 3681 + }, + { + "epoch": 1.998914223669924, + "grad_norm": 7.9594907003419495, + "learning_rate": 1.3647017808052135e-05, + "loss": 0.9257, + "step": 3682 + }, + { + "epoch": 1.999457111834962, + "grad_norm": 9.821777823228977, + "learning_rate": 1.3643742967660859e-05, + "loss": 0.8787, + "step": 3683 + }, + { + "epoch": 2.0, + "grad_norm": 8.766507481527489, + "learning_rate": 1.3640467676605648e-05, + "loss": 0.8472, + "step": 3684 + }, + { + "epoch": 2.000542888165038, + "grad_norm": 9.493942149196295, + "learning_rate": 1.3637191935291596e-05, + "loss": 0.8147, + "step": 3685 + }, + { + "epoch": 2.001085776330076, + "grad_norm": 9.782638508357579, + "learning_rate": 1.3633915744123844e-05, + "loss": 0.8788, + "step": 3686 + }, + { + "epoch": 2.001628664495114, + "grad_norm": 8.46210638839138, + "learning_rate": 1.3630639103507604e-05, + "loss": 1.078, + "step": 3687 + }, + { + "epoch": 2.002171552660152, + "grad_norm": 7.345658753018104, + "learning_rate": 1.3627362013848134e-05, + "loss": 0.6838, + "step": 3688 + }, + { + "epoch": 2.00271444082519, + "grad_norm": 7.329182601936008, + "learning_rate": 1.3624084475550743e-05, + "loss": 0.7249, + "step": 3689 + }, + { + "epoch": 2.003257328990228, + "grad_norm": 9.89510466995638, + "learning_rate": 1.3620806489020813e-05, + "loss": 0.8018, + "step": 3690 + }, + { + "epoch": 2.003800217155266, + "grad_norm": 6.836942603575798, + "learning_rate": 1.3617528054663764e-05, + "loss": 0.6645, + "step": 3691 + }, + { + "epoch": 2.004343105320304, + "grad_norm": 8.842045073299985, + "learning_rate": 1.3614249172885081e-05, + "loss": 0.5143, + "step": 3692 + }, + { + "epoch": 2.004885993485342, + "grad_norm": 7.671200527248645, + "learning_rate": 1.36109698440903e-05, + "loss": 0.8038, + "step": 3693 + }, + { + "epoch": 2.00542888165038, + "grad_norm": 9.544464428119456, + "learning_rate": 1.3607690068685013e-05, + "loss": 0.6396, + "step": 3694 + }, + { + "epoch": 2.005971769815418, + "grad_norm": 8.537202930183216, + "learning_rate": 1.3604409847074868e-05, + "loss": 0.7104, + "step": 3695 + }, + { + "epoch": 2.006514657980456, + "grad_norm": 7.121091005738386, + "learning_rate": 1.3601129179665572e-05, + "loss": 0.4337, + "step": 3696 + }, + { + "epoch": 2.007057546145494, + "grad_norm": 9.740494539210806, + "learning_rate": 1.3597848066862875e-05, + "loss": 0.9114, + "step": 3697 + }, + { + "epoch": 2.007600434310532, + "grad_norm": 8.100951122902112, + "learning_rate": 1.3594566509072599e-05, + "loss": 0.722, + "step": 3698 + }, + { + "epoch": 2.00814332247557, + "grad_norm": 6.959347625774498, + "learning_rate": 1.3591284506700606e-05, + "loss": 0.5429, + "step": 3699 + }, + { + "epoch": 2.008686210640608, + "grad_norm": 9.993493206941414, + "learning_rate": 1.3588002060152822e-05, + "loss": 0.8251, + "step": 3700 + }, + { + "epoch": 2.009229098805646, + "grad_norm": 9.966526343412045, + "learning_rate": 1.3584719169835226e-05, + "loss": 0.4847, + "step": 3701 + }, + { + "epoch": 2.009771986970684, + "grad_norm": 7.888503709599375, + "learning_rate": 1.3581435836153847e-05, + "loss": 0.7941, + "step": 3702 + }, + { + "epoch": 2.010314875135722, + "grad_norm": 9.057972693252164, + "learning_rate": 1.3578152059514778e-05, + "loss": 0.8963, + "step": 3703 + }, + { + "epoch": 2.01085776330076, + "grad_norm": 10.717910066603423, + "learning_rate": 1.3574867840324157e-05, + "loss": 0.6485, + "step": 3704 + }, + { + "epoch": 2.011400651465798, + "grad_norm": 7.257931643777971, + "learning_rate": 1.3571583178988188e-05, + "loss": 0.4569, + "step": 3705 + }, + { + "epoch": 2.011943539630836, + "grad_norm": 16.546104792345954, + "learning_rate": 1.3568298075913119e-05, + "loss": 1.0473, + "step": 3706 + }, + { + "epoch": 2.012486427795874, + "grad_norm": 9.882729561546991, + "learning_rate": 1.3565012531505252e-05, + "loss": 0.9729, + "step": 3707 + }, + { + "epoch": 2.013029315960912, + "grad_norm": 7.060037163975289, + "learning_rate": 1.3561726546170956e-05, + "loss": 0.5935, + "step": 3708 + }, + { + "epoch": 2.01357220412595, + "grad_norm": 7.081117948590212, + "learning_rate": 1.3558440120316644e-05, + "loss": 0.6676, + "step": 3709 + }, + { + "epoch": 2.014115092290988, + "grad_norm": 10.048606586375149, + "learning_rate": 1.3555153254348788e-05, + "loss": 0.6307, + "step": 3710 + }, + { + "epoch": 2.014657980456026, + "grad_norm": 7.720688031890385, + "learning_rate": 1.3551865948673912e-05, + "loss": 0.7882, + "step": 3711 + }, + { + "epoch": 2.015200868621064, + "grad_norm": 7.644495073412038, + "learning_rate": 1.3548578203698592e-05, + "loss": 0.8904, + "step": 3712 + }, + { + "epoch": 2.015743756786102, + "grad_norm": 10.13216917020047, + "learning_rate": 1.3545290019829466e-05, + "loss": 0.7832, + "step": 3713 + }, + { + "epoch": 2.01628664495114, + "grad_norm": 6.321686691568546, + "learning_rate": 1.3542001397473219e-05, + "loss": 0.5682, + "step": 3714 + }, + { + "epoch": 2.016829533116178, + "grad_norm": 7.966895819853924, + "learning_rate": 1.3538712337036594e-05, + "loss": 0.7704, + "step": 3715 + }, + { + "epoch": 2.017372421281216, + "grad_norm": 8.827713158118389, + "learning_rate": 1.3535422838926389e-05, + "loss": 0.6741, + "step": 3716 + }, + { + "epoch": 2.017915309446254, + "grad_norm": 7.404288938912365, + "learning_rate": 1.3532132903549453e-05, + "loss": 0.5325, + "step": 3717 + }, + { + "epoch": 2.018458197611292, + "grad_norm": 9.869324924245625, + "learning_rate": 1.352884253131269e-05, + "loss": 0.7326, + "step": 3718 + }, + { + "epoch": 2.01900108577633, + "grad_norm": 7.824783976445888, + "learning_rate": 1.3525551722623056e-05, + "loss": 0.4683, + "step": 3719 + }, + { + "epoch": 2.019543973941368, + "grad_norm": 10.516375854488775, + "learning_rate": 1.3522260477887566e-05, + "loss": 0.839, + "step": 3720 + }, + { + "epoch": 2.020086862106406, + "grad_norm": 10.698678918246005, + "learning_rate": 1.3518968797513288e-05, + "loss": 1.0513, + "step": 3721 + }, + { + "epoch": 2.020629750271444, + "grad_norm": 9.430167728870957, + "learning_rate": 1.351567668190734e-05, + "loss": 0.6696, + "step": 3722 + }, + { + "epoch": 2.021172638436482, + "grad_norm": 9.619270238605017, + "learning_rate": 1.3512384131476897e-05, + "loss": 0.4608, + "step": 3723 + }, + { + "epoch": 2.02171552660152, + "grad_norm": 6.9328636261250525, + "learning_rate": 1.350909114662919e-05, + "loss": 0.3952, + "step": 3724 + }, + { + "epoch": 2.022258414766558, + "grad_norm": 11.694555068035188, + "learning_rate": 1.3505797727771493e-05, + "loss": 0.7526, + "step": 3725 + }, + { + "epoch": 2.022801302931596, + "grad_norm": 7.417427417334792, + "learning_rate": 1.3502503875311149e-05, + "loss": 0.5239, + "step": 3726 + }, + { + "epoch": 2.023344191096634, + "grad_norm": 10.99931861711161, + "learning_rate": 1.349920958965554e-05, + "loss": 0.5504, + "step": 3727 + }, + { + "epoch": 2.023887079261672, + "grad_norm": 9.418315087755559, + "learning_rate": 1.3495914871212113e-05, + "loss": 0.6791, + "step": 3728 + }, + { + "epoch": 2.02442996742671, + "grad_norm": 12.589472190346669, + "learning_rate": 1.3492619720388363e-05, + "loss": 0.815, + "step": 3729 + }, + { + "epoch": 2.024972855591748, + "grad_norm": 6.60691672927324, + "learning_rate": 1.348932413759184e-05, + "loss": 0.4112, + "step": 3730 + }, + { + "epoch": 2.025515743756786, + "grad_norm": 10.970405529985833, + "learning_rate": 1.3486028123230145e-05, + "loss": 0.6164, + "step": 3731 + }, + { + "epoch": 2.026058631921824, + "grad_norm": 9.419473243249211, + "learning_rate": 1.3482731677710938e-05, + "loss": 0.829, + "step": 3732 + }, + { + "epoch": 2.026601520086862, + "grad_norm": 8.809487998193466, + "learning_rate": 1.3479434801441925e-05, + "loss": 0.6386, + "step": 3733 + }, + { + "epoch": 2.0271444082519, + "grad_norm": 8.586869223296738, + "learning_rate": 1.3476137494830872e-05, + "loss": 0.4586, + "step": 3734 + }, + { + "epoch": 2.027687296416938, + "grad_norm": 10.792985384487755, + "learning_rate": 1.3472839758285595e-05, + "loss": 0.68, + "step": 3735 + }, + { + "epoch": 2.028230184581976, + "grad_norm": 11.494014730485675, + "learning_rate": 1.346954159221396e-05, + "loss": 0.9465, + "step": 3736 + }, + { + "epoch": 2.028773072747014, + "grad_norm": 9.047474116723855, + "learning_rate": 1.3466242997023891e-05, + "loss": 0.637, + "step": 3737 + }, + { + "epoch": 2.029315960912052, + "grad_norm": 7.32591412143244, + "learning_rate": 1.3462943973123362e-05, + "loss": 0.5071, + "step": 3738 + }, + { + "epoch": 2.02985884907709, + "grad_norm": 8.462579835722188, + "learning_rate": 1.3459644520920405e-05, + "loss": 0.5568, + "step": 3739 + }, + { + "epoch": 2.030401737242128, + "grad_norm": 11.667085873745426, + "learning_rate": 1.34563446408231e-05, + "loss": 0.579, + "step": 3740 + }, + { + "epoch": 2.030944625407166, + "grad_norm": 6.778347309469529, + "learning_rate": 1.3453044333239577e-05, + "loss": 0.3399, + "step": 3741 + }, + { + "epoch": 2.031487513572204, + "grad_norm": 7.280157509234148, + "learning_rate": 1.3449743598578033e-05, + "loss": 0.4236, + "step": 3742 + }, + { + "epoch": 2.032030401737242, + "grad_norm": 10.902366947251025, + "learning_rate": 1.34464424372467e-05, + "loss": 0.7984, + "step": 3743 + }, + { + "epoch": 2.03257328990228, + "grad_norm": 7.729938792603154, + "learning_rate": 1.344314084965388e-05, + "loss": 0.5115, + "step": 3744 + }, + { + "epoch": 2.033116178067318, + "grad_norm": 9.034690624481534, + "learning_rate": 1.3439838836207905e-05, + "loss": 0.3528, + "step": 3745 + }, + { + "epoch": 2.033659066232356, + "grad_norm": 10.813371264470662, + "learning_rate": 1.3436536397317183e-05, + "loss": 0.4972, + "step": 3746 + }, + { + "epoch": 2.034201954397394, + "grad_norm": 12.116428074423327, + "learning_rate": 1.343323353339016e-05, + "loss": 0.9414, + "step": 3747 + }, + { + "epoch": 2.034744842562432, + "grad_norm": 11.714995187138708, + "learning_rate": 1.3429930244835343e-05, + "loss": 0.7842, + "step": 3748 + }, + { + "epoch": 2.03528773072747, + "grad_norm": 9.956102297397003, + "learning_rate": 1.3426626532061287e-05, + "loss": 0.7304, + "step": 3749 + }, + { + "epoch": 2.035830618892508, + "grad_norm": 10.670014393137768, + "learning_rate": 1.34233223954766e-05, + "loss": 0.746, + "step": 3750 + }, + { + "epoch": 2.036373507057546, + "grad_norm": 10.266437034084714, + "learning_rate": 1.3420017835489945e-05, + "loss": 1.0871, + "step": 3751 + }, + { + "epoch": 2.036916395222584, + "grad_norm": 11.75190588784515, + "learning_rate": 1.3416712852510033e-05, + "loss": 0.7606, + "step": 3752 + }, + { + "epoch": 2.037459283387622, + "grad_norm": 11.54288048263274, + "learning_rate": 1.3413407446945627e-05, + "loss": 0.5887, + "step": 3753 + }, + { + "epoch": 2.03800217155266, + "grad_norm": 18.002644323614724, + "learning_rate": 1.3410101619205552e-05, + "loss": 1.0593, + "step": 3754 + }, + { + "epoch": 2.038545059717698, + "grad_norm": 8.97759172274721, + "learning_rate": 1.3406795369698671e-05, + "loss": 0.7194, + "step": 3755 + }, + { + "epoch": 2.039087947882736, + "grad_norm": 8.793507308104033, + "learning_rate": 1.3403488698833912e-05, + "loss": 0.4891, + "step": 3756 + }, + { + "epoch": 2.039630836047774, + "grad_norm": 9.263725202719241, + "learning_rate": 1.3400181607020243e-05, + "loss": 0.8258, + "step": 3757 + }, + { + "epoch": 2.040173724212812, + "grad_norm": 10.568214203997627, + "learning_rate": 1.3396874094666694e-05, + "loss": 0.5841, + "step": 3758 + }, + { + "epoch": 2.04071661237785, + "grad_norm": 10.237711379932081, + "learning_rate": 1.3393566162182346e-05, + "loss": 0.6901, + "step": 3759 + }, + { + "epoch": 2.041259500542888, + "grad_norm": 11.264856346485661, + "learning_rate": 1.3390257809976322e-05, + "loss": 0.8029, + "step": 3760 + }, + { + "epoch": 2.041802388707926, + "grad_norm": 7.504173750209935, + "learning_rate": 1.3386949038457813e-05, + "loss": 0.3739, + "step": 3761 + }, + { + "epoch": 2.042345276872964, + "grad_norm": 8.099570741006275, + "learning_rate": 1.3383639848036044e-05, + "loss": 0.7457, + "step": 3762 + }, + { + "epoch": 2.042888165038002, + "grad_norm": 9.87563463161881, + "learning_rate": 1.3380330239120313e-05, + "loss": 0.632, + "step": 3763 + }, + { + "epoch": 2.04343105320304, + "grad_norm": 9.304965814712034, + "learning_rate": 1.3377020212119946e-05, + "loss": 0.657, + "step": 3764 + }, + { + "epoch": 2.043973941368078, + "grad_norm": 7.738549848691211, + "learning_rate": 1.3373709767444339e-05, + "loss": 0.4198, + "step": 3765 + }, + { + "epoch": 2.044516829533116, + "grad_norm": 11.805824611419341, + "learning_rate": 1.3370398905502928e-05, + "loss": 0.8173, + "step": 3766 + }, + { + "epoch": 2.045059717698154, + "grad_norm": 10.425380865754251, + "learning_rate": 1.3367087626705211e-05, + "loss": 0.7149, + "step": 3767 + }, + { + "epoch": 2.045602605863192, + "grad_norm": 11.884153232685467, + "learning_rate": 1.336377593146073e-05, + "loss": 0.723, + "step": 3768 + }, + { + "epoch": 2.04614549402823, + "grad_norm": 9.663581488842127, + "learning_rate": 1.336046382017908e-05, + "loss": 0.6457, + "step": 3769 + }, + { + "epoch": 2.046688382193268, + "grad_norm": 9.184284700313228, + "learning_rate": 1.335715129326991e-05, + "loss": 0.6645, + "step": 3770 + }, + { + "epoch": 2.047231270358306, + "grad_norm": 7.398657595562234, + "learning_rate": 1.3353838351142915e-05, + "loss": 0.7288, + "step": 3771 + }, + { + "epoch": 2.047774158523344, + "grad_norm": 10.416946087073072, + "learning_rate": 1.335052499420785e-05, + "loss": 0.4974, + "step": 3772 + }, + { + "epoch": 2.048317046688382, + "grad_norm": 8.676282075018099, + "learning_rate": 1.3347211222874514e-05, + "loss": 0.7128, + "step": 3773 + }, + { + "epoch": 2.04885993485342, + "grad_norm": 10.339910800835435, + "learning_rate": 1.3343897037552758e-05, + "loss": 0.6674, + "step": 3774 + }, + { + "epoch": 2.049402823018458, + "grad_norm": 10.976308424624598, + "learning_rate": 1.3340582438652488e-05, + "loss": 0.7683, + "step": 3775 + }, + { + "epoch": 2.049945711183496, + "grad_norm": 11.874329810000683, + "learning_rate": 1.3337267426583658e-05, + "loss": 1.2199, + "step": 3776 + }, + { + "epoch": 2.050488599348534, + "grad_norm": 9.306473243935873, + "learning_rate": 1.3333952001756272e-05, + "loss": 0.8681, + "step": 3777 + }, + { + "epoch": 2.0510314875135722, + "grad_norm": 9.083116427983501, + "learning_rate": 1.333063616458039e-05, + "loss": 0.7512, + "step": 3778 + }, + { + "epoch": 2.05157437567861, + "grad_norm": 5.762711234835306, + "learning_rate": 1.3327319915466119e-05, + "loss": 0.3384, + "step": 3779 + }, + { + "epoch": 2.0521172638436482, + "grad_norm": 7.431272084519616, + "learning_rate": 1.332400325482362e-05, + "loss": 0.5188, + "step": 3780 + }, + { + "epoch": 2.052660152008686, + "grad_norm": 5.736965263179954, + "learning_rate": 1.3320686183063096e-05, + "loss": 0.3476, + "step": 3781 + }, + { + "epoch": 2.0532030401737242, + "grad_norm": 7.130870597882685, + "learning_rate": 1.3317368700594815e-05, + "loss": 0.7929, + "step": 3782 + }, + { + "epoch": 2.053745928338762, + "grad_norm": 7.713657404988891, + "learning_rate": 1.3314050807829088e-05, + "loss": 0.422, + "step": 3783 + }, + { + "epoch": 2.0542888165038002, + "grad_norm": 6.383542293487279, + "learning_rate": 1.3310732505176276e-05, + "loss": 0.568, + "step": 3784 + }, + { + "epoch": 2.054831704668838, + "grad_norm": 12.326747003957852, + "learning_rate": 1.3307413793046787e-05, + "loss": 0.8061, + "step": 3785 + }, + { + "epoch": 2.0553745928338762, + "grad_norm": 10.668098962086818, + "learning_rate": 1.330409467185109e-05, + "loss": 1.4344, + "step": 3786 + }, + { + "epoch": 2.055917480998914, + "grad_norm": 9.844537720560512, + "learning_rate": 1.3300775141999698e-05, + "loss": 1.0089, + "step": 3787 + }, + { + "epoch": 2.0564603691639523, + "grad_norm": 9.714672703692647, + "learning_rate": 1.3297455203903176e-05, + "loss": 0.6069, + "step": 3788 + }, + { + "epoch": 2.05700325732899, + "grad_norm": 10.27882170416886, + "learning_rate": 1.3294134857972139e-05, + "loss": 0.7279, + "step": 3789 + }, + { + "epoch": 2.0575461454940283, + "grad_norm": 10.99914755541959, + "learning_rate": 1.3290814104617253e-05, + "loss": 0.6787, + "step": 3790 + }, + { + "epoch": 2.058089033659066, + "grad_norm": 13.627823006472678, + "learning_rate": 1.3287492944249234e-05, + "loss": 0.9052, + "step": 3791 + }, + { + "epoch": 2.0586319218241043, + "grad_norm": 7.734783593910536, + "learning_rate": 1.3284171377278849e-05, + "loss": 0.7141, + "step": 3792 + }, + { + "epoch": 2.059174809989142, + "grad_norm": 9.247896244835857, + "learning_rate": 1.3280849404116913e-05, + "loss": 0.6071, + "step": 3793 + }, + { + "epoch": 2.0597176981541803, + "grad_norm": 9.004013607687236, + "learning_rate": 1.3277527025174295e-05, + "loss": 0.5887, + "step": 3794 + }, + { + "epoch": 2.060260586319218, + "grad_norm": 9.62029412051341, + "learning_rate": 1.3274204240861908e-05, + "loss": 0.8847, + "step": 3795 + }, + { + "epoch": 2.0608034744842563, + "grad_norm": 7.500348953035017, + "learning_rate": 1.3270881051590725e-05, + "loss": 0.4769, + "step": 3796 + }, + { + "epoch": 2.061346362649294, + "grad_norm": 10.153144580789224, + "learning_rate": 1.326755745777176e-05, + "loss": 0.645, + "step": 3797 + }, + { + "epoch": 2.0618892508143323, + "grad_norm": 6.813155627292939, + "learning_rate": 1.326423345981608e-05, + "loss": 0.6711, + "step": 3798 + }, + { + "epoch": 2.06243213897937, + "grad_norm": 9.362898780009592, + "learning_rate": 1.32609090581348e-05, + "loss": 0.5464, + "step": 3799 + }, + { + "epoch": 2.0629750271444083, + "grad_norm": 8.540638734313639, + "learning_rate": 1.3257584253139096e-05, + "loss": 0.6523, + "step": 3800 + }, + { + "epoch": 2.063517915309446, + "grad_norm": 9.000753471311711, + "learning_rate": 1.3254259045240176e-05, + "loss": 0.5362, + "step": 3801 + }, + { + "epoch": 2.0640608034744843, + "grad_norm": 10.275669325762056, + "learning_rate": 1.3250933434849316e-05, + "loss": 0.6061, + "step": 3802 + }, + { + "epoch": 2.064603691639522, + "grad_norm": 11.615294601299743, + "learning_rate": 1.3247607422377823e-05, + "loss": 0.673, + "step": 3803 + }, + { + "epoch": 2.0651465798045603, + "grad_norm": 10.316289286330115, + "learning_rate": 1.324428100823707e-05, + "loss": 0.7997, + "step": 3804 + }, + { + "epoch": 2.065689467969598, + "grad_norm": 14.555740784089037, + "learning_rate": 1.3240954192838472e-05, + "loss": 0.9957, + "step": 3805 + }, + { + "epoch": 2.0662323561346363, + "grad_norm": 7.674873455902871, + "learning_rate": 1.323762697659349e-05, + "loss": 0.4771, + "step": 3806 + }, + { + "epoch": 2.066775244299674, + "grad_norm": 8.530555441214027, + "learning_rate": 1.3234299359913647e-05, + "loss": 0.8666, + "step": 3807 + }, + { + "epoch": 2.0673181324647123, + "grad_norm": 6.463731480876421, + "learning_rate": 1.3230971343210503e-05, + "loss": 0.5742, + "step": 3808 + }, + { + "epoch": 2.06786102062975, + "grad_norm": 8.074017879990592, + "learning_rate": 1.3227642926895676e-05, + "loss": 0.3637, + "step": 3809 + }, + { + "epoch": 2.0684039087947883, + "grad_norm": 9.989752857060338, + "learning_rate": 1.3224314111380828e-05, + "loss": 0.7389, + "step": 3810 + }, + { + "epoch": 2.068946796959826, + "grad_norm": 12.892065053280843, + "learning_rate": 1.3220984897077669e-05, + "loss": 1.2959, + "step": 3811 + }, + { + "epoch": 2.0694896851248643, + "grad_norm": 7.497956491794191, + "learning_rate": 1.3217655284397965e-05, + "loss": 0.4384, + "step": 3812 + }, + { + "epoch": 2.070032573289902, + "grad_norm": 14.719876212901973, + "learning_rate": 1.3214325273753528e-05, + "loss": 1.3054, + "step": 3813 + }, + { + "epoch": 2.0705754614549403, + "grad_norm": 10.767865599283661, + "learning_rate": 1.3210994865556219e-05, + "loss": 0.9961, + "step": 3814 + }, + { + "epoch": 2.071118349619978, + "grad_norm": 9.198818655810543, + "learning_rate": 1.3207664060217946e-05, + "loss": 0.5916, + "step": 3815 + }, + { + "epoch": 2.0716612377850163, + "grad_norm": 9.598428080121797, + "learning_rate": 1.320433285815067e-05, + "loss": 0.7907, + "step": 3816 + }, + { + "epoch": 2.072204125950054, + "grad_norm": 9.48389213525687, + "learning_rate": 1.32010012597664e-05, + "loss": 0.8385, + "step": 3817 + }, + { + "epoch": 2.0727470141150923, + "grad_norm": 7.562927002528845, + "learning_rate": 1.3197669265477191e-05, + "loss": 0.6748, + "step": 3818 + }, + { + "epoch": 2.07328990228013, + "grad_norm": 7.5241368226931185, + "learning_rate": 1.319433687569515e-05, + "loss": 0.5605, + "step": 3819 + }, + { + "epoch": 2.0738327904451683, + "grad_norm": 7.952800933557159, + "learning_rate": 1.3191004090832436e-05, + "loss": 0.6624, + "step": 3820 + }, + { + "epoch": 2.074375678610206, + "grad_norm": 8.02814530035842, + "learning_rate": 1.318767091130125e-05, + "loss": 0.488, + "step": 3821 + }, + { + "epoch": 2.0749185667752443, + "grad_norm": 7.000417608550192, + "learning_rate": 1.3184337337513849e-05, + "loss": 0.7194, + "step": 3822 + }, + { + "epoch": 2.075461454940282, + "grad_norm": 8.965861432524546, + "learning_rate": 1.3181003369882527e-05, + "loss": 0.4823, + "step": 3823 + }, + { + "epoch": 2.0760043431053203, + "grad_norm": 8.717885349674289, + "learning_rate": 1.3177669008819635e-05, + "loss": 0.6482, + "step": 3824 + }, + { + "epoch": 2.076547231270358, + "grad_norm": 9.217767101253152, + "learning_rate": 1.317433425473758e-05, + "loss": 0.5276, + "step": 3825 + }, + { + "epoch": 2.0770901194353963, + "grad_norm": 8.84305940258517, + "learning_rate": 1.3170999108048804e-05, + "loss": 0.6579, + "step": 3826 + }, + { + "epoch": 2.077633007600434, + "grad_norm": 7.501762934992334, + "learning_rate": 1.3167663569165803e-05, + "loss": 0.6633, + "step": 3827 + }, + { + "epoch": 2.0781758957654723, + "grad_norm": 7.815925996775423, + "learning_rate": 1.3164327638501126e-05, + "loss": 0.6881, + "step": 3828 + }, + { + "epoch": 2.07871878393051, + "grad_norm": 11.789396888417095, + "learning_rate": 1.3160991316467362e-05, + "loss": 0.5761, + "step": 3829 + }, + { + "epoch": 2.0792616720955484, + "grad_norm": 8.450719528350264, + "learning_rate": 1.3157654603477155e-05, + "loss": 0.7263, + "step": 3830 + }, + { + "epoch": 2.079804560260586, + "grad_norm": 9.544428310813693, + "learning_rate": 1.315431749994319e-05, + "loss": 0.6408, + "step": 3831 + }, + { + "epoch": 2.0803474484256244, + "grad_norm": 7.454950880630089, + "learning_rate": 1.3150980006278219e-05, + "loss": 0.4425, + "step": 3832 + }, + { + "epoch": 2.080890336590662, + "grad_norm": 11.462923993924347, + "learning_rate": 1.314764212289501e-05, + "loss": 0.5349, + "step": 3833 + }, + { + "epoch": 2.0814332247557004, + "grad_norm": 8.900231872199269, + "learning_rate": 1.314430385020641e-05, + "loss": 0.8022, + "step": 3834 + }, + { + "epoch": 2.081976112920738, + "grad_norm": 7.035528648373333, + "learning_rate": 1.3140965188625299e-05, + "loss": 0.6513, + "step": 3835 + }, + { + "epoch": 2.0825190010857764, + "grad_norm": 12.327229999200162, + "learning_rate": 1.3137626138564606e-05, + "loss": 1.0623, + "step": 3836 + }, + { + "epoch": 2.083061889250814, + "grad_norm": 8.942692654829559, + "learning_rate": 1.3134286700437308e-05, + "loss": 0.4689, + "step": 3837 + }, + { + "epoch": 2.0836047774158524, + "grad_norm": 11.092455380734007, + "learning_rate": 1.313094687465644e-05, + "loss": 0.6808, + "step": 3838 + }, + { + "epoch": 2.08414766558089, + "grad_norm": 8.918018917563646, + "learning_rate": 1.3127606661635075e-05, + "loss": 0.523, + "step": 3839 + }, + { + "epoch": 2.0846905537459284, + "grad_norm": 13.878014157853947, + "learning_rate": 1.312426606178633e-05, + "loss": 1.0231, + "step": 3840 + }, + { + "epoch": 2.085233441910966, + "grad_norm": 10.432596711564004, + "learning_rate": 1.3120925075523379e-05, + "loss": 0.8011, + "step": 3841 + }, + { + "epoch": 2.0857763300760044, + "grad_norm": 8.126214858686065, + "learning_rate": 1.3117583703259445e-05, + "loss": 0.5272, + "step": 3842 + }, + { + "epoch": 2.086319218241042, + "grad_norm": 9.596541824389169, + "learning_rate": 1.3114241945407783e-05, + "loss": 0.6585, + "step": 3843 + }, + { + "epoch": 2.0868621064060804, + "grad_norm": 10.438282441705661, + "learning_rate": 1.3110899802381718e-05, + "loss": 0.8474, + "step": 3844 + }, + { + "epoch": 2.087404994571118, + "grad_norm": 11.98868629313521, + "learning_rate": 1.3107557274594607e-05, + "loss": 1.0492, + "step": 3845 + }, + { + "epoch": 2.0879478827361564, + "grad_norm": 9.734289423556795, + "learning_rate": 1.310421436245986e-05, + "loss": 0.6573, + "step": 3846 + }, + { + "epoch": 2.088490770901194, + "grad_norm": 10.196913105458549, + "learning_rate": 1.310087106639093e-05, + "loss": 0.8964, + "step": 3847 + }, + { + "epoch": 2.0890336590662324, + "grad_norm": 9.155222462665467, + "learning_rate": 1.3097527386801327e-05, + "loss": 0.6668, + "step": 3848 + }, + { + "epoch": 2.08957654723127, + "grad_norm": 8.512351137441192, + "learning_rate": 1.3094183324104602e-05, + "loss": 0.6084, + "step": 3849 + }, + { + "epoch": 2.0901194353963084, + "grad_norm": 7.62701351090082, + "learning_rate": 1.3090838878714349e-05, + "loss": 0.4199, + "step": 3850 + }, + { + "epoch": 2.090662323561346, + "grad_norm": 8.57862015121728, + "learning_rate": 1.3087494051044218e-05, + "loss": 0.3694, + "step": 3851 + }, + { + "epoch": 2.0912052117263844, + "grad_norm": 11.407572249271688, + "learning_rate": 1.3084148841507904e-05, + "loss": 0.8861, + "step": 3852 + }, + { + "epoch": 2.091748099891422, + "grad_norm": 9.297247466948287, + "learning_rate": 1.3080803250519142e-05, + "loss": 0.8432, + "step": 3853 + }, + { + "epoch": 2.0922909880564604, + "grad_norm": 7.909660106539073, + "learning_rate": 1.3077457278491728e-05, + "loss": 0.5103, + "step": 3854 + }, + { + "epoch": 2.092833876221498, + "grad_norm": 11.013011688454522, + "learning_rate": 1.3074110925839491e-05, + "loss": 1.0117, + "step": 3855 + }, + { + "epoch": 2.0933767643865364, + "grad_norm": 10.249008356283914, + "learning_rate": 1.3070764192976315e-05, + "loss": 0.5248, + "step": 3856 + }, + { + "epoch": 2.093919652551574, + "grad_norm": 8.240438304204865, + "learning_rate": 1.306741708031613e-05, + "loss": 0.6811, + "step": 3857 + }, + { + "epoch": 2.0944625407166124, + "grad_norm": 9.542880797186081, + "learning_rate": 1.3064069588272913e-05, + "loss": 0.4746, + "step": 3858 + }, + { + "epoch": 2.09500542888165, + "grad_norm": 9.251214264026915, + "learning_rate": 1.3060721717260685e-05, + "loss": 0.6405, + "step": 3859 + }, + { + "epoch": 2.0955483170466884, + "grad_norm": 12.366199541130733, + "learning_rate": 1.3057373467693515e-05, + "loss": 0.8378, + "step": 3860 + }, + { + "epoch": 2.096091205211726, + "grad_norm": 10.566968498265949, + "learning_rate": 1.3054024839985526e-05, + "loss": 0.5975, + "step": 3861 + }, + { + "epoch": 2.0966340933767644, + "grad_norm": 9.58134406953771, + "learning_rate": 1.3050675834550872e-05, + "loss": 0.6846, + "step": 3862 + }, + { + "epoch": 2.097176981541802, + "grad_norm": 8.437576527492594, + "learning_rate": 1.3047326451803772e-05, + "loss": 0.6301, + "step": 3863 + }, + { + "epoch": 2.0977198697068404, + "grad_norm": 8.68957421381677, + "learning_rate": 1.304397669215848e-05, + "loss": 0.6121, + "step": 3864 + }, + { + "epoch": 2.098262757871878, + "grad_norm": 12.281098753982102, + "learning_rate": 1.30406265560293e-05, + "loss": 1.174, + "step": 3865 + }, + { + "epoch": 2.0988056460369164, + "grad_norm": 12.64390360725308, + "learning_rate": 1.303727604383058e-05, + "loss": 0.727, + "step": 3866 + }, + { + "epoch": 2.099348534201954, + "grad_norm": 8.982715547555616, + "learning_rate": 1.3033925155976718e-05, + "loss": 0.4946, + "step": 3867 + }, + { + "epoch": 2.0998914223669924, + "grad_norm": 8.77924533690129, + "learning_rate": 1.303057389288216e-05, + "loss": 0.5027, + "step": 3868 + }, + { + "epoch": 2.1004343105320302, + "grad_norm": 12.246059108799978, + "learning_rate": 1.302722225496139e-05, + "loss": 0.6522, + "step": 3869 + }, + { + "epoch": 2.1009771986970684, + "grad_norm": 13.95610327170378, + "learning_rate": 1.3023870242628944e-05, + "loss": 1.043, + "step": 3870 + }, + { + "epoch": 2.1015200868621062, + "grad_norm": 10.355142745873641, + "learning_rate": 1.3020517856299413e-05, + "loss": 0.9167, + "step": 3871 + }, + { + "epoch": 2.1020629750271445, + "grad_norm": 8.43332484970558, + "learning_rate": 1.3017165096387419e-05, + "loss": 0.5952, + "step": 3872 + }, + { + "epoch": 2.1026058631921822, + "grad_norm": 9.224204311836482, + "learning_rate": 1.3013811963307634e-05, + "loss": 0.5631, + "step": 3873 + }, + { + "epoch": 2.1031487513572205, + "grad_norm": 9.719137418718985, + "learning_rate": 1.3010458457474784e-05, + "loss": 0.9293, + "step": 3874 + }, + { + "epoch": 2.1036916395222582, + "grad_norm": 8.234100299546002, + "learning_rate": 1.3007104579303636e-05, + "loss": 0.7041, + "step": 3875 + }, + { + "epoch": 2.1042345276872965, + "grad_norm": 11.0180135357339, + "learning_rate": 1.3003750329208995e-05, + "loss": 0.6068, + "step": 3876 + }, + { + "epoch": 2.1047774158523342, + "grad_norm": 8.790070650912016, + "learning_rate": 1.300039570760573e-05, + "loss": 0.4956, + "step": 3877 + }, + { + "epoch": 2.1053203040173725, + "grad_norm": 8.43754180724615, + "learning_rate": 1.2997040714908742e-05, + "loss": 0.3965, + "step": 3878 + }, + { + "epoch": 2.1058631921824102, + "grad_norm": 11.815520632897917, + "learning_rate": 1.299368535153298e-05, + "loss": 0.5963, + "step": 3879 + }, + { + "epoch": 2.1064060803474485, + "grad_norm": 14.168765746771772, + "learning_rate": 1.2990329617893445e-05, + "loss": 0.8111, + "step": 3880 + }, + { + "epoch": 2.1069489685124863, + "grad_norm": 5.706418305089391, + "learning_rate": 1.2986973514405176e-05, + "loss": 0.41, + "step": 3881 + }, + { + "epoch": 2.1074918566775245, + "grad_norm": 10.870432547178469, + "learning_rate": 1.2983617041483261e-05, + "loss": 0.7805, + "step": 3882 + }, + { + "epoch": 2.1080347448425623, + "grad_norm": 9.440648088759033, + "learning_rate": 1.2980260199542838e-05, + "loss": 0.6047, + "step": 3883 + }, + { + "epoch": 2.1085776330076005, + "grad_norm": 12.145529985557056, + "learning_rate": 1.2976902988999081e-05, + "loss": 0.8366, + "step": 3884 + }, + { + "epoch": 2.1091205211726383, + "grad_norm": 9.17886258135476, + "learning_rate": 1.2973545410267218e-05, + "loss": 0.6402, + "step": 3885 + }, + { + "epoch": 2.1096634093376765, + "grad_norm": 9.613840291682566, + "learning_rate": 1.297018746376252e-05, + "loss": 0.8175, + "step": 3886 + }, + { + "epoch": 2.1102062975027143, + "grad_norm": 12.898243756363112, + "learning_rate": 1.2966829149900304e-05, + "loss": 0.8514, + "step": 3887 + }, + { + "epoch": 2.1107491856677525, + "grad_norm": 8.825284004643853, + "learning_rate": 1.2963470469095928e-05, + "loss": 0.4508, + "step": 3888 + }, + { + "epoch": 2.1112920738327903, + "grad_norm": 9.953568896257451, + "learning_rate": 1.2960111421764803e-05, + "loss": 0.5657, + "step": 3889 + }, + { + "epoch": 2.1118349619978285, + "grad_norm": 8.988870757227582, + "learning_rate": 1.2956752008322378e-05, + "loss": 0.7242, + "step": 3890 + }, + { + "epoch": 2.1123778501628663, + "grad_norm": 9.68770492429417, + "learning_rate": 1.2953392229184156e-05, + "loss": 0.9403, + "step": 3891 + }, + { + "epoch": 2.1129207383279045, + "grad_norm": 11.903522492661367, + "learning_rate": 1.2950032084765674e-05, + "loss": 0.8743, + "step": 3892 + }, + { + "epoch": 2.1134636264929423, + "grad_norm": 12.24198259147124, + "learning_rate": 1.294667157548252e-05, + "loss": 0.9861, + "step": 3893 + }, + { + "epoch": 2.1140065146579805, + "grad_norm": 13.105368254945933, + "learning_rate": 1.2943310701750331e-05, + "loss": 0.8815, + "step": 3894 + }, + { + "epoch": 2.1145494028230183, + "grad_norm": 9.322992215038694, + "learning_rate": 1.2939949463984782e-05, + "loss": 0.4986, + "step": 3895 + }, + { + "epoch": 2.1150922909880565, + "grad_norm": 7.59330739725767, + "learning_rate": 1.29365878626016e-05, + "loss": 0.4343, + "step": 3896 + }, + { + "epoch": 2.1156351791530943, + "grad_norm": 9.782670203398764, + "learning_rate": 1.293322589801655e-05, + "loss": 0.4843, + "step": 3897 + }, + { + "epoch": 2.1161780673181325, + "grad_norm": 7.130500958222279, + "learning_rate": 1.2929863570645446e-05, + "loss": 0.3784, + "step": 3898 + }, + { + "epoch": 2.1167209554831703, + "grad_norm": 7.336707110452157, + "learning_rate": 1.2926500880904147e-05, + "loss": 0.4938, + "step": 3899 + }, + { + "epoch": 2.1172638436482085, + "grad_norm": 7.047478722817777, + "learning_rate": 1.2923137829208555e-05, + "loss": 0.3002, + "step": 3900 + }, + { + "epoch": 2.1178067318132463, + "grad_norm": 9.533093539252569, + "learning_rate": 1.2919774415974616e-05, + "loss": 0.711, + "step": 3901 + }, + { + "epoch": 2.1183496199782845, + "grad_norm": 8.468755965185016, + "learning_rate": 1.2916410641618324e-05, + "loss": 0.5491, + "step": 3902 + }, + { + "epoch": 2.1188925081433223, + "grad_norm": 10.351769959990103, + "learning_rate": 1.2913046506555715e-05, + "loss": 0.6269, + "step": 3903 + }, + { + "epoch": 2.1194353963083605, + "grad_norm": 10.2212813809041, + "learning_rate": 1.2909682011202875e-05, + "loss": 0.7249, + "step": 3904 + }, + { + "epoch": 2.1199782844733983, + "grad_norm": 9.344205468199693, + "learning_rate": 1.2906317155975922e-05, + "loss": 0.6277, + "step": 3905 + }, + { + "epoch": 2.1205211726384365, + "grad_norm": 8.704990474328149, + "learning_rate": 1.2902951941291035e-05, + "loss": 0.6317, + "step": 3906 + }, + { + "epoch": 2.1210640608034743, + "grad_norm": 11.644285794605782, + "learning_rate": 1.2899586367564422e-05, + "loss": 0.8585, + "step": 3907 + }, + { + "epoch": 2.1216069489685125, + "grad_norm": 8.462034849045626, + "learning_rate": 1.2896220435212347e-05, + "loss": 0.5174, + "step": 3908 + }, + { + "epoch": 2.1221498371335503, + "grad_norm": 10.023149647380059, + "learning_rate": 1.2892854144651112e-05, + "loss": 0.6044, + "step": 3909 + }, + { + "epoch": 2.1226927252985885, + "grad_norm": 9.437366874691593, + "learning_rate": 1.2889487496297068e-05, + "loss": 0.5081, + "step": 3910 + }, + { + "epoch": 2.1232356134636263, + "grad_norm": 11.06235798573597, + "learning_rate": 1.2886120490566607e-05, + "loss": 0.7385, + "step": 3911 + }, + { + "epoch": 2.1237785016286646, + "grad_norm": 9.45153797141243, + "learning_rate": 1.288275312787616e-05, + "loss": 0.6165, + "step": 3912 + }, + { + "epoch": 2.1243213897937023, + "grad_norm": 8.907808417684745, + "learning_rate": 1.2879385408642214e-05, + "loss": 0.3956, + "step": 3913 + }, + { + "epoch": 2.1248642779587406, + "grad_norm": 12.022235830361943, + "learning_rate": 1.2876017333281288e-05, + "loss": 0.7492, + "step": 3914 + }, + { + "epoch": 2.1254071661237783, + "grad_norm": 10.18229463165741, + "learning_rate": 1.2872648902209957e-05, + "loss": 1.0188, + "step": 3915 + }, + { + "epoch": 2.1259500542888166, + "grad_norm": 8.068860605172112, + "learning_rate": 1.2869280115844831e-05, + "loss": 0.4258, + "step": 3916 + }, + { + "epoch": 2.1264929424538543, + "grad_norm": 7.300141006448243, + "learning_rate": 1.2865910974602568e-05, + "loss": 0.414, + "step": 3917 + }, + { + "epoch": 2.1270358306188926, + "grad_norm": 9.447874959421453, + "learning_rate": 1.2862541478899868e-05, + "loss": 0.6323, + "step": 3918 + }, + { + "epoch": 2.1275787187839303, + "grad_norm": 11.845175678539553, + "learning_rate": 1.2859171629153476e-05, + "loss": 0.8539, + "step": 3919 + }, + { + "epoch": 2.1281216069489686, + "grad_norm": 8.29306868561595, + "learning_rate": 1.2855801425780179e-05, + "loss": 0.7127, + "step": 3920 + }, + { + "epoch": 2.1286644951140063, + "grad_norm": 8.040765832896788, + "learning_rate": 1.285243086919681e-05, + "loss": 0.495, + "step": 3921 + }, + { + "epoch": 2.1292073832790446, + "grad_norm": 13.042470321623425, + "learning_rate": 1.2849059959820248e-05, + "loss": 0.616, + "step": 3922 + }, + { + "epoch": 2.1297502714440824, + "grad_norm": 8.585317997471494, + "learning_rate": 1.2845688698067406e-05, + "loss": 0.5599, + "step": 3923 + }, + { + "epoch": 2.1302931596091206, + "grad_norm": 9.356896896219936, + "learning_rate": 1.2842317084355251e-05, + "loss": 0.605, + "step": 3924 + }, + { + "epoch": 2.1308360477741584, + "grad_norm": 17.057557558601207, + "learning_rate": 1.283894511910079e-05, + "loss": 0.9396, + "step": 3925 + }, + { + "epoch": 2.1313789359391966, + "grad_norm": 10.659876178353805, + "learning_rate": 1.2835572802721072e-05, + "loss": 0.765, + "step": 3926 + }, + { + "epoch": 2.1319218241042344, + "grad_norm": 9.903600273604221, + "learning_rate": 1.2832200135633191e-05, + "loss": 0.4745, + "step": 3927 + }, + { + "epoch": 2.1324647122692726, + "grad_norm": 12.075983241730247, + "learning_rate": 1.2828827118254279e-05, + "loss": 0.8613, + "step": 3928 + }, + { + "epoch": 2.1330076004343104, + "grad_norm": 10.13381566988187, + "learning_rate": 1.2825453751001526e-05, + "loss": 0.7342, + "step": 3929 + }, + { + "epoch": 2.1335504885993486, + "grad_norm": 12.671225163451464, + "learning_rate": 1.282208003429215e-05, + "loss": 0.7765, + "step": 3930 + }, + { + "epoch": 2.1340933767643864, + "grad_norm": 10.503393267034513, + "learning_rate": 1.2818705968543417e-05, + "loss": 0.7042, + "step": 3931 + }, + { + "epoch": 2.1346362649294246, + "grad_norm": 7.8720699552641635, + "learning_rate": 1.2815331554172634e-05, + "loss": 0.4393, + "step": 3932 + }, + { + "epoch": 2.1351791530944624, + "grad_norm": 14.057115653429742, + "learning_rate": 1.281195679159716e-05, + "loss": 0.8676, + "step": 3933 + }, + { + "epoch": 2.1357220412595006, + "grad_norm": 12.006814229853843, + "learning_rate": 1.2808581681234387e-05, + "loss": 0.8764, + "step": 3934 + }, + { + "epoch": 2.1362649294245384, + "grad_norm": 11.511042628784308, + "learning_rate": 1.2805206223501756e-05, + "loss": 0.6034, + "step": 3935 + }, + { + "epoch": 2.1368078175895766, + "grad_norm": 6.922027915468186, + "learning_rate": 1.2801830418816749e-05, + "loss": 0.4303, + "step": 3936 + }, + { + "epoch": 2.1373507057546144, + "grad_norm": 8.583211919241508, + "learning_rate": 1.2798454267596892e-05, + "loss": 0.6059, + "step": 3937 + }, + { + "epoch": 2.1378935939196526, + "grad_norm": 10.400540758946061, + "learning_rate": 1.2795077770259749e-05, + "loss": 0.7327, + "step": 3938 + }, + { + "epoch": 2.1384364820846904, + "grad_norm": 11.253848098269156, + "learning_rate": 1.2791700927222932e-05, + "loss": 0.7247, + "step": 3939 + }, + { + "epoch": 2.1389793702497286, + "grad_norm": 10.650957738366092, + "learning_rate": 1.2788323738904098e-05, + "loss": 0.8589, + "step": 3940 + }, + { + "epoch": 2.1395222584147664, + "grad_norm": 7.284559079353519, + "learning_rate": 1.2784946205720936e-05, + "loss": 0.3361, + "step": 3941 + }, + { + "epoch": 2.1400651465798046, + "grad_norm": 10.532878125436458, + "learning_rate": 1.2781568328091192e-05, + "loss": 0.8466, + "step": 3942 + }, + { + "epoch": 2.1406080347448424, + "grad_norm": 16.5717179039462, + "learning_rate": 1.2778190106432643e-05, + "loss": 1.1453, + "step": 3943 + }, + { + "epoch": 2.1411509229098806, + "grad_norm": 8.591001506273539, + "learning_rate": 1.2774811541163114e-05, + "loss": 0.5379, + "step": 3944 + }, + { + "epoch": 2.1416938110749184, + "grad_norm": 11.173192436567485, + "learning_rate": 1.2771432632700471e-05, + "loss": 1.0715, + "step": 3945 + }, + { + "epoch": 2.1422366992399566, + "grad_norm": 11.381968938636014, + "learning_rate": 1.2768053381462625e-05, + "loss": 0.8903, + "step": 3946 + }, + { + "epoch": 2.1427795874049944, + "grad_norm": 9.847481766271917, + "learning_rate": 1.2764673787867519e-05, + "loss": 0.5874, + "step": 3947 + }, + { + "epoch": 2.1433224755700326, + "grad_norm": 9.183549346287741, + "learning_rate": 1.2761293852333156e-05, + "loss": 0.6745, + "step": 3948 + }, + { + "epoch": 2.1438653637350704, + "grad_norm": 9.610890909648008, + "learning_rate": 1.2757913575277572e-05, + "loss": 0.9781, + "step": 3949 + }, + { + "epoch": 2.1444082519001086, + "grad_norm": 9.814885217738354, + "learning_rate": 1.275453295711884e-05, + "loss": 0.9977, + "step": 3950 + }, + { + "epoch": 2.1449511400651464, + "grad_norm": 10.984041832783662, + "learning_rate": 1.275115199827508e-05, + "loss": 0.7822, + "step": 3951 + }, + { + "epoch": 2.1454940282301846, + "grad_norm": 9.634528803708669, + "learning_rate": 1.2747770699164457e-05, + "loss": 1.1719, + "step": 3952 + }, + { + "epoch": 2.1460369163952224, + "grad_norm": 9.767304548901404, + "learning_rate": 1.2744389060205173e-05, + "loss": 0.6284, + "step": 3953 + }, + { + "epoch": 2.1465798045602607, + "grad_norm": 10.373642481495464, + "learning_rate": 1.2741007081815478e-05, + "loss": 0.6742, + "step": 3954 + }, + { + "epoch": 2.1471226927252984, + "grad_norm": 10.390187789918086, + "learning_rate": 1.2737624764413659e-05, + "loss": 0.6647, + "step": 3955 + }, + { + "epoch": 2.1476655808903367, + "grad_norm": 10.617840626931814, + "learning_rate": 1.2734242108418044e-05, + "loss": 0.877, + "step": 3956 + }, + { + "epoch": 2.1482084690553744, + "grad_norm": 9.898468473630299, + "learning_rate": 1.273085911424701e-05, + "loss": 0.7335, + "step": 3957 + }, + { + "epoch": 2.1487513572204127, + "grad_norm": 10.409230729626815, + "learning_rate": 1.2727475782318966e-05, + "loss": 0.7169, + "step": 3958 + }, + { + "epoch": 2.1492942453854504, + "grad_norm": 10.425915655294723, + "learning_rate": 1.2724092113052372e-05, + "loss": 0.6578, + "step": 3959 + }, + { + "epoch": 2.1498371335504887, + "grad_norm": 6.3375473265246995, + "learning_rate": 1.2720708106865722e-05, + "loss": 0.3471, + "step": 3960 + }, + { + "epoch": 2.1503800217155264, + "grad_norm": 9.396081485808933, + "learning_rate": 1.2717323764177559e-05, + "loss": 0.5232, + "step": 3961 + }, + { + "epoch": 2.1509229098805647, + "grad_norm": 13.419504298012779, + "learning_rate": 1.2713939085406461e-05, + "loss": 1.8096, + "step": 3962 + }, + { + "epoch": 2.1514657980456025, + "grad_norm": 9.635837991132963, + "learning_rate": 1.2710554070971053e-05, + "loss": 0.5949, + "step": 3963 + }, + { + "epoch": 2.1520086862106407, + "grad_norm": 10.111061783908927, + "learning_rate": 1.2707168721289995e-05, + "loss": 0.8731, + "step": 3964 + }, + { + "epoch": 2.1525515743756785, + "grad_norm": 8.256139809775819, + "learning_rate": 1.2703783036781998e-05, + "loss": 0.726, + "step": 3965 + }, + { + "epoch": 2.1530944625407167, + "grad_norm": 8.385511442711055, + "learning_rate": 1.2700397017865802e-05, + "loss": 0.6209, + "step": 3966 + }, + { + "epoch": 2.1536373507057545, + "grad_norm": 8.556891915186654, + "learning_rate": 1.2697010664960201e-05, + "loss": 0.6746, + "step": 3967 + }, + { + "epoch": 2.1541802388707927, + "grad_norm": 8.219611855039698, + "learning_rate": 1.2693623978484022e-05, + "loss": 0.7542, + "step": 3968 + }, + { + "epoch": 2.1547231270358305, + "grad_norm": 9.425031325511556, + "learning_rate": 1.2690236958856139e-05, + "loss": 0.6735, + "step": 3969 + }, + { + "epoch": 2.1552660152008687, + "grad_norm": 9.21346190912602, + "learning_rate": 1.2686849606495461e-05, + "loss": 0.5311, + "step": 3970 + }, + { + "epoch": 2.1558089033659065, + "grad_norm": 12.261438418537729, + "learning_rate": 1.268346192182094e-05, + "loss": 0.8556, + "step": 3971 + }, + { + "epoch": 2.1563517915309447, + "grad_norm": 8.630907790930149, + "learning_rate": 1.268007390525157e-05, + "loss": 0.7789, + "step": 3972 + }, + { + "epoch": 2.1568946796959825, + "grad_norm": 10.654367663061162, + "learning_rate": 1.267668555720639e-05, + "loss": 1.1708, + "step": 3973 + }, + { + "epoch": 2.1574375678610207, + "grad_norm": 15.113883332023352, + "learning_rate": 1.2673296878104472e-05, + "loss": 1.2273, + "step": 3974 + }, + { + "epoch": 2.1579804560260585, + "grad_norm": 10.205014286438582, + "learning_rate": 1.266990786836494e-05, + "loss": 1.2711, + "step": 3975 + }, + { + "epoch": 2.1585233441910967, + "grad_norm": 8.32550465973343, + "learning_rate": 1.2666518528406944e-05, + "loss": 0.6387, + "step": 3976 + }, + { + "epoch": 2.1590662323561345, + "grad_norm": 8.80250087611585, + "learning_rate": 1.2663128858649686e-05, + "loss": 0.5669, + "step": 3977 + }, + { + "epoch": 2.1596091205211727, + "grad_norm": 7.977255679001254, + "learning_rate": 1.2659738859512411e-05, + "loss": 0.4443, + "step": 3978 + }, + { + "epoch": 2.1601520086862105, + "grad_norm": 6.086622373481402, + "learning_rate": 1.265634853141439e-05, + "loss": 0.3029, + "step": 3979 + }, + { + "epoch": 2.1606948968512487, + "grad_norm": 16.46377636281654, + "learning_rate": 1.2652957874774953e-05, + "loss": 1.09, + "step": 3980 + }, + { + "epoch": 2.1612377850162865, + "grad_norm": 10.805039141935024, + "learning_rate": 1.2649566890013456e-05, + "loss": 0.672, + "step": 3981 + }, + { + "epoch": 2.1617806731813247, + "grad_norm": 10.731407400272554, + "learning_rate": 1.2646175577549303e-05, + "loss": 0.6618, + "step": 3982 + }, + { + "epoch": 2.1623235613463625, + "grad_norm": 9.572301578766334, + "learning_rate": 1.2642783937801937e-05, + "loss": 0.6608, + "step": 3983 + }, + { + "epoch": 2.1628664495114007, + "grad_norm": 8.415534879772189, + "learning_rate": 1.2639391971190842e-05, + "loss": 0.5314, + "step": 3984 + }, + { + "epoch": 2.1634093376764385, + "grad_norm": 8.290688286613944, + "learning_rate": 1.263599967813554e-05, + "loss": 0.6736, + "step": 3985 + }, + { + "epoch": 2.1639522258414767, + "grad_norm": 8.92112434155926, + "learning_rate": 1.2632607059055597e-05, + "loss": 0.5572, + "step": 3986 + }, + { + "epoch": 2.1644951140065145, + "grad_norm": 9.82202342503724, + "learning_rate": 1.2629214114370618e-05, + "loss": 0.7838, + "step": 3987 + }, + { + "epoch": 2.1650380021715527, + "grad_norm": 7.20619661733841, + "learning_rate": 1.262582084450025e-05, + "loss": 0.5577, + "step": 3988 + }, + { + "epoch": 2.1655808903365905, + "grad_norm": 8.640482408220107, + "learning_rate": 1.2622427249864172e-05, + "loss": 0.819, + "step": 3989 + }, + { + "epoch": 2.1661237785016287, + "grad_norm": 9.245400712044432, + "learning_rate": 1.2619033330882114e-05, + "loss": 0.6009, + "step": 3990 + }, + { + "epoch": 2.1666666666666665, + "grad_norm": 8.891547209594252, + "learning_rate": 1.2615639087973835e-05, + "loss": 0.6085, + "step": 3991 + }, + { + "epoch": 2.1672095548317047, + "grad_norm": 12.196798951360252, + "learning_rate": 1.2612244521559148e-05, + "loss": 0.745, + "step": 3992 + }, + { + "epoch": 2.1677524429967425, + "grad_norm": 9.610928616001752, + "learning_rate": 1.2608849632057895e-05, + "loss": 0.7709, + "step": 3993 + }, + { + "epoch": 2.1682953311617807, + "grad_norm": 9.454740121591225, + "learning_rate": 1.2605454419889962e-05, + "loss": 0.6455, + "step": 3994 + }, + { + "epoch": 2.1688382193268185, + "grad_norm": 13.245987794421069, + "learning_rate": 1.2602058885475273e-05, + "loss": 0.7104, + "step": 3995 + }, + { + "epoch": 2.1693811074918568, + "grad_norm": 10.51068899063252, + "learning_rate": 1.2598663029233794e-05, + "loss": 0.5747, + "step": 3996 + }, + { + "epoch": 2.1699239956568945, + "grad_norm": 10.589584490390159, + "learning_rate": 1.2595266851585532e-05, + "loss": 0.7878, + "step": 3997 + }, + { + "epoch": 2.1704668838219328, + "grad_norm": 7.818930212111502, + "learning_rate": 1.2591870352950528e-05, + "loss": 0.5179, + "step": 3998 + }, + { + "epoch": 2.1710097719869705, + "grad_norm": 8.222376749362372, + "learning_rate": 1.2588473533748868e-05, + "loss": 0.4722, + "step": 3999 + }, + { + "epoch": 2.1715526601520088, + "grad_norm": 12.451645756942625, + "learning_rate": 1.2585076394400675e-05, + "loss": 0.7742, + "step": 4000 + }, + { + "epoch": 2.1720955483170465, + "grad_norm": 11.33368393512842, + "learning_rate": 1.2581678935326117e-05, + "loss": 0.3945, + "step": 4001 + }, + { + "epoch": 2.1726384364820848, + "grad_norm": 10.914311466768305, + "learning_rate": 1.2578281156945389e-05, + "loss": 0.6345, + "step": 4002 + }, + { + "epoch": 2.1731813246471225, + "grad_norm": 12.867661375339019, + "learning_rate": 1.2574883059678742e-05, + "loss": 1.1569, + "step": 4003 + }, + { + "epoch": 2.1737242128121608, + "grad_norm": 11.4420969393128, + "learning_rate": 1.2571484643946452e-05, + "loss": 0.4927, + "step": 4004 + }, + { + "epoch": 2.1742671009771986, + "grad_norm": 10.234468502706436, + "learning_rate": 1.2568085910168842e-05, + "loss": 0.463, + "step": 4005 + }, + { + "epoch": 2.1748099891422368, + "grad_norm": 7.649699643480241, + "learning_rate": 1.2564686858766275e-05, + "loss": 0.3981, + "step": 4006 + }, + { + "epoch": 2.1753528773072746, + "grad_norm": 9.503268181536553, + "learning_rate": 1.2561287490159151e-05, + "loss": 0.4162, + "step": 4007 + }, + { + "epoch": 2.175895765472313, + "grad_norm": 8.81726258724681, + "learning_rate": 1.2557887804767907e-05, + "loss": 0.92, + "step": 4008 + }, + { + "epoch": 2.1764386536373506, + "grad_norm": 9.625823227183314, + "learning_rate": 1.2554487803013027e-05, + "loss": 0.7527, + "step": 4009 + }, + { + "epoch": 2.176981541802389, + "grad_norm": 11.452435192265499, + "learning_rate": 1.2551087485315017e-05, + "loss": 0.8611, + "step": 4010 + }, + { + "epoch": 2.1775244299674266, + "grad_norm": 9.123700040373722, + "learning_rate": 1.2547686852094445e-05, + "loss": 0.475, + "step": 4011 + }, + { + "epoch": 2.178067318132465, + "grad_norm": 9.306092992521913, + "learning_rate": 1.2544285903771902e-05, + "loss": 0.6063, + "step": 4012 + }, + { + "epoch": 2.1786102062975026, + "grad_norm": 14.283804972037375, + "learning_rate": 1.2540884640768022e-05, + "loss": 0.9142, + "step": 4013 + }, + { + "epoch": 2.179153094462541, + "grad_norm": 14.299740055929474, + "learning_rate": 1.2537483063503483e-05, + "loss": 0.8124, + "step": 4014 + }, + { + "epoch": 2.1796959826275786, + "grad_norm": 9.637221226214642, + "learning_rate": 1.2534081172398993e-05, + "loss": 0.6278, + "step": 4015 + }, + { + "epoch": 2.180238870792617, + "grad_norm": 6.674062415604907, + "learning_rate": 1.2530678967875304e-05, + "loss": 0.5078, + "step": 4016 + }, + { + "epoch": 2.1807817589576546, + "grad_norm": 11.68250334366375, + "learning_rate": 1.252727645035321e-05, + "loss": 0.7393, + "step": 4017 + }, + { + "epoch": 2.181324647122693, + "grad_norm": 9.642668530719417, + "learning_rate": 1.2523873620253535e-05, + "loss": 0.6218, + "step": 4018 + }, + { + "epoch": 2.1818675352877306, + "grad_norm": 9.955103488106618, + "learning_rate": 1.2520470477997146e-05, + "loss": 0.899, + "step": 4019 + }, + { + "epoch": 2.182410423452769, + "grad_norm": 13.184397199493825, + "learning_rate": 1.2517067024004955e-05, + "loss": 0.8437, + "step": 4020 + }, + { + "epoch": 2.1829533116178066, + "grad_norm": 11.142111068408965, + "learning_rate": 1.2513663258697901e-05, + "loss": 0.6064, + "step": 4021 + }, + { + "epoch": 2.183496199782845, + "grad_norm": 9.336503366824324, + "learning_rate": 1.251025918249697e-05, + "loss": 0.6649, + "step": 4022 + }, + { + "epoch": 2.1840390879478826, + "grad_norm": 11.35476662379509, + "learning_rate": 1.250685479582318e-05, + "loss": 0.8573, + "step": 4023 + }, + { + "epoch": 2.184581976112921, + "grad_norm": 9.049806197407708, + "learning_rate": 1.2503450099097594e-05, + "loss": 0.3839, + "step": 4024 + }, + { + "epoch": 2.1851248642779586, + "grad_norm": 12.739046308083323, + "learning_rate": 1.250004509274131e-05, + "loss": 0.9344, + "step": 4025 + }, + { + "epoch": 2.185667752442997, + "grad_norm": 7.539200963039068, + "learning_rate": 1.2496639777175465e-05, + "loss": 0.3691, + "step": 4026 + }, + { + "epoch": 2.1862106406080346, + "grad_norm": 10.659038181144831, + "learning_rate": 1.2493234152821234e-05, + "loss": 0.9462, + "step": 4027 + }, + { + "epoch": 2.186753528773073, + "grad_norm": 8.047186696422612, + "learning_rate": 1.2489828220099831e-05, + "loss": 0.4974, + "step": 4028 + }, + { + "epoch": 2.1872964169381106, + "grad_norm": 11.048564474051437, + "learning_rate": 1.2486421979432503e-05, + "loss": 0.6221, + "step": 4029 + }, + { + "epoch": 2.187839305103149, + "grad_norm": 9.808559304060044, + "learning_rate": 1.2483015431240542e-05, + "loss": 0.5221, + "step": 4030 + }, + { + "epoch": 2.1883821932681866, + "grad_norm": 7.720481766274275, + "learning_rate": 1.2479608575945274e-05, + "loss": 0.5248, + "step": 4031 + }, + { + "epoch": 2.188925081433225, + "grad_norm": 7.791803772948659, + "learning_rate": 1.2476201413968068e-05, + "loss": 0.5313, + "step": 4032 + }, + { + "epoch": 2.1894679695982626, + "grad_norm": 8.280464830936529, + "learning_rate": 1.2472793945730323e-05, + "loss": 0.5155, + "step": 4033 + }, + { + "epoch": 2.190010857763301, + "grad_norm": 6.500045291722395, + "learning_rate": 1.2469386171653483e-05, + "loss": 0.6057, + "step": 4034 + }, + { + "epoch": 2.1905537459283386, + "grad_norm": 6.735518823264981, + "learning_rate": 1.2465978092159025e-05, + "loss": 0.5207, + "step": 4035 + }, + { + "epoch": 2.191096634093377, + "grad_norm": 12.977966385021237, + "learning_rate": 1.2462569707668468e-05, + "loss": 1.0472, + "step": 4036 + }, + { + "epoch": 2.1916395222584146, + "grad_norm": 12.805295393279883, + "learning_rate": 1.245916101860336e-05, + "loss": 0.7486, + "step": 4037 + }, + { + "epoch": 2.192182410423453, + "grad_norm": 11.04531935760714, + "learning_rate": 1.2455752025385304e-05, + "loss": 0.6107, + "step": 4038 + }, + { + "epoch": 2.1927252985884906, + "grad_norm": 12.12096270238078, + "learning_rate": 1.245234272843592e-05, + "loss": 0.6264, + "step": 4039 + }, + { + "epoch": 2.193268186753529, + "grad_norm": 14.202650747934204, + "learning_rate": 1.244893312817688e-05, + "loss": 0.7965, + "step": 4040 + }, + { + "epoch": 2.1938110749185666, + "grad_norm": 10.409453866661186, + "learning_rate": 1.2445523225029887e-05, + "loss": 0.6012, + "step": 4041 + }, + { + "epoch": 2.194353963083605, + "grad_norm": 8.68329891299571, + "learning_rate": 1.2442113019416683e-05, + "loss": 0.4393, + "step": 4042 + }, + { + "epoch": 2.1948968512486426, + "grad_norm": 11.77894618341781, + "learning_rate": 1.2438702511759049e-05, + "loss": 0.7838, + "step": 4043 + }, + { + "epoch": 2.195439739413681, + "grad_norm": 8.31374036003515, + "learning_rate": 1.2435291702478802e-05, + "loss": 0.6043, + "step": 4044 + }, + { + "epoch": 2.1959826275787186, + "grad_norm": 10.219603175661236, + "learning_rate": 1.2431880591997799e-05, + "loss": 0.9292, + "step": 4045 + }, + { + "epoch": 2.196525515743757, + "grad_norm": 10.662478372576002, + "learning_rate": 1.2428469180737923e-05, + "loss": 0.5951, + "step": 4046 + }, + { + "epoch": 2.1970684039087947, + "grad_norm": 15.414471746340716, + "learning_rate": 1.2425057469121113e-05, + "loss": 1.0091, + "step": 4047 + }, + { + "epoch": 2.197611292073833, + "grad_norm": 8.553095532374297, + "learning_rate": 1.242164545756933e-05, + "loss": 0.4494, + "step": 4048 + }, + { + "epoch": 2.1981541802388707, + "grad_norm": 10.520590758759942, + "learning_rate": 1.2418233146504575e-05, + "loss": 0.5277, + "step": 4049 + }, + { + "epoch": 2.198697068403909, + "grad_norm": 10.792986797196212, + "learning_rate": 1.241482053634889e-05, + "loss": 0.7036, + "step": 4050 + }, + { + "epoch": 2.1992399565689467, + "grad_norm": 10.425002068049416, + "learning_rate": 1.2411407627524354e-05, + "loss": 0.8283, + "step": 4051 + }, + { + "epoch": 2.199782844733985, + "grad_norm": 10.421667203553737, + "learning_rate": 1.240799442045308e-05, + "loss": 0.8689, + "step": 4052 + }, + { + "epoch": 2.2003257328990227, + "grad_norm": 9.866413466645723, + "learning_rate": 1.2404580915557217e-05, + "loss": 0.517, + "step": 4053 + }, + { + "epoch": 2.200868621064061, + "grad_norm": 9.806755744401848, + "learning_rate": 1.2401167113258954e-05, + "loss": 0.7176, + "step": 4054 + }, + { + "epoch": 2.2014115092290987, + "grad_norm": 12.103200009209074, + "learning_rate": 1.2397753013980516e-05, + "loss": 0.8563, + "step": 4055 + }, + { + "epoch": 2.201954397394137, + "grad_norm": 9.310089002310463, + "learning_rate": 1.2394338618144162e-05, + "loss": 0.5478, + "step": 4056 + }, + { + "epoch": 2.2024972855591747, + "grad_norm": 8.147499111803365, + "learning_rate": 1.2390923926172194e-05, + "loss": 0.4616, + "step": 4057 + }, + { + "epoch": 2.203040173724213, + "grad_norm": 10.363627124726912, + "learning_rate": 1.2387508938486945e-05, + "loss": 0.6039, + "step": 4058 + }, + { + "epoch": 2.2035830618892507, + "grad_norm": 11.09580557172493, + "learning_rate": 1.2384093655510785e-05, + "loss": 0.6772, + "step": 4059 + }, + { + "epoch": 2.204125950054289, + "grad_norm": 10.762880100606296, + "learning_rate": 1.238067807766612e-05, + "loss": 0.8104, + "step": 4060 + }, + { + "epoch": 2.2046688382193267, + "grad_norm": 10.590539714326667, + "learning_rate": 1.2377262205375398e-05, + "loss": 0.6766, + "step": 4061 + }, + { + "epoch": 2.205211726384365, + "grad_norm": 10.060689580140192, + "learning_rate": 1.2373846039061095e-05, + "loss": 1.0282, + "step": 4062 + }, + { + "epoch": 2.2057546145494027, + "grad_norm": 7.612672548635284, + "learning_rate": 1.237042957914573e-05, + "loss": 0.449, + "step": 4063 + }, + { + "epoch": 2.206297502714441, + "grad_norm": 12.239451929254052, + "learning_rate": 1.2367012826051861e-05, + "loss": 0.5385, + "step": 4064 + }, + { + "epoch": 2.2068403908794787, + "grad_norm": 14.199356139948698, + "learning_rate": 1.236359578020207e-05, + "loss": 0.7867, + "step": 4065 + }, + { + "epoch": 2.207383279044517, + "grad_norm": 8.357180617953833, + "learning_rate": 1.2360178442018989e-05, + "loss": 0.6803, + "step": 4066 + }, + { + "epoch": 2.2079261672095547, + "grad_norm": 9.203584379654275, + "learning_rate": 1.2356760811925277e-05, + "loss": 0.677, + "step": 4067 + }, + { + "epoch": 2.208469055374593, + "grad_norm": 8.063224350440471, + "learning_rate": 1.2353342890343626e-05, + "loss": 0.644, + "step": 4068 + }, + { + "epoch": 2.2090119435396307, + "grad_norm": 9.352085551954278, + "learning_rate": 1.2349924677696781e-05, + "loss": 0.9888, + "step": 4069 + }, + { + "epoch": 2.209554831704669, + "grad_norm": 10.54696840423902, + "learning_rate": 1.2346506174407505e-05, + "loss": 0.5895, + "step": 4070 + }, + { + "epoch": 2.2100977198697067, + "grad_norm": 8.568183164428444, + "learning_rate": 1.2343087380898604e-05, + "loss": 0.6287, + "step": 4071 + }, + { + "epoch": 2.210640608034745, + "grad_norm": 11.191021916852304, + "learning_rate": 1.2339668297592924e-05, + "loss": 0.8216, + "step": 4072 + }, + { + "epoch": 2.2111834961997827, + "grad_norm": 7.67795379285754, + "learning_rate": 1.233624892491334e-05, + "loss": 0.423, + "step": 4073 + }, + { + "epoch": 2.211726384364821, + "grad_norm": 8.108373133947762, + "learning_rate": 1.2332829263282764e-05, + "loss": 0.6548, + "step": 4074 + }, + { + "epoch": 2.2122692725298587, + "grad_norm": 9.0962841262343, + "learning_rate": 1.2329409313124146e-05, + "loss": 0.5506, + "step": 4075 + }, + { + "epoch": 2.212812160694897, + "grad_norm": 8.85949097192624, + "learning_rate": 1.2325989074860472e-05, + "loss": 0.5258, + "step": 4076 + }, + { + "epoch": 2.2133550488599347, + "grad_norm": 7.051914022367066, + "learning_rate": 1.2322568548914764e-05, + "loss": 0.5242, + "step": 4077 + }, + { + "epoch": 2.213897937024973, + "grad_norm": 8.287257693111446, + "learning_rate": 1.2319147735710076e-05, + "loss": 0.4769, + "step": 4078 + }, + { + "epoch": 2.2144408251900107, + "grad_norm": 9.473240892632637, + "learning_rate": 1.2315726635669498e-05, + "loss": 1.0853, + "step": 4079 + }, + { + "epoch": 2.214983713355049, + "grad_norm": 10.460874830183508, + "learning_rate": 1.231230524921616e-05, + "loss": 0.8907, + "step": 4080 + }, + { + "epoch": 2.2155266015200867, + "grad_norm": 9.17618869023913, + "learning_rate": 1.2308883576773223e-05, + "loss": 0.6355, + "step": 4081 + }, + { + "epoch": 2.216069489685125, + "grad_norm": 10.389188046047401, + "learning_rate": 1.2305461618763886e-05, + "loss": 0.9455, + "step": 4082 + }, + { + "epoch": 2.2166123778501627, + "grad_norm": 8.41220072964392, + "learning_rate": 1.230203937561138e-05, + "loss": 0.5574, + "step": 4083 + }, + { + "epoch": 2.217155266015201, + "grad_norm": 8.193062738430084, + "learning_rate": 1.2298616847738978e-05, + "loss": 0.548, + "step": 4084 + }, + { + "epoch": 2.2176981541802387, + "grad_norm": 8.55082220874788, + "learning_rate": 1.2295194035569979e-05, + "loss": 0.6087, + "step": 4085 + }, + { + "epoch": 2.218241042345277, + "grad_norm": 10.482752217531827, + "learning_rate": 1.2291770939527725e-05, + "loss": 0.5404, + "step": 4086 + }, + { + "epoch": 2.2187839305103148, + "grad_norm": 9.786686689665721, + "learning_rate": 1.2288347560035587e-05, + "loss": 0.6795, + "step": 4087 + }, + { + "epoch": 2.219326818675353, + "grad_norm": 9.696178175990593, + "learning_rate": 1.2284923897516978e-05, + "loss": 1.2474, + "step": 4088 + }, + { + "epoch": 2.2198697068403908, + "grad_norm": 9.135935376096977, + "learning_rate": 1.2281499952395336e-05, + "loss": 0.6746, + "step": 4089 + }, + { + "epoch": 2.220412595005429, + "grad_norm": 10.341327194770198, + "learning_rate": 1.2278075725094147e-05, + "loss": 0.6501, + "step": 4090 + }, + { + "epoch": 2.2209554831704668, + "grad_norm": 11.750926749777816, + "learning_rate": 1.2274651216036921e-05, + "loss": 0.6421, + "step": 4091 + }, + { + "epoch": 2.221498371335505, + "grad_norm": 15.264249369051182, + "learning_rate": 1.2271226425647207e-05, + "loss": 0.9053, + "step": 4092 + }, + { + "epoch": 2.2220412595005428, + "grad_norm": 10.996946312108877, + "learning_rate": 1.2267801354348589e-05, + "loss": 0.6776, + "step": 4093 + }, + { + "epoch": 2.222584147665581, + "grad_norm": 9.01852590664764, + "learning_rate": 1.2264376002564687e-05, + "loss": 0.7342, + "step": 4094 + }, + { + "epoch": 2.2231270358306188, + "grad_norm": 9.776717004980412, + "learning_rate": 1.2260950370719146e-05, + "loss": 0.7353, + "step": 4095 + }, + { + "epoch": 2.223669923995657, + "grad_norm": 12.928031483014212, + "learning_rate": 1.2257524459235666e-05, + "loss": 0.8649, + "step": 4096 + }, + { + "epoch": 2.2242128121606948, + "grad_norm": 9.549545586005916, + "learning_rate": 1.2254098268537962e-05, + "loss": 0.8056, + "step": 4097 + }, + { + "epoch": 2.224755700325733, + "grad_norm": 10.729492841694832, + "learning_rate": 1.2250671799049791e-05, + "loss": 0.5817, + "step": 4098 + }, + { + "epoch": 2.225298588490771, + "grad_norm": 14.13032585380651, + "learning_rate": 1.2247245051194944e-05, + "loss": 0.5998, + "step": 4099 + }, + { + "epoch": 2.225841476655809, + "grad_norm": 9.143101156244217, + "learning_rate": 1.2243818025397247e-05, + "loss": 0.5633, + "step": 4100 + }, + { + "epoch": 2.226384364820847, + "grad_norm": 9.010721857216959, + "learning_rate": 1.224039072208056e-05, + "loss": 0.5855, + "step": 4101 + }, + { + "epoch": 2.226927252985885, + "grad_norm": 11.257457693848512, + "learning_rate": 1.2236963141668778e-05, + "loss": 0.5672, + "step": 4102 + }, + { + "epoch": 2.227470141150923, + "grad_norm": 9.238231248800124, + "learning_rate": 1.2233535284585831e-05, + "loss": 0.4633, + "step": 4103 + }, + { + "epoch": 2.228013029315961, + "grad_norm": 8.93957445803232, + "learning_rate": 1.223010715125568e-05, + "loss": 0.8019, + "step": 4104 + }, + { + "epoch": 2.228555917480999, + "grad_norm": 10.869791394117428, + "learning_rate": 1.2226678742102322e-05, + "loss": 0.4235, + "step": 4105 + }, + { + "epoch": 2.229098805646037, + "grad_norm": 11.71960282550904, + "learning_rate": 1.2223250057549789e-05, + "loss": 0.8825, + "step": 4106 + }, + { + "epoch": 2.229641693811075, + "grad_norm": 9.133591574750724, + "learning_rate": 1.2219821098022146e-05, + "loss": 0.456, + "step": 4107 + }, + { + "epoch": 2.230184581976113, + "grad_norm": 7.713253727965175, + "learning_rate": 1.2216391863943492e-05, + "loss": 0.407, + "step": 4108 + }, + { + "epoch": 2.230727470141151, + "grad_norm": 8.351720726424757, + "learning_rate": 1.2212962355737959e-05, + "loss": 0.6003, + "step": 4109 + }, + { + "epoch": 2.231270358306189, + "grad_norm": 12.012727761361187, + "learning_rate": 1.2209532573829714e-05, + "loss": 0.731, + "step": 4110 + }, + { + "epoch": 2.231813246471227, + "grad_norm": 12.902339895008394, + "learning_rate": 1.220610251864296e-05, + "loss": 0.9873, + "step": 4111 + }, + { + "epoch": 2.232356134636265, + "grad_norm": 8.935941303438996, + "learning_rate": 1.220267219060193e-05, + "loss": 0.7549, + "step": 4112 + }, + { + "epoch": 2.232899022801303, + "grad_norm": 9.69521468611751, + "learning_rate": 1.2199241590130892e-05, + "loss": 0.6099, + "step": 4113 + }, + { + "epoch": 2.233441910966341, + "grad_norm": 7.437030925496973, + "learning_rate": 1.2195810717654148e-05, + "loss": 0.4743, + "step": 4114 + }, + { + "epoch": 2.233984799131379, + "grad_norm": 7.435823614079823, + "learning_rate": 1.2192379573596036e-05, + "loss": 0.5019, + "step": 4115 + }, + { + "epoch": 2.234527687296417, + "grad_norm": 7.563032718583443, + "learning_rate": 1.2188948158380927e-05, + "loss": 0.7176, + "step": 4116 + }, + { + "epoch": 2.235070575461455, + "grad_norm": 10.449779515276235, + "learning_rate": 1.2185516472433218e-05, + "loss": 0.7921, + "step": 4117 + }, + { + "epoch": 2.235613463626493, + "grad_norm": 9.088046167756731, + "learning_rate": 1.2182084516177351e-05, + "loss": 0.811, + "step": 4118 + }, + { + "epoch": 2.236156351791531, + "grad_norm": 9.664269830378425, + "learning_rate": 1.217865229003779e-05, + "loss": 0.8024, + "step": 4119 + }, + { + "epoch": 2.236699239956569, + "grad_norm": 8.055354430405984, + "learning_rate": 1.2175219794439044e-05, + "loss": 0.5747, + "step": 4120 + }, + { + "epoch": 2.237242128121607, + "grad_norm": 6.428342674450092, + "learning_rate": 1.2171787029805646e-05, + "loss": 0.5673, + "step": 4121 + }, + { + "epoch": 2.237785016286645, + "grad_norm": 7.502074776894542, + "learning_rate": 1.2168353996562167e-05, + "loss": 0.566, + "step": 4122 + }, + { + "epoch": 2.238327904451683, + "grad_norm": 10.380110541434405, + "learning_rate": 1.2164920695133212e-05, + "loss": 1.0609, + "step": 4123 + }, + { + "epoch": 2.238870792616721, + "grad_norm": 13.199317832444493, + "learning_rate": 1.2161487125943413e-05, + "loss": 0.7662, + "step": 4124 + }, + { + "epoch": 2.239413680781759, + "grad_norm": 9.067564902950002, + "learning_rate": 1.2158053289417439e-05, + "loss": 0.6325, + "step": 4125 + }, + { + "epoch": 2.239956568946797, + "grad_norm": 8.074724230493178, + "learning_rate": 1.215461918598e-05, + "loss": 0.6854, + "step": 4126 + }, + { + "epoch": 2.240499457111835, + "grad_norm": 7.916509466850638, + "learning_rate": 1.2151184816055822e-05, + "loss": 0.3852, + "step": 4127 + }, + { + "epoch": 2.241042345276873, + "grad_norm": 8.647089691838135, + "learning_rate": 1.214775018006968e-05, + "loss": 0.6291, + "step": 4128 + }, + { + "epoch": 2.241585233441911, + "grad_norm": 9.238872667328287, + "learning_rate": 1.214431527844637e-05, + "loss": 0.6117, + "step": 4129 + }, + { + "epoch": 2.242128121606949, + "grad_norm": 12.537117099920934, + "learning_rate": 1.214088011161073e-05, + "loss": 0.745, + "step": 4130 + }, + { + "epoch": 2.242671009771987, + "grad_norm": 9.99529796933279, + "learning_rate": 1.2137444679987627e-05, + "loss": 0.612, + "step": 4131 + }, + { + "epoch": 2.243213897937025, + "grad_norm": 6.673816908475189, + "learning_rate": 1.2134008984001959e-05, + "loss": 0.3219, + "step": 4132 + }, + { + "epoch": 2.243756786102063, + "grad_norm": 13.936566760747976, + "learning_rate": 1.2130573024078656e-05, + "loss": 0.8689, + "step": 4133 + }, + { + "epoch": 2.244299674267101, + "grad_norm": 7.800578548383095, + "learning_rate": 1.2127136800642689e-05, + "loss": 0.3804, + "step": 4134 + }, + { + "epoch": 2.244842562432139, + "grad_norm": 9.283785411998215, + "learning_rate": 1.2123700314119052e-05, + "loss": 0.699, + "step": 4135 + }, + { + "epoch": 2.245385450597177, + "grad_norm": 8.613596910195628, + "learning_rate": 1.2120263564932775e-05, + "loss": 0.7155, + "step": 4136 + }, + { + "epoch": 2.245928338762215, + "grad_norm": 8.522662049603209, + "learning_rate": 1.2116826553508923e-05, + "loss": 0.5705, + "step": 4137 + }, + { + "epoch": 2.246471226927253, + "grad_norm": 9.023598184433123, + "learning_rate": 1.2113389280272587e-05, + "loss": 0.4392, + "step": 4138 + }, + { + "epoch": 2.247014115092291, + "grad_norm": 10.959276458237868, + "learning_rate": 1.2109951745648894e-05, + "loss": 0.7827, + "step": 4139 + }, + { + "epoch": 2.247557003257329, + "grad_norm": 11.660193747264707, + "learning_rate": 1.210651395006301e-05, + "loss": 0.9824, + "step": 4140 + }, + { + "epoch": 2.248099891422367, + "grad_norm": 18.121219443467986, + "learning_rate": 1.2103075893940122e-05, + "loss": 1.2339, + "step": 4141 + }, + { + "epoch": 2.248642779587405, + "grad_norm": 15.059306090767514, + "learning_rate": 1.2099637577705457e-05, + "loss": 0.7434, + "step": 4142 + }, + { + "epoch": 2.249185667752443, + "grad_norm": 10.999940106944315, + "learning_rate": 1.2096199001784268e-05, + "loss": 0.6422, + "step": 4143 + }, + { + "epoch": 2.249728555917481, + "grad_norm": 6.704480729179984, + "learning_rate": 1.2092760166601848e-05, + "loss": 0.4829, + "step": 4144 + }, + { + "epoch": 2.250271444082519, + "grad_norm": 11.849496386091012, + "learning_rate": 1.2089321072583512e-05, + "loss": 0.8969, + "step": 4145 + }, + { + "epoch": 2.250814332247557, + "grad_norm": 8.596596068581539, + "learning_rate": 1.2085881720154617e-05, + "loss": 0.6148, + "step": 4146 + }, + { + "epoch": 2.251357220412595, + "grad_norm": 7.50181808267645, + "learning_rate": 1.2082442109740548e-05, + "loss": 0.5499, + "step": 4147 + }, + { + "epoch": 2.251900108577633, + "grad_norm": 10.178327453067736, + "learning_rate": 1.207900224176672e-05, + "loss": 0.5582, + "step": 4148 + }, + { + "epoch": 2.252442996742671, + "grad_norm": 10.74128468394649, + "learning_rate": 1.2075562116658581e-05, + "loss": 0.8023, + "step": 4149 + }, + { + "epoch": 2.252985884907709, + "grad_norm": 10.195557872174708, + "learning_rate": 1.207212173484161e-05, + "loss": 0.6092, + "step": 4150 + }, + { + "epoch": 2.253528773072747, + "grad_norm": 9.843275181764184, + "learning_rate": 1.2068681096741322e-05, + "loss": 0.8451, + "step": 4151 + }, + { + "epoch": 2.254071661237785, + "grad_norm": 7.859122806337064, + "learning_rate": 1.2065240202783259e-05, + "loss": 0.6158, + "step": 4152 + }, + { + "epoch": 2.254614549402823, + "grad_norm": 11.4233860908698, + "learning_rate": 1.2061799053392994e-05, + "loss": 0.6552, + "step": 4153 + }, + { + "epoch": 2.255157437567861, + "grad_norm": 7.339673441521232, + "learning_rate": 1.2058357648996138e-05, + "loss": 0.5103, + "step": 4154 + }, + { + "epoch": 2.255700325732899, + "grad_norm": 11.603888679814519, + "learning_rate": 1.2054915990018326e-05, + "loss": 0.8117, + "step": 4155 + }, + { + "epoch": 2.256243213897937, + "grad_norm": 9.757430713637703, + "learning_rate": 1.2051474076885232e-05, + "loss": 0.6942, + "step": 4156 + }, + { + "epoch": 2.256786102062975, + "grad_norm": 10.286384353132442, + "learning_rate": 1.2048031910022553e-05, + "loss": 0.8958, + "step": 4157 + }, + { + "epoch": 2.257328990228013, + "grad_norm": 7.6945320206947185, + "learning_rate": 1.2044589489856023e-05, + "loss": 0.4508, + "step": 4158 + }, + { + "epoch": 2.257871878393051, + "grad_norm": 9.546205961324182, + "learning_rate": 1.2041146816811406e-05, + "loss": 0.6889, + "step": 4159 + }, + { + "epoch": 2.258414766558089, + "grad_norm": 6.440627468414342, + "learning_rate": 1.2037703891314498e-05, + "loss": 0.5146, + "step": 4160 + }, + { + "epoch": 2.258957654723127, + "grad_norm": 8.769926030437105, + "learning_rate": 1.2034260713791124e-05, + "loss": 0.539, + "step": 4161 + }, + { + "epoch": 2.259500542888165, + "grad_norm": 8.385311572142552, + "learning_rate": 1.2030817284667145e-05, + "loss": 0.6471, + "step": 4162 + }, + { + "epoch": 2.260043431053203, + "grad_norm": 10.687000132395355, + "learning_rate": 1.2027373604368446e-05, + "loss": 0.4047, + "step": 4163 + }, + { + "epoch": 2.260586319218241, + "grad_norm": 9.22667613364525, + "learning_rate": 1.2023929673320952e-05, + "loss": 0.6039, + "step": 4164 + }, + { + "epoch": 2.261129207383279, + "grad_norm": 8.822764152626888, + "learning_rate": 1.2020485491950607e-05, + "loss": 0.5333, + "step": 4165 + }, + { + "epoch": 2.261672095548317, + "grad_norm": 8.505680622817092, + "learning_rate": 1.2017041060683394e-05, + "loss": 0.6679, + "step": 4166 + }, + { + "epoch": 2.262214983713355, + "grad_norm": 12.980914094830798, + "learning_rate": 1.2013596379945331e-05, + "loss": 0.7265, + "step": 4167 + }, + { + "epoch": 2.262757871878393, + "grad_norm": 12.278007785084847, + "learning_rate": 1.201015145016246e-05, + "loss": 0.7175, + "step": 4168 + }, + { + "epoch": 2.263300760043431, + "grad_norm": 8.871985541095814, + "learning_rate": 1.2006706271760851e-05, + "loss": 0.4937, + "step": 4169 + }, + { + "epoch": 2.263843648208469, + "grad_norm": 10.070308144714824, + "learning_rate": 1.2003260845166613e-05, + "loss": 0.5994, + "step": 4170 + }, + { + "epoch": 2.264386536373507, + "grad_norm": 8.846376302708403, + "learning_rate": 1.1999815170805882e-05, + "loss": 0.5795, + "step": 4171 + }, + { + "epoch": 2.264929424538545, + "grad_norm": 11.984661783852376, + "learning_rate": 1.199636924910482e-05, + "loss": 1.1718, + "step": 4172 + }, + { + "epoch": 2.265472312703583, + "grad_norm": 10.035244756765978, + "learning_rate": 1.1992923080489629e-05, + "loss": 0.8101, + "step": 4173 + }, + { + "epoch": 2.266015200868621, + "grad_norm": 7.429999917349929, + "learning_rate": 1.1989476665386537e-05, + "loss": 0.4121, + "step": 4174 + }, + { + "epoch": 2.266558089033659, + "grad_norm": 8.464944728590373, + "learning_rate": 1.1986030004221802e-05, + "loss": 0.5553, + "step": 4175 + }, + { + "epoch": 2.267100977198697, + "grad_norm": 11.414355640267063, + "learning_rate": 1.1982583097421707e-05, + "loss": 0.8705, + "step": 4176 + }, + { + "epoch": 2.267643865363735, + "grad_norm": 10.094383752442985, + "learning_rate": 1.1979135945412576e-05, + "loss": 0.5512, + "step": 4177 + }, + { + "epoch": 2.268186753528773, + "grad_norm": 10.131657295101125, + "learning_rate": 1.1975688548620756e-05, + "loss": 0.5678, + "step": 4178 + }, + { + "epoch": 2.268729641693811, + "grad_norm": 11.675191701181866, + "learning_rate": 1.197224090747263e-05, + "loss": 0.5967, + "step": 4179 + }, + { + "epoch": 2.269272529858849, + "grad_norm": 9.879570041096912, + "learning_rate": 1.1968793022394603e-05, + "loss": 1.0245, + "step": 4180 + }, + { + "epoch": 2.269815418023887, + "grad_norm": 12.255119190489571, + "learning_rate": 1.1965344893813117e-05, + "loss": 0.9963, + "step": 4181 + }, + { + "epoch": 2.270358306188925, + "grad_norm": 8.870271458097042, + "learning_rate": 1.1961896522154643e-05, + "loss": 0.3658, + "step": 4182 + }, + { + "epoch": 2.270901194353963, + "grad_norm": 9.50086592138889, + "learning_rate": 1.1958447907845684e-05, + "loss": 0.588, + "step": 4183 + }, + { + "epoch": 2.271444082519001, + "grad_norm": 8.92062913157286, + "learning_rate": 1.1954999051312762e-05, + "loss": 0.6038, + "step": 4184 + }, + { + "epoch": 2.271986970684039, + "grad_norm": 11.511359098427171, + "learning_rate": 1.1951549952982442e-05, + "loss": 0.7837, + "step": 4185 + }, + { + "epoch": 2.272529858849077, + "grad_norm": 7.7230189119336075, + "learning_rate": 1.1948100613281315e-05, + "loss": 0.5235, + "step": 4186 + }, + { + "epoch": 2.273072747014115, + "grad_norm": 8.85303744288786, + "learning_rate": 1.1944651032635997e-05, + "loss": 0.5194, + "step": 4187 + }, + { + "epoch": 2.273615635179153, + "grad_norm": 9.931552012214697, + "learning_rate": 1.1941201211473142e-05, + "loss": 0.6419, + "step": 4188 + }, + { + "epoch": 2.274158523344191, + "grad_norm": 11.635637385284054, + "learning_rate": 1.1937751150219428e-05, + "loss": 0.7948, + "step": 4189 + }, + { + "epoch": 2.274701411509229, + "grad_norm": 7.783494952264655, + "learning_rate": 1.1934300849301561e-05, + "loss": 0.568, + "step": 4190 + }, + { + "epoch": 2.275244299674267, + "grad_norm": 7.979443655042872, + "learning_rate": 1.1930850309146283e-05, + "loss": 0.531, + "step": 4191 + }, + { + "epoch": 2.2757871878393052, + "grad_norm": 12.081321074807914, + "learning_rate": 1.1927399530180359e-05, + "loss": 0.9917, + "step": 4192 + }, + { + "epoch": 2.276330076004343, + "grad_norm": 9.376910406413142, + "learning_rate": 1.1923948512830595e-05, + "loss": 0.6587, + "step": 4193 + }, + { + "epoch": 2.2768729641693812, + "grad_norm": 9.576589478578342, + "learning_rate": 1.1920497257523809e-05, + "loss": 0.5703, + "step": 4194 + }, + { + "epoch": 2.277415852334419, + "grad_norm": 9.129784891039465, + "learning_rate": 1.1917045764686863e-05, + "loss": 0.668, + "step": 4195 + }, + { + "epoch": 2.2779587404994572, + "grad_norm": 11.723388678968943, + "learning_rate": 1.191359403474664e-05, + "loss": 0.6293, + "step": 4196 + }, + { + "epoch": 2.278501628664495, + "grad_norm": 10.662032565840782, + "learning_rate": 1.1910142068130058e-05, + "loss": 1.0251, + "step": 4197 + }, + { + "epoch": 2.2790445168295332, + "grad_norm": 5.976705265828664, + "learning_rate": 1.1906689865264058e-05, + "loss": 0.3892, + "step": 4198 + }, + { + "epoch": 2.279587404994571, + "grad_norm": 9.77428172632025, + "learning_rate": 1.190323742657562e-05, + "loss": 0.8023, + "step": 4199 + }, + { + "epoch": 2.2801302931596092, + "grad_norm": 9.701826347676784, + "learning_rate": 1.1899784752491743e-05, + "loss": 0.6668, + "step": 4200 + }, + { + "epoch": 2.280673181324647, + "grad_norm": 10.939483572398276, + "learning_rate": 1.1896331843439459e-05, + "loss": 0.8751, + "step": 4201 + }, + { + "epoch": 2.2812160694896852, + "grad_norm": 12.656192710489524, + "learning_rate": 1.1892878699845832e-05, + "loss": 0.7874, + "step": 4202 + }, + { + "epoch": 2.281758957654723, + "grad_norm": 8.821175958313784, + "learning_rate": 1.1889425322137948e-05, + "loss": 0.7599, + "step": 4203 + }, + { + "epoch": 2.2823018458197613, + "grad_norm": 11.535196460357659, + "learning_rate": 1.188597171074293e-05, + "loss": 0.7448, + "step": 4204 + }, + { + "epoch": 2.282844733984799, + "grad_norm": 9.221866003570257, + "learning_rate": 1.1882517866087926e-05, + "loss": 0.7176, + "step": 4205 + }, + { + "epoch": 2.2833876221498373, + "grad_norm": 10.791249990779047, + "learning_rate": 1.1879063788600113e-05, + "loss": 0.6905, + "step": 4206 + }, + { + "epoch": 2.283930510314875, + "grad_norm": 11.953403018554845, + "learning_rate": 1.1875609478706694e-05, + "loss": 0.7006, + "step": 4207 + }, + { + "epoch": 2.2844733984799133, + "grad_norm": 10.830659589867402, + "learning_rate": 1.1872154936834905e-05, + "loss": 0.6872, + "step": 4208 + }, + { + "epoch": 2.285016286644951, + "grad_norm": 10.8147041656128, + "learning_rate": 1.186870016341201e-05, + "loss": 0.6429, + "step": 4209 + }, + { + "epoch": 2.2855591748099893, + "grad_norm": 10.70508214004579, + "learning_rate": 1.18652451588653e-05, + "loss": 0.9063, + "step": 4210 + }, + { + "epoch": 2.286102062975027, + "grad_norm": 8.333893553672915, + "learning_rate": 1.1861789923622099e-05, + "loss": 0.5245, + "step": 4211 + }, + { + "epoch": 2.2866449511400653, + "grad_norm": 9.62415991156623, + "learning_rate": 1.1858334458109747e-05, + "loss": 0.734, + "step": 4212 + }, + { + "epoch": 2.287187839305103, + "grad_norm": 15.893278058161348, + "learning_rate": 1.1854878762755633e-05, + "loss": 0.898, + "step": 4213 + }, + { + "epoch": 2.2877307274701413, + "grad_norm": 8.12705480823925, + "learning_rate": 1.1851422837987159e-05, + "loss": 0.5586, + "step": 4214 + }, + { + "epoch": 2.288273615635179, + "grad_norm": 19.440766463719637, + "learning_rate": 1.1847966684231754e-05, + "loss": 1.0309, + "step": 4215 + }, + { + "epoch": 2.2888165038002173, + "grad_norm": 10.590183829770972, + "learning_rate": 1.1844510301916883e-05, + "loss": 0.8545, + "step": 4216 + }, + { + "epoch": 2.289359391965255, + "grad_norm": 9.825370158957117, + "learning_rate": 1.184105369147004e-05, + "loss": 0.7426, + "step": 4217 + }, + { + "epoch": 2.2899022801302933, + "grad_norm": 10.341219416998992, + "learning_rate": 1.1837596853318742e-05, + "loss": 0.8584, + "step": 4218 + }, + { + "epoch": 2.290445168295331, + "grad_norm": 11.121322184848584, + "learning_rate": 1.1834139787890538e-05, + "loss": 0.6598, + "step": 4219 + }, + { + "epoch": 2.2909880564603693, + "grad_norm": 9.600608599739235, + "learning_rate": 1.1830682495613e-05, + "loss": 0.5885, + "step": 4220 + }, + { + "epoch": 2.291530944625407, + "grad_norm": 10.517374576113433, + "learning_rate": 1.1827224976913732e-05, + "loss": 0.751, + "step": 4221 + }, + { + "epoch": 2.2920738327904453, + "grad_norm": 9.666522143402908, + "learning_rate": 1.1823767232220369e-05, + "loss": 0.7602, + "step": 4222 + }, + { + "epoch": 2.292616720955483, + "grad_norm": 9.390811834354404, + "learning_rate": 1.1820309261960567e-05, + "loss": 0.6187, + "step": 4223 + }, + { + "epoch": 2.2931596091205213, + "grad_norm": 8.772653626817414, + "learning_rate": 1.1816851066562017e-05, + "loss": 0.5221, + "step": 4224 + }, + { + "epoch": 2.293702497285559, + "grad_norm": 11.957278593216312, + "learning_rate": 1.1813392646452429e-05, + "loss": 0.9231, + "step": 4225 + }, + { + "epoch": 2.2942453854505973, + "grad_norm": 9.58191568158383, + "learning_rate": 1.1809934002059547e-05, + "loss": 0.6747, + "step": 4226 + }, + { + "epoch": 2.294788273615635, + "grad_norm": 8.242154702381846, + "learning_rate": 1.1806475133811146e-05, + "loss": 0.4053, + "step": 4227 + }, + { + "epoch": 2.2953311617806733, + "grad_norm": 8.83768826569311, + "learning_rate": 1.1803016042135021e-05, + "loss": 0.4177, + "step": 4228 + }, + { + "epoch": 2.295874049945711, + "grad_norm": 7.496802017426867, + "learning_rate": 1.1799556727458996e-05, + "loss": 0.6402, + "step": 4229 + }, + { + "epoch": 2.2964169381107493, + "grad_norm": 10.15598151517949, + "learning_rate": 1.1796097190210927e-05, + "loss": 0.6301, + "step": 4230 + }, + { + "epoch": 2.296959826275787, + "grad_norm": 9.100517578174307, + "learning_rate": 1.1792637430818695e-05, + "loss": 0.6059, + "step": 4231 + }, + { + "epoch": 2.2975027144408253, + "grad_norm": 9.341766570492698, + "learning_rate": 1.178917744971021e-05, + "loss": 0.7462, + "step": 4232 + }, + { + "epoch": 2.298045602605863, + "grad_norm": 9.790098160117006, + "learning_rate": 1.1785717247313407e-05, + "loss": 0.7422, + "step": 4233 + }, + { + "epoch": 2.2985884907709013, + "grad_norm": 11.785065190423747, + "learning_rate": 1.1782256824056249e-05, + "loss": 0.8544, + "step": 4234 + }, + { + "epoch": 2.299131378935939, + "grad_norm": 11.165339576639699, + "learning_rate": 1.1778796180366722e-05, + "loss": 0.6492, + "step": 4235 + }, + { + "epoch": 2.2996742671009773, + "grad_norm": 9.026810304485423, + "learning_rate": 1.1775335316672854e-05, + "loss": 0.751, + "step": 4236 + }, + { + "epoch": 2.300217155266015, + "grad_norm": 10.085898624762818, + "learning_rate": 1.1771874233402682e-05, + "loss": 0.6052, + "step": 4237 + }, + { + "epoch": 2.3007600434310533, + "grad_norm": 9.302471978496024, + "learning_rate": 1.176841293098428e-05, + "loss": 0.7269, + "step": 4238 + }, + { + "epoch": 2.301302931596091, + "grad_norm": 11.406933536282311, + "learning_rate": 1.176495140984575e-05, + "loss": 0.7287, + "step": 4239 + }, + { + "epoch": 2.3018458197611293, + "grad_norm": 8.172755818720033, + "learning_rate": 1.1761489670415218e-05, + "loss": 0.4692, + "step": 4240 + }, + { + "epoch": 2.302388707926167, + "grad_norm": 11.729275499093248, + "learning_rate": 1.1758027713120833e-05, + "loss": 0.8301, + "step": 4241 + }, + { + "epoch": 2.3029315960912053, + "grad_norm": 7.31457707003035, + "learning_rate": 1.1754565538390782e-05, + "loss": 0.5372, + "step": 4242 + }, + { + "epoch": 2.303474484256243, + "grad_norm": 10.15517896282683, + "learning_rate": 1.1751103146653265e-05, + "loss": 0.8744, + "step": 4243 + }, + { + "epoch": 2.3040173724212814, + "grad_norm": 7.780368081203029, + "learning_rate": 1.1747640538336525e-05, + "loss": 0.4743, + "step": 4244 + }, + { + "epoch": 2.304560260586319, + "grad_norm": 7.814816774582812, + "learning_rate": 1.1744177713868814e-05, + "loss": 0.628, + "step": 4245 + }, + { + "epoch": 2.3051031487513574, + "grad_norm": 8.531245398152228, + "learning_rate": 1.1740714673678425e-05, + "loss": 0.4419, + "step": 4246 + }, + { + "epoch": 2.305646036916395, + "grad_norm": 8.091385303458143, + "learning_rate": 1.1737251418193669e-05, + "loss": 0.6249, + "step": 4247 + }, + { + "epoch": 2.3061889250814334, + "grad_norm": 10.253051982681455, + "learning_rate": 1.1733787947842892e-05, + "loss": 0.8892, + "step": 4248 + }, + { + "epoch": 2.306731813246471, + "grad_norm": 8.95406502670429, + "learning_rate": 1.1730324263054453e-05, + "loss": 0.7452, + "step": 4249 + }, + { + "epoch": 2.3072747014115094, + "grad_norm": 10.491159901998326, + "learning_rate": 1.1726860364256756e-05, + "loss": 0.4995, + "step": 4250 + }, + { + "epoch": 2.307817589576547, + "grad_norm": 9.241135791509704, + "learning_rate": 1.1723396251878214e-05, + "loss": 0.7167, + "step": 4251 + }, + { + "epoch": 2.3083604777415854, + "grad_norm": 9.466621171091207, + "learning_rate": 1.1719931926347279e-05, + "loss": 0.5849, + "step": 4252 + }, + { + "epoch": 2.308903365906623, + "grad_norm": 10.933415330757905, + "learning_rate": 1.1716467388092423e-05, + "loss": 1.0411, + "step": 4253 + }, + { + "epoch": 2.3094462540716614, + "grad_norm": 10.009779873858768, + "learning_rate": 1.1713002637542146e-05, + "loss": 0.5456, + "step": 4254 + }, + { + "epoch": 2.309989142236699, + "grad_norm": 11.678038797864954, + "learning_rate": 1.1709537675124966e-05, + "loss": 0.5537, + "step": 4255 + }, + { + "epoch": 2.3105320304017374, + "grad_norm": 11.578035258644414, + "learning_rate": 1.1706072501269448e-05, + "loss": 0.7804, + "step": 4256 + }, + { + "epoch": 2.311074918566775, + "grad_norm": 8.678307181037598, + "learning_rate": 1.1702607116404163e-05, + "loss": 0.5378, + "step": 4257 + }, + { + "epoch": 2.3116178067318134, + "grad_norm": 8.487670920129334, + "learning_rate": 1.1699141520957716e-05, + "loss": 0.558, + "step": 4258 + }, + { + "epoch": 2.312160694896851, + "grad_norm": 10.92371973791016, + "learning_rate": 1.1695675715358737e-05, + "loss": 0.6615, + "step": 4259 + }, + { + "epoch": 2.3127035830618894, + "grad_norm": 9.360985487621843, + "learning_rate": 1.169220970003588e-05, + "loss": 0.5896, + "step": 4260 + }, + { + "epoch": 2.313246471226927, + "grad_norm": 10.382914656511591, + "learning_rate": 1.1688743475417836e-05, + "loss": 0.8869, + "step": 4261 + }, + { + "epoch": 2.3137893593919654, + "grad_norm": 10.27408683638607, + "learning_rate": 1.1685277041933303e-05, + "loss": 1.0929, + "step": 4262 + }, + { + "epoch": 2.314332247557003, + "grad_norm": 10.738574683790112, + "learning_rate": 1.168181040001102e-05, + "loss": 0.845, + "step": 4263 + }, + { + "epoch": 2.3148751357220414, + "grad_norm": 11.783082834204276, + "learning_rate": 1.1678343550079751e-05, + "loss": 0.8358, + "step": 4264 + }, + { + "epoch": 2.315418023887079, + "grad_norm": 9.277577078134165, + "learning_rate": 1.1674876492568272e-05, + "loss": 0.4774, + "step": 4265 + }, + { + "epoch": 2.3159609120521174, + "grad_norm": 10.232983577395474, + "learning_rate": 1.1671409227905401e-05, + "loss": 0.6358, + "step": 4266 + }, + { + "epoch": 2.316503800217155, + "grad_norm": 7.971516257815676, + "learning_rate": 1.1667941756519973e-05, + "loss": 0.5724, + "step": 4267 + }, + { + "epoch": 2.3170466883821934, + "grad_norm": 6.657326238165826, + "learning_rate": 1.1664474078840848e-05, + "loss": 0.3377, + "step": 4268 + }, + { + "epoch": 2.317589576547231, + "grad_norm": 9.831416988328497, + "learning_rate": 1.1661006195296918e-05, + "loss": 0.7115, + "step": 4269 + }, + { + "epoch": 2.3181324647122694, + "grad_norm": 7.2473297381695225, + "learning_rate": 1.1657538106317094e-05, + "loss": 0.4828, + "step": 4270 + }, + { + "epoch": 2.318675352877307, + "grad_norm": 11.474212650892378, + "learning_rate": 1.1654069812330315e-05, + "loss": 0.66, + "step": 4271 + }, + { + "epoch": 2.3192182410423454, + "grad_norm": 9.354650178037069, + "learning_rate": 1.1650601313765547e-05, + "loss": 0.6651, + "step": 4272 + }, + { + "epoch": 2.319761129207383, + "grad_norm": 8.945160340560053, + "learning_rate": 1.1647132611051779e-05, + "loss": 0.6422, + "step": 4273 + }, + { + "epoch": 2.3203040173724214, + "grad_norm": 8.790681463696455, + "learning_rate": 1.164366370461802e-05, + "loss": 0.6387, + "step": 4274 + }, + { + "epoch": 2.320846905537459, + "grad_norm": 8.950872212457588, + "learning_rate": 1.1640194594893317e-05, + "loss": 0.4455, + "step": 4275 + }, + { + "epoch": 2.3213897937024974, + "grad_norm": 9.879748686678086, + "learning_rate": 1.1636725282306733e-05, + "loss": 0.5539, + "step": 4276 + }, + { + "epoch": 2.321932681867535, + "grad_norm": 8.351885716181597, + "learning_rate": 1.1633255767287356e-05, + "loss": 0.4273, + "step": 4277 + }, + { + "epoch": 2.3224755700325734, + "grad_norm": 12.165689696691171, + "learning_rate": 1.1629786050264304e-05, + "loss": 0.6877, + "step": 4278 + }, + { + "epoch": 2.323018458197611, + "grad_norm": 10.70545900286579, + "learning_rate": 1.1626316131666714e-05, + "loss": 0.6667, + "step": 4279 + }, + { + "epoch": 2.3235613463626494, + "grad_norm": 7.521776321190906, + "learning_rate": 1.1622846011923756e-05, + "loss": 0.6538, + "step": 4280 + }, + { + "epoch": 2.324104234527687, + "grad_norm": 10.429691195494106, + "learning_rate": 1.1619375691464614e-05, + "loss": 0.7918, + "step": 4281 + }, + { + "epoch": 2.3246471226927254, + "grad_norm": 13.40241359482942, + "learning_rate": 1.1615905170718507e-05, + "loss": 0.8492, + "step": 4282 + }, + { + "epoch": 2.3251900108577632, + "grad_norm": 9.180074174181755, + "learning_rate": 1.1612434450114674e-05, + "loss": 0.9197, + "step": 4283 + }, + { + "epoch": 2.3257328990228014, + "grad_norm": 8.662388876238891, + "learning_rate": 1.1608963530082375e-05, + "loss": 0.4911, + "step": 4284 + }, + { + "epoch": 2.3262757871878392, + "grad_norm": 12.915716152146215, + "learning_rate": 1.1605492411050907e-05, + "loss": 1.2342, + "step": 4285 + }, + { + "epoch": 2.3268186753528775, + "grad_norm": 15.064512853420336, + "learning_rate": 1.1602021093449577e-05, + "loss": 0.9441, + "step": 4286 + }, + { + "epoch": 2.3273615635179152, + "grad_norm": 8.152078540931088, + "learning_rate": 1.159854957770772e-05, + "loss": 0.4407, + "step": 4287 + }, + { + "epoch": 2.3279044516829535, + "grad_norm": 13.440334362031807, + "learning_rate": 1.1595077864254708e-05, + "loss": 1.0186, + "step": 4288 + }, + { + "epoch": 2.3284473398479912, + "grad_norm": 14.960404464650807, + "learning_rate": 1.1591605953519925e-05, + "loss": 0.854, + "step": 4289 + }, + { + "epoch": 2.3289902280130295, + "grad_norm": 10.871106066518651, + "learning_rate": 1.1588133845932779e-05, + "loss": 0.8522, + "step": 4290 + }, + { + "epoch": 2.3295331161780672, + "grad_norm": 9.218798486570465, + "learning_rate": 1.1584661541922706e-05, + "loss": 0.5529, + "step": 4291 + }, + { + "epoch": 2.3300760043431055, + "grad_norm": 9.644924391242567, + "learning_rate": 1.158118904191917e-05, + "loss": 0.619, + "step": 4292 + }, + { + "epoch": 2.3306188925081432, + "grad_norm": 7.573505575520159, + "learning_rate": 1.1577716346351655e-05, + "loss": 0.5931, + "step": 4293 + }, + { + "epoch": 2.3311617806731815, + "grad_norm": 11.909869183875244, + "learning_rate": 1.1574243455649666e-05, + "loss": 0.8922, + "step": 4294 + }, + { + "epoch": 2.3317046688382193, + "grad_norm": 13.963856423636752, + "learning_rate": 1.1570770370242737e-05, + "loss": 0.9243, + "step": 4295 + }, + { + "epoch": 2.3322475570032575, + "grad_norm": 8.732100079298183, + "learning_rate": 1.1567297090560425e-05, + "loss": 0.5075, + "step": 4296 + }, + { + "epoch": 2.3327904451682953, + "grad_norm": 8.823180159543773, + "learning_rate": 1.1563823617032313e-05, + "loss": 0.5471, + "step": 4297 + }, + { + "epoch": 2.3333333333333335, + "grad_norm": 8.566225161651246, + "learning_rate": 1.1560349950088001e-05, + "loss": 0.5719, + "step": 4298 + }, + { + "epoch": 2.3338762214983713, + "grad_norm": 10.549198011052681, + "learning_rate": 1.155687609015712e-05, + "loss": 0.4775, + "step": 4299 + }, + { + "epoch": 2.3344191096634095, + "grad_norm": 9.79276879932382, + "learning_rate": 1.1553402037669326e-05, + "loss": 0.5429, + "step": 4300 + }, + { + "epoch": 2.3349619978284473, + "grad_norm": 12.065070131397079, + "learning_rate": 1.154992779305429e-05, + "loss": 0.9434, + "step": 4301 + }, + { + "epoch": 2.3355048859934855, + "grad_norm": 11.007790011549671, + "learning_rate": 1.1546453356741714e-05, + "loss": 0.6199, + "step": 4302 + }, + { + "epoch": 2.3360477741585233, + "grad_norm": 12.466762411503309, + "learning_rate": 1.1542978729161326e-05, + "loss": 0.7941, + "step": 4303 + }, + { + "epoch": 2.3365906623235615, + "grad_norm": 11.55197462123976, + "learning_rate": 1.1539503910742866e-05, + "loss": 0.6766, + "step": 4304 + }, + { + "epoch": 2.3371335504885993, + "grad_norm": 8.462143561329682, + "learning_rate": 1.153602890191611e-05, + "loss": 0.5183, + "step": 4305 + }, + { + "epoch": 2.3376764386536375, + "grad_norm": 13.818285051654057, + "learning_rate": 1.1532553703110852e-05, + "loss": 0.8363, + "step": 4306 + }, + { + "epoch": 2.3382193268186753, + "grad_norm": 8.676120343621662, + "learning_rate": 1.1529078314756905e-05, + "loss": 0.6976, + "step": 4307 + }, + { + "epoch": 2.3387622149837135, + "grad_norm": 10.024443385726268, + "learning_rate": 1.1525602737284119e-05, + "loss": 0.4879, + "step": 4308 + }, + { + "epoch": 2.3393051031487513, + "grad_norm": 11.965265440206402, + "learning_rate": 1.1522126971122355e-05, + "loss": 0.8655, + "step": 4309 + }, + { + "epoch": 2.3398479913137895, + "grad_norm": 9.503916256652449, + "learning_rate": 1.15186510167015e-05, + "loss": 0.8966, + "step": 4310 + }, + { + "epoch": 2.3403908794788273, + "grad_norm": 11.942859003506952, + "learning_rate": 1.1515174874451466e-05, + "loss": 0.7769, + "step": 4311 + }, + { + "epoch": 2.3409337676438655, + "grad_norm": 9.50234633554061, + "learning_rate": 1.1511698544802193e-05, + "loss": 0.552, + "step": 4312 + }, + { + "epoch": 2.3414766558089033, + "grad_norm": 8.495987313736284, + "learning_rate": 1.1508222028183632e-05, + "loss": 0.8566, + "step": 4313 + }, + { + "epoch": 2.3420195439739415, + "grad_norm": 11.854978708479615, + "learning_rate": 1.1504745325025766e-05, + "loss": 0.6736, + "step": 4314 + }, + { + "epoch": 2.3425624321389793, + "grad_norm": 11.169148481350696, + "learning_rate": 1.1501268435758602e-05, + "loss": 0.8075, + "step": 4315 + }, + { + "epoch": 2.3431053203040175, + "grad_norm": 11.983506203130691, + "learning_rate": 1.1497791360812163e-05, + "loss": 0.6407, + "step": 4316 + }, + { + "epoch": 2.3436482084690553, + "grad_norm": 8.507852970222672, + "learning_rate": 1.1494314100616502e-05, + "loss": 0.6247, + "step": 4317 + }, + { + "epoch": 2.3441910966340935, + "grad_norm": 19.964294777139003, + "learning_rate": 1.1490836655601692e-05, + "loss": 0.9219, + "step": 4318 + }, + { + "epoch": 2.3447339847991313, + "grad_norm": 8.202074828225685, + "learning_rate": 1.1487359026197828e-05, + "loss": 0.4887, + "step": 4319 + }, + { + "epoch": 2.3452768729641695, + "grad_norm": 10.22404099258637, + "learning_rate": 1.1483881212835029e-05, + "loss": 0.4028, + "step": 4320 + }, + { + "epoch": 2.3458197611292073, + "grad_norm": 8.30587529218613, + "learning_rate": 1.1480403215943436e-05, + "loss": 0.4047, + "step": 4321 + }, + { + "epoch": 2.3463626492942455, + "grad_norm": 7.121924947407594, + "learning_rate": 1.1476925035953217e-05, + "loss": 0.4922, + "step": 4322 + }, + { + "epoch": 2.3469055374592833, + "grad_norm": 10.58793763221932, + "learning_rate": 1.1473446673294552e-05, + "loss": 0.9691, + "step": 4323 + }, + { + "epoch": 2.3474484256243215, + "grad_norm": 11.51770882137354, + "learning_rate": 1.1469968128397656e-05, + "loss": 0.6713, + "step": 4324 + }, + { + "epoch": 2.3479913137893593, + "grad_norm": 9.960125063324698, + "learning_rate": 1.146648940169276e-05, + "loss": 0.8266, + "step": 4325 + }, + { + "epoch": 2.3485342019543975, + "grad_norm": 14.574982915069748, + "learning_rate": 1.1463010493610114e-05, + "loss": 1.4129, + "step": 4326 + }, + { + "epoch": 2.3490770901194353, + "grad_norm": 7.9866843222052, + "learning_rate": 1.145953140458e-05, + "loss": 0.6368, + "step": 4327 + }, + { + "epoch": 2.3496199782844736, + "grad_norm": 14.879386143441858, + "learning_rate": 1.1456052135032717e-05, + "loss": 0.579, + "step": 4328 + }, + { + "epoch": 2.3501628664495113, + "grad_norm": 9.951154243493171, + "learning_rate": 1.1452572685398586e-05, + "loss": 0.6989, + "step": 4329 + }, + { + "epoch": 2.3507057546145496, + "grad_norm": 8.242511521892395, + "learning_rate": 1.144909305610795e-05, + "loss": 0.5213, + "step": 4330 + }, + { + "epoch": 2.3512486427795873, + "grad_norm": 10.259159837692009, + "learning_rate": 1.1445613247591174e-05, + "loss": 0.9849, + "step": 4331 + }, + { + "epoch": 2.3517915309446256, + "grad_norm": 7.97479592417325, + "learning_rate": 1.1442133260278648e-05, + "loss": 0.5867, + "step": 4332 + }, + { + "epoch": 2.3523344191096633, + "grad_norm": 8.493125656661944, + "learning_rate": 1.143865309460078e-05, + "loss": 0.9095, + "step": 4333 + }, + { + "epoch": 2.3528773072747016, + "grad_norm": 12.49067508129187, + "learning_rate": 1.1435172750988009e-05, + "loss": 0.8484, + "step": 4334 + }, + { + "epoch": 2.3534201954397393, + "grad_norm": 11.134649233778442, + "learning_rate": 1.1431692229870782e-05, + "loss": 0.8422, + "step": 4335 + }, + { + "epoch": 2.3539630836047776, + "grad_norm": 6.99064506687852, + "learning_rate": 1.1428211531679576e-05, + "loss": 0.5388, + "step": 4336 + }, + { + "epoch": 2.3545059717698154, + "grad_norm": 7.798255243454909, + "learning_rate": 1.1424730656844895e-05, + "loss": 0.593, + "step": 4337 + }, + { + "epoch": 2.3550488599348536, + "grad_norm": 8.178641043550261, + "learning_rate": 1.1421249605797255e-05, + "loss": 0.4778, + "step": 4338 + }, + { + "epoch": 2.3555917480998914, + "grad_norm": 8.917024765042555, + "learning_rate": 1.1417768378967196e-05, + "loss": 0.5782, + "step": 4339 + }, + { + "epoch": 2.3561346362649296, + "grad_norm": 9.742472427719518, + "learning_rate": 1.1414286976785286e-05, + "loss": 0.8709, + "step": 4340 + }, + { + "epoch": 2.3566775244299674, + "grad_norm": 10.08457903035175, + "learning_rate": 1.141080539968211e-05, + "loss": 0.5582, + "step": 4341 + }, + { + "epoch": 2.3572204125950056, + "grad_norm": 10.74958408516018, + "learning_rate": 1.1407323648088275e-05, + "loss": 0.7498, + "step": 4342 + }, + { + "epoch": 2.3577633007600434, + "grad_norm": 12.356830179446728, + "learning_rate": 1.1403841722434405e-05, + "loss": 0.9357, + "step": 4343 + }, + { + "epoch": 2.3583061889250816, + "grad_norm": 8.220220873503408, + "learning_rate": 1.1400359623151153e-05, + "loss": 0.5554, + "step": 4344 + }, + { + "epoch": 2.3588490770901194, + "grad_norm": 10.916598567149965, + "learning_rate": 1.1396877350669194e-05, + "loss": 0.8239, + "step": 4345 + }, + { + "epoch": 2.3593919652551576, + "grad_norm": 10.51236366107464, + "learning_rate": 1.1393394905419215e-05, + "loss": 0.5945, + "step": 4346 + }, + { + "epoch": 2.3599348534201954, + "grad_norm": 10.534153676139686, + "learning_rate": 1.1389912287831935e-05, + "loss": 0.575, + "step": 4347 + }, + { + "epoch": 2.3604777415852336, + "grad_norm": 10.06758846687261, + "learning_rate": 1.1386429498338088e-05, + "loss": 0.702, + "step": 4348 + }, + { + "epoch": 2.3610206297502714, + "grad_norm": 9.282335539362101, + "learning_rate": 1.1382946537368432e-05, + "loss": 0.5645, + "step": 4349 + }, + { + "epoch": 2.3615635179153096, + "grad_norm": 8.137108078430895, + "learning_rate": 1.1379463405353741e-05, + "loss": 0.5057, + "step": 4350 + }, + { + "epoch": 2.3621064060803474, + "grad_norm": 10.651520296934317, + "learning_rate": 1.137598010272482e-05, + "loss": 0.762, + "step": 4351 + }, + { + "epoch": 2.3626492942453856, + "grad_norm": 10.001083727438939, + "learning_rate": 1.1372496629912486e-05, + "loss": 0.7255, + "step": 4352 + }, + { + "epoch": 2.3631921824104234, + "grad_norm": 12.180927670825834, + "learning_rate": 1.1369012987347582e-05, + "loss": 0.7295, + "step": 4353 + }, + { + "epoch": 2.3637350705754616, + "grad_norm": 10.697210588401258, + "learning_rate": 1.136552917546097e-05, + "loss": 0.5159, + "step": 4354 + }, + { + "epoch": 2.3642779587404994, + "grad_norm": 13.551389197850602, + "learning_rate": 1.1362045194683534e-05, + "loss": 0.9265, + "step": 4355 + }, + { + "epoch": 2.3648208469055376, + "grad_norm": 12.38857946888456, + "learning_rate": 1.1358561045446175e-05, + "loss": 0.7577, + "step": 4356 + }, + { + "epoch": 2.3653637350705754, + "grad_norm": 10.138108604403497, + "learning_rate": 1.1355076728179824e-05, + "loss": 0.5328, + "step": 4357 + }, + { + "epoch": 2.3659066232356136, + "grad_norm": 9.823785539850283, + "learning_rate": 1.1351592243315416e-05, + "loss": 0.6116, + "step": 4358 + }, + { + "epoch": 2.3664495114006514, + "grad_norm": 8.238880239030014, + "learning_rate": 1.1348107591283932e-05, + "loss": 0.3863, + "step": 4359 + }, + { + "epoch": 2.3669923995656896, + "grad_norm": 11.34031175123055, + "learning_rate": 1.1344622772516352e-05, + "loss": 0.9601, + "step": 4360 + }, + { + "epoch": 2.3675352877307274, + "grad_norm": 11.270983899241404, + "learning_rate": 1.1341137787443683e-05, + "loss": 0.7598, + "step": 4361 + }, + { + "epoch": 2.3680781758957656, + "grad_norm": 10.649130568340542, + "learning_rate": 1.1337652636496959e-05, + "loss": 0.6547, + "step": 4362 + }, + { + "epoch": 2.3686210640608034, + "grad_norm": 9.462979596192575, + "learning_rate": 1.1334167320107224e-05, + "loss": 0.5247, + "step": 4363 + }, + { + "epoch": 2.3691639522258416, + "grad_norm": 11.496313370817235, + "learning_rate": 1.1330681838705544e-05, + "loss": 0.5465, + "step": 4364 + }, + { + "epoch": 2.3697068403908794, + "grad_norm": 10.125563730301518, + "learning_rate": 1.1327196192723016e-05, + "loss": 0.5687, + "step": 4365 + }, + { + "epoch": 2.3702497285559176, + "grad_norm": 14.511657264928447, + "learning_rate": 1.1323710382590747e-05, + "loss": 0.7274, + "step": 4366 + }, + { + "epoch": 2.3707926167209554, + "grad_norm": 10.08856219742556, + "learning_rate": 1.1320224408739868e-05, + "loss": 0.4633, + "step": 4367 + }, + { + "epoch": 2.3713355048859937, + "grad_norm": 9.943874647753937, + "learning_rate": 1.131673827160153e-05, + "loss": 0.7042, + "step": 4368 + }, + { + "epoch": 2.3718783930510314, + "grad_norm": 10.550654748070723, + "learning_rate": 1.1313251971606906e-05, + "loss": 0.6336, + "step": 4369 + }, + { + "epoch": 2.3724212812160697, + "grad_norm": 14.84394771322204, + "learning_rate": 1.1309765509187182e-05, + "loss": 0.6591, + "step": 4370 + }, + { + "epoch": 2.3729641693811074, + "grad_norm": 10.962774929698096, + "learning_rate": 1.1306278884773576e-05, + "loss": 0.9648, + "step": 4371 + }, + { + "epoch": 2.3735070575461457, + "grad_norm": 9.063486875880555, + "learning_rate": 1.1302792098797312e-05, + "loss": 0.4618, + "step": 4372 + }, + { + "epoch": 2.3740499457111834, + "grad_norm": 10.458070080982525, + "learning_rate": 1.1299305151689648e-05, + "loss": 0.6216, + "step": 4373 + }, + { + "epoch": 2.3745928338762217, + "grad_norm": 10.041421519621146, + "learning_rate": 1.1295818043881852e-05, + "loss": 0.6642, + "step": 4374 + }, + { + "epoch": 2.3751357220412594, + "grad_norm": 8.817266669804289, + "learning_rate": 1.1292330775805214e-05, + "loss": 0.4725, + "step": 4375 + }, + { + "epoch": 2.3756786102062977, + "grad_norm": 12.823614653687773, + "learning_rate": 1.1288843347891045e-05, + "loss": 0.553, + "step": 4376 + }, + { + "epoch": 2.3762214983713354, + "grad_norm": 9.281557160359569, + "learning_rate": 1.1285355760570677e-05, + "loss": 0.6436, + "step": 4377 + }, + { + "epoch": 2.3767643865363737, + "grad_norm": 13.234297666234115, + "learning_rate": 1.1281868014275458e-05, + "loss": 0.7665, + "step": 4378 + }, + { + "epoch": 2.3773072747014115, + "grad_norm": 13.013112559574942, + "learning_rate": 1.1278380109436764e-05, + "loss": 0.9087, + "step": 4379 + }, + { + "epoch": 2.3778501628664497, + "grad_norm": 9.413433369666237, + "learning_rate": 1.1274892046485979e-05, + "loss": 0.5248, + "step": 4380 + }, + { + "epoch": 2.3783930510314875, + "grad_norm": 10.865873661981059, + "learning_rate": 1.1271403825854516e-05, + "loss": 0.7398, + "step": 4381 + }, + { + "epoch": 2.3789359391965257, + "grad_norm": 8.516596816721957, + "learning_rate": 1.12679154479738e-05, + "loss": 0.4652, + "step": 4382 + }, + { + "epoch": 2.3794788273615635, + "grad_norm": 7.9464803374429085, + "learning_rate": 1.1264426913275277e-05, + "loss": 0.6952, + "step": 4383 + }, + { + "epoch": 2.3800217155266017, + "grad_norm": 9.171877282099603, + "learning_rate": 1.1260938222190423e-05, + "loss": 0.7534, + "step": 4384 + }, + { + "epoch": 2.3805646036916395, + "grad_norm": 10.833526250938174, + "learning_rate": 1.1257449375150717e-05, + "loss": 0.7966, + "step": 4385 + }, + { + "epoch": 2.3811074918566777, + "grad_norm": 7.087015821603363, + "learning_rate": 1.1253960372587669e-05, + "loss": 0.5145, + "step": 4386 + }, + { + "epoch": 2.3816503800217155, + "grad_norm": 9.355325600419727, + "learning_rate": 1.1250471214932805e-05, + "loss": 0.6084, + "step": 4387 + }, + { + "epoch": 2.3821932681867537, + "grad_norm": 11.679312988029317, + "learning_rate": 1.1246981902617666e-05, + "loss": 0.7734, + "step": 4388 + }, + { + "epoch": 2.3827361563517915, + "grad_norm": 11.696985760230557, + "learning_rate": 1.124349243607382e-05, + "loss": 0.7843, + "step": 4389 + }, + { + "epoch": 2.3832790445168297, + "grad_norm": 11.045261072625825, + "learning_rate": 1.1240002815732847e-05, + "loss": 0.8102, + "step": 4390 + }, + { + "epoch": 2.3838219326818675, + "grad_norm": 13.853716237422205, + "learning_rate": 1.123651304202635e-05, + "loss": 0.6548, + "step": 4391 + }, + { + "epoch": 2.3843648208469057, + "grad_norm": 9.256284103603257, + "learning_rate": 1.1233023115385949e-05, + "loss": 0.5769, + "step": 4392 + }, + { + "epoch": 2.3849077090119435, + "grad_norm": 11.719289709502613, + "learning_rate": 1.1229533036243283e-05, + "loss": 0.9445, + "step": 4393 + }, + { + "epoch": 2.3854505971769817, + "grad_norm": 13.181528127455582, + "learning_rate": 1.1226042805030014e-05, + "loss": 0.78, + "step": 4394 + }, + { + "epoch": 2.3859934853420195, + "grad_norm": 10.610212858388932, + "learning_rate": 1.1222552422177813e-05, + "loss": 0.9325, + "step": 4395 + }, + { + "epoch": 2.3865363735070577, + "grad_norm": 8.750069163521957, + "learning_rate": 1.1219061888118381e-05, + "loss": 0.5576, + "step": 4396 + }, + { + "epoch": 2.3870792616720955, + "grad_norm": 11.697241415039246, + "learning_rate": 1.121557120328343e-05, + "loss": 0.6042, + "step": 4397 + }, + { + "epoch": 2.3876221498371337, + "grad_norm": 14.150909202475516, + "learning_rate": 1.1212080368104699e-05, + "loss": 0.8888, + "step": 4398 + }, + { + "epoch": 2.3881650380021715, + "grad_norm": 8.45338495407854, + "learning_rate": 1.1208589383013934e-05, + "loss": 0.5394, + "step": 4399 + }, + { + "epoch": 2.3887079261672097, + "grad_norm": 8.581855639244504, + "learning_rate": 1.120509824844291e-05, + "loss": 0.4231, + "step": 4400 + }, + { + "epoch": 2.3892508143322475, + "grad_norm": 9.723553293714806, + "learning_rate": 1.1201606964823415e-05, + "loss": 0.5184, + "step": 4401 + }, + { + "epoch": 2.3897937024972857, + "grad_norm": 11.437338247112033, + "learning_rate": 1.1198115532587254e-05, + "loss": 0.9033, + "step": 4402 + }, + { + "epoch": 2.3903365906623235, + "grad_norm": 11.185644591472458, + "learning_rate": 1.1194623952166252e-05, + "loss": 0.7607, + "step": 4403 + }, + { + "epoch": 2.3908794788273617, + "grad_norm": 11.125764861054968, + "learning_rate": 1.1191132223992258e-05, + "loss": 0.7844, + "step": 4404 + }, + { + "epoch": 2.3914223669923995, + "grad_norm": 10.915426846928037, + "learning_rate": 1.1187640348497131e-05, + "loss": 0.5862, + "step": 4405 + }, + { + "epoch": 2.3919652551574377, + "grad_norm": 8.032417531111763, + "learning_rate": 1.1184148326112753e-05, + "loss": 0.5992, + "step": 4406 + }, + { + "epoch": 2.3925081433224755, + "grad_norm": 10.322050440175012, + "learning_rate": 1.1180656157271026e-05, + "loss": 0.6275, + "step": 4407 + }, + { + "epoch": 2.3930510314875137, + "grad_norm": 5.756218508975568, + "learning_rate": 1.117716384240386e-05, + "loss": 0.3605, + "step": 4408 + }, + { + "epoch": 2.3935939196525515, + "grad_norm": 13.072320066260373, + "learning_rate": 1.1173671381943197e-05, + "loss": 1.1771, + "step": 4409 + }, + { + "epoch": 2.3941368078175898, + "grad_norm": 8.390483377503248, + "learning_rate": 1.1170178776320982e-05, + "loss": 0.484, + "step": 4410 + }, + { + "epoch": 2.3946796959826275, + "grad_norm": 9.168734569918955, + "learning_rate": 1.1166686025969196e-05, + "loss": 0.4637, + "step": 4411 + }, + { + "epoch": 2.3952225841476658, + "grad_norm": 11.945726197992448, + "learning_rate": 1.116319313131982e-05, + "loss": 0.6501, + "step": 4412 + }, + { + "epoch": 2.3957654723127035, + "grad_norm": 9.925140956274975, + "learning_rate": 1.1159700092804866e-05, + "loss": 0.5583, + "step": 4413 + }, + { + "epoch": 2.3963083604777418, + "grad_norm": 10.0377125619045, + "learning_rate": 1.1156206910856353e-05, + "loss": 0.5783, + "step": 4414 + }, + { + "epoch": 2.3968512486427795, + "grad_norm": 13.878494510161136, + "learning_rate": 1.115271358590633e-05, + "loss": 0.7029, + "step": 4415 + }, + { + "epoch": 2.3973941368078178, + "grad_norm": 8.414419089776398, + "learning_rate": 1.114922011838685e-05, + "loss": 0.4398, + "step": 4416 + }, + { + "epoch": 2.3979370249728555, + "grad_norm": 10.455474671185716, + "learning_rate": 1.1145726508729994e-05, + "loss": 0.5462, + "step": 4417 + }, + { + "epoch": 2.3984799131378938, + "grad_norm": 8.80305307358664, + "learning_rate": 1.1142232757367861e-05, + "loss": 0.3885, + "step": 4418 + }, + { + "epoch": 2.3990228013029316, + "grad_norm": 9.381551586854757, + "learning_rate": 1.1138738864732559e-05, + "loss": 0.6646, + "step": 4419 + }, + { + "epoch": 2.3995656894679698, + "grad_norm": 9.409317053539013, + "learning_rate": 1.113524483125622e-05, + "loss": 0.7772, + "step": 4420 + }, + { + "epoch": 2.4001085776330076, + "grad_norm": 7.414798924114551, + "learning_rate": 1.113175065737099e-05, + "loss": 0.4508, + "step": 4421 + }, + { + "epoch": 2.400651465798046, + "grad_norm": 10.931216387643186, + "learning_rate": 1.1128256343509032e-05, + "loss": 0.6434, + "step": 4422 + }, + { + "epoch": 2.4011943539630836, + "grad_norm": 9.811349462515302, + "learning_rate": 1.1124761890102535e-05, + "loss": 0.5104, + "step": 4423 + }, + { + "epoch": 2.401737242128122, + "grad_norm": 10.856233458965448, + "learning_rate": 1.1121267297583694e-05, + "loss": 0.6535, + "step": 4424 + }, + { + "epoch": 2.4022801302931596, + "grad_norm": 10.2119482085421, + "learning_rate": 1.1117772566384725e-05, + "loss": 0.593, + "step": 4425 + }, + { + "epoch": 2.402823018458198, + "grad_norm": 11.327455944037217, + "learning_rate": 1.1114277696937865e-05, + "loss": 0.9199, + "step": 4426 + }, + { + "epoch": 2.4033659066232356, + "grad_norm": 10.64004734338448, + "learning_rate": 1.1110782689675365e-05, + "loss": 0.793, + "step": 4427 + }, + { + "epoch": 2.403908794788274, + "grad_norm": 5.71867532259649, + "learning_rate": 1.110728754502949e-05, + "loss": 0.2726, + "step": 4428 + }, + { + "epoch": 2.4044516829533116, + "grad_norm": 10.33820400695821, + "learning_rate": 1.1103792263432525e-05, + "loss": 0.787, + "step": 4429 + }, + { + "epoch": 2.40499457111835, + "grad_norm": 10.02381386485252, + "learning_rate": 1.1100296845316777e-05, + "loss": 0.7701, + "step": 4430 + }, + { + "epoch": 2.4055374592833876, + "grad_norm": 12.111852512562214, + "learning_rate": 1.1096801291114561e-05, + "loss": 0.6123, + "step": 4431 + }, + { + "epoch": 2.406080347448426, + "grad_norm": 8.432777542688768, + "learning_rate": 1.1093305601258213e-05, + "loss": 0.4817, + "step": 4432 + }, + { + "epoch": 2.4066232356134636, + "grad_norm": 12.648769211741241, + "learning_rate": 1.1089809776180084e-05, + "loss": 0.7328, + "step": 4433 + }, + { + "epoch": 2.407166123778502, + "grad_norm": 9.260727895387426, + "learning_rate": 1.1086313816312546e-05, + "loss": 0.2951, + "step": 4434 + }, + { + "epoch": 2.4077090119435396, + "grad_norm": 7.801968230318713, + "learning_rate": 1.1082817722087981e-05, + "loss": 0.6204, + "step": 4435 + }, + { + "epoch": 2.408251900108578, + "grad_norm": 18.28104514211005, + "learning_rate": 1.10793214939388e-05, + "loss": 1.2717, + "step": 4436 + }, + { + "epoch": 2.4087947882736156, + "grad_norm": 8.074911353186037, + "learning_rate": 1.1075825132297414e-05, + "loss": 0.391, + "step": 4437 + }, + { + "epoch": 2.409337676438654, + "grad_norm": 10.223302145592744, + "learning_rate": 1.1072328637596259e-05, + "loss": 0.8764, + "step": 4438 + }, + { + "epoch": 2.4098805646036916, + "grad_norm": 8.909676311441572, + "learning_rate": 1.1068832010267792e-05, + "loss": 0.5866, + "step": 4439 + }, + { + "epoch": 2.41042345276873, + "grad_norm": 8.270875830864977, + "learning_rate": 1.1065335250744478e-05, + "loss": 0.4935, + "step": 4440 + }, + { + "epoch": 2.4109663409337676, + "grad_norm": 8.125694150069458, + "learning_rate": 1.1061838359458801e-05, + "loss": 0.5855, + "step": 4441 + }, + { + "epoch": 2.411509229098806, + "grad_norm": 9.00158598137614, + "learning_rate": 1.1058341336843264e-05, + "loss": 0.4464, + "step": 4442 + }, + { + "epoch": 2.4120521172638436, + "grad_norm": 10.568261973334824, + "learning_rate": 1.1054844183330385e-05, + "loss": 0.844, + "step": 4443 + }, + { + "epoch": 2.412595005428882, + "grad_norm": 10.29648068473729, + "learning_rate": 1.1051346899352693e-05, + "loss": 0.7083, + "step": 4444 + }, + { + "epoch": 2.4131378935939196, + "grad_norm": 10.658053637486233, + "learning_rate": 1.1047849485342744e-05, + "loss": 0.917, + "step": 4445 + }, + { + "epoch": 2.413680781758958, + "grad_norm": 10.85927844636501, + "learning_rate": 1.1044351941733097e-05, + "loss": 0.7737, + "step": 4446 + }, + { + "epoch": 2.4142236699239956, + "grad_norm": 13.105614857394716, + "learning_rate": 1.1040854268956343e-05, + "loss": 0.7626, + "step": 4447 + }, + { + "epoch": 2.414766558089034, + "grad_norm": 12.088913830608213, + "learning_rate": 1.1037356467445071e-05, + "loss": 0.7439, + "step": 4448 + }, + { + "epoch": 2.4153094462540716, + "grad_norm": 13.524031933483466, + "learning_rate": 1.1033858537631894e-05, + "loss": 0.6172, + "step": 4449 + }, + { + "epoch": 2.41585233441911, + "grad_norm": 12.25037908572634, + "learning_rate": 1.1030360479949452e-05, + "loss": 0.7959, + "step": 4450 + }, + { + "epoch": 2.4163952225841476, + "grad_norm": 10.26766943217781, + "learning_rate": 1.102686229483038e-05, + "loss": 0.8928, + "step": 4451 + }, + { + "epoch": 2.416938110749186, + "grad_norm": 10.66247714207056, + "learning_rate": 1.1023363982707345e-05, + "loss": 0.8498, + "step": 4452 + }, + { + "epoch": 2.4174809989142236, + "grad_norm": 9.37397925289073, + "learning_rate": 1.101986554401302e-05, + "loss": 0.5633, + "step": 4453 + }, + { + "epoch": 2.418023887079262, + "grad_norm": 8.656197915403546, + "learning_rate": 1.10163669791801e-05, + "loss": 0.6403, + "step": 4454 + }, + { + "epoch": 2.4185667752442996, + "grad_norm": 10.199878069425685, + "learning_rate": 1.101286828864129e-05, + "loss": 0.7139, + "step": 4455 + }, + { + "epoch": 2.419109663409338, + "grad_norm": 8.176125775602177, + "learning_rate": 1.1009369472829316e-05, + "loss": 0.5322, + "step": 4456 + }, + { + "epoch": 2.4196525515743756, + "grad_norm": 9.441655350085792, + "learning_rate": 1.1005870532176918e-05, + "loss": 0.8574, + "step": 4457 + }, + { + "epoch": 2.420195439739414, + "grad_norm": 10.487844747641686, + "learning_rate": 1.100237146711685e-05, + "loss": 0.7399, + "step": 4458 + }, + { + "epoch": 2.4207383279044516, + "grad_norm": 8.456961156540515, + "learning_rate": 1.0998872278081883e-05, + "loss": 0.5795, + "step": 4459 + }, + { + "epoch": 2.42128121606949, + "grad_norm": 10.247575624097415, + "learning_rate": 1.0995372965504798e-05, + "loss": 0.7048, + "step": 4460 + }, + { + "epoch": 2.4218241042345277, + "grad_norm": 8.127194271974078, + "learning_rate": 1.09918735298184e-05, + "loss": 0.6339, + "step": 4461 + }, + { + "epoch": 2.422366992399566, + "grad_norm": 11.080151666019892, + "learning_rate": 1.0988373971455502e-05, + "loss": 0.6, + "step": 4462 + }, + { + "epoch": 2.4229098805646037, + "grad_norm": 9.249791609753641, + "learning_rate": 1.0984874290848935e-05, + "loss": 0.5823, + "step": 4463 + }, + { + "epoch": 2.423452768729642, + "grad_norm": 8.673439204574109, + "learning_rate": 1.0981374488431546e-05, + "loss": 0.4486, + "step": 4464 + }, + { + "epoch": 2.4239956568946797, + "grad_norm": 8.530441262511525, + "learning_rate": 1.0977874564636199e-05, + "loss": 0.4685, + "step": 4465 + }, + { + "epoch": 2.424538545059718, + "grad_norm": 11.975099667402777, + "learning_rate": 1.0974374519895765e-05, + "loss": 0.5608, + "step": 4466 + }, + { + "epoch": 2.4250814332247557, + "grad_norm": 12.28650683165383, + "learning_rate": 1.097087435464314e-05, + "loss": 0.6428, + "step": 4467 + }, + { + "epoch": 2.425624321389794, + "grad_norm": 11.443904338187624, + "learning_rate": 1.0967374069311224e-05, + "loss": 0.5093, + "step": 4468 + }, + { + "epoch": 2.4261672095548317, + "grad_norm": 12.11642038793139, + "learning_rate": 1.0963873664332944e-05, + "loss": 0.7573, + "step": 4469 + }, + { + "epoch": 2.42671009771987, + "grad_norm": 12.716855768902285, + "learning_rate": 1.0960373140141232e-05, + "loss": 1.0742, + "step": 4470 + }, + { + "epoch": 2.4272529858849077, + "grad_norm": 8.060615842022353, + "learning_rate": 1.095687249716904e-05, + "loss": 0.5299, + "step": 4471 + }, + { + "epoch": 2.427795874049946, + "grad_norm": 10.663178608260651, + "learning_rate": 1.0953371735849329e-05, + "loss": 0.6575, + "step": 4472 + }, + { + "epoch": 2.4283387622149837, + "grad_norm": 11.883215571850693, + "learning_rate": 1.0949870856615088e-05, + "loss": 0.7516, + "step": 4473 + }, + { + "epoch": 2.428881650380022, + "grad_norm": 8.725361323128087, + "learning_rate": 1.0946369859899299e-05, + "loss": 0.5543, + "step": 4474 + }, + { + "epoch": 2.4294245385450597, + "grad_norm": 7.964297132373328, + "learning_rate": 1.0942868746134983e-05, + "loss": 0.3799, + "step": 4475 + }, + { + "epoch": 2.429967426710098, + "grad_norm": 11.129935043446988, + "learning_rate": 1.0939367515755155e-05, + "loss": 0.5274, + "step": 4476 + }, + { + "epoch": 2.4305103148751357, + "grad_norm": 8.84506268400313, + "learning_rate": 1.0935866169192858e-05, + "loss": 0.6314, + "step": 4477 + }, + { + "epoch": 2.431053203040174, + "grad_norm": 9.898346545614544, + "learning_rate": 1.0932364706881138e-05, + "loss": 0.4733, + "step": 4478 + }, + { + "epoch": 2.4315960912052117, + "grad_norm": 12.708528067939865, + "learning_rate": 1.092886312925307e-05, + "loss": 1.1591, + "step": 4479 + }, + { + "epoch": 2.43213897937025, + "grad_norm": 8.852913949596426, + "learning_rate": 1.0925361436741726e-05, + "loss": 0.4934, + "step": 4480 + }, + { + "epoch": 2.4326818675352877, + "grad_norm": 9.976618553528468, + "learning_rate": 1.0921859629780205e-05, + "loss": 0.5726, + "step": 4481 + }, + { + "epoch": 2.433224755700326, + "grad_norm": 10.591185923462703, + "learning_rate": 1.0918357708801617e-05, + "loss": 0.9099, + "step": 4482 + }, + { + "epoch": 2.4337676438653637, + "grad_norm": 9.782319605062739, + "learning_rate": 1.0914855674239083e-05, + "loss": 0.7016, + "step": 4483 + }, + { + "epoch": 2.434310532030402, + "grad_norm": 7.962337485760134, + "learning_rate": 1.0911353526525741e-05, + "loss": 0.3899, + "step": 4484 + }, + { + "epoch": 2.4348534201954397, + "grad_norm": 10.926665567713723, + "learning_rate": 1.0907851266094742e-05, + "loss": 1.0416, + "step": 4485 + }, + { + "epoch": 2.435396308360478, + "grad_norm": 10.774706653267238, + "learning_rate": 1.0904348893379252e-05, + "loss": 0.8539, + "step": 4486 + }, + { + "epoch": 2.4359391965255157, + "grad_norm": 9.962550119554402, + "learning_rate": 1.0900846408812449e-05, + "loss": 0.6202, + "step": 4487 + }, + { + "epoch": 2.436482084690554, + "grad_norm": 7.425101710467455, + "learning_rate": 1.0897343812827528e-05, + "loss": 0.4138, + "step": 4488 + }, + { + "epoch": 2.4370249728555917, + "grad_norm": 11.159990654051901, + "learning_rate": 1.0893841105857695e-05, + "loss": 0.7927, + "step": 4489 + }, + { + "epoch": 2.4375678610206295, + "grad_norm": 8.584412414000917, + "learning_rate": 1.0890338288336168e-05, + "loss": 0.4596, + "step": 4490 + }, + { + "epoch": 2.4381107491856677, + "grad_norm": 13.182969196780135, + "learning_rate": 1.0886835360696182e-05, + "loss": 0.7052, + "step": 4491 + }, + { + "epoch": 2.438653637350706, + "grad_norm": 12.643144928400368, + "learning_rate": 1.0883332323370986e-05, + "loss": 0.7273, + "step": 4492 + }, + { + "epoch": 2.4391965255157437, + "grad_norm": 11.758017246786887, + "learning_rate": 1.087982917679384e-05, + "loss": 0.6389, + "step": 4493 + }, + { + "epoch": 2.4397394136807815, + "grad_norm": 10.332831381965816, + "learning_rate": 1.0876325921398016e-05, + "loss": 0.49, + "step": 4494 + }, + { + "epoch": 2.4402823018458197, + "grad_norm": 12.79909493300538, + "learning_rate": 1.0872822557616811e-05, + "loss": 1.027, + "step": 4495 + }, + { + "epoch": 2.440825190010858, + "grad_norm": 10.225359261113429, + "learning_rate": 1.0869319085883518e-05, + "loss": 0.6319, + "step": 4496 + }, + { + "epoch": 2.4413680781758957, + "grad_norm": 10.600390533226815, + "learning_rate": 1.0865815506631456e-05, + "loss": 0.5513, + "step": 4497 + }, + { + "epoch": 2.4419109663409335, + "grad_norm": 9.069709981087057, + "learning_rate": 1.0862311820293951e-05, + "loss": 0.4358, + "step": 4498 + }, + { + "epoch": 2.4424538545059717, + "grad_norm": 10.228930326876846, + "learning_rate": 1.0858808027304346e-05, + "loss": 1.0149, + "step": 4499 + }, + { + "epoch": 2.44299674267101, + "grad_norm": 6.981580424693794, + "learning_rate": 1.0855304128095997e-05, + "loss": 0.3823, + "step": 4500 + }, + { + "epoch": 2.4435396308360477, + "grad_norm": 12.347328486710586, + "learning_rate": 1.085180012310227e-05, + "loss": 0.8202, + "step": 4501 + }, + { + "epoch": 2.4440825190010855, + "grad_norm": 7.794475194787575, + "learning_rate": 1.0848296012756545e-05, + "loss": 0.4972, + "step": 4502 + }, + { + "epoch": 2.4446254071661238, + "grad_norm": 8.640639888495457, + "learning_rate": 1.0844791797492218e-05, + "loss": 0.4174, + "step": 4503 + }, + { + "epoch": 2.445168295331162, + "grad_norm": 9.363787480425108, + "learning_rate": 1.0841287477742695e-05, + "loss": 0.5783, + "step": 4504 + }, + { + "epoch": 2.4457111834961998, + "grad_norm": 9.455569437628846, + "learning_rate": 1.0837783053941396e-05, + "loss": 0.5611, + "step": 4505 + }, + { + "epoch": 2.4462540716612375, + "grad_norm": 7.587727857256139, + "learning_rate": 1.083427852652175e-05, + "loss": 0.4129, + "step": 4506 + }, + { + "epoch": 2.4467969598262758, + "grad_norm": 11.579317957056839, + "learning_rate": 1.083077389591721e-05, + "loss": 0.6628, + "step": 4507 + }, + { + "epoch": 2.447339847991314, + "grad_norm": 9.130000079525795, + "learning_rate": 1.0827269162561229e-05, + "loss": 0.4626, + "step": 4508 + }, + { + "epoch": 2.4478827361563518, + "grad_norm": 8.791608822057356, + "learning_rate": 1.0823764326887283e-05, + "loss": 0.7441, + "step": 4509 + }, + { + "epoch": 2.4484256243213895, + "grad_norm": 8.698379661142646, + "learning_rate": 1.0820259389328852e-05, + "loss": 0.4924, + "step": 4510 + }, + { + "epoch": 2.4489685124864278, + "grad_norm": 12.811407852860862, + "learning_rate": 1.0816754350319427e-05, + "loss": 0.9829, + "step": 4511 + }, + { + "epoch": 2.449511400651466, + "grad_norm": 11.588653893201888, + "learning_rate": 1.0813249210292525e-05, + "loss": 0.7218, + "step": 4512 + }, + { + "epoch": 2.450054288816504, + "grad_norm": 12.334703169006005, + "learning_rate": 1.0809743969681665e-05, + "loss": 0.8737, + "step": 4513 + }, + { + "epoch": 2.4505971769815416, + "grad_norm": 12.583313163331177, + "learning_rate": 1.080623862892038e-05, + "loss": 0.7226, + "step": 4514 + }, + { + "epoch": 2.45114006514658, + "grad_norm": 12.11103490873811, + "learning_rate": 1.0802733188442216e-05, + "loss": 0.7867, + "step": 4515 + }, + { + "epoch": 2.451682953311618, + "grad_norm": 11.222788743446143, + "learning_rate": 1.0799227648680732e-05, + "loss": 0.5457, + "step": 4516 + }, + { + "epoch": 2.452225841476656, + "grad_norm": 7.685503695913846, + "learning_rate": 1.07957220100695e-05, + "loss": 0.6483, + "step": 4517 + }, + { + "epoch": 2.4527687296416936, + "grad_norm": 8.407197016108878, + "learning_rate": 1.0792216273042099e-05, + "loss": 0.4895, + "step": 4518 + }, + { + "epoch": 2.453311617806732, + "grad_norm": 9.426708870733849, + "learning_rate": 1.0788710438032131e-05, + "loss": 1.0719, + "step": 4519 + }, + { + "epoch": 2.45385450597177, + "grad_norm": 13.638918388289357, + "learning_rate": 1.0785204505473197e-05, + "loss": 0.8828, + "step": 4520 + }, + { + "epoch": 2.454397394136808, + "grad_norm": 12.478693804182608, + "learning_rate": 1.0781698475798917e-05, + "loss": 0.6527, + "step": 4521 + }, + { + "epoch": 2.4549402823018456, + "grad_norm": 10.20424250017949, + "learning_rate": 1.0778192349442923e-05, + "loss": 0.9303, + "step": 4522 + }, + { + "epoch": 2.455483170466884, + "grad_norm": 12.175885097025716, + "learning_rate": 1.0774686126838863e-05, + "loss": 0.6391, + "step": 4523 + }, + { + "epoch": 2.456026058631922, + "grad_norm": 10.568732307533892, + "learning_rate": 1.0771179808420385e-05, + "loss": 0.5524, + "step": 4524 + }, + { + "epoch": 2.45656894679696, + "grad_norm": 7.314447097275954, + "learning_rate": 1.0767673394621159e-05, + "loss": 0.4976, + "step": 4525 + }, + { + "epoch": 2.4571118349619976, + "grad_norm": 11.397811881243696, + "learning_rate": 1.0764166885874862e-05, + "loss": 0.4918, + "step": 4526 + }, + { + "epoch": 2.457654723127036, + "grad_norm": 9.691957338181856, + "learning_rate": 1.0760660282615191e-05, + "loss": 0.6725, + "step": 4527 + }, + { + "epoch": 2.458197611292074, + "grad_norm": 7.510379855172465, + "learning_rate": 1.0757153585275844e-05, + "loss": 0.4684, + "step": 4528 + }, + { + "epoch": 2.458740499457112, + "grad_norm": 7.751570484202345, + "learning_rate": 1.0753646794290534e-05, + "loss": 0.5495, + "step": 4529 + }, + { + "epoch": 2.4592833876221496, + "grad_norm": 7.390182612219503, + "learning_rate": 1.0750139910092987e-05, + "loss": 0.4538, + "step": 4530 + }, + { + "epoch": 2.459826275787188, + "grad_norm": 10.180132351193016, + "learning_rate": 1.074663293311694e-05, + "loss": 0.5579, + "step": 4531 + }, + { + "epoch": 2.460369163952226, + "grad_norm": 8.909847677295557, + "learning_rate": 1.074312586379614e-05, + "loss": 0.5023, + "step": 4532 + }, + { + "epoch": 2.460912052117264, + "grad_norm": 10.091297535262147, + "learning_rate": 1.073961870256435e-05, + "loss": 0.7068, + "step": 4533 + }, + { + "epoch": 2.4614549402823016, + "grad_norm": 11.778234555569338, + "learning_rate": 1.0736111449855341e-05, + "loss": 0.7687, + "step": 4534 + }, + { + "epoch": 2.46199782844734, + "grad_norm": 11.871444812620142, + "learning_rate": 1.0732604106102895e-05, + "loss": 0.8333, + "step": 4535 + }, + { + "epoch": 2.462540716612378, + "grad_norm": 6.164790402663993, + "learning_rate": 1.0729096671740806e-05, + "loss": 0.3601, + "step": 4536 + }, + { + "epoch": 2.463083604777416, + "grad_norm": 9.524060445013603, + "learning_rate": 1.0725589147202877e-05, + "loss": 0.6859, + "step": 4537 + }, + { + "epoch": 2.4636264929424536, + "grad_norm": 11.266171306725775, + "learning_rate": 1.0722081532922925e-05, + "loss": 0.7459, + "step": 4538 + }, + { + "epoch": 2.464169381107492, + "grad_norm": 8.945835373858307, + "learning_rate": 1.0718573829334782e-05, + "loss": 0.3787, + "step": 4539 + }, + { + "epoch": 2.46471226927253, + "grad_norm": 6.838806611638338, + "learning_rate": 1.071506603687228e-05, + "loss": 0.4995, + "step": 4540 + }, + { + "epoch": 2.465255157437568, + "grad_norm": 11.246624755560871, + "learning_rate": 1.071155815596927e-05, + "loss": 0.8699, + "step": 4541 + }, + { + "epoch": 2.4657980456026056, + "grad_norm": 7.324342918236285, + "learning_rate": 1.0708050187059615e-05, + "loss": 0.4778, + "step": 4542 + }, + { + "epoch": 2.466340933767644, + "grad_norm": 10.063475303955434, + "learning_rate": 1.0704542130577185e-05, + "loss": 0.5774, + "step": 4543 + }, + { + "epoch": 2.466883821932682, + "grad_norm": 12.644825932835714, + "learning_rate": 1.070103398695586e-05, + "loss": 0.8158, + "step": 4544 + }, + { + "epoch": 2.46742671009772, + "grad_norm": 10.332286780133, + "learning_rate": 1.0697525756629531e-05, + "loss": 0.6518, + "step": 4545 + }, + { + "epoch": 2.4679695982627576, + "grad_norm": 9.98899797230979, + "learning_rate": 1.069401744003211e-05, + "loss": 0.6167, + "step": 4546 + }, + { + "epoch": 2.468512486427796, + "grad_norm": 13.445015438013797, + "learning_rate": 1.0690509037597502e-05, + "loss": 0.8229, + "step": 4547 + }, + { + "epoch": 2.469055374592834, + "grad_norm": 13.637667335089724, + "learning_rate": 1.0687000549759641e-05, + "loss": 0.7186, + "step": 4548 + }, + { + "epoch": 2.469598262757872, + "grad_norm": 10.021629994296763, + "learning_rate": 1.0683491976952454e-05, + "loss": 0.6209, + "step": 4549 + }, + { + "epoch": 2.4701411509229096, + "grad_norm": 10.510874896443969, + "learning_rate": 1.067998331960989e-05, + "loss": 0.9009, + "step": 4550 + }, + { + "epoch": 2.470684039087948, + "grad_norm": 9.53841539843291, + "learning_rate": 1.0676474578165903e-05, + "loss": 0.7187, + "step": 4551 + }, + { + "epoch": 2.471226927252986, + "grad_norm": 11.693787891306421, + "learning_rate": 1.0672965753054466e-05, + "loss": 0.5568, + "step": 4552 + }, + { + "epoch": 2.471769815418024, + "grad_norm": 11.523811853237843, + "learning_rate": 1.0669456844709549e-05, + "loss": 0.8465, + "step": 4553 + }, + { + "epoch": 2.4723127035830617, + "grad_norm": 9.165552336797202, + "learning_rate": 1.0665947853565144e-05, + "loss": 0.7148, + "step": 4554 + }, + { + "epoch": 2.4728555917481, + "grad_norm": 10.390907422337587, + "learning_rate": 1.0662438780055247e-05, + "loss": 0.5271, + "step": 4555 + }, + { + "epoch": 2.473398479913138, + "grad_norm": 10.514660404488053, + "learning_rate": 1.0658929624613866e-05, + "loss": 0.7972, + "step": 4556 + }, + { + "epoch": 2.473941368078176, + "grad_norm": 8.804604897853023, + "learning_rate": 1.065542038767502e-05, + "loss": 0.4391, + "step": 4557 + }, + { + "epoch": 2.4744842562432137, + "grad_norm": 12.58202482779834, + "learning_rate": 1.0651911069672733e-05, + "loss": 0.601, + "step": 4558 + }, + { + "epoch": 2.475027144408252, + "grad_norm": 8.510520469826693, + "learning_rate": 1.0648401671041048e-05, + "loss": 0.6638, + "step": 4559 + }, + { + "epoch": 2.47557003257329, + "grad_norm": 10.132731434552191, + "learning_rate": 1.0644892192214012e-05, + "loss": 0.547, + "step": 4560 + }, + { + "epoch": 2.476112920738328, + "grad_norm": 9.486912806484277, + "learning_rate": 1.0641382633625678e-05, + "loss": 0.6903, + "step": 4561 + }, + { + "epoch": 2.4766558089033657, + "grad_norm": 8.656529900389968, + "learning_rate": 1.0637872995710122e-05, + "loss": 0.6445, + "step": 4562 + }, + { + "epoch": 2.477198697068404, + "grad_norm": 8.377471382601675, + "learning_rate": 1.0634363278901413e-05, + "loss": 0.475, + "step": 4563 + }, + { + "epoch": 2.477741585233442, + "grad_norm": 11.941789514685023, + "learning_rate": 1.0630853483633643e-05, + "loss": 0.8705, + "step": 4564 + }, + { + "epoch": 2.47828447339848, + "grad_norm": 8.551632836218364, + "learning_rate": 1.0627343610340912e-05, + "loss": 0.3577, + "step": 4565 + }, + { + "epoch": 2.4788273615635177, + "grad_norm": 10.220220103336683, + "learning_rate": 1.0623833659457322e-05, + "loss": 0.7002, + "step": 4566 + }, + { + "epoch": 2.479370249728556, + "grad_norm": 10.783849224421678, + "learning_rate": 1.0620323631416994e-05, + "loss": 0.6969, + "step": 4567 + }, + { + "epoch": 2.479913137893594, + "grad_norm": 9.018283325771, + "learning_rate": 1.0616813526654047e-05, + "loss": 0.6844, + "step": 4568 + }, + { + "epoch": 2.480456026058632, + "grad_norm": 9.570551013210757, + "learning_rate": 1.061330334560262e-05, + "loss": 0.5063, + "step": 4569 + }, + { + "epoch": 2.4809989142236697, + "grad_norm": 10.106289919073877, + "learning_rate": 1.0609793088696856e-05, + "loss": 0.6131, + "step": 4570 + }, + { + "epoch": 2.481541802388708, + "grad_norm": 10.842801269455231, + "learning_rate": 1.060628275637091e-05, + "loss": 0.6637, + "step": 4571 + }, + { + "epoch": 2.482084690553746, + "grad_norm": 11.953941591792438, + "learning_rate": 1.0602772349058952e-05, + "loss": 0.7631, + "step": 4572 + }, + { + "epoch": 2.482627578718784, + "grad_norm": 8.194683405468492, + "learning_rate": 1.0599261867195146e-05, + "loss": 0.5936, + "step": 4573 + }, + { + "epoch": 2.4831704668838217, + "grad_norm": 10.216659522745752, + "learning_rate": 1.0595751311213678e-05, + "loss": 0.7017, + "step": 4574 + }, + { + "epoch": 2.48371335504886, + "grad_norm": 11.0782985653065, + "learning_rate": 1.0592240681548738e-05, + "loss": 0.7285, + "step": 4575 + }, + { + "epoch": 2.484256243213898, + "grad_norm": 7.740854250685081, + "learning_rate": 1.0588729978634527e-05, + "loss": 0.4544, + "step": 4576 + }, + { + "epoch": 2.484799131378936, + "grad_norm": 6.9863990778515745, + "learning_rate": 1.0585219202905255e-05, + "loss": 0.2831, + "step": 4577 + }, + { + "epoch": 2.4853420195439737, + "grad_norm": 9.687327462597597, + "learning_rate": 1.0581708354795136e-05, + "loss": 0.5028, + "step": 4578 + }, + { + "epoch": 2.485884907709012, + "grad_norm": 8.009180785791076, + "learning_rate": 1.0578197434738405e-05, + "loss": 0.342, + "step": 4579 + }, + { + "epoch": 2.48642779587405, + "grad_norm": 8.101662777133065, + "learning_rate": 1.057468644316929e-05, + "loss": 0.4146, + "step": 4580 + }, + { + "epoch": 2.486970684039088, + "grad_norm": 9.370936142076715, + "learning_rate": 1.0571175380522041e-05, + "loss": 0.6006, + "step": 4581 + }, + { + "epoch": 2.4875135722041257, + "grad_norm": 11.164770907813724, + "learning_rate": 1.0567664247230911e-05, + "loss": 0.6787, + "step": 4582 + }, + { + "epoch": 2.488056460369164, + "grad_norm": 12.573930877103525, + "learning_rate": 1.056415304373016e-05, + "loss": 0.8051, + "step": 4583 + }, + { + "epoch": 2.488599348534202, + "grad_norm": 8.395060833285095, + "learning_rate": 1.0560641770454063e-05, + "loss": 0.5596, + "step": 4584 + }, + { + "epoch": 2.48914223669924, + "grad_norm": 11.015478890741145, + "learning_rate": 1.0557130427836899e-05, + "loss": 0.5442, + "step": 4585 + }, + { + "epoch": 2.4896851248642777, + "grad_norm": 9.420432906299153, + "learning_rate": 1.0553619016312955e-05, + "loss": 0.4917, + "step": 4586 + }, + { + "epoch": 2.490228013029316, + "grad_norm": 11.026358328990858, + "learning_rate": 1.0550107536316532e-05, + "loss": 0.4291, + "step": 4587 + }, + { + "epoch": 2.490770901194354, + "grad_norm": 16.545112067555284, + "learning_rate": 1.0546595988281928e-05, + "loss": 0.9417, + "step": 4588 + }, + { + "epoch": 2.491313789359392, + "grad_norm": 7.252757680637398, + "learning_rate": 1.0543084372643462e-05, + "loss": 0.7908, + "step": 4589 + }, + { + "epoch": 2.4918566775244297, + "grad_norm": 7.598562688636616, + "learning_rate": 1.0539572689835455e-05, + "loss": 0.3229, + "step": 4590 + }, + { + "epoch": 2.492399565689468, + "grad_norm": 12.449529704914275, + "learning_rate": 1.0536060940292238e-05, + "loss": 0.7373, + "step": 4591 + }, + { + "epoch": 2.492942453854506, + "grad_norm": 11.706159828307792, + "learning_rate": 1.0532549124448151e-05, + "loss": 0.6525, + "step": 4592 + }, + { + "epoch": 2.493485342019544, + "grad_norm": 8.000677691767947, + "learning_rate": 1.0529037242737538e-05, + "loss": 0.3613, + "step": 4593 + }, + { + "epoch": 2.4940282301845818, + "grad_norm": 9.529463393752437, + "learning_rate": 1.0525525295594755e-05, + "loss": 0.5921, + "step": 4594 + }, + { + "epoch": 2.49457111834962, + "grad_norm": 11.71769497271457, + "learning_rate": 1.0522013283454169e-05, + "loss": 0.5507, + "step": 4595 + }, + { + "epoch": 2.495114006514658, + "grad_norm": 8.546908524296436, + "learning_rate": 1.0518501206750145e-05, + "loss": 0.4564, + "step": 4596 + }, + { + "epoch": 2.495656894679696, + "grad_norm": 14.968617219210655, + "learning_rate": 1.0514989065917068e-05, + "loss": 0.5593, + "step": 4597 + }, + { + "epoch": 2.4961997828447338, + "grad_norm": 13.0682164313941, + "learning_rate": 1.051147686138932e-05, + "loss": 0.8812, + "step": 4598 + }, + { + "epoch": 2.496742671009772, + "grad_norm": 15.01014358585534, + "learning_rate": 1.05079645936013e-05, + "loss": 1.0756, + "step": 4599 + }, + { + "epoch": 2.49728555917481, + "grad_norm": 7.598240677419264, + "learning_rate": 1.0504452262987407e-05, + "loss": 0.3936, + "step": 4600 + }, + { + "epoch": 2.497828447339848, + "grad_norm": 10.39263281879832, + "learning_rate": 1.0500939869982058e-05, + "loss": 0.4944, + "step": 4601 + }, + { + "epoch": 2.4983713355048858, + "grad_norm": 12.104880157441473, + "learning_rate": 1.0497427415019665e-05, + "loss": 0.7498, + "step": 4602 + }, + { + "epoch": 2.498914223669924, + "grad_norm": 9.18522035093318, + "learning_rate": 1.0493914898534656e-05, + "loss": 0.6921, + "step": 4603 + }, + { + "epoch": 2.499457111834962, + "grad_norm": 15.812637613525052, + "learning_rate": 1.0490402320961466e-05, + "loss": 0.937, + "step": 4604 + }, + { + "epoch": 2.5, + "grad_norm": 11.400921058050775, + "learning_rate": 1.0486889682734535e-05, + "loss": 0.6453, + "step": 4605 + }, + { + "epoch": 2.500542888165038, + "grad_norm": 8.270129268701748, + "learning_rate": 1.0483376984288314e-05, + "loss": 0.3921, + "step": 4606 + }, + { + "epoch": 2.501085776330076, + "grad_norm": 15.017103371329522, + "learning_rate": 1.0479864226057258e-05, + "loss": 1.0397, + "step": 4607 + }, + { + "epoch": 2.5016286644951142, + "grad_norm": 8.346121650934089, + "learning_rate": 1.047635140847583e-05, + "loss": 0.3872, + "step": 4608 + }, + { + "epoch": 2.502171552660152, + "grad_norm": 14.17791420098929, + "learning_rate": 1.0472838531978495e-05, + "loss": 0.6932, + "step": 4609 + }, + { + "epoch": 2.50271444082519, + "grad_norm": 12.940166215438651, + "learning_rate": 1.046932559699974e-05, + "loss": 0.7045, + "step": 4610 + }, + { + "epoch": 2.503257328990228, + "grad_norm": 8.154813733233013, + "learning_rate": 1.0465812603974048e-05, + "loss": 0.4225, + "step": 4611 + }, + { + "epoch": 2.5038002171552662, + "grad_norm": 13.356750387494248, + "learning_rate": 1.0462299553335911e-05, + "loss": 0.7951, + "step": 4612 + }, + { + "epoch": 2.504343105320304, + "grad_norm": 8.683404141286427, + "learning_rate": 1.0458786445519827e-05, + "loss": 0.5871, + "step": 4613 + }, + { + "epoch": 2.504885993485342, + "grad_norm": 5.750758875120564, + "learning_rate": 1.0455273280960306e-05, + "loss": 0.3815, + "step": 4614 + }, + { + "epoch": 2.50542888165038, + "grad_norm": 9.328766968136149, + "learning_rate": 1.0451760060091857e-05, + "loss": 0.6451, + "step": 4615 + }, + { + "epoch": 2.5059717698154182, + "grad_norm": 9.704349389930364, + "learning_rate": 1.0448246783349002e-05, + "loss": 0.5136, + "step": 4616 + }, + { + "epoch": 2.506514657980456, + "grad_norm": 10.438938086465326, + "learning_rate": 1.0444733451166277e-05, + "loss": 0.6278, + "step": 4617 + }, + { + "epoch": 2.507057546145494, + "grad_norm": 11.019760442005925, + "learning_rate": 1.0441220063978208e-05, + "loss": 0.7477, + "step": 4618 + }, + { + "epoch": 2.507600434310532, + "grad_norm": 11.738067010881544, + "learning_rate": 1.0437706622219337e-05, + "loss": 0.5768, + "step": 4619 + }, + { + "epoch": 2.5081433224755703, + "grad_norm": 9.601242936349621, + "learning_rate": 1.0434193126324215e-05, + "loss": 0.4236, + "step": 4620 + }, + { + "epoch": 2.508686210640608, + "grad_norm": 8.041097216565952, + "learning_rate": 1.0430679576727392e-05, + "loss": 0.494, + "step": 4621 + }, + { + "epoch": 2.509229098805646, + "grad_norm": 11.144286090694623, + "learning_rate": 1.0427165973863434e-05, + "loss": 0.7882, + "step": 4622 + }, + { + "epoch": 2.509771986970684, + "grad_norm": 14.369854225910128, + "learning_rate": 1.0423652318166905e-05, + "loss": 0.7112, + "step": 4623 + }, + { + "epoch": 2.5103148751357223, + "grad_norm": 9.688103222900846, + "learning_rate": 1.0420138610072386e-05, + "loss": 0.4564, + "step": 4624 + }, + { + "epoch": 2.51085776330076, + "grad_norm": 8.2914031160658, + "learning_rate": 1.0416624850014454e-05, + "loss": 0.4106, + "step": 4625 + }, + { + "epoch": 2.511400651465798, + "grad_norm": 13.145683230031937, + "learning_rate": 1.0413111038427696e-05, + "loss": 0.905, + "step": 4626 + }, + { + "epoch": 2.511943539630836, + "grad_norm": 10.544391990984453, + "learning_rate": 1.0409597175746709e-05, + "loss": 0.5791, + "step": 4627 + }, + { + "epoch": 2.5124864277958743, + "grad_norm": 8.830789770145405, + "learning_rate": 1.0406083262406084e-05, + "loss": 0.4806, + "step": 4628 + }, + { + "epoch": 2.513029315960912, + "grad_norm": 9.186020814962909, + "learning_rate": 1.0402569298840437e-05, + "loss": 0.7227, + "step": 4629 + }, + { + "epoch": 2.51357220412595, + "grad_norm": 8.395511535015437, + "learning_rate": 1.0399055285484378e-05, + "loss": 0.5372, + "step": 4630 + }, + { + "epoch": 2.514115092290988, + "grad_norm": 9.969741133621088, + "learning_rate": 1.0395541222772524e-05, + "loss": 0.5151, + "step": 4631 + }, + { + "epoch": 2.5146579804560263, + "grad_norm": 11.64674593889149, + "learning_rate": 1.0392027111139505e-05, + "loss": 0.6159, + "step": 4632 + }, + { + "epoch": 2.515200868621064, + "grad_norm": 12.194306814356107, + "learning_rate": 1.0388512951019945e-05, + "loss": 0.6977, + "step": 4633 + }, + { + "epoch": 2.515743756786102, + "grad_norm": 9.063328798639105, + "learning_rate": 1.0384998742848485e-05, + "loss": 0.6518, + "step": 4634 + }, + { + "epoch": 2.51628664495114, + "grad_norm": 8.709419510130173, + "learning_rate": 1.0381484487059764e-05, + "loss": 0.5571, + "step": 4635 + }, + { + "epoch": 2.5168295331161783, + "grad_norm": 10.875178380627338, + "learning_rate": 1.0377970184088438e-05, + "loss": 0.8384, + "step": 4636 + }, + { + "epoch": 2.517372421281216, + "grad_norm": 8.42478873312803, + "learning_rate": 1.0374455834369155e-05, + "loss": 0.6209, + "step": 4637 + }, + { + "epoch": 2.517915309446254, + "grad_norm": 13.875461781038107, + "learning_rate": 1.0370941438336577e-05, + "loss": 0.8787, + "step": 4638 + }, + { + "epoch": 2.518458197611292, + "grad_norm": 11.372446219871874, + "learning_rate": 1.036742699642537e-05, + "loss": 0.569, + "step": 4639 + }, + { + "epoch": 2.5190010857763303, + "grad_norm": 10.725098253853016, + "learning_rate": 1.036391250907021e-05, + "loss": 0.8431, + "step": 4640 + }, + { + "epoch": 2.519543973941368, + "grad_norm": 8.268642654673096, + "learning_rate": 1.0360397976705767e-05, + "loss": 0.5833, + "step": 4641 + }, + { + "epoch": 2.520086862106406, + "grad_norm": 10.560802722537995, + "learning_rate": 1.0356883399766731e-05, + "loss": 0.7495, + "step": 4642 + }, + { + "epoch": 2.520629750271444, + "grad_norm": 11.789754393456334, + "learning_rate": 1.0353368778687784e-05, + "loss": 0.644, + "step": 4643 + }, + { + "epoch": 2.5211726384364823, + "grad_norm": 10.236138559776574, + "learning_rate": 1.0349854113903625e-05, + "loss": 0.6544, + "step": 4644 + }, + { + "epoch": 2.52171552660152, + "grad_norm": 8.552964039446971, + "learning_rate": 1.0346339405848953e-05, + "loss": 0.5093, + "step": 4645 + }, + { + "epoch": 2.522258414766558, + "grad_norm": 8.301860195209034, + "learning_rate": 1.0342824654958471e-05, + "loss": 0.6285, + "step": 4646 + }, + { + "epoch": 2.522801302931596, + "grad_norm": 9.238558119037732, + "learning_rate": 1.0339309861666889e-05, + "loss": 0.6519, + "step": 4647 + }, + { + "epoch": 2.5233441910966343, + "grad_norm": 9.840812219132841, + "learning_rate": 1.0335795026408922e-05, + "loss": 0.5874, + "step": 4648 + }, + { + "epoch": 2.523887079261672, + "grad_norm": 8.108461441633077, + "learning_rate": 1.0332280149619291e-05, + "loss": 0.5793, + "step": 4649 + }, + { + "epoch": 2.52442996742671, + "grad_norm": 9.446694020377457, + "learning_rate": 1.032876523173272e-05, + "loss": 0.607, + "step": 4650 + }, + { + "epoch": 2.524972855591748, + "grad_norm": 11.88681485670853, + "learning_rate": 1.0325250273183944e-05, + "loss": 1.0745, + "step": 4651 + }, + { + "epoch": 2.5255157437567863, + "grad_norm": 8.347193432264572, + "learning_rate": 1.0321735274407694e-05, + "loss": 0.8431, + "step": 4652 + }, + { + "epoch": 2.526058631921824, + "grad_norm": 9.639244837832804, + "learning_rate": 1.0318220235838715e-05, + "loss": 0.7827, + "step": 4653 + }, + { + "epoch": 2.526601520086862, + "grad_norm": 12.239750504764533, + "learning_rate": 1.0314705157911744e-05, + "loss": 0.6735, + "step": 4654 + }, + { + "epoch": 2.5271444082519, + "grad_norm": 15.826940901953726, + "learning_rate": 1.0311190041061543e-05, + "loss": 0.9182, + "step": 4655 + }, + { + "epoch": 2.5276872964169383, + "grad_norm": 11.015508331511823, + "learning_rate": 1.0307674885722863e-05, + "loss": 0.7834, + "step": 4656 + }, + { + "epoch": 2.528230184581976, + "grad_norm": 11.163753697057714, + "learning_rate": 1.030415969233046e-05, + "loss": 0.6613, + "step": 4657 + }, + { + "epoch": 2.528773072747014, + "grad_norm": 8.747908439433298, + "learning_rate": 1.03006444613191e-05, + "loss": 0.6974, + "step": 4658 + }, + { + "epoch": 2.529315960912052, + "grad_norm": 8.370456762868022, + "learning_rate": 1.0297129193123555e-05, + "loss": 0.5343, + "step": 4659 + }, + { + "epoch": 2.5298588490770904, + "grad_norm": 13.46061984772866, + "learning_rate": 1.0293613888178597e-05, + "loss": 1.3054, + "step": 4660 + }, + { + "epoch": 2.530401737242128, + "grad_norm": 10.948007938187232, + "learning_rate": 1.0290098546919007e-05, + "loss": 0.7028, + "step": 4661 + }, + { + "epoch": 2.530944625407166, + "grad_norm": 8.833095994203507, + "learning_rate": 1.0286583169779568e-05, + "loss": 0.7663, + "step": 4662 + }, + { + "epoch": 2.531487513572204, + "grad_norm": 10.283493007620729, + "learning_rate": 1.0283067757195062e-05, + "loss": 0.7804, + "step": 4663 + }, + { + "epoch": 2.5320304017372424, + "grad_norm": 13.55508996688301, + "learning_rate": 1.0279552309600288e-05, + "loss": 0.656, + "step": 4664 + }, + { + "epoch": 2.53257328990228, + "grad_norm": 8.738397704085456, + "learning_rate": 1.0276036827430036e-05, + "loss": 0.4902, + "step": 4665 + }, + { + "epoch": 2.533116178067318, + "grad_norm": 10.260011622402263, + "learning_rate": 1.0272521311119113e-05, + "loss": 0.4798, + "step": 4666 + }, + { + "epoch": 2.533659066232356, + "grad_norm": 10.743076748133591, + "learning_rate": 1.0269005761102315e-05, + "loss": 0.9671, + "step": 4667 + }, + { + "epoch": 2.5342019543973944, + "grad_norm": 11.846945314251228, + "learning_rate": 1.0265490177814458e-05, + "loss": 0.7168, + "step": 4668 + }, + { + "epoch": 2.534744842562432, + "grad_norm": 9.129703791341635, + "learning_rate": 1.0261974561690356e-05, + "loss": 0.6008, + "step": 4669 + }, + { + "epoch": 2.53528773072747, + "grad_norm": 8.464074733631888, + "learning_rate": 1.025845891316482e-05, + "loss": 0.4327, + "step": 4670 + }, + { + "epoch": 2.535830618892508, + "grad_norm": 10.894552922917367, + "learning_rate": 1.0254943232672671e-05, + "loss": 0.5627, + "step": 4671 + }, + { + "epoch": 2.5363735070575464, + "grad_norm": 12.365454725760225, + "learning_rate": 1.025142752064874e-05, + "loss": 0.606, + "step": 4672 + }, + { + "epoch": 2.536916395222584, + "grad_norm": 10.066952200044973, + "learning_rate": 1.0247911777527852e-05, + "loss": 0.7503, + "step": 4673 + }, + { + "epoch": 2.537459283387622, + "grad_norm": 8.386860314534095, + "learning_rate": 1.0244396003744836e-05, + "loss": 0.7663, + "step": 4674 + }, + { + "epoch": 2.53800217155266, + "grad_norm": 9.309667750090998, + "learning_rate": 1.024088019973454e-05, + "loss": 0.5658, + "step": 4675 + }, + { + "epoch": 2.5385450597176984, + "grad_norm": 13.530278392237364, + "learning_rate": 1.0237364365931794e-05, + "loss": 0.799, + "step": 4676 + }, + { + "epoch": 2.539087947882736, + "grad_norm": 11.344034257453236, + "learning_rate": 1.0233848502771444e-05, + "loss": 0.5931, + "step": 4677 + }, + { + "epoch": 2.539630836047774, + "grad_norm": 9.26197757297852, + "learning_rate": 1.0230332610688338e-05, + "loss": 0.5419, + "step": 4678 + }, + { + "epoch": 2.540173724212812, + "grad_norm": 12.620486559012823, + "learning_rate": 1.022681669011733e-05, + "loss": 0.6746, + "step": 4679 + }, + { + "epoch": 2.5407166123778504, + "grad_norm": 13.640092752737832, + "learning_rate": 1.0223300741493268e-05, + "loss": 0.8806, + "step": 4680 + }, + { + "epoch": 2.541259500542888, + "grad_norm": 10.435820181571753, + "learning_rate": 1.0219784765251018e-05, + "loss": 0.6876, + "step": 4681 + }, + { + "epoch": 2.541802388707926, + "grad_norm": 11.624678093656978, + "learning_rate": 1.0216268761825436e-05, + "loss": 0.7377, + "step": 4682 + }, + { + "epoch": 2.542345276872964, + "grad_norm": 10.361224010713974, + "learning_rate": 1.021275273165139e-05, + "loss": 0.5911, + "step": 4683 + }, + { + "epoch": 2.5428881650380024, + "grad_norm": 6.869744257762391, + "learning_rate": 1.0209236675163747e-05, + "loss": 0.4204, + "step": 4684 + }, + { + "epoch": 2.54343105320304, + "grad_norm": 16.790885412103638, + "learning_rate": 1.020572059279738e-05, + "loss": 0.8837, + "step": 4685 + }, + { + "epoch": 2.543973941368078, + "grad_norm": 9.226978433564653, + "learning_rate": 1.0202204484987159e-05, + "loss": 0.4686, + "step": 4686 + }, + { + "epoch": 2.544516829533116, + "grad_norm": 7.391909229375343, + "learning_rate": 1.019868835216797e-05, + "loss": 0.6996, + "step": 4687 + }, + { + "epoch": 2.5450597176981544, + "grad_norm": 9.103089555285077, + "learning_rate": 1.0195172194774684e-05, + "loss": 0.7243, + "step": 4688 + }, + { + "epoch": 2.545602605863192, + "grad_norm": 7.27509790187114, + "learning_rate": 1.0191656013242192e-05, + "loss": 0.4612, + "step": 4689 + }, + { + "epoch": 2.54614549402823, + "grad_norm": 6.939999239823949, + "learning_rate": 1.0188139808005379e-05, + "loss": 0.3483, + "step": 4690 + }, + { + "epoch": 2.546688382193268, + "grad_norm": 9.036119044447904, + "learning_rate": 1.0184623579499133e-05, + "loss": 0.5822, + "step": 4691 + }, + { + "epoch": 2.5472312703583064, + "grad_norm": 15.325233133671158, + "learning_rate": 1.018110732815835e-05, + "loss": 0.9375, + "step": 4692 + }, + { + "epoch": 2.547774158523344, + "grad_norm": 13.335480222530155, + "learning_rate": 1.017759105441792e-05, + "loss": 0.8165, + "step": 4693 + }, + { + "epoch": 2.548317046688382, + "grad_norm": 10.097959894991517, + "learning_rate": 1.0174074758712751e-05, + "loss": 0.7773, + "step": 4694 + }, + { + "epoch": 2.54885993485342, + "grad_norm": 9.7484700487623, + "learning_rate": 1.0170558441477738e-05, + "loss": 0.7805, + "step": 4695 + }, + { + "epoch": 2.5494028230184584, + "grad_norm": 9.687849482731272, + "learning_rate": 1.0167042103147784e-05, + "loss": 0.6675, + "step": 4696 + }, + { + "epoch": 2.549945711183496, + "grad_norm": 12.578144763595915, + "learning_rate": 1.0163525744157796e-05, + "loss": 0.7757, + "step": 4697 + }, + { + "epoch": 2.550488599348534, + "grad_norm": 9.03070780679602, + "learning_rate": 1.0160009364942684e-05, + "loss": 0.5321, + "step": 4698 + }, + { + "epoch": 2.5510314875135722, + "grad_norm": 10.632448946778698, + "learning_rate": 1.0156492965937357e-05, + "loss": 0.5589, + "step": 4699 + }, + { + "epoch": 2.5515743756786105, + "grad_norm": 10.053170332804227, + "learning_rate": 1.015297654757673e-05, + "loss": 0.4974, + "step": 4700 + }, + { + "epoch": 2.5521172638436482, + "grad_norm": 9.833516973903698, + "learning_rate": 1.0149460110295722e-05, + "loss": 0.6022, + "step": 4701 + }, + { + "epoch": 2.552660152008686, + "grad_norm": 16.704821665027783, + "learning_rate": 1.0145943654529248e-05, + "loss": 0.776, + "step": 4702 + }, + { + "epoch": 2.5532030401737242, + "grad_norm": 12.589948540180313, + "learning_rate": 1.0142427180712231e-05, + "loss": 0.8549, + "step": 4703 + }, + { + "epoch": 2.5537459283387625, + "grad_norm": 10.414828597120934, + "learning_rate": 1.0138910689279596e-05, + "loss": 0.7209, + "step": 4704 + }, + { + "epoch": 2.5542888165038002, + "grad_norm": 10.158714848928406, + "learning_rate": 1.0135394180666261e-05, + "loss": 0.6835, + "step": 4705 + }, + { + "epoch": 2.554831704668838, + "grad_norm": 8.255885711809917, + "learning_rate": 1.013187765530716e-05, + "loss": 0.4542, + "step": 4706 + }, + { + "epoch": 2.5553745928338762, + "grad_norm": 10.386471672795576, + "learning_rate": 1.012836111363722e-05, + "loss": 0.9478, + "step": 4707 + }, + { + "epoch": 2.5559174809989145, + "grad_norm": 11.828466617801858, + "learning_rate": 1.0124844556091374e-05, + "loss": 0.7975, + "step": 4708 + }, + { + "epoch": 2.5564603691639523, + "grad_norm": 10.742130491540307, + "learning_rate": 1.0121327983104555e-05, + "loss": 0.9712, + "step": 4709 + }, + { + "epoch": 2.55700325732899, + "grad_norm": 13.314461670532042, + "learning_rate": 1.0117811395111695e-05, + "loss": 0.9351, + "step": 4710 + }, + { + "epoch": 2.5575461454940283, + "grad_norm": 10.15371491897671, + "learning_rate": 1.0114294792547733e-05, + "loss": 0.6056, + "step": 4711 + }, + { + "epoch": 2.5580890336590665, + "grad_norm": 8.317356068473407, + "learning_rate": 1.011077817584761e-05, + "loss": 0.6411, + "step": 4712 + }, + { + "epoch": 2.5586319218241043, + "grad_norm": 11.215419343504372, + "learning_rate": 1.0107261545446267e-05, + "loss": 0.5411, + "step": 4713 + }, + { + "epoch": 2.559174809989142, + "grad_norm": 10.343171151673559, + "learning_rate": 1.0103744901778645e-05, + "loss": 0.5395, + "step": 4714 + }, + { + "epoch": 2.5597176981541803, + "grad_norm": 14.089833480392985, + "learning_rate": 1.0100228245279688e-05, + "loss": 0.9855, + "step": 4715 + }, + { + "epoch": 2.5602605863192185, + "grad_norm": 14.836069456421422, + "learning_rate": 1.0096711576384344e-05, + "loss": 0.8219, + "step": 4716 + }, + { + "epoch": 2.5608034744842563, + "grad_norm": 16.933747543735457, + "learning_rate": 1.0093194895527556e-05, + "loss": 0.9341, + "step": 4717 + }, + { + "epoch": 2.561346362649294, + "grad_norm": 13.574017237717102, + "learning_rate": 1.008967820314427e-05, + "loss": 1.1551, + "step": 4718 + }, + { + "epoch": 2.5618892508143323, + "grad_norm": 8.877246747020687, + "learning_rate": 1.0086161499669447e-05, + "loss": 0.6341, + "step": 4719 + }, + { + "epoch": 2.5624321389793705, + "grad_norm": 9.391996392121968, + "learning_rate": 1.0082644785538031e-05, + "loss": 0.5285, + "step": 4720 + }, + { + "epoch": 2.5629750271444083, + "grad_norm": 10.876272607307573, + "learning_rate": 1.0079128061184977e-05, + "loss": 0.5678, + "step": 4721 + }, + { + "epoch": 2.563517915309446, + "grad_norm": 12.516236640792583, + "learning_rate": 1.0075611327045236e-05, + "loss": 0.8382, + "step": 4722 + }, + { + "epoch": 2.5640608034744843, + "grad_norm": 6.395711230996481, + "learning_rate": 1.0072094583553769e-05, + "loss": 0.3785, + "step": 4723 + }, + { + "epoch": 2.5646036916395225, + "grad_norm": 8.671727273867468, + "learning_rate": 1.0068577831145526e-05, + "loss": 0.5039, + "step": 4724 + }, + { + "epoch": 2.5651465798045603, + "grad_norm": 12.77911107984442, + "learning_rate": 1.0065061070255469e-05, + "loss": 0.7647, + "step": 4725 + }, + { + "epoch": 2.565689467969598, + "grad_norm": 9.108473588917152, + "learning_rate": 1.0061544301318556e-05, + "loss": 0.798, + "step": 4726 + }, + { + "epoch": 2.5662323561346363, + "grad_norm": 8.776410877174934, + "learning_rate": 1.0058027524769742e-05, + "loss": 0.6388, + "step": 4727 + }, + { + "epoch": 2.5667752442996745, + "grad_norm": 10.392346204038525, + "learning_rate": 1.0054510741043995e-05, + "loss": 0.8178, + "step": 4728 + }, + { + "epoch": 2.5673181324647123, + "grad_norm": 8.611264791773579, + "learning_rate": 1.005099395057627e-05, + "loss": 0.5022, + "step": 4729 + }, + { + "epoch": 2.56786102062975, + "grad_norm": 8.228466341722912, + "learning_rate": 1.0047477153801534e-05, + "loss": 0.6572, + "step": 4730 + }, + { + "epoch": 2.5684039087947883, + "grad_norm": 8.410912812963819, + "learning_rate": 1.0043960351154745e-05, + "loss": 0.4526, + "step": 4731 + }, + { + "epoch": 2.5689467969598265, + "grad_norm": 8.24917306612946, + "learning_rate": 1.0040443543070872e-05, + "loss": 0.5492, + "step": 4732 + }, + { + "epoch": 2.5694896851248643, + "grad_norm": 7.578877723490418, + "learning_rate": 1.0036926729984878e-05, + "loss": 0.5482, + "step": 4733 + }, + { + "epoch": 2.570032573289902, + "grad_norm": 8.464719447395414, + "learning_rate": 1.003340991233173e-05, + "loss": 0.5236, + "step": 4734 + }, + { + "epoch": 2.5705754614549403, + "grad_norm": 10.297009632330814, + "learning_rate": 1.0029893090546385e-05, + "loss": 0.6747, + "step": 4735 + }, + { + "epoch": 2.5711183496199785, + "grad_norm": 9.895183624345862, + "learning_rate": 1.0026376265063818e-05, + "loss": 0.7036, + "step": 4736 + }, + { + "epoch": 2.5716612377850163, + "grad_norm": 16.034084078355107, + "learning_rate": 1.0022859436318988e-05, + "loss": 0.8583, + "step": 4737 + }, + { + "epoch": 2.572204125950054, + "grad_norm": 10.881395091863821, + "learning_rate": 1.0019342604746871e-05, + "loss": 0.7214, + "step": 4738 + }, + { + "epoch": 2.5727470141150923, + "grad_norm": 9.051906164481618, + "learning_rate": 1.0015825770782428e-05, + "loss": 0.36, + "step": 4739 + }, + { + "epoch": 2.5732899022801305, + "grad_norm": 11.71081741433651, + "learning_rate": 1.001230893486063e-05, + "loss": 0.8467, + "step": 4740 + }, + { + "epoch": 2.5738327904451683, + "grad_norm": 8.4727788666991, + "learning_rate": 1.0008792097416442e-05, + "loss": 0.5312, + "step": 4741 + }, + { + "epoch": 2.574375678610206, + "grad_norm": 10.308791039753308, + "learning_rate": 1.0005275258884835e-05, + "loss": 0.7132, + "step": 4742 + }, + { + "epoch": 2.5749185667752443, + "grad_norm": 13.591326014197278, + "learning_rate": 1.0001758419700773e-05, + "loss": 0.7087, + "step": 4743 + }, + { + "epoch": 2.5754614549402826, + "grad_norm": 9.610601406060551, + "learning_rate": 9.998241580299229e-06, + "loss": 0.6708, + "step": 4744 + }, + { + "epoch": 2.5760043431053203, + "grad_norm": 10.04740706892737, + "learning_rate": 9.994724741115169e-06, + "loss": 0.6644, + "step": 4745 + }, + { + "epoch": 2.576547231270358, + "grad_norm": 7.760395046863199, + "learning_rate": 9.991207902583558e-06, + "loss": 0.4361, + "step": 4746 + }, + { + "epoch": 2.5770901194353963, + "grad_norm": 9.097495186854097, + "learning_rate": 9.987691065139373e-06, + "loss": 0.5613, + "step": 4747 + }, + { + "epoch": 2.5776330076004346, + "grad_norm": 9.584167359092984, + "learning_rate": 9.984174229217572e-06, + "loss": 0.6408, + "step": 4748 + }, + { + "epoch": 2.5781758957654723, + "grad_norm": 10.329246788617224, + "learning_rate": 9.980657395253132e-06, + "loss": 0.6776, + "step": 4749 + }, + { + "epoch": 2.57871878393051, + "grad_norm": 10.000264678013584, + "learning_rate": 9.977140563681015e-06, + "loss": 0.7489, + "step": 4750 + }, + { + "epoch": 2.5792616720955484, + "grad_norm": 6.80240569067885, + "learning_rate": 9.973623734936185e-06, + "loss": 0.4639, + "step": 4751 + }, + { + "epoch": 2.5798045602605866, + "grad_norm": 7.578175941165148, + "learning_rate": 9.97010690945362e-06, + "loss": 0.38, + "step": 4752 + }, + { + "epoch": 2.5803474484256244, + "grad_norm": 12.48681779678407, + "learning_rate": 9.966590087668274e-06, + "loss": 0.8046, + "step": 4753 + }, + { + "epoch": 2.580890336590662, + "grad_norm": 13.31142746407252, + "learning_rate": 9.963073270015126e-06, + "loss": 0.7671, + "step": 4754 + }, + { + "epoch": 2.5814332247557004, + "grad_norm": 8.987640426318515, + "learning_rate": 9.959556456929128e-06, + "loss": 0.4691, + "step": 4755 + }, + { + "epoch": 2.5819761129207386, + "grad_norm": 9.551368066054666, + "learning_rate": 9.956039648845257e-06, + "loss": 0.6207, + "step": 4756 + }, + { + "epoch": 2.5825190010857764, + "grad_norm": 17.060876787385865, + "learning_rate": 9.952522846198468e-06, + "loss": 0.9276, + "step": 4757 + }, + { + "epoch": 2.583061889250814, + "grad_norm": 9.760247190255836, + "learning_rate": 9.949006049423731e-06, + "loss": 0.539, + "step": 4758 + }, + { + "epoch": 2.5836047774158524, + "grad_norm": 14.929770896239804, + "learning_rate": 9.94548925895601e-06, + "loss": 1.0896, + "step": 4759 + }, + { + "epoch": 2.5841476655808906, + "grad_norm": 14.902014520576888, + "learning_rate": 9.94197247523026e-06, + "loss": 1.0207, + "step": 4760 + }, + { + "epoch": 2.5846905537459284, + "grad_norm": 9.630889650448959, + "learning_rate": 9.93845569868145e-06, + "loss": 0.6322, + "step": 4761 + }, + { + "epoch": 2.585233441910966, + "grad_norm": 6.534627620002312, + "learning_rate": 9.934938929744535e-06, + "loss": 0.2627, + "step": 4762 + }, + { + "epoch": 2.5857763300760044, + "grad_norm": 10.701147725996508, + "learning_rate": 9.931422168854476e-06, + "loss": 0.6789, + "step": 4763 + }, + { + "epoch": 2.5863192182410426, + "grad_norm": 9.539331065006966, + "learning_rate": 9.927905416446233e-06, + "loss": 0.7109, + "step": 4764 + }, + { + "epoch": 2.5868621064060804, + "grad_norm": 9.979058002114003, + "learning_rate": 9.924388672954766e-06, + "loss": 0.6497, + "step": 4765 + }, + { + "epoch": 2.587404994571118, + "grad_norm": 9.638452627150874, + "learning_rate": 9.920871938815024e-06, + "loss": 0.6695, + "step": 4766 + }, + { + "epoch": 2.5879478827361564, + "grad_norm": 12.326422327522844, + "learning_rate": 9.91735521446197e-06, + "loss": 0.6354, + "step": 4767 + }, + { + "epoch": 2.5884907709011946, + "grad_norm": 8.571541287165285, + "learning_rate": 9.913838500330553e-06, + "loss": 0.3939, + "step": 4768 + }, + { + "epoch": 2.5890336590662324, + "grad_norm": 9.709310867338898, + "learning_rate": 9.910321796855732e-06, + "loss": 0.5866, + "step": 4769 + }, + { + "epoch": 2.58957654723127, + "grad_norm": 8.48642736575235, + "learning_rate": 9.90680510447245e-06, + "loss": 0.8167, + "step": 4770 + }, + { + "epoch": 2.5901194353963084, + "grad_norm": 9.153780615106163, + "learning_rate": 9.90328842361566e-06, + "loss": 0.6551, + "step": 4771 + }, + { + "epoch": 2.5906623235613466, + "grad_norm": 9.331431330636146, + "learning_rate": 9.899771754720315e-06, + "loss": 0.598, + "step": 4772 + }, + { + "epoch": 2.5912052117263844, + "grad_norm": 10.71941498109364, + "learning_rate": 9.896255098221357e-06, + "loss": 0.6996, + "step": 4773 + }, + { + "epoch": 2.591748099891422, + "grad_norm": 11.415708336876241, + "learning_rate": 9.892738454553736e-06, + "loss": 0.8247, + "step": 4774 + }, + { + "epoch": 2.5922909880564604, + "grad_norm": 9.526809613609492, + "learning_rate": 9.889221824152391e-06, + "loss": 0.5088, + "step": 4775 + }, + { + "epoch": 2.5928338762214986, + "grad_norm": 8.911159845652058, + "learning_rate": 9.885705207452268e-06, + "loss": 0.3377, + "step": 4776 + }, + { + "epoch": 2.5933767643865364, + "grad_norm": 13.451504414556261, + "learning_rate": 9.882188604888307e-06, + "loss": 0.9247, + "step": 4777 + }, + { + "epoch": 2.593919652551574, + "grad_norm": 10.251509263282227, + "learning_rate": 9.87867201689545e-06, + "loss": 0.4384, + "step": 4778 + }, + { + "epoch": 2.5944625407166124, + "grad_norm": 9.955424126921221, + "learning_rate": 9.875155443908631e-06, + "loss": 0.6318, + "step": 4779 + }, + { + "epoch": 2.5950054288816506, + "grad_norm": 8.584868222198555, + "learning_rate": 9.871638886362782e-06, + "loss": 0.5148, + "step": 4780 + }, + { + "epoch": 2.5955483170466884, + "grad_norm": 16.858447088677025, + "learning_rate": 9.868122344692846e-06, + "loss": 1.2564, + "step": 4781 + }, + { + "epoch": 2.596091205211726, + "grad_norm": 9.7339544501354, + "learning_rate": 9.86460581933374e-06, + "loss": 0.7158, + "step": 4782 + }, + { + "epoch": 2.5966340933767644, + "grad_norm": 11.03551156903236, + "learning_rate": 9.861089310720409e-06, + "loss": 0.5136, + "step": 4783 + }, + { + "epoch": 2.5971769815418027, + "grad_norm": 11.173351784755749, + "learning_rate": 9.857572819287768e-06, + "loss": 0.6439, + "step": 4784 + }, + { + "epoch": 2.5977198697068404, + "grad_norm": 14.318584084756125, + "learning_rate": 9.854056345470754e-06, + "loss": 1.2827, + "step": 4785 + }, + { + "epoch": 2.598262757871878, + "grad_norm": 11.132635136137834, + "learning_rate": 9.850539889704278e-06, + "loss": 0.499, + "step": 4786 + }, + { + "epoch": 2.5988056460369164, + "grad_norm": 11.710993594625656, + "learning_rate": 9.847023452423272e-06, + "loss": 0.466, + "step": 4787 + }, + { + "epoch": 2.5993485342019547, + "grad_norm": 12.05431003500401, + "learning_rate": 9.843507034062646e-06, + "loss": 1.002, + "step": 4788 + }, + { + "epoch": 2.5998914223669924, + "grad_norm": 12.651344078635795, + "learning_rate": 9.83999063505732e-06, + "loss": 0.6906, + "step": 4789 + }, + { + "epoch": 2.6004343105320302, + "grad_norm": 11.592009871908612, + "learning_rate": 9.83647425584221e-06, + "loss": 0.908, + "step": 4790 + }, + { + "epoch": 2.6009771986970684, + "grad_norm": 11.164823794029719, + "learning_rate": 9.83295789685222e-06, + "loss": 1.1006, + "step": 4791 + }, + { + "epoch": 2.6015200868621067, + "grad_norm": 18.55453736068596, + "learning_rate": 9.829441558522267e-06, + "loss": 0.6774, + "step": 4792 + }, + { + "epoch": 2.6020629750271445, + "grad_norm": 9.004957503789628, + "learning_rate": 9.825925241287249e-06, + "loss": 0.612, + "step": 4793 + }, + { + "epoch": 2.6026058631921822, + "grad_norm": 15.331324393349687, + "learning_rate": 9.822408945582081e-06, + "loss": 0.588, + "step": 4794 + }, + { + "epoch": 2.6031487513572205, + "grad_norm": 11.30698173094918, + "learning_rate": 9.818892671841653e-06, + "loss": 0.8413, + "step": 4795 + }, + { + "epoch": 2.6036916395222587, + "grad_norm": 10.970237049386885, + "learning_rate": 9.815376420500869e-06, + "loss": 0.5687, + "step": 4796 + }, + { + "epoch": 2.6042345276872965, + "grad_norm": 12.062898313733237, + "learning_rate": 9.811860191994623e-06, + "loss": 0.88, + "step": 4797 + }, + { + "epoch": 2.6047774158523342, + "grad_norm": 13.760783812579094, + "learning_rate": 9.808343986757811e-06, + "loss": 0.7319, + "step": 4798 + }, + { + "epoch": 2.6053203040173725, + "grad_norm": 10.142158775926303, + "learning_rate": 9.80482780522532e-06, + "loss": 0.5404, + "step": 4799 + }, + { + "epoch": 2.6058631921824107, + "grad_norm": 14.235611160734313, + "learning_rate": 9.801311647832035e-06, + "loss": 0.6849, + "step": 4800 + }, + { + "epoch": 2.6064060803474485, + "grad_norm": 14.33484497964673, + "learning_rate": 9.797795515012843e-06, + "loss": 1.0677, + "step": 4801 + }, + { + "epoch": 2.6069489685124863, + "grad_norm": 12.238534475065178, + "learning_rate": 9.794279407202624e-06, + "loss": 0.8279, + "step": 4802 + }, + { + "epoch": 2.6074918566775245, + "grad_norm": 13.51950414574984, + "learning_rate": 9.790763324836256e-06, + "loss": 0.8372, + "step": 4803 + }, + { + "epoch": 2.6080347448425627, + "grad_norm": 9.360760584245645, + "learning_rate": 9.78724726834861e-06, + "loss": 0.666, + "step": 4804 + }, + { + "epoch": 2.6085776330076005, + "grad_norm": 10.793352407977682, + "learning_rate": 9.783731238174566e-06, + "loss": 0.8308, + "step": 4805 + }, + { + "epoch": 2.6091205211726383, + "grad_norm": 13.862491939748782, + "learning_rate": 9.780215234748984e-06, + "loss": 0.8018, + "step": 4806 + }, + { + "epoch": 2.6096634093376765, + "grad_norm": 8.019652931555422, + "learning_rate": 9.776699258506734e-06, + "loss": 0.3801, + "step": 4807 + }, + { + "epoch": 2.6102062975027147, + "grad_norm": 10.161487802712038, + "learning_rate": 9.773183309882675e-06, + "loss": 0.6723, + "step": 4808 + }, + { + "epoch": 2.6107491856677525, + "grad_norm": 9.64703000959591, + "learning_rate": 9.769667389311664e-06, + "loss": 0.6294, + "step": 4809 + }, + { + "epoch": 2.6112920738327903, + "grad_norm": 9.792927662736604, + "learning_rate": 9.766151497228561e-06, + "loss": 0.4596, + "step": 4810 + }, + { + "epoch": 2.6118349619978285, + "grad_norm": 10.336511067045754, + "learning_rate": 9.76263563406821e-06, + "loss": 0.6494, + "step": 4811 + }, + { + "epoch": 2.6123778501628667, + "grad_norm": 12.121823172890695, + "learning_rate": 9.759119800265464e-06, + "loss": 1.0394, + "step": 4812 + }, + { + "epoch": 2.6129207383279045, + "grad_norm": 10.891668605270631, + "learning_rate": 9.755603996255164e-06, + "loss": 0.6212, + "step": 4813 + }, + { + "epoch": 2.6134636264929423, + "grad_norm": 9.599793914114507, + "learning_rate": 9.752088222472151e-06, + "loss": 0.6182, + "step": 4814 + }, + { + "epoch": 2.6140065146579805, + "grad_norm": 10.448317247236231, + "learning_rate": 9.748572479351263e-06, + "loss": 0.6779, + "step": 4815 + }, + { + "epoch": 2.6145494028230187, + "grad_norm": 12.703205416876392, + "learning_rate": 9.74505676732733e-06, + "loss": 0.8093, + "step": 4816 + }, + { + "epoch": 2.6150922909880565, + "grad_norm": 10.14294888595903, + "learning_rate": 9.741541086835182e-06, + "loss": 0.6493, + "step": 4817 + }, + { + "epoch": 2.6156351791530943, + "grad_norm": 11.455318908063175, + "learning_rate": 9.738025438309649e-06, + "loss": 0.9375, + "step": 4818 + }, + { + "epoch": 2.6161780673181325, + "grad_norm": 8.90202557317791, + "learning_rate": 9.734509822185545e-06, + "loss": 0.5978, + "step": 4819 + }, + { + "epoch": 2.6167209554831707, + "grad_norm": 10.38484629007874, + "learning_rate": 9.730994238897687e-06, + "loss": 0.9772, + "step": 4820 + }, + { + "epoch": 2.6172638436482085, + "grad_norm": 10.650781833239714, + "learning_rate": 9.72747868888089e-06, + "loss": 0.7354, + "step": 4821 + }, + { + "epoch": 2.6178067318132463, + "grad_norm": 11.462972783418381, + "learning_rate": 9.723963172569964e-06, + "loss": 0.8748, + "step": 4822 + }, + { + "epoch": 2.6183496199782845, + "grad_norm": 7.4804704893307505, + "learning_rate": 9.720447690399716e-06, + "loss": 0.4853, + "step": 4823 + }, + { + "epoch": 2.6188925081433228, + "grad_norm": 8.032622743372611, + "learning_rate": 9.716932242804938e-06, + "loss": 0.6484, + "step": 4824 + }, + { + "epoch": 2.6194353963083605, + "grad_norm": 10.141962204854094, + "learning_rate": 9.713416830220436e-06, + "loss": 0.669, + "step": 4825 + }, + { + "epoch": 2.6199782844733983, + "grad_norm": 11.48618310533414, + "learning_rate": 9.709901453080993e-06, + "loss": 0.6279, + "step": 4826 + }, + { + "epoch": 2.6205211726384365, + "grad_norm": 9.789449860925128, + "learning_rate": 9.706386111821406e-06, + "loss": 0.6072, + "step": 4827 + }, + { + "epoch": 2.6210640608034748, + "grad_norm": 8.10222084311329, + "learning_rate": 9.702870806876448e-06, + "loss": 0.6476, + "step": 4828 + }, + { + "epoch": 2.6216069489685125, + "grad_norm": 10.30951278757575, + "learning_rate": 9.699355538680902e-06, + "loss": 0.4691, + "step": 4829 + }, + { + "epoch": 2.6221498371335503, + "grad_norm": 10.960835467770405, + "learning_rate": 9.695840307669547e-06, + "loss": 0.7962, + "step": 4830 + }, + { + "epoch": 2.6226927252985885, + "grad_norm": 6.691408537531251, + "learning_rate": 9.69232511427714e-06, + "loss": 0.3644, + "step": 4831 + }, + { + "epoch": 2.6232356134636268, + "grad_norm": 7.019904615334962, + "learning_rate": 9.68880995893846e-06, + "loss": 0.4594, + "step": 4832 + }, + { + "epoch": 2.6237785016286646, + "grad_norm": 11.780644240657415, + "learning_rate": 9.685294842088256e-06, + "loss": 1.0625, + "step": 4833 + }, + { + "epoch": 2.6243213897937023, + "grad_norm": 7.34603375318287, + "learning_rate": 9.681779764161289e-06, + "loss": 0.3502, + "step": 4834 + }, + { + "epoch": 2.6248642779587406, + "grad_norm": 9.787656886442962, + "learning_rate": 9.678264725592306e-06, + "loss": 0.5054, + "step": 4835 + }, + { + "epoch": 2.6254071661237783, + "grad_norm": 9.687484513116981, + "learning_rate": 9.674749726816058e-06, + "loss": 0.8678, + "step": 4836 + }, + { + "epoch": 2.6259500542888166, + "grad_norm": 7.620468702427845, + "learning_rate": 9.671234768267284e-06, + "loss": 0.4672, + "step": 4837 + }, + { + "epoch": 2.6264929424538543, + "grad_norm": 14.051025442281276, + "learning_rate": 9.667719850380712e-06, + "loss": 0.6462, + "step": 4838 + }, + { + "epoch": 2.6270358306188926, + "grad_norm": 11.579748119704858, + "learning_rate": 9.664204973591081e-06, + "loss": 0.8856, + "step": 4839 + }, + { + "epoch": 2.6275787187839303, + "grad_norm": 8.348102354642592, + "learning_rate": 9.660690138333114e-06, + "loss": 0.5623, + "step": 4840 + }, + { + "epoch": 2.6281216069489686, + "grad_norm": 9.351950984929493, + "learning_rate": 9.657175345041532e-06, + "loss": 0.5809, + "step": 4841 + }, + { + "epoch": 2.6286644951140063, + "grad_norm": 12.421448876984805, + "learning_rate": 9.653660594151047e-06, + "loss": 0.7831, + "step": 4842 + }, + { + "epoch": 2.6292073832790446, + "grad_norm": 11.981212578199576, + "learning_rate": 9.650145886096376e-06, + "loss": 0.8767, + "step": 4843 + }, + { + "epoch": 2.6297502714440824, + "grad_norm": 10.659578351567276, + "learning_rate": 9.646631221312216e-06, + "loss": 0.7313, + "step": 4844 + }, + { + "epoch": 2.6302931596091206, + "grad_norm": 12.480207661204222, + "learning_rate": 9.643116600233274e-06, + "loss": 0.7412, + "step": 4845 + }, + { + "epoch": 2.6308360477741584, + "grad_norm": 9.131072461148838, + "learning_rate": 9.639602023294233e-06, + "loss": 0.5518, + "step": 4846 + }, + { + "epoch": 2.6313789359391966, + "grad_norm": 7.489089256780018, + "learning_rate": 9.636087490929793e-06, + "loss": 0.4321, + "step": 4847 + }, + { + "epoch": 2.6319218241042344, + "grad_norm": 12.12610683915495, + "learning_rate": 9.632573003574634e-06, + "loss": 0.8265, + "step": 4848 + }, + { + "epoch": 2.6324647122692726, + "grad_norm": 7.965586818160769, + "learning_rate": 9.629058561663426e-06, + "loss": 0.373, + "step": 4849 + }, + { + "epoch": 2.6330076004343104, + "grad_norm": 9.348866579305472, + "learning_rate": 9.62554416563085e-06, + "loss": 0.4274, + "step": 4850 + }, + { + "epoch": 2.6335504885993486, + "grad_norm": 10.13050086693858, + "learning_rate": 9.622029815911566e-06, + "loss": 0.5276, + "step": 4851 + }, + { + "epoch": 2.6340933767643864, + "grad_norm": 10.19180911105006, + "learning_rate": 9.61851551294024e-06, + "loss": 0.5871, + "step": 4852 + }, + { + "epoch": 2.6346362649294246, + "grad_norm": 11.184621464328854, + "learning_rate": 9.61500125715152e-06, + "loss": 0.6047, + "step": 4853 + }, + { + "epoch": 2.6351791530944624, + "grad_norm": 14.484355041392742, + "learning_rate": 9.611487048980058e-06, + "loss": 0.8628, + "step": 4854 + }, + { + "epoch": 2.6357220412595006, + "grad_norm": 13.070986825901516, + "learning_rate": 9.607972888860497e-06, + "loss": 0.7793, + "step": 4855 + }, + { + "epoch": 2.6362649294245384, + "grad_norm": 9.640708751060075, + "learning_rate": 9.604458777227477e-06, + "loss": 0.6016, + "step": 4856 + }, + { + "epoch": 2.6368078175895766, + "grad_norm": 11.05523237140924, + "learning_rate": 9.600944714515627e-06, + "loss": 0.5999, + "step": 4857 + }, + { + "epoch": 2.6373507057546144, + "grad_norm": 13.240640169530616, + "learning_rate": 9.597430701159565e-06, + "loss": 0.6779, + "step": 4858 + }, + { + "epoch": 2.6378935939196526, + "grad_norm": 10.878893821343203, + "learning_rate": 9.593916737593919e-06, + "loss": 0.6119, + "step": 4859 + }, + { + "epoch": 2.6384364820846904, + "grad_norm": 10.820526634571284, + "learning_rate": 9.590402824253295e-06, + "loss": 0.839, + "step": 4860 + }, + { + "epoch": 2.6389793702497286, + "grad_norm": 12.938750622065529, + "learning_rate": 9.586888961572307e-06, + "loss": 0.517, + "step": 4861 + }, + { + "epoch": 2.6395222584147664, + "grad_norm": 9.725430374953143, + "learning_rate": 9.583375149985547e-06, + "loss": 0.7604, + "step": 4862 + }, + { + "epoch": 2.6400651465798046, + "grad_norm": 11.840312760690757, + "learning_rate": 9.579861389927615e-06, + "loss": 0.7614, + "step": 4863 + }, + { + "epoch": 2.6406080347448424, + "grad_norm": 11.047185391206618, + "learning_rate": 9.576347681833093e-06, + "loss": 0.6351, + "step": 4864 + }, + { + "epoch": 2.6411509229098806, + "grad_norm": 9.986349266550256, + "learning_rate": 9.572834026136571e-06, + "loss": 0.5549, + "step": 4865 + }, + { + "epoch": 2.6416938110749184, + "grad_norm": 14.455847384105025, + "learning_rate": 9.56932042327261e-06, + "loss": 0.8206, + "step": 4866 + }, + { + "epoch": 2.6422366992399566, + "grad_norm": 10.377988559117977, + "learning_rate": 9.56580687367579e-06, + "loss": 0.5275, + "step": 4867 + }, + { + "epoch": 2.6427795874049944, + "grad_norm": 10.71910613137521, + "learning_rate": 9.562293377780668e-06, + "loss": 0.6958, + "step": 4868 + }, + { + "epoch": 2.6433224755700326, + "grad_norm": 8.635819291033194, + "learning_rate": 9.558779936021795e-06, + "loss": 0.3946, + "step": 4869 + }, + { + "epoch": 2.6438653637350704, + "grad_norm": 16.60173101533996, + "learning_rate": 9.555266548833728e-06, + "loss": 1.066, + "step": 4870 + }, + { + "epoch": 2.6444082519001086, + "grad_norm": 14.726517961222527, + "learning_rate": 9.551753216650998e-06, + "loss": 1.1356, + "step": 4871 + }, + { + "epoch": 2.6449511400651464, + "grad_norm": 11.123184991281361, + "learning_rate": 9.548239939908146e-06, + "loss": 0.757, + "step": 4872 + }, + { + "epoch": 2.6454940282301846, + "grad_norm": 11.666285426731331, + "learning_rate": 9.544726719039699e-06, + "loss": 0.7109, + "step": 4873 + }, + { + "epoch": 2.6460369163952224, + "grad_norm": 9.657892010966115, + "learning_rate": 9.541213554480176e-06, + "loss": 0.6191, + "step": 4874 + }, + { + "epoch": 2.6465798045602607, + "grad_norm": 9.984832056257446, + "learning_rate": 9.53770044666409e-06, + "loss": 0.8649, + "step": 4875 + }, + { + "epoch": 2.6471226927252984, + "grad_norm": 9.143852173958573, + "learning_rate": 9.534187396025955e-06, + "loss": 0.5282, + "step": 4876 + }, + { + "epoch": 2.6476655808903367, + "grad_norm": 8.129441981336987, + "learning_rate": 9.530674403000266e-06, + "loss": 0.3783, + "step": 4877 + }, + { + "epoch": 2.6482084690553744, + "grad_norm": 10.147175407433437, + "learning_rate": 9.527161468021508e-06, + "loss": 0.6676, + "step": 4878 + }, + { + "epoch": 2.6487513572204127, + "grad_norm": 11.015226101188722, + "learning_rate": 9.523648591524176e-06, + "loss": 0.5218, + "step": 4879 + }, + { + "epoch": 2.6492942453854504, + "grad_norm": 9.374140703922862, + "learning_rate": 9.520135773942743e-06, + "loss": 0.6754, + "step": 4880 + }, + { + "epoch": 2.6498371335504887, + "grad_norm": 8.84122429163583, + "learning_rate": 9.516623015711688e-06, + "loss": 0.5236, + "step": 4881 + }, + { + "epoch": 2.6503800217155264, + "grad_norm": 10.659047956662723, + "learning_rate": 9.513110317265463e-06, + "loss": 0.4742, + "step": 4882 + }, + { + "epoch": 2.6509229098805647, + "grad_norm": 8.798273983585764, + "learning_rate": 9.509597679038536e-06, + "loss": 0.4949, + "step": 4883 + }, + { + "epoch": 2.6514657980456025, + "grad_norm": 14.827321458575321, + "learning_rate": 9.506085101465344e-06, + "loss": 0.967, + "step": 4884 + }, + { + "epoch": 2.6520086862106407, + "grad_norm": 13.212543627120823, + "learning_rate": 9.502572584980338e-06, + "loss": 0.904, + "step": 4885 + }, + { + "epoch": 2.6525515743756785, + "grad_norm": 12.151700192108333, + "learning_rate": 9.499060130017947e-06, + "loss": 0.6512, + "step": 4886 + }, + { + "epoch": 2.6530944625407167, + "grad_norm": 9.051653631541132, + "learning_rate": 9.495547737012594e-06, + "loss": 0.5264, + "step": 4887 + }, + { + "epoch": 2.6536373507057545, + "grad_norm": 11.286333388065833, + "learning_rate": 9.492035406398706e-06, + "loss": 0.5864, + "step": 4888 + }, + { + "epoch": 2.6541802388707927, + "grad_norm": 8.303712468523708, + "learning_rate": 9.488523138610684e-06, + "loss": 0.5887, + "step": 4889 + }, + { + "epoch": 2.6547231270358305, + "grad_norm": 8.762643235318839, + "learning_rate": 9.485010934082939e-06, + "loss": 0.5079, + "step": 4890 + }, + { + "epoch": 2.6552660152008687, + "grad_norm": 13.211351989906888, + "learning_rate": 9.481498793249859e-06, + "loss": 0.8983, + "step": 4891 + }, + { + "epoch": 2.6558089033659065, + "grad_norm": 11.087558391384889, + "learning_rate": 9.477986716545834e-06, + "loss": 1.1671, + "step": 4892 + }, + { + "epoch": 2.6563517915309447, + "grad_norm": 6.825801048730381, + "learning_rate": 9.474474704405245e-06, + "loss": 0.3664, + "step": 4893 + }, + { + "epoch": 2.6568946796959825, + "grad_norm": 6.583440546159338, + "learning_rate": 9.470962757262465e-06, + "loss": 0.5464, + "step": 4894 + }, + { + "epoch": 2.6574375678610207, + "grad_norm": 8.987234989797157, + "learning_rate": 9.46745087555185e-06, + "loss": 0.5954, + "step": 4895 + }, + { + "epoch": 2.6579804560260585, + "grad_norm": 10.699367568639158, + "learning_rate": 9.463939059707763e-06, + "loss": 0.632, + "step": 4896 + }, + { + "epoch": 2.6585233441910967, + "grad_norm": 10.246298042329498, + "learning_rate": 9.460427310164548e-06, + "loss": 0.6766, + "step": 4897 + }, + { + "epoch": 2.6590662323561345, + "grad_norm": 12.300177238674, + "learning_rate": 9.456915627356542e-06, + "loss": 0.7277, + "step": 4898 + }, + { + "epoch": 2.6596091205211727, + "grad_norm": 9.893644041111745, + "learning_rate": 9.453404011718077e-06, + "loss": 0.644, + "step": 4899 + }, + { + "epoch": 2.6601520086862105, + "grad_norm": 10.80727748895235, + "learning_rate": 9.449892463683471e-06, + "loss": 0.5844, + "step": 4900 + }, + { + "epoch": 2.6606948968512487, + "grad_norm": 8.350836469669698, + "learning_rate": 9.446380983687046e-06, + "loss": 0.4745, + "step": 4901 + }, + { + "epoch": 2.6612377850162865, + "grad_norm": 10.697471057638213, + "learning_rate": 9.442869572163101e-06, + "loss": 0.5749, + "step": 4902 + }, + { + "epoch": 2.6617806731813247, + "grad_norm": 7.582465634030485, + "learning_rate": 9.43935822954594e-06, + "loss": 0.4481, + "step": 4903 + }, + { + "epoch": 2.6623235613463625, + "grad_norm": 10.01801356389796, + "learning_rate": 9.435846956269841e-06, + "loss": 0.5817, + "step": 4904 + }, + { + "epoch": 2.6628664495114007, + "grad_norm": 9.383471785643394, + "learning_rate": 9.432335752769092e-06, + "loss": 0.526, + "step": 4905 + }, + { + "epoch": 2.6634093376764385, + "grad_norm": 9.555536887097055, + "learning_rate": 9.428824619477964e-06, + "loss": 0.5056, + "step": 4906 + }, + { + "epoch": 2.6639522258414767, + "grad_norm": 7.636054626405326, + "learning_rate": 9.425313556830713e-06, + "loss": 0.4383, + "step": 4907 + }, + { + "epoch": 2.6644951140065145, + "grad_norm": 9.842725681691716, + "learning_rate": 9.421802565261602e-06, + "loss": 0.548, + "step": 4908 + }, + { + "epoch": 2.6650380021715527, + "grad_norm": 12.891942531084513, + "learning_rate": 9.418291645204865e-06, + "loss": 0.6489, + "step": 4909 + }, + { + "epoch": 2.6655808903365905, + "grad_norm": 15.298821348423894, + "learning_rate": 9.41478079709475e-06, + "loss": 0.8286, + "step": 4910 + }, + { + "epoch": 2.6661237785016287, + "grad_norm": 10.80698538838323, + "learning_rate": 9.411270021365475e-06, + "loss": 0.7704, + "step": 4911 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 10.837498285149955, + "learning_rate": 9.407759318451264e-06, + "loss": 0.5384, + "step": 4912 + }, + { + "epoch": 2.6672095548317047, + "grad_norm": 11.666848847579812, + "learning_rate": 9.404248688786322e-06, + "loss": 0.6402, + "step": 4913 + }, + { + "epoch": 2.6677524429967425, + "grad_norm": 12.038932116915525, + "learning_rate": 9.400738132804856e-06, + "loss": 0.559, + "step": 4914 + }, + { + "epoch": 2.6682953311617807, + "grad_norm": 13.411348889811451, + "learning_rate": 9.397227650941048e-06, + "loss": 0.9587, + "step": 4915 + }, + { + "epoch": 2.6688382193268185, + "grad_norm": 8.79354274091956, + "learning_rate": 9.393717243629091e-06, + "loss": 0.5627, + "step": 4916 + }, + { + "epoch": 2.6693811074918568, + "grad_norm": 8.956388294197298, + "learning_rate": 9.390206911303148e-06, + "loss": 0.5694, + "step": 4917 + }, + { + "epoch": 2.6699239956568945, + "grad_norm": 11.296789676569125, + "learning_rate": 9.386696654397384e-06, + "loss": 0.8356, + "step": 4918 + }, + { + "epoch": 2.6704668838219328, + "grad_norm": 8.778967068719545, + "learning_rate": 9.38318647334596e-06, + "loss": 0.5283, + "step": 4919 + }, + { + "epoch": 2.6710097719869705, + "grad_norm": 10.041165955923448, + "learning_rate": 9.379676368583011e-06, + "loss": 0.7492, + "step": 4920 + }, + { + "epoch": 2.6715526601520088, + "grad_norm": 7.152774382953927, + "learning_rate": 9.376166340542682e-06, + "loss": 0.4563, + "step": 4921 + }, + { + "epoch": 2.6720955483170465, + "grad_norm": 10.189040304576451, + "learning_rate": 9.37265638965909e-06, + "loss": 0.5869, + "step": 4922 + }, + { + "epoch": 2.6726384364820848, + "grad_norm": 8.47728562416046, + "learning_rate": 9.369146516366359e-06, + "loss": 0.6168, + "step": 4923 + }, + { + "epoch": 2.6731813246471225, + "grad_norm": 9.47866332824828, + "learning_rate": 9.365636721098588e-06, + "loss": 0.5233, + "step": 4924 + }, + { + "epoch": 2.6737242128121608, + "grad_norm": 11.264475439058582, + "learning_rate": 9.362127004289883e-06, + "loss": 0.693, + "step": 4925 + }, + { + "epoch": 2.6742671009771986, + "grad_norm": 11.495205201206721, + "learning_rate": 9.358617366374327e-06, + "loss": 0.587, + "step": 4926 + }, + { + "epoch": 2.6748099891422368, + "grad_norm": 10.081706772004408, + "learning_rate": 9.355107807785993e-06, + "loss": 0.65, + "step": 4927 + }, + { + "epoch": 2.6753528773072746, + "grad_norm": 8.912371373305913, + "learning_rate": 9.351598328958957e-06, + "loss": 0.6475, + "step": 4928 + }, + { + "epoch": 2.675895765472313, + "grad_norm": 8.434539074965906, + "learning_rate": 9.348088930327269e-06, + "loss": 0.4321, + "step": 4929 + }, + { + "epoch": 2.6764386536373506, + "grad_norm": 10.356652945785703, + "learning_rate": 9.344579612324984e-06, + "loss": 0.4243, + "step": 4930 + }, + { + "epoch": 2.676981541802389, + "grad_norm": 10.975361657654028, + "learning_rate": 9.341070375386134e-06, + "loss": 0.6042, + "step": 4931 + }, + { + "epoch": 2.6775244299674266, + "grad_norm": 10.912795228418068, + "learning_rate": 9.337561219944756e-06, + "loss": 0.6655, + "step": 4932 + }, + { + "epoch": 2.678067318132465, + "grad_norm": 9.852947274867201, + "learning_rate": 9.334052146434857e-06, + "loss": 0.7696, + "step": 4933 + }, + { + "epoch": 2.6786102062975026, + "grad_norm": 7.323342130507832, + "learning_rate": 9.330543155290454e-06, + "loss": 0.3453, + "step": 4934 + }, + { + "epoch": 2.679153094462541, + "grad_norm": 7.335485317101141, + "learning_rate": 9.32703424694554e-06, + "loss": 0.4291, + "step": 4935 + }, + { + "epoch": 2.6796959826275786, + "grad_norm": 12.700387251668076, + "learning_rate": 9.3235254218341e-06, + "loss": 0.7156, + "step": 4936 + }, + { + "epoch": 2.680238870792617, + "grad_norm": 10.828431765432423, + "learning_rate": 9.320016680390115e-06, + "loss": 0.7293, + "step": 4937 + }, + { + "epoch": 2.6807817589576546, + "grad_norm": 9.224454671275938, + "learning_rate": 9.316508023047548e-06, + "loss": 0.4157, + "step": 4938 + }, + { + "epoch": 2.681324647122693, + "grad_norm": 13.095737517917204, + "learning_rate": 9.312999450240362e-06, + "loss": 0.7928, + "step": 4939 + }, + { + "epoch": 2.6818675352877306, + "grad_norm": 10.338830489426515, + "learning_rate": 9.309490962402498e-06, + "loss": 0.6826, + "step": 4940 + }, + { + "epoch": 2.682410423452769, + "grad_norm": 11.650643384237476, + "learning_rate": 9.305982559967894e-06, + "loss": 0.6709, + "step": 4941 + }, + { + "epoch": 2.6829533116178066, + "grad_norm": 9.153121673948778, + "learning_rate": 9.302474243370469e-06, + "loss": 0.4998, + "step": 4942 + }, + { + "epoch": 2.683496199782845, + "grad_norm": 11.760747525125, + "learning_rate": 9.298966013044144e-06, + "loss": 1.0389, + "step": 4943 + }, + { + "epoch": 2.6840390879478826, + "grad_norm": 13.657927099528628, + "learning_rate": 9.295457869422818e-06, + "loss": 0.9807, + "step": 4944 + }, + { + "epoch": 2.684581976112921, + "grad_norm": 12.726310962899635, + "learning_rate": 9.291949812940387e-06, + "loss": 0.6495, + "step": 4945 + }, + { + "epoch": 2.6851248642779586, + "grad_norm": 11.439120086220523, + "learning_rate": 9.288441844030735e-06, + "loss": 0.6616, + "step": 4946 + }, + { + "epoch": 2.685667752442997, + "grad_norm": 10.384831951074661, + "learning_rate": 9.284933963127724e-06, + "loss": 0.536, + "step": 4947 + }, + { + "epoch": 2.6862106406080346, + "grad_norm": 8.803102808116847, + "learning_rate": 9.281426170665225e-06, + "loss": 0.6621, + "step": 4948 + }, + { + "epoch": 2.686753528773073, + "grad_norm": 14.516361098495818, + "learning_rate": 9.277918467077077e-06, + "loss": 1.0579, + "step": 4949 + }, + { + "epoch": 2.6872964169381106, + "grad_norm": 11.08498885907847, + "learning_rate": 9.274410852797126e-06, + "loss": 0.463, + "step": 4950 + }, + { + "epoch": 2.687839305103149, + "grad_norm": 10.827937520590183, + "learning_rate": 9.270903328259195e-06, + "loss": 0.5015, + "step": 4951 + }, + { + "epoch": 2.6883821932681866, + "grad_norm": 11.572829724801146, + "learning_rate": 9.267395893897107e-06, + "loss": 0.663, + "step": 4952 + }, + { + "epoch": 2.688925081433225, + "grad_norm": 9.3016488903525, + "learning_rate": 9.26388855014466e-06, + "loss": 0.6876, + "step": 4953 + }, + { + "epoch": 2.6894679695982626, + "grad_norm": 17.324535315422423, + "learning_rate": 9.260381297435652e-06, + "loss": 1.1741, + "step": 4954 + }, + { + "epoch": 2.690010857763301, + "grad_norm": 11.465003825461562, + "learning_rate": 9.256874136203864e-06, + "loss": 0.6867, + "step": 4955 + }, + { + "epoch": 2.6905537459283386, + "grad_norm": 8.846935307480056, + "learning_rate": 9.253367066883063e-06, + "loss": 0.4518, + "step": 4956 + }, + { + "epoch": 2.691096634093377, + "grad_norm": 11.007513424878196, + "learning_rate": 9.249860089907018e-06, + "loss": 0.7987, + "step": 4957 + }, + { + "epoch": 2.6916395222584146, + "grad_norm": 12.297659996678936, + "learning_rate": 9.246353205709468e-06, + "loss": 0.588, + "step": 4958 + }, + { + "epoch": 2.692182410423453, + "grad_norm": 10.549862062131211, + "learning_rate": 9.24284641472416e-06, + "loss": 0.737, + "step": 4959 + }, + { + "epoch": 2.6927252985884906, + "grad_norm": 10.712460113916158, + "learning_rate": 9.23933971738481e-06, + "loss": 0.6092, + "step": 4960 + }, + { + "epoch": 2.693268186753529, + "grad_norm": 14.104109743551868, + "learning_rate": 9.235833114125141e-06, + "loss": 1.008, + "step": 4961 + }, + { + "epoch": 2.6938110749185666, + "grad_norm": 7.755916952898445, + "learning_rate": 9.232326605378843e-06, + "loss": 0.4708, + "step": 4962 + }, + { + "epoch": 2.694353963083605, + "grad_norm": 11.677016459785008, + "learning_rate": 9.228820191579618e-06, + "loss": 0.7099, + "step": 4963 + }, + { + "epoch": 2.6948968512486426, + "grad_norm": 10.070884129580872, + "learning_rate": 9.225313873161139e-06, + "loss": 0.6546, + "step": 4964 + }, + { + "epoch": 2.695439739413681, + "grad_norm": 8.108710814711873, + "learning_rate": 9.221807650557078e-06, + "loss": 0.6102, + "step": 4965 + }, + { + "epoch": 2.6959826275787186, + "grad_norm": 10.279519502018594, + "learning_rate": 9.218301524201087e-06, + "loss": 0.5783, + "step": 4966 + }, + { + "epoch": 2.696525515743757, + "grad_norm": 11.882105089319197, + "learning_rate": 9.214795494526806e-06, + "loss": 1.1489, + "step": 4967 + }, + { + "epoch": 2.6970684039087947, + "grad_norm": 9.767493447568013, + "learning_rate": 9.211289561967872e-06, + "loss": 0.6051, + "step": 4968 + }, + { + "epoch": 2.697611292073833, + "grad_norm": 11.690415242614575, + "learning_rate": 9.207783726957903e-06, + "loss": 0.592, + "step": 4969 + }, + { + "epoch": 2.6981541802388707, + "grad_norm": 7.427254360689941, + "learning_rate": 9.204277989930502e-06, + "loss": 0.4672, + "step": 4970 + }, + { + "epoch": 2.698697068403909, + "grad_norm": 9.41305388818184, + "learning_rate": 9.200772351319266e-06, + "loss": 0.515, + "step": 4971 + }, + { + "epoch": 2.6992399565689467, + "grad_norm": 11.02585368985108, + "learning_rate": 9.197266811557787e-06, + "loss": 0.7867, + "step": 4972 + }, + { + "epoch": 2.699782844733985, + "grad_norm": 10.462642998204657, + "learning_rate": 9.193761371079622e-06, + "loss": 0.6689, + "step": 4973 + }, + { + "epoch": 2.7003257328990227, + "grad_norm": 9.964585703249949, + "learning_rate": 9.190256030318339e-06, + "loss": 0.6676, + "step": 4974 + }, + { + "epoch": 2.700868621064061, + "grad_norm": 12.383650152937822, + "learning_rate": 9.186750789707478e-06, + "loss": 0.9124, + "step": 4975 + }, + { + "epoch": 2.7014115092290987, + "grad_norm": 7.428651587888086, + "learning_rate": 9.183245649680574e-06, + "loss": 0.6241, + "step": 4976 + }, + { + "epoch": 2.701954397394137, + "grad_norm": 8.171702461332401, + "learning_rate": 9.179740610671155e-06, + "loss": 0.468, + "step": 4977 + }, + { + "epoch": 2.7024972855591747, + "grad_norm": 15.630545262276021, + "learning_rate": 9.176235673112719e-06, + "loss": 1.4177, + "step": 4978 + }, + { + "epoch": 2.703040173724213, + "grad_norm": 9.04535779717474, + "learning_rate": 9.172730837438774e-06, + "loss": 0.504, + "step": 4979 + }, + { + "epoch": 2.7035830618892507, + "grad_norm": 9.263908210238581, + "learning_rate": 9.169226104082792e-06, + "loss": 0.4043, + "step": 4980 + }, + { + "epoch": 2.704125950054289, + "grad_norm": 10.022418469090024, + "learning_rate": 9.165721473478253e-06, + "loss": 0.6372, + "step": 4981 + }, + { + "epoch": 2.7046688382193267, + "grad_norm": 12.924869609819282, + "learning_rate": 9.16221694605861e-06, + "loss": 0.9433, + "step": 4982 + }, + { + "epoch": 2.705211726384365, + "grad_norm": 8.79512920135937, + "learning_rate": 9.158712522257309e-06, + "loss": 0.4725, + "step": 4983 + }, + { + "epoch": 2.7057546145494027, + "grad_norm": 11.753381756239166, + "learning_rate": 9.155208202507789e-06, + "loss": 0.7395, + "step": 4984 + }, + { + "epoch": 2.706297502714441, + "grad_norm": 6.454949124720575, + "learning_rate": 9.151703987243459e-06, + "loss": 0.3108, + "step": 4985 + }, + { + "epoch": 2.7068403908794787, + "grad_norm": 15.244281036714051, + "learning_rate": 9.148199876897737e-06, + "loss": 0.802, + "step": 4986 + }, + { + "epoch": 2.707383279044517, + "grad_norm": 11.093235618680396, + "learning_rate": 9.144695871904005e-06, + "loss": 0.6252, + "step": 4987 + }, + { + "epoch": 2.7079261672095547, + "grad_norm": 10.703835311171614, + "learning_rate": 9.141191972695655e-06, + "loss": 0.6642, + "step": 4988 + }, + { + "epoch": 2.708469055374593, + "grad_norm": 11.327820449600674, + "learning_rate": 9.137688179706049e-06, + "loss": 1.1613, + "step": 4989 + }, + { + "epoch": 2.7090119435396307, + "grad_norm": 7.527861958807383, + "learning_rate": 9.134184493368548e-06, + "loss": 0.4991, + "step": 4990 + }, + { + "epoch": 2.709554831704669, + "grad_norm": 10.199776895453677, + "learning_rate": 9.130680914116484e-06, + "loss": 0.977, + "step": 4991 + }, + { + "epoch": 2.7100977198697067, + "grad_norm": 9.778998506495286, + "learning_rate": 9.127177442383192e-06, + "loss": 0.6729, + "step": 4992 + }, + { + "epoch": 2.710640608034745, + "grad_norm": 10.831437132716289, + "learning_rate": 9.123674078601984e-06, + "loss": 0.6252, + "step": 4993 + }, + { + "epoch": 2.7111834961997827, + "grad_norm": 9.722012246624114, + "learning_rate": 9.120170823206165e-06, + "loss": 0.6169, + "step": 4994 + }, + { + "epoch": 2.711726384364821, + "grad_norm": 9.696836781330795, + "learning_rate": 9.116667676629019e-06, + "loss": 0.7824, + "step": 4995 + }, + { + "epoch": 2.7122692725298587, + "grad_norm": 8.532030770830737, + "learning_rate": 9.11316463930382e-06, + "loss": 0.7649, + "step": 4996 + }, + { + "epoch": 2.712812160694897, + "grad_norm": 10.155581062168379, + "learning_rate": 9.109661711663837e-06, + "loss": 0.5581, + "step": 4997 + }, + { + "epoch": 2.7133550488599347, + "grad_norm": 11.913092610845165, + "learning_rate": 9.106158894142307e-06, + "loss": 0.6256, + "step": 4998 + }, + { + "epoch": 2.713897937024973, + "grad_norm": 10.326179923874385, + "learning_rate": 9.102656187172475e-06, + "loss": 0.8703, + "step": 4999 + }, + { + "epoch": 2.7144408251900107, + "grad_norm": 10.51344451860866, + "learning_rate": 9.099153591187553e-06, + "loss": 0.8281, + "step": 5000 + }, + { + "epoch": 2.714983713355049, + "grad_norm": 9.101512118782772, + "learning_rate": 9.09565110662075e-06, + "loss": 0.4127, + "step": 5001 + }, + { + "epoch": 2.7155266015200867, + "grad_norm": 14.912595390103256, + "learning_rate": 9.092148733905257e-06, + "loss": 0.828, + "step": 5002 + }, + { + "epoch": 2.716069489685125, + "grad_norm": 10.567445107841705, + "learning_rate": 9.088646473474262e-06, + "loss": 0.8542, + "step": 5003 + }, + { + "epoch": 2.7166123778501627, + "grad_norm": 11.727344925695448, + "learning_rate": 9.085144325760922e-06, + "loss": 0.7593, + "step": 5004 + }, + { + "epoch": 2.717155266015201, + "grad_norm": 10.964390265146204, + "learning_rate": 9.081642291198387e-06, + "loss": 0.696, + "step": 5005 + }, + { + "epoch": 2.7176981541802387, + "grad_norm": 8.594728435704301, + "learning_rate": 9.0781403702198e-06, + "loss": 0.3582, + "step": 5006 + }, + { + "epoch": 2.718241042345277, + "grad_norm": 10.662428429321805, + "learning_rate": 9.074638563258279e-06, + "loss": 0.6213, + "step": 5007 + }, + { + "epoch": 2.7187839305103148, + "grad_norm": 13.853700163509458, + "learning_rate": 9.071136870746934e-06, + "loss": 0.653, + "step": 5008 + }, + { + "epoch": 2.719326818675353, + "grad_norm": 9.421600748659138, + "learning_rate": 9.067635293118862e-06, + "loss": 0.4653, + "step": 5009 + }, + { + "epoch": 2.7198697068403908, + "grad_norm": 15.40532102462592, + "learning_rate": 9.064133830807147e-06, + "loss": 0.8502, + "step": 5010 + }, + { + "epoch": 2.720412595005429, + "grad_norm": 14.246891389640892, + "learning_rate": 9.060632484244845e-06, + "loss": 0.711, + "step": 5011 + }, + { + "epoch": 2.7209554831704668, + "grad_norm": 13.657358501095473, + "learning_rate": 9.057131253865022e-06, + "loss": 1.0586, + "step": 5012 + }, + { + "epoch": 2.721498371335505, + "grad_norm": 11.08278909637767, + "learning_rate": 9.053630140100701e-06, + "loss": 0.5658, + "step": 5013 + }, + { + "epoch": 2.7220412595005428, + "grad_norm": 11.922765678431421, + "learning_rate": 9.050129143384917e-06, + "loss": 0.6929, + "step": 5014 + }, + { + "epoch": 2.722584147665581, + "grad_norm": 10.93390692730677, + "learning_rate": 9.046628264150674e-06, + "loss": 1.417, + "step": 5015 + }, + { + "epoch": 2.7231270358306188, + "grad_norm": 12.86011863530373, + "learning_rate": 9.043127502830964e-06, + "loss": 1.0364, + "step": 5016 + }, + { + "epoch": 2.723669923995657, + "grad_norm": 7.848749292562499, + "learning_rate": 9.039626859858773e-06, + "loss": 0.6696, + "step": 5017 + }, + { + "epoch": 2.7242128121606948, + "grad_norm": 11.212923870670298, + "learning_rate": 9.036126335667059e-06, + "loss": 0.5751, + "step": 5018 + }, + { + "epoch": 2.724755700325733, + "grad_norm": 9.666644723437658, + "learning_rate": 9.032625930688781e-06, + "loss": 0.5699, + "step": 5019 + }, + { + "epoch": 2.725298588490771, + "grad_norm": 8.860936490743164, + "learning_rate": 9.029125645356864e-06, + "loss": 0.4262, + "step": 5020 + }, + { + "epoch": 2.725841476655809, + "grad_norm": 8.597880980516297, + "learning_rate": 9.025625480104238e-06, + "loss": 0.5646, + "step": 5021 + }, + { + "epoch": 2.726384364820847, + "grad_norm": 11.334362425216222, + "learning_rate": 9.022125435363803e-06, + "loss": 0.7424, + "step": 5022 + }, + { + "epoch": 2.726927252985885, + "grad_norm": 10.551777615550284, + "learning_rate": 9.018625511568456e-06, + "loss": 0.4961, + "step": 5023 + }, + { + "epoch": 2.727470141150923, + "grad_norm": 9.061655577456406, + "learning_rate": 9.015125709151069e-06, + "loss": 0.6466, + "step": 5024 + }, + { + "epoch": 2.728013029315961, + "grad_norm": 12.51368822693772, + "learning_rate": 9.011626028544502e-06, + "loss": 0.8461, + "step": 5025 + }, + { + "epoch": 2.728555917480999, + "grad_norm": 9.972383117450178, + "learning_rate": 9.008126470181605e-06, + "loss": 0.7941, + "step": 5026 + }, + { + "epoch": 2.729098805646037, + "grad_norm": 9.372092186313578, + "learning_rate": 9.004627034495204e-06, + "loss": 0.5291, + "step": 5027 + }, + { + "epoch": 2.729641693811075, + "grad_norm": 11.649348346812042, + "learning_rate": 9.00112772191812e-06, + "loss": 0.658, + "step": 5028 + }, + { + "epoch": 2.730184581976113, + "grad_norm": 9.4217070731629, + "learning_rate": 8.997628532883149e-06, + "loss": 0.7391, + "step": 5029 + }, + { + "epoch": 2.730727470141151, + "grad_norm": 9.294886195505088, + "learning_rate": 8.994129467823083e-06, + "loss": 0.7955, + "step": 5030 + }, + { + "epoch": 2.731270358306189, + "grad_norm": 9.046521616766688, + "learning_rate": 8.990630527170684e-06, + "loss": 0.6163, + "step": 5031 + }, + { + "epoch": 2.731813246471227, + "grad_norm": 13.078763070358596, + "learning_rate": 8.987131711358714e-06, + "loss": 0.6352, + "step": 5032 + }, + { + "epoch": 2.732356134636265, + "grad_norm": 8.372819066707777, + "learning_rate": 8.983633020819906e-06, + "loss": 0.5139, + "step": 5033 + }, + { + "epoch": 2.732899022801303, + "grad_norm": 7.908976790945395, + "learning_rate": 8.980134455986982e-06, + "loss": 0.4263, + "step": 5034 + }, + { + "epoch": 2.733441910966341, + "grad_norm": 9.508847688216756, + "learning_rate": 8.97663601729266e-06, + "loss": 0.614, + "step": 5035 + }, + { + "epoch": 2.733984799131379, + "grad_norm": 9.135340041370506, + "learning_rate": 8.973137705169621e-06, + "loss": 0.5803, + "step": 5036 + }, + { + "epoch": 2.734527687296417, + "grad_norm": 10.900299047123902, + "learning_rate": 8.969639520050553e-06, + "loss": 0.4833, + "step": 5037 + }, + { + "epoch": 2.735070575461455, + "grad_norm": 9.243890767443602, + "learning_rate": 8.966141462368106e-06, + "loss": 0.4405, + "step": 5038 + }, + { + "epoch": 2.735613463626493, + "grad_norm": 11.372121051541363, + "learning_rate": 8.962643532554934e-06, + "loss": 0.6618, + "step": 5039 + }, + { + "epoch": 2.736156351791531, + "grad_norm": 11.318259460716959, + "learning_rate": 8.95914573104366e-06, + "loss": 1.0367, + "step": 5040 + }, + { + "epoch": 2.736699239956569, + "grad_norm": 9.055918919260264, + "learning_rate": 8.955648058266904e-06, + "loss": 0.5463, + "step": 5041 + }, + { + "epoch": 2.737242128121607, + "grad_norm": 12.452665937187591, + "learning_rate": 8.952150514657258e-06, + "loss": 0.9772, + "step": 5042 + }, + { + "epoch": 2.737785016286645, + "grad_norm": 11.532813733829784, + "learning_rate": 8.94865310064731e-06, + "loss": 0.6372, + "step": 5043 + }, + { + "epoch": 2.738327904451683, + "grad_norm": 9.800145767151978, + "learning_rate": 8.945155816669622e-06, + "loss": 0.6384, + "step": 5044 + }, + { + "epoch": 2.738870792616721, + "grad_norm": 13.619886494303222, + "learning_rate": 8.94165866315674e-06, + "loss": 0.7295, + "step": 5045 + }, + { + "epoch": 2.739413680781759, + "grad_norm": 13.097331078127143, + "learning_rate": 8.938161640541202e-06, + "loss": 1.1075, + "step": 5046 + }, + { + "epoch": 2.739956568946797, + "grad_norm": 17.458162898454454, + "learning_rate": 8.934664749255524e-06, + "loss": 1.0174, + "step": 5047 + }, + { + "epoch": 2.740499457111835, + "grad_norm": 7.51493556495456, + "learning_rate": 8.931167989732212e-06, + "loss": 0.6407, + "step": 5048 + }, + { + "epoch": 2.741042345276873, + "grad_norm": 11.717405055058808, + "learning_rate": 8.927671362403741e-06, + "loss": 0.6899, + "step": 5049 + }, + { + "epoch": 2.741585233441911, + "grad_norm": 11.433736258508228, + "learning_rate": 8.924174867702591e-06, + "loss": 0.8481, + "step": 5050 + }, + { + "epoch": 2.742128121606949, + "grad_norm": 11.800039466500984, + "learning_rate": 8.920678506061202e-06, + "loss": 0.6044, + "step": 5051 + }, + { + "epoch": 2.742671009771987, + "grad_norm": 8.499003916350551, + "learning_rate": 8.91718227791202e-06, + "loss": 0.475, + "step": 5052 + }, + { + "epoch": 2.743213897937025, + "grad_norm": 10.586652955325345, + "learning_rate": 8.913686183687459e-06, + "loss": 0.5853, + "step": 5053 + }, + { + "epoch": 2.743756786102063, + "grad_norm": 14.066921486210767, + "learning_rate": 8.910190223819919e-06, + "loss": 0.8164, + "step": 5054 + }, + { + "epoch": 2.744299674267101, + "grad_norm": 9.77990149419258, + "learning_rate": 8.906694398741792e-06, + "loss": 0.6053, + "step": 5055 + }, + { + "epoch": 2.744842562432139, + "grad_norm": 15.328290145831401, + "learning_rate": 8.903198708885442e-06, + "loss": 1.3555, + "step": 5056 + }, + { + "epoch": 2.745385450597177, + "grad_norm": 11.626317894758118, + "learning_rate": 8.899703154683228e-06, + "loss": 0.6692, + "step": 5057 + }, + { + "epoch": 2.745928338762215, + "grad_norm": 11.566102899846406, + "learning_rate": 8.896207736567476e-06, + "loss": 0.6589, + "step": 5058 + }, + { + "epoch": 2.746471226927253, + "grad_norm": 9.521589192518187, + "learning_rate": 8.892712454970512e-06, + "loss": 0.7559, + "step": 5059 + }, + { + "epoch": 2.747014115092291, + "grad_norm": 8.719647496284242, + "learning_rate": 8.889217310324636e-06, + "loss": 0.4123, + "step": 5060 + }, + { + "epoch": 2.747557003257329, + "grad_norm": 6.709750992678039, + "learning_rate": 8.885722303062136e-06, + "loss": 0.3492, + "step": 5061 + }, + { + "epoch": 2.748099891422367, + "grad_norm": 11.273080989021693, + "learning_rate": 8.882227433615275e-06, + "loss": 0.7416, + "step": 5062 + }, + { + "epoch": 2.748642779587405, + "grad_norm": 13.864388014548679, + "learning_rate": 8.87873270241631e-06, + "loss": 0.8197, + "step": 5063 + }, + { + "epoch": 2.749185667752443, + "grad_norm": 9.492558149681301, + "learning_rate": 8.875238109897468e-06, + "loss": 0.6517, + "step": 5064 + }, + { + "epoch": 2.749728555917481, + "grad_norm": 10.634433093222732, + "learning_rate": 8.87174365649097e-06, + "loss": 0.5431, + "step": 5065 + }, + { + "epoch": 2.750271444082519, + "grad_norm": 8.469745230774542, + "learning_rate": 8.868249342629015e-06, + "loss": 0.5189, + "step": 5066 + }, + { + "epoch": 2.750814332247557, + "grad_norm": 8.182617911488233, + "learning_rate": 8.864755168743783e-06, + "loss": 0.4322, + "step": 5067 + }, + { + "epoch": 2.751357220412595, + "grad_norm": 7.6335690540689685, + "learning_rate": 8.861261135267444e-06, + "loss": 0.5268, + "step": 5068 + }, + { + "epoch": 2.751900108577633, + "grad_norm": 11.52493444073719, + "learning_rate": 8.85776724263214e-06, + "loss": 0.9757, + "step": 5069 + }, + { + "epoch": 2.752442996742671, + "grad_norm": 9.591556669015672, + "learning_rate": 8.854273491270008e-06, + "loss": 0.5527, + "step": 5070 + }, + { + "epoch": 2.752985884907709, + "grad_norm": 11.178077098634855, + "learning_rate": 8.850779881613151e-06, + "loss": 0.581, + "step": 5071 + }, + { + "epoch": 2.753528773072747, + "grad_norm": 8.921814933147093, + "learning_rate": 8.847286414093673e-06, + "loss": 0.5604, + "step": 5072 + }, + { + "epoch": 2.754071661237785, + "grad_norm": 11.887605361931898, + "learning_rate": 8.84379308914365e-06, + "loss": 0.6055, + "step": 5073 + }, + { + "epoch": 2.754614549402823, + "grad_norm": 14.12418845494581, + "learning_rate": 8.840299907195137e-06, + "loss": 0.9614, + "step": 5074 + }, + { + "epoch": 2.755157437567861, + "grad_norm": 10.094252933928091, + "learning_rate": 8.836806868680185e-06, + "loss": 0.6836, + "step": 5075 + }, + { + "epoch": 2.755700325732899, + "grad_norm": 10.048762466683991, + "learning_rate": 8.833313974030807e-06, + "loss": 0.5722, + "step": 5076 + }, + { + "epoch": 2.756243213897937, + "grad_norm": 12.51909069586841, + "learning_rate": 8.829821223679022e-06, + "loss": 1.2238, + "step": 5077 + }, + { + "epoch": 2.756786102062975, + "grad_norm": 9.50434691504725, + "learning_rate": 8.826328618056808e-06, + "loss": 0.6393, + "step": 5078 + }, + { + "epoch": 2.757328990228013, + "grad_norm": 10.737059961404412, + "learning_rate": 8.822836157596141e-06, + "loss": 0.7401, + "step": 5079 + }, + { + "epoch": 2.757871878393051, + "grad_norm": 9.16480609854369, + "learning_rate": 8.819343842728976e-06, + "loss": 0.5289, + "step": 5080 + }, + { + "epoch": 2.758414766558089, + "grad_norm": 10.849970138344764, + "learning_rate": 8.815851673887248e-06, + "loss": 0.6157, + "step": 5081 + }, + { + "epoch": 2.758957654723127, + "grad_norm": 8.609887529945185, + "learning_rate": 8.812359651502872e-06, + "loss": 0.5961, + "step": 5082 + }, + { + "epoch": 2.759500542888165, + "grad_norm": 10.765964781119916, + "learning_rate": 8.808867776007745e-06, + "loss": 1.1942, + "step": 5083 + }, + { + "epoch": 2.760043431053203, + "grad_norm": 10.466671227569803, + "learning_rate": 8.80537604783375e-06, + "loss": 0.7117, + "step": 5084 + }, + { + "epoch": 2.760586319218241, + "grad_norm": 10.079228354003783, + "learning_rate": 8.801884467412747e-06, + "loss": 0.4238, + "step": 5085 + }, + { + "epoch": 2.761129207383279, + "grad_norm": 11.422859396303908, + "learning_rate": 8.798393035176588e-06, + "loss": 0.8841, + "step": 5086 + }, + { + "epoch": 2.761672095548317, + "grad_norm": 9.736302153055926, + "learning_rate": 8.79490175155709e-06, + "loss": 0.6414, + "step": 5087 + }, + { + "epoch": 2.762214983713355, + "grad_norm": 8.513897139101958, + "learning_rate": 8.791410616986067e-06, + "loss": 0.3052, + "step": 5088 + }, + { + "epoch": 2.762757871878393, + "grad_norm": 10.101572440599828, + "learning_rate": 8.787919631895301e-06, + "loss": 0.5305, + "step": 5089 + }, + { + "epoch": 2.763300760043431, + "grad_norm": 12.17319240146357, + "learning_rate": 8.784428796716571e-06, + "loss": 0.6585, + "step": 5090 + }, + { + "epoch": 2.763843648208469, + "grad_norm": 10.930394499045825, + "learning_rate": 8.78093811188162e-06, + "loss": 0.5984, + "step": 5091 + }, + { + "epoch": 2.764386536373507, + "grad_norm": 9.550475415502788, + "learning_rate": 8.77744757782219e-06, + "loss": 0.5634, + "step": 5092 + }, + { + "epoch": 2.764929424538545, + "grad_norm": 12.890085749205962, + "learning_rate": 8.773957194969993e-06, + "loss": 0.9274, + "step": 5093 + }, + { + "epoch": 2.765472312703583, + "grad_norm": 12.38000962026181, + "learning_rate": 8.77046696375672e-06, + "loss": 0.6988, + "step": 5094 + }, + { + "epoch": 2.766015200868621, + "grad_norm": 9.523692226826551, + "learning_rate": 8.766976884614056e-06, + "loss": 0.4856, + "step": 5095 + }, + { + "epoch": 2.766558089033659, + "grad_norm": 9.26829606100349, + "learning_rate": 8.763486957973652e-06, + "loss": 0.6863, + "step": 5096 + }, + { + "epoch": 2.767100977198697, + "grad_norm": 11.227870751984687, + "learning_rate": 8.759997184267155e-06, + "loss": 0.6344, + "step": 5097 + }, + { + "epoch": 2.767643865363735, + "grad_norm": 11.653609442391305, + "learning_rate": 8.756507563926182e-06, + "loss": 0.5884, + "step": 5098 + }, + { + "epoch": 2.768186753528773, + "grad_norm": 11.59192980688709, + "learning_rate": 8.753018097382336e-06, + "loss": 0.8214, + "step": 5099 + }, + { + "epoch": 2.768729641693811, + "grad_norm": 8.668298021688548, + "learning_rate": 8.749528785067196e-06, + "loss": 0.6487, + "step": 5100 + }, + { + "epoch": 2.769272529858849, + "grad_norm": 5.456725122843202, + "learning_rate": 8.746039627412333e-06, + "loss": 0.2868, + "step": 5101 + }, + { + "epoch": 2.769815418023887, + "grad_norm": 12.29265212176352, + "learning_rate": 8.742550624849288e-06, + "loss": 0.512, + "step": 5102 + }, + { + "epoch": 2.770358306188925, + "grad_norm": 14.125650271042089, + "learning_rate": 8.73906177780958e-06, + "loss": 0.8903, + "step": 5103 + }, + { + "epoch": 2.770901194353963, + "grad_norm": 12.046023026774584, + "learning_rate": 8.735573086724725e-06, + "loss": 0.801, + "step": 5104 + }, + { + "epoch": 2.771444082519001, + "grad_norm": 10.153944591634119, + "learning_rate": 8.732084552026203e-06, + "loss": 0.6888, + "step": 5105 + }, + { + "epoch": 2.771986970684039, + "grad_norm": 10.581737293617769, + "learning_rate": 8.72859617414549e-06, + "loss": 0.7775, + "step": 5106 + }, + { + "epoch": 2.772529858849077, + "grad_norm": 8.304371234513637, + "learning_rate": 8.725107953514021e-06, + "loss": 0.4704, + "step": 5107 + }, + { + "epoch": 2.773072747014115, + "grad_norm": 12.579288268640527, + "learning_rate": 8.72161989056324e-06, + "loss": 0.5861, + "step": 5108 + }, + { + "epoch": 2.773615635179153, + "grad_norm": 10.68813978428288, + "learning_rate": 8.718131985724542e-06, + "loss": 0.6367, + "step": 5109 + }, + { + "epoch": 2.774158523344191, + "grad_norm": 8.921584649335454, + "learning_rate": 8.714644239429326e-06, + "loss": 0.5597, + "step": 5110 + }, + { + "epoch": 2.774701411509229, + "grad_norm": 10.473739328784887, + "learning_rate": 8.711156652108957e-06, + "loss": 0.8587, + "step": 5111 + }, + { + "epoch": 2.775244299674267, + "grad_norm": 10.905504725358476, + "learning_rate": 8.70766922419479e-06, + "loss": 0.5729, + "step": 5112 + }, + { + "epoch": 2.7757871878393052, + "grad_norm": 7.88310879226585, + "learning_rate": 8.704181956118153e-06, + "loss": 0.4998, + "step": 5113 + }, + { + "epoch": 2.776330076004343, + "grad_norm": 8.595679686197345, + "learning_rate": 8.700694848310354e-06, + "loss": 0.699, + "step": 5114 + }, + { + "epoch": 2.7768729641693812, + "grad_norm": 9.784504057993846, + "learning_rate": 8.697207901202691e-06, + "loss": 0.6509, + "step": 5115 + }, + { + "epoch": 2.777415852334419, + "grad_norm": 9.537890155200703, + "learning_rate": 8.693721115226427e-06, + "loss": 0.4484, + "step": 5116 + }, + { + "epoch": 2.7779587404994572, + "grad_norm": 10.110968424420188, + "learning_rate": 8.69023449081282e-06, + "loss": 0.6746, + "step": 5117 + }, + { + "epoch": 2.778501628664495, + "grad_norm": 14.837427287717714, + "learning_rate": 8.686748028393096e-06, + "loss": 1.3601, + "step": 5118 + }, + { + "epoch": 2.7790445168295332, + "grad_norm": 11.553555350476039, + "learning_rate": 8.683261728398472e-06, + "loss": 0.6054, + "step": 5119 + }, + { + "epoch": 2.779587404994571, + "grad_norm": 12.684609984344696, + "learning_rate": 8.679775591260132e-06, + "loss": 0.947, + "step": 5120 + }, + { + "epoch": 2.7801302931596092, + "grad_norm": 9.50144977278699, + "learning_rate": 8.676289617409256e-06, + "loss": 0.6401, + "step": 5121 + }, + { + "epoch": 2.780673181324647, + "grad_norm": 9.588585115620386, + "learning_rate": 8.672803807276988e-06, + "loss": 0.7214, + "step": 5122 + }, + { + "epoch": 2.7812160694896852, + "grad_norm": 14.597044499431716, + "learning_rate": 8.66931816129446e-06, + "loss": 1.0037, + "step": 5123 + }, + { + "epoch": 2.781758957654723, + "grad_norm": 10.552996924508232, + "learning_rate": 8.665832679892783e-06, + "loss": 0.5097, + "step": 5124 + }, + { + "epoch": 2.7823018458197613, + "grad_norm": 8.368460915764464, + "learning_rate": 8.662347363503043e-06, + "loss": 0.5289, + "step": 5125 + }, + { + "epoch": 2.782844733984799, + "grad_norm": 7.412774968306893, + "learning_rate": 8.658862212556318e-06, + "loss": 0.4117, + "step": 5126 + }, + { + "epoch": 2.7833876221498373, + "grad_norm": 7.965326482678092, + "learning_rate": 8.655377227483648e-06, + "loss": 0.4492, + "step": 5127 + }, + { + "epoch": 2.783930510314875, + "grad_norm": 10.16388699920939, + "learning_rate": 8.65189240871607e-06, + "loss": 0.6847, + "step": 5128 + }, + { + "epoch": 2.7844733984799133, + "grad_norm": 8.606167058304848, + "learning_rate": 8.648407756684582e-06, + "loss": 0.5952, + "step": 5129 + }, + { + "epoch": 2.785016286644951, + "grad_norm": 16.346182786192927, + "learning_rate": 8.64492327182018e-06, + "loss": 1.0626, + "step": 5130 + }, + { + "epoch": 2.7855591748099893, + "grad_norm": 9.397161435588243, + "learning_rate": 8.64143895455383e-06, + "loss": 0.6416, + "step": 5131 + }, + { + "epoch": 2.786102062975027, + "grad_norm": 10.215541582389136, + "learning_rate": 8.63795480531647e-06, + "loss": 0.5311, + "step": 5132 + }, + { + "epoch": 2.7866449511400653, + "grad_norm": 13.362954088182768, + "learning_rate": 8.634470824539035e-06, + "loss": 1.0249, + "step": 5133 + }, + { + "epoch": 2.787187839305103, + "grad_norm": 13.843076908870096, + "learning_rate": 8.630987012652421e-06, + "loss": 0.767, + "step": 5134 + }, + { + "epoch": 2.7877307274701413, + "grad_norm": 10.581731921492887, + "learning_rate": 8.627503370087519e-06, + "loss": 0.6027, + "step": 5135 + }, + { + "epoch": 2.788273615635179, + "grad_norm": 11.161057087658808, + "learning_rate": 8.624019897275184e-06, + "loss": 0.8193, + "step": 5136 + }, + { + "epoch": 2.7888165038002173, + "grad_norm": 7.972214509867566, + "learning_rate": 8.620536594646262e-06, + "loss": 0.4776, + "step": 5137 + }, + { + "epoch": 2.789359391965255, + "grad_norm": 12.57635432127455, + "learning_rate": 8.61705346263157e-06, + "loss": 0.7194, + "step": 5138 + }, + { + "epoch": 2.7899022801302933, + "grad_norm": 7.077115343788877, + "learning_rate": 8.613570501661915e-06, + "loss": 0.3571, + "step": 5139 + }, + { + "epoch": 2.790445168295331, + "grad_norm": 13.158625298591346, + "learning_rate": 8.610087712168065e-06, + "loss": 0.5872, + "step": 5140 + }, + { + "epoch": 2.7909880564603693, + "grad_norm": 8.652483474203775, + "learning_rate": 8.606605094580788e-06, + "loss": 0.624, + "step": 5141 + }, + { + "epoch": 2.791530944625407, + "grad_norm": 8.73276991998644, + "learning_rate": 8.603122649330811e-06, + "loss": 0.4358, + "step": 5142 + }, + { + "epoch": 2.7920738327904453, + "grad_norm": 12.749713905257105, + "learning_rate": 8.599640376848849e-06, + "loss": 0.7574, + "step": 5143 + }, + { + "epoch": 2.792616720955483, + "grad_norm": 12.054416741196698, + "learning_rate": 8.5961582775656e-06, + "loss": 0.886, + "step": 5144 + }, + { + "epoch": 2.7931596091205213, + "grad_norm": 10.958538797565106, + "learning_rate": 8.592676351911728e-06, + "loss": 0.7312, + "step": 5145 + }, + { + "epoch": 2.793702497285559, + "grad_norm": 10.124234819891349, + "learning_rate": 8.589194600317894e-06, + "loss": 0.5059, + "step": 5146 + }, + { + "epoch": 2.7942453854505973, + "grad_norm": 12.84934263063866, + "learning_rate": 8.585713023214715e-06, + "loss": 0.7016, + "step": 5147 + }, + { + "epoch": 2.794788273615635, + "grad_norm": 11.859965090765057, + "learning_rate": 8.582231621032807e-06, + "loss": 0.7158, + "step": 5148 + }, + { + "epoch": 2.7953311617806733, + "grad_norm": 11.57843340583913, + "learning_rate": 8.578750394202749e-06, + "loss": 0.7103, + "step": 5149 + }, + { + "epoch": 2.795874049945711, + "grad_norm": 7.143782607145388, + "learning_rate": 8.575269343155108e-06, + "loss": 0.5275, + "step": 5150 + }, + { + "epoch": 2.7964169381107493, + "grad_norm": 12.095108158973952, + "learning_rate": 8.571788468320427e-06, + "loss": 0.7308, + "step": 5151 + }, + { + "epoch": 2.796959826275787, + "grad_norm": 11.153004271537196, + "learning_rate": 8.568307770129223e-06, + "loss": 0.7576, + "step": 5152 + }, + { + "epoch": 2.7975027144408253, + "grad_norm": 8.326263454089252, + "learning_rate": 8.564827249011998e-06, + "loss": 0.408, + "step": 5153 + }, + { + "epoch": 2.798045602605863, + "grad_norm": 13.46902871500607, + "learning_rate": 8.561346905399221e-06, + "loss": 0.6624, + "step": 5154 + }, + { + "epoch": 2.7985884907709013, + "grad_norm": 8.925567749596054, + "learning_rate": 8.557866739721356e-06, + "loss": 0.4557, + "step": 5155 + }, + { + "epoch": 2.799131378935939, + "grad_norm": 11.700305884529763, + "learning_rate": 8.554386752408827e-06, + "loss": 0.6818, + "step": 5156 + }, + { + "epoch": 2.7996742671009773, + "grad_norm": 9.49767460620272, + "learning_rate": 8.550906943892054e-06, + "loss": 0.7256, + "step": 5157 + }, + { + "epoch": 2.800217155266015, + "grad_norm": 11.749650388906772, + "learning_rate": 8.547427314601416e-06, + "loss": 0.6933, + "step": 5158 + }, + { + "epoch": 2.8007600434310533, + "grad_norm": 11.830065328109, + "learning_rate": 8.543947864967286e-06, + "loss": 0.4282, + "step": 5159 + }, + { + "epoch": 2.801302931596091, + "grad_norm": 12.11100989733807, + "learning_rate": 8.540468595419999e-06, + "loss": 0.3617, + "step": 5160 + }, + { + "epoch": 2.8018458197611293, + "grad_norm": 9.923764364092053, + "learning_rate": 8.536989506389889e-06, + "loss": 0.6469, + "step": 5161 + }, + { + "epoch": 2.802388707926167, + "grad_norm": 10.555923377692814, + "learning_rate": 8.533510598307244e-06, + "loss": 0.9672, + "step": 5162 + }, + { + "epoch": 2.8029315960912053, + "grad_norm": 12.544866122081839, + "learning_rate": 8.530031871602345e-06, + "loss": 0.6885, + "step": 5163 + }, + { + "epoch": 2.803474484256243, + "grad_norm": 11.680102099467787, + "learning_rate": 8.526553326705452e-06, + "loss": 0.5571, + "step": 5164 + }, + { + "epoch": 2.8040173724212814, + "grad_norm": 11.769615311559438, + "learning_rate": 8.523074964046785e-06, + "loss": 0.6176, + "step": 5165 + }, + { + "epoch": 2.804560260586319, + "grad_norm": 10.249168095085986, + "learning_rate": 8.519596784056567e-06, + "loss": 0.7974, + "step": 5166 + }, + { + "epoch": 2.8051031487513574, + "grad_norm": 7.602995989085923, + "learning_rate": 8.516118787164973e-06, + "loss": 0.4132, + "step": 5167 + }, + { + "epoch": 2.805646036916395, + "grad_norm": 9.32078805379704, + "learning_rate": 8.512640973802175e-06, + "loss": 0.6492, + "step": 5168 + }, + { + "epoch": 2.8061889250814334, + "grad_norm": 13.80437475122657, + "learning_rate": 8.50916334439831e-06, + "loss": 0.6663, + "step": 5169 + }, + { + "epoch": 2.806731813246471, + "grad_norm": 9.121579447999185, + "learning_rate": 8.5056858993835e-06, + "loss": 0.4546, + "step": 5170 + }, + { + "epoch": 2.8072747014115094, + "grad_norm": 8.073186873408805, + "learning_rate": 8.502208639187842e-06, + "loss": 0.4467, + "step": 5171 + }, + { + "epoch": 2.807817589576547, + "grad_norm": 10.498644561860338, + "learning_rate": 8.498731564241403e-06, + "loss": 0.4907, + "step": 5172 + }, + { + "epoch": 2.8083604777415854, + "grad_norm": 9.027361205852865, + "learning_rate": 8.495254674974239e-06, + "loss": 0.4581, + "step": 5173 + }, + { + "epoch": 2.808903365906623, + "grad_norm": 9.692775527105594, + "learning_rate": 8.491777971816372e-06, + "loss": 0.5898, + "step": 5174 + }, + { + "epoch": 2.8094462540716614, + "grad_norm": 9.171182553280271, + "learning_rate": 8.48830145519781e-06, + "loss": 0.4129, + "step": 5175 + }, + { + "epoch": 2.809989142236699, + "grad_norm": 11.635096708230561, + "learning_rate": 8.484825125548532e-06, + "loss": 0.7632, + "step": 5176 + }, + { + "epoch": 2.8105320304017374, + "grad_norm": 12.18174333972076, + "learning_rate": 8.481348983298503e-06, + "loss": 0.6158, + "step": 5177 + }, + { + "epoch": 2.811074918566775, + "grad_norm": 8.880576943532684, + "learning_rate": 8.477873028877645e-06, + "loss": 0.374, + "step": 5178 + }, + { + "epoch": 2.8116178067318134, + "grad_norm": 11.735757011642823, + "learning_rate": 8.474397262715884e-06, + "loss": 0.6772, + "step": 5179 + }, + { + "epoch": 2.812160694896851, + "grad_norm": 9.50535973508058, + "learning_rate": 8.470921685243098e-06, + "loss": 0.6021, + "step": 5180 + }, + { + "epoch": 2.8127035830618894, + "grad_norm": 11.780091573431722, + "learning_rate": 8.467446296889151e-06, + "loss": 0.6944, + "step": 5181 + }, + { + "epoch": 2.813246471226927, + "grad_norm": 10.177875512109438, + "learning_rate": 8.463971098083896e-06, + "loss": 0.4312, + "step": 5182 + }, + { + "epoch": 2.8137893593919654, + "grad_norm": 16.588608629478436, + "learning_rate": 8.460496089257136e-06, + "loss": 0.818, + "step": 5183 + }, + { + "epoch": 2.814332247557003, + "grad_norm": 12.416295723031903, + "learning_rate": 8.45702127083868e-06, + "loss": 0.5531, + "step": 5184 + }, + { + "epoch": 2.8148751357220414, + "grad_norm": 13.657036134609482, + "learning_rate": 8.453546643258286e-06, + "loss": 0.7343, + "step": 5185 + }, + { + "epoch": 2.815418023887079, + "grad_norm": 9.833034529955514, + "learning_rate": 8.450072206945715e-06, + "loss": 0.6592, + "step": 5186 + }, + { + "epoch": 2.8159609120521174, + "grad_norm": 10.036112921946827, + "learning_rate": 8.446597962330675e-06, + "loss": 0.8127, + "step": 5187 + }, + { + "epoch": 2.816503800217155, + "grad_norm": 12.901123274951528, + "learning_rate": 8.443123909842882e-06, + "loss": 1.0507, + "step": 5188 + }, + { + "epoch": 2.8170466883821934, + "grad_norm": 10.692152694001132, + "learning_rate": 8.439650049911999e-06, + "loss": 0.5602, + "step": 5189 + }, + { + "epoch": 2.817589576547231, + "grad_norm": 19.731345695292617, + "learning_rate": 8.436176382967692e-06, + "loss": 1.0473, + "step": 5190 + }, + { + "epoch": 2.8181324647122694, + "grad_norm": 14.567179107047348, + "learning_rate": 8.432702909439579e-06, + "loss": 0.6359, + "step": 5191 + }, + { + "epoch": 2.818675352877307, + "grad_norm": 13.853200529175792, + "learning_rate": 8.429229629757266e-06, + "loss": 0.8277, + "step": 5192 + }, + { + "epoch": 2.8192182410423454, + "grad_norm": 9.811701731637578, + "learning_rate": 8.425756544350338e-06, + "loss": 0.533, + "step": 5193 + }, + { + "epoch": 2.819761129207383, + "grad_norm": 11.486229326022617, + "learning_rate": 8.422283653648348e-06, + "loss": 0.9078, + "step": 5194 + }, + { + "epoch": 2.8203040173724214, + "grad_norm": 11.086905618662822, + "learning_rate": 8.418810958080832e-06, + "loss": 0.7101, + "step": 5195 + }, + { + "epoch": 2.820846905537459, + "grad_norm": 8.065915758013317, + "learning_rate": 8.415338458077293e-06, + "loss": 0.5744, + "step": 5196 + }, + { + "epoch": 2.8213897937024974, + "grad_norm": 9.063098353131359, + "learning_rate": 8.411866154067224e-06, + "loss": 0.5008, + "step": 5197 + }, + { + "epoch": 2.821932681867535, + "grad_norm": 12.457981789036095, + "learning_rate": 8.408394046480077e-06, + "loss": 0.6377, + "step": 5198 + }, + { + "epoch": 2.8224755700325734, + "grad_norm": 10.806638216082748, + "learning_rate": 8.404922135745295e-06, + "loss": 0.6137, + "step": 5199 + }, + { + "epoch": 2.823018458197611, + "grad_norm": 10.78339089292669, + "learning_rate": 8.401450422292281e-06, + "loss": 0.7984, + "step": 5200 + }, + { + "epoch": 2.8235613463626494, + "grad_norm": 9.574355357184858, + "learning_rate": 8.397978906550429e-06, + "loss": 0.6307, + "step": 5201 + }, + { + "epoch": 2.824104234527687, + "grad_norm": 9.457238410125305, + "learning_rate": 8.3945075889491e-06, + "loss": 0.6782, + "step": 5202 + }, + { + "epoch": 2.8246471226927254, + "grad_norm": 15.132207636686802, + "learning_rate": 8.391036469917626e-06, + "loss": 0.7843, + "step": 5203 + }, + { + "epoch": 2.8251900108577632, + "grad_norm": 12.425847301998207, + "learning_rate": 8.387565549885331e-06, + "loss": 0.9953, + "step": 5204 + }, + { + "epoch": 2.8257328990228014, + "grad_norm": 9.948899666849417, + "learning_rate": 8.384094829281495e-06, + "loss": 0.6601, + "step": 5205 + }, + { + "epoch": 2.8262757871878392, + "grad_norm": 9.412819573011207, + "learning_rate": 8.38062430853539e-06, + "loss": 0.3747, + "step": 5206 + }, + { + "epoch": 2.8268186753528775, + "grad_norm": 8.137925073552552, + "learning_rate": 8.377153988076247e-06, + "loss": 0.3818, + "step": 5207 + }, + { + "epoch": 2.8273615635179152, + "grad_norm": 8.994244424095461, + "learning_rate": 8.373683868333287e-06, + "loss": 0.796, + "step": 5208 + }, + { + "epoch": 2.8279044516829535, + "grad_norm": 7.570203929319876, + "learning_rate": 8.370213949735696e-06, + "loss": 0.5587, + "step": 5209 + }, + { + "epoch": 2.8284473398479912, + "grad_norm": 13.134200586671845, + "learning_rate": 8.366744232712645e-06, + "loss": 0.8274, + "step": 5210 + }, + { + "epoch": 2.8289902280130295, + "grad_norm": 9.833443120274195, + "learning_rate": 8.363274717693272e-06, + "loss": 0.7984, + "step": 5211 + }, + { + "epoch": 2.8295331161780672, + "grad_norm": 11.623259569780537, + "learning_rate": 8.359805405106685e-06, + "loss": 0.7305, + "step": 5212 + }, + { + "epoch": 2.8300760043431055, + "grad_norm": 10.494299983717905, + "learning_rate": 8.356336295381981e-06, + "loss": 0.6299, + "step": 5213 + }, + { + "epoch": 2.8306188925081432, + "grad_norm": 7.662506144803812, + "learning_rate": 8.352867388948223e-06, + "loss": 0.4432, + "step": 5214 + }, + { + "epoch": 2.8311617806731815, + "grad_norm": 11.250218590721435, + "learning_rate": 8.349398686234455e-06, + "loss": 0.5805, + "step": 5215 + }, + { + "epoch": 2.8317046688382193, + "grad_norm": 12.63080713053447, + "learning_rate": 8.345930187669685e-06, + "loss": 0.7236, + "step": 5216 + }, + { + "epoch": 2.8322475570032575, + "grad_norm": 12.902831533172366, + "learning_rate": 8.342461893682908e-06, + "loss": 0.6462, + "step": 5217 + }, + { + "epoch": 2.8327904451682953, + "grad_norm": 8.440334577273788, + "learning_rate": 8.338993804703082e-06, + "loss": 0.6557, + "step": 5218 + }, + { + "epoch": 2.8333333333333335, + "grad_norm": 8.946600771683327, + "learning_rate": 8.335525921159155e-06, + "loss": 0.4964, + "step": 5219 + }, + { + "epoch": 2.8338762214983713, + "grad_norm": 11.427720200073038, + "learning_rate": 8.332058243480032e-06, + "loss": 0.5832, + "step": 5220 + }, + { + "epoch": 2.8344191096634095, + "grad_norm": 10.663320470599537, + "learning_rate": 8.328590772094602e-06, + "loss": 0.385, + "step": 5221 + }, + { + "epoch": 2.8349619978284473, + "grad_norm": 10.091874752226794, + "learning_rate": 8.325123507431732e-06, + "loss": 0.8347, + "step": 5222 + }, + { + "epoch": 2.8355048859934855, + "grad_norm": 11.844959588854952, + "learning_rate": 8.321656449920252e-06, + "loss": 0.5577, + "step": 5223 + }, + { + "epoch": 2.8360477741585233, + "grad_norm": 10.895554793526724, + "learning_rate": 8.318189599988982e-06, + "loss": 0.7961, + "step": 5224 + }, + { + "epoch": 2.8365906623235615, + "grad_norm": 15.457894886801157, + "learning_rate": 8.314722958066699e-06, + "loss": 0.9784, + "step": 5225 + }, + { + "epoch": 2.8371335504885993, + "grad_norm": 9.287014113188532, + "learning_rate": 8.311256524582167e-06, + "loss": 0.5947, + "step": 5226 + }, + { + "epoch": 2.8376764386536375, + "grad_norm": 12.953808801302275, + "learning_rate": 8.307790299964119e-06, + "loss": 0.6073, + "step": 5227 + }, + { + "epoch": 2.8382193268186753, + "grad_norm": 12.53202947557571, + "learning_rate": 8.304324284641268e-06, + "loss": 0.8199, + "step": 5228 + }, + { + "epoch": 2.8387622149837135, + "grad_norm": 10.62038819495681, + "learning_rate": 8.300858479042291e-06, + "loss": 0.488, + "step": 5229 + }, + { + "epoch": 2.8393051031487513, + "grad_norm": 10.914681492949684, + "learning_rate": 8.29739288359584e-06, + "loss": 0.6159, + "step": 5230 + }, + { + "epoch": 2.8398479913137895, + "grad_norm": 9.514892864934003, + "learning_rate": 8.293927498730556e-06, + "loss": 0.8757, + "step": 5231 + }, + { + "epoch": 2.8403908794788273, + "grad_norm": 10.004218198768447, + "learning_rate": 8.290462324875036e-06, + "loss": 0.4543, + "step": 5232 + }, + { + "epoch": 2.8409337676438655, + "grad_norm": 7.955044675492623, + "learning_rate": 8.286997362457859e-06, + "loss": 0.3227, + "step": 5233 + }, + { + "epoch": 2.8414766558089033, + "grad_norm": 5.674531662978047, + "learning_rate": 8.283532611907577e-06, + "loss": 0.2674, + "step": 5234 + }, + { + "epoch": 2.8420195439739415, + "grad_norm": 10.825015665199876, + "learning_rate": 8.280068073652723e-06, + "loss": 0.7373, + "step": 5235 + }, + { + "epoch": 2.8425624321389793, + "grad_norm": 11.937523350465815, + "learning_rate": 8.276603748121786e-06, + "loss": 0.6019, + "step": 5236 + }, + { + "epoch": 2.8431053203040175, + "grad_norm": 11.660316243586271, + "learning_rate": 8.273139635743249e-06, + "loss": 0.8063, + "step": 5237 + }, + { + "epoch": 2.8436482084690553, + "grad_norm": 10.978384303044228, + "learning_rate": 8.269675736945547e-06, + "loss": 0.6178, + "step": 5238 + }, + { + "epoch": 2.8441910966340935, + "grad_norm": 10.489917172678325, + "learning_rate": 8.266212052157113e-06, + "loss": 0.4382, + "step": 5239 + }, + { + "epoch": 2.8447339847991313, + "grad_norm": 9.430697797795169, + "learning_rate": 8.262748581806336e-06, + "loss": 0.4811, + "step": 5240 + }, + { + "epoch": 2.8452768729641695, + "grad_norm": 13.552717272826865, + "learning_rate": 8.259285326321579e-06, + "loss": 0.6866, + "step": 5241 + }, + { + "epoch": 2.8458197611292073, + "grad_norm": 12.883094545932988, + "learning_rate": 8.25582228613119e-06, + "loss": 0.9637, + "step": 5242 + }, + { + "epoch": 2.8463626492942455, + "grad_norm": 18.04770906924292, + "learning_rate": 8.25235946166348e-06, + "loss": 0.801, + "step": 5243 + }, + { + "epoch": 2.8469055374592833, + "grad_norm": 12.482559439953985, + "learning_rate": 8.248896853346738e-06, + "loss": 0.7968, + "step": 5244 + }, + { + "epoch": 2.8474484256243215, + "grad_norm": 10.728211626455765, + "learning_rate": 8.245434461609221e-06, + "loss": 0.5192, + "step": 5245 + }, + { + "epoch": 2.8479913137893593, + "grad_norm": 10.793312471383818, + "learning_rate": 8.241972286879168e-06, + "loss": 0.6121, + "step": 5246 + }, + { + "epoch": 2.8485342019543975, + "grad_norm": 12.822804552881774, + "learning_rate": 8.238510329584782e-06, + "loss": 0.7807, + "step": 5247 + }, + { + "epoch": 2.8490770901194353, + "grad_norm": 14.208687960885756, + "learning_rate": 8.23504859015425e-06, + "loss": 0.73, + "step": 5248 + }, + { + "epoch": 2.8496199782844736, + "grad_norm": 13.561299495458854, + "learning_rate": 8.231587069015723e-06, + "loss": 1.269, + "step": 5249 + }, + { + "epoch": 2.8501628664495113, + "grad_norm": 12.542912517352129, + "learning_rate": 8.22812576659732e-06, + "loss": 0.5758, + "step": 5250 + }, + { + "epoch": 2.8507057546145496, + "grad_norm": 9.184884000564415, + "learning_rate": 8.22466468332715e-06, + "loss": 0.5323, + "step": 5251 + }, + { + "epoch": 2.8512486427795873, + "grad_norm": 10.234457404263107, + "learning_rate": 8.22120381963328e-06, + "loss": 0.5455, + "step": 5252 + }, + { + "epoch": 2.8517915309446256, + "grad_norm": 13.327787107696029, + "learning_rate": 8.217743175943756e-06, + "loss": 0.6097, + "step": 5253 + }, + { + "epoch": 2.8523344191096633, + "grad_norm": 10.375306748712186, + "learning_rate": 8.214282752686595e-06, + "loss": 0.9622, + "step": 5254 + }, + { + "epoch": 2.8528773072747016, + "grad_norm": 10.445896680603816, + "learning_rate": 8.210822550289794e-06, + "loss": 0.6347, + "step": 5255 + }, + { + "epoch": 2.8534201954397393, + "grad_norm": 8.375912125960554, + "learning_rate": 8.207362569181305e-06, + "loss": 0.419, + "step": 5256 + }, + { + "epoch": 2.8539630836047776, + "grad_norm": 12.786819256474223, + "learning_rate": 8.203902809789078e-06, + "loss": 1.0155, + "step": 5257 + }, + { + "epoch": 2.8545059717698154, + "grad_norm": 9.916362057542294, + "learning_rate": 8.200443272541007e-06, + "loss": 0.5501, + "step": 5258 + }, + { + "epoch": 2.8550488599348536, + "grad_norm": 12.303081851739867, + "learning_rate": 8.196983957864984e-06, + "loss": 0.5965, + "step": 5259 + }, + { + "epoch": 2.8555917480998914, + "grad_norm": 11.992761200744544, + "learning_rate": 8.19352486618886e-06, + "loss": 0.4804, + "step": 5260 + }, + { + "epoch": 2.8561346362649296, + "grad_norm": 11.342631954988665, + "learning_rate": 8.190065997940455e-06, + "loss": 0.6241, + "step": 5261 + }, + { + "epoch": 2.8566775244299674, + "grad_norm": 14.06120317718869, + "learning_rate": 8.186607353547578e-06, + "loss": 1.3883, + "step": 5262 + }, + { + "epoch": 2.8572204125950056, + "grad_norm": 9.985876763230308, + "learning_rate": 8.183148933437988e-06, + "loss": 0.6843, + "step": 5263 + }, + { + "epoch": 2.8577633007600434, + "grad_norm": 13.658693433537799, + "learning_rate": 8.179690738039436e-06, + "loss": 1.027, + "step": 5264 + }, + { + "epoch": 2.8583061889250816, + "grad_norm": 13.070549937452688, + "learning_rate": 8.176232767779634e-06, + "loss": 0.9171, + "step": 5265 + }, + { + "epoch": 2.8588490770901194, + "grad_norm": 10.202999590557772, + "learning_rate": 8.17277502308627e-06, + "loss": 0.6412, + "step": 5266 + }, + { + "epoch": 2.8593919652551576, + "grad_norm": 11.358682840396053, + "learning_rate": 8.169317504387002e-06, + "loss": 0.8102, + "step": 5267 + }, + { + "epoch": 2.8599348534201954, + "grad_norm": 12.795930497437475, + "learning_rate": 8.165860212109467e-06, + "loss": 0.7885, + "step": 5268 + }, + { + "epoch": 2.8604777415852336, + "grad_norm": 12.540977748153804, + "learning_rate": 8.162403146681263e-06, + "loss": 0.453, + "step": 5269 + }, + { + "epoch": 2.8610206297502714, + "grad_norm": 9.835742357069565, + "learning_rate": 8.158946308529964e-06, + "loss": 0.553, + "step": 5270 + }, + { + "epoch": 2.8615635179153096, + "grad_norm": 8.768245847600197, + "learning_rate": 8.15548969808312e-06, + "loss": 0.6084, + "step": 5271 + }, + { + "epoch": 2.8621064060803474, + "grad_norm": 11.27517155767251, + "learning_rate": 8.152033315768248e-06, + "loss": 0.708, + "step": 5272 + }, + { + "epoch": 2.8626492942453856, + "grad_norm": 10.175542619256326, + "learning_rate": 8.148577162012848e-06, + "loss": 0.567, + "step": 5273 + }, + { + "epoch": 2.8631921824104234, + "grad_norm": 13.49575371946775, + "learning_rate": 8.145121237244367e-06, + "loss": 0.7178, + "step": 5274 + }, + { + "epoch": 2.8637350705754616, + "grad_norm": 12.848120779354371, + "learning_rate": 8.141665541890254e-06, + "loss": 0.7601, + "step": 5275 + }, + { + "epoch": 2.8642779587404994, + "grad_norm": 9.660459046348308, + "learning_rate": 8.138210076377903e-06, + "loss": 0.5867, + "step": 5276 + }, + { + "epoch": 2.8648208469055376, + "grad_norm": 12.014506663974018, + "learning_rate": 8.134754841134704e-06, + "loss": 0.6157, + "step": 5277 + }, + { + "epoch": 2.8653637350705754, + "grad_norm": 9.006817793845983, + "learning_rate": 8.131299836587994e-06, + "loss": 0.417, + "step": 5278 + }, + { + "epoch": 2.8659066232356136, + "grad_norm": 14.075391102924492, + "learning_rate": 8.127845063165097e-06, + "loss": 0.9453, + "step": 5279 + }, + { + "epoch": 2.8664495114006514, + "grad_norm": 11.384071845529691, + "learning_rate": 8.124390521293311e-06, + "loss": 0.757, + "step": 5280 + }, + { + "epoch": 2.8669923995656896, + "grad_norm": 11.412348591705651, + "learning_rate": 8.12093621139989e-06, + "loss": 0.7149, + "step": 5281 + }, + { + "epoch": 2.8675352877307274, + "grad_norm": 10.533279510688539, + "learning_rate": 8.117482133912078e-06, + "loss": 0.4981, + "step": 5282 + }, + { + "epoch": 2.8680781758957656, + "grad_norm": 10.800216526898991, + "learning_rate": 8.114028289257071e-06, + "loss": 0.6033, + "step": 5283 + }, + { + "epoch": 2.8686210640608034, + "grad_norm": 10.834462944873792, + "learning_rate": 8.110574677862054e-06, + "loss": 0.5554, + "step": 5284 + }, + { + "epoch": 2.8691639522258416, + "grad_norm": 13.02642177423764, + "learning_rate": 8.10712130015417e-06, + "loss": 0.7303, + "step": 5285 + }, + { + "epoch": 2.8697068403908794, + "grad_norm": 13.210484866681897, + "learning_rate": 8.103668156560543e-06, + "loss": 0.9579, + "step": 5286 + }, + { + "epoch": 2.8702497285559176, + "grad_norm": 7.866855822388311, + "learning_rate": 8.100215247508259e-06, + "loss": 0.5878, + "step": 5287 + }, + { + "epoch": 2.8707926167209554, + "grad_norm": 10.857793278838274, + "learning_rate": 8.096762573424384e-06, + "loss": 0.6734, + "step": 5288 + }, + { + "epoch": 2.8713355048859937, + "grad_norm": 11.519175454268337, + "learning_rate": 8.093310134735945e-06, + "loss": 0.7203, + "step": 5289 + }, + { + "epoch": 2.8718783930510314, + "grad_norm": 12.16192132511212, + "learning_rate": 8.089857931869947e-06, + "loss": 0.5887, + "step": 5290 + }, + { + "epoch": 2.8724212812160697, + "grad_norm": 8.545914428475232, + "learning_rate": 8.086405965253364e-06, + "loss": 0.677, + "step": 5291 + }, + { + "epoch": 2.8729641693811074, + "grad_norm": 12.463412130388328, + "learning_rate": 8.08295423531314e-06, + "loss": 0.8954, + "step": 5292 + }, + { + "epoch": 2.8735070575461457, + "grad_norm": 10.296132022290973, + "learning_rate": 8.079502742476195e-06, + "loss": 0.6582, + "step": 5293 + }, + { + "epoch": 2.8740499457111834, + "grad_norm": 9.170071716678386, + "learning_rate": 8.076051487169407e-06, + "loss": 0.5406, + "step": 5294 + }, + { + "epoch": 2.8745928338762217, + "grad_norm": 14.43641149914591, + "learning_rate": 8.072600469819643e-06, + "loss": 0.7459, + "step": 5295 + }, + { + "epoch": 2.8751357220412594, + "grad_norm": 9.705510058137687, + "learning_rate": 8.069149690853719e-06, + "loss": 0.6341, + "step": 5296 + }, + { + "epoch": 2.8756786102062977, + "grad_norm": 11.720382904656155, + "learning_rate": 8.065699150698442e-06, + "loss": 0.6016, + "step": 5297 + }, + { + "epoch": 2.8762214983713354, + "grad_norm": 14.838955909579008, + "learning_rate": 8.062248849780578e-06, + "loss": 0.8757, + "step": 5298 + }, + { + "epoch": 2.8767643865363732, + "grad_norm": 10.594386765695665, + "learning_rate": 8.05879878852686e-06, + "loss": 0.4864, + "step": 5299 + }, + { + "epoch": 2.8773072747014115, + "grad_norm": 9.83553313698661, + "learning_rate": 8.055348967364007e-06, + "loss": 0.5116, + "step": 5300 + }, + { + "epoch": 2.8778501628664497, + "grad_norm": 12.500251504139728, + "learning_rate": 8.05189938671869e-06, + "loss": 0.891, + "step": 5301 + }, + { + "epoch": 2.8783930510314875, + "grad_norm": 8.988571871064325, + "learning_rate": 8.048450047017563e-06, + "loss": 0.3699, + "step": 5302 + }, + { + "epoch": 2.8789359391965252, + "grad_norm": 10.893701604379793, + "learning_rate": 8.045000948687242e-06, + "loss": 0.5933, + "step": 5303 + }, + { + "epoch": 2.8794788273615635, + "grad_norm": 10.474072412748654, + "learning_rate": 8.041552092154321e-06, + "loss": 0.8787, + "step": 5304 + }, + { + "epoch": 2.8800217155266017, + "grad_norm": 8.555744151752272, + "learning_rate": 8.038103477845357e-06, + "loss": 0.3793, + "step": 5305 + }, + { + "epoch": 2.8805646036916395, + "grad_norm": 9.792363571958322, + "learning_rate": 8.034655106186884e-06, + "loss": 0.5347, + "step": 5306 + }, + { + "epoch": 2.8811074918566772, + "grad_norm": 8.810561156197549, + "learning_rate": 8.031206977605399e-06, + "loss": 0.5128, + "step": 5307 + }, + { + "epoch": 2.8816503800217155, + "grad_norm": 12.383361133227218, + "learning_rate": 8.027759092527374e-06, + "loss": 0.7834, + "step": 5308 + }, + { + "epoch": 2.8821932681867537, + "grad_norm": 9.51185389588206, + "learning_rate": 8.024311451379247e-06, + "loss": 0.4661, + "step": 5309 + }, + { + "epoch": 2.8827361563517915, + "grad_norm": 10.564107506597352, + "learning_rate": 8.020864054587426e-06, + "loss": 0.8199, + "step": 5310 + }, + { + "epoch": 2.8832790445168293, + "grad_norm": 14.579672874726727, + "learning_rate": 8.017416902578296e-06, + "loss": 1.1114, + "step": 5311 + }, + { + "epoch": 2.8838219326818675, + "grad_norm": 10.739540875321001, + "learning_rate": 8.013969995778201e-06, + "loss": 0.5384, + "step": 5312 + }, + { + "epoch": 2.8843648208469057, + "grad_norm": 11.955074827477672, + "learning_rate": 8.010523334613466e-06, + "loss": 0.8475, + "step": 5313 + }, + { + "epoch": 2.8849077090119435, + "grad_norm": 12.842414101367579, + "learning_rate": 8.007076919510371e-06, + "loss": 0.7096, + "step": 5314 + }, + { + "epoch": 2.8854505971769813, + "grad_norm": 9.17766217916239, + "learning_rate": 8.003630750895183e-06, + "loss": 0.4673, + "step": 5315 + }, + { + "epoch": 2.8859934853420195, + "grad_norm": 13.581449358552742, + "learning_rate": 8.000184829194121e-06, + "loss": 0.7501, + "step": 5316 + }, + { + "epoch": 2.8865363735070577, + "grad_norm": 6.157129192897698, + "learning_rate": 7.99673915483339e-06, + "loss": 0.2521, + "step": 5317 + }, + { + "epoch": 2.8870792616720955, + "grad_norm": 9.179432551374715, + "learning_rate": 7.993293728239154e-06, + "loss": 0.5682, + "step": 5318 + }, + { + "epoch": 2.8876221498371333, + "grad_norm": 8.458089104581813, + "learning_rate": 7.989848549837544e-06, + "loss": 0.7131, + "step": 5319 + }, + { + "epoch": 2.8881650380021715, + "grad_norm": 11.039294062728864, + "learning_rate": 7.986403620054674e-06, + "loss": 0.5421, + "step": 5320 + }, + { + "epoch": 2.8887079261672097, + "grad_norm": 8.296966992960693, + "learning_rate": 7.982958939316607e-06, + "loss": 0.5809, + "step": 5321 + }, + { + "epoch": 2.8892508143322475, + "grad_norm": 9.862357699114755, + "learning_rate": 7.979514508049398e-06, + "loss": 0.7253, + "step": 5322 + }, + { + "epoch": 2.8897937024972853, + "grad_norm": 10.60481802559805, + "learning_rate": 7.976070326679053e-06, + "loss": 0.5593, + "step": 5323 + }, + { + "epoch": 2.8903365906623235, + "grad_norm": 12.452210665582081, + "learning_rate": 7.972626395631556e-06, + "loss": 0.9987, + "step": 5324 + }, + { + "epoch": 2.8908794788273617, + "grad_norm": 8.161513982505882, + "learning_rate": 7.969182715332855e-06, + "loss": 0.5242, + "step": 5325 + }, + { + "epoch": 2.8914223669923995, + "grad_norm": 9.073053317988423, + "learning_rate": 7.965739286208878e-06, + "loss": 0.5015, + "step": 5326 + }, + { + "epoch": 2.8919652551574373, + "grad_norm": 8.632672314975157, + "learning_rate": 7.962296108685507e-06, + "loss": 0.4398, + "step": 5327 + }, + { + "epoch": 2.8925081433224755, + "grad_norm": 12.024378540292089, + "learning_rate": 7.958853183188597e-06, + "loss": 0.7876, + "step": 5328 + }, + { + "epoch": 2.8930510314875137, + "grad_norm": 9.943181911450774, + "learning_rate": 7.955410510143982e-06, + "loss": 0.6408, + "step": 5329 + }, + { + "epoch": 2.8935939196525515, + "grad_norm": 11.580662527673274, + "learning_rate": 7.951968089977449e-06, + "loss": 0.6484, + "step": 5330 + }, + { + "epoch": 2.8941368078175893, + "grad_norm": 10.477418271397871, + "learning_rate": 7.948525923114773e-06, + "loss": 0.7197, + "step": 5331 + }, + { + "epoch": 2.8946796959826275, + "grad_norm": 8.959473056544534, + "learning_rate": 7.945084009981674e-06, + "loss": 0.5344, + "step": 5332 + }, + { + "epoch": 2.8952225841476658, + "grad_norm": 7.339268198721023, + "learning_rate": 7.941642351003867e-06, + "loss": 0.4373, + "step": 5333 + }, + { + "epoch": 2.8957654723127035, + "grad_norm": 11.710220973199451, + "learning_rate": 7.938200946607008e-06, + "loss": 0.7354, + "step": 5334 + }, + { + "epoch": 2.8963083604777413, + "grad_norm": 10.548034804633037, + "learning_rate": 7.934759797216744e-06, + "loss": 1.0912, + "step": 5335 + }, + { + "epoch": 2.8968512486427795, + "grad_norm": 7.427458411514844, + "learning_rate": 7.93131890325868e-06, + "loss": 0.4515, + "step": 5336 + }, + { + "epoch": 2.8973941368078178, + "grad_norm": 7.5336162811756475, + "learning_rate": 7.927878265158391e-06, + "loss": 0.4106, + "step": 5337 + }, + { + "epoch": 2.8979370249728555, + "grad_norm": 9.845844831585431, + "learning_rate": 7.924437883341424e-06, + "loss": 0.394, + "step": 5338 + }, + { + "epoch": 2.8984799131378933, + "grad_norm": 11.91533381613983, + "learning_rate": 7.920997758233282e-06, + "loss": 0.5863, + "step": 5339 + }, + { + "epoch": 2.8990228013029316, + "grad_norm": 7.229828747210386, + "learning_rate": 7.917557890259457e-06, + "loss": 0.5663, + "step": 5340 + }, + { + "epoch": 2.8995656894679698, + "grad_norm": 11.378354320604084, + "learning_rate": 7.914118279845385e-06, + "loss": 0.7622, + "step": 5341 + }, + { + "epoch": 2.9001085776330076, + "grad_norm": 7.591103634044025, + "learning_rate": 7.91067892741649e-06, + "loss": 0.3933, + "step": 5342 + }, + { + "epoch": 2.9006514657980453, + "grad_norm": 12.040725731955204, + "learning_rate": 7.907239833398154e-06, + "loss": 0.7143, + "step": 5343 + }, + { + "epoch": 2.9011943539630836, + "grad_norm": 14.022224272834306, + "learning_rate": 7.903800998215735e-06, + "loss": 0.7389, + "step": 5344 + }, + { + "epoch": 2.901737242128122, + "grad_norm": 9.316170523577583, + "learning_rate": 7.900362422294545e-06, + "loss": 0.6061, + "step": 5345 + }, + { + "epoch": 2.9022801302931596, + "grad_norm": 8.20692587330132, + "learning_rate": 7.896924106059882e-06, + "loss": 0.4652, + "step": 5346 + }, + { + "epoch": 2.9028230184581973, + "grad_norm": 12.086544028526333, + "learning_rate": 7.893486049936993e-06, + "loss": 0.6579, + "step": 5347 + }, + { + "epoch": 2.9033659066232356, + "grad_norm": 11.562435941032952, + "learning_rate": 7.89004825435111e-06, + "loss": 0.6619, + "step": 5348 + }, + { + "epoch": 2.903908794788274, + "grad_norm": 9.535770499152292, + "learning_rate": 7.88661071972742e-06, + "loss": 0.4646, + "step": 5349 + }, + { + "epoch": 2.9044516829533116, + "grad_norm": 9.216711023311518, + "learning_rate": 7.88317344649108e-06, + "loss": 0.8006, + "step": 5350 + }, + { + "epoch": 2.9049945711183494, + "grad_norm": 9.577156917577227, + "learning_rate": 7.879736435067228e-06, + "loss": 0.7352, + "step": 5351 + }, + { + "epoch": 2.9055374592833876, + "grad_norm": 7.559386359727195, + "learning_rate": 7.87629968588095e-06, + "loss": 0.3841, + "step": 5352 + }, + { + "epoch": 2.906080347448426, + "grad_norm": 13.704096139083338, + "learning_rate": 7.872863199357315e-06, + "loss": 0.992, + "step": 5353 + }, + { + "epoch": 2.9066232356134636, + "grad_norm": 9.988539472899605, + "learning_rate": 7.869426975921344e-06, + "loss": 0.455, + "step": 5354 + }, + { + "epoch": 2.9071661237785014, + "grad_norm": 11.253082081483269, + "learning_rate": 7.865991015998043e-06, + "loss": 0.8226, + "step": 5355 + }, + { + "epoch": 2.9077090119435396, + "grad_norm": 10.526429101674665, + "learning_rate": 7.862555320012373e-06, + "loss": 0.5357, + "step": 5356 + }, + { + "epoch": 2.908251900108578, + "grad_norm": 9.646031093569851, + "learning_rate": 7.859119888389271e-06, + "loss": 0.4366, + "step": 5357 + }, + { + "epoch": 2.9087947882736156, + "grad_norm": 11.695549963620051, + "learning_rate": 7.855684721553633e-06, + "loss": 0.5834, + "step": 5358 + }, + { + "epoch": 2.9093376764386534, + "grad_norm": 13.315809341009361, + "learning_rate": 7.852249819930324e-06, + "loss": 1.0645, + "step": 5359 + }, + { + "epoch": 2.9098805646036916, + "grad_norm": 12.349221876871807, + "learning_rate": 7.848815183944183e-06, + "loss": 0.577, + "step": 5360 + }, + { + "epoch": 2.91042345276873, + "grad_norm": 16.63305773978975, + "learning_rate": 7.845380814020004e-06, + "loss": 0.9412, + "step": 5361 + }, + { + "epoch": 2.9109663409337676, + "grad_norm": 8.450548593743887, + "learning_rate": 7.841946710582563e-06, + "loss": 0.4955, + "step": 5362 + }, + { + "epoch": 2.9115092290988054, + "grad_norm": 11.465063371310775, + "learning_rate": 7.838512874056589e-06, + "loss": 0.6399, + "step": 5363 + }, + { + "epoch": 2.9120521172638436, + "grad_norm": 8.712777448735437, + "learning_rate": 7.835079304866793e-06, + "loss": 0.309, + "step": 5364 + }, + { + "epoch": 2.912595005428882, + "grad_norm": 9.44188587266759, + "learning_rate": 7.831646003437835e-06, + "loss": 0.435, + "step": 5365 + }, + { + "epoch": 2.9131378935939196, + "grad_norm": 12.094488103222616, + "learning_rate": 7.828212970194357e-06, + "loss": 0.7245, + "step": 5366 + }, + { + "epoch": 2.9136807817589574, + "grad_norm": 11.219679589625631, + "learning_rate": 7.82478020556096e-06, + "loss": 0.8716, + "step": 5367 + }, + { + "epoch": 2.9142236699239956, + "grad_norm": 10.091805149105353, + "learning_rate": 7.821347709962211e-06, + "loss": 0.553, + "step": 5368 + }, + { + "epoch": 2.914766558089034, + "grad_norm": 12.047544009676127, + "learning_rate": 7.817915483822654e-06, + "loss": 0.7283, + "step": 5369 + }, + { + "epoch": 2.9153094462540716, + "grad_norm": 10.891198338724942, + "learning_rate": 7.814483527566783e-06, + "loss": 0.6233, + "step": 5370 + }, + { + "epoch": 2.9158523344191094, + "grad_norm": 8.372829160348498, + "learning_rate": 7.811051841619077e-06, + "loss": 0.3781, + "step": 5371 + }, + { + "epoch": 2.9163952225841476, + "grad_norm": 7.8341866326850464, + "learning_rate": 7.807620426403964e-06, + "loss": 0.3455, + "step": 5372 + }, + { + "epoch": 2.916938110749186, + "grad_norm": 11.836979275353194, + "learning_rate": 7.804189282345855e-06, + "loss": 0.5573, + "step": 5373 + }, + { + "epoch": 2.9174809989142236, + "grad_norm": 8.195935708953526, + "learning_rate": 7.800758409869111e-06, + "loss": 0.4219, + "step": 5374 + }, + { + "epoch": 2.9180238870792614, + "grad_norm": 12.93514826751349, + "learning_rate": 7.797327809398074e-06, + "loss": 1.0471, + "step": 5375 + }, + { + "epoch": 2.9185667752442996, + "grad_norm": 9.384641490219252, + "learning_rate": 7.793897481357047e-06, + "loss": 0.5004, + "step": 5376 + }, + { + "epoch": 2.919109663409338, + "grad_norm": 10.40559875162228, + "learning_rate": 7.79046742617029e-06, + "loss": 0.5434, + "step": 5377 + }, + { + "epoch": 2.9196525515743756, + "grad_norm": 10.11339680443781, + "learning_rate": 7.787037644262048e-06, + "loss": 0.6003, + "step": 5378 + }, + { + "epoch": 2.9201954397394134, + "grad_norm": 12.37057409853331, + "learning_rate": 7.783608136056512e-06, + "loss": 0.685, + "step": 5379 + }, + { + "epoch": 2.9207383279044516, + "grad_norm": 11.589913705537441, + "learning_rate": 7.780178901977857e-06, + "loss": 0.7646, + "step": 5380 + }, + { + "epoch": 2.92128121606949, + "grad_norm": 12.462452174896484, + "learning_rate": 7.77674994245021e-06, + "loss": 0.6442, + "step": 5381 + }, + { + "epoch": 2.9218241042345277, + "grad_norm": 9.567483949166586, + "learning_rate": 7.77332125789768e-06, + "loss": 0.4645, + "step": 5382 + }, + { + "epoch": 2.9223669923995654, + "grad_norm": 11.079212789248007, + "learning_rate": 7.76989284874432e-06, + "loss": 0.7285, + "step": 5383 + }, + { + "epoch": 2.9229098805646037, + "grad_norm": 14.661322701022721, + "learning_rate": 7.76646471541417e-06, + "loss": 0.7481, + "step": 5384 + }, + { + "epoch": 2.923452768729642, + "grad_norm": 13.968723101600672, + "learning_rate": 7.763036858331222e-06, + "loss": 0.6632, + "step": 5385 + }, + { + "epoch": 2.9239956568946797, + "grad_norm": 12.145893524297147, + "learning_rate": 7.759609277919442e-06, + "loss": 0.8389, + "step": 5386 + }, + { + "epoch": 2.9245385450597174, + "grad_norm": 10.052846347427483, + "learning_rate": 7.756181974602757e-06, + "loss": 0.5351, + "step": 5387 + }, + { + "epoch": 2.9250814332247557, + "grad_norm": 11.030654536959574, + "learning_rate": 7.75275494880506e-06, + "loss": 0.6469, + "step": 5388 + }, + { + "epoch": 2.925624321389794, + "grad_norm": 13.670311382356902, + "learning_rate": 7.749328200950215e-06, + "loss": 0.984, + "step": 5389 + }, + { + "epoch": 2.9261672095548317, + "grad_norm": 10.129392847361466, + "learning_rate": 7.745901731462041e-06, + "loss": 0.8149, + "step": 5390 + }, + { + "epoch": 2.9267100977198695, + "grad_norm": 7.867905327161546, + "learning_rate": 7.742475540764339e-06, + "loss": 0.4957, + "step": 5391 + }, + { + "epoch": 2.9272529858849077, + "grad_norm": 8.467285225937513, + "learning_rate": 7.739049629280854e-06, + "loss": 0.579, + "step": 5392 + }, + { + "epoch": 2.927795874049946, + "grad_norm": 8.215885752516279, + "learning_rate": 7.735623997435318e-06, + "loss": 0.4073, + "step": 5393 + }, + { + "epoch": 2.9283387622149837, + "grad_norm": 8.51797242171681, + "learning_rate": 7.732198645651413e-06, + "loss": 0.63, + "step": 5394 + }, + { + "epoch": 2.9288816503800215, + "grad_norm": 15.898718458126094, + "learning_rate": 7.728773574352795e-06, + "loss": 0.7468, + "step": 5395 + }, + { + "epoch": 2.9294245385450597, + "grad_norm": 11.412421000549452, + "learning_rate": 7.725348783963084e-06, + "loss": 0.9522, + "step": 5396 + }, + { + "epoch": 2.929967426710098, + "grad_norm": 9.428473061604137, + "learning_rate": 7.721924274905855e-06, + "loss": 0.5388, + "step": 5397 + }, + { + "epoch": 2.9305103148751357, + "grad_norm": 10.083367258939811, + "learning_rate": 7.718500047604667e-06, + "loss": 0.6605, + "step": 5398 + }, + { + "epoch": 2.9310532030401735, + "grad_norm": 12.85805147473796, + "learning_rate": 7.715076102483026e-06, + "loss": 0.7407, + "step": 5399 + }, + { + "epoch": 2.9315960912052117, + "grad_norm": 9.866769374539361, + "learning_rate": 7.711652439964415e-06, + "loss": 0.6007, + "step": 5400 + }, + { + "epoch": 2.93213897937025, + "grad_norm": 12.367813318049626, + "learning_rate": 7.708229060472277e-06, + "loss": 0.5462, + "step": 5401 + }, + { + "epoch": 2.9326818675352877, + "grad_norm": 13.515093387837554, + "learning_rate": 7.704805964430023e-06, + "loss": 1.0205, + "step": 5402 + }, + { + "epoch": 2.9332247557003255, + "grad_norm": 10.667525494454107, + "learning_rate": 7.701383152261022e-06, + "loss": 0.7873, + "step": 5403 + }, + { + "epoch": 2.9337676438653637, + "grad_norm": 13.024983806475014, + "learning_rate": 7.697960624388621e-06, + "loss": 0.6388, + "step": 5404 + }, + { + "epoch": 2.934310532030402, + "grad_norm": 7.885758920671933, + "learning_rate": 7.694538381236115e-06, + "loss": 0.44, + "step": 5405 + }, + { + "epoch": 2.9348534201954397, + "grad_norm": 9.862140672970398, + "learning_rate": 7.69111642322678e-06, + "loss": 0.4065, + "step": 5406 + }, + { + "epoch": 2.9353963083604775, + "grad_norm": 11.605538896695489, + "learning_rate": 7.687694750783844e-06, + "loss": 0.6574, + "step": 5407 + }, + { + "epoch": 2.9359391965255157, + "grad_norm": 9.384935873492674, + "learning_rate": 7.684273364330505e-06, + "loss": 0.573, + "step": 5408 + }, + { + "epoch": 2.936482084690554, + "grad_norm": 7.506351724565119, + "learning_rate": 7.68085226428993e-06, + "loss": 0.4421, + "step": 5409 + }, + { + "epoch": 2.9370249728555917, + "grad_norm": 9.783370092521682, + "learning_rate": 7.677431451085238e-06, + "loss": 0.49, + "step": 5410 + }, + { + "epoch": 2.9375678610206295, + "grad_norm": 10.612737595311591, + "learning_rate": 7.674010925139533e-06, + "loss": 0.6855, + "step": 5411 + }, + { + "epoch": 2.9381107491856677, + "grad_norm": 10.77127188283207, + "learning_rate": 7.670590686875856e-06, + "loss": 0.4701, + "step": 5412 + }, + { + "epoch": 2.938653637350706, + "grad_norm": 9.009892989770995, + "learning_rate": 7.66717073671724e-06, + "loss": 0.3802, + "step": 5413 + }, + { + "epoch": 2.9391965255157437, + "grad_norm": 9.718396629861862, + "learning_rate": 7.663751075086662e-06, + "loss": 0.6433, + "step": 5414 + }, + { + "epoch": 2.9397394136807815, + "grad_norm": 8.544955706387817, + "learning_rate": 7.66033170240708e-06, + "loss": 0.3449, + "step": 5415 + }, + { + "epoch": 2.9402823018458197, + "grad_norm": 12.174166457717448, + "learning_rate": 7.656912619101401e-06, + "loss": 0.4685, + "step": 5416 + }, + { + "epoch": 2.940825190010858, + "grad_norm": 8.829471274310071, + "learning_rate": 7.6534938255925e-06, + "loss": 0.6213, + "step": 5417 + }, + { + "epoch": 2.9413680781758957, + "grad_norm": 11.089620150396728, + "learning_rate": 7.650075322303222e-06, + "loss": 0.5682, + "step": 5418 + }, + { + "epoch": 2.9419109663409335, + "grad_norm": 7.099888540087664, + "learning_rate": 7.646657109656375e-06, + "loss": 0.2937, + "step": 5419 + }, + { + "epoch": 2.9424538545059717, + "grad_norm": 9.8313977098716, + "learning_rate": 7.64323918807473e-06, + "loss": 0.6224, + "step": 5420 + }, + { + "epoch": 2.94299674267101, + "grad_norm": 14.892485492602605, + "learning_rate": 7.639821557981013e-06, + "loss": 1.0678, + "step": 5421 + }, + { + "epoch": 2.9435396308360477, + "grad_norm": 8.648910922847081, + "learning_rate": 7.636404219797931e-06, + "loss": 0.3605, + "step": 5422 + }, + { + "epoch": 2.9440825190010855, + "grad_norm": 12.093378946161085, + "learning_rate": 7.63298717394814e-06, + "loss": 0.9133, + "step": 5423 + }, + { + "epoch": 2.9446254071661238, + "grad_norm": 9.302461570715176, + "learning_rate": 7.629570420854271e-06, + "loss": 0.5305, + "step": 5424 + }, + { + "epoch": 2.945168295331162, + "grad_norm": 10.492242594764365, + "learning_rate": 7.626153960938909e-06, + "loss": 0.6007, + "step": 5425 + }, + { + "epoch": 2.9457111834961998, + "grad_norm": 11.212789857802488, + "learning_rate": 7.622737794624605e-06, + "loss": 0.4854, + "step": 5426 + }, + { + "epoch": 2.9462540716612375, + "grad_norm": 11.607518266622822, + "learning_rate": 7.619321922333884e-06, + "loss": 0.5543, + "step": 5427 + }, + { + "epoch": 2.9467969598262758, + "grad_norm": 11.268089998182159, + "learning_rate": 7.615906344489219e-06, + "loss": 0.6998, + "step": 5428 + }, + { + "epoch": 2.947339847991314, + "grad_norm": 12.383131636281371, + "learning_rate": 7.61249106151306e-06, + "loss": 0.8179, + "step": 5429 + }, + { + "epoch": 2.9478827361563518, + "grad_norm": 11.603914807740104, + "learning_rate": 7.609076073827808e-06, + "loss": 0.9006, + "step": 5430 + }, + { + "epoch": 2.9484256243213895, + "grad_norm": 9.674307468707637, + "learning_rate": 7.6056613818558415e-06, + "loss": 0.6638, + "step": 5431 + }, + { + "epoch": 2.9489685124864278, + "grad_norm": 13.055177059547185, + "learning_rate": 7.602246986019487e-06, + "loss": 0.9642, + "step": 5432 + }, + { + "epoch": 2.949511400651466, + "grad_norm": 11.00057075160912, + "learning_rate": 7.598832886741049e-06, + "loss": 0.4843, + "step": 5433 + }, + { + "epoch": 2.950054288816504, + "grad_norm": 11.15214171419394, + "learning_rate": 7.5954190844427834e-06, + "loss": 0.6722, + "step": 5434 + }, + { + "epoch": 2.9505971769815416, + "grad_norm": 8.91145682769725, + "learning_rate": 7.592005579546924e-06, + "loss": 0.6258, + "step": 5435 + }, + { + "epoch": 2.95114006514658, + "grad_norm": 12.685207726915342, + "learning_rate": 7.588592372475651e-06, + "loss": 0.8481, + "step": 5436 + }, + { + "epoch": 2.951682953311618, + "grad_norm": 13.10934163141668, + "learning_rate": 7.585179463651112e-06, + "loss": 0.6655, + "step": 5437 + }, + { + "epoch": 2.952225841476656, + "grad_norm": 10.993826657178786, + "learning_rate": 7.5817668534954295e-06, + "loss": 0.5687, + "step": 5438 + }, + { + "epoch": 2.9527687296416936, + "grad_norm": 8.424743777393193, + "learning_rate": 7.578354542430673e-06, + "loss": 0.5485, + "step": 5439 + }, + { + "epoch": 2.953311617806732, + "grad_norm": 14.730970588587084, + "learning_rate": 7.574942530878891e-06, + "loss": 0.8759, + "step": 5440 + }, + { + "epoch": 2.95385450597177, + "grad_norm": 8.667781572678388, + "learning_rate": 7.571530819262077e-06, + "loss": 0.3704, + "step": 5441 + }, + { + "epoch": 2.954397394136808, + "grad_norm": 6.739834901590041, + "learning_rate": 7.568119408002206e-06, + "loss": 0.3619, + "step": 5442 + }, + { + "epoch": 2.9549402823018456, + "grad_norm": 5.9221212094491795, + "learning_rate": 7.564708297521197e-06, + "loss": 0.3177, + "step": 5443 + }, + { + "epoch": 2.955483170466884, + "grad_norm": 9.008687743468037, + "learning_rate": 7.561297488240953e-06, + "loss": 0.5756, + "step": 5444 + }, + { + "epoch": 2.956026058631922, + "grad_norm": 10.49570531884228, + "learning_rate": 7.55788698058332e-06, + "loss": 0.609, + "step": 5445 + }, + { + "epoch": 2.95656894679696, + "grad_norm": 11.053313913817204, + "learning_rate": 7.554476774970115e-06, + "loss": 0.6893, + "step": 5446 + }, + { + "epoch": 2.9571118349619976, + "grad_norm": 10.747800237741414, + "learning_rate": 7.551066871823125e-06, + "loss": 0.6361, + "step": 5447 + }, + { + "epoch": 2.957654723127036, + "grad_norm": 12.158436457559148, + "learning_rate": 7.547657271564083e-06, + "loss": 0.6427, + "step": 5448 + }, + { + "epoch": 2.958197611292074, + "grad_norm": 11.021084054397756, + "learning_rate": 7.544247974614701e-06, + "loss": 0.4399, + "step": 5449 + }, + { + "epoch": 2.958740499457112, + "grad_norm": 8.337942068125022, + "learning_rate": 7.540838981396641e-06, + "loss": 0.3781, + "step": 5450 + }, + { + "epoch": 2.9592833876221496, + "grad_norm": 9.86638573791189, + "learning_rate": 7.537430292331536e-06, + "loss": 0.4257, + "step": 5451 + }, + { + "epoch": 2.959826275787188, + "grad_norm": 11.512553822962575, + "learning_rate": 7.534021907840975e-06, + "loss": 0.64, + "step": 5452 + }, + { + "epoch": 2.960369163952226, + "grad_norm": 10.540338627981921, + "learning_rate": 7.530613828346519e-06, + "loss": 0.6669, + "step": 5453 + }, + { + "epoch": 2.960912052117264, + "grad_norm": 12.092382555989715, + "learning_rate": 7.5272060542696765e-06, + "loss": 0.6346, + "step": 5454 + }, + { + "epoch": 2.9614549402823016, + "grad_norm": 11.429168527550765, + "learning_rate": 7.5237985860319344e-06, + "loss": 0.7454, + "step": 5455 + }, + { + "epoch": 2.96199782844734, + "grad_norm": 10.331080624952307, + "learning_rate": 7.52039142405473e-06, + "loss": 0.5472, + "step": 5456 + }, + { + "epoch": 2.962540716612378, + "grad_norm": 16.484391160691512, + "learning_rate": 7.516984568759461e-06, + "loss": 1.3039, + "step": 5457 + }, + { + "epoch": 2.963083604777416, + "grad_norm": 12.853254391289898, + "learning_rate": 7.5135780205675e-06, + "loss": 0.6846, + "step": 5458 + }, + { + "epoch": 2.9636264929424536, + "grad_norm": 12.062646480747443, + "learning_rate": 7.510171779900171e-06, + "loss": 0.5656, + "step": 5459 + }, + { + "epoch": 2.964169381107492, + "grad_norm": 8.21924446991787, + "learning_rate": 7.506765847178768e-06, + "loss": 0.3401, + "step": 5460 + }, + { + "epoch": 2.96471226927253, + "grad_norm": 11.116098821110214, + "learning_rate": 7.503360222824535e-06, + "loss": 0.4931, + "step": 5461 + }, + { + "epoch": 2.965255157437568, + "grad_norm": 9.759573410672228, + "learning_rate": 7.499954907258693e-06, + "loss": 0.5363, + "step": 5462 + }, + { + "epoch": 2.9657980456026056, + "grad_norm": 15.195018271421278, + "learning_rate": 7.496549900902408e-06, + "loss": 0.5349, + "step": 5463 + }, + { + "epoch": 2.966340933767644, + "grad_norm": 11.555768004386108, + "learning_rate": 7.493145204176823e-06, + "loss": 0.8273, + "step": 5464 + }, + { + "epoch": 2.966883821932682, + "grad_norm": 12.516961882115927, + "learning_rate": 7.4897408175030366e-06, + "loss": 0.7826, + "step": 5465 + }, + { + "epoch": 2.96742671009772, + "grad_norm": 11.67450620684471, + "learning_rate": 7.486336741302103e-06, + "loss": 0.5689, + "step": 5466 + }, + { + "epoch": 2.9679695982627576, + "grad_norm": 13.075852566862391, + "learning_rate": 7.48293297599505e-06, + "loss": 0.8355, + "step": 5467 + }, + { + "epoch": 2.968512486427796, + "grad_norm": 8.943584587537828, + "learning_rate": 7.479529522002855e-06, + "loss": 0.3921, + "step": 5468 + }, + { + "epoch": 2.969055374592834, + "grad_norm": 10.91179454178022, + "learning_rate": 7.476126379746471e-06, + "loss": 0.7276, + "step": 5469 + }, + { + "epoch": 2.969598262757872, + "grad_norm": 11.79182407811329, + "learning_rate": 7.472723549646793e-06, + "loss": 1.012, + "step": 5470 + }, + { + "epoch": 2.9701411509229096, + "grad_norm": 12.476871332438998, + "learning_rate": 7.469321032124697e-06, + "loss": 0.7994, + "step": 5471 + }, + { + "epoch": 2.970684039087948, + "grad_norm": 14.350758882559228, + "learning_rate": 7.465918827601008e-06, + "loss": 0.8477, + "step": 5472 + }, + { + "epoch": 2.971226927252986, + "grad_norm": 13.79963097595528, + "learning_rate": 7.46251693649652e-06, + "loss": 1.0897, + "step": 5473 + }, + { + "epoch": 2.971769815418024, + "grad_norm": 9.061867840853571, + "learning_rate": 7.459115359231977e-06, + "loss": 0.2972, + "step": 5474 + }, + { + "epoch": 2.9723127035830617, + "grad_norm": 11.193898477701971, + "learning_rate": 7.455714096228102e-06, + "loss": 0.9238, + "step": 5475 + }, + { + "epoch": 2.9728555917481, + "grad_norm": 13.156649778942683, + "learning_rate": 7.452313147905559e-06, + "loss": 0.9945, + "step": 5476 + }, + { + "epoch": 2.973398479913138, + "grad_norm": 10.673119691620487, + "learning_rate": 7.448912514684985e-06, + "loss": 0.6296, + "step": 5477 + }, + { + "epoch": 2.973941368078176, + "grad_norm": 13.208883224921669, + "learning_rate": 7.44551219698698e-06, + "loss": 0.5018, + "step": 5478 + }, + { + "epoch": 2.9744842562432137, + "grad_norm": 10.477928943580013, + "learning_rate": 7.442112195232093e-06, + "loss": 0.4782, + "step": 5479 + }, + { + "epoch": 2.975027144408252, + "grad_norm": 10.052094162546567, + "learning_rate": 7.438712509840852e-06, + "loss": 0.5661, + "step": 5480 + }, + { + "epoch": 2.97557003257329, + "grad_norm": 14.921523553916886, + "learning_rate": 7.435313141233724e-06, + "loss": 1.3182, + "step": 5481 + }, + { + "epoch": 2.976112920738328, + "grad_norm": 9.3079131640488, + "learning_rate": 7.43191408983116e-06, + "loss": 0.5672, + "step": 5482 + }, + { + "epoch": 2.9766558089033657, + "grad_norm": 8.76037752811841, + "learning_rate": 7.428515356053551e-06, + "loss": 0.6026, + "step": 5483 + }, + { + "epoch": 2.977198697068404, + "grad_norm": 11.037798492427186, + "learning_rate": 7.425116940321262e-06, + "loss": 0.7053, + "step": 5484 + }, + { + "epoch": 2.977741585233442, + "grad_norm": 12.329608171901635, + "learning_rate": 7.421718843054615e-06, + "loss": 0.6576, + "step": 5485 + }, + { + "epoch": 2.97828447339848, + "grad_norm": 14.664205153544405, + "learning_rate": 7.4183210646738875e-06, + "loss": 0.7546, + "step": 5486 + }, + { + "epoch": 2.9788273615635177, + "grad_norm": 16.28337626285995, + "learning_rate": 7.414923605599329e-06, + "loss": 1.2774, + "step": 5487 + }, + { + "epoch": 2.979370249728556, + "grad_norm": 11.936294560502736, + "learning_rate": 7.411526466251135e-06, + "loss": 0.9397, + "step": 5488 + }, + { + "epoch": 2.979913137893594, + "grad_norm": 15.443918342772207, + "learning_rate": 7.408129647049474e-06, + "loss": 0.7591, + "step": 5489 + }, + { + "epoch": 2.980456026058632, + "grad_norm": 9.303363471188012, + "learning_rate": 7.404733148414471e-06, + "loss": 0.4917, + "step": 5490 + }, + { + "epoch": 2.9809989142236697, + "grad_norm": 16.111589693524394, + "learning_rate": 7.401336970766207e-06, + "loss": 0.8724, + "step": 5491 + }, + { + "epoch": 2.981541802388708, + "grad_norm": 15.913107093535366, + "learning_rate": 7.397941114524727e-06, + "loss": 1.0915, + "step": 5492 + }, + { + "epoch": 2.982084690553746, + "grad_norm": 12.432387434616404, + "learning_rate": 7.3945455801100404e-06, + "loss": 1.1143, + "step": 5493 + }, + { + "epoch": 2.982627578718784, + "grad_norm": 11.65910861056599, + "learning_rate": 7.39115036794211e-06, + "loss": 0.7012, + "step": 5494 + }, + { + "epoch": 2.9831704668838217, + "grad_norm": 10.645886415744744, + "learning_rate": 7.387755478440855e-06, + "loss": 0.5327, + "step": 5495 + }, + { + "epoch": 2.98371335504886, + "grad_norm": 11.340492191845808, + "learning_rate": 7.384360912026167e-06, + "loss": 0.8393, + "step": 5496 + }, + { + "epoch": 2.984256243213898, + "grad_norm": 11.222712079595647, + "learning_rate": 7.38096666911789e-06, + "loss": 0.4731, + "step": 5497 + }, + { + "epoch": 2.984799131378936, + "grad_norm": 14.06199638971423, + "learning_rate": 7.377572750135833e-06, + "loss": 1.0083, + "step": 5498 + }, + { + "epoch": 2.9853420195439737, + "grad_norm": 11.609516661749161, + "learning_rate": 7.374179155499752e-06, + "loss": 0.5397, + "step": 5499 + }, + { + "epoch": 2.985884907709012, + "grad_norm": 11.809076211376697, + "learning_rate": 7.370785885629384e-06, + "loss": 0.5726, + "step": 5500 + }, + { + "epoch": 2.98642779587405, + "grad_norm": 12.767099228571903, + "learning_rate": 7.367392940944403e-06, + "loss": 0.8919, + "step": 5501 + }, + { + "epoch": 2.986970684039088, + "grad_norm": 8.511059646511987, + "learning_rate": 7.364000321864464e-06, + "loss": 0.4431, + "step": 5502 + }, + { + "epoch": 2.9875135722041257, + "grad_norm": 14.166198872023001, + "learning_rate": 7.360608028809161e-06, + "loss": 0.8709, + "step": 5503 + }, + { + "epoch": 2.988056460369164, + "grad_norm": 11.774867588892974, + "learning_rate": 7.357216062198066e-06, + "loss": 0.4857, + "step": 5504 + }, + { + "epoch": 2.988599348534202, + "grad_norm": 9.622067683785492, + "learning_rate": 7.353824422450702e-06, + "loss": 0.4385, + "step": 5505 + }, + { + "epoch": 2.98914223669924, + "grad_norm": 9.959533869172905, + "learning_rate": 7.3504331099865474e-06, + "loss": 0.6041, + "step": 5506 + }, + { + "epoch": 2.9896851248642777, + "grad_norm": 12.611515654376284, + "learning_rate": 7.347042125225052e-06, + "loss": 0.5667, + "step": 5507 + }, + { + "epoch": 2.990228013029316, + "grad_norm": 16.346151732068346, + "learning_rate": 7.343651468585611e-06, + "loss": 0.8017, + "step": 5508 + }, + { + "epoch": 2.990770901194354, + "grad_norm": 14.039933672423732, + "learning_rate": 7.340261140487593e-06, + "loss": 0.6905, + "step": 5509 + }, + { + "epoch": 2.991313789359392, + "grad_norm": 9.68037751098424, + "learning_rate": 7.336871141350313e-06, + "loss": 0.7553, + "step": 5510 + }, + { + "epoch": 2.9918566775244297, + "grad_norm": 11.6044953546391, + "learning_rate": 7.333481471593058e-06, + "loss": 0.7525, + "step": 5511 + }, + { + "epoch": 2.992399565689468, + "grad_norm": 10.257246759089158, + "learning_rate": 7.330092131635061e-06, + "loss": 0.5479, + "step": 5512 + }, + { + "epoch": 2.992942453854506, + "grad_norm": 9.37947259585779, + "learning_rate": 7.326703121895528e-06, + "loss": 0.6117, + "step": 5513 + }, + { + "epoch": 2.993485342019544, + "grad_norm": 10.344183596900065, + "learning_rate": 7.323314442793614e-06, + "loss": 0.5074, + "step": 5514 + }, + { + "epoch": 2.9940282301845818, + "grad_norm": 8.940259405187144, + "learning_rate": 7.319926094748433e-06, + "loss": 0.2863, + "step": 5515 + }, + { + "epoch": 2.99457111834962, + "grad_norm": 10.920924886109798, + "learning_rate": 7.316538078179065e-06, + "loss": 0.7706, + "step": 5516 + }, + { + "epoch": 2.995114006514658, + "grad_norm": 4.8493484123114445, + "learning_rate": 7.3131503935045424e-06, + "loss": 0.2066, + "step": 5517 + }, + { + "epoch": 2.995656894679696, + "grad_norm": 11.385458894978902, + "learning_rate": 7.309763041143865e-06, + "loss": 0.5706, + "step": 5518 + }, + { + "epoch": 2.9961997828447338, + "grad_norm": 8.263697834815542, + "learning_rate": 7.306376021515977e-06, + "loss": 0.667, + "step": 5519 + }, + { + "epoch": 2.996742671009772, + "grad_norm": 11.30709409496411, + "learning_rate": 7.302989335039801e-06, + "loss": 0.5522, + "step": 5520 + }, + { + "epoch": 2.99728555917481, + "grad_norm": 9.914253293704078, + "learning_rate": 7.299602982134199e-06, + "loss": 0.6109, + "step": 5521 + }, + { + "epoch": 2.997828447339848, + "grad_norm": 15.471798238376472, + "learning_rate": 7.2962169632180055e-06, + "loss": 0.8321, + "step": 5522 + }, + { + "epoch": 2.9983713355048858, + "grad_norm": 10.956384342902831, + "learning_rate": 7.292831278710007e-06, + "loss": 0.5791, + "step": 5523 + }, + { + "epoch": 2.998914223669924, + "grad_norm": 8.8527583386149, + "learning_rate": 7.289445929028951e-06, + "loss": 0.5807, + "step": 5524 + }, + { + "epoch": 2.999457111834962, + "grad_norm": 10.520949488289421, + "learning_rate": 7.286060914593544e-06, + "loss": 0.6861, + "step": 5525 + }, + { + "epoch": 3.0, + "grad_norm": 11.327416720286067, + "learning_rate": 7.282676235822444e-06, + "loss": 0.4657, + "step": 5526 + } + ], + "logging_steps": 1.0, + "max_steps": 9210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6593413474418688.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/training_args.bin b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b16de63f342ea53d59c4d97aa8a3bd463a79531f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9801fdd09c6ba798e12b77f77221db3b5747b6e1296785789d44cf2bc492ff9e +size 7160 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/zero_to_fp32.py b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-5526/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/README.md b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/README.md new file mode 100644 index 0000000000000000000000000000000000000000..972a036203d877262d3c6673f4d81814e7409dc5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/README.md @@ -0,0 +1,202 @@ +--- +base_model: llava-hf/LLaVA-NeXT-Video-7B-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/adapter_config.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..719f6e47b29561ff93d519680ccd115b480d825e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/adapter_config.json @@ -0,0 +1,132 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "llava-hf/LLaVA-NeXT-Video-7B-hf", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 8, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [], + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "model.layers.17.self_attn.q_proj", + "model.layers.17.self_attn.k_proj", + "model.layers.13.self_attn.k_proj", + "model.layers.20.self_attn.q_proj", + "model.layers.18.self_attn.k_proj", + "29.self_attn.k_proj", + "model.layers.16.self_attn.v_proj", + "model.layers.14.self_attn.q_proj", + "29.self_attn.v_proj", + "24.self_attn.k_proj", + "model.layers.18.self_attn.v_proj", + "model.layers.5.self_attn.q_proj", + "model.layers.11.self_attn.k_proj", + "27.self_attn.v_proj", + "model.layers.16.self_attn.q_proj", + "model.layers.10.self_attn.q_proj", + "model.layers.12.self_attn.v_proj", + "model.layers.20.self_attn.v_proj", + "model.layers.0.self_attn.v_proj", + "model.layers.7.self_attn.q_proj", + "model.layers.23.self_attn.q_proj", + "model.layers.1.self_attn.q_proj", + "25.self_attn.v_proj", + "model.layers.4.self_attn.v_proj", + "model.layers.19.self_attn.q_proj", + "model.layers.10.self_attn.k_proj", + "model.layers.22.self_attn.k_proj", + "model.layers.14.self_attn.v_proj", + "model.layers.13.self_attn.v_proj", + "25.self_attn.k_proj", + "24.self_attn.v_proj", + "model.layers.21.self_attn.q_proj", + "model.layers.8.self_attn.q_proj", + "model.layers.6.self_attn.q_proj", + "model.layers.6.self_attn.k_proj", + "model.layers.1.self_attn.v_proj", + "model.layers.9.self_attn.k_proj", + "model.layers.13.self_attn.q_proj", + "o_proj", + "model.layers.16.self_attn.k_proj", + "model.layers.6.self_attn.v_proj", + "model.layers.21.self_attn.k_proj", + "model.layers.5.self_attn.k_proj", + "27.self_attn.q_proj", + "model.layers.20.self_attn.k_proj", + "model.layers.11.self_attn.v_proj", + "model.layers.22.self_attn.q_proj", + "model.layers.14.self_attn.k_proj", + "28.self_attn.v_proj", + "model.layers.22.self_attn.v_proj", + "gate_proj", + "model.layers.0.self_attn.k_proj", + "model.layers.5.self_attn.v_proj", + "30.self_attn.q_proj", + "model.layers.3.self_attn.k_proj", + "29.self_attn.q_proj", + "model.layers.19.self_attn.k_proj", + "model.layers.10.self_attn.v_proj", + "model.layers.15.self_attn.k_proj", + "model.layers.7.self_attn.v_proj", + "model.layers.9.self_attn.v_proj", + "up_proj", + "model.layers.4.self_attn.q_proj", + "28.self_attn.q_proj", + "model.layers.3.self_attn.q_proj", + "24.self_attn.q_proj", + "30.self_attn.v_proj", + "model.layers.12.self_attn.k_proj", + "model.layers.23.self_attn.k_proj", + "26.self_attn.q_proj", + "28.self_attn.k_proj", + "31.self_attn.k_proj", + "down_proj", + "model.layers.2.self_attn.q_proj", + "31.self_attn.q_proj", + "model.layers.15.self_attn.q_proj", + "model.layers.12.self_attn.q_proj", + "model.layers.0.self_attn.q_proj", + "model.layers.2.self_attn.k_proj", + "model.layers.3.self_attn.v_proj", + "model.layers.11.self_attn.q_proj", + "25.self_attn.q_proj", + "model.layers.23.self_attn.v_proj", + "30.self_attn.k_proj", + "31.self_attn.v_proj", + "model.layers.15.self_attn.v_proj", + "model.layers.8.self_attn.v_proj", + "26.self_attn.k_proj", + "model.layers.19.self_attn.v_proj", + "model.layers.2.self_attn.v_proj", + "26.self_attn.v_proj", + "model.layers.21.self_attn.v_proj", + "model.layers.17.self_attn.v_proj", + "model.layers.8.self_attn.k_proj", + "model.layers.18.self_attn.q_proj", + "model.layers.7.self_attn.k_proj", + "model.layers.4.self_attn.k_proj", + "model.layers.9.self_attn.q_proj", + "27.self_attn.k_proj", + "model.layers.1.self_attn.k_proj" + ], + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/adapter_model.safetensors b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee4295200d409db0212dbd52de1fe10129805c52 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:291400300815f186e4dfd0ebf16b1044097816690b93a4ddfe1d19b08af85446 +size 40043208 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..5971f49845f12953926ccfcf3c9fa046cccfbd3c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb95adc1a5a697c2b89fc0344924a27b89ffab56dcd0876bbe047ed4c1c248de +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..d08fb8522d86e21f01bc385bfa927e4a1ad96259 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abbda07b4b16ed46fcd276063a60f39449bf83d635fbf0c4b36369bf94a30491 +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/zero_pp_rank_0_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/zero_pp_rank_0_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..bf7c11d472801a085b88ab4c8139ff4fd50c511f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/zero_pp_rank_0_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d3a8bafc07c399f5e310771888e35e55f933f7c9de1f0d9ef640255bd339eac +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/zero_pp_rank_1_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/zero_pp_rank_1_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..dcb7db4c940d7220c092e2cffbb9f62fa74c363c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/global_step7368/zero_pp_rank_1_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9a1cb0f1d9536384519933ee976eaca05dcd1c94f812a3152cbb6fba65defec +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/latest b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/latest new file mode 100644 index 0000000000000000000000000000000000000000..0bc800786d75dd4ef0664dc7f03761b0edad9f3b --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/latest @@ -0,0 +1 @@ +global_step7368 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/rng_state_0.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..da67ba346fcc5751f942dd42da9a5af7fd026601 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d7885a44dae4456621f6cbeef8535eaa0afd5c0813cd6e2384b4681e3a5fe3f +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/rng_state_1.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..f98be9b00808a389d74dd5ec1a3c9a7c6131e5bc --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:319db49ba74bcbbdf4be433f7f773bd0cf3bc032c0da6c8771487d0e15537f04 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/scheduler.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..ac0555b11c1d27f14a819e126668d50da0cca490 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39eba225fe24ce87732a98e90019303fbaa59882193de992d7d7d27702feaca4 +size 1064 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/trainer_state.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..2abc8daf7de73e05d180db3fa95ac4603b57dd91 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/trainer_state.json @@ -0,0 +1,51609 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.0, + "eval_steps": 500, + "global_step": 7368, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005428881650380022, + "grad_norm": 2.1320506281889067, + "learning_rate": 7.220216606498195e-08, + "loss": 5.1936, + "step": 1 + }, + { + "epoch": 0.0010857763300760044, + "grad_norm": 1.8633488766761366, + "learning_rate": 1.444043321299639e-07, + "loss": 4.6995, + "step": 2 + }, + { + "epoch": 0.0016286644951140066, + "grad_norm": 2.063286098879776, + "learning_rate": 2.1660649819494586e-07, + "loss": 4.4902, + "step": 3 + }, + { + "epoch": 0.002171552660152009, + "grad_norm": 1.7123050230907495, + "learning_rate": 2.888086642599278e-07, + "loss": 4.9125, + "step": 4 + }, + { + "epoch": 0.0027144408251900108, + "grad_norm": 1.7499657276799891, + "learning_rate": 3.610108303249098e-07, + "loss": 3.7648, + "step": 5 + }, + { + "epoch": 0.003257328990228013, + "grad_norm": 2.0419045636066886, + "learning_rate": 4.332129963898917e-07, + "loss": 4.7842, + "step": 6 + }, + { + "epoch": 0.003800217155266015, + "grad_norm": 1.4960282529105982, + "learning_rate": 5.054151624548736e-07, + "loss": 4.9333, + "step": 7 + }, + { + "epoch": 0.004343105320304018, + "grad_norm": 1.640954967476452, + "learning_rate": 5.776173285198556e-07, + "loss": 3.9576, + "step": 8 + }, + { + "epoch": 0.004885993485342019, + "grad_norm": 2.0590500459610244, + "learning_rate": 6.498194945848375e-07, + "loss": 4.3833, + "step": 9 + }, + { + "epoch": 0.0054288816503800215, + "grad_norm": 1.07709353257816, + "learning_rate": 7.220216606498196e-07, + "loss": 3.7669, + "step": 10 + }, + { + "epoch": 0.005971769815418024, + "grad_norm": 1.9506235655782926, + "learning_rate": 7.942238267148016e-07, + "loss": 4.5975, + "step": 11 + }, + { + "epoch": 0.006514657980456026, + "grad_norm": 1.411983096954528, + "learning_rate": 8.664259927797834e-07, + "loss": 3.9706, + "step": 12 + }, + { + "epoch": 0.007057546145494028, + "grad_norm": 1.7110053131090028, + "learning_rate": 9.386281588447654e-07, + "loss": 4.9081, + "step": 13 + }, + { + "epoch": 0.00760043431053203, + "grad_norm": 1.3689315757199692, + "learning_rate": 1.0108303249097473e-06, + "loss": 4.2426, + "step": 14 + }, + { + "epoch": 0.008143322475570033, + "grad_norm": 1.5612091671030481, + "learning_rate": 1.0830324909747293e-06, + "loss": 4.5144, + "step": 15 + }, + { + "epoch": 0.008686210640608035, + "grad_norm": 1.3725168953865754, + "learning_rate": 1.1552346570397112e-06, + "loss": 4.9378, + "step": 16 + }, + { + "epoch": 0.009229098805646038, + "grad_norm": 2.027359913808926, + "learning_rate": 1.2274368231046932e-06, + "loss": 3.7477, + "step": 17 + }, + { + "epoch": 0.009771986970684038, + "grad_norm": 2.1695456962160224, + "learning_rate": 1.299638989169675e-06, + "loss": 5.7636, + "step": 18 + }, + { + "epoch": 0.01031487513572204, + "grad_norm": 3.2637002287680428, + "learning_rate": 1.3718411552346572e-06, + "loss": 6.1883, + "step": 19 + }, + { + "epoch": 0.010857763300760043, + "grad_norm": 2.5723616120643857, + "learning_rate": 1.4440433212996392e-06, + "loss": 4.0267, + "step": 20 + }, + { + "epoch": 0.011400651465798045, + "grad_norm": 1.4473754577877487, + "learning_rate": 1.516245487364621e-06, + "loss": 4.526, + "step": 21 + }, + { + "epoch": 0.011943539630836048, + "grad_norm": 1.6098208731761, + "learning_rate": 1.5884476534296031e-06, + "loss": 4.5163, + "step": 22 + }, + { + "epoch": 0.01248642779587405, + "grad_norm": 2.899624682573202, + "learning_rate": 1.6606498194945849e-06, + "loss": 4.8188, + "step": 23 + }, + { + "epoch": 0.013029315960912053, + "grad_norm": 2.273571202553131, + "learning_rate": 1.7328519855595669e-06, + "loss": 5.0074, + "step": 24 + }, + { + "epoch": 0.013572204125950055, + "grad_norm": 1.1877323869171315, + "learning_rate": 1.805054151624549e-06, + "loss": 4.4331, + "step": 25 + }, + { + "epoch": 0.014115092290988056, + "grad_norm": 1.9365501019385494, + "learning_rate": 1.8772563176895308e-06, + "loss": 5.0525, + "step": 26 + }, + { + "epoch": 0.014657980456026058, + "grad_norm": 1.7710485953105986, + "learning_rate": 1.949458483754513e-06, + "loss": 4.1346, + "step": 27 + }, + { + "epoch": 0.01520086862106406, + "grad_norm": 1.231848881292397, + "learning_rate": 2.0216606498194946e-06, + "loss": 4.2431, + "step": 28 + }, + { + "epoch": 0.015743756786102063, + "grad_norm": 1.6372838462475754, + "learning_rate": 2.0938628158844768e-06, + "loss": 5.4317, + "step": 29 + }, + { + "epoch": 0.016286644951140065, + "grad_norm": 1.7326568199523718, + "learning_rate": 2.1660649819494585e-06, + "loss": 4.4774, + "step": 30 + }, + { + "epoch": 0.016829533116178068, + "grad_norm": 2.1315641946923387, + "learning_rate": 2.2382671480144407e-06, + "loss": 4.1418, + "step": 31 + }, + { + "epoch": 0.01737242128121607, + "grad_norm": 1.916538366799816, + "learning_rate": 2.3104693140794225e-06, + "loss": 5.7926, + "step": 32 + }, + { + "epoch": 0.017915309446254073, + "grad_norm": 2.057531274363787, + "learning_rate": 2.3826714801444047e-06, + "loss": 4.2862, + "step": 33 + }, + { + "epoch": 0.018458197611292075, + "grad_norm": 2.6727877745045023, + "learning_rate": 2.4548736462093864e-06, + "loss": 5.1323, + "step": 34 + }, + { + "epoch": 0.019001085776330078, + "grad_norm": 1.5043525736330898, + "learning_rate": 2.527075812274368e-06, + "loss": 4.0634, + "step": 35 + }, + { + "epoch": 0.019543973941368076, + "grad_norm": 2.6324248758076862, + "learning_rate": 2.59927797833935e-06, + "loss": 5.4211, + "step": 36 + }, + { + "epoch": 0.02008686210640608, + "grad_norm": 1.6581905646911415, + "learning_rate": 2.6714801444043326e-06, + "loss": 4.9825, + "step": 37 + }, + { + "epoch": 0.02062975027144408, + "grad_norm": 1.6667970880898455, + "learning_rate": 2.7436823104693144e-06, + "loss": 4.3219, + "step": 38 + }, + { + "epoch": 0.021172638436482084, + "grad_norm": 1.7448150724822071, + "learning_rate": 2.815884476534296e-06, + "loss": 4.3912, + "step": 39 + }, + { + "epoch": 0.021715526601520086, + "grad_norm": 3.720503777726443, + "learning_rate": 2.8880866425992783e-06, + "loss": 6.0592, + "step": 40 + }, + { + "epoch": 0.02225841476655809, + "grad_norm": 1.6853001074010774, + "learning_rate": 2.96028880866426e-06, + "loss": 3.8395, + "step": 41 + }, + { + "epoch": 0.02280130293159609, + "grad_norm": 1.797029067943667, + "learning_rate": 3.032490974729242e-06, + "loss": 4.4651, + "step": 42 + }, + { + "epoch": 0.023344191096634093, + "grad_norm": 1.6298824221707349, + "learning_rate": 3.1046931407942245e-06, + "loss": 4.4507, + "step": 43 + }, + { + "epoch": 0.023887079261672096, + "grad_norm": 1.4793189209059368, + "learning_rate": 3.1768953068592062e-06, + "loss": 4.7849, + "step": 44 + }, + { + "epoch": 0.024429967426710098, + "grad_norm": 1.1741346570472608, + "learning_rate": 3.249097472924188e-06, + "loss": 3.5897, + "step": 45 + }, + { + "epoch": 0.0249728555917481, + "grad_norm": 2.1468867306169064, + "learning_rate": 3.3212996389891698e-06, + "loss": 3.8969, + "step": 46 + }, + { + "epoch": 0.025515743756786103, + "grad_norm": 1.657494000697021, + "learning_rate": 3.393501805054152e-06, + "loss": 4.3412, + "step": 47 + }, + { + "epoch": 0.026058631921824105, + "grad_norm": 1.7125864055710684, + "learning_rate": 3.4657039711191337e-06, + "loss": 3.8059, + "step": 48 + }, + { + "epoch": 0.026601520086862108, + "grad_norm": 2.9515347000925622, + "learning_rate": 3.5379061371841155e-06, + "loss": 4.1069, + "step": 49 + }, + { + "epoch": 0.02714440825190011, + "grad_norm": 3.0185188918098262, + "learning_rate": 3.610108303249098e-06, + "loss": 5.2646, + "step": 50 + }, + { + "epoch": 0.02768729641693811, + "grad_norm": 1.5546972252207274, + "learning_rate": 3.68231046931408e-06, + "loss": 3.7212, + "step": 51 + }, + { + "epoch": 0.02823018458197611, + "grad_norm": 2.6376587426282474, + "learning_rate": 3.7545126353790616e-06, + "loss": 4.2862, + "step": 52 + }, + { + "epoch": 0.028773072747014114, + "grad_norm": 2.053905428064986, + "learning_rate": 3.826714801444043e-06, + "loss": 3.8304, + "step": 53 + }, + { + "epoch": 0.029315960912052116, + "grad_norm": 2.9083183656849725, + "learning_rate": 3.898916967509026e-06, + "loss": 4.8223, + "step": 54 + }, + { + "epoch": 0.02985884907709012, + "grad_norm": 1.9919239162425852, + "learning_rate": 3.971119133574007e-06, + "loss": 3.9054, + "step": 55 + }, + { + "epoch": 0.03040173724212812, + "grad_norm": 2.4076122411223135, + "learning_rate": 4.043321299638989e-06, + "loss": 4.8568, + "step": 56 + }, + { + "epoch": 0.030944625407166124, + "grad_norm": 3.854434602830178, + "learning_rate": 4.115523465703971e-06, + "loss": 5.7771, + "step": 57 + }, + { + "epoch": 0.031487513572204126, + "grad_norm": 2.510303564095041, + "learning_rate": 4.1877256317689535e-06, + "loss": 5.5851, + "step": 58 + }, + { + "epoch": 0.03203040173724213, + "grad_norm": 2.8884448366576305, + "learning_rate": 4.259927797833936e-06, + "loss": 4.1774, + "step": 59 + }, + { + "epoch": 0.03257328990228013, + "grad_norm": 2.0893778739721, + "learning_rate": 4.332129963898917e-06, + "loss": 4.4679, + "step": 60 + }, + { + "epoch": 0.03311617806731813, + "grad_norm": 1.5816439039579229, + "learning_rate": 4.404332129963899e-06, + "loss": 4.1999, + "step": 61 + }, + { + "epoch": 0.033659066232356136, + "grad_norm": 1.9591574137365761, + "learning_rate": 4.4765342960288814e-06, + "loss": 4.1539, + "step": 62 + }, + { + "epoch": 0.03420195439739414, + "grad_norm": 3.1533909413995924, + "learning_rate": 4.548736462093864e-06, + "loss": 4.5842, + "step": 63 + }, + { + "epoch": 0.03474484256243214, + "grad_norm": 2.543715437360663, + "learning_rate": 4.620938628158845e-06, + "loss": 4.1473, + "step": 64 + }, + { + "epoch": 0.03528773072747014, + "grad_norm": 2.7608093188531955, + "learning_rate": 4.693140794223827e-06, + "loss": 4.9354, + "step": 65 + }, + { + "epoch": 0.035830618892508145, + "grad_norm": 4.192163162864289, + "learning_rate": 4.765342960288809e-06, + "loss": 5.4263, + "step": 66 + }, + { + "epoch": 0.03637350705754615, + "grad_norm": 2.809206875027472, + "learning_rate": 4.837545126353791e-06, + "loss": 4.4259, + "step": 67 + }, + { + "epoch": 0.03691639522258415, + "grad_norm": 3.9318073942863534, + "learning_rate": 4.909747292418773e-06, + "loss": 4.1483, + "step": 68 + }, + { + "epoch": 0.03745928338762215, + "grad_norm": 2.935065385226121, + "learning_rate": 4.981949458483755e-06, + "loss": 4.6854, + "step": 69 + }, + { + "epoch": 0.038002171552660155, + "grad_norm": 2.3687140876301305, + "learning_rate": 5.054151624548736e-06, + "loss": 4.5875, + "step": 70 + }, + { + "epoch": 0.03854505971769816, + "grad_norm": 2.987709377109457, + "learning_rate": 5.126353790613719e-06, + "loss": 5.5784, + "step": 71 + }, + { + "epoch": 0.03908794788273615, + "grad_norm": 2.618514747422557, + "learning_rate": 5.1985559566787e-06, + "loss": 3.9901, + "step": 72 + }, + { + "epoch": 0.039630836047774155, + "grad_norm": 3.729239515680902, + "learning_rate": 5.270758122743683e-06, + "loss": 4.6192, + "step": 73 + }, + { + "epoch": 0.04017372421281216, + "grad_norm": 2.488658936956188, + "learning_rate": 5.342960288808665e-06, + "loss": 3.9889, + "step": 74 + }, + { + "epoch": 0.04071661237785016, + "grad_norm": 2.6055640979236223, + "learning_rate": 5.415162454873647e-06, + "loss": 4.3166, + "step": 75 + }, + { + "epoch": 0.04125950054288816, + "grad_norm": 2.091905540809132, + "learning_rate": 5.487364620938629e-06, + "loss": 3.5848, + "step": 76 + }, + { + "epoch": 0.041802388707926165, + "grad_norm": 4.153140315460812, + "learning_rate": 5.559566787003611e-06, + "loss": 4.4889, + "step": 77 + }, + { + "epoch": 0.04234527687296417, + "grad_norm": 4.828033123535034, + "learning_rate": 5.631768953068592e-06, + "loss": 4.886, + "step": 78 + }, + { + "epoch": 0.04288816503800217, + "grad_norm": 3.106146715385438, + "learning_rate": 5.7039711191335744e-06, + "loss": 4.3698, + "step": 79 + }, + { + "epoch": 0.04343105320304017, + "grad_norm": 4.458698539182903, + "learning_rate": 5.776173285198557e-06, + "loss": 4.7018, + "step": 80 + }, + { + "epoch": 0.043973941368078175, + "grad_norm": 4.619902873416999, + "learning_rate": 5.848375451263538e-06, + "loss": 4.3449, + "step": 81 + }, + { + "epoch": 0.04451682953311618, + "grad_norm": 2.411841615517987, + "learning_rate": 5.92057761732852e-06, + "loss": 3.9952, + "step": 82 + }, + { + "epoch": 0.04505971769815418, + "grad_norm": 3.6594837742651705, + "learning_rate": 5.992779783393502e-06, + "loss": 4.1308, + "step": 83 + }, + { + "epoch": 0.04560260586319218, + "grad_norm": 2.3760888273966074, + "learning_rate": 6.064981949458484e-06, + "loss": 4.0315, + "step": 84 + }, + { + "epoch": 0.046145494028230184, + "grad_norm": 4.579655616004927, + "learning_rate": 6.137184115523466e-06, + "loss": 4.1051, + "step": 85 + }, + { + "epoch": 0.04668838219326819, + "grad_norm": 3.5909339610827846, + "learning_rate": 6.209386281588449e-06, + "loss": 4.3126, + "step": 86 + }, + { + "epoch": 0.04723127035830619, + "grad_norm": 3.1030422711991625, + "learning_rate": 6.28158844765343e-06, + "loss": 4.3038, + "step": 87 + }, + { + "epoch": 0.04777415852334419, + "grad_norm": 4.134211324790428, + "learning_rate": 6.3537906137184125e-06, + "loss": 4.906, + "step": 88 + }, + { + "epoch": 0.048317046688382194, + "grad_norm": 3.538946913288816, + "learning_rate": 6.425992779783395e-06, + "loss": 4.2182, + "step": 89 + }, + { + "epoch": 0.048859934853420196, + "grad_norm": 1.9828152543749473, + "learning_rate": 6.498194945848376e-06, + "loss": 3.8487, + "step": 90 + }, + { + "epoch": 0.0494028230184582, + "grad_norm": 2.127969936938475, + "learning_rate": 6.570397111913358e-06, + "loss": 3.9221, + "step": 91 + }, + { + "epoch": 0.0499457111834962, + "grad_norm": 2.740653026382279, + "learning_rate": 6.6425992779783395e-06, + "loss": 4.4661, + "step": 92 + }, + { + "epoch": 0.050488599348534204, + "grad_norm": 3.278496788216742, + "learning_rate": 6.714801444043322e-06, + "loss": 4.1995, + "step": 93 + }, + { + "epoch": 0.051031487513572206, + "grad_norm": 1.5455749978381657, + "learning_rate": 6.787003610108304e-06, + "loss": 3.3941, + "step": 94 + }, + { + "epoch": 0.05157437567861021, + "grad_norm": 2.6938438745410975, + "learning_rate": 6.859205776173285e-06, + "loss": 3.9863, + "step": 95 + }, + { + "epoch": 0.05211726384364821, + "grad_norm": 2.754193515919841, + "learning_rate": 6.9314079422382674e-06, + "loss": 4.1621, + "step": 96 + }, + { + "epoch": 0.05266015200868621, + "grad_norm": 1.787757508585929, + "learning_rate": 7.00361010830325e-06, + "loss": 3.728, + "step": 97 + }, + { + "epoch": 0.053203040173724216, + "grad_norm": 1.6756343581979376, + "learning_rate": 7.075812274368231e-06, + "loss": 3.4483, + "step": 98 + }, + { + "epoch": 0.05374592833876222, + "grad_norm": 3.0418610323064317, + "learning_rate": 7.148014440433214e-06, + "loss": 4.1131, + "step": 99 + }, + { + "epoch": 0.05428881650380022, + "grad_norm": 2.710846119482301, + "learning_rate": 7.220216606498196e-06, + "loss": 5.0449, + "step": 100 + }, + { + "epoch": 0.054831704668838216, + "grad_norm": 1.906382949957388, + "learning_rate": 7.2924187725631776e-06, + "loss": 3.3181, + "step": 101 + }, + { + "epoch": 0.05537459283387622, + "grad_norm": 2.4669878697837, + "learning_rate": 7.36462093862816e-06, + "loss": 3.7996, + "step": 102 + }, + { + "epoch": 0.05591748099891422, + "grad_norm": 2.3755842712374586, + "learning_rate": 7.436823104693142e-06, + "loss": 4.0321, + "step": 103 + }, + { + "epoch": 0.05646036916395222, + "grad_norm": 2.032793864149989, + "learning_rate": 7.509025270758123e-06, + "loss": 3.4121, + "step": 104 + }, + { + "epoch": 0.057003257328990226, + "grad_norm": 2.1561554352810686, + "learning_rate": 7.5812274368231055e-06, + "loss": 4.2585, + "step": 105 + }, + { + "epoch": 0.05754614549402823, + "grad_norm": 1.9627108654626784, + "learning_rate": 7.653429602888087e-06, + "loss": 4.3085, + "step": 106 + }, + { + "epoch": 0.05808903365906623, + "grad_norm": 1.8622753095278954, + "learning_rate": 7.72563176895307e-06, + "loss": 3.8459, + "step": 107 + }, + { + "epoch": 0.05863192182410423, + "grad_norm": 1.4438765470100654, + "learning_rate": 7.797833935018051e-06, + "loss": 3.4914, + "step": 108 + }, + { + "epoch": 0.059174809989142235, + "grad_norm": 1.6164085905580508, + "learning_rate": 7.870036101083033e-06, + "loss": 3.4958, + "step": 109 + }, + { + "epoch": 0.05971769815418024, + "grad_norm": 1.6075014519578008, + "learning_rate": 7.942238267148014e-06, + "loss": 3.202, + "step": 110 + }, + { + "epoch": 0.06026058631921824, + "grad_norm": 1.7630994896021268, + "learning_rate": 8.014440433212997e-06, + "loss": 3.0916, + "step": 111 + }, + { + "epoch": 0.06080347448425624, + "grad_norm": 1.7600666006428431, + "learning_rate": 8.086642599277978e-06, + "loss": 3.5791, + "step": 112 + }, + { + "epoch": 0.061346362649294245, + "grad_norm": 2.324243905399748, + "learning_rate": 8.158844765342961e-06, + "loss": 3.6247, + "step": 113 + }, + { + "epoch": 0.06188925081433225, + "grad_norm": 1.2715449173786617, + "learning_rate": 8.231046931407943e-06, + "loss": 3.2381, + "step": 114 + }, + { + "epoch": 0.06243213897937025, + "grad_norm": 1.9029014476619814, + "learning_rate": 8.303249097472926e-06, + "loss": 3.5269, + "step": 115 + }, + { + "epoch": 0.06297502714440825, + "grad_norm": 1.8782574049467182, + "learning_rate": 8.375451263537907e-06, + "loss": 2.8698, + "step": 116 + }, + { + "epoch": 0.06351791530944625, + "grad_norm": 1.8522383705853351, + "learning_rate": 8.447653429602888e-06, + "loss": 3.5003, + "step": 117 + }, + { + "epoch": 0.06406080347448426, + "grad_norm": 1.6480852413427443, + "learning_rate": 8.519855595667871e-06, + "loss": 3.5665, + "step": 118 + }, + { + "epoch": 0.06460369163952226, + "grad_norm": 2.1628448626122685, + "learning_rate": 8.592057761732853e-06, + "loss": 3.343, + "step": 119 + }, + { + "epoch": 0.06514657980456026, + "grad_norm": 1.568456113423694, + "learning_rate": 8.664259927797834e-06, + "loss": 2.474, + "step": 120 + }, + { + "epoch": 0.06568946796959826, + "grad_norm": 1.8054357065473743, + "learning_rate": 8.736462093862817e-06, + "loss": 2.9228, + "step": 121 + }, + { + "epoch": 0.06623235613463627, + "grad_norm": 1.9418450135728782, + "learning_rate": 8.808664259927798e-06, + "loss": 3.1995, + "step": 122 + }, + { + "epoch": 0.06677524429967427, + "grad_norm": 1.321851521267606, + "learning_rate": 8.88086642599278e-06, + "loss": 3.5515, + "step": 123 + }, + { + "epoch": 0.06731813246471227, + "grad_norm": 1.390198419578882, + "learning_rate": 8.953068592057763e-06, + "loss": 2.7741, + "step": 124 + }, + { + "epoch": 0.06786102062975027, + "grad_norm": 2.119370139523294, + "learning_rate": 9.025270758122744e-06, + "loss": 3.3427, + "step": 125 + }, + { + "epoch": 0.06840390879478828, + "grad_norm": 2.042189734142807, + "learning_rate": 9.097472924187727e-06, + "loss": 3.1414, + "step": 126 + }, + { + "epoch": 0.06894679695982628, + "grad_norm": 2.0399787373405123, + "learning_rate": 9.169675090252709e-06, + "loss": 3.0862, + "step": 127 + }, + { + "epoch": 0.06948968512486428, + "grad_norm": 2.0523727295223915, + "learning_rate": 9.24187725631769e-06, + "loss": 2.6461, + "step": 128 + }, + { + "epoch": 0.07003257328990228, + "grad_norm": 1.459190908283032, + "learning_rate": 9.314079422382673e-06, + "loss": 2.9907, + "step": 129 + }, + { + "epoch": 0.07057546145494029, + "grad_norm": 1.7403521896723462, + "learning_rate": 9.386281588447654e-06, + "loss": 3.0681, + "step": 130 + }, + { + "epoch": 0.07111834961997829, + "grad_norm": 2.2283604183714383, + "learning_rate": 9.458483754512636e-06, + "loss": 3.4619, + "step": 131 + }, + { + "epoch": 0.07166123778501629, + "grad_norm": 1.3291461832292721, + "learning_rate": 9.530685920577619e-06, + "loss": 3.096, + "step": 132 + }, + { + "epoch": 0.0722041259500543, + "grad_norm": 1.7839357457835638, + "learning_rate": 9.6028880866426e-06, + "loss": 2.7512, + "step": 133 + }, + { + "epoch": 0.0727470141150923, + "grad_norm": 1.7836465130823984, + "learning_rate": 9.675090252707581e-06, + "loss": 3.3118, + "step": 134 + }, + { + "epoch": 0.0732899022801303, + "grad_norm": 2.0407172747925135, + "learning_rate": 9.747292418772564e-06, + "loss": 2.1032, + "step": 135 + }, + { + "epoch": 0.0738327904451683, + "grad_norm": 2.335226908691354, + "learning_rate": 9.819494584837546e-06, + "loss": 2.9039, + "step": 136 + }, + { + "epoch": 0.0743756786102063, + "grad_norm": 2.567081845339778, + "learning_rate": 9.891696750902527e-06, + "loss": 3.1071, + "step": 137 + }, + { + "epoch": 0.0749185667752443, + "grad_norm": 1.9549184920601244, + "learning_rate": 9.96389891696751e-06, + "loss": 2.8253, + "step": 138 + }, + { + "epoch": 0.07546145494028231, + "grad_norm": 1.6389815727698025, + "learning_rate": 1.0036101083032491e-05, + "loss": 2.5701, + "step": 139 + }, + { + "epoch": 0.07600434310532031, + "grad_norm": 1.6558924195158018, + "learning_rate": 1.0108303249097473e-05, + "loss": 2.6428, + "step": 140 + }, + { + "epoch": 0.07654723127035831, + "grad_norm": 2.0257343427634087, + "learning_rate": 1.0180505415162456e-05, + "loss": 2.2828, + "step": 141 + }, + { + "epoch": 0.07709011943539631, + "grad_norm": 1.879991343961764, + "learning_rate": 1.0252707581227437e-05, + "loss": 3.1436, + "step": 142 + }, + { + "epoch": 0.07763300760043432, + "grad_norm": 1.8718236498893788, + "learning_rate": 1.032490974729242e-05, + "loss": 3.2434, + "step": 143 + }, + { + "epoch": 0.0781758957654723, + "grad_norm": 1.8252190287923398, + "learning_rate": 1.03971119133574e-05, + "loss": 2.4851, + "step": 144 + }, + { + "epoch": 0.07871878393051031, + "grad_norm": 2.6564413093838213, + "learning_rate": 1.0469314079422383e-05, + "loss": 2.7315, + "step": 145 + }, + { + "epoch": 0.07926167209554831, + "grad_norm": 2.054556112475997, + "learning_rate": 1.0541516245487366e-05, + "loss": 2.647, + "step": 146 + }, + { + "epoch": 0.07980456026058631, + "grad_norm": 1.072343898976271, + "learning_rate": 1.0613718411552347e-05, + "loss": 2.5868, + "step": 147 + }, + { + "epoch": 0.08034744842562432, + "grad_norm": 1.1443585178571194, + "learning_rate": 1.068592057761733e-05, + "loss": 2.3247, + "step": 148 + }, + { + "epoch": 0.08089033659066232, + "grad_norm": 2.0424670761496646, + "learning_rate": 1.0758122743682312e-05, + "loss": 1.989, + "step": 149 + }, + { + "epoch": 0.08143322475570032, + "grad_norm": 1.3715615162918033, + "learning_rate": 1.0830324909747295e-05, + "loss": 2.3989, + "step": 150 + }, + { + "epoch": 0.08197611292073832, + "grad_norm": 1.3740116291532023, + "learning_rate": 1.0902527075812274e-05, + "loss": 1.8726, + "step": 151 + }, + { + "epoch": 0.08251900108577633, + "grad_norm": 2.5824759155435655, + "learning_rate": 1.0974729241877257e-05, + "loss": 2.3019, + "step": 152 + }, + { + "epoch": 0.08306188925081433, + "grad_norm": 1.5584514490971877, + "learning_rate": 1.1046931407942239e-05, + "loss": 2.4209, + "step": 153 + }, + { + "epoch": 0.08360477741585233, + "grad_norm": 1.6184456176727229, + "learning_rate": 1.1119133574007222e-05, + "loss": 2.2377, + "step": 154 + }, + { + "epoch": 0.08414766558089033, + "grad_norm": 2.6665151568462586, + "learning_rate": 1.1191335740072201e-05, + "loss": 2.6381, + "step": 155 + }, + { + "epoch": 0.08469055374592833, + "grad_norm": 1.442929928712006, + "learning_rate": 1.1263537906137184e-05, + "loss": 2.3405, + "step": 156 + }, + { + "epoch": 0.08523344191096634, + "grad_norm": 1.906547769868643, + "learning_rate": 1.1335740072202166e-05, + "loss": 2.4391, + "step": 157 + }, + { + "epoch": 0.08577633007600434, + "grad_norm": 1.2153727696829317, + "learning_rate": 1.1407942238267149e-05, + "loss": 1.8401, + "step": 158 + }, + { + "epoch": 0.08631921824104234, + "grad_norm": 1.4810729195972283, + "learning_rate": 1.1480144404332132e-05, + "loss": 2.7248, + "step": 159 + }, + { + "epoch": 0.08686210640608034, + "grad_norm": 2.9401118999625058, + "learning_rate": 1.1552346570397113e-05, + "loss": 2.6582, + "step": 160 + }, + { + "epoch": 0.08740499457111835, + "grad_norm": 1.145106458573816, + "learning_rate": 1.1624548736462096e-05, + "loss": 2.6648, + "step": 161 + }, + { + "epoch": 0.08794788273615635, + "grad_norm": 1.9228128025198137, + "learning_rate": 1.1696750902527076e-05, + "loss": 2.3702, + "step": 162 + }, + { + "epoch": 0.08849077090119435, + "grad_norm": 1.318978856783607, + "learning_rate": 1.1768953068592059e-05, + "loss": 1.9388, + "step": 163 + }, + { + "epoch": 0.08903365906623235, + "grad_norm": 1.58104775083856, + "learning_rate": 1.184115523465704e-05, + "loss": 2.0774, + "step": 164 + }, + { + "epoch": 0.08957654723127036, + "grad_norm": 1.8206646323260607, + "learning_rate": 1.1913357400722023e-05, + "loss": 2.5221, + "step": 165 + }, + { + "epoch": 0.09011943539630836, + "grad_norm": 1.0815960351491627, + "learning_rate": 1.1985559566787005e-05, + "loss": 2.128, + "step": 166 + }, + { + "epoch": 0.09066232356134636, + "grad_norm": 1.3341486432977188, + "learning_rate": 1.2057761732851988e-05, + "loss": 1.9418, + "step": 167 + }, + { + "epoch": 0.09120521172638436, + "grad_norm": 1.3772486036170464, + "learning_rate": 1.2129963898916967e-05, + "loss": 1.84, + "step": 168 + }, + { + "epoch": 0.09174809989142237, + "grad_norm": 1.42874152479176, + "learning_rate": 1.220216606498195e-05, + "loss": 2.2172, + "step": 169 + }, + { + "epoch": 0.09229098805646037, + "grad_norm": 1.9691394708584433, + "learning_rate": 1.2274368231046932e-05, + "loss": 2.9247, + "step": 170 + }, + { + "epoch": 0.09283387622149837, + "grad_norm": 1.4773974635910447, + "learning_rate": 1.2346570397111915e-05, + "loss": 2.1028, + "step": 171 + }, + { + "epoch": 0.09337676438653637, + "grad_norm": 1.5937344901476087, + "learning_rate": 1.2418772563176898e-05, + "loss": 2.4073, + "step": 172 + }, + { + "epoch": 0.09391965255157438, + "grad_norm": 1.630783959399718, + "learning_rate": 1.2490974729241878e-05, + "loss": 2.6978, + "step": 173 + }, + { + "epoch": 0.09446254071661238, + "grad_norm": 1.44736447594164, + "learning_rate": 1.256317689530686e-05, + "loss": 1.8851, + "step": 174 + }, + { + "epoch": 0.09500542888165038, + "grad_norm": 1.3909459856510267, + "learning_rate": 1.2635379061371842e-05, + "loss": 1.576, + "step": 175 + }, + { + "epoch": 0.09554831704668838, + "grad_norm": 1.7838393677575515, + "learning_rate": 1.2707581227436825e-05, + "loss": 2.3895, + "step": 176 + }, + { + "epoch": 0.09609120521172639, + "grad_norm": 1.1294127385799255, + "learning_rate": 1.2779783393501806e-05, + "loss": 2.4855, + "step": 177 + }, + { + "epoch": 0.09663409337676439, + "grad_norm": 1.259662421149648, + "learning_rate": 1.285198555956679e-05, + "loss": 2.5469, + "step": 178 + }, + { + "epoch": 0.09717698154180239, + "grad_norm": 1.4621367635995284, + "learning_rate": 1.2924187725631769e-05, + "loss": 1.9015, + "step": 179 + }, + { + "epoch": 0.09771986970684039, + "grad_norm": 1.6960126954903645, + "learning_rate": 1.2996389891696752e-05, + "loss": 2.9545, + "step": 180 + }, + { + "epoch": 0.0982627578718784, + "grad_norm": 1.5419022503429414, + "learning_rate": 1.3068592057761733e-05, + "loss": 1.9912, + "step": 181 + }, + { + "epoch": 0.0988056460369164, + "grad_norm": 1.4073422496120207, + "learning_rate": 1.3140794223826716e-05, + "loss": 2.3353, + "step": 182 + }, + { + "epoch": 0.0993485342019544, + "grad_norm": 1.2881659254471862, + "learning_rate": 1.3212996389891696e-05, + "loss": 2.7784, + "step": 183 + }, + { + "epoch": 0.0998914223669924, + "grad_norm": 1.361969663712106, + "learning_rate": 1.3285198555956679e-05, + "loss": 2.523, + "step": 184 + }, + { + "epoch": 0.1004343105320304, + "grad_norm": 2.5841731119141316, + "learning_rate": 1.3357400722021662e-05, + "loss": 2.43, + "step": 185 + }, + { + "epoch": 0.10097719869706841, + "grad_norm": 1.3203537847943392, + "learning_rate": 1.3429602888086643e-05, + "loss": 1.9404, + "step": 186 + }, + { + "epoch": 0.10152008686210641, + "grad_norm": 1.393097901493087, + "learning_rate": 1.3501805054151626e-05, + "loss": 2.6202, + "step": 187 + }, + { + "epoch": 0.10206297502714441, + "grad_norm": 0.9414377380652101, + "learning_rate": 1.3574007220216608e-05, + "loss": 2.0879, + "step": 188 + }, + { + "epoch": 0.10260586319218241, + "grad_norm": 1.6547581806701488, + "learning_rate": 1.3646209386281591e-05, + "loss": 2.8912, + "step": 189 + }, + { + "epoch": 0.10314875135722042, + "grad_norm": 1.1236261864015666, + "learning_rate": 1.371841155234657e-05, + "loss": 2.1802, + "step": 190 + }, + { + "epoch": 0.10369163952225842, + "grad_norm": 1.1766384878462062, + "learning_rate": 1.3790613718411554e-05, + "loss": 1.9127, + "step": 191 + }, + { + "epoch": 0.10423452768729642, + "grad_norm": 2.0453724866488656, + "learning_rate": 1.3862815884476535e-05, + "loss": 1.8522, + "step": 192 + }, + { + "epoch": 0.10477741585233442, + "grad_norm": 1.550964370734908, + "learning_rate": 1.3935018050541518e-05, + "loss": 2.5852, + "step": 193 + }, + { + "epoch": 0.10532030401737243, + "grad_norm": 1.303227180334529, + "learning_rate": 1.40072202166065e-05, + "loss": 2.407, + "step": 194 + }, + { + "epoch": 0.10586319218241043, + "grad_norm": 1.2009193115283474, + "learning_rate": 1.4079422382671482e-05, + "loss": 1.8978, + "step": 195 + }, + { + "epoch": 0.10640608034744843, + "grad_norm": 1.353648963417667, + "learning_rate": 1.4151624548736462e-05, + "loss": 2.283, + "step": 196 + }, + { + "epoch": 0.10694896851248643, + "grad_norm": 1.266154745219424, + "learning_rate": 1.4223826714801445e-05, + "loss": 2.0679, + "step": 197 + }, + { + "epoch": 0.10749185667752444, + "grad_norm": 4.186447043886807, + "learning_rate": 1.4296028880866428e-05, + "loss": 1.9175, + "step": 198 + }, + { + "epoch": 0.10803474484256244, + "grad_norm": 1.147065238667786, + "learning_rate": 1.436823104693141e-05, + "loss": 2.2748, + "step": 199 + }, + { + "epoch": 0.10857763300760044, + "grad_norm": 1.2760699256137895, + "learning_rate": 1.4440433212996392e-05, + "loss": 2.3031, + "step": 200 + }, + { + "epoch": 0.10912052117263844, + "grad_norm": 0.9452152631580967, + "learning_rate": 1.4512635379061372e-05, + "loss": 1.4833, + "step": 201 + }, + { + "epoch": 0.10966340933767643, + "grad_norm": 1.3980490034032074, + "learning_rate": 1.4584837545126355e-05, + "loss": 1.5865, + "step": 202 + }, + { + "epoch": 0.11020629750271443, + "grad_norm": 1.6440838011007093, + "learning_rate": 1.4657039711191336e-05, + "loss": 1.632, + "step": 203 + }, + { + "epoch": 0.11074918566775244, + "grad_norm": 1.3969764794458068, + "learning_rate": 1.472924187725632e-05, + "loss": 1.4301, + "step": 204 + }, + { + "epoch": 0.11129207383279044, + "grad_norm": 1.223663090997016, + "learning_rate": 1.48014440433213e-05, + "loss": 2.3993, + "step": 205 + }, + { + "epoch": 0.11183496199782844, + "grad_norm": 1.183881027275754, + "learning_rate": 1.4873646209386284e-05, + "loss": 1.4987, + "step": 206 + }, + { + "epoch": 0.11237785016286644, + "grad_norm": 1.3788483357578956, + "learning_rate": 1.4945848375451264e-05, + "loss": 2.7742, + "step": 207 + }, + { + "epoch": 0.11292073832790445, + "grad_norm": 1.1260622559595288, + "learning_rate": 1.5018050541516247e-05, + "loss": 1.755, + "step": 208 + }, + { + "epoch": 0.11346362649294245, + "grad_norm": 1.8439396062381286, + "learning_rate": 1.5090252707581228e-05, + "loss": 2.9875, + "step": 209 + }, + { + "epoch": 0.11400651465798045, + "grad_norm": 1.2789617727573501, + "learning_rate": 1.5162454873646211e-05, + "loss": 2.0119, + "step": 210 + }, + { + "epoch": 0.11454940282301845, + "grad_norm": 1.3385913561767113, + "learning_rate": 1.5234657039711192e-05, + "loss": 1.7693, + "step": 211 + }, + { + "epoch": 0.11509229098805646, + "grad_norm": 1.049162702021749, + "learning_rate": 1.5306859205776174e-05, + "loss": 1.7541, + "step": 212 + }, + { + "epoch": 0.11563517915309446, + "grad_norm": 1.0797407129441892, + "learning_rate": 1.537906137184116e-05, + "loss": 2.0427, + "step": 213 + }, + { + "epoch": 0.11617806731813246, + "grad_norm": 1.3661735952693994, + "learning_rate": 1.545126353790614e-05, + "loss": 2.722, + "step": 214 + }, + { + "epoch": 0.11672095548317046, + "grad_norm": 1.7084933140661187, + "learning_rate": 1.552346570397112e-05, + "loss": 2.2125, + "step": 215 + }, + { + "epoch": 0.11726384364820847, + "grad_norm": 1.3814428141700028, + "learning_rate": 1.5595667870036102e-05, + "loss": 2.448, + "step": 216 + }, + { + "epoch": 0.11780673181324647, + "grad_norm": 1.298309196615407, + "learning_rate": 1.5667870036101084e-05, + "loss": 2.2706, + "step": 217 + }, + { + "epoch": 0.11834961997828447, + "grad_norm": 1.4553148411944858, + "learning_rate": 1.5740072202166065e-05, + "loss": 2.0087, + "step": 218 + }, + { + "epoch": 0.11889250814332247, + "grad_norm": 1.5083115470870359, + "learning_rate": 1.581227436823105e-05, + "loss": 1.7179, + "step": 219 + }, + { + "epoch": 0.11943539630836048, + "grad_norm": 1.2335179878839475, + "learning_rate": 1.5884476534296028e-05, + "loss": 2.5337, + "step": 220 + }, + { + "epoch": 0.11997828447339848, + "grad_norm": 1.1729282902419653, + "learning_rate": 1.5956678700361013e-05, + "loss": 1.9604, + "step": 221 + }, + { + "epoch": 0.12052117263843648, + "grad_norm": 1.5700347334422509, + "learning_rate": 1.6028880866425994e-05, + "loss": 2.3655, + "step": 222 + }, + { + "epoch": 0.12106406080347448, + "grad_norm": 1.435764746128967, + "learning_rate": 1.6101083032490975e-05, + "loss": 2.109, + "step": 223 + }, + { + "epoch": 0.12160694896851248, + "grad_norm": 1.0866216936301165, + "learning_rate": 1.6173285198555957e-05, + "loss": 2.3334, + "step": 224 + }, + { + "epoch": 0.12214983713355049, + "grad_norm": 1.8060789361277199, + "learning_rate": 1.624548736462094e-05, + "loss": 1.6605, + "step": 225 + }, + { + "epoch": 0.12269272529858849, + "grad_norm": 1.539101200174871, + "learning_rate": 1.6317689530685923e-05, + "loss": 2.2597, + "step": 226 + }, + { + "epoch": 0.12323561346362649, + "grad_norm": 1.7392984350243024, + "learning_rate": 1.6389891696750904e-05, + "loss": 2.5979, + "step": 227 + }, + { + "epoch": 0.1237785016286645, + "grad_norm": 1.6792982939949113, + "learning_rate": 1.6462093862815885e-05, + "loss": 2.0626, + "step": 228 + }, + { + "epoch": 0.1243213897937025, + "grad_norm": 2.067856070276011, + "learning_rate": 1.6534296028880867e-05, + "loss": 2.056, + "step": 229 + }, + { + "epoch": 0.1248642779587405, + "grad_norm": 1.2773224600362139, + "learning_rate": 1.660649819494585e-05, + "loss": 1.5852, + "step": 230 + }, + { + "epoch": 0.1254071661237785, + "grad_norm": 1.381245750855955, + "learning_rate": 1.6678700361010833e-05, + "loss": 2.0675, + "step": 231 + }, + { + "epoch": 0.1259500542888165, + "grad_norm": 1.4215755121774605, + "learning_rate": 1.6750902527075814e-05, + "loss": 2.1002, + "step": 232 + }, + { + "epoch": 0.1264929424538545, + "grad_norm": 1.4581673093175722, + "learning_rate": 1.6823104693140795e-05, + "loss": 1.9279, + "step": 233 + }, + { + "epoch": 0.1270358306188925, + "grad_norm": 1.2656102558648035, + "learning_rate": 1.6895306859205777e-05, + "loss": 1.4913, + "step": 234 + }, + { + "epoch": 0.1275787187839305, + "grad_norm": 1.3985246778048182, + "learning_rate": 1.6967509025270758e-05, + "loss": 2.1629, + "step": 235 + }, + { + "epoch": 0.1281216069489685, + "grad_norm": 1.757194392177075, + "learning_rate": 1.7039711191335743e-05, + "loss": 2.5688, + "step": 236 + }, + { + "epoch": 0.12866449511400652, + "grad_norm": 1.3766112824147358, + "learning_rate": 1.711191335740072e-05, + "loss": 2.3344, + "step": 237 + }, + { + "epoch": 0.12920738327904452, + "grad_norm": 1.2985043105700416, + "learning_rate": 1.7184115523465706e-05, + "loss": 2.0626, + "step": 238 + }, + { + "epoch": 0.12975027144408252, + "grad_norm": 1.715796674825951, + "learning_rate": 1.7256317689530687e-05, + "loss": 2.1559, + "step": 239 + }, + { + "epoch": 0.13029315960912052, + "grad_norm": 1.6374209306607368, + "learning_rate": 1.7328519855595668e-05, + "loss": 2.0801, + "step": 240 + }, + { + "epoch": 0.13083604777415853, + "grad_norm": 1.244682506367769, + "learning_rate": 1.7400722021660653e-05, + "loss": 1.6125, + "step": 241 + }, + { + "epoch": 0.13137893593919653, + "grad_norm": 1.3001838781366861, + "learning_rate": 1.7472924187725634e-05, + "loss": 2.0934, + "step": 242 + }, + { + "epoch": 0.13192182410423453, + "grad_norm": 1.1600879910417528, + "learning_rate": 1.7545126353790616e-05, + "loss": 1.7862, + "step": 243 + }, + { + "epoch": 0.13246471226927253, + "grad_norm": 1.3503312968411385, + "learning_rate": 1.7617328519855597e-05, + "loss": 2.1636, + "step": 244 + }, + { + "epoch": 0.13300760043431054, + "grad_norm": 2.1299761628528167, + "learning_rate": 1.768953068592058e-05, + "loss": 2.4826, + "step": 245 + }, + { + "epoch": 0.13355048859934854, + "grad_norm": 1.3232052544377066, + "learning_rate": 1.776173285198556e-05, + "loss": 2.0497, + "step": 246 + }, + { + "epoch": 0.13409337676438654, + "grad_norm": 1.5079892467642568, + "learning_rate": 1.7833935018050544e-05, + "loss": 2.4552, + "step": 247 + }, + { + "epoch": 0.13463626492942454, + "grad_norm": 1.716355318914646, + "learning_rate": 1.7906137184115526e-05, + "loss": 1.5309, + "step": 248 + }, + { + "epoch": 0.13517915309446255, + "grad_norm": 2.1523878955969193, + "learning_rate": 1.7978339350180507e-05, + "loss": 2.7403, + "step": 249 + }, + { + "epoch": 0.13572204125950055, + "grad_norm": 1.8169301927997485, + "learning_rate": 1.805054151624549e-05, + "loss": 2.3324, + "step": 250 + }, + { + "epoch": 0.13626492942453855, + "grad_norm": 1.4464924096148426, + "learning_rate": 1.812274368231047e-05, + "loss": 1.8399, + "step": 251 + }, + { + "epoch": 0.13680781758957655, + "grad_norm": 1.9818298103877654, + "learning_rate": 1.8194945848375454e-05, + "loss": 2.4237, + "step": 252 + }, + { + "epoch": 0.13735070575461455, + "grad_norm": 1.6371003739671923, + "learning_rate": 1.8267148014440436e-05, + "loss": 2.0477, + "step": 253 + }, + { + "epoch": 0.13789359391965256, + "grad_norm": 1.2944832148952543, + "learning_rate": 1.8339350180505417e-05, + "loss": 2.2269, + "step": 254 + }, + { + "epoch": 0.13843648208469056, + "grad_norm": 1.4284289887933213, + "learning_rate": 1.84115523465704e-05, + "loss": 2.2465, + "step": 255 + }, + { + "epoch": 0.13897937024972856, + "grad_norm": 1.4633226693151502, + "learning_rate": 1.848375451263538e-05, + "loss": 2.1582, + "step": 256 + }, + { + "epoch": 0.13952225841476656, + "grad_norm": 1.6169546473154692, + "learning_rate": 1.855595667870036e-05, + "loss": 1.9526, + "step": 257 + }, + { + "epoch": 0.14006514657980457, + "grad_norm": 1.47082651679944, + "learning_rate": 1.8628158844765346e-05, + "loss": 2.0314, + "step": 258 + }, + { + "epoch": 0.14060803474484257, + "grad_norm": 1.2279964838302115, + "learning_rate": 1.8700361010830327e-05, + "loss": 1.4345, + "step": 259 + }, + { + "epoch": 0.14115092290988057, + "grad_norm": 1.5727648286246085, + "learning_rate": 1.877256317689531e-05, + "loss": 2.7371, + "step": 260 + }, + { + "epoch": 0.14169381107491857, + "grad_norm": 1.5536143039045338, + "learning_rate": 1.884476534296029e-05, + "loss": 1.6691, + "step": 261 + }, + { + "epoch": 0.14223669923995658, + "grad_norm": 1.504953118413033, + "learning_rate": 1.891696750902527e-05, + "loss": 1.8422, + "step": 262 + }, + { + "epoch": 0.14277958740499458, + "grad_norm": 1.6092868309857633, + "learning_rate": 1.8989169675090253e-05, + "loss": 1.6446, + "step": 263 + }, + { + "epoch": 0.14332247557003258, + "grad_norm": 1.434484546738067, + "learning_rate": 1.9061371841155237e-05, + "loss": 1.5945, + "step": 264 + }, + { + "epoch": 0.14386536373507058, + "grad_norm": 1.3472204852366787, + "learning_rate": 1.913357400722022e-05, + "loss": 1.4511, + "step": 265 + }, + { + "epoch": 0.1444082519001086, + "grad_norm": 2.493842366649246, + "learning_rate": 1.92057761732852e-05, + "loss": 2.1144, + "step": 266 + }, + { + "epoch": 0.1449511400651466, + "grad_norm": 1.3138428233046642, + "learning_rate": 1.927797833935018e-05, + "loss": 1.6592, + "step": 267 + }, + { + "epoch": 0.1454940282301846, + "grad_norm": 1.7201458829291465, + "learning_rate": 1.9350180505415163e-05, + "loss": 1.6895, + "step": 268 + }, + { + "epoch": 0.1460369163952226, + "grad_norm": 1.9777013455620547, + "learning_rate": 1.9422382671480147e-05, + "loss": 2.2628, + "step": 269 + }, + { + "epoch": 0.1465798045602606, + "grad_norm": 1.8559862870685497, + "learning_rate": 1.949458483754513e-05, + "loss": 1.3374, + "step": 270 + }, + { + "epoch": 0.1471226927252986, + "grad_norm": 1.919404449298866, + "learning_rate": 1.956678700361011e-05, + "loss": 2.3805, + "step": 271 + }, + { + "epoch": 0.1476655808903366, + "grad_norm": 1.4760010543412088, + "learning_rate": 1.963898916967509e-05, + "loss": 1.9791, + "step": 272 + }, + { + "epoch": 0.1482084690553746, + "grad_norm": 1.4491842831658432, + "learning_rate": 1.9711191335740073e-05, + "loss": 1.8119, + "step": 273 + }, + { + "epoch": 0.1487513572204126, + "grad_norm": 1.4098796806726177, + "learning_rate": 1.9783393501805054e-05, + "loss": 1.7398, + "step": 274 + }, + { + "epoch": 0.1492942453854506, + "grad_norm": 1.5009369942229867, + "learning_rate": 1.985559566787004e-05, + "loss": 1.5194, + "step": 275 + }, + { + "epoch": 0.1498371335504886, + "grad_norm": 1.5073462671342261, + "learning_rate": 1.992779783393502e-05, + "loss": 1.8125, + "step": 276 + }, + { + "epoch": 0.1503800217155266, + "grad_norm": 1.9315604890805302, + "learning_rate": 2e-05, + "loss": 1.5429, + "step": 277 + }, + { + "epoch": 0.15092290988056462, + "grad_norm": 1.5059106777628657, + "learning_rate": 1.999999938159203e-05, + "loss": 1.6402, + "step": 278 + }, + { + "epoch": 0.15146579804560262, + "grad_norm": 1.7255630320192266, + "learning_rate": 1.9999997526368205e-05, + "loss": 2.2806, + "step": 279 + }, + { + "epoch": 0.15200868621064062, + "grad_norm": 1.7985407961594548, + "learning_rate": 1.999999443432874e-05, + "loss": 2.5213, + "step": 280 + }, + { + "epoch": 0.15255157437567862, + "grad_norm": 1.801471230172198, + "learning_rate": 1.999999010547403e-05, + "loss": 1.8362, + "step": 281 + }, + { + "epoch": 0.15309446254071662, + "grad_norm": 3.236881896883975, + "learning_rate": 1.999998453980461e-05, + "loss": 2.7148, + "step": 282 + }, + { + "epoch": 0.15363735070575463, + "grad_norm": 1.7343786362856708, + "learning_rate": 1.9999977737321156e-05, + "loss": 1.8195, + "step": 283 + }, + { + "epoch": 0.15418023887079263, + "grad_norm": 1.5006942215232695, + "learning_rate": 1.999996969802452e-05, + "loss": 1.8744, + "step": 284 + }, + { + "epoch": 0.15472312703583063, + "grad_norm": 4.171967924853298, + "learning_rate": 1.99999604219157e-05, + "loss": 2.4329, + "step": 285 + }, + { + "epoch": 0.15526601520086863, + "grad_norm": 1.7546486556973808, + "learning_rate": 1.9999949908995832e-05, + "loss": 1.3931, + "step": 286 + }, + { + "epoch": 0.15580890336590664, + "grad_norm": 2.0223702721102375, + "learning_rate": 1.999993815926622e-05, + "loss": 1.6704, + "step": 287 + }, + { + "epoch": 0.1563517915309446, + "grad_norm": 1.8736762651082781, + "learning_rate": 1.9999925172728324e-05, + "loss": 1.9895, + "step": 288 + }, + { + "epoch": 0.15689467969598261, + "grad_norm": 1.8878827291463698, + "learning_rate": 1.9999910949383742e-05, + "loss": 2.4939, + "step": 289 + }, + { + "epoch": 0.15743756786102062, + "grad_norm": 1.7277080078786893, + "learning_rate": 1.9999895489234234e-05, + "loss": 1.9265, + "step": 290 + }, + { + "epoch": 0.15798045602605862, + "grad_norm": 1.7402172969699101, + "learning_rate": 1.999987879228172e-05, + "loss": 1.8101, + "step": 291 + }, + { + "epoch": 0.15852334419109662, + "grad_norm": 2.077278713661093, + "learning_rate": 1.9999860858528257e-05, + "loss": 2.1763, + "step": 292 + }, + { + "epoch": 0.15906623235613462, + "grad_norm": 2.247624069618552, + "learning_rate": 1.9999841687976067e-05, + "loss": 1.9727, + "step": 293 + }, + { + "epoch": 0.15960912052117263, + "grad_norm": 1.6712153667666279, + "learning_rate": 1.999982128062752e-05, + "loss": 1.9921, + "step": 294 + }, + { + "epoch": 0.16015200868621063, + "grad_norm": 1.7126188149003763, + "learning_rate": 1.999979963648514e-05, + "loss": 1.9375, + "step": 295 + }, + { + "epoch": 0.16069489685124863, + "grad_norm": 1.949228534239786, + "learning_rate": 1.99997767555516e-05, + "loss": 2.1628, + "step": 296 + }, + { + "epoch": 0.16123778501628663, + "grad_norm": 1.8253305155157105, + "learning_rate": 1.9999752637829734e-05, + "loss": 1.8293, + "step": 297 + }, + { + "epoch": 0.16178067318132464, + "grad_norm": 1.5881557560095252, + "learning_rate": 1.9999727283322524e-05, + "loss": 1.3688, + "step": 298 + }, + { + "epoch": 0.16232356134636264, + "grad_norm": 2.4008202914116503, + "learning_rate": 1.9999700692033112e-05, + "loss": 1.5919, + "step": 299 + }, + { + "epoch": 0.16286644951140064, + "grad_norm": 1.610839420604411, + "learning_rate": 1.9999672863964778e-05, + "loss": 1.8724, + "step": 300 + }, + { + "epoch": 0.16340933767643864, + "grad_norm": 1.4663439746230646, + "learning_rate": 1.9999643799120964e-05, + "loss": 1.4628, + "step": 301 + }, + { + "epoch": 0.16395222584147665, + "grad_norm": 1.7051101810518976, + "learning_rate": 1.9999613497505272e-05, + "loss": 1.2916, + "step": 302 + }, + { + "epoch": 0.16449511400651465, + "grad_norm": 1.8967399031374104, + "learning_rate": 1.9999581959121443e-05, + "loss": 1.7559, + "step": 303 + }, + { + "epoch": 0.16503800217155265, + "grad_norm": 1.6206520761555416, + "learning_rate": 1.9999549183973382e-05, + "loss": 1.9196, + "step": 304 + }, + { + "epoch": 0.16558089033659065, + "grad_norm": 2.3545652344963544, + "learning_rate": 1.999951517206514e-05, + "loss": 1.2731, + "step": 305 + }, + { + "epoch": 0.16612377850162866, + "grad_norm": 1.7952736257095967, + "learning_rate": 1.9999479923400926e-05, + "loss": 2.0502, + "step": 306 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 1.7199420678903319, + "learning_rate": 1.99994434379851e-05, + "loss": 1.7059, + "step": 307 + }, + { + "epoch": 0.16720955483170466, + "grad_norm": 2.3318553047346, + "learning_rate": 1.9999405715822167e-05, + "loss": 1.4667, + "step": 308 + }, + { + "epoch": 0.16775244299674266, + "grad_norm": 1.618961362672766, + "learning_rate": 1.9999366756916804e-05, + "loss": 1.1866, + "step": 309 + }, + { + "epoch": 0.16829533116178066, + "grad_norm": 2.0805239071290114, + "learning_rate": 1.999932656127382e-05, + "loss": 1.7601, + "step": 310 + }, + { + "epoch": 0.16883821932681867, + "grad_norm": 1.782022796323131, + "learning_rate": 1.9999285128898193e-05, + "loss": 1.6207, + "step": 311 + }, + { + "epoch": 0.16938110749185667, + "grad_norm": 2.2438870289045028, + "learning_rate": 1.9999242459795045e-05, + "loss": 2.1329, + "step": 312 + }, + { + "epoch": 0.16992399565689467, + "grad_norm": 2.2550840823235587, + "learning_rate": 1.9999198553969652e-05, + "loss": 1.6055, + "step": 313 + }, + { + "epoch": 0.17046688382193267, + "grad_norm": 1.8012459587153988, + "learning_rate": 1.9999153411427445e-05, + "loss": 2.018, + "step": 314 + }, + { + "epoch": 0.17100977198697068, + "grad_norm": 1.8498840013582993, + "learning_rate": 1.9999107032174007e-05, + "loss": 1.8113, + "step": 315 + }, + { + "epoch": 0.17155266015200868, + "grad_norm": 2.2799557163953965, + "learning_rate": 1.9999059416215078e-05, + "loss": 1.9494, + "step": 316 + }, + { + "epoch": 0.17209554831704668, + "grad_norm": 2.383891228288174, + "learning_rate": 1.999901056355654e-05, + "loss": 1.8323, + "step": 317 + }, + { + "epoch": 0.17263843648208468, + "grad_norm": 1.9081077971876759, + "learning_rate": 1.9998960474204443e-05, + "loss": 1.5476, + "step": 318 + }, + { + "epoch": 0.1731813246471227, + "grad_norm": 1.9010783437994137, + "learning_rate": 1.999890914816498e-05, + "loss": 1.4602, + "step": 319 + }, + { + "epoch": 0.1737242128121607, + "grad_norm": 2.295573334252978, + "learning_rate": 1.9998856585444493e-05, + "loss": 1.0965, + "step": 320 + }, + { + "epoch": 0.1742671009771987, + "grad_norm": 2.1014332124066954, + "learning_rate": 1.999880278604949e-05, + "loss": 1.3726, + "step": 321 + }, + { + "epoch": 0.1748099891422367, + "grad_norm": 2.0418722317428206, + "learning_rate": 1.9998747749986625e-05, + "loss": 1.8875, + "step": 322 + }, + { + "epoch": 0.1753528773072747, + "grad_norm": 1.768897707524193, + "learning_rate": 1.99986914772627e-05, + "loss": 1.8841, + "step": 323 + }, + { + "epoch": 0.1758957654723127, + "grad_norm": 2.603335158781435, + "learning_rate": 1.9998633967884676e-05, + "loss": 1.4262, + "step": 324 + }, + { + "epoch": 0.1764386536373507, + "grad_norm": 2.1485447383197607, + "learning_rate": 1.999857522185967e-05, + "loss": 1.9283, + "step": 325 + }, + { + "epoch": 0.1769815418023887, + "grad_norm": 1.8398472624384716, + "learning_rate": 1.9998515239194945e-05, + "loss": 1.9089, + "step": 326 + }, + { + "epoch": 0.1775244299674267, + "grad_norm": 2.6988014220525374, + "learning_rate": 1.9998454019897918e-05, + "loss": 1.3636, + "step": 327 + }, + { + "epoch": 0.1780673181324647, + "grad_norm": 2.182794188412918, + "learning_rate": 1.9998391563976166e-05, + "loss": 1.7041, + "step": 328 + }, + { + "epoch": 0.1786102062975027, + "grad_norm": 1.970858895313084, + "learning_rate": 1.9998327871437405e-05, + "loss": 1.4917, + "step": 329 + }, + { + "epoch": 0.1791530944625407, + "grad_norm": 1.829655959511745, + "learning_rate": 1.9998262942289524e-05, + "loss": 1.9694, + "step": 330 + }, + { + "epoch": 0.17969598262757872, + "grad_norm": 1.9990017706312653, + "learning_rate": 1.9998196776540545e-05, + "loss": 2.067, + "step": 331 + }, + { + "epoch": 0.18023887079261672, + "grad_norm": 2.2829395740777234, + "learning_rate": 1.9998129374198655e-05, + "loss": 1.6707, + "step": 332 + }, + { + "epoch": 0.18078175895765472, + "grad_norm": 2.174263401724357, + "learning_rate": 1.9998060735272186e-05, + "loss": 1.6897, + "step": 333 + }, + { + "epoch": 0.18132464712269272, + "grad_norm": 1.8438064757899841, + "learning_rate": 1.9997990859769633e-05, + "loss": 1.4446, + "step": 334 + }, + { + "epoch": 0.18186753528773072, + "grad_norm": 2.0988396037795507, + "learning_rate": 1.9997919747699638e-05, + "loss": 1.8837, + "step": 335 + }, + { + "epoch": 0.18241042345276873, + "grad_norm": 2.582616969632433, + "learning_rate": 1.999784739907099e-05, + "loss": 2.459, + "step": 336 + }, + { + "epoch": 0.18295331161780673, + "grad_norm": 2.5198677585207707, + "learning_rate": 1.9997773813892644e-05, + "loss": 1.5117, + "step": 337 + }, + { + "epoch": 0.18349619978284473, + "grad_norm": 2.1361317951692635, + "learning_rate": 1.9997698992173697e-05, + "loss": 1.5892, + "step": 338 + }, + { + "epoch": 0.18403908794788273, + "grad_norm": 1.7284990063627486, + "learning_rate": 1.9997622933923406e-05, + "loss": 1.5151, + "step": 339 + }, + { + "epoch": 0.18458197611292074, + "grad_norm": 2.076752898704788, + "learning_rate": 1.9997545639151176e-05, + "loss": 2.3424, + "step": 340 + }, + { + "epoch": 0.18512486427795874, + "grad_norm": 2.479369276757794, + "learning_rate": 1.999746710786657e-05, + "loss": 1.5923, + "step": 341 + }, + { + "epoch": 0.18566775244299674, + "grad_norm": 1.9193986648366674, + "learning_rate": 1.9997387340079294e-05, + "loss": 1.6222, + "step": 342 + }, + { + "epoch": 0.18621064060803474, + "grad_norm": 1.8622986153150223, + "learning_rate": 1.999730633579922e-05, + "loss": 1.8402, + "step": 343 + }, + { + "epoch": 0.18675352877307275, + "grad_norm": 2.4012194427686238, + "learning_rate": 1.999722409503637e-05, + "loss": 1.9406, + "step": 344 + }, + { + "epoch": 0.18729641693811075, + "grad_norm": 2.3641518463127538, + "learning_rate": 1.9997140617800907e-05, + "loss": 2.0108, + "step": 345 + }, + { + "epoch": 0.18783930510314875, + "grad_norm": 2.406749148558192, + "learning_rate": 1.9997055904103156e-05, + "loss": 1.1323, + "step": 346 + }, + { + "epoch": 0.18838219326818675, + "grad_norm": 2.002785642517938, + "learning_rate": 1.99969699539536e-05, + "loss": 1.8745, + "step": 347 + }, + { + "epoch": 0.18892508143322476, + "grad_norm": 2.0002882431712816, + "learning_rate": 1.9996882767362874e-05, + "loss": 1.6311, + "step": 348 + }, + { + "epoch": 0.18946796959826276, + "grad_norm": 2.14418043520414, + "learning_rate": 1.9996794344341744e-05, + "loss": 1.7789, + "step": 349 + }, + { + "epoch": 0.19001085776330076, + "grad_norm": 2.0976983604919144, + "learning_rate": 1.9996704684901163e-05, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.19055374592833876, + "grad_norm": 2.017704706993719, + "learning_rate": 1.9996613789052214e-05, + "loss": 1.7974, + "step": 351 + }, + { + "epoch": 0.19109663409337677, + "grad_norm": 1.9907769296520095, + "learning_rate": 1.999652165680614e-05, + "loss": 1.9079, + "step": 352 + }, + { + "epoch": 0.19163952225841477, + "grad_norm": 2.3031451756435746, + "learning_rate": 1.999642828817433e-05, + "loss": 1.7022, + "step": 353 + }, + { + "epoch": 0.19218241042345277, + "grad_norm": 1.6722822353766713, + "learning_rate": 1.9996333683168342e-05, + "loss": 1.6655, + "step": 354 + }, + { + "epoch": 0.19272529858849077, + "grad_norm": 2.2974087830498027, + "learning_rate": 1.9996237841799874e-05, + "loss": 1.761, + "step": 355 + }, + { + "epoch": 0.19326818675352878, + "grad_norm": 1.9181254747298617, + "learning_rate": 1.9996140764080777e-05, + "loss": 1.8259, + "step": 356 + }, + { + "epoch": 0.19381107491856678, + "grad_norm": 2.07933834321742, + "learning_rate": 1.9996042450023053e-05, + "loss": 1.3783, + "step": 357 + }, + { + "epoch": 0.19435396308360478, + "grad_norm": 2.8332787295887436, + "learning_rate": 1.9995942899638875e-05, + "loss": 1.9796, + "step": 358 + }, + { + "epoch": 0.19489685124864278, + "grad_norm": 1.7924418183793758, + "learning_rate": 1.9995842112940545e-05, + "loss": 1.5063, + "step": 359 + }, + { + "epoch": 0.19543973941368079, + "grad_norm": 2.78627522830663, + "learning_rate": 1.9995740089940532e-05, + "loss": 1.891, + "step": 360 + }, + { + "epoch": 0.1959826275787188, + "grad_norm": 2.067311381839208, + "learning_rate": 1.9995636830651453e-05, + "loss": 1.9235, + "step": 361 + }, + { + "epoch": 0.1965255157437568, + "grad_norm": 2.539112412358468, + "learning_rate": 1.9995532335086078e-05, + "loss": 2.3062, + "step": 362 + }, + { + "epoch": 0.1970684039087948, + "grad_norm": 1.9040915545393904, + "learning_rate": 1.999542660325734e-05, + "loss": 1.9263, + "step": 363 + }, + { + "epoch": 0.1976112920738328, + "grad_norm": 1.7664296640025938, + "learning_rate": 1.9995319635178305e-05, + "loss": 1.5562, + "step": 364 + }, + { + "epoch": 0.1981541802388708, + "grad_norm": 1.687746217581272, + "learning_rate": 1.9995211430862206e-05, + "loss": 1.738, + "step": 365 + }, + { + "epoch": 0.1986970684039088, + "grad_norm": 1.9292930460598687, + "learning_rate": 1.9995101990322428e-05, + "loss": 1.5446, + "step": 366 + }, + { + "epoch": 0.1992399565689468, + "grad_norm": 1.9890390069401638, + "learning_rate": 1.9994991313572508e-05, + "loss": 1.9086, + "step": 367 + }, + { + "epoch": 0.1997828447339848, + "grad_norm": 1.8632430045841502, + "learning_rate": 1.999487940062613e-05, + "loss": 1.9747, + "step": 368 + }, + { + "epoch": 0.2003257328990228, + "grad_norm": 2.0677818303742237, + "learning_rate": 1.999476625149714e-05, + "loss": 1.4377, + "step": 369 + }, + { + "epoch": 0.2008686210640608, + "grad_norm": 1.7412205305358945, + "learning_rate": 1.9994651866199527e-05, + "loss": 1.9352, + "step": 370 + }, + { + "epoch": 0.2014115092290988, + "grad_norm": 1.8562175126383418, + "learning_rate": 1.9994536244747448e-05, + "loss": 1.5469, + "step": 371 + }, + { + "epoch": 0.20195439739413681, + "grad_norm": 1.8688345857586197, + "learning_rate": 1.9994419387155194e-05, + "loss": 1.8327, + "step": 372 + }, + { + "epoch": 0.20249728555917482, + "grad_norm": 2.1538407513899647, + "learning_rate": 1.9994301293437223e-05, + "loss": 1.9621, + "step": 373 + }, + { + "epoch": 0.20304017372421282, + "grad_norm": 2.4849664096044752, + "learning_rate": 1.999418196360814e-05, + "loss": 1.8179, + "step": 374 + }, + { + "epoch": 0.20358306188925082, + "grad_norm": 2.098949652063271, + "learning_rate": 1.99940613976827e-05, + "loss": 1.8719, + "step": 375 + }, + { + "epoch": 0.20412595005428882, + "grad_norm": 1.8802547433747465, + "learning_rate": 1.999393959567582e-05, + "loss": 1.5697, + "step": 376 + }, + { + "epoch": 0.20466883821932683, + "grad_norm": 2.090861017905227, + "learning_rate": 1.9993816557602567e-05, + "loss": 1.8249, + "step": 377 + }, + { + "epoch": 0.20521172638436483, + "grad_norm": 1.775481861127165, + "learning_rate": 1.999369228347815e-05, + "loss": 1.3095, + "step": 378 + }, + { + "epoch": 0.20575461454940283, + "grad_norm": 2.260330600420479, + "learning_rate": 1.999356677331794e-05, + "loss": 1.9153, + "step": 379 + }, + { + "epoch": 0.20629750271444083, + "grad_norm": 2.0769016143551484, + "learning_rate": 1.999344002713747e-05, + "loss": 1.958, + "step": 380 + }, + { + "epoch": 0.20684039087947884, + "grad_norm": 2.063584775140761, + "learning_rate": 1.9993312044952408e-05, + "loss": 1.7887, + "step": 381 + }, + { + "epoch": 0.20738327904451684, + "grad_norm": 1.9638736053910157, + "learning_rate": 1.9993182826778588e-05, + "loss": 1.3178, + "step": 382 + }, + { + "epoch": 0.20792616720955484, + "grad_norm": 1.7752329715548703, + "learning_rate": 1.9993052372631988e-05, + "loss": 1.4473, + "step": 383 + }, + { + "epoch": 0.20846905537459284, + "grad_norm": 2.1002641405528952, + "learning_rate": 1.999292068252874e-05, + "loss": 1.9698, + "step": 384 + }, + { + "epoch": 0.20901194353963085, + "grad_norm": 2.757803188261632, + "learning_rate": 1.999278775648514e-05, + "loss": 1.1517, + "step": 385 + }, + { + "epoch": 0.20955483170466885, + "grad_norm": 3.0341284449704378, + "learning_rate": 1.9992653594517624e-05, + "loss": 1.987, + "step": 386 + }, + { + "epoch": 0.21009771986970685, + "grad_norm": 2.445830333821562, + "learning_rate": 1.9992518196642786e-05, + "loss": 1.7656, + "step": 387 + }, + { + "epoch": 0.21064060803474485, + "grad_norm": 2.737783962568244, + "learning_rate": 1.9992381562877368e-05, + "loss": 2.3012, + "step": 388 + }, + { + "epoch": 0.21118349619978286, + "grad_norm": 2.0108432149096793, + "learning_rate": 1.9992243693238275e-05, + "loss": 2.1096, + "step": 389 + }, + { + "epoch": 0.21172638436482086, + "grad_norm": 2.1448465445226397, + "learning_rate": 1.9992104587742558e-05, + "loss": 1.6912, + "step": 390 + }, + { + "epoch": 0.21226927252985886, + "grad_norm": 2.842739011518123, + "learning_rate": 1.999196424640742e-05, + "loss": 1.9987, + "step": 391 + }, + { + "epoch": 0.21281216069489686, + "grad_norm": 2.026900017823922, + "learning_rate": 1.9991822669250216e-05, + "loss": 1.6251, + "step": 392 + }, + { + "epoch": 0.21335504885993486, + "grad_norm": 2.139818571931489, + "learning_rate": 1.9991679856288462e-05, + "loss": 1.4181, + "step": 393 + }, + { + "epoch": 0.21389793702497287, + "grad_norm": 2.061327407258535, + "learning_rate": 1.999153580753982e-05, + "loss": 2.0976, + "step": 394 + }, + { + "epoch": 0.21444082519001087, + "grad_norm": 1.9510646261617408, + "learning_rate": 1.9991390523022105e-05, + "loss": 1.3091, + "step": 395 + }, + { + "epoch": 0.21498371335504887, + "grad_norm": 2.0892427860245664, + "learning_rate": 1.9991244002753287e-05, + "loss": 1.3693, + "step": 396 + }, + { + "epoch": 0.21552660152008687, + "grad_norm": 1.9528708238712815, + "learning_rate": 1.9991096246751483e-05, + "loss": 1.2807, + "step": 397 + }, + { + "epoch": 0.21606948968512488, + "grad_norm": 2.0084632909809983, + "learning_rate": 1.9990947255034977e-05, + "loss": 1.7429, + "step": 398 + }, + { + "epoch": 0.21661237785016288, + "grad_norm": 2.19414527352725, + "learning_rate": 1.999079702762219e-05, + "loss": 1.5962, + "step": 399 + }, + { + "epoch": 0.21715526601520088, + "grad_norm": 2.184484203373308, + "learning_rate": 1.9990645564531702e-05, + "loss": 1.6887, + "step": 400 + }, + { + "epoch": 0.21769815418023888, + "grad_norm": 2.214473402714986, + "learning_rate": 1.9990492865782248e-05, + "loss": 1.5226, + "step": 401 + }, + { + "epoch": 0.2182410423452769, + "grad_norm": 2.1467748369867414, + "learning_rate": 1.9990338931392714e-05, + "loss": 1.5634, + "step": 402 + }, + { + "epoch": 0.21878393051031486, + "grad_norm": 2.1969353910959684, + "learning_rate": 1.999018376138214e-05, + "loss": 1.8652, + "step": 403 + }, + { + "epoch": 0.21932681867535286, + "grad_norm": 2.245031544698939, + "learning_rate": 1.9990027355769715e-05, + "loss": 2.1811, + "step": 404 + }, + { + "epoch": 0.21986970684039087, + "grad_norm": 2.381275606702572, + "learning_rate": 1.9989869714574784e-05, + "loss": 1.6356, + "step": 405 + }, + { + "epoch": 0.22041259500542887, + "grad_norm": 2.8847531540936053, + "learning_rate": 1.9989710837816846e-05, + "loss": 1.9953, + "step": 406 + }, + { + "epoch": 0.22095548317046687, + "grad_norm": 2.383074182943421, + "learning_rate": 1.9989550725515553e-05, + "loss": 1.5589, + "step": 407 + }, + { + "epoch": 0.22149837133550487, + "grad_norm": 2.3867744628533987, + "learning_rate": 1.99893893776907e-05, + "loss": 1.5351, + "step": 408 + }, + { + "epoch": 0.22204125950054288, + "grad_norm": 2.4139767232435463, + "learning_rate": 1.998922679436225e-05, + "loss": 1.5949, + "step": 409 + }, + { + "epoch": 0.22258414766558088, + "grad_norm": 3.034840719531937, + "learning_rate": 1.9989062975550313e-05, + "loss": 1.7994, + "step": 410 + }, + { + "epoch": 0.22312703583061888, + "grad_norm": 2.6245673280102544, + "learning_rate": 1.9988897921275144e-05, + "loss": 2.2593, + "step": 411 + }, + { + "epoch": 0.22366992399565688, + "grad_norm": 2.403934048820897, + "learning_rate": 1.998873163155716e-05, + "loss": 1.3179, + "step": 412 + }, + { + "epoch": 0.22421281216069489, + "grad_norm": 2.0174357446052524, + "learning_rate": 1.998856410641693e-05, + "loss": 1.8698, + "step": 413 + }, + { + "epoch": 0.2247557003257329, + "grad_norm": 2.25869172419699, + "learning_rate": 1.998839534587517e-05, + "loss": 1.596, + "step": 414 + }, + { + "epoch": 0.2252985884907709, + "grad_norm": 3.384596068816443, + "learning_rate": 1.9988225349952758e-05, + "loss": 1.5232, + "step": 415 + }, + { + "epoch": 0.2258414766558089, + "grad_norm": 2.2109577343653246, + "learning_rate": 1.9988054118670712e-05, + "loss": 1.2403, + "step": 416 + }, + { + "epoch": 0.2263843648208469, + "grad_norm": 2.1825505717325053, + "learning_rate": 1.9987881652050215e-05, + "loss": 1.7205, + "step": 417 + }, + { + "epoch": 0.2269272529858849, + "grad_norm": 2.7992033856118583, + "learning_rate": 1.99877079501126e-05, + "loss": 1.7486, + "step": 418 + }, + { + "epoch": 0.2274701411509229, + "grad_norm": 2.6351781282432793, + "learning_rate": 1.9987533012879344e-05, + "loss": 1.6693, + "step": 419 + }, + { + "epoch": 0.2280130293159609, + "grad_norm": 2.2385207734696384, + "learning_rate": 1.9987356840372088e-05, + "loss": 1.1296, + "step": 420 + }, + { + "epoch": 0.2285559174809989, + "grad_norm": 2.4921881643709107, + "learning_rate": 1.998717943261262e-05, + "loss": 1.1379, + "step": 421 + }, + { + "epoch": 0.2290988056460369, + "grad_norm": 2.842685080646979, + "learning_rate": 1.9987000789622884e-05, + "loss": 1.7963, + "step": 422 + }, + { + "epoch": 0.2296416938110749, + "grad_norm": 2.4776027353414145, + "learning_rate": 1.9986820911424972e-05, + "loss": 1.6939, + "step": 423 + }, + { + "epoch": 0.2301845819761129, + "grad_norm": 2.688785422525701, + "learning_rate": 1.9986639798041134e-05, + "loss": 1.2282, + "step": 424 + }, + { + "epoch": 0.23072747014115091, + "grad_norm": 3.4758865457335495, + "learning_rate": 1.998645744949377e-05, + "loss": 1.7664, + "step": 425 + }, + { + "epoch": 0.23127035830618892, + "grad_norm": 2.5586740557561605, + "learning_rate": 1.9986273865805432e-05, + "loss": 1.5637, + "step": 426 + }, + { + "epoch": 0.23181324647122692, + "grad_norm": 2.645306116903878, + "learning_rate": 1.9986089046998827e-05, + "loss": 1.5331, + "step": 427 + }, + { + "epoch": 0.23235613463626492, + "grad_norm": 2.748145883999443, + "learning_rate": 1.998590299309681e-05, + "loss": 1.7316, + "step": 428 + }, + { + "epoch": 0.23289902280130292, + "grad_norm": 2.4357923833746438, + "learning_rate": 1.99857157041224e-05, + "loss": 1.1479, + "step": 429 + }, + { + "epoch": 0.23344191096634093, + "grad_norm": 2.7322633978331377, + "learning_rate": 1.9985527180098755e-05, + "loss": 1.6089, + "step": 430 + }, + { + "epoch": 0.23398479913137893, + "grad_norm": 2.625531215380387, + "learning_rate": 1.9985337421049193e-05, + "loss": 1.8973, + "step": 431 + }, + { + "epoch": 0.23452768729641693, + "grad_norm": 2.1586487666504754, + "learning_rate": 1.9985146426997185e-05, + "loss": 1.664, + "step": 432 + }, + { + "epoch": 0.23507057546145493, + "grad_norm": 2.748704307086674, + "learning_rate": 1.9984954197966355e-05, + "loss": 1.677, + "step": 433 + }, + { + "epoch": 0.23561346362649294, + "grad_norm": 2.345532526959197, + "learning_rate": 1.9984760733980476e-05, + "loss": 1.5133, + "step": 434 + }, + { + "epoch": 0.23615635179153094, + "grad_norm": 2.1655585346308848, + "learning_rate": 1.9984566035063473e-05, + "loss": 1.4206, + "step": 435 + }, + { + "epoch": 0.23669923995656894, + "grad_norm": 2.697584333049271, + "learning_rate": 1.9984370101239434e-05, + "loss": 1.8131, + "step": 436 + }, + { + "epoch": 0.23724212812160694, + "grad_norm": 2.235069158786981, + "learning_rate": 1.9984172932532583e-05, + "loss": 1.7839, + "step": 437 + }, + { + "epoch": 0.23778501628664495, + "grad_norm": 2.548562657856099, + "learning_rate": 1.998397452896731e-05, + "loss": 1.4876, + "step": 438 + }, + { + "epoch": 0.23832790445168295, + "grad_norm": 2.1698874138883673, + "learning_rate": 1.9983774890568163e-05, + "loss": 1.2916, + "step": 439 + }, + { + "epoch": 0.23887079261672095, + "grad_norm": 2.3237954527681084, + "learning_rate": 1.998357401735982e-05, + "loss": 1.398, + "step": 440 + }, + { + "epoch": 0.23941368078175895, + "grad_norm": 2.448364938314344, + "learning_rate": 1.9983371909367135e-05, + "loss": 1.2663, + "step": 441 + }, + { + "epoch": 0.23995656894679696, + "grad_norm": 1.935899726785714, + "learning_rate": 1.99831685666151e-05, + "loss": 1.2448, + "step": 442 + }, + { + "epoch": 0.24049945711183496, + "grad_norm": 2.14357016947003, + "learning_rate": 1.9982963989128864e-05, + "loss": 1.4786, + "step": 443 + }, + { + "epoch": 0.24104234527687296, + "grad_norm": 2.5830135196717148, + "learning_rate": 1.998275817693373e-05, + "loss": 1.8324, + "step": 444 + }, + { + "epoch": 0.24158523344191096, + "grad_norm": 3.159289299734819, + "learning_rate": 1.9982551130055157e-05, + "loss": 1.9091, + "step": 445 + }, + { + "epoch": 0.24212812160694897, + "grad_norm": 3.2712239674501755, + "learning_rate": 1.9982342848518753e-05, + "loss": 1.7869, + "step": 446 + }, + { + "epoch": 0.24267100977198697, + "grad_norm": 3.35834841128443, + "learning_rate": 1.998213333235027e-05, + "loss": 1.2772, + "step": 447 + }, + { + "epoch": 0.24321389793702497, + "grad_norm": 2.9379666136057354, + "learning_rate": 1.998192258157563e-05, + "loss": 1.3497, + "step": 448 + }, + { + "epoch": 0.24375678610206297, + "grad_norm": 2.334119075845609, + "learning_rate": 1.9981710596220897e-05, + "loss": 1.5541, + "step": 449 + }, + { + "epoch": 0.24429967426710097, + "grad_norm": 2.781469856763384, + "learning_rate": 1.998149737631229e-05, + "loss": 1.9413, + "step": 450 + }, + { + "epoch": 0.24484256243213898, + "grad_norm": 3.0196781534130452, + "learning_rate": 1.9981282921876177e-05, + "loss": 1.3238, + "step": 451 + }, + { + "epoch": 0.24538545059717698, + "grad_norm": 3.547169321727429, + "learning_rate": 1.9981067232939086e-05, + "loss": 1.9952, + "step": 452 + }, + { + "epoch": 0.24592833876221498, + "grad_norm": 3.2411499530913535, + "learning_rate": 1.9980850309527693e-05, + "loss": 1.8244, + "step": 453 + }, + { + "epoch": 0.24647122692725298, + "grad_norm": 2.670313260104859, + "learning_rate": 1.9980632151668822e-05, + "loss": 1.607, + "step": 454 + }, + { + "epoch": 0.247014115092291, + "grad_norm": 2.8822100638306143, + "learning_rate": 1.9980412759389468e-05, + "loss": 1.8868, + "step": 455 + }, + { + "epoch": 0.247557003257329, + "grad_norm": 2.4776152417583317, + "learning_rate": 1.9980192132716748e-05, + "loss": 1.778, + "step": 456 + }, + { + "epoch": 0.248099891422367, + "grad_norm": 2.2001923672712076, + "learning_rate": 1.9979970271677967e-05, + "loss": 1.3544, + "step": 457 + }, + { + "epoch": 0.248642779587405, + "grad_norm": 2.7694932683911837, + "learning_rate": 1.9979747176300553e-05, + "loss": 1.6521, + "step": 458 + }, + { + "epoch": 0.249185667752443, + "grad_norm": 2.8464573075472845, + "learning_rate": 1.99795228466121e-05, + "loss": 1.8803, + "step": 459 + }, + { + "epoch": 0.249728555917481, + "grad_norm": 2.577989994947286, + "learning_rate": 1.9979297282640365e-05, + "loss": 1.8838, + "step": 460 + }, + { + "epoch": 0.250271444082519, + "grad_norm": 2.9859834231033164, + "learning_rate": 1.997907048441323e-05, + "loss": 1.5122, + "step": 461 + }, + { + "epoch": 0.250814332247557, + "grad_norm": 3.0857580735568098, + "learning_rate": 1.9978842451958757e-05, + "loss": 1.7789, + "step": 462 + }, + { + "epoch": 0.251357220412595, + "grad_norm": 2.523783848348806, + "learning_rate": 1.9978613185305145e-05, + "loss": 1.5815, + "step": 463 + }, + { + "epoch": 0.251900108577633, + "grad_norm": 2.699094362013039, + "learning_rate": 1.9978382684480747e-05, + "loss": 1.7448, + "step": 464 + }, + { + "epoch": 0.252442996742671, + "grad_norm": 2.736268550628698, + "learning_rate": 1.997815094951408e-05, + "loss": 1.7052, + "step": 465 + }, + { + "epoch": 0.252985884907709, + "grad_norm": 3.1150828428090014, + "learning_rate": 1.99779179804338e-05, + "loss": 1.7743, + "step": 466 + }, + { + "epoch": 0.253528773072747, + "grad_norm": 3.3513164619888482, + "learning_rate": 1.997768377726872e-05, + "loss": 2.3905, + "step": 467 + }, + { + "epoch": 0.254071661237785, + "grad_norm": 2.400886982379507, + "learning_rate": 1.9977448340047808e-05, + "loss": 1.6096, + "step": 468 + }, + { + "epoch": 0.254614549402823, + "grad_norm": 2.633533044966171, + "learning_rate": 1.9977211668800186e-05, + "loss": 1.7796, + "step": 469 + }, + { + "epoch": 0.255157437567861, + "grad_norm": 3.9297139007235042, + "learning_rate": 1.997697376355512e-05, + "loss": 1.4602, + "step": 470 + }, + { + "epoch": 0.255700325732899, + "grad_norm": 5.588260619709643, + "learning_rate": 1.9976734624342044e-05, + "loss": 2.0389, + "step": 471 + }, + { + "epoch": 0.256243213897937, + "grad_norm": 2.600723171476426, + "learning_rate": 1.9976494251190522e-05, + "loss": 1.3676, + "step": 472 + }, + { + "epoch": 0.25678610206297503, + "grad_norm": 2.61945002649116, + "learning_rate": 1.9976252644130297e-05, + "loss": 1.7902, + "step": 473 + }, + { + "epoch": 0.25732899022801303, + "grad_norm": 3.7513561790803838, + "learning_rate": 1.997600980319124e-05, + "loss": 1.5997, + "step": 474 + }, + { + "epoch": 0.25787187839305103, + "grad_norm": 2.8832359552778737, + "learning_rate": 1.9975765728403395e-05, + "loss": 1.9636, + "step": 475 + }, + { + "epoch": 0.25841476655808904, + "grad_norm": 3.0975018752600243, + "learning_rate": 1.9975520419796942e-05, + "loss": 1.1165, + "step": 476 + }, + { + "epoch": 0.25895765472312704, + "grad_norm": 2.121708502818221, + "learning_rate": 1.9975273877402227e-05, + "loss": 1.5108, + "step": 477 + }, + { + "epoch": 0.25950054288816504, + "grad_norm": 2.4073592870530116, + "learning_rate": 1.997502610124974e-05, + "loss": 1.7828, + "step": 478 + }, + { + "epoch": 0.26004343105320304, + "grad_norm": 2.6858679986632974, + "learning_rate": 1.997477709137013e-05, + "loss": 1.8483, + "step": 479 + }, + { + "epoch": 0.26058631921824105, + "grad_norm": 3.283752190131325, + "learning_rate": 1.997452684779419e-05, + "loss": 1.6105, + "step": 480 + }, + { + "epoch": 0.26112920738327905, + "grad_norm": 2.500181185675909, + "learning_rate": 1.997427537055287e-05, + "loss": 1.5475, + "step": 481 + }, + { + "epoch": 0.26167209554831705, + "grad_norm": 2.3992324550953885, + "learning_rate": 1.9974022659677278e-05, + "loss": 1.6062, + "step": 482 + }, + { + "epoch": 0.26221498371335505, + "grad_norm": 2.7692293822867837, + "learning_rate": 1.9973768715198667e-05, + "loss": 1.4995, + "step": 483 + }, + { + "epoch": 0.26275787187839306, + "grad_norm": 2.971423364277874, + "learning_rate": 1.9973513537148447e-05, + "loss": 1.7904, + "step": 484 + }, + { + "epoch": 0.26330076004343106, + "grad_norm": 2.2769736321644105, + "learning_rate": 1.9973257125558177e-05, + "loss": 1.3121, + "step": 485 + }, + { + "epoch": 0.26384364820846906, + "grad_norm": 2.5949046051899254, + "learning_rate": 1.997299948045957e-05, + "loss": 1.4555, + "step": 486 + }, + { + "epoch": 0.26438653637350706, + "grad_norm": 3.0210593683445204, + "learning_rate": 1.997274060188449e-05, + "loss": 1.8211, + "step": 487 + }, + { + "epoch": 0.26492942453854507, + "grad_norm": 3.076626224616319, + "learning_rate": 1.9972480489864962e-05, + "loss": 1.6366, + "step": 488 + }, + { + "epoch": 0.26547231270358307, + "grad_norm": 3.1369728365663536, + "learning_rate": 1.9972219144433148e-05, + "loss": 1.5027, + "step": 489 + }, + { + "epoch": 0.26601520086862107, + "grad_norm": 2.2870450349164635, + "learning_rate": 1.9971956565621383e-05, + "loss": 1.2784, + "step": 490 + }, + { + "epoch": 0.2665580890336591, + "grad_norm": 3.0130036065633776, + "learning_rate": 1.9971692753462134e-05, + "loss": 1.2083, + "step": 491 + }, + { + "epoch": 0.2671009771986971, + "grad_norm": 2.6063157323029733, + "learning_rate": 1.9971427707988034e-05, + "loss": 1.0083, + "step": 492 + }, + { + "epoch": 0.2676438653637351, + "grad_norm": 2.579371053895234, + "learning_rate": 1.997116142923186e-05, + "loss": 1.1937, + "step": 493 + }, + { + "epoch": 0.2681867535287731, + "grad_norm": 2.532537971800688, + "learning_rate": 1.9970893917226554e-05, + "loss": 1.4735, + "step": 494 + }, + { + "epoch": 0.2687296416938111, + "grad_norm": 2.3483150144294105, + "learning_rate": 1.997062517200519e-05, + "loss": 1.7269, + "step": 495 + }, + { + "epoch": 0.2692725298588491, + "grad_norm": 2.594809867192747, + "learning_rate": 1.997035519360102e-05, + "loss": 1.8283, + "step": 496 + }, + { + "epoch": 0.2698154180238871, + "grad_norm": 2.613813750609998, + "learning_rate": 1.9970083982047428e-05, + "loss": 1.2302, + "step": 497 + }, + { + "epoch": 0.2703583061889251, + "grad_norm": 2.648279162964909, + "learning_rate": 1.9969811537377956e-05, + "loss": 1.6225, + "step": 498 + }, + { + "epoch": 0.2709011943539631, + "grad_norm": 2.3790190706794325, + "learning_rate": 1.9969537859626308e-05, + "loss": 1.5172, + "step": 499 + }, + { + "epoch": 0.2714440825190011, + "grad_norm": 2.7054998578606364, + "learning_rate": 1.9969262948826326e-05, + "loss": 1.4525, + "step": 500 + }, + { + "epoch": 0.2719869706840391, + "grad_norm": 2.412151508264948, + "learning_rate": 1.9968986805012012e-05, + "loss": 1.3299, + "step": 501 + }, + { + "epoch": 0.2725298588490771, + "grad_norm": 2.4836460319285414, + "learning_rate": 1.9968709428217525e-05, + "loss": 1.5217, + "step": 502 + }, + { + "epoch": 0.2730727470141151, + "grad_norm": 3.3112944949184606, + "learning_rate": 1.9968430818477168e-05, + "loss": 2.0643, + "step": 503 + }, + { + "epoch": 0.2736156351791531, + "grad_norm": 3.266395629954733, + "learning_rate": 1.9968150975825397e-05, + "loss": 1.1719, + "step": 504 + }, + { + "epoch": 0.2741585233441911, + "grad_norm": 2.3487351103507073, + "learning_rate": 1.996786990029683e-05, + "loss": 1.6876, + "step": 505 + }, + { + "epoch": 0.2747014115092291, + "grad_norm": 3.0672442719402673, + "learning_rate": 1.9967587591926227e-05, + "loss": 1.3946, + "step": 506 + }, + { + "epoch": 0.2752442996742671, + "grad_norm": 2.84320621483769, + "learning_rate": 1.99673040507485e-05, + "loss": 1.3259, + "step": 507 + }, + { + "epoch": 0.2757871878393051, + "grad_norm": 2.5078449617010707, + "learning_rate": 1.9967019276798728e-05, + "loss": 0.8478, + "step": 508 + }, + { + "epoch": 0.2763300760043431, + "grad_norm": 3.137784699454796, + "learning_rate": 1.9966733270112126e-05, + "loss": 1.1688, + "step": 509 + }, + { + "epoch": 0.2768729641693811, + "grad_norm": 3.7277826491955017, + "learning_rate": 1.996644603072407e-05, + "loss": 1.1091, + "step": 510 + }, + { + "epoch": 0.2774158523344191, + "grad_norm": 2.5276781604415635, + "learning_rate": 1.996615755867008e-05, + "loss": 1.1299, + "step": 511 + }, + { + "epoch": 0.2779587404994571, + "grad_norm": 3.1012700661738744, + "learning_rate": 1.996586785398584e-05, + "loss": 1.7218, + "step": 512 + }, + { + "epoch": 0.2785016286644951, + "grad_norm": 2.285166712515903, + "learning_rate": 1.9965576916707182e-05, + "loss": 1.2868, + "step": 513 + }, + { + "epoch": 0.27904451682953313, + "grad_norm": 2.33097906349044, + "learning_rate": 1.9965284746870088e-05, + "loss": 0.9887, + "step": 514 + }, + { + "epoch": 0.27958740499457113, + "grad_norm": 2.6473787082237927, + "learning_rate": 1.9964991344510697e-05, + "loss": 1.8543, + "step": 515 + }, + { + "epoch": 0.28013029315960913, + "grad_norm": 2.4628160599533366, + "learning_rate": 1.996469670966529e-05, + "loss": 1.2263, + "step": 516 + }, + { + "epoch": 0.28067318132464714, + "grad_norm": 3.2897847068350905, + "learning_rate": 1.9964400842370314e-05, + "loss": 1.6338, + "step": 517 + }, + { + "epoch": 0.28121606948968514, + "grad_norm": 2.4439319341540324, + "learning_rate": 1.9964103742662363e-05, + "loss": 1.0836, + "step": 518 + }, + { + "epoch": 0.28175895765472314, + "grad_norm": 2.3221991020412003, + "learning_rate": 1.996380541057818e-05, + "loss": 1.2331, + "step": 519 + }, + { + "epoch": 0.28230184581976114, + "grad_norm": 2.9571040634251564, + "learning_rate": 1.9963505846154662e-05, + "loss": 1.3066, + "step": 520 + }, + { + "epoch": 0.28284473398479915, + "grad_norm": 3.7512706020225624, + "learning_rate": 1.996320504942886e-05, + "loss": 1.7482, + "step": 521 + }, + { + "epoch": 0.28338762214983715, + "grad_norm": 2.4620109793388267, + "learning_rate": 1.9962903020437983e-05, + "loss": 1.5334, + "step": 522 + }, + { + "epoch": 0.28393051031487515, + "grad_norm": 3.030374272795485, + "learning_rate": 1.9962599759219383e-05, + "loss": 1.8957, + "step": 523 + }, + { + "epoch": 0.28447339847991315, + "grad_norm": 2.452389821491403, + "learning_rate": 1.9962295265810563e-05, + "loss": 1.5438, + "step": 524 + }, + { + "epoch": 0.28501628664495116, + "grad_norm": 2.713028369466205, + "learning_rate": 1.996198954024919e-05, + "loss": 1.4272, + "step": 525 + }, + { + "epoch": 0.28555917480998916, + "grad_norm": 3.810321275175567, + "learning_rate": 1.996168258257307e-05, + "loss": 1.9028, + "step": 526 + }, + { + "epoch": 0.28610206297502716, + "grad_norm": 2.7774100977441236, + "learning_rate": 1.9961374392820173e-05, + "loss": 1.5644, + "step": 527 + }, + { + "epoch": 0.28664495114006516, + "grad_norm": 2.6798712089104186, + "learning_rate": 1.9961064971028616e-05, + "loss": 1.188, + "step": 528 + }, + { + "epoch": 0.28718783930510317, + "grad_norm": 2.889465990486677, + "learning_rate": 1.9960754317236666e-05, + "loss": 1.5393, + "step": 529 + }, + { + "epoch": 0.28773072747014117, + "grad_norm": 3.465251366831076, + "learning_rate": 1.996044243148275e-05, + "loss": 2.1899, + "step": 530 + }, + { + "epoch": 0.28827361563517917, + "grad_norm": 2.645941940974219, + "learning_rate": 1.9960129313805437e-05, + "loss": 1.3691, + "step": 531 + }, + { + "epoch": 0.2888165038002172, + "grad_norm": 2.3914199977194293, + "learning_rate": 1.9959814964243455e-05, + "loss": 1.5219, + "step": 532 + }, + { + "epoch": 0.2893593919652552, + "grad_norm": 2.3023810529281343, + "learning_rate": 1.995949938283569e-05, + "loss": 1.5147, + "step": 533 + }, + { + "epoch": 0.2899022801302932, + "grad_norm": 2.7362205671791155, + "learning_rate": 1.9959182569621164e-05, + "loss": 1.7571, + "step": 534 + }, + { + "epoch": 0.2904451682953312, + "grad_norm": 3.971162331076012, + "learning_rate": 1.9958864524639066e-05, + "loss": 1.3425, + "step": 535 + }, + { + "epoch": 0.2909880564603692, + "grad_norm": 3.4144928239616514, + "learning_rate": 1.9958545247928727e-05, + "loss": 1.6962, + "step": 536 + }, + { + "epoch": 0.2915309446254072, + "grad_norm": 3.5063126675319043, + "learning_rate": 1.9958224739529647e-05, + "loss": 1.6406, + "step": 537 + }, + { + "epoch": 0.2920738327904452, + "grad_norm": 2.9013783116047547, + "learning_rate": 1.995790299948146e-05, + "loss": 1.6376, + "step": 538 + }, + { + "epoch": 0.2926167209554832, + "grad_norm": 2.4827691033904693, + "learning_rate": 1.9957580027823957e-05, + "loss": 1.8672, + "step": 539 + }, + { + "epoch": 0.2931596091205212, + "grad_norm": 3.1245563354940242, + "learning_rate": 1.9957255824597087e-05, + "loss": 1.885, + "step": 540 + }, + { + "epoch": 0.2937024972855592, + "grad_norm": 2.7941860466759896, + "learning_rate": 1.9956930389840945e-05, + "loss": 1.0903, + "step": 541 + }, + { + "epoch": 0.2942453854505972, + "grad_norm": 3.1348104146717772, + "learning_rate": 1.9956603723595784e-05, + "loss": 2.1446, + "step": 542 + }, + { + "epoch": 0.2947882736156352, + "grad_norm": 2.4248222663066747, + "learning_rate": 1.995627582590201e-05, + "loss": 1.3503, + "step": 543 + }, + { + "epoch": 0.2953311617806732, + "grad_norm": 3.1538245864476337, + "learning_rate": 1.995594669680017e-05, + "loss": 1.3275, + "step": 544 + }, + { + "epoch": 0.2958740499457112, + "grad_norm": 3.436905240480997, + "learning_rate": 1.9955616336330976e-05, + "loss": 1.7249, + "step": 545 + }, + { + "epoch": 0.2964169381107492, + "grad_norm": 3.9145352507266393, + "learning_rate": 1.9955284744535287e-05, + "loss": 2.0089, + "step": 546 + }, + { + "epoch": 0.2969598262757872, + "grad_norm": 2.8600509325831185, + "learning_rate": 1.9954951921454113e-05, + "loss": 1.4527, + "step": 547 + }, + { + "epoch": 0.2975027144408252, + "grad_norm": 2.865760632888347, + "learning_rate": 1.995461786712862e-05, + "loss": 1.4553, + "step": 548 + }, + { + "epoch": 0.2980456026058632, + "grad_norm": 3.3559882259900706, + "learning_rate": 1.9954282581600127e-05, + "loss": 2.0456, + "step": 549 + }, + { + "epoch": 0.2985884907709012, + "grad_norm": 3.3424656149019008, + "learning_rate": 1.9953946064910098e-05, + "loss": 1.8253, + "step": 550 + }, + { + "epoch": 0.2991313789359392, + "grad_norm": 2.569119318410074, + "learning_rate": 1.9953608317100153e-05, + "loss": 1.3623, + "step": 551 + }, + { + "epoch": 0.2996742671009772, + "grad_norm": 2.8918953822102424, + "learning_rate": 1.995326933821207e-05, + "loss": 1.7521, + "step": 552 + }, + { + "epoch": 0.3002171552660152, + "grad_norm": 2.7627900710192246, + "learning_rate": 1.995292912828777e-05, + "loss": 1.775, + "step": 553 + }, + { + "epoch": 0.3007600434310532, + "grad_norm": 2.8053609300694804, + "learning_rate": 1.9952587687369334e-05, + "loss": 1.7536, + "step": 554 + }, + { + "epoch": 0.30130293159609123, + "grad_norm": 2.9775715301146803, + "learning_rate": 1.995224501549899e-05, + "loss": 1.6715, + "step": 555 + }, + { + "epoch": 0.30184581976112923, + "grad_norm": 3.146995410263436, + "learning_rate": 1.9951901112719123e-05, + "loss": 1.1032, + "step": 556 + }, + { + "epoch": 0.30238870792616723, + "grad_norm": 3.219197817112143, + "learning_rate": 1.9951555979072266e-05, + "loss": 1.6326, + "step": 557 + }, + { + "epoch": 0.30293159609120524, + "grad_norm": 2.676508071644292, + "learning_rate": 1.99512096146011e-05, + "loss": 1.4836, + "step": 558 + }, + { + "epoch": 0.30347448425624324, + "grad_norm": 3.8806900857620374, + "learning_rate": 1.9950862019348474e-05, + "loss": 1.7794, + "step": 559 + }, + { + "epoch": 0.30401737242128124, + "grad_norm": 2.547409032322543, + "learning_rate": 1.995051319335737e-05, + "loss": 1.3263, + "step": 560 + }, + { + "epoch": 0.30456026058631924, + "grad_norm": 3.025189850713409, + "learning_rate": 1.995016313667094e-05, + "loss": 1.2409, + "step": 561 + }, + { + "epoch": 0.30510314875135724, + "grad_norm": 3.3644665856402614, + "learning_rate": 1.9949811849332476e-05, + "loss": 1.2988, + "step": 562 + }, + { + "epoch": 0.30564603691639525, + "grad_norm": 3.4126932100522755, + "learning_rate": 1.9949459331385422e-05, + "loss": 1.6126, + "step": 563 + }, + { + "epoch": 0.30618892508143325, + "grad_norm": 3.3751971677759416, + "learning_rate": 1.994910558287338e-05, + "loss": 1.6243, + "step": 564 + }, + { + "epoch": 0.30673181324647125, + "grad_norm": 3.0948952637866105, + "learning_rate": 1.9948750603840102e-05, + "loss": 1.6553, + "step": 565 + }, + { + "epoch": 0.30727470141150925, + "grad_norm": 3.1116929004314224, + "learning_rate": 1.9948394394329494e-05, + "loss": 1.0466, + "step": 566 + }, + { + "epoch": 0.30781758957654726, + "grad_norm": 2.649244034262683, + "learning_rate": 1.9948036954385613e-05, + "loss": 1.2914, + "step": 567 + }, + { + "epoch": 0.30836047774158526, + "grad_norm": 3.9338023936000965, + "learning_rate": 1.9947678284052667e-05, + "loss": 1.7532, + "step": 568 + }, + { + "epoch": 0.30890336590662326, + "grad_norm": 2.887133065759567, + "learning_rate": 1.9947318383375017e-05, + "loss": 1.8001, + "step": 569 + }, + { + "epoch": 0.30944625407166126, + "grad_norm": 3.6996119456889915, + "learning_rate": 1.9946957252397173e-05, + "loss": 2.4852, + "step": 570 + }, + { + "epoch": 0.30998914223669927, + "grad_norm": 2.829855772438557, + "learning_rate": 1.9946594891163808e-05, + "loss": 1.6048, + "step": 571 + }, + { + "epoch": 0.31053203040173727, + "grad_norm": 4.288772356209683, + "learning_rate": 1.9946231299719732e-05, + "loss": 1.4841, + "step": 572 + }, + { + "epoch": 0.31107491856677527, + "grad_norm": 2.37958986026152, + "learning_rate": 1.9945866478109914e-05, + "loss": 1.1797, + "step": 573 + }, + { + "epoch": 0.3116178067318133, + "grad_norm": 2.9977970906442932, + "learning_rate": 1.9945500426379483e-05, + "loss": 1.224, + "step": 574 + }, + { + "epoch": 0.3121606948968513, + "grad_norm": 3.406596416686285, + "learning_rate": 1.9945133144573705e-05, + "loss": 1.4793, + "step": 575 + }, + { + "epoch": 0.3127035830618892, + "grad_norm": 3.3772239188722244, + "learning_rate": 1.994476463273801e-05, + "loss": 1.3696, + "step": 576 + }, + { + "epoch": 0.3132464712269272, + "grad_norm": 2.729299768057245, + "learning_rate": 1.9944394890917977e-05, + "loss": 1.448, + "step": 577 + }, + { + "epoch": 0.31378935939196523, + "grad_norm": 2.6526406591248297, + "learning_rate": 1.9944023919159335e-05, + "loss": 1.6905, + "step": 578 + }, + { + "epoch": 0.31433224755700323, + "grad_norm": 2.9512501195242944, + "learning_rate": 1.9943651717507965e-05, + "loss": 1.6277, + "step": 579 + }, + { + "epoch": 0.31487513572204123, + "grad_norm": 2.8292191327941723, + "learning_rate": 1.9943278286009903e-05, + "loss": 1.0532, + "step": 580 + }, + { + "epoch": 0.31541802388707924, + "grad_norm": 3.2105870294745436, + "learning_rate": 1.9942903624711335e-05, + "loss": 1.1823, + "step": 581 + }, + { + "epoch": 0.31596091205211724, + "grad_norm": 3.224137115744835, + "learning_rate": 1.9942527733658602e-05, + "loss": 1.5409, + "step": 582 + }, + { + "epoch": 0.31650380021715524, + "grad_norm": 3.126134123334164, + "learning_rate": 1.9942150612898194e-05, + "loss": 1.2423, + "step": 583 + }, + { + "epoch": 0.31704668838219324, + "grad_norm": 2.709954974374804, + "learning_rate": 1.994177226247675e-05, + "loss": 1.3419, + "step": 584 + }, + { + "epoch": 0.31758957654723124, + "grad_norm": 2.9030367366777927, + "learning_rate": 1.9941392682441066e-05, + "loss": 1.427, + "step": 585 + }, + { + "epoch": 0.31813246471226925, + "grad_norm": 3.272875981776567, + "learning_rate": 1.9941011872838092e-05, + "loss": 2.0196, + "step": 586 + }, + { + "epoch": 0.31867535287730725, + "grad_norm": 3.3480185179867, + "learning_rate": 1.994062983371493e-05, + "loss": 1.6038, + "step": 587 + }, + { + "epoch": 0.31921824104234525, + "grad_norm": 2.801738772846361, + "learning_rate": 1.9940246565118822e-05, + "loss": 1.7505, + "step": 588 + }, + { + "epoch": 0.31976112920738325, + "grad_norm": 2.743550967049156, + "learning_rate": 1.993986206709718e-05, + "loss": 1.3914, + "step": 589 + }, + { + "epoch": 0.32030401737242126, + "grad_norm": 2.481845489278486, + "learning_rate": 1.9939476339697555e-05, + "loss": 1.6927, + "step": 590 + }, + { + "epoch": 0.32084690553745926, + "grad_norm": 3.145019330878407, + "learning_rate": 1.993908938296765e-05, + "loss": 1.42, + "step": 591 + }, + { + "epoch": 0.32138979370249726, + "grad_norm": 3.326427662456394, + "learning_rate": 1.9938701196955335e-05, + "loss": 1.2516, + "step": 592 + }, + { + "epoch": 0.32193268186753526, + "grad_norm": 3.195815443331326, + "learning_rate": 1.9938311781708616e-05, + "loss": 2.1428, + "step": 593 + }, + { + "epoch": 0.32247557003257327, + "grad_norm": 2.461395361556941, + "learning_rate": 1.9937921137275657e-05, + "loss": 1.4448, + "step": 594 + }, + { + "epoch": 0.32301845819761127, + "grad_norm": 3.1793351214549794, + "learning_rate": 1.993752926370477e-05, + "loss": 1.4609, + "step": 595 + }, + { + "epoch": 0.32356134636264927, + "grad_norm": 2.783909288864463, + "learning_rate": 1.9937136161044427e-05, + "loss": 1.3355, + "step": 596 + }, + { + "epoch": 0.3241042345276873, + "grad_norm": 2.8156929696256734, + "learning_rate": 1.9936741829343247e-05, + "loss": 2.101, + "step": 597 + }, + { + "epoch": 0.3246471226927253, + "grad_norm": 3.793352093788154, + "learning_rate": 1.993634626865e-05, + "loss": 2.19, + "step": 598 + }, + { + "epoch": 0.3251900108577633, + "grad_norm": 2.6680863350639545, + "learning_rate": 1.993594947901361e-05, + "loss": 1.5199, + "step": 599 + }, + { + "epoch": 0.3257328990228013, + "grad_norm": 2.6498586813134297, + "learning_rate": 1.9935551460483155e-05, + "loss": 0.9282, + "step": 600 + }, + { + "epoch": 0.3262757871878393, + "grad_norm": 3.443063689484479, + "learning_rate": 1.993515221310786e-05, + "loss": 1.9107, + "step": 601 + }, + { + "epoch": 0.3268186753528773, + "grad_norm": 2.45686855619251, + "learning_rate": 1.9934751736937103e-05, + "loss": 1.2929, + "step": 602 + }, + { + "epoch": 0.3273615635179153, + "grad_norm": 3.3974155688994077, + "learning_rate": 1.9934350032020417e-05, + "loss": 1.546, + "step": 603 + }, + { + "epoch": 0.3279044516829533, + "grad_norm": 2.3692191456624783, + "learning_rate": 1.993394709840749e-05, + "loss": 1.2239, + "step": 604 + }, + { + "epoch": 0.3284473398479913, + "grad_norm": 4.044352925058249, + "learning_rate": 1.993354293614815e-05, + "loss": 1.801, + "step": 605 + }, + { + "epoch": 0.3289902280130293, + "grad_norm": 4.051192057617293, + "learning_rate": 1.993313754529239e-05, + "loss": 1.8043, + "step": 606 + }, + { + "epoch": 0.3295331161780673, + "grad_norm": 3.3680281079594634, + "learning_rate": 1.9932730925890344e-05, + "loss": 1.4915, + "step": 607 + }, + { + "epoch": 0.3300760043431053, + "grad_norm": 2.993677842102555, + "learning_rate": 1.9932323077992312e-05, + "loss": 1.3457, + "step": 608 + }, + { + "epoch": 0.3306188925081433, + "grad_norm": 3.325666927669253, + "learning_rate": 1.9931914001648726e-05, + "loss": 1.4221, + "step": 609 + }, + { + "epoch": 0.3311617806731813, + "grad_norm": 2.459382097232459, + "learning_rate": 1.993150369691019e-05, + "loss": 1.2551, + "step": 610 + }, + { + "epoch": 0.3317046688382193, + "grad_norm": 5.281638549560053, + "learning_rate": 1.993109216382745e-05, + "loss": 1.142, + "step": 611 + }, + { + "epoch": 0.3322475570032573, + "grad_norm": 2.6931033883715374, + "learning_rate": 1.99306794024514e-05, + "loss": 1.2573, + "step": 612 + }, + { + "epoch": 0.3327904451682953, + "grad_norm": 2.8066393604642714, + "learning_rate": 1.9930265412833097e-05, + "loss": 0.9847, + "step": 613 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 3.682172765988149, + "learning_rate": 1.992985019502374e-05, + "loss": 1.4482, + "step": 614 + }, + { + "epoch": 0.3338762214983713, + "grad_norm": 3.7190248655227873, + "learning_rate": 1.9929433749074684e-05, + "loss": 1.3682, + "step": 615 + }, + { + "epoch": 0.3344191096634093, + "grad_norm": 2.8594798198555527, + "learning_rate": 1.9929016075037438e-05, + "loss": 1.2396, + "step": 616 + }, + { + "epoch": 0.3349619978284473, + "grad_norm": 3.5439348644341035, + "learning_rate": 1.992859717296366e-05, + "loss": 1.5843, + "step": 617 + }, + { + "epoch": 0.3355048859934853, + "grad_norm": 2.3447575030783283, + "learning_rate": 1.992817704290516e-05, + "loss": 1.2025, + "step": 618 + }, + { + "epoch": 0.3360477741585233, + "grad_norm": 3.027039186876229, + "learning_rate": 1.99277556849139e-05, + "loss": 1.7133, + "step": 619 + }, + { + "epoch": 0.33659066232356133, + "grad_norm": 2.9868315770124343, + "learning_rate": 1.9927333099041992e-05, + "loss": 1.2309, + "step": 620 + }, + { + "epoch": 0.33713355048859933, + "grad_norm": 3.215889472723888, + "learning_rate": 1.9926909285341705e-05, + "loss": 1.2217, + "step": 621 + }, + { + "epoch": 0.33767643865363733, + "grad_norm": 3.293119071439456, + "learning_rate": 1.992648424386546e-05, + "loss": 1.7828, + "step": 622 + }, + { + "epoch": 0.33821932681867534, + "grad_norm": 2.7023629491691623, + "learning_rate": 1.992605797466582e-05, + "loss": 1.3745, + "step": 623 + }, + { + "epoch": 0.33876221498371334, + "grad_norm": 4.234534944070885, + "learning_rate": 1.9925630477795514e-05, + "loss": 1.3415, + "step": 624 + }, + { + "epoch": 0.33930510314875134, + "grad_norm": 3.6994706259321815, + "learning_rate": 1.9925201753307406e-05, + "loss": 1.4234, + "step": 625 + }, + { + "epoch": 0.33984799131378934, + "grad_norm": 3.156992028368943, + "learning_rate": 1.992477180125453e-05, + "loss": 1.4845, + "step": 626 + }, + { + "epoch": 0.34039087947882735, + "grad_norm": 3.1286393814611397, + "learning_rate": 1.9924340621690058e-05, + "loss": 0.9685, + "step": 627 + }, + { + "epoch": 0.34093376764386535, + "grad_norm": 3.336970766050665, + "learning_rate": 1.9923908214667323e-05, + "loss": 1.3268, + "step": 628 + }, + { + "epoch": 0.34147665580890335, + "grad_norm": 4.290496076646803, + "learning_rate": 1.99234745802398e-05, + "loss": 1.2785, + "step": 629 + }, + { + "epoch": 0.34201954397394135, + "grad_norm": 2.7491777959300023, + "learning_rate": 1.9923039718461127e-05, + "loss": 1.3798, + "step": 630 + }, + { + "epoch": 0.34256243213897936, + "grad_norm": 5.312529540414799, + "learning_rate": 1.992260362938509e-05, + "loss": 2.2512, + "step": 631 + }, + { + "epoch": 0.34310532030401736, + "grad_norm": 3.9557409230221676, + "learning_rate": 1.9922166313065618e-05, + "loss": 1.849, + "step": 632 + }, + { + "epoch": 0.34364820846905536, + "grad_norm": 3.7327574664165786, + "learning_rate": 1.9921727769556805e-05, + "loss": 1.4078, + "step": 633 + }, + { + "epoch": 0.34419109663409336, + "grad_norm": 3.7644055867642594, + "learning_rate": 1.992128799891289e-05, + "loss": 1.714, + "step": 634 + }, + { + "epoch": 0.34473398479913137, + "grad_norm": 3.923952524278675, + "learning_rate": 1.9920847001188258e-05, + "loss": 1.7613, + "step": 635 + }, + { + "epoch": 0.34527687296416937, + "grad_norm": 3.1523646374054155, + "learning_rate": 1.992040477643746e-05, + "loss": 1.2563, + "step": 636 + }, + { + "epoch": 0.34581976112920737, + "grad_norm": 3.6157151039083444, + "learning_rate": 1.991996132471519e-05, + "loss": 1.1795, + "step": 637 + }, + { + "epoch": 0.3463626492942454, + "grad_norm": 4.222007576307266, + "learning_rate": 1.991951664607629e-05, + "loss": 1.6605, + "step": 638 + }, + { + "epoch": 0.3469055374592834, + "grad_norm": 3.1555785318795544, + "learning_rate": 1.9919070740575764e-05, + "loss": 1.526, + "step": 639 + }, + { + "epoch": 0.3474484256243214, + "grad_norm": 3.2645269107818677, + "learning_rate": 1.991862360826876e-05, + "loss": 1.3182, + "step": 640 + }, + { + "epoch": 0.3479913137893594, + "grad_norm": 3.53044955128402, + "learning_rate": 1.991817524921058e-05, + "loss": 1.5583, + "step": 641 + }, + { + "epoch": 0.3485342019543974, + "grad_norm": 3.24710260787145, + "learning_rate": 1.9917725663456678e-05, + "loss": 0.9878, + "step": 642 + }, + { + "epoch": 0.3490770901194354, + "grad_norm": 3.9154742975791654, + "learning_rate": 1.991727485106266e-05, + "loss": 1.6034, + "step": 643 + }, + { + "epoch": 0.3496199782844734, + "grad_norm": 3.8199965805178318, + "learning_rate": 1.9916822812084282e-05, + "loss": 1.3768, + "step": 644 + }, + { + "epoch": 0.3501628664495114, + "grad_norm": 3.198793670499035, + "learning_rate": 1.9916369546577455e-05, + "loss": 1.4308, + "step": 645 + }, + { + "epoch": 0.3507057546145494, + "grad_norm": 2.908177540585374, + "learning_rate": 1.9915915054598237e-05, + "loss": 1.3964, + "step": 646 + }, + { + "epoch": 0.3512486427795874, + "grad_norm": 2.920889003436479, + "learning_rate": 1.9915459336202844e-05, + "loss": 1.056, + "step": 647 + }, + { + "epoch": 0.3517915309446254, + "grad_norm": 3.7046400158340864, + "learning_rate": 1.991500239144763e-05, + "loss": 1.9052, + "step": 648 + }, + { + "epoch": 0.3523344191096634, + "grad_norm": 4.412988121206581, + "learning_rate": 1.9914544220389124e-05, + "loss": 1.944, + "step": 649 + }, + { + "epoch": 0.3528773072747014, + "grad_norm": 3.87124697251994, + "learning_rate": 1.9914084823083988e-05, + "loss": 1.4951, + "step": 650 + }, + { + "epoch": 0.3534201954397394, + "grad_norm": 3.8734663848108584, + "learning_rate": 1.9913624199589037e-05, + "loss": 1.5462, + "step": 651 + }, + { + "epoch": 0.3539630836047774, + "grad_norm": 2.7283359997443126, + "learning_rate": 1.9913162349961248e-05, + "loss": 0.9188, + "step": 652 + }, + { + "epoch": 0.3545059717698154, + "grad_norm": 2.885333878853398, + "learning_rate": 1.991269927425774e-05, + "loss": 1.2292, + "step": 653 + }, + { + "epoch": 0.3550488599348534, + "grad_norm": 3.2425772460053257, + "learning_rate": 1.9912234972535788e-05, + "loss": 1.2863, + "step": 654 + }, + { + "epoch": 0.3555917480998914, + "grad_norm": 3.060966260943164, + "learning_rate": 1.991176944485281e-05, + "loss": 1.5056, + "step": 655 + }, + { + "epoch": 0.3561346362649294, + "grad_norm": 4.533553635387312, + "learning_rate": 1.99113026912664e-05, + "loss": 1.8229, + "step": 656 + }, + { + "epoch": 0.3566775244299674, + "grad_norm": 2.8582150809054045, + "learning_rate": 1.9910834711834267e-05, + "loss": 1.3233, + "step": 657 + }, + { + "epoch": 0.3572204125950054, + "grad_norm": 4.930680768753159, + "learning_rate": 1.9910365506614308e-05, + "loss": 1.5997, + "step": 658 + }, + { + "epoch": 0.3577633007600434, + "grad_norm": 3.0733335483268083, + "learning_rate": 1.9909895075664545e-05, + "loss": 0.9206, + "step": 659 + }, + { + "epoch": 0.3583061889250814, + "grad_norm": 3.4659704922712686, + "learning_rate": 1.990942341904317e-05, + "loss": 1.2752, + "step": 660 + }, + { + "epoch": 0.35884907709011943, + "grad_norm": 3.493850558129149, + "learning_rate": 1.9908950536808508e-05, + "loss": 1.588, + "step": 661 + }, + { + "epoch": 0.35939196525515743, + "grad_norm": 3.61256229272583, + "learning_rate": 1.9908476429019056e-05, + "loss": 1.469, + "step": 662 + }, + { + "epoch": 0.35993485342019543, + "grad_norm": 3.3430037514089803, + "learning_rate": 1.9908001095733445e-05, + "loss": 1.3038, + "step": 663 + }, + { + "epoch": 0.36047774158523344, + "grad_norm": 3.8918915484374557, + "learning_rate": 1.9907524537010467e-05, + "loss": 1.3683, + "step": 664 + }, + { + "epoch": 0.36102062975027144, + "grad_norm": 3.6032272045245053, + "learning_rate": 1.9907046752909064e-05, + "loss": 1.8694, + "step": 665 + }, + { + "epoch": 0.36156351791530944, + "grad_norm": 4.546268377661869, + "learning_rate": 1.9906567743488326e-05, + "loss": 1.2871, + "step": 666 + }, + { + "epoch": 0.36210640608034744, + "grad_norm": 3.5283169402860777, + "learning_rate": 1.9906087508807504e-05, + "loss": 1.5334, + "step": 667 + }, + { + "epoch": 0.36264929424538545, + "grad_norm": 2.7984191262784504, + "learning_rate": 1.9905606048925993e-05, + "loss": 1.1924, + "step": 668 + }, + { + "epoch": 0.36319218241042345, + "grad_norm": 3.54703835950394, + "learning_rate": 1.9905123363903335e-05, + "loss": 1.5972, + "step": 669 + }, + { + "epoch": 0.36373507057546145, + "grad_norm": 3.0126938064772855, + "learning_rate": 1.9904639453799236e-05, + "loss": 1.2294, + "step": 670 + }, + { + "epoch": 0.36427795874049945, + "grad_norm": 3.16358553989926, + "learning_rate": 1.990415431867354e-05, + "loss": 1.3564, + "step": 671 + }, + { + "epoch": 0.36482084690553745, + "grad_norm": 4.118910539945259, + "learning_rate": 1.990366795858626e-05, + "loss": 1.6838, + "step": 672 + }, + { + "epoch": 0.36536373507057546, + "grad_norm": 3.8466533902335502, + "learning_rate": 1.9903180373597534e-05, + "loss": 1.7986, + "step": 673 + }, + { + "epoch": 0.36590662323561346, + "grad_norm": 3.6195384409794684, + "learning_rate": 1.990269156376768e-05, + "loss": 1.6113, + "step": 674 + }, + { + "epoch": 0.36644951140065146, + "grad_norm": 3.509724379422402, + "learning_rate": 1.9902201529157152e-05, + "loss": 1.5496, + "step": 675 + }, + { + "epoch": 0.36699239956568946, + "grad_norm": 3.579737463592409, + "learning_rate": 1.9901710269826554e-05, + "loss": 1.4856, + "step": 676 + }, + { + "epoch": 0.36753528773072747, + "grad_norm": 3.780431786449365, + "learning_rate": 1.9901217785836655e-05, + "loss": 1.8519, + "step": 677 + }, + { + "epoch": 0.36807817589576547, + "grad_norm": 4.0659352627131735, + "learning_rate": 1.9900724077248354e-05, + "loss": 1.4382, + "step": 678 + }, + { + "epoch": 0.36862106406080347, + "grad_norm": 2.9524656376446274, + "learning_rate": 1.9900229144122723e-05, + "loss": 1.4541, + "step": 679 + }, + { + "epoch": 0.3691639522258415, + "grad_norm": 2.388269642575342, + "learning_rate": 1.989973298652097e-05, + "loss": 0.9567, + "step": 680 + }, + { + "epoch": 0.3697068403908795, + "grad_norm": 2.8322419431529453, + "learning_rate": 1.9899235604504467e-05, + "loss": 1.1187, + "step": 681 + }, + { + "epoch": 0.3702497285559175, + "grad_norm": 3.0469435898757613, + "learning_rate": 1.9898736998134726e-05, + "loss": 1.5571, + "step": 682 + }, + { + "epoch": 0.3707926167209555, + "grad_norm": 3.623982705749655, + "learning_rate": 1.9898237167473416e-05, + "loss": 1.2047, + "step": 683 + }, + { + "epoch": 0.3713355048859935, + "grad_norm": 3.669362267695381, + "learning_rate": 1.9897736112582357e-05, + "loss": 1.7747, + "step": 684 + }, + { + "epoch": 0.3718783930510315, + "grad_norm": 2.9109325710576353, + "learning_rate": 1.989723383352352e-05, + "loss": 1.0044, + "step": 685 + }, + { + "epoch": 0.3724212812160695, + "grad_norm": 3.4478434843760146, + "learning_rate": 1.9896730330359032e-05, + "loss": 1.1245, + "step": 686 + }, + { + "epoch": 0.3729641693811075, + "grad_norm": 3.2957514959337275, + "learning_rate": 1.989622560315116e-05, + "loss": 1.3243, + "step": 687 + }, + { + "epoch": 0.3735070575461455, + "grad_norm": 3.5500423086355988, + "learning_rate": 1.989571965196234e-05, + "loss": 2.0478, + "step": 688 + }, + { + "epoch": 0.3740499457111835, + "grad_norm": 2.621987663839103, + "learning_rate": 1.9895212476855136e-05, + "loss": 1.3135, + "step": 689 + }, + { + "epoch": 0.3745928338762215, + "grad_norm": 4.055390555970487, + "learning_rate": 1.989470407789228e-05, + "loss": 1.5165, + "step": 690 + }, + { + "epoch": 0.3751357220412595, + "grad_norm": 2.94204110816024, + "learning_rate": 1.989419445513666e-05, + "loss": 1.3458, + "step": 691 + }, + { + "epoch": 0.3756786102062975, + "grad_norm": 3.5714132672245977, + "learning_rate": 1.98936836086513e-05, + "loss": 1.3034, + "step": 692 + }, + { + "epoch": 0.3762214983713355, + "grad_norm": 3.659137782783938, + "learning_rate": 1.9893171538499382e-05, + "loss": 1.6203, + "step": 693 + }, + { + "epoch": 0.3767643865363735, + "grad_norm": 5.116097300755018, + "learning_rate": 1.9892658244744236e-05, + "loss": 2.1071, + "step": 694 + }, + { + "epoch": 0.3773072747014115, + "grad_norm": 2.8449662018307005, + "learning_rate": 1.9892143727449357e-05, + "loss": 1.0477, + "step": 695 + }, + { + "epoch": 0.3778501628664495, + "grad_norm": 4.021030963638319, + "learning_rate": 1.989162798667838e-05, + "loss": 1.9528, + "step": 696 + }, + { + "epoch": 0.3783930510314875, + "grad_norm": 3.377368666687089, + "learning_rate": 1.989111102249508e-05, + "loss": 1.2481, + "step": 697 + }, + { + "epoch": 0.3789359391965255, + "grad_norm": 3.403268816169458, + "learning_rate": 1.9890592834963406e-05, + "loss": 1.1864, + "step": 698 + }, + { + "epoch": 0.3794788273615635, + "grad_norm": 3.082879601892987, + "learning_rate": 1.9890073424147453e-05, + "loss": 1.7365, + "step": 699 + }, + { + "epoch": 0.3800217155266015, + "grad_norm": 4.122610250215441, + "learning_rate": 1.988955279011145e-05, + "loss": 1.4139, + "step": 700 + }, + { + "epoch": 0.3805646036916395, + "grad_norm": 3.298087120576001, + "learning_rate": 1.98890309329198e-05, + "loss": 1.5904, + "step": 701 + }, + { + "epoch": 0.3811074918566775, + "grad_norm": 3.1183850173888703, + "learning_rate": 1.9888507852637043e-05, + "loss": 1.2146, + "step": 702 + }, + { + "epoch": 0.38165038002171553, + "grad_norm": 3.128042971411966, + "learning_rate": 1.9887983549327873e-05, + "loss": 1.2751, + "step": 703 + }, + { + "epoch": 0.38219326818675353, + "grad_norm": 3.362555156234201, + "learning_rate": 1.988745802305714e-05, + "loss": 1.1363, + "step": 704 + }, + { + "epoch": 0.38273615635179153, + "grad_norm": 3.29574344438116, + "learning_rate": 1.988693127388984e-05, + "loss": 1.227, + "step": 705 + }, + { + "epoch": 0.38327904451682954, + "grad_norm": 4.2418853227437205, + "learning_rate": 1.9886403301891123e-05, + "loss": 1.7091, + "step": 706 + }, + { + "epoch": 0.38382193268186754, + "grad_norm": 4.165071133964158, + "learning_rate": 1.9885874107126287e-05, + "loss": 1.9403, + "step": 707 + }, + { + "epoch": 0.38436482084690554, + "grad_norm": 4.639609386348944, + "learning_rate": 1.9885343689660787e-05, + "loss": 1.221, + "step": 708 + }, + { + "epoch": 0.38490770901194354, + "grad_norm": 3.0746775894794034, + "learning_rate": 1.9884812049560226e-05, + "loss": 1.4958, + "step": 709 + }, + { + "epoch": 0.38545059717698155, + "grad_norm": 4.822804859325608, + "learning_rate": 1.9884279186890357e-05, + "loss": 1.6928, + "step": 710 + }, + { + "epoch": 0.38599348534201955, + "grad_norm": 4.299811491105524, + "learning_rate": 1.9883745101717084e-05, + "loss": 1.3075, + "step": 711 + }, + { + "epoch": 0.38653637350705755, + "grad_norm": 3.882170994492157, + "learning_rate": 1.9883209794106464e-05, + "loss": 1.9712, + "step": 712 + }, + { + "epoch": 0.38707926167209555, + "grad_norm": 4.164019029255246, + "learning_rate": 1.9882673264124705e-05, + "loss": 1.6109, + "step": 713 + }, + { + "epoch": 0.38762214983713356, + "grad_norm": 3.691226317297577, + "learning_rate": 1.9882135511838167e-05, + "loss": 1.3225, + "step": 714 + }, + { + "epoch": 0.38816503800217156, + "grad_norm": 5.777282505586183, + "learning_rate": 1.988159653731336e-05, + "loss": 1.7945, + "step": 715 + }, + { + "epoch": 0.38870792616720956, + "grad_norm": 3.3926512133724165, + "learning_rate": 1.9881056340616944e-05, + "loss": 1.5797, + "step": 716 + }, + { + "epoch": 0.38925081433224756, + "grad_norm": 3.7756965718668467, + "learning_rate": 1.988051492181573e-05, + "loss": 1.5515, + "step": 717 + }, + { + "epoch": 0.38979370249728557, + "grad_norm": 3.5509658648264613, + "learning_rate": 1.987997228097668e-05, + "loss": 1.2004, + "step": 718 + }, + { + "epoch": 0.39033659066232357, + "grad_norm": 4.682493032945398, + "learning_rate": 1.987942841816692e-05, + "loss": 1.2957, + "step": 719 + }, + { + "epoch": 0.39087947882736157, + "grad_norm": 2.923319971884201, + "learning_rate": 1.9878883333453704e-05, + "loss": 0.753, + "step": 720 + }, + { + "epoch": 0.3914223669923996, + "grad_norm": 2.8706465146507845, + "learning_rate": 1.987833702690445e-05, + "loss": 1.1057, + "step": 721 + }, + { + "epoch": 0.3919652551574376, + "grad_norm": 3.4516078574834057, + "learning_rate": 1.987778949858673e-05, + "loss": 1.1275, + "step": 722 + }, + { + "epoch": 0.3925081433224756, + "grad_norm": 3.3400553957950567, + "learning_rate": 1.9877240748568263e-05, + "loss": 1.1538, + "step": 723 + }, + { + "epoch": 0.3930510314875136, + "grad_norm": 3.155540544638446, + "learning_rate": 1.987669077691692e-05, + "loss": 1.2486, + "step": 724 + }, + { + "epoch": 0.3935939196525516, + "grad_norm": 3.1450276486107054, + "learning_rate": 1.987613958370072e-05, + "loss": 1.265, + "step": 725 + }, + { + "epoch": 0.3941368078175896, + "grad_norm": 3.0493327058864406, + "learning_rate": 1.9875587168987834e-05, + "loss": 1.2097, + "step": 726 + }, + { + "epoch": 0.3946796959826276, + "grad_norm": 3.716259137065701, + "learning_rate": 1.987503353284659e-05, + "loss": 1.5386, + "step": 727 + }, + { + "epoch": 0.3952225841476656, + "grad_norm": 3.441725300410296, + "learning_rate": 1.9874478675345458e-05, + "loss": 1.3936, + "step": 728 + }, + { + "epoch": 0.3957654723127036, + "grad_norm": 3.773353353497139, + "learning_rate": 1.9873922596553067e-05, + "loss": 1.731, + "step": 729 + }, + { + "epoch": 0.3963083604777416, + "grad_norm": 3.527645597828566, + "learning_rate": 1.987336529653819e-05, + "loss": 1.708, + "step": 730 + }, + { + "epoch": 0.3968512486427796, + "grad_norm": 3.8440953147603643, + "learning_rate": 1.9872806775369762e-05, + "loss": 1.6102, + "step": 731 + }, + { + "epoch": 0.3973941368078176, + "grad_norm": 3.4036622029265966, + "learning_rate": 1.9872247033116855e-05, + "loss": 0.9298, + "step": 732 + }, + { + "epoch": 0.3979370249728556, + "grad_norm": 4.034027632823603, + "learning_rate": 1.98716860698487e-05, + "loss": 1.78, + "step": 733 + }, + { + "epoch": 0.3984799131378936, + "grad_norm": 4.10524690369694, + "learning_rate": 1.987112388563468e-05, + "loss": 1.2354, + "step": 734 + }, + { + "epoch": 0.3990228013029316, + "grad_norm": 4.059289261823474, + "learning_rate": 1.9870560480544325e-05, + "loss": 1.804, + "step": 735 + }, + { + "epoch": 0.3995656894679696, + "grad_norm": 3.2544839104292667, + "learning_rate": 1.986999585464732e-05, + "loss": 1.1228, + "step": 736 + }, + { + "epoch": 0.4001085776330076, + "grad_norm": 3.741117309135591, + "learning_rate": 1.9869430008013496e-05, + "loss": 1.329, + "step": 737 + }, + { + "epoch": 0.4006514657980456, + "grad_norm": 4.446330579981585, + "learning_rate": 1.9868862940712838e-05, + "loss": 1.6506, + "step": 738 + }, + { + "epoch": 0.4011943539630836, + "grad_norm": 3.139296317304318, + "learning_rate": 1.9868294652815483e-05, + "loss": 1.144, + "step": 739 + }, + { + "epoch": 0.4017372421281216, + "grad_norm": 4.025270331554587, + "learning_rate": 1.986772514439172e-05, + "loss": 1.336, + "step": 740 + }, + { + "epoch": 0.4022801302931596, + "grad_norm": 3.4545518889756557, + "learning_rate": 1.986715441551198e-05, + "loss": 1.6006, + "step": 741 + }, + { + "epoch": 0.4028230184581976, + "grad_norm": 3.748598610621174, + "learning_rate": 1.986658246624686e-05, + "loss": 1.4812, + "step": 742 + }, + { + "epoch": 0.4033659066232356, + "grad_norm": 5.126058311631635, + "learning_rate": 1.9866009296667093e-05, + "loss": 2.0129, + "step": 743 + }, + { + "epoch": 0.40390879478827363, + "grad_norm": 8.145692299678405, + "learning_rate": 1.9865434906843574e-05, + "loss": 1.8321, + "step": 744 + }, + { + "epoch": 0.40445168295331163, + "grad_norm": 4.23918362592579, + "learning_rate": 1.9864859296847343e-05, + "loss": 1.5765, + "step": 745 + }, + { + "epoch": 0.40499457111834963, + "grad_norm": 4.85145039296263, + "learning_rate": 1.986428246674959e-05, + "loss": 1.6486, + "step": 746 + }, + { + "epoch": 0.40553745928338764, + "grad_norm": 3.7681072947070415, + "learning_rate": 1.986370441662166e-05, + "loss": 1.4466, + "step": 747 + }, + { + "epoch": 0.40608034744842564, + "grad_norm": 4.206136090938125, + "learning_rate": 1.986312514653505e-05, + "loss": 1.6717, + "step": 748 + }, + { + "epoch": 0.40662323561346364, + "grad_norm": 3.7336223866799965, + "learning_rate": 1.9862544656561403e-05, + "loss": 0.9599, + "step": 749 + }, + { + "epoch": 0.40716612377850164, + "grad_norm": 4.037759008136602, + "learning_rate": 1.986196294677251e-05, + "loss": 1.9387, + "step": 750 + }, + { + "epoch": 0.40770901194353965, + "grad_norm": 3.8832525632606156, + "learning_rate": 1.9861380017240324e-05, + "loss": 2.0184, + "step": 751 + }, + { + "epoch": 0.40825190010857765, + "grad_norm": 3.9521376069923546, + "learning_rate": 1.986079586803694e-05, + "loss": 1.0926, + "step": 752 + }, + { + "epoch": 0.40879478827361565, + "grad_norm": 5.128973903987411, + "learning_rate": 1.986021049923461e-05, + "loss": 1.5943, + "step": 753 + }, + { + "epoch": 0.40933767643865365, + "grad_norm": 4.5372892013581865, + "learning_rate": 1.9859623910905728e-05, + "loss": 1.643, + "step": 754 + }, + { + "epoch": 0.40988056460369166, + "grad_norm": 4.2137978897196415, + "learning_rate": 1.985903610312285e-05, + "loss": 1.7237, + "step": 755 + }, + { + "epoch": 0.41042345276872966, + "grad_norm": 2.713265124864733, + "learning_rate": 1.985844707595867e-05, + "loss": 1.0521, + "step": 756 + }, + { + "epoch": 0.41096634093376766, + "grad_norm": 3.362872283003281, + "learning_rate": 1.9857856829486045e-05, + "loss": 1.2879, + "step": 757 + }, + { + "epoch": 0.41150922909880566, + "grad_norm": 4.022235417044952, + "learning_rate": 1.9857265363777975e-05, + "loss": 1.398, + "step": 758 + }, + { + "epoch": 0.41205211726384366, + "grad_norm": 4.083280210123792, + "learning_rate": 1.9856672678907616e-05, + "loss": 1.6652, + "step": 759 + }, + { + "epoch": 0.41259500542888167, + "grad_norm": 4.469771898345589, + "learning_rate": 1.985607877494827e-05, + "loss": 1.6242, + "step": 760 + }, + { + "epoch": 0.41313789359391967, + "grad_norm": 3.0490808076449016, + "learning_rate": 1.9855483651973396e-05, + "loss": 1.0783, + "step": 761 + }, + { + "epoch": 0.41368078175895767, + "grad_norm": 3.488230005104396, + "learning_rate": 1.9854887310056593e-05, + "loss": 1.3953, + "step": 762 + }, + { + "epoch": 0.4142236699239957, + "grad_norm": 3.5610333355465653, + "learning_rate": 1.9854289749271624e-05, + "loss": 1.3692, + "step": 763 + }, + { + "epoch": 0.4147665580890337, + "grad_norm": 3.5572639096784533, + "learning_rate": 1.9853690969692393e-05, + "loss": 1.646, + "step": 764 + }, + { + "epoch": 0.4153094462540717, + "grad_norm": 3.811741925703646, + "learning_rate": 1.9853090971392953e-05, + "loss": 1.6637, + "step": 765 + }, + { + "epoch": 0.4158523344191097, + "grad_norm": 2.912609699248709, + "learning_rate": 1.9852489754447526e-05, + "loss": 0.9086, + "step": 766 + }, + { + "epoch": 0.4163952225841477, + "grad_norm": 3.7608762982374646, + "learning_rate": 1.985188731893046e-05, + "loss": 1.5131, + "step": 767 + }, + { + "epoch": 0.4169381107491857, + "grad_norm": 3.776000457813264, + "learning_rate": 1.985128366491627e-05, + "loss": 1.5929, + "step": 768 + }, + { + "epoch": 0.4174809989142237, + "grad_norm": 4.129873624081247, + "learning_rate": 1.9850678792479613e-05, + "loss": 1.5461, + "step": 769 + }, + { + "epoch": 0.4180238870792617, + "grad_norm": 3.2233425858963645, + "learning_rate": 1.9850072701695306e-05, + "loss": 0.9614, + "step": 770 + }, + { + "epoch": 0.4185667752442997, + "grad_norm": 3.8215876727885054, + "learning_rate": 1.984946539263831e-05, + "loss": 1.2549, + "step": 771 + }, + { + "epoch": 0.4191096634093377, + "grad_norm": 3.0062072101032125, + "learning_rate": 1.9848856865383732e-05, + "loss": 0.9796, + "step": 772 + }, + { + "epoch": 0.4196525515743757, + "grad_norm": 2.937513473620578, + "learning_rate": 1.984824712000684e-05, + "loss": 1.0204, + "step": 773 + }, + { + "epoch": 0.4201954397394137, + "grad_norm": 2.949279428268976, + "learning_rate": 1.984763615658305e-05, + "loss": 1.332, + "step": 774 + }, + { + "epoch": 0.4207383279044517, + "grad_norm": 3.455520053658832, + "learning_rate": 1.9847023975187925e-05, + "loss": 1.3786, + "step": 775 + }, + { + "epoch": 0.4212812160694897, + "grad_norm": 3.8352233275650733, + "learning_rate": 1.9846410575897183e-05, + "loss": 1.4897, + "step": 776 + }, + { + "epoch": 0.4218241042345277, + "grad_norm": 3.262791090623995, + "learning_rate": 1.984579595878669e-05, + "loss": 1.5097, + "step": 777 + }, + { + "epoch": 0.4223669923995657, + "grad_norm": 3.603809745585289, + "learning_rate": 1.9845180123932456e-05, + "loss": 1.8419, + "step": 778 + }, + { + "epoch": 0.4229098805646037, + "grad_norm": 3.9648489260279196, + "learning_rate": 1.9844563071410656e-05, + "loss": 1.5159, + "step": 779 + }, + { + "epoch": 0.4234527687296417, + "grad_norm": 4.576987136894564, + "learning_rate": 1.9843944801297605e-05, + "loss": 1.6866, + "step": 780 + }, + { + "epoch": 0.4239956568946797, + "grad_norm": 3.4164331578089406, + "learning_rate": 1.9843325313669774e-05, + "loss": 1.5533, + "step": 781 + }, + { + "epoch": 0.4245385450597177, + "grad_norm": 3.2121441901059367, + "learning_rate": 1.9842704608603774e-05, + "loss": 0.8879, + "step": 782 + }, + { + "epoch": 0.4250814332247557, + "grad_norm": 3.6213877563860537, + "learning_rate": 1.9842082686176388e-05, + "loss": 1.3098, + "step": 783 + }, + { + "epoch": 0.4256243213897937, + "grad_norm": 3.719993086277978, + "learning_rate": 1.9841459546464527e-05, + "loss": 1.331, + "step": 784 + }, + { + "epoch": 0.4261672095548317, + "grad_norm": 2.7952357248876516, + "learning_rate": 1.9840835189545266e-05, + "loss": 1.1298, + "step": 785 + }, + { + "epoch": 0.42671009771986973, + "grad_norm": 3.450597586563723, + "learning_rate": 1.9840209615495822e-05, + "loss": 1.0023, + "step": 786 + }, + { + "epoch": 0.42725298588490773, + "grad_norm": 4.338480682869939, + "learning_rate": 1.983958282439357e-05, + "loss": 1.7044, + "step": 787 + }, + { + "epoch": 0.42779587404994573, + "grad_norm": 3.4831324579282765, + "learning_rate": 1.983895481631603e-05, + "loss": 1.4948, + "step": 788 + }, + { + "epoch": 0.42833876221498374, + "grad_norm": 6.382810905837672, + "learning_rate": 1.9838325591340885e-05, + "loss": 2.3008, + "step": 789 + }, + { + "epoch": 0.42888165038002174, + "grad_norm": 3.063848408601439, + "learning_rate": 1.9837695149545945e-05, + "loss": 1.2045, + "step": 790 + }, + { + "epoch": 0.42942453854505974, + "grad_norm": 4.80361465821882, + "learning_rate": 1.9837063491009193e-05, + "loss": 1.2552, + "step": 791 + }, + { + "epoch": 0.42996742671009774, + "grad_norm": 4.16884813133452, + "learning_rate": 1.9836430615808745e-05, + "loss": 1.6618, + "step": 792 + }, + { + "epoch": 0.43051031487513575, + "grad_norm": 4.215747683493222, + "learning_rate": 1.9835796524022886e-05, + "loss": 1.2571, + "step": 793 + }, + { + "epoch": 0.43105320304017375, + "grad_norm": 4.894853560761769, + "learning_rate": 1.9835161215730038e-05, + "loss": 0.793, + "step": 794 + }, + { + "epoch": 0.43159609120521175, + "grad_norm": 3.264854272979086, + "learning_rate": 1.983452469100877e-05, + "loss": 1.0721, + "step": 795 + }, + { + "epoch": 0.43213897937024975, + "grad_norm": 3.2346237132359335, + "learning_rate": 1.9833886949937823e-05, + "loss": 1.2396, + "step": 796 + }, + { + "epoch": 0.43268186753528776, + "grad_norm": 3.7265266370894023, + "learning_rate": 1.9833247992596058e-05, + "loss": 1.8845, + "step": 797 + }, + { + "epoch": 0.43322475570032576, + "grad_norm": 3.0655010730082126, + "learning_rate": 1.9832607819062513e-05, + "loss": 1.556, + "step": 798 + }, + { + "epoch": 0.43376764386536376, + "grad_norm": 3.209417077142042, + "learning_rate": 1.983196642941636e-05, + "loss": 1.2183, + "step": 799 + }, + { + "epoch": 0.43431053203040176, + "grad_norm": 4.384697511143271, + "learning_rate": 1.9831323823736933e-05, + "loss": 2.0889, + "step": 800 + }, + { + "epoch": 0.43485342019543977, + "grad_norm": 3.2773759507081084, + "learning_rate": 1.9830680002103703e-05, + "loss": 0.919, + "step": 801 + }, + { + "epoch": 0.43539630836047777, + "grad_norm": 4.1436406822638805, + "learning_rate": 1.9830034964596304e-05, + "loss": 1.7759, + "step": 802 + }, + { + "epoch": 0.43593919652551577, + "grad_norm": 5.715436881791243, + "learning_rate": 1.9829388711294512e-05, + "loss": 1.4277, + "step": 803 + }, + { + "epoch": 0.4364820846905538, + "grad_norm": 3.166985124043999, + "learning_rate": 1.982874124227826e-05, + "loss": 1.56, + "step": 804 + }, + { + "epoch": 0.4370249728555918, + "grad_norm": 4.237249076660815, + "learning_rate": 1.9828092557627626e-05, + "loss": 1.2366, + "step": 805 + }, + { + "epoch": 0.4375678610206297, + "grad_norm": 3.073398749866342, + "learning_rate": 1.982744265742284e-05, + "loss": 1.1475, + "step": 806 + }, + { + "epoch": 0.4381107491856677, + "grad_norm": 3.356777818996544, + "learning_rate": 1.9826791541744285e-05, + "loss": 1.2666, + "step": 807 + }, + { + "epoch": 0.4386536373507057, + "grad_norm": 3.171383453112648, + "learning_rate": 1.982613921067249e-05, + "loss": 1.1407, + "step": 808 + }, + { + "epoch": 0.43919652551574373, + "grad_norm": 3.334493176451811, + "learning_rate": 1.9825485664288138e-05, + "loss": 1.3309, + "step": 809 + }, + { + "epoch": 0.43973941368078173, + "grad_norm": 3.199045603289457, + "learning_rate": 1.982483090267206e-05, + "loss": 1.1634, + "step": 810 + }, + { + "epoch": 0.44028230184581973, + "grad_norm": 2.9970299057571856, + "learning_rate": 1.9824174925905235e-05, + "loss": 0.9642, + "step": 811 + }, + { + "epoch": 0.44082519001085774, + "grad_norm": 3.9851247078132994, + "learning_rate": 1.98235177340688e-05, + "loss": 1.4926, + "step": 812 + }, + { + "epoch": 0.44136807817589574, + "grad_norm": 3.586604601363014, + "learning_rate": 1.9822859327244034e-05, + "loss": 1.3105, + "step": 813 + }, + { + "epoch": 0.44191096634093374, + "grad_norm": 4.198634753759357, + "learning_rate": 1.9822199705512372e-05, + "loss": 1.4437, + "step": 814 + }, + { + "epoch": 0.44245385450597174, + "grad_norm": 3.6499375531563096, + "learning_rate": 1.9821538868955394e-05, + "loss": 1.2063, + "step": 815 + }, + { + "epoch": 0.44299674267100975, + "grad_norm": 4.3973905860052716, + "learning_rate": 1.9820876817654836e-05, + "loss": 1.5002, + "step": 816 + }, + { + "epoch": 0.44353963083604775, + "grad_norm": 2.7871960857121287, + "learning_rate": 1.9820213551692585e-05, + "loss": 0.6589, + "step": 817 + }, + { + "epoch": 0.44408251900108575, + "grad_norm": 3.4126665543063184, + "learning_rate": 1.981954907115067e-05, + "loss": 1.1651, + "step": 818 + }, + { + "epoch": 0.44462540716612375, + "grad_norm": 4.229651941034759, + "learning_rate": 1.981888337611127e-05, + "loss": 1.6444, + "step": 819 + }, + { + "epoch": 0.44516829533116176, + "grad_norm": 3.4135993440364056, + "learning_rate": 1.981821646665673e-05, + "loss": 1.2614, + "step": 820 + }, + { + "epoch": 0.44571118349619976, + "grad_norm": 4.274078711060124, + "learning_rate": 1.9817548342869527e-05, + "loss": 1.66, + "step": 821 + }, + { + "epoch": 0.44625407166123776, + "grad_norm": 2.9406863812447903, + "learning_rate": 1.98168790048323e-05, + "loss": 1.0412, + "step": 822 + }, + { + "epoch": 0.44679695982627576, + "grad_norm": 4.300279758003301, + "learning_rate": 1.981620845262783e-05, + "loss": 1.3689, + "step": 823 + }, + { + "epoch": 0.44733984799131377, + "grad_norm": 3.130525530503507, + "learning_rate": 1.9815536686339056e-05, + "loss": 1.1771, + "step": 824 + }, + { + "epoch": 0.44788273615635177, + "grad_norm": 3.6606030046920597, + "learning_rate": 1.981486370604906e-05, + "loss": 1.0323, + "step": 825 + }, + { + "epoch": 0.44842562432138977, + "grad_norm": 3.495865966877518, + "learning_rate": 1.981418951184108e-05, + "loss": 1.4493, + "step": 826 + }, + { + "epoch": 0.4489685124864278, + "grad_norm": 3.621229661878418, + "learning_rate": 1.9813514103798498e-05, + "loss": 1.5392, + "step": 827 + }, + { + "epoch": 0.4495114006514658, + "grad_norm": 3.6798995423063485, + "learning_rate": 1.9812837482004853e-05, + "loss": 1.0593, + "step": 828 + }, + { + "epoch": 0.4500542888165038, + "grad_norm": 3.4514040085362954, + "learning_rate": 1.9812159646543824e-05, + "loss": 0.9262, + "step": 829 + }, + { + "epoch": 0.4505971769815418, + "grad_norm": 3.195454430712678, + "learning_rate": 1.9811480597499257e-05, + "loss": 1.1469, + "step": 830 + }, + { + "epoch": 0.4511400651465798, + "grad_norm": 4.590874916307442, + "learning_rate": 1.981080033495513e-05, + "loss": 1.6534, + "step": 831 + }, + { + "epoch": 0.4516829533116178, + "grad_norm": 3.668702667937113, + "learning_rate": 1.981011885899558e-05, + "loss": 1.0597, + "step": 832 + }, + { + "epoch": 0.4522258414766558, + "grad_norm": 2.7669639542025046, + "learning_rate": 1.98094361697049e-05, + "loss": 1.0387, + "step": 833 + }, + { + "epoch": 0.4527687296416938, + "grad_norm": 4.734510249253564, + "learning_rate": 1.9808752267167515e-05, + "loss": 1.6757, + "step": 834 + }, + { + "epoch": 0.4533116178067318, + "grad_norm": 9.006078348758516, + "learning_rate": 1.9808067151468018e-05, + "loss": 2.1559, + "step": 835 + }, + { + "epoch": 0.4538545059717698, + "grad_norm": 4.169764802563779, + "learning_rate": 1.9807380822691146e-05, + "loss": 1.2557, + "step": 836 + }, + { + "epoch": 0.4543973941368078, + "grad_norm": 3.970541817152741, + "learning_rate": 1.9806693280921783e-05, + "loss": 1.5586, + "step": 837 + }, + { + "epoch": 0.4549402823018458, + "grad_norm": 5.042448582457387, + "learning_rate": 1.9806004526244966e-05, + "loss": 1.3191, + "step": 838 + }, + { + "epoch": 0.4554831704668838, + "grad_norm": 3.7467475303782236, + "learning_rate": 1.9805314558745876e-05, + "loss": 1.3318, + "step": 839 + }, + { + "epoch": 0.4560260586319218, + "grad_norm": 4.214820131429537, + "learning_rate": 1.980462337850986e-05, + "loss": 1.3958, + "step": 840 + }, + { + "epoch": 0.4565689467969598, + "grad_norm": 3.9655409992603965, + "learning_rate": 1.980393098562239e-05, + "loss": 0.9978, + "step": 841 + }, + { + "epoch": 0.4571118349619978, + "grad_norm": 5.522638140748038, + "learning_rate": 1.980323738016912e-05, + "loss": 1.4151, + "step": 842 + }, + { + "epoch": 0.4576547231270358, + "grad_norm": 3.84359618797871, + "learning_rate": 1.9802542562235822e-05, + "loss": 1.3567, + "step": 843 + }, + { + "epoch": 0.4581976112920738, + "grad_norm": 5.115138943098397, + "learning_rate": 1.9801846531908437e-05, + "loss": 1.6437, + "step": 844 + }, + { + "epoch": 0.4587404994571118, + "grad_norm": 3.811847477591368, + "learning_rate": 1.9801149289273054e-05, + "loss": 1.2335, + "step": 845 + }, + { + "epoch": 0.4592833876221498, + "grad_norm": 4.4009848681180115, + "learning_rate": 1.98004508344159e-05, + "loss": 1.2147, + "step": 846 + }, + { + "epoch": 0.4598262757871878, + "grad_norm": 3.850991850744113, + "learning_rate": 1.9799751167423376e-05, + "loss": 1.4865, + "step": 847 + }, + { + "epoch": 0.4603691639522258, + "grad_norm": 3.1303042065802917, + "learning_rate": 1.9799050288382004e-05, + "loss": 1.4926, + "step": 848 + }, + { + "epoch": 0.4609120521172638, + "grad_norm": 4.225513214699549, + "learning_rate": 1.9798348197378475e-05, + "loss": 1.2044, + "step": 849 + }, + { + "epoch": 0.46145494028230183, + "grad_norm": 4.139979880818486, + "learning_rate": 1.9797644894499625e-05, + "loss": 1.5682, + "step": 850 + }, + { + "epoch": 0.46199782844733983, + "grad_norm": 4.414209735978709, + "learning_rate": 1.979694037983244e-05, + "loss": 1.9033, + "step": 851 + }, + { + "epoch": 0.46254071661237783, + "grad_norm": 5.079031167256292, + "learning_rate": 1.9796234653464057e-05, + "loss": 1.8745, + "step": 852 + }, + { + "epoch": 0.46308360477741584, + "grad_norm": 6.59971413476838, + "learning_rate": 1.9795527715481755e-05, + "loss": 1.2449, + "step": 853 + }, + { + "epoch": 0.46362649294245384, + "grad_norm": 3.863740127637264, + "learning_rate": 1.9794819565972973e-05, + "loss": 1.3912, + "step": 854 + }, + { + "epoch": 0.46416938110749184, + "grad_norm": 3.8578186204961327, + "learning_rate": 1.9794110205025302e-05, + "loss": 1.3487, + "step": 855 + }, + { + "epoch": 0.46471226927252984, + "grad_norm": 4.19751015099396, + "learning_rate": 1.9793399632726466e-05, + "loss": 1.129, + "step": 856 + }, + { + "epoch": 0.46525515743756785, + "grad_norm": 4.261483190159906, + "learning_rate": 1.979268784916436e-05, + "loss": 1.7298, + "step": 857 + }, + { + "epoch": 0.46579804560260585, + "grad_norm": 4.1774094914047435, + "learning_rate": 1.9791974854427008e-05, + "loss": 0.8716, + "step": 858 + }, + { + "epoch": 0.46634093376764385, + "grad_norm": 3.845723101011648, + "learning_rate": 1.9791260648602603e-05, + "loss": 0.9086, + "step": 859 + }, + { + "epoch": 0.46688382193268185, + "grad_norm": 3.4922208632687393, + "learning_rate": 1.9790545231779476e-05, + "loss": 1.3688, + "step": 860 + }, + { + "epoch": 0.46742671009771986, + "grad_norm": 2.976761133296452, + "learning_rate": 1.978982860404611e-05, + "loss": 0.7865, + "step": 861 + }, + { + "epoch": 0.46796959826275786, + "grad_norm": 3.934765402497862, + "learning_rate": 1.978911076549114e-05, + "loss": 1.2993, + "step": 862 + }, + { + "epoch": 0.46851248642779586, + "grad_norm": 3.029198772099951, + "learning_rate": 1.978839171620335e-05, + "loss": 1.1737, + "step": 863 + }, + { + "epoch": 0.46905537459283386, + "grad_norm": 3.5637545602445946, + "learning_rate": 1.978767145627167e-05, + "loss": 1.1932, + "step": 864 + }, + { + "epoch": 0.46959826275787186, + "grad_norm": 3.131289093388952, + "learning_rate": 1.9786949985785187e-05, + "loss": 0.69, + "step": 865 + }, + { + "epoch": 0.47014115092290987, + "grad_norm": 3.522468470152659, + "learning_rate": 1.978622730483313e-05, + "loss": 1.0277, + "step": 866 + }, + { + "epoch": 0.47068403908794787, + "grad_norm": 4.046439598363532, + "learning_rate": 1.9785503413504883e-05, + "loss": 1.2917, + "step": 867 + }, + { + "epoch": 0.47122692725298587, + "grad_norm": 5.55259287782346, + "learning_rate": 1.978477831188998e-05, + "loss": 1.866, + "step": 868 + }, + { + "epoch": 0.4717698154180239, + "grad_norm": 5.874766516228883, + "learning_rate": 1.97840520000781e-05, + "loss": 1.6221, + "step": 869 + }, + { + "epoch": 0.4723127035830619, + "grad_norm": 5.530080877073305, + "learning_rate": 1.9783324478159074e-05, + "loss": 1.7617, + "step": 870 + }, + { + "epoch": 0.4728555917480999, + "grad_norm": 4.996268624326209, + "learning_rate": 1.9782595746222886e-05, + "loss": 1.5794, + "step": 871 + }, + { + "epoch": 0.4733984799131379, + "grad_norm": 3.934049315300115, + "learning_rate": 1.9781865804359663e-05, + "loss": 1.2603, + "step": 872 + }, + { + "epoch": 0.4739413680781759, + "grad_norm": 4.534039775579426, + "learning_rate": 1.978113465265969e-05, + "loss": 1.2766, + "step": 873 + }, + { + "epoch": 0.4744842562432139, + "grad_norm": 4.281791787646871, + "learning_rate": 1.9780402291213393e-05, + "loss": 1.6424, + "step": 874 + }, + { + "epoch": 0.4750271444082519, + "grad_norm": 4.443571747878234, + "learning_rate": 1.977966872011135e-05, + "loss": 1.514, + "step": 875 + }, + { + "epoch": 0.4755700325732899, + "grad_norm": 4.432653652742011, + "learning_rate": 1.9778933939444298e-05, + "loss": 1.2405, + "step": 876 + }, + { + "epoch": 0.4761129207383279, + "grad_norm": 3.408851551186198, + "learning_rate": 1.9778197949303107e-05, + "loss": 0.9597, + "step": 877 + }, + { + "epoch": 0.4766558089033659, + "grad_norm": 3.13258933410851, + "learning_rate": 1.9777460749778812e-05, + "loss": 0.8015, + "step": 878 + }, + { + "epoch": 0.4771986970684039, + "grad_norm": 3.535707436018844, + "learning_rate": 1.977672234096259e-05, + "loss": 1.3375, + "step": 879 + }, + { + "epoch": 0.4777415852334419, + "grad_norm": 5.210333639552661, + "learning_rate": 1.9775982722945764e-05, + "loss": 1.5039, + "step": 880 + }, + { + "epoch": 0.4782844733984799, + "grad_norm": 4.14115837125392, + "learning_rate": 1.9775241895819818e-05, + "loss": 1.0267, + "step": 881 + }, + { + "epoch": 0.4788273615635179, + "grad_norm": 4.390843049257208, + "learning_rate": 1.977449985967637e-05, + "loss": 1.6639, + "step": 882 + }, + { + "epoch": 0.4793702497285559, + "grad_norm": 3.5906251649150405, + "learning_rate": 1.9773756614607205e-05, + "loss": 1.2382, + "step": 883 + }, + { + "epoch": 0.4799131378935939, + "grad_norm": 4.78880370244169, + "learning_rate": 1.9773012160704242e-05, + "loss": 1.9436, + "step": 884 + }, + { + "epoch": 0.4804560260586319, + "grad_norm": 4.024679917575935, + "learning_rate": 1.977226649805956e-05, + "loss": 1.7126, + "step": 885 + }, + { + "epoch": 0.4809989142236699, + "grad_norm": 4.789782875875197, + "learning_rate": 1.9771519626765384e-05, + "loss": 1.4639, + "step": 886 + }, + { + "epoch": 0.4815418023887079, + "grad_norm": 4.188325588019175, + "learning_rate": 1.9770771546914088e-05, + "loss": 0.9698, + "step": 887 + }, + { + "epoch": 0.4820846905537459, + "grad_norm": 4.491436069698424, + "learning_rate": 1.9770022258598192e-05, + "loss": 1.5504, + "step": 888 + }, + { + "epoch": 0.4826275787187839, + "grad_norm": 4.889270563723842, + "learning_rate": 1.9769271761910373e-05, + "loss": 1.6052, + "step": 889 + }, + { + "epoch": 0.4831704668838219, + "grad_norm": 3.7615089990618493, + "learning_rate": 1.9768520056943454e-05, + "loss": 1.2445, + "step": 890 + }, + { + "epoch": 0.4837133550488599, + "grad_norm": 4.153654659347469, + "learning_rate": 1.9767767143790403e-05, + "loss": 1.4304, + "step": 891 + }, + { + "epoch": 0.48425624321389793, + "grad_norm": 4.040032964648755, + "learning_rate": 1.9767013022544346e-05, + "loss": 1.2906, + "step": 892 + }, + { + "epoch": 0.48479913137893593, + "grad_norm": 4.6559235863058275, + "learning_rate": 1.976625769329855e-05, + "loss": 1.5214, + "step": 893 + }, + { + "epoch": 0.48534201954397393, + "grad_norm": 4.224494142526306, + "learning_rate": 1.976550115614644e-05, + "loss": 1.2299, + "step": 894 + }, + { + "epoch": 0.48588490770901194, + "grad_norm": 5.225621473809677, + "learning_rate": 1.9764743411181585e-05, + "loss": 1.3489, + "step": 895 + }, + { + "epoch": 0.48642779587404994, + "grad_norm": 3.298440710545404, + "learning_rate": 1.97639844584977e-05, + "loss": 1.2107, + "step": 896 + }, + { + "epoch": 0.48697068403908794, + "grad_norm": 4.8373671778516485, + "learning_rate": 1.9763224298188652e-05, + "loss": 1.7748, + "step": 897 + }, + { + "epoch": 0.48751357220412594, + "grad_norm": 3.9332967347674224, + "learning_rate": 1.9762462930348465e-05, + "loss": 1.1036, + "step": 898 + }, + { + "epoch": 0.48805646036916395, + "grad_norm": 3.555513906222738, + "learning_rate": 1.9761700355071306e-05, + "loss": 1.5988, + "step": 899 + }, + { + "epoch": 0.48859934853420195, + "grad_norm": 4.115001448738943, + "learning_rate": 1.9760936572451487e-05, + "loss": 0.9789, + "step": 900 + }, + { + "epoch": 0.48914223669923995, + "grad_norm": 3.955269044937872, + "learning_rate": 1.9760171582583476e-05, + "loss": 1.3833, + "step": 901 + }, + { + "epoch": 0.48968512486427795, + "grad_norm": 2.718909991804504, + "learning_rate": 1.975940538556189e-05, + "loss": 1.1708, + "step": 902 + }, + { + "epoch": 0.49022801302931596, + "grad_norm": 4.234679090051759, + "learning_rate": 1.975863798148149e-05, + "loss": 1.3002, + "step": 903 + }, + { + "epoch": 0.49077090119435396, + "grad_norm": 4.162638017444828, + "learning_rate": 1.975786937043719e-05, + "loss": 1.246, + "step": 904 + }, + { + "epoch": 0.49131378935939196, + "grad_norm": 4.466027820294199, + "learning_rate": 1.975709955252406e-05, + "loss": 1.7345, + "step": 905 + }, + { + "epoch": 0.49185667752442996, + "grad_norm": 4.120456111509664, + "learning_rate": 1.97563285278373e-05, + "loss": 1.1339, + "step": 906 + }, + { + "epoch": 0.49239956568946797, + "grad_norm": 3.7932999801551346, + "learning_rate": 1.9755556296472286e-05, + "loss": 1.241, + "step": 907 + }, + { + "epoch": 0.49294245385450597, + "grad_norm": 3.3311865780622107, + "learning_rate": 1.9754782858524515e-05, + "loss": 1.2329, + "step": 908 + }, + { + "epoch": 0.49348534201954397, + "grad_norm": 4.21189840914184, + "learning_rate": 1.9754008214089654e-05, + "loss": 1.089, + "step": 909 + }, + { + "epoch": 0.494028230184582, + "grad_norm": 4.882048591403942, + "learning_rate": 1.9753232363263513e-05, + "loss": 1.3579, + "step": 910 + }, + { + "epoch": 0.49457111834962, + "grad_norm": 3.8495412115497833, + "learning_rate": 1.9752455306142052e-05, + "loss": 0.876, + "step": 911 + }, + { + "epoch": 0.495114006514658, + "grad_norm": 5.35378911928701, + "learning_rate": 1.975167704282137e-05, + "loss": 1.5025, + "step": 912 + }, + { + "epoch": 0.495656894679696, + "grad_norm": 3.4776107222093176, + "learning_rate": 1.9750897573397733e-05, + "loss": 1.1576, + "step": 913 + }, + { + "epoch": 0.496199782844734, + "grad_norm": 4.3680755226003525, + "learning_rate": 1.9750116897967545e-05, + "loss": 1.2519, + "step": 914 + }, + { + "epoch": 0.496742671009772, + "grad_norm": 4.154395835162648, + "learning_rate": 1.9749335016627355e-05, + "loss": 1.3397, + "step": 915 + }, + { + "epoch": 0.49728555917481, + "grad_norm": 5.465210053229008, + "learning_rate": 1.9748551929473873e-05, + "loss": 1.7555, + "step": 916 + }, + { + "epoch": 0.497828447339848, + "grad_norm": 4.391089608848125, + "learning_rate": 1.9747767636603953e-05, + "loss": 0.9685, + "step": 917 + }, + { + "epoch": 0.498371335504886, + "grad_norm": 4.215127655197256, + "learning_rate": 1.9746982138114597e-05, + "loss": 1.3889, + "step": 918 + }, + { + "epoch": 0.498914223669924, + "grad_norm": 4.559291789845839, + "learning_rate": 1.9746195434102956e-05, + "loss": 1.2411, + "step": 919 + }, + { + "epoch": 0.499457111834962, + "grad_norm": 4.226267506705079, + "learning_rate": 1.9745407524666327e-05, + "loss": 0.9517, + "step": 920 + }, + { + "epoch": 0.5, + "grad_norm": 4.289151336000736, + "learning_rate": 1.9744618409902164e-05, + "loss": 0.9123, + "step": 921 + }, + { + "epoch": 0.500542888165038, + "grad_norm": 5.356043714827742, + "learning_rate": 1.9743828089908067e-05, + "loss": 1.4039, + "step": 922 + }, + { + "epoch": 0.501085776330076, + "grad_norm": 4.867388642606108, + "learning_rate": 1.9743036564781785e-05, + "loss": 1.877, + "step": 923 + }, + { + "epoch": 0.501628664495114, + "grad_norm": 4.034273796102648, + "learning_rate": 1.9742243834621207e-05, + "loss": 1.6151, + "step": 924 + }, + { + "epoch": 0.502171552660152, + "grad_norm": 4.67131234305618, + "learning_rate": 1.9741449899524393e-05, + "loss": 1.6415, + "step": 925 + }, + { + "epoch": 0.50271444082519, + "grad_norm": 4.609211834325983, + "learning_rate": 1.9740654759589524e-05, + "loss": 1.2702, + "step": 926 + }, + { + "epoch": 0.503257328990228, + "grad_norm": 3.8693712873725064, + "learning_rate": 1.973985841491495e-05, + "loss": 1.0518, + "step": 927 + }, + { + "epoch": 0.503800217155266, + "grad_norm": 4.224951197105749, + "learning_rate": 1.9739060865599163e-05, + "loss": 1.1367, + "step": 928 + }, + { + "epoch": 0.504343105320304, + "grad_norm": 4.851656117293343, + "learning_rate": 1.973826211174081e-05, + "loss": 1.2753, + "step": 929 + }, + { + "epoch": 0.504885993485342, + "grad_norm": 4.255261482914816, + "learning_rate": 1.9737462153438675e-05, + "loss": 1.1723, + "step": 930 + }, + { + "epoch": 0.50542888165038, + "grad_norm": 3.8038614228711616, + "learning_rate": 1.9736660990791705e-05, + "loss": 1.1919, + "step": 931 + }, + { + "epoch": 0.505971769815418, + "grad_norm": 3.934227455395002, + "learning_rate": 1.9735858623898984e-05, + "loss": 1.1311, + "step": 932 + }, + { + "epoch": 0.506514657980456, + "grad_norm": 6.087081162565673, + "learning_rate": 1.9735055052859752e-05, + "loss": 1.4226, + "step": 933 + }, + { + "epoch": 0.507057546145494, + "grad_norm": 3.901680463400142, + "learning_rate": 1.9734250277773398e-05, + "loss": 0.8336, + "step": 934 + }, + { + "epoch": 0.507600434310532, + "grad_norm": 4.29804205692996, + "learning_rate": 1.9733444298739454e-05, + "loss": 1.5767, + "step": 935 + }, + { + "epoch": 0.50814332247557, + "grad_norm": 4.114842645062852, + "learning_rate": 1.9732637115857606e-05, + "loss": 1.3353, + "step": 936 + }, + { + "epoch": 0.508686210640608, + "grad_norm": 3.8525228530533706, + "learning_rate": 1.9731828729227686e-05, + "loss": 0.8208, + "step": 937 + }, + { + "epoch": 0.509229098805646, + "grad_norm": 4.592070280997817, + "learning_rate": 1.973101913894968e-05, + "loss": 1.3908, + "step": 938 + }, + { + "epoch": 0.509771986970684, + "grad_norm": 3.5842386277465628, + "learning_rate": 1.9730208345123718e-05, + "loss": 1.1996, + "step": 939 + }, + { + "epoch": 0.510314875135722, + "grad_norm": 6.053647281932304, + "learning_rate": 1.9729396347850082e-05, + "loss": 1.565, + "step": 940 + }, + { + "epoch": 0.51085776330076, + "grad_norm": 4.515664489508703, + "learning_rate": 1.9728583147229196e-05, + "loss": 1.4823, + "step": 941 + }, + { + "epoch": 0.511400651465798, + "grad_norm": 4.7752750595730475, + "learning_rate": 1.9727768743361644e-05, + "loss": 1.3867, + "step": 942 + }, + { + "epoch": 0.511943539630836, + "grad_norm": 3.983262894008015, + "learning_rate": 1.972695313634815e-05, + "loss": 1.1327, + "step": 943 + }, + { + "epoch": 0.512486427795874, + "grad_norm": 3.8510372454489326, + "learning_rate": 1.9726136326289586e-05, + "loss": 1.0233, + "step": 944 + }, + { + "epoch": 0.5130293159609121, + "grad_norm": 3.7881258264140634, + "learning_rate": 1.9725318313286985e-05, + "loss": 1.0239, + "step": 945 + }, + { + "epoch": 0.5135722041259501, + "grad_norm": 5.532602386318144, + "learning_rate": 1.9724499097441513e-05, + "loss": 1.8325, + "step": 946 + }, + { + "epoch": 0.5141150922909881, + "grad_norm": 5.978249809716352, + "learning_rate": 1.9723678678854493e-05, + "loss": 1.4899, + "step": 947 + }, + { + "epoch": 0.5146579804560261, + "grad_norm": 4.164217442125778, + "learning_rate": 1.9722857057627398e-05, + "loss": 1.5428, + "step": 948 + }, + { + "epoch": 0.5152008686210641, + "grad_norm": 3.315839086509451, + "learning_rate": 1.9722034233861848e-05, + "loss": 1.161, + "step": 949 + }, + { + "epoch": 0.5157437567861021, + "grad_norm": 4.362448119896709, + "learning_rate": 1.9721210207659608e-05, + "loss": 1.3402, + "step": 950 + }, + { + "epoch": 0.5162866449511401, + "grad_norm": 4.196652705893848, + "learning_rate": 1.9720384979122594e-05, + "loss": 1.4782, + "step": 951 + }, + { + "epoch": 0.5168295331161781, + "grad_norm": 4.093486911730782, + "learning_rate": 1.9719558548352876e-05, + "loss": 0.8508, + "step": 952 + }, + { + "epoch": 0.5173724212812161, + "grad_norm": 3.623412868803193, + "learning_rate": 1.9718730915452664e-05, + "loss": 1.4979, + "step": 953 + }, + { + "epoch": 0.5179153094462541, + "grad_norm": 3.4964911024452845, + "learning_rate": 1.9717902080524324e-05, + "loss": 1.0495, + "step": 954 + }, + { + "epoch": 0.5184581976112921, + "grad_norm": 4.021867363378223, + "learning_rate": 1.9717072043670367e-05, + "loss": 1.5284, + "step": 955 + }, + { + "epoch": 0.5190010857763301, + "grad_norm": 4.383918034859983, + "learning_rate": 1.9716240804993454e-05, + "loss": 1.3092, + "step": 956 + }, + { + "epoch": 0.5195439739413681, + "grad_norm": 4.61200677033352, + "learning_rate": 1.971540836459639e-05, + "loss": 1.8809, + "step": 957 + }, + { + "epoch": 0.5200868621064061, + "grad_norm": 4.490685725956694, + "learning_rate": 1.9714574722582142e-05, + "loss": 1.4004, + "step": 958 + }, + { + "epoch": 0.5206297502714441, + "grad_norm": 3.4814367857792736, + "learning_rate": 1.9713739879053802e-05, + "loss": 0.7918, + "step": 959 + }, + { + "epoch": 0.5211726384364821, + "grad_norm": 4.751597566480749, + "learning_rate": 1.9712903834114635e-05, + "loss": 1.6595, + "step": 960 + }, + { + "epoch": 0.5217155266015201, + "grad_norm": 4.535333205317335, + "learning_rate": 1.9712066587868042e-05, + "loss": 1.5676, + "step": 961 + }, + { + "epoch": 0.5222584147665581, + "grad_norm": 4.285001993552699, + "learning_rate": 1.9711228140417577e-05, + "loss": 1.0156, + "step": 962 + }, + { + "epoch": 0.5228013029315961, + "grad_norm": 4.61361421766516, + "learning_rate": 1.9710388491866934e-05, + "loss": 0.9552, + "step": 963 + }, + { + "epoch": 0.5233441910966341, + "grad_norm": 5.760482951437034, + "learning_rate": 1.9709547642319968e-05, + "loss": 2.2573, + "step": 964 + }, + { + "epoch": 0.5238870792616721, + "grad_norm": 4.38974180509626, + "learning_rate": 1.9708705591880674e-05, + "loss": 0.9939, + "step": 965 + }, + { + "epoch": 0.5244299674267101, + "grad_norm": 4.856793718321527, + "learning_rate": 1.97078623406532e-05, + "loss": 1.6161, + "step": 966 + }, + { + "epoch": 0.5249728555917481, + "grad_norm": 4.164767885626303, + "learning_rate": 1.9707017888741838e-05, + "loss": 1.4327, + "step": 967 + }, + { + "epoch": 0.5255157437567861, + "grad_norm": 5.163022804287371, + "learning_rate": 1.970617223625104e-05, + "loss": 1.2063, + "step": 968 + }, + { + "epoch": 0.5260586319218241, + "grad_norm": 4.826390195584734, + "learning_rate": 1.9705325383285384e-05, + "loss": 1.609, + "step": 969 + }, + { + "epoch": 0.5266015200868621, + "grad_norm": 6.3570394668292405, + "learning_rate": 1.9704477329949617e-05, + "loss": 1.5845, + "step": 970 + }, + { + "epoch": 0.5271444082519001, + "grad_norm": 5.176208926963624, + "learning_rate": 1.9703628076348628e-05, + "loss": 1.5792, + "step": 971 + }, + { + "epoch": 0.5276872964169381, + "grad_norm": 4.968725075311641, + "learning_rate": 1.9702777622587452e-05, + "loss": 1.774, + "step": 972 + }, + { + "epoch": 0.5282301845819761, + "grad_norm": 3.733580423042607, + "learning_rate": 1.9701925968771277e-05, + "loss": 0.8424, + "step": 973 + }, + { + "epoch": 0.5287730727470141, + "grad_norm": 4.330993358163065, + "learning_rate": 1.9701073115005437e-05, + "loss": 1.6198, + "step": 974 + }, + { + "epoch": 0.5293159609120521, + "grad_norm": 4.64032700965104, + "learning_rate": 1.9700219061395408e-05, + "loss": 1.2958, + "step": 975 + }, + { + "epoch": 0.5298588490770901, + "grad_norm": 4.08538215893539, + "learning_rate": 1.969936380804683e-05, + "loss": 1.5371, + "step": 976 + }, + { + "epoch": 0.5304017372421281, + "grad_norm": 3.7672102191405012, + "learning_rate": 1.9698507355065478e-05, + "loss": 1.4831, + "step": 977 + }, + { + "epoch": 0.5309446254071661, + "grad_norm": 4.237599031471779, + "learning_rate": 1.969764970255728e-05, + "loss": 1.4561, + "step": 978 + }, + { + "epoch": 0.5314875135722041, + "grad_norm": 4.296489590459447, + "learning_rate": 1.9696790850628308e-05, + "loss": 1.3813, + "step": 979 + }, + { + "epoch": 0.5320304017372421, + "grad_norm": 4.037479773042692, + "learning_rate": 1.9695930799384787e-05, + "loss": 1.103, + "step": 980 + }, + { + "epoch": 0.5325732899022801, + "grad_norm": 3.388491089813824, + "learning_rate": 1.9695069548933097e-05, + "loss": 1.0416, + "step": 981 + }, + { + "epoch": 0.5331161780673181, + "grad_norm": 3.6808593992848873, + "learning_rate": 1.969420709937975e-05, + "loss": 1.3158, + "step": 982 + }, + { + "epoch": 0.5336590662323561, + "grad_norm": 3.574434479685402, + "learning_rate": 1.969334345083142e-05, + "loss": 0.9542, + "step": 983 + }, + { + "epoch": 0.5342019543973942, + "grad_norm": 4.205886917864195, + "learning_rate": 1.9692478603394926e-05, + "loss": 1.1542, + "step": 984 + }, + { + "epoch": 0.5347448425624322, + "grad_norm": 3.89660120144171, + "learning_rate": 1.9691612557177225e-05, + "loss": 1.2374, + "step": 985 + }, + { + "epoch": 0.5352877307274702, + "grad_norm": 5.508260629607227, + "learning_rate": 1.9690745312285443e-05, + "loss": 2.0836, + "step": 986 + }, + { + "epoch": 0.5358306188925082, + "grad_norm": 5.482641917586033, + "learning_rate": 1.968987686882683e-05, + "loss": 1.5608, + "step": 987 + }, + { + "epoch": 0.5363735070575462, + "grad_norm": 5.4640893579957845, + "learning_rate": 1.9689007226908807e-05, + "loss": 1.3632, + "step": 988 + }, + { + "epoch": 0.5369163952225842, + "grad_norm": 4.163793092253212, + "learning_rate": 1.9688136386638926e-05, + "loss": 1.0509, + "step": 989 + }, + { + "epoch": 0.5374592833876222, + "grad_norm": 7.431185891169344, + "learning_rate": 1.96872643481249e-05, + "loss": 2.2147, + "step": 990 + }, + { + "epoch": 0.5380021715526602, + "grad_norm": 4.243102861037138, + "learning_rate": 1.9686391111474574e-05, + "loss": 0.9978, + "step": 991 + }, + { + "epoch": 0.5385450597176982, + "grad_norm": 4.250454886057412, + "learning_rate": 1.968551667679596e-05, + "loss": 1.5265, + "step": 992 + }, + { + "epoch": 0.5390879478827362, + "grad_norm": 5.989466882905416, + "learning_rate": 1.9684641044197207e-05, + "loss": 1.3344, + "step": 993 + }, + { + "epoch": 0.5396308360477742, + "grad_norm": 6.003524822087355, + "learning_rate": 1.9683764213786617e-05, + "loss": 2.0785, + "step": 994 + }, + { + "epoch": 0.5401737242128122, + "grad_norm": 4.535858561115482, + "learning_rate": 1.9682886185672633e-05, + "loss": 1.7654, + "step": 995 + }, + { + "epoch": 0.5407166123778502, + "grad_norm": 3.304287633610362, + "learning_rate": 1.9682006959963854e-05, + "loss": 0.9151, + "step": 996 + }, + { + "epoch": 0.5412595005428882, + "grad_norm": 3.6154194334954797, + "learning_rate": 1.9681126536769022e-05, + "loss": 1.1871, + "step": 997 + }, + { + "epoch": 0.5418023887079262, + "grad_norm": 5.596808777173772, + "learning_rate": 1.968024491619703e-05, + "loss": 1.1773, + "step": 998 + }, + { + "epoch": 0.5423452768729642, + "grad_norm": 4.486728043355676, + "learning_rate": 1.9679362098356923e-05, + "loss": 1.6495, + "step": 999 + }, + { + "epoch": 0.5428881650380022, + "grad_norm": 5.254525660355539, + "learning_rate": 1.9678478083357882e-05, + "loss": 1.7632, + "step": 1000 + }, + { + "epoch": 0.5434310532030402, + "grad_norm": 4.084222439189464, + "learning_rate": 1.9677592871309248e-05, + "loss": 1.4108, + "step": 1001 + }, + { + "epoch": 0.5439739413680782, + "grad_norm": 3.7647794378542248, + "learning_rate": 1.9676706462320504e-05, + "loss": 1.0893, + "step": 1002 + }, + { + "epoch": 0.5445168295331162, + "grad_norm": 3.658128786211824, + "learning_rate": 1.967581885650128e-05, + "loss": 1.0626, + "step": 1003 + }, + { + "epoch": 0.5450597176981542, + "grad_norm": 4.111525087755858, + "learning_rate": 1.967493005396136e-05, + "loss": 1.1915, + "step": 1004 + }, + { + "epoch": 0.5456026058631922, + "grad_norm": 4.130761699023899, + "learning_rate": 1.967404005481067e-05, + "loss": 1.1007, + "step": 1005 + }, + { + "epoch": 0.5461454940282302, + "grad_norm": 4.610164776763781, + "learning_rate": 1.9673148859159292e-05, + "loss": 1.3626, + "step": 1006 + }, + { + "epoch": 0.5466883821932682, + "grad_norm": 3.3516729554956686, + "learning_rate": 1.9672256467117445e-05, + "loss": 1.1138, + "step": 1007 + }, + { + "epoch": 0.5472312703583062, + "grad_norm": 5.185808250719794, + "learning_rate": 1.9671362878795502e-05, + "loss": 1.5639, + "step": 1008 + }, + { + "epoch": 0.5477741585233442, + "grad_norm": 3.9549787550673323, + "learning_rate": 1.9670468094303983e-05, + "loss": 1.0094, + "step": 1009 + }, + { + "epoch": 0.5483170466883822, + "grad_norm": 5.025417088271204, + "learning_rate": 1.966957211375356e-05, + "loss": 1.9741, + "step": 1010 + }, + { + "epoch": 0.5488599348534202, + "grad_norm": 5.1560673391175, + "learning_rate": 1.9668674937255044e-05, + "loss": 0.9653, + "step": 1011 + }, + { + "epoch": 0.5494028230184582, + "grad_norm": 4.843540796107876, + "learning_rate": 1.9667776564919404e-05, + "loss": 1.4745, + "step": 1012 + }, + { + "epoch": 0.5499457111834962, + "grad_norm": 3.8379573247832193, + "learning_rate": 1.966687699685775e-05, + "loss": 1.537, + "step": 1013 + }, + { + "epoch": 0.5504885993485342, + "grad_norm": 4.291804816024517, + "learning_rate": 1.9665976233181342e-05, + "loss": 1.6859, + "step": 1014 + }, + { + "epoch": 0.5510314875135722, + "grad_norm": 4.827630960232388, + "learning_rate": 1.966507427400159e-05, + "loss": 1.3205, + "step": 1015 + }, + { + "epoch": 0.5515743756786102, + "grad_norm": 4.251540690047392, + "learning_rate": 1.9664171119430044e-05, + "loss": 0.8213, + "step": 1016 + }, + { + "epoch": 0.5521172638436482, + "grad_norm": 3.7693342275615476, + "learning_rate": 1.9663266769578414e-05, + "loss": 1.0349, + "step": 1017 + }, + { + "epoch": 0.5526601520086862, + "grad_norm": 4.075472936638236, + "learning_rate": 1.966236122455855e-05, + "loss": 1.0853, + "step": 1018 + }, + { + "epoch": 0.5532030401737242, + "grad_norm": 4.771167333111926, + "learning_rate": 1.9661454484482448e-05, + "loss": 1.2135, + "step": 1019 + }, + { + "epoch": 0.5537459283387622, + "grad_norm": 3.834854303591261, + "learning_rate": 1.9660546549462262e-05, + "loss": 1.0972, + "step": 1020 + }, + { + "epoch": 0.5542888165038002, + "grad_norm": 4.712560042125338, + "learning_rate": 1.9659637419610278e-05, + "loss": 0.8779, + "step": 1021 + }, + { + "epoch": 0.5548317046688382, + "grad_norm": 3.8174143787104566, + "learning_rate": 1.9658727095038942e-05, + "loss": 0.7395, + "step": 1022 + }, + { + "epoch": 0.5553745928338762, + "grad_norm": 5.970646625648977, + "learning_rate": 1.965781557586085e-05, + "loss": 1.4233, + "step": 1023 + }, + { + "epoch": 0.5559174809989142, + "grad_norm": 3.872798187838392, + "learning_rate": 1.9656902862188732e-05, + "loss": 0.8147, + "step": 1024 + }, + { + "epoch": 0.5564603691639523, + "grad_norm": 5.632609234869673, + "learning_rate": 1.9655988954135473e-05, + "loss": 1.1581, + "step": 1025 + }, + { + "epoch": 0.5570032573289903, + "grad_norm": 4.141188191911578, + "learning_rate": 1.9655073851814117e-05, + "loss": 1.1707, + "step": 1026 + }, + { + "epoch": 0.5575461454940283, + "grad_norm": 5.7835323867879564, + "learning_rate": 1.9654157555337837e-05, + "loss": 1.2383, + "step": 1027 + }, + { + "epoch": 0.5580890336590663, + "grad_norm": 6.410603262158535, + "learning_rate": 1.9653240064819965e-05, + "loss": 2.3461, + "step": 1028 + }, + { + "epoch": 0.5586319218241043, + "grad_norm": 4.661944777700264, + "learning_rate": 1.9652321380373974e-05, + "loss": 1.1718, + "step": 1029 + }, + { + "epoch": 0.5591748099891423, + "grad_norm": 4.224334614804915, + "learning_rate": 1.9651401502113497e-05, + "loss": 1.0937, + "step": 1030 + }, + { + "epoch": 0.5597176981541803, + "grad_norm": 4.292561821167698, + "learning_rate": 1.9650480430152295e-05, + "loss": 0.8306, + "step": 1031 + }, + { + "epoch": 0.5602605863192183, + "grad_norm": 4.515070808084738, + "learning_rate": 1.9649558164604293e-05, + "loss": 1.4042, + "step": 1032 + }, + { + "epoch": 0.5608034744842563, + "grad_norm": 4.906478484057087, + "learning_rate": 1.964863470558356e-05, + "loss": 1.3396, + "step": 1033 + }, + { + "epoch": 0.5613463626492943, + "grad_norm": 3.9530548146664306, + "learning_rate": 1.9647710053204307e-05, + "loss": 0.917, + "step": 1034 + }, + { + "epoch": 0.5618892508143323, + "grad_norm": 4.640160859146004, + "learning_rate": 1.96467842075809e-05, + "loss": 1.6625, + "step": 1035 + }, + { + "epoch": 0.5624321389793703, + "grad_norm": 4.921887921875984, + "learning_rate": 1.964585716882785e-05, + "loss": 1.5133, + "step": 1036 + }, + { + "epoch": 0.5629750271444083, + "grad_norm": 4.667679101301949, + "learning_rate": 1.964492893705981e-05, + "loss": 0.9713, + "step": 1037 + }, + { + "epoch": 0.5635179153094463, + "grad_norm": 6.0457527064758825, + "learning_rate": 1.9643999512391586e-05, + "loss": 1.1101, + "step": 1038 + }, + { + "epoch": 0.5640608034744843, + "grad_norm": 5.188403972218186, + "learning_rate": 1.964306889493813e-05, + "loss": 1.8276, + "step": 1039 + }, + { + "epoch": 0.5646036916395223, + "grad_norm": 4.449370010668479, + "learning_rate": 1.9642137084814548e-05, + "loss": 1.0797, + "step": 1040 + }, + { + "epoch": 0.5651465798045603, + "grad_norm": 5.97660491495462, + "learning_rate": 1.9641204082136085e-05, + "loss": 1.913, + "step": 1041 + }, + { + "epoch": 0.5656894679695983, + "grad_norm": 5.611282387277076, + "learning_rate": 1.9640269887018135e-05, + "loss": 2.4593, + "step": 1042 + }, + { + "epoch": 0.5662323561346363, + "grad_norm": 3.6226559029227996, + "learning_rate": 1.9639334499576237e-05, + "loss": 0.7939, + "step": 1043 + }, + { + "epoch": 0.5667752442996743, + "grad_norm": 4.701569628633213, + "learning_rate": 1.963839791992609e-05, + "loss": 1.4092, + "step": 1044 + }, + { + "epoch": 0.5673181324647123, + "grad_norm": 5.185222096624812, + "learning_rate": 1.9637460148183525e-05, + "loss": 1.3851, + "step": 1045 + }, + { + "epoch": 0.5678610206297503, + "grad_norm": 4.094772946058271, + "learning_rate": 1.963652118446453e-05, + "loss": 1.1872, + "step": 1046 + }, + { + "epoch": 0.5684039087947883, + "grad_norm": 4.653748340189311, + "learning_rate": 1.9635581028885233e-05, + "loss": 1.1546, + "step": 1047 + }, + { + "epoch": 0.5689467969598263, + "grad_norm": 3.9426831069951844, + "learning_rate": 1.9634639681561924e-05, + "loss": 0.9324, + "step": 1048 + }, + { + "epoch": 0.5694896851248643, + "grad_norm": 4.949304407204856, + "learning_rate": 1.9633697142611017e-05, + "loss": 1.2372, + "step": 1049 + }, + { + "epoch": 0.5700325732899023, + "grad_norm": 5.364689744345368, + "learning_rate": 1.9632753412149096e-05, + "loss": 1.5104, + "step": 1050 + }, + { + "epoch": 0.5705754614549403, + "grad_norm": 5.307969223857254, + "learning_rate": 1.9631808490292884e-05, + "loss": 1.212, + "step": 1051 + }, + { + "epoch": 0.5711183496199783, + "grad_norm": 5.197310781163775, + "learning_rate": 1.963086237715924e-05, + "loss": 1.0175, + "step": 1052 + }, + { + "epoch": 0.5716612377850163, + "grad_norm": 4.095170540464123, + "learning_rate": 1.9629915072865194e-05, + "loss": 1.2711, + "step": 1053 + }, + { + "epoch": 0.5722041259500543, + "grad_norm": 4.816474093418903, + "learning_rate": 1.9628966577527902e-05, + "loss": 0.7274, + "step": 1054 + }, + { + "epoch": 0.5727470141150923, + "grad_norm": 4.000604469041167, + "learning_rate": 1.962801689126468e-05, + "loss": 1.2913, + "step": 1055 + }, + { + "epoch": 0.5732899022801303, + "grad_norm": 4.31668064977018, + "learning_rate": 1.962706601419298e-05, + "loss": 1.2598, + "step": 1056 + }, + { + "epoch": 0.5738327904451683, + "grad_norm": 5.033983318277113, + "learning_rate": 1.9626113946430414e-05, + "loss": 1.5149, + "step": 1057 + }, + { + "epoch": 0.5743756786102063, + "grad_norm": 4.51738649586986, + "learning_rate": 1.9625160688094733e-05, + "loss": 1.2389, + "step": 1058 + }, + { + "epoch": 0.5749185667752443, + "grad_norm": 3.6899120907028853, + "learning_rate": 1.9624206239303837e-05, + "loss": 1.1632, + "step": 1059 + }, + { + "epoch": 0.5754614549402823, + "grad_norm": 4.423539955787714, + "learning_rate": 1.9623250600175775e-05, + "loss": 0.9526, + "step": 1060 + }, + { + "epoch": 0.5760043431053203, + "grad_norm": 5.062986942342047, + "learning_rate": 1.962229377082874e-05, + "loss": 1.6768, + "step": 1061 + }, + { + "epoch": 0.5765472312703583, + "grad_norm": 5.044415439938055, + "learning_rate": 1.962133575138108e-05, + "loss": 1.3808, + "step": 1062 + }, + { + "epoch": 0.5770901194353963, + "grad_norm": 3.1405283495786587, + "learning_rate": 1.962037654195128e-05, + "loss": 0.7588, + "step": 1063 + }, + { + "epoch": 0.5776330076004343, + "grad_norm": 5.449873922495929, + "learning_rate": 1.9619416142657974e-05, + "loss": 1.1544, + "step": 1064 + }, + { + "epoch": 0.5781758957654723, + "grad_norm": 4.268527750000251, + "learning_rate": 1.961845455361995e-05, + "loss": 1.1941, + "step": 1065 + }, + { + "epoch": 0.5787187839305103, + "grad_norm": 4.6641878464575095, + "learning_rate": 1.9617491774956137e-05, + "loss": 1.7162, + "step": 1066 + }, + { + "epoch": 0.5792616720955484, + "grad_norm": 4.583539478633983, + "learning_rate": 1.9616527806785612e-05, + "loss": 1.1415, + "step": 1067 + }, + { + "epoch": 0.5798045602605864, + "grad_norm": 4.325093393868848, + "learning_rate": 1.96155626492276e-05, + "loss": 1.3284, + "step": 1068 + }, + { + "epoch": 0.5803474484256244, + "grad_norm": 4.180109104666535, + "learning_rate": 1.9614596302401478e-05, + "loss": 1.136, + "step": 1069 + }, + { + "epoch": 0.5808903365906624, + "grad_norm": 4.108700840520527, + "learning_rate": 1.9613628766426762e-05, + "loss": 1.6142, + "step": 1070 + }, + { + "epoch": 0.5814332247557004, + "grad_norm": 5.1931558422083555, + "learning_rate": 1.9612660041423116e-05, + "loss": 0.8479, + "step": 1071 + }, + { + "epoch": 0.5819761129207384, + "grad_norm": 6.008601497593343, + "learning_rate": 1.9611690127510358e-05, + "loss": 1.4743, + "step": 1072 + }, + { + "epoch": 0.5825190010857764, + "grad_norm": 4.531622764164067, + "learning_rate": 1.9610719024808444e-05, + "loss": 1.5755, + "step": 1073 + }, + { + "epoch": 0.5830618892508144, + "grad_norm": 6.18647874774261, + "learning_rate": 1.9609746733437486e-05, + "loss": 1.7369, + "step": 1074 + }, + { + "epoch": 0.5836047774158524, + "grad_norm": 5.5880249590216025, + "learning_rate": 1.9608773253517738e-05, + "loss": 1.4402, + "step": 1075 + }, + { + "epoch": 0.5841476655808904, + "grad_norm": 4.798692544278573, + "learning_rate": 1.9607798585169595e-05, + "loss": 1.2257, + "step": 1076 + }, + { + "epoch": 0.5846905537459284, + "grad_norm": 5.635786862796798, + "learning_rate": 1.960682272851361e-05, + "loss": 1.183, + "step": 1077 + }, + { + "epoch": 0.5852334419109664, + "grad_norm": 3.3979344506068037, + "learning_rate": 1.9605845683670484e-05, + "loss": 0.6463, + "step": 1078 + }, + { + "epoch": 0.5857763300760044, + "grad_norm": 4.5152680576801405, + "learning_rate": 1.960486745076105e-05, + "loss": 0.8476, + "step": 1079 + }, + { + "epoch": 0.5863192182410424, + "grad_norm": 4.4606140059612835, + "learning_rate": 1.9603888029906305e-05, + "loss": 1.185, + "step": 1080 + }, + { + "epoch": 0.5868621064060804, + "grad_norm": 4.229219175605004, + "learning_rate": 1.960290742122738e-05, + "loss": 1.3672, + "step": 1081 + }, + { + "epoch": 0.5874049945711184, + "grad_norm": 4.964172517040214, + "learning_rate": 1.960192562484556e-05, + "loss": 1.0561, + "step": 1082 + }, + { + "epoch": 0.5879478827361564, + "grad_norm": 4.8348336861653785, + "learning_rate": 1.960094264088228e-05, + "loss": 1.1005, + "step": 1083 + }, + { + "epoch": 0.5884907709011944, + "grad_norm": 4.3153540955607514, + "learning_rate": 1.959995846945911e-05, + "loss": 1.6513, + "step": 1084 + }, + { + "epoch": 0.5890336590662324, + "grad_norm": 4.733352525701168, + "learning_rate": 1.9598973110697773e-05, + "loss": 1.6162, + "step": 1085 + }, + { + "epoch": 0.5895765472312704, + "grad_norm": 5.402815147148315, + "learning_rate": 1.959798656472015e-05, + "loss": 1.4178, + "step": 1086 + }, + { + "epoch": 0.5901194353963084, + "grad_norm": 3.8276911935691147, + "learning_rate": 1.9596998831648247e-05, + "loss": 1.1554, + "step": 1087 + }, + { + "epoch": 0.5906623235613464, + "grad_norm": 4.499779496087119, + "learning_rate": 1.9596009911604232e-05, + "loss": 0.9786, + "step": 1088 + }, + { + "epoch": 0.5912052117263844, + "grad_norm": 4.3705352636049595, + "learning_rate": 1.959501980471042e-05, + "loss": 1.0827, + "step": 1089 + }, + { + "epoch": 0.5917480998914224, + "grad_norm": 4.9770508472387025, + "learning_rate": 1.9594028511089264e-05, + "loss": 1.2512, + "step": 1090 + }, + { + "epoch": 0.5922909880564604, + "grad_norm": 4.23068807802663, + "learning_rate": 1.9593036030863376e-05, + "loss": 0.7528, + "step": 1091 + }, + { + "epoch": 0.5928338762214984, + "grad_norm": 4.321323716192747, + "learning_rate": 1.9592042364155496e-05, + "loss": 1.0773, + "step": 1092 + }, + { + "epoch": 0.5933767643865364, + "grad_norm": 3.380036112638087, + "learning_rate": 1.9591047511088535e-05, + "loss": 0.7238, + "step": 1093 + }, + { + "epoch": 0.5939196525515744, + "grad_norm": 3.9722637658815447, + "learning_rate": 1.959005147178553e-05, + "loss": 0.8132, + "step": 1094 + }, + { + "epoch": 0.5944625407166124, + "grad_norm": 4.354488865716506, + "learning_rate": 1.9589054246369673e-05, + "loss": 1.1214, + "step": 1095 + }, + { + "epoch": 0.5950054288816504, + "grad_norm": 4.337355429292092, + "learning_rate": 1.9588055834964307e-05, + "loss": 0.9039, + "step": 1096 + }, + { + "epoch": 0.5955483170466884, + "grad_norm": 5.864047590425705, + "learning_rate": 1.9587056237692912e-05, + "loss": 1.6648, + "step": 1097 + }, + { + "epoch": 0.5960912052117264, + "grad_norm": 4.904173131706714, + "learning_rate": 1.9586055454679123e-05, + "loss": 1.1465, + "step": 1098 + }, + { + "epoch": 0.5966340933767644, + "grad_norm": 4.411877930208604, + "learning_rate": 1.9585053486046717e-05, + "loss": 1.4579, + "step": 1099 + }, + { + "epoch": 0.5971769815418024, + "grad_norm": 4.985171448881133, + "learning_rate": 1.9584050331919616e-05, + "loss": 1.5497, + "step": 1100 + }, + { + "epoch": 0.5977198697068404, + "grad_norm": 4.950532301400581, + "learning_rate": 1.9583045992421902e-05, + "loss": 1.3811, + "step": 1101 + }, + { + "epoch": 0.5982627578718784, + "grad_norm": 4.154119544907876, + "learning_rate": 1.9582040467677782e-05, + "loss": 1.186, + "step": 1102 + }, + { + "epoch": 0.5988056460369164, + "grad_norm": 4.069018604381756, + "learning_rate": 1.9581033757811628e-05, + "loss": 0.9312, + "step": 1103 + }, + { + "epoch": 0.5993485342019544, + "grad_norm": 5.33070083129983, + "learning_rate": 1.9580025862947948e-05, + "loss": 1.4274, + "step": 1104 + }, + { + "epoch": 0.5998914223669924, + "grad_norm": 4.855606632567906, + "learning_rate": 1.95790167832114e-05, + "loss": 1.0875, + "step": 1105 + }, + { + "epoch": 0.6004343105320304, + "grad_norm": 5.30182267546013, + "learning_rate": 1.957800651872679e-05, + "loss": 1.6501, + "step": 1106 + }, + { + "epoch": 0.6009771986970684, + "grad_norm": 4.433903391255917, + "learning_rate": 1.957699506961907e-05, + "loss": 1.3897, + "step": 1107 + }, + { + "epoch": 0.6015200868621065, + "grad_norm": 4.687653782723915, + "learning_rate": 1.9575982436013335e-05, + "loss": 0.8484, + "step": 1108 + }, + { + "epoch": 0.6020629750271445, + "grad_norm": 5.650426140105992, + "learning_rate": 1.957496861803483e-05, + "loss": 1.923, + "step": 1109 + }, + { + "epoch": 0.6026058631921825, + "grad_norm": 4.024650772512973, + "learning_rate": 1.957395361580895e-05, + "loss": 1.2713, + "step": 1110 + }, + { + "epoch": 0.6031487513572205, + "grad_norm": 4.343921133786845, + "learning_rate": 1.9572937429461223e-05, + "loss": 1.3152, + "step": 1111 + }, + { + "epoch": 0.6036916395222585, + "grad_norm": 4.1467176317930265, + "learning_rate": 1.957192005911734e-05, + "loss": 1.3325, + "step": 1112 + }, + { + "epoch": 0.6042345276872965, + "grad_norm": 4.305465221711973, + "learning_rate": 1.9570901504903128e-05, + "loss": 0.9237, + "step": 1113 + }, + { + "epoch": 0.6047774158523345, + "grad_norm": 4.101045495900469, + "learning_rate": 1.9569881766944564e-05, + "loss": 1.3932, + "step": 1114 + }, + { + "epoch": 0.6053203040173725, + "grad_norm": 4.0920748344339835, + "learning_rate": 1.956886084536777e-05, + "loss": 1.0067, + "step": 1115 + }, + { + "epoch": 0.6058631921824105, + "grad_norm": 4.844431281019285, + "learning_rate": 1.956783874029902e-05, + "loss": 1.2726, + "step": 1116 + }, + { + "epoch": 0.6064060803474485, + "grad_norm": 4.329792546128024, + "learning_rate": 1.9566815451864723e-05, + "loss": 1.1144, + "step": 1117 + }, + { + "epoch": 0.6069489685124865, + "grad_norm": 5.430322578642178, + "learning_rate": 1.9565790980191447e-05, + "loss": 1.3307, + "step": 1118 + }, + { + "epoch": 0.6074918566775245, + "grad_norm": 3.672369600344093, + "learning_rate": 1.9564765325405895e-05, + "loss": 0.8514, + "step": 1119 + }, + { + "epoch": 0.6080347448425625, + "grad_norm": 3.6433282223345334, + "learning_rate": 1.9563738487634924e-05, + "loss": 1.1972, + "step": 1120 + }, + { + "epoch": 0.6085776330076005, + "grad_norm": 4.23249445053715, + "learning_rate": 1.956271046700553e-05, + "loss": 1.0638, + "step": 1121 + }, + { + "epoch": 0.6091205211726385, + "grad_norm": 3.9162807435580134, + "learning_rate": 1.956168126364487e-05, + "loss": 1.121, + "step": 1122 + }, + { + "epoch": 0.6096634093376765, + "grad_norm": 3.3651337591577812, + "learning_rate": 1.956065087768023e-05, + "loss": 0.9604, + "step": 1123 + }, + { + "epoch": 0.6102062975027145, + "grad_norm": 5.5905307606778685, + "learning_rate": 1.955961930923905e-05, + "loss": 1.4509, + "step": 1124 + }, + { + "epoch": 0.6107491856677525, + "grad_norm": 3.6120233107886404, + "learning_rate": 1.955858655844892e-05, + "loss": 1.0106, + "step": 1125 + }, + { + "epoch": 0.6112920738327905, + "grad_norm": 3.6242967727445468, + "learning_rate": 1.9557552625437574e-05, + "loss": 1.0768, + "step": 1126 + }, + { + "epoch": 0.6118349619978285, + "grad_norm": 4.856763858492221, + "learning_rate": 1.9556517510332883e-05, + "loss": 1.6372, + "step": 1127 + }, + { + "epoch": 0.6123778501628665, + "grad_norm": 4.648427809898363, + "learning_rate": 1.9555481213262873e-05, + "loss": 1.1622, + "step": 1128 + }, + { + "epoch": 0.6129207383279045, + "grad_norm": 4.228089396781667, + "learning_rate": 1.9554443734355723e-05, + "loss": 1.067, + "step": 1129 + }, + { + "epoch": 0.6134636264929425, + "grad_norm": 4.069786144676624, + "learning_rate": 1.9553405073739743e-05, + "loss": 1.1427, + "step": 1130 + }, + { + "epoch": 0.6140065146579805, + "grad_norm": 3.112182391585594, + "learning_rate": 1.9552365231543395e-05, + "loss": 0.7862, + "step": 1131 + }, + { + "epoch": 0.6145494028230185, + "grad_norm": 4.783518734299315, + "learning_rate": 1.955132420789529e-05, + "loss": 1.0366, + "step": 1132 + }, + { + "epoch": 0.6150922909880565, + "grad_norm": 4.905128048667836, + "learning_rate": 1.9550282002924187e-05, + "loss": 1.2566, + "step": 1133 + }, + { + "epoch": 0.6156351791530945, + "grad_norm": 4.643678450407587, + "learning_rate": 1.9549238616758987e-05, + "loss": 1.2749, + "step": 1134 + }, + { + "epoch": 0.6161780673181325, + "grad_norm": 4.18363664236292, + "learning_rate": 1.954819404952873e-05, + "loss": 1.1523, + "step": 1135 + }, + { + "epoch": 0.6167209554831705, + "grad_norm": 3.9776065520060335, + "learning_rate": 1.9547148301362623e-05, + "loss": 0.8672, + "step": 1136 + }, + { + "epoch": 0.6172638436482085, + "grad_norm": 5.417170242762362, + "learning_rate": 1.9546101372389994e-05, + "loss": 1.6434, + "step": 1137 + }, + { + "epoch": 0.6178067318132465, + "grad_norm": 3.3325133642660068, + "learning_rate": 1.9545053262740335e-05, + "loss": 1.1447, + "step": 1138 + }, + { + "epoch": 0.6183496199782845, + "grad_norm": 5.4727757064949545, + "learning_rate": 1.9544003972543273e-05, + "loss": 1.5013, + "step": 1139 + }, + { + "epoch": 0.6188925081433225, + "grad_norm": 4.619714236951128, + "learning_rate": 1.954295350192859e-05, + "loss": 0.9758, + "step": 1140 + }, + { + "epoch": 0.6194353963083605, + "grad_norm": 5.189787937143953, + "learning_rate": 1.954190185102621e-05, + "loss": 1.2771, + "step": 1141 + }, + { + "epoch": 0.6199782844733985, + "grad_norm": 6.571114312185869, + "learning_rate": 1.9540849019966198e-05, + "loss": 1.2805, + "step": 1142 + }, + { + "epoch": 0.6205211726384365, + "grad_norm": 5.6343515388196455, + "learning_rate": 1.9539795008878774e-05, + "loss": 1.4061, + "step": 1143 + }, + { + "epoch": 0.6210640608034745, + "grad_norm": 5.582702517092888, + "learning_rate": 1.9538739817894302e-05, + "loss": 1.613, + "step": 1144 + }, + { + "epoch": 0.6216069489685125, + "grad_norm": 6.194616996988272, + "learning_rate": 1.9537683447143287e-05, + "loss": 1.114, + "step": 1145 + }, + { + "epoch": 0.6221498371335505, + "grad_norm": 6.0395907363668515, + "learning_rate": 1.9536625896756377e-05, + "loss": 1.5233, + "step": 1146 + }, + { + "epoch": 0.6226927252985885, + "grad_norm": 4.672511820840566, + "learning_rate": 1.9535567166864382e-05, + "loss": 1.1491, + "step": 1147 + }, + { + "epoch": 0.6232356134636265, + "grad_norm": 4.954133576744497, + "learning_rate": 1.9534507257598244e-05, + "loss": 1.1974, + "step": 1148 + }, + { + "epoch": 0.6237785016286646, + "grad_norm": 4.447205012745297, + "learning_rate": 1.953344616908905e-05, + "loss": 1.019, + "step": 1149 + }, + { + "epoch": 0.6243213897937026, + "grad_norm": 4.284087771325503, + "learning_rate": 1.9532383901468038e-05, + "loss": 1.0814, + "step": 1150 + }, + { + "epoch": 0.6248642779587406, + "grad_norm": 3.576067836847051, + "learning_rate": 1.9531320454866595e-05, + "loss": 1.1025, + "step": 1151 + }, + { + "epoch": 0.6254071661237784, + "grad_norm": 3.6831292979487777, + "learning_rate": 1.9530255829416246e-05, + "loss": 0.967, + "step": 1152 + }, + { + "epoch": 0.6259500542888164, + "grad_norm": 3.842496728015195, + "learning_rate": 1.952919002524867e-05, + "loss": 1.0903, + "step": 1153 + }, + { + "epoch": 0.6264929424538545, + "grad_norm": 4.077633480381933, + "learning_rate": 1.952812304249568e-05, + "loss": 0.9227, + "step": 1154 + }, + { + "epoch": 0.6270358306188925, + "grad_norm": 4.157004955121598, + "learning_rate": 1.952705488128925e-05, + "loss": 0.9427, + "step": 1155 + }, + { + "epoch": 0.6275787187839305, + "grad_norm": 5.267614104282471, + "learning_rate": 1.952598554176149e-05, + "loss": 1.9875, + "step": 1156 + }, + { + "epoch": 0.6281216069489685, + "grad_norm": 3.8938350314376735, + "learning_rate": 1.952491502404465e-05, + "loss": 1.1662, + "step": 1157 + }, + { + "epoch": 0.6286644951140065, + "grad_norm": 4.851409026807682, + "learning_rate": 1.9523843328271144e-05, + "loss": 1.9559, + "step": 1158 + }, + { + "epoch": 0.6292073832790445, + "grad_norm": 4.4104613077358845, + "learning_rate": 1.9522770454573513e-05, + "loss": 0.8011, + "step": 1159 + }, + { + "epoch": 0.6297502714440825, + "grad_norm": 3.6883335641947856, + "learning_rate": 1.952169640308446e-05, + "loss": 1.1838, + "step": 1160 + }, + { + "epoch": 0.6302931596091205, + "grad_norm": 3.728203039459734, + "learning_rate": 1.9520621173936818e-05, + "loss": 1.1423, + "step": 1161 + }, + { + "epoch": 0.6308360477741585, + "grad_norm": 3.6272612276564296, + "learning_rate": 1.9519544767263574e-05, + "loss": 0.8376, + "step": 1162 + }, + { + "epoch": 0.6313789359391965, + "grad_norm": 3.928276333701112, + "learning_rate": 1.951846718319786e-05, + "loss": 1.0749, + "step": 1163 + }, + { + "epoch": 0.6319218241042345, + "grad_norm": 5.941949883397345, + "learning_rate": 1.951738842187296e-05, + "loss": 1.8535, + "step": 1164 + }, + { + "epoch": 0.6324647122692725, + "grad_norm": 3.4741200080360644, + "learning_rate": 1.951630848342229e-05, + "loss": 0.7537, + "step": 1165 + }, + { + "epoch": 0.6330076004343105, + "grad_norm": 5.392480994365815, + "learning_rate": 1.9515227367979416e-05, + "loss": 0.8776, + "step": 1166 + }, + { + "epoch": 0.6335504885993485, + "grad_norm": 5.203169892005258, + "learning_rate": 1.951414507567806e-05, + "loss": 1.2678, + "step": 1167 + }, + { + "epoch": 0.6340933767643865, + "grad_norm": 7.499226723318992, + "learning_rate": 1.9513061606652076e-05, + "loss": 1.3295, + "step": 1168 + }, + { + "epoch": 0.6346362649294245, + "grad_norm": 4.355172848230307, + "learning_rate": 1.9511976961035474e-05, + "loss": 0.9188, + "step": 1169 + }, + { + "epoch": 0.6351791530944625, + "grad_norm": 4.983220943925928, + "learning_rate": 1.9510891138962398e-05, + "loss": 1.1951, + "step": 1170 + }, + { + "epoch": 0.6357220412595005, + "grad_norm": 5.3098934312394395, + "learning_rate": 1.950980414056715e-05, + "loss": 0.855, + "step": 1171 + }, + { + "epoch": 0.6362649294245385, + "grad_norm": 5.125683116898323, + "learning_rate": 1.950871596598417e-05, + "loss": 1.1371, + "step": 1172 + }, + { + "epoch": 0.6368078175895765, + "grad_norm": 4.3010814264850215, + "learning_rate": 1.950762661534804e-05, + "loss": 1.7401, + "step": 1173 + }, + { + "epoch": 0.6373507057546145, + "grad_norm": 5.860625964437982, + "learning_rate": 1.95065360887935e-05, + "loss": 1.3557, + "step": 1174 + }, + { + "epoch": 0.6378935939196525, + "grad_norm": 6.3356570754666155, + "learning_rate": 1.9505444386455426e-05, + "loss": 1.2281, + "step": 1175 + }, + { + "epoch": 0.6384364820846905, + "grad_norm": 4.788366086551355, + "learning_rate": 1.9504351508468842e-05, + "loss": 1.3458, + "step": 1176 + }, + { + "epoch": 0.6389793702497285, + "grad_norm": 4.5553889359557065, + "learning_rate": 1.9503257454968914e-05, + "loss": 0.803, + "step": 1177 + }, + { + "epoch": 0.6395222584147665, + "grad_norm": 4.9063187966995345, + "learning_rate": 1.950216222609096e-05, + "loss": 0.8306, + "step": 1178 + }, + { + "epoch": 0.6400651465798045, + "grad_norm": 4.775138114897886, + "learning_rate": 1.9501065821970435e-05, + "loss": 0.9783, + "step": 1179 + }, + { + "epoch": 0.6406080347448425, + "grad_norm": 4.53672284240203, + "learning_rate": 1.9499968242742948e-05, + "loss": 1.3076, + "step": 1180 + }, + { + "epoch": 0.6411509229098805, + "grad_norm": 4.539860050199487, + "learning_rate": 1.949886948854425e-05, + "loss": 1.4212, + "step": 1181 + }, + { + "epoch": 0.6416938110749185, + "grad_norm": 4.251462432005476, + "learning_rate": 1.9497769559510232e-05, + "loss": 1.0274, + "step": 1182 + }, + { + "epoch": 0.6422366992399565, + "grad_norm": 4.49791289564495, + "learning_rate": 1.9496668455776938e-05, + "loss": 1.7531, + "step": 1183 + }, + { + "epoch": 0.6427795874049945, + "grad_norm": 4.996740483321485, + "learning_rate": 1.9495566177480555e-05, + "loss": 1.201, + "step": 1184 + }, + { + "epoch": 0.6433224755700325, + "grad_norm": 5.347352236818569, + "learning_rate": 1.9494462724757413e-05, + "loss": 1.5723, + "step": 1185 + }, + { + "epoch": 0.6438653637350705, + "grad_norm": 4.2242558589374015, + "learning_rate": 1.9493358097743988e-05, + "loss": 1.8014, + "step": 1186 + }, + { + "epoch": 0.6444082519001085, + "grad_norm": 5.042556201755935, + "learning_rate": 1.9492252296576906e-05, + "loss": 1.5466, + "step": 1187 + }, + { + "epoch": 0.6449511400651465, + "grad_norm": 4.284314238035697, + "learning_rate": 1.949114532139293e-05, + "loss": 0.9515, + "step": 1188 + }, + { + "epoch": 0.6454940282301845, + "grad_norm": 4.699826119048575, + "learning_rate": 1.9490037172328974e-05, + "loss": 1.1677, + "step": 1189 + }, + { + "epoch": 0.6460369163952225, + "grad_norm": 6.035247135554916, + "learning_rate": 1.9488927849522095e-05, + "loss": 1.957, + "step": 1190 + }, + { + "epoch": 0.6465798045602605, + "grad_norm": 4.042004235065528, + "learning_rate": 1.94878173531095e-05, + "loss": 1.2182, + "step": 1191 + }, + { + "epoch": 0.6471226927252985, + "grad_norm": 4.114859597350168, + "learning_rate": 1.948670568322853e-05, + "loss": 1.1787, + "step": 1192 + }, + { + "epoch": 0.6476655808903365, + "grad_norm": 4.380177461001504, + "learning_rate": 1.9485592840016682e-05, + "loss": 1.3912, + "step": 1193 + }, + { + "epoch": 0.6482084690553745, + "grad_norm": 5.160710982722015, + "learning_rate": 1.948447882361159e-05, + "loss": 1.2642, + "step": 1194 + }, + { + "epoch": 0.6487513572204126, + "grad_norm": 3.950160510485583, + "learning_rate": 1.9483363634151046e-05, + "loss": 1.1246, + "step": 1195 + }, + { + "epoch": 0.6492942453854506, + "grad_norm": 4.587449618376974, + "learning_rate": 1.9482247271772974e-05, + "loss": 1.3275, + "step": 1196 + }, + { + "epoch": 0.6498371335504886, + "grad_norm": 4.919831513668322, + "learning_rate": 1.9481129736615445e-05, + "loss": 1.4007, + "step": 1197 + }, + { + "epoch": 0.6503800217155266, + "grad_norm": 4.298954423192263, + "learning_rate": 1.948001102881668e-05, + "loss": 0.8076, + "step": 1198 + }, + { + "epoch": 0.6509229098805646, + "grad_norm": 5.525772014179547, + "learning_rate": 1.9478891148515043e-05, + "loss": 1.4558, + "step": 1199 + }, + { + "epoch": 0.6514657980456026, + "grad_norm": 5.473549669240299, + "learning_rate": 1.947777009584904e-05, + "loss": 1.0976, + "step": 1200 + }, + { + "epoch": 0.6520086862106406, + "grad_norm": 5.360010284326786, + "learning_rate": 1.947664787095733e-05, + "loss": 1.0462, + "step": 1201 + }, + { + "epoch": 0.6525515743756786, + "grad_norm": 4.696691081444868, + "learning_rate": 1.9475524473978705e-05, + "loss": 0.933, + "step": 1202 + }, + { + "epoch": 0.6530944625407166, + "grad_norm": 6.0503072087093965, + "learning_rate": 1.947439990505211e-05, + "loss": 0.9628, + "step": 1203 + }, + { + "epoch": 0.6536373507057546, + "grad_norm": 5.476272854511399, + "learning_rate": 1.9473274164316637e-05, + "loss": 0.9716, + "step": 1204 + }, + { + "epoch": 0.6541802388707926, + "grad_norm": 3.910448715873912, + "learning_rate": 1.9472147251911517e-05, + "loss": 1.0732, + "step": 1205 + }, + { + "epoch": 0.6547231270358306, + "grad_norm": 4.596114022138067, + "learning_rate": 1.9471019167976126e-05, + "loss": 0.9399, + "step": 1206 + }, + { + "epoch": 0.6552660152008686, + "grad_norm": 6.764840749864346, + "learning_rate": 1.946988991264999e-05, + "loss": 2.0529, + "step": 1207 + }, + { + "epoch": 0.6558089033659066, + "grad_norm": 5.217328230116913, + "learning_rate": 1.9468759486072778e-05, + "loss": 1.0477, + "step": 1208 + }, + { + "epoch": 0.6563517915309446, + "grad_norm": 4.03429219354135, + "learning_rate": 1.9467627888384303e-05, + "loss": 1.1103, + "step": 1209 + }, + { + "epoch": 0.6568946796959826, + "grad_norm": 4.301388894222403, + "learning_rate": 1.946649511972452e-05, + "loss": 0.8515, + "step": 1210 + }, + { + "epoch": 0.6574375678610206, + "grad_norm": 5.285446092583105, + "learning_rate": 1.9465361180233536e-05, + "loss": 1.2881, + "step": 1211 + }, + { + "epoch": 0.6579804560260586, + "grad_norm": 5.879897026479874, + "learning_rate": 1.9464226070051593e-05, + "loss": 0.9362, + "step": 1212 + }, + { + "epoch": 0.6585233441910966, + "grad_norm": 5.335257600571856, + "learning_rate": 1.9463089789319083e-05, + "loss": 1.4372, + "step": 1213 + }, + { + "epoch": 0.6590662323561346, + "grad_norm": 5.272070125431448, + "learning_rate": 1.9461952338176552e-05, + "loss": 1.2063, + "step": 1214 + }, + { + "epoch": 0.6596091205211726, + "grad_norm": 5.188765715525815, + "learning_rate": 1.946081371676467e-05, + "loss": 0.8912, + "step": 1215 + }, + { + "epoch": 0.6601520086862106, + "grad_norm": 4.747539421286818, + "learning_rate": 1.9459673925224275e-05, + "loss": 1.0381, + "step": 1216 + }, + { + "epoch": 0.6606948968512486, + "grad_norm": 4.486254331013965, + "learning_rate": 1.945853296369633e-05, + "loss": 1.2584, + "step": 1217 + }, + { + "epoch": 0.6612377850162866, + "grad_norm": 4.730012415000151, + "learning_rate": 1.945739083232195e-05, + "loss": 1.1061, + "step": 1218 + }, + { + "epoch": 0.6617806731813246, + "grad_norm": 5.713566521874704, + "learning_rate": 1.9456247531242405e-05, + "loss": 1.4628, + "step": 1219 + }, + { + "epoch": 0.6623235613463626, + "grad_norm": 6.08134803588137, + "learning_rate": 1.9455103060599093e-05, + "loss": 1.6743, + "step": 1220 + }, + { + "epoch": 0.6628664495114006, + "grad_norm": 4.778151760566848, + "learning_rate": 1.9453957420533562e-05, + "loss": 0.8454, + "step": 1221 + }, + { + "epoch": 0.6634093376764386, + "grad_norm": 5.007171333863363, + "learning_rate": 1.945281061118751e-05, + "loss": 1.079, + "step": 1222 + }, + { + "epoch": 0.6639522258414766, + "grad_norm": 4.034543591270571, + "learning_rate": 1.945166263270278e-05, + "loss": 0.8378, + "step": 1223 + }, + { + "epoch": 0.6644951140065146, + "grad_norm": 4.191899722379318, + "learning_rate": 1.9450513485221352e-05, + "loss": 1.3602, + "step": 1224 + }, + { + "epoch": 0.6650380021715526, + "grad_norm": 6.349672663190246, + "learning_rate": 1.944936316888535e-05, + "loss": 1.861, + "step": 1225 + }, + { + "epoch": 0.6655808903365906, + "grad_norm": 5.991894404952759, + "learning_rate": 1.9448211683837055e-05, + "loss": 1.0931, + "step": 1226 + }, + { + "epoch": 0.6661237785016286, + "grad_norm": 4.405247018509644, + "learning_rate": 1.9447059030218876e-05, + "loss": 1.2239, + "step": 1227 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 4.810361623786047, + "learning_rate": 1.9445905208173387e-05, + "loss": 1.1468, + "step": 1228 + }, + { + "epoch": 0.6672095548317046, + "grad_norm": 4.6967843893557015, + "learning_rate": 1.944475021784328e-05, + "loss": 1.5304, + "step": 1229 + }, + { + "epoch": 0.6677524429967426, + "grad_norm": 5.325911382516086, + "learning_rate": 1.9443594059371417e-05, + "loss": 1.2218, + "step": 1230 + }, + { + "epoch": 0.6682953311617806, + "grad_norm": 4.416437363913103, + "learning_rate": 1.9442436732900788e-05, + "loss": 1.223, + "step": 1231 + }, + { + "epoch": 0.6688382193268186, + "grad_norm": 3.720560087272861, + "learning_rate": 1.9441278238574537e-05, + "loss": 0.8155, + "step": 1232 + }, + { + "epoch": 0.6693811074918566, + "grad_norm": 5.273159297494141, + "learning_rate": 1.9440118576535947e-05, + "loss": 1.1958, + "step": 1233 + }, + { + "epoch": 0.6699239956568946, + "grad_norm": 5.684867015184436, + "learning_rate": 1.9438957746928443e-05, + "loss": 1.2384, + "step": 1234 + }, + { + "epoch": 0.6704668838219326, + "grad_norm": 4.952541482610055, + "learning_rate": 1.9437795749895604e-05, + "loss": 1.337, + "step": 1235 + }, + { + "epoch": 0.6710097719869706, + "grad_norm": 4.788863413585636, + "learning_rate": 1.9436632585581145e-05, + "loss": 0.5182, + "step": 1236 + }, + { + "epoch": 0.6715526601520087, + "grad_norm": 5.723878805619219, + "learning_rate": 1.9435468254128925e-05, + "loss": 1.1444, + "step": 1237 + }, + { + "epoch": 0.6720955483170467, + "grad_norm": 4.3068839361839, + "learning_rate": 1.9434302755682958e-05, + "loss": 1.1286, + "step": 1238 + }, + { + "epoch": 0.6726384364820847, + "grad_norm": 3.7588872381668477, + "learning_rate": 1.9433136090387384e-05, + "loss": 0.9718, + "step": 1239 + }, + { + "epoch": 0.6731813246471227, + "grad_norm": 6.802207677110681, + "learning_rate": 1.9431968258386508e-05, + "loss": 1.2389, + "step": 1240 + }, + { + "epoch": 0.6737242128121607, + "grad_norm": 3.6180158875124735, + "learning_rate": 1.9430799259824766e-05, + "loss": 0.644, + "step": 1241 + }, + { + "epoch": 0.6742671009771987, + "grad_norm": 5.559822163300458, + "learning_rate": 1.9429629094846742e-05, + "loss": 1.1157, + "step": 1242 + }, + { + "epoch": 0.6748099891422367, + "grad_norm": 4.084355750224745, + "learning_rate": 1.942845776359716e-05, + "loss": 0.91, + "step": 1243 + }, + { + "epoch": 0.6753528773072747, + "grad_norm": 3.9402358983627583, + "learning_rate": 1.9427285266220895e-05, + "loss": 0.9409, + "step": 1244 + }, + { + "epoch": 0.6758957654723127, + "grad_norm": 5.093172430302549, + "learning_rate": 1.9426111602862968e-05, + "loss": 1.5905, + "step": 1245 + }, + { + "epoch": 0.6764386536373507, + "grad_norm": 4.02536040473524, + "learning_rate": 1.942493677366853e-05, + "loss": 0.9277, + "step": 1246 + }, + { + "epoch": 0.6769815418023887, + "grad_norm": 4.741533426757492, + "learning_rate": 1.942376077878289e-05, + "loss": 1.1027, + "step": 1247 + }, + { + "epoch": 0.6775244299674267, + "grad_norm": 5.365058237804207, + "learning_rate": 1.9422583618351503e-05, + "loss": 1.072, + "step": 1248 + }, + { + "epoch": 0.6780673181324647, + "grad_norm": 4.479339123691231, + "learning_rate": 1.9421405292519956e-05, + "loss": 1.474, + "step": 1249 + }, + { + "epoch": 0.6786102062975027, + "grad_norm": 7.79086989645888, + "learning_rate": 1.942022580143398e-05, + "loss": 1.6973, + "step": 1250 + }, + { + "epoch": 0.6791530944625407, + "grad_norm": 5.621467578044424, + "learning_rate": 1.9419045145239474e-05, + "loss": 1.5432, + "step": 1251 + }, + { + "epoch": 0.6796959826275787, + "grad_norm": 5.041038447258229, + "learning_rate": 1.9417863324082444e-05, + "loss": 1.2977, + "step": 1252 + }, + { + "epoch": 0.6802388707926167, + "grad_norm": 4.717189339301973, + "learning_rate": 1.941668033810907e-05, + "loss": 0.924, + "step": 1253 + }, + { + "epoch": 0.6807817589576547, + "grad_norm": 5.086833595497465, + "learning_rate": 1.9415496187465667e-05, + "loss": 1.5545, + "step": 1254 + }, + { + "epoch": 0.6813246471226927, + "grad_norm": 4.790403842984273, + "learning_rate": 1.941431087229869e-05, + "loss": 1.2545, + "step": 1255 + }, + { + "epoch": 0.6818675352877307, + "grad_norm": 4.470785518257313, + "learning_rate": 1.9413124392754735e-05, + "loss": 1.2036, + "step": 1256 + }, + { + "epoch": 0.6824104234527687, + "grad_norm": 3.820710537112263, + "learning_rate": 1.9411936748980557e-05, + "loss": 1.0856, + "step": 1257 + }, + { + "epoch": 0.6829533116178067, + "grad_norm": 4.453720544384727, + "learning_rate": 1.9410747941123036e-05, + "loss": 1.1645, + "step": 1258 + }, + { + "epoch": 0.6834961997828447, + "grad_norm": 4.692503519890958, + "learning_rate": 1.9409557969329218e-05, + "loss": 1.1174, + "step": 1259 + }, + { + "epoch": 0.6840390879478827, + "grad_norm": 5.666971506629844, + "learning_rate": 1.940836683374627e-05, + "loss": 1.4904, + "step": 1260 + }, + { + "epoch": 0.6845819761129207, + "grad_norm": 5.468580615919992, + "learning_rate": 1.940717453452152e-05, + "loss": 1.2834, + "step": 1261 + }, + { + "epoch": 0.6851248642779587, + "grad_norm": 6.42319990350808, + "learning_rate": 1.9405981071802425e-05, + "loss": 1.1144, + "step": 1262 + }, + { + "epoch": 0.6856677524429967, + "grad_norm": 6.078674980286373, + "learning_rate": 1.9404786445736607e-05, + "loss": 1.4214, + "step": 1263 + }, + { + "epoch": 0.6862106406080347, + "grad_norm": 4.465631588283012, + "learning_rate": 1.9403590656471806e-05, + "loss": 0.7813, + "step": 1264 + }, + { + "epoch": 0.6867535287730727, + "grad_norm": 5.220582351132811, + "learning_rate": 1.940239370415593e-05, + "loss": 1.2112, + "step": 1265 + }, + { + "epoch": 0.6872964169381107, + "grad_norm": 5.1374760487841895, + "learning_rate": 1.9401195588937014e-05, + "loss": 1.2402, + "step": 1266 + }, + { + "epoch": 0.6878393051031487, + "grad_norm": 5.106543779494583, + "learning_rate": 1.9399996310963243e-05, + "loss": 1.1846, + "step": 1267 + }, + { + "epoch": 0.6883821932681867, + "grad_norm": 3.9521299273320363, + "learning_rate": 1.939879587038295e-05, + "loss": 0.8975, + "step": 1268 + }, + { + "epoch": 0.6889250814332247, + "grad_norm": 5.398700470714175, + "learning_rate": 1.9397594267344604e-05, + "loss": 1.4686, + "step": 1269 + }, + { + "epoch": 0.6894679695982627, + "grad_norm": 5.8287973073589265, + "learning_rate": 1.939639150199682e-05, + "loss": 1.011, + "step": 1270 + }, + { + "epoch": 0.6900108577633007, + "grad_norm": 4.812101240482426, + "learning_rate": 1.9395187574488358e-05, + "loss": 0.948, + "step": 1271 + }, + { + "epoch": 0.6905537459283387, + "grad_norm": 5.828731649028842, + "learning_rate": 1.939398248496813e-05, + "loss": 1.2233, + "step": 1272 + }, + { + "epoch": 0.6910966340933767, + "grad_norm": 4.890080173319348, + "learning_rate": 1.9392776233585167e-05, + "loss": 0.8859, + "step": 1273 + }, + { + "epoch": 0.6916395222584147, + "grad_norm": 4.920508250004042, + "learning_rate": 1.9391568820488674e-05, + "loss": 0.73, + "step": 1274 + }, + { + "epoch": 0.6921824104234527, + "grad_norm": 5.635401553567158, + "learning_rate": 1.9390360245827983e-05, + "loss": 1.1432, + "step": 1275 + }, + { + "epoch": 0.6927252985884907, + "grad_norm": 4.588336191443314, + "learning_rate": 1.9389150509752566e-05, + "loss": 1.3326, + "step": 1276 + }, + { + "epoch": 0.6932681867535287, + "grad_norm": 4.631694372660094, + "learning_rate": 1.9387939612412056e-05, + "loss": 1.3762, + "step": 1277 + }, + { + "epoch": 0.6938110749185668, + "grad_norm": 5.585985863722577, + "learning_rate": 1.938672755395621e-05, + "loss": 1.3606, + "step": 1278 + }, + { + "epoch": 0.6943539630836048, + "grad_norm": 4.764783952643876, + "learning_rate": 1.938551433453494e-05, + "loss": 0.9858, + "step": 1279 + }, + { + "epoch": 0.6948968512486428, + "grad_norm": 4.234252918062867, + "learning_rate": 1.9384299954298297e-05, + "loss": 0.7702, + "step": 1280 + }, + { + "epoch": 0.6954397394136808, + "grad_norm": 4.36117548276201, + "learning_rate": 1.938308441339648e-05, + "loss": 1.1846, + "step": 1281 + }, + { + "epoch": 0.6959826275787188, + "grad_norm": 4.311957696864392, + "learning_rate": 1.938186771197983e-05, + "loss": 0.9934, + "step": 1282 + }, + { + "epoch": 0.6965255157437568, + "grad_norm": 5.117807004985933, + "learning_rate": 1.9380649850198824e-05, + "loss": 1.4474, + "step": 1283 + }, + { + "epoch": 0.6970684039087948, + "grad_norm": 5.695329527120276, + "learning_rate": 1.93794308282041e-05, + "loss": 1.3001, + "step": 1284 + }, + { + "epoch": 0.6976112920738328, + "grad_norm": 4.305244126285458, + "learning_rate": 1.937821064614642e-05, + "loss": 1.1421, + "step": 1285 + }, + { + "epoch": 0.6981541802388708, + "grad_norm": 4.574797635236639, + "learning_rate": 1.93769893041767e-05, + "loss": 1.1104, + "step": 1286 + }, + { + "epoch": 0.6986970684039088, + "grad_norm": 5.051193103711606, + "learning_rate": 1.9375766802446002e-05, + "loss": 1.2858, + "step": 1287 + }, + { + "epoch": 0.6992399565689468, + "grad_norm": 5.002235181102662, + "learning_rate": 1.9374543141105518e-05, + "loss": 1.2931, + "step": 1288 + }, + { + "epoch": 0.6997828447339848, + "grad_norm": 5.259386594746194, + "learning_rate": 1.93733183203066e-05, + "loss": 1.0432, + "step": 1289 + }, + { + "epoch": 0.7003257328990228, + "grad_norm": 3.7218449672815117, + "learning_rate": 1.9372092340200736e-05, + "loss": 0.9913, + "step": 1290 + }, + { + "epoch": 0.7008686210640608, + "grad_norm": 4.45109753373239, + "learning_rate": 1.937086520093955e-05, + "loss": 1.2012, + "step": 1291 + }, + { + "epoch": 0.7014115092290988, + "grad_norm": 4.73529586440727, + "learning_rate": 1.9369636902674823e-05, + "loss": 1.3995, + "step": 1292 + }, + { + "epoch": 0.7019543973941368, + "grad_norm": 3.845443933095641, + "learning_rate": 1.936840744555847e-05, + "loss": 1.0971, + "step": 1293 + }, + { + "epoch": 0.7024972855591748, + "grad_norm": 6.67046211091574, + "learning_rate": 1.9367176829742553e-05, + "loss": 1.5273, + "step": 1294 + }, + { + "epoch": 0.7030401737242128, + "grad_norm": 4.454761961182613, + "learning_rate": 1.9365945055379275e-05, + "loss": 1.1282, + "step": 1295 + }, + { + "epoch": 0.7035830618892508, + "grad_norm": 6.552953195473038, + "learning_rate": 1.936471212262099e-05, + "loss": 2.171, + "step": 1296 + }, + { + "epoch": 0.7041259500542888, + "grad_norm": 4.894073730847946, + "learning_rate": 1.9363478031620182e-05, + "loss": 1.2303, + "step": 1297 + }, + { + "epoch": 0.7046688382193268, + "grad_norm": 3.863202427039748, + "learning_rate": 1.936224278252949e-05, + "loss": 0.8515, + "step": 1298 + }, + { + "epoch": 0.7052117263843648, + "grad_norm": 5.503562663348476, + "learning_rate": 1.9361006375501685e-05, + "loss": 1.1834, + "step": 1299 + }, + { + "epoch": 0.7057546145494028, + "grad_norm": 5.318422186143503, + "learning_rate": 1.9359768810689697e-05, + "loss": 1.3378, + "step": 1300 + }, + { + "epoch": 0.7062975027144408, + "grad_norm": 4.46194766163203, + "learning_rate": 1.9358530088246582e-05, + "loss": 1.3318, + "step": 1301 + }, + { + "epoch": 0.7068403908794788, + "grad_norm": 5.6623901137247765, + "learning_rate": 1.9357290208325552e-05, + "loss": 1.3512, + "step": 1302 + }, + { + "epoch": 0.7073832790445168, + "grad_norm": 5.805010217787694, + "learning_rate": 1.9356049171079957e-05, + "loss": 1.2383, + "step": 1303 + }, + { + "epoch": 0.7079261672095548, + "grad_norm": 4.847713425075124, + "learning_rate": 1.9354806976663286e-05, + "loss": 0.7329, + "step": 1304 + }, + { + "epoch": 0.7084690553745928, + "grad_norm": 4.207200307039602, + "learning_rate": 1.935356362522918e-05, + "loss": 1.0119, + "step": 1305 + }, + { + "epoch": 0.7090119435396308, + "grad_norm": 5.76731204400203, + "learning_rate": 1.9352319116931417e-05, + "loss": 1.7383, + "step": 1306 + }, + { + "epoch": 0.7095548317046688, + "grad_norm": 4.989804434976344, + "learning_rate": 1.935107345192392e-05, + "loss": 1.1958, + "step": 1307 + }, + { + "epoch": 0.7100977198697068, + "grad_norm": 4.6099552434159135, + "learning_rate": 1.9349826630360757e-05, + "loss": 0.9933, + "step": 1308 + }, + { + "epoch": 0.7106406080347448, + "grad_norm": 4.693361205108708, + "learning_rate": 1.9348578652396136e-05, + "loss": 1.3553, + "step": 1309 + }, + { + "epoch": 0.7111834961997828, + "grad_norm": 5.473441077550542, + "learning_rate": 1.9347329518184406e-05, + "loss": 1.2202, + "step": 1310 + }, + { + "epoch": 0.7117263843648208, + "grad_norm": 5.885376904454794, + "learning_rate": 1.9346079227880062e-05, + "loss": 1.3698, + "step": 1311 + }, + { + "epoch": 0.7122692725298588, + "grad_norm": 4.115405735194196, + "learning_rate": 1.9344827781637744e-05, + "loss": 1.3361, + "step": 1312 + }, + { + "epoch": 0.7128121606948968, + "grad_norm": 6.745730370060005, + "learning_rate": 1.9343575179612236e-05, + "loss": 1.6411, + "step": 1313 + }, + { + "epoch": 0.7133550488599348, + "grad_norm": 4.363445628534919, + "learning_rate": 1.9342321421958455e-05, + "loss": 0.792, + "step": 1314 + }, + { + "epoch": 0.7138979370249728, + "grad_norm": 6.167581448099132, + "learning_rate": 1.9341066508831472e-05, + "loss": 1.4264, + "step": 1315 + }, + { + "epoch": 0.7144408251900108, + "grad_norm": 4.973571230650737, + "learning_rate": 1.9339810440386495e-05, + "loss": 0.9769, + "step": 1316 + }, + { + "epoch": 0.7149837133550488, + "grad_norm": 5.2637238951774945, + "learning_rate": 1.933855321677888e-05, + "loss": 1.2478, + "step": 1317 + }, + { + "epoch": 0.7155266015200868, + "grad_norm": 5.2296234862723505, + "learning_rate": 1.9337294838164118e-05, + "loss": 1.3999, + "step": 1318 + }, + { + "epoch": 0.7160694896851248, + "grad_norm": 4.720322767524389, + "learning_rate": 1.9336035304697848e-05, + "loss": 1.3561, + "step": 1319 + }, + { + "epoch": 0.7166123778501629, + "grad_norm": 4.2513159313917015, + "learning_rate": 1.9334774616535854e-05, + "loss": 1.176, + "step": 1320 + }, + { + "epoch": 0.7171552660152009, + "grad_norm": 4.5267349508423225, + "learning_rate": 1.9333512773834057e-05, + "loss": 1.1211, + "step": 1321 + }, + { + "epoch": 0.7176981541802389, + "grad_norm": 5.384909301403834, + "learning_rate": 1.9332249776748523e-05, + "loss": 1.3904, + "step": 1322 + }, + { + "epoch": 0.7182410423452769, + "grad_norm": 4.67839864320759, + "learning_rate": 1.9330985625435468e-05, + "loss": 0.8608, + "step": 1323 + }, + { + "epoch": 0.7187839305103149, + "grad_norm": 5.193826637693982, + "learning_rate": 1.9329720320051233e-05, + "loss": 1.3172, + "step": 1324 + }, + { + "epoch": 0.7193268186753529, + "grad_norm": 4.8605293911870096, + "learning_rate": 1.9328453860752324e-05, + "loss": 1.3134, + "step": 1325 + }, + { + "epoch": 0.7198697068403909, + "grad_norm": 5.685265101122876, + "learning_rate": 1.9327186247695377e-05, + "loss": 1.2036, + "step": 1326 + }, + { + "epoch": 0.7204125950054289, + "grad_norm": 5.159715270999642, + "learning_rate": 1.9325917481037164e-05, + "loss": 1.4898, + "step": 1327 + }, + { + "epoch": 0.7209554831704669, + "grad_norm": 5.793178164657722, + "learning_rate": 1.9324647560934613e-05, + "loss": 1.3909, + "step": 1328 + }, + { + "epoch": 0.7214983713355049, + "grad_norm": 3.7442056178189267, + "learning_rate": 1.9323376487544795e-05, + "loss": 0.774, + "step": 1329 + }, + { + "epoch": 0.7220412595005429, + "grad_norm": 4.962252385233631, + "learning_rate": 1.9322104261024912e-05, + "loss": 1.3528, + "step": 1330 + }, + { + "epoch": 0.7225841476655809, + "grad_norm": 5.4534463408495, + "learning_rate": 1.9320830881532316e-05, + "loss": 1.6754, + "step": 1331 + }, + { + "epoch": 0.7231270358306189, + "grad_norm": 6.301262380555263, + "learning_rate": 1.93195563492245e-05, + "loss": 1.5583, + "step": 1332 + }, + { + "epoch": 0.7236699239956569, + "grad_norm": 3.970618104957748, + "learning_rate": 1.9318280664259103e-05, + "loss": 0.8964, + "step": 1333 + }, + { + "epoch": 0.7242128121606949, + "grad_norm": 3.490105469488415, + "learning_rate": 1.9317003826793904e-05, + "loss": 0.8098, + "step": 1334 + }, + { + "epoch": 0.7247557003257329, + "grad_norm": 5.240119278714349, + "learning_rate": 1.9315725836986822e-05, + "loss": 1.1688, + "step": 1335 + }, + { + "epoch": 0.7252985884907709, + "grad_norm": 4.550130688769963, + "learning_rate": 1.931444669499592e-05, + "loss": 1.2569, + "step": 1336 + }, + { + "epoch": 0.7258414766558089, + "grad_norm": 5.058353172472431, + "learning_rate": 1.9313166400979404e-05, + "loss": 1.3471, + "step": 1337 + }, + { + "epoch": 0.7263843648208469, + "grad_norm": 4.528967594836655, + "learning_rate": 1.931188495509563e-05, + "loss": 1.1831, + "step": 1338 + }, + { + "epoch": 0.7269272529858849, + "grad_norm": 4.493564615513058, + "learning_rate": 1.931060235750308e-05, + "loss": 1.4889, + "step": 1339 + }, + { + "epoch": 0.7274701411509229, + "grad_norm": 5.176344629525538, + "learning_rate": 1.9309318608360392e-05, + "loss": 1.6704, + "step": 1340 + }, + { + "epoch": 0.7280130293159609, + "grad_norm": 6.738195156936204, + "learning_rate": 1.930803370782634e-05, + "loss": 1.2147, + "step": 1341 + }, + { + "epoch": 0.7285559174809989, + "grad_norm": 4.248439922934747, + "learning_rate": 1.9306747656059847e-05, + "loss": 1.1943, + "step": 1342 + }, + { + "epoch": 0.7290988056460369, + "grad_norm": 5.175389864390461, + "learning_rate": 1.930546045321997e-05, + "loss": 1.4587, + "step": 1343 + }, + { + "epoch": 0.7296416938110749, + "grad_norm": 5.19934365393201, + "learning_rate": 1.9304172099465914e-05, + "loss": 0.9433, + "step": 1344 + }, + { + "epoch": 0.7301845819761129, + "grad_norm": 6.18681550522946, + "learning_rate": 1.9302882594957025e-05, + "loss": 1.5576, + "step": 1345 + }, + { + "epoch": 0.7307274701411509, + "grad_norm": 4.311510847746277, + "learning_rate": 1.930159193985279e-05, + "loss": 1.327, + "step": 1346 + }, + { + "epoch": 0.7312703583061889, + "grad_norm": 4.044345899710192, + "learning_rate": 1.9300300134312838e-05, + "loss": 1.2106, + "step": 1347 + }, + { + "epoch": 0.7318132464712269, + "grad_norm": 4.181894752624144, + "learning_rate": 1.929900717849694e-05, + "loss": 0.8703, + "step": 1348 + }, + { + "epoch": 0.7323561346362649, + "grad_norm": 5.37566626429507, + "learning_rate": 1.929771307256502e-05, + "loss": 1.446, + "step": 1349 + }, + { + "epoch": 0.7328990228013029, + "grad_norm": 5.37247658500195, + "learning_rate": 1.9296417816677123e-05, + "loss": 0.9239, + "step": 1350 + }, + { + "epoch": 0.7334419109663409, + "grad_norm": 5.749510968066806, + "learning_rate": 1.929512141099346e-05, + "loss": 1.4799, + "step": 1351 + }, + { + "epoch": 0.7339847991313789, + "grad_norm": 5.079635654383841, + "learning_rate": 1.929382385567436e-05, + "loss": 1.2571, + "step": 1352 + }, + { + "epoch": 0.7345276872964169, + "grad_norm": 4.923400923431618, + "learning_rate": 1.929252515088032e-05, + "loss": 1.1967, + "step": 1353 + }, + { + "epoch": 0.7350705754614549, + "grad_norm": 4.397561354835364, + "learning_rate": 1.9291225296771957e-05, + "loss": 0.9088, + "step": 1354 + }, + { + "epoch": 0.7356134636264929, + "grad_norm": 4.13744065040371, + "learning_rate": 1.9289924293510037e-05, + "loss": 1.023, + "step": 1355 + }, + { + "epoch": 0.7361563517915309, + "grad_norm": 4.266553618249588, + "learning_rate": 1.9288622141255477e-05, + "loss": 1.019, + "step": 1356 + }, + { + "epoch": 0.7366992399565689, + "grad_norm": 4.52588883698767, + "learning_rate": 1.928731884016933e-05, + "loss": 1.0935, + "step": 1357 + }, + { + "epoch": 0.7372421281216069, + "grad_norm": 5.46382933116543, + "learning_rate": 1.9286014390412786e-05, + "loss": 1.2833, + "step": 1358 + }, + { + "epoch": 0.737785016286645, + "grad_norm": 4.88651312061201, + "learning_rate": 1.928470879214718e-05, + "loss": 1.6064, + "step": 1359 + }, + { + "epoch": 0.738327904451683, + "grad_norm": 5.403882256899124, + "learning_rate": 1.9283402045533995e-05, + "loss": 1.1537, + "step": 1360 + }, + { + "epoch": 0.738870792616721, + "grad_norm": 6.381532932652649, + "learning_rate": 1.928209415073485e-05, + "loss": 1.527, + "step": 1361 + }, + { + "epoch": 0.739413680781759, + "grad_norm": 4.795244725313847, + "learning_rate": 1.9280785107911505e-05, + "loss": 1.4696, + "step": 1362 + }, + { + "epoch": 0.739956568946797, + "grad_norm": 5.8579061053451955, + "learning_rate": 1.9279474917225866e-05, + "loss": 1.5566, + "step": 1363 + }, + { + "epoch": 0.740499457111835, + "grad_norm": 4.5018596923036185, + "learning_rate": 1.927816357883998e-05, + "loss": 0.9842, + "step": 1364 + }, + { + "epoch": 0.741042345276873, + "grad_norm": 3.596935443077127, + "learning_rate": 1.927685109291604e-05, + "loss": 0.632, + "step": 1365 + }, + { + "epoch": 0.741585233441911, + "grad_norm": 4.535193817325199, + "learning_rate": 1.9275537459616364e-05, + "loss": 1.332, + "step": 1366 + }, + { + "epoch": 0.742128121606949, + "grad_norm": 6.678844775309246, + "learning_rate": 1.9274222679103437e-05, + "loss": 1.6826, + "step": 1367 + }, + { + "epoch": 0.742671009771987, + "grad_norm": 4.444554760823683, + "learning_rate": 1.927290675153987e-05, + "loss": 0.9597, + "step": 1368 + }, + { + "epoch": 0.743213897937025, + "grad_norm": 5.19917061501543, + "learning_rate": 1.927158967708841e-05, + "loss": 1.4333, + "step": 1369 + }, + { + "epoch": 0.743756786102063, + "grad_norm": 4.742331177279521, + "learning_rate": 1.927027145591197e-05, + "loss": 1.1659, + "step": 1370 + }, + { + "epoch": 0.744299674267101, + "grad_norm": 3.713496752478956, + "learning_rate": 1.926895208817358e-05, + "loss": 0.8688, + "step": 1371 + }, + { + "epoch": 0.744842562432139, + "grad_norm": 4.424173231823696, + "learning_rate": 1.9267631574036417e-05, + "loss": 1.0751, + "step": 1372 + }, + { + "epoch": 0.745385450597177, + "grad_norm": 4.7028219143924055, + "learning_rate": 1.9266309913663815e-05, + "loss": 1.3478, + "step": 1373 + }, + { + "epoch": 0.745928338762215, + "grad_norm": 5.498103807374834, + "learning_rate": 1.9264987107219237e-05, + "loss": 1.168, + "step": 1374 + }, + { + "epoch": 0.746471226927253, + "grad_norm": 4.326699254091958, + "learning_rate": 1.9263663154866285e-05, + "loss": 1.0713, + "step": 1375 + }, + { + "epoch": 0.747014115092291, + "grad_norm": 5.58697502498226, + "learning_rate": 1.926233805676871e-05, + "loss": 1.2474, + "step": 1376 + }, + { + "epoch": 0.747557003257329, + "grad_norm": 6.475832093356038, + "learning_rate": 1.92610118130904e-05, + "loss": 1.2805, + "step": 1377 + }, + { + "epoch": 0.748099891422367, + "grad_norm": 5.194166453296571, + "learning_rate": 1.925968442399539e-05, + "loss": 0.9736, + "step": 1378 + }, + { + "epoch": 0.748642779587405, + "grad_norm": 3.867607548943643, + "learning_rate": 1.9258355889647855e-05, + "loss": 1.2325, + "step": 1379 + }, + { + "epoch": 0.749185667752443, + "grad_norm": 5.602424635230674, + "learning_rate": 1.925702621021211e-05, + "loss": 1.0153, + "step": 1380 + }, + { + "epoch": 0.749728555917481, + "grad_norm": 5.011820196133583, + "learning_rate": 1.9255695385852604e-05, + "loss": 1.1823, + "step": 1381 + }, + { + "epoch": 0.750271444082519, + "grad_norm": 5.238515717778492, + "learning_rate": 1.9254363416733944e-05, + "loss": 1.4279, + "step": 1382 + }, + { + "epoch": 0.750814332247557, + "grad_norm": 5.600089025591521, + "learning_rate": 1.925303030302087e-05, + "loss": 0.9212, + "step": 1383 + }, + { + "epoch": 0.751357220412595, + "grad_norm": 4.318444011047328, + "learning_rate": 1.9251696044878255e-05, + "loss": 0.9623, + "step": 1384 + }, + { + "epoch": 0.751900108577633, + "grad_norm": 5.471464995723797, + "learning_rate": 1.925036064247113e-05, + "loss": 1.5446, + "step": 1385 + }, + { + "epoch": 0.752442996742671, + "grad_norm": 5.153911794581426, + "learning_rate": 1.9249024095964663e-05, + "loss": 1.7223, + "step": 1386 + }, + { + "epoch": 0.752985884907709, + "grad_norm": 3.7202924563691537, + "learning_rate": 1.924768640552415e-05, + "loss": 1.1469, + "step": 1387 + }, + { + "epoch": 0.753528773072747, + "grad_norm": 4.612135079041308, + "learning_rate": 1.9246347571315043e-05, + "loss": 0.9572, + "step": 1388 + }, + { + "epoch": 0.754071661237785, + "grad_norm": 4.618898432554816, + "learning_rate": 1.9245007593502937e-05, + "loss": 1.0443, + "step": 1389 + }, + { + "epoch": 0.754614549402823, + "grad_norm": 5.2787203720857345, + "learning_rate": 1.9243666472253554e-05, + "loss": 1.2034, + "step": 1390 + }, + { + "epoch": 0.755157437567861, + "grad_norm": 4.510719578546688, + "learning_rate": 1.9242324207732766e-05, + "loss": 0.8055, + "step": 1391 + }, + { + "epoch": 0.755700325732899, + "grad_norm": 6.256234902874701, + "learning_rate": 1.9240980800106596e-05, + "loss": 1.5985, + "step": 1392 + }, + { + "epoch": 0.756243213897937, + "grad_norm": 9.01213198061367, + "learning_rate": 1.923963624954119e-05, + "loss": 1.7921, + "step": 1393 + }, + { + "epoch": 0.756786102062975, + "grad_norm": 5.149498439254725, + "learning_rate": 1.923829055620285e-05, + "loss": 1.2406, + "step": 1394 + }, + { + "epoch": 0.757328990228013, + "grad_norm": 5.083183084254609, + "learning_rate": 1.9236943720258007e-05, + "loss": 0.9887, + "step": 1395 + }, + { + "epoch": 0.757871878393051, + "grad_norm": 4.5118989088500685, + "learning_rate": 1.9235595741873247e-05, + "loss": 1.3528, + "step": 1396 + }, + { + "epoch": 0.758414766558089, + "grad_norm": 4.583236692880461, + "learning_rate": 1.923424662121528e-05, + "loss": 1.0185, + "step": 1397 + }, + { + "epoch": 0.758957654723127, + "grad_norm": 5.3602373686113625, + "learning_rate": 1.9232896358450976e-05, + "loss": 1.1827, + "step": 1398 + }, + { + "epoch": 0.759500542888165, + "grad_norm": 4.510149132944334, + "learning_rate": 1.9231544953747336e-05, + "loss": 0.9981, + "step": 1399 + }, + { + "epoch": 0.760043431053203, + "grad_norm": 5.008678356958532, + "learning_rate": 1.9230192407271506e-05, + "loss": 1.4957, + "step": 1400 + }, + { + "epoch": 0.760586319218241, + "grad_norm": 4.6938647576746995, + "learning_rate": 1.9228838719190765e-05, + "loss": 1.361, + "step": 1401 + }, + { + "epoch": 0.761129207383279, + "grad_norm": 5.263462060803471, + "learning_rate": 1.9227483889672544e-05, + "loss": 1.1716, + "step": 1402 + }, + { + "epoch": 0.761672095548317, + "grad_norm": 4.737410394333335, + "learning_rate": 1.9226127918884407e-05, + "loss": 1.3924, + "step": 1403 + }, + { + "epoch": 0.762214983713355, + "grad_norm": 5.117476933198257, + "learning_rate": 1.9224770806994066e-05, + "loss": 1.1215, + "step": 1404 + }, + { + "epoch": 0.7627578718783931, + "grad_norm": 4.525492833460315, + "learning_rate": 1.922341255416937e-05, + "loss": 1.0626, + "step": 1405 + }, + { + "epoch": 0.7633007600434311, + "grad_norm": 4.928417801176309, + "learning_rate": 1.9222053160578312e-05, + "loss": 0.9576, + "step": 1406 + }, + { + "epoch": 0.7638436482084691, + "grad_norm": 4.44076437857558, + "learning_rate": 1.9220692626389018e-05, + "loss": 0.9186, + "step": 1407 + }, + { + "epoch": 0.7643865363735071, + "grad_norm": 4.779705057681976, + "learning_rate": 1.9219330951769763e-05, + "loss": 1.3392, + "step": 1408 + }, + { + "epoch": 0.7649294245385451, + "grad_norm": 5.37856562129718, + "learning_rate": 1.9217968136888965e-05, + "loss": 1.043, + "step": 1409 + }, + { + "epoch": 0.7654723127035831, + "grad_norm": 5.596447438931628, + "learning_rate": 1.9216604181915178e-05, + "loss": 1.3223, + "step": 1410 + }, + { + "epoch": 0.7660152008686211, + "grad_norm": 5.267245970848837, + "learning_rate": 1.9215239087017093e-05, + "loss": 1.0484, + "step": 1411 + }, + { + "epoch": 0.7665580890336591, + "grad_norm": 6.124227707475327, + "learning_rate": 1.9213872852363552e-05, + "loss": 1.5361, + "step": 1412 + }, + { + "epoch": 0.7671009771986971, + "grad_norm": 5.632804630433347, + "learning_rate": 1.9212505478123532e-05, + "loss": 1.2227, + "step": 1413 + }, + { + "epoch": 0.7676438653637351, + "grad_norm": 4.903635376409972, + "learning_rate": 1.9211136964466152e-05, + "loss": 0.9045, + "step": 1414 + }, + { + "epoch": 0.7681867535287731, + "grad_norm": 4.995962297168909, + "learning_rate": 1.9209767311560673e-05, + "loss": 1.2364, + "step": 1415 + }, + { + "epoch": 0.7687296416938111, + "grad_norm": 4.76685077713632, + "learning_rate": 1.9208396519576494e-05, + "loss": 1.4849, + "step": 1416 + }, + { + "epoch": 0.7692725298588491, + "grad_norm": 5.66289239913894, + "learning_rate": 1.9207024588683158e-05, + "loss": 1.389, + "step": 1417 + }, + { + "epoch": 0.7698154180238871, + "grad_norm": 4.044195406366437, + "learning_rate": 1.920565151905035e-05, + "loss": 0.5736, + "step": 1418 + }, + { + "epoch": 0.7703583061889251, + "grad_norm": 5.610527750585898, + "learning_rate": 1.9204277310847887e-05, + "loss": 1.5147, + "step": 1419 + }, + { + "epoch": 0.7709011943539631, + "grad_norm": 3.8629108074125424, + "learning_rate": 1.9202901964245734e-05, + "loss": 0.9184, + "step": 1420 + }, + { + "epoch": 0.7714440825190011, + "grad_norm": 4.582445031278247, + "learning_rate": 1.9201525479414e-05, + "loss": 1.071, + "step": 1421 + }, + { + "epoch": 0.7719869706840391, + "grad_norm": 5.016846104390101, + "learning_rate": 1.9200147856522933e-05, + "loss": 1.3673, + "step": 1422 + }, + { + "epoch": 0.7725298588490771, + "grad_norm": 4.798189213061551, + "learning_rate": 1.9198769095742914e-05, + "loss": 1.3483, + "step": 1423 + }, + { + "epoch": 0.7730727470141151, + "grad_norm": 5.871902023790772, + "learning_rate": 1.9197389197244473e-05, + "loss": 1.7625, + "step": 1424 + }, + { + "epoch": 0.7736156351791531, + "grad_norm": 5.895934775040147, + "learning_rate": 1.9196008161198277e-05, + "loss": 0.999, + "step": 1425 + }, + { + "epoch": 0.7741585233441911, + "grad_norm": 5.0199402476408155, + "learning_rate": 1.9194625987775138e-05, + "loss": 1.3251, + "step": 1426 + }, + { + "epoch": 0.7747014115092291, + "grad_norm": 6.650397672217608, + "learning_rate": 1.9193242677146e-05, + "loss": 1.2162, + "step": 1427 + }, + { + "epoch": 0.7752442996742671, + "grad_norm": 5.515267491505962, + "learning_rate": 1.9191858229481958e-05, + "loss": 1.3849, + "step": 1428 + }, + { + "epoch": 0.7757871878393051, + "grad_norm": 5.694611687374825, + "learning_rate": 1.9190472644954236e-05, + "loss": 1.0831, + "step": 1429 + }, + { + "epoch": 0.7763300760043431, + "grad_norm": 4.548114219835821, + "learning_rate": 1.9189085923734215e-05, + "loss": 1.2549, + "step": 1430 + }, + { + "epoch": 0.7768729641693811, + "grad_norm": 6.119997613777156, + "learning_rate": 1.9187698065993398e-05, + "loss": 1.6137, + "step": 1431 + }, + { + "epoch": 0.7774158523344191, + "grad_norm": 4.574150272616086, + "learning_rate": 1.9186309071903445e-05, + "loss": 1.3015, + "step": 1432 + }, + { + "epoch": 0.7779587404994571, + "grad_norm": 4.77854731853541, + "learning_rate": 1.9184918941636142e-05, + "loss": 0.6973, + "step": 1433 + }, + { + "epoch": 0.7785016286644951, + "grad_norm": 6.926122760031406, + "learning_rate": 1.9183527675363425e-05, + "loss": 1.4034, + "step": 1434 + }, + { + "epoch": 0.7790445168295331, + "grad_norm": 4.6748890605309645, + "learning_rate": 1.9182135273257372e-05, + "loss": 0.9854, + "step": 1435 + }, + { + "epoch": 0.7795874049945711, + "grad_norm": 5.072338861625223, + "learning_rate": 1.9180741735490194e-05, + "loss": 1.1604, + "step": 1436 + }, + { + "epoch": 0.7801302931596091, + "grad_norm": 5.995183838581222, + "learning_rate": 1.9179347062234245e-05, + "loss": 1.3913, + "step": 1437 + }, + { + "epoch": 0.7806731813246471, + "grad_norm": 5.494006248546126, + "learning_rate": 1.917795125366202e-05, + "loss": 1.2541, + "step": 1438 + }, + { + "epoch": 0.7812160694896851, + "grad_norm": 3.8223879163574694, + "learning_rate": 1.917655430994616e-05, + "loss": 0.7292, + "step": 1439 + }, + { + "epoch": 0.7817589576547231, + "grad_norm": 4.1476284953657405, + "learning_rate": 1.9175156231259434e-05, + "loss": 0.608, + "step": 1440 + }, + { + "epoch": 0.7823018458197611, + "grad_norm": 5.5144637244676495, + "learning_rate": 1.9173757017774764e-05, + "loss": 1.2674, + "step": 1441 + }, + { + "epoch": 0.7828447339847991, + "grad_norm": 5.4660654663594945, + "learning_rate": 1.9172356669665206e-05, + "loss": 1.3043, + "step": 1442 + }, + { + "epoch": 0.7833876221498371, + "grad_norm": 3.526312966998694, + "learning_rate": 1.9170955187103957e-05, + "loss": 0.8721, + "step": 1443 + }, + { + "epoch": 0.7839305103148752, + "grad_norm": 4.395429767339641, + "learning_rate": 1.9169552570264355e-05, + "loss": 0.985, + "step": 1444 + }, + { + "epoch": 0.7844733984799132, + "grad_norm": 4.608635000415997, + "learning_rate": 1.9168148819319874e-05, + "loss": 1.3492, + "step": 1445 + }, + { + "epoch": 0.7850162866449512, + "grad_norm": 5.729965497961937, + "learning_rate": 1.9166743934444137e-05, + "loss": 1.3405, + "step": 1446 + }, + { + "epoch": 0.7855591748099892, + "grad_norm": 6.60908414105679, + "learning_rate": 1.91653379158109e-05, + "loss": 1.5271, + "step": 1447 + }, + { + "epoch": 0.7861020629750272, + "grad_norm": 4.91386728282805, + "learning_rate": 1.916393076359406e-05, + "loss": 1.475, + "step": 1448 + }, + { + "epoch": 0.7866449511400652, + "grad_norm": 5.286158051966196, + "learning_rate": 1.916252247796766e-05, + "loss": 1.5235, + "step": 1449 + }, + { + "epoch": 0.7871878393051032, + "grad_norm": 4.699459928019414, + "learning_rate": 1.916111305910588e-05, + "loss": 1.151, + "step": 1450 + }, + { + "epoch": 0.7877307274701412, + "grad_norm": 5.125659016648778, + "learning_rate": 1.915970250718303e-05, + "loss": 1.2952, + "step": 1451 + }, + { + "epoch": 0.7882736156351792, + "grad_norm": 4.976170683421487, + "learning_rate": 1.915829082237358e-05, + "loss": 1.3291, + "step": 1452 + }, + { + "epoch": 0.7888165038002172, + "grad_norm": 5.237853683905863, + "learning_rate": 1.9156878004852123e-05, + "loss": 1.4775, + "step": 1453 + }, + { + "epoch": 0.7893593919652552, + "grad_norm": 4.914941394388547, + "learning_rate": 1.9155464054793404e-05, + "loss": 1.2151, + "step": 1454 + }, + { + "epoch": 0.7899022801302932, + "grad_norm": 5.050785426148085, + "learning_rate": 1.9154048972372293e-05, + "loss": 1.163, + "step": 1455 + }, + { + "epoch": 0.7904451682953312, + "grad_norm": 4.2016259300832255, + "learning_rate": 1.915263275776382e-05, + "loss": 0.9601, + "step": 1456 + }, + { + "epoch": 0.7909880564603692, + "grad_norm": 5.48804064939896, + "learning_rate": 1.915121541114314e-05, + "loss": 1.3026, + "step": 1457 + }, + { + "epoch": 0.7915309446254072, + "grad_norm": 5.230495684608947, + "learning_rate": 1.9149796932685552e-05, + "loss": 1.1923, + "step": 1458 + }, + { + "epoch": 0.7920738327904452, + "grad_norm": 4.788655104859546, + "learning_rate": 1.91483773225665e-05, + "loss": 1.2437, + "step": 1459 + }, + { + "epoch": 0.7926167209554832, + "grad_norm": 4.506216689801701, + "learning_rate": 1.9146956580961556e-05, + "loss": 0.9364, + "step": 1460 + }, + { + "epoch": 0.7931596091205212, + "grad_norm": 4.2335852812311865, + "learning_rate": 1.9145534708046446e-05, + "loss": 0.7104, + "step": 1461 + }, + { + "epoch": 0.7937024972855592, + "grad_norm": 6.204688311211956, + "learning_rate": 1.914411170399703e-05, + "loss": 1.0825, + "step": 1462 + }, + { + "epoch": 0.7942453854505972, + "grad_norm": 3.445489329210515, + "learning_rate": 1.91426875689893e-05, + "loss": 0.9921, + "step": 1463 + }, + { + "epoch": 0.7947882736156352, + "grad_norm": 5.204416925095863, + "learning_rate": 1.9141262303199403e-05, + "loss": 1.3043, + "step": 1464 + }, + { + "epoch": 0.7953311617806732, + "grad_norm": 6.3486214559668985, + "learning_rate": 1.9139835906803612e-05, + "loss": 1.3193, + "step": 1465 + }, + { + "epoch": 0.7958740499457112, + "grad_norm": 5.610159156463615, + "learning_rate": 1.913840837997835e-05, + "loss": 1.0455, + "step": 1466 + }, + { + "epoch": 0.7964169381107492, + "grad_norm": 4.625080334899242, + "learning_rate": 1.913697972290018e-05, + "loss": 0.7981, + "step": 1467 + }, + { + "epoch": 0.7969598262757872, + "grad_norm": 5.993005746484773, + "learning_rate": 1.9135549935745792e-05, + "loss": 1.1674, + "step": 1468 + }, + { + "epoch": 0.7975027144408252, + "grad_norm": 5.084136125969368, + "learning_rate": 1.913411901869203e-05, + "loss": 1.0362, + "step": 1469 + }, + { + "epoch": 0.7980456026058632, + "grad_norm": 5.701736254232889, + "learning_rate": 1.913268697191587e-05, + "loss": 1.2159, + "step": 1470 + }, + { + "epoch": 0.7985884907709012, + "grad_norm": 4.715190322082246, + "learning_rate": 1.9131253795594428e-05, + "loss": 0.9848, + "step": 1471 + }, + { + "epoch": 0.7991313789359392, + "grad_norm": 5.375233653105075, + "learning_rate": 1.9129819489904964e-05, + "loss": 1.0476, + "step": 1472 + }, + { + "epoch": 0.7996742671009772, + "grad_norm": 6.352674291337691, + "learning_rate": 1.9128384055024874e-05, + "loss": 1.2362, + "step": 1473 + }, + { + "epoch": 0.8002171552660152, + "grad_norm": 5.46716259791096, + "learning_rate": 1.91269474911317e-05, + "loss": 1.2949, + "step": 1474 + }, + { + "epoch": 0.8007600434310532, + "grad_norm": 5.3150808603597826, + "learning_rate": 1.912550979840311e-05, + "loss": 1.1587, + "step": 1475 + }, + { + "epoch": 0.8013029315960912, + "grad_norm": 4.46814887402293, + "learning_rate": 1.9124070977016926e-05, + "loss": 0.9649, + "step": 1476 + }, + { + "epoch": 0.8018458197611292, + "grad_norm": 4.916267658604107, + "learning_rate": 1.9122631027151103e-05, + "loss": 1.2117, + "step": 1477 + }, + { + "epoch": 0.8023887079261672, + "grad_norm": 5.506988713852874, + "learning_rate": 1.9121189948983733e-05, + "loss": 1.3387, + "step": 1478 + }, + { + "epoch": 0.8029315960912052, + "grad_norm": 4.7704152930487895, + "learning_rate": 1.911974774269305e-05, + "loss": 1.3379, + "step": 1479 + }, + { + "epoch": 0.8034744842562432, + "grad_norm": 5.667769720352476, + "learning_rate": 1.9118304408457435e-05, + "loss": 1.0552, + "step": 1480 + }, + { + "epoch": 0.8040173724212812, + "grad_norm": 4.9471232370904925, + "learning_rate": 1.91168599464554e-05, + "loss": 1.0013, + "step": 1481 + }, + { + "epoch": 0.8045602605863192, + "grad_norm": 4.856623754785127, + "learning_rate": 1.9115414356865594e-05, + "loss": 0.9001, + "step": 1482 + }, + { + "epoch": 0.8051031487513572, + "grad_norm": 4.004976425614515, + "learning_rate": 1.9113967639866815e-05, + "loss": 0.8114, + "step": 1483 + }, + { + "epoch": 0.8056460369163952, + "grad_norm": 5.983237168687411, + "learning_rate": 1.911251979563799e-05, + "loss": 1.1678, + "step": 1484 + }, + { + "epoch": 0.8061889250814332, + "grad_norm": 4.808701495369871, + "learning_rate": 1.9111070824358196e-05, + "loss": 0.9181, + "step": 1485 + }, + { + "epoch": 0.8067318132464713, + "grad_norm": 5.080557830592386, + "learning_rate": 1.910962072620664e-05, + "loss": 1.0751, + "step": 1486 + }, + { + "epoch": 0.8072747014115093, + "grad_norm": 5.730412968009966, + "learning_rate": 1.9108169501362674e-05, + "loss": 1.2727, + "step": 1487 + }, + { + "epoch": 0.8078175895765473, + "grad_norm": 6.47838124684283, + "learning_rate": 1.9106717150005785e-05, + "loss": 1.6491, + "step": 1488 + }, + { + "epoch": 0.8083604777415853, + "grad_norm": 5.311235841870027, + "learning_rate": 1.910526367231561e-05, + "loss": 0.8382, + "step": 1489 + }, + { + "epoch": 0.8089033659066233, + "grad_norm": 5.43072146168114, + "learning_rate": 1.9103809068471914e-05, + "loss": 1.3026, + "step": 1490 + }, + { + "epoch": 0.8094462540716613, + "grad_norm": 4.46156225654375, + "learning_rate": 1.9102353338654597e-05, + "loss": 0.8071, + "step": 1491 + }, + { + "epoch": 0.8099891422366993, + "grad_norm": 5.297520452299967, + "learning_rate": 1.9100896483043714e-05, + "loss": 1.0625, + "step": 1492 + }, + { + "epoch": 0.8105320304017373, + "grad_norm": 6.736735698357962, + "learning_rate": 1.909943850181945e-05, + "loss": 1.6497, + "step": 1493 + }, + { + "epoch": 0.8110749185667753, + "grad_norm": 4.8675910913183955, + "learning_rate": 1.9097979395162132e-05, + "loss": 1.0822, + "step": 1494 + }, + { + "epoch": 0.8116178067318133, + "grad_norm": 4.5489525578867305, + "learning_rate": 1.909651916325222e-05, + "loss": 0.7908, + "step": 1495 + }, + { + "epoch": 0.8121606948968513, + "grad_norm": 6.002121574887706, + "learning_rate": 1.909505780627032e-05, + "loss": 0.5178, + "step": 1496 + }, + { + "epoch": 0.8127035830618893, + "grad_norm": 4.34564271493731, + "learning_rate": 1.9093595324397175e-05, + "loss": 0.9818, + "step": 1497 + }, + { + "epoch": 0.8132464712269273, + "grad_norm": 5.0382562486474525, + "learning_rate": 1.9092131717813668e-05, + "loss": 1.0997, + "step": 1498 + }, + { + "epoch": 0.8137893593919653, + "grad_norm": 6.8289191133749, + "learning_rate": 1.909066698670082e-05, + "loss": 1.0046, + "step": 1499 + }, + { + "epoch": 0.8143322475570033, + "grad_norm": 6.6968746410007585, + "learning_rate": 1.908920113123979e-05, + "loss": 1.5191, + "step": 1500 + }, + { + "epoch": 0.8148751357220413, + "grad_norm": 5.548259576405749, + "learning_rate": 1.9087734151611877e-05, + "loss": 1.3272, + "step": 1501 + }, + { + "epoch": 0.8154180238870793, + "grad_norm": 5.684602861026239, + "learning_rate": 1.9086266047998522e-05, + "loss": 1.0698, + "step": 1502 + }, + { + "epoch": 0.8159609120521173, + "grad_norm": 6.832627276636624, + "learning_rate": 1.90847968205813e-05, + "loss": 1.5796, + "step": 1503 + }, + { + "epoch": 0.8165038002171553, + "grad_norm": 5.605724679057613, + "learning_rate": 1.908332646954193e-05, + "loss": 1.1709, + "step": 1504 + }, + { + "epoch": 0.8170466883821933, + "grad_norm": 4.607345782065147, + "learning_rate": 1.908185499506226e-05, + "loss": 0.8686, + "step": 1505 + }, + { + "epoch": 0.8175895765472313, + "grad_norm": 4.687252800873217, + "learning_rate": 1.9080382397324296e-05, + "loss": 0.8881, + "step": 1506 + }, + { + "epoch": 0.8181324647122693, + "grad_norm": 6.357714385620256, + "learning_rate": 1.907890867651016e-05, + "loss": 1.1948, + "step": 1507 + }, + { + "epoch": 0.8186753528773073, + "grad_norm": 4.8608334992124425, + "learning_rate": 1.9077433832802135e-05, + "loss": 1.0311, + "step": 1508 + }, + { + "epoch": 0.8192182410423453, + "grad_norm": 5.948864135911491, + "learning_rate": 1.9075957866382623e-05, + "loss": 1.1314, + "step": 1509 + }, + { + "epoch": 0.8197611292073833, + "grad_norm": 4.807850657807276, + "learning_rate": 1.9074480777434178e-05, + "loss": 0.9478, + "step": 1510 + }, + { + "epoch": 0.8203040173724213, + "grad_norm": 4.26222909817726, + "learning_rate": 1.9073002566139486e-05, + "loss": 0.8541, + "step": 1511 + }, + { + "epoch": 0.8208469055374593, + "grad_norm": 6.6130092616391005, + "learning_rate": 1.9071523232681382e-05, + "loss": 1.2754, + "step": 1512 + }, + { + "epoch": 0.8213897937024973, + "grad_norm": 5.255155673043404, + "learning_rate": 1.907004277724282e-05, + "loss": 1.0813, + "step": 1513 + }, + { + "epoch": 0.8219326818675353, + "grad_norm": 5.079098182105948, + "learning_rate": 1.9068561200006917e-05, + "loss": 1.0016, + "step": 1514 + }, + { + "epoch": 0.8224755700325733, + "grad_norm": 5.491172879434626, + "learning_rate": 1.906707850115691e-05, + "loss": 1.2884, + "step": 1515 + }, + { + "epoch": 0.8230184581976113, + "grad_norm": 6.718962166599785, + "learning_rate": 1.9065594680876182e-05, + "loss": 1.4973, + "step": 1516 + }, + { + "epoch": 0.8235613463626493, + "grad_norm": 5.348428262646105, + "learning_rate": 1.9064109739348257e-05, + "loss": 1.1113, + "step": 1517 + }, + { + "epoch": 0.8241042345276873, + "grad_norm": 4.604059650726469, + "learning_rate": 1.906262367675679e-05, + "loss": 1.1614, + "step": 1518 + }, + { + "epoch": 0.8246471226927253, + "grad_norm": 5.895860662978225, + "learning_rate": 1.9061136493285586e-05, + "loss": 1.1532, + "step": 1519 + }, + { + "epoch": 0.8251900108577633, + "grad_norm": 5.621847434524929, + "learning_rate": 1.905964818911858e-05, + "loss": 0.9277, + "step": 1520 + }, + { + "epoch": 0.8257328990228013, + "grad_norm": 4.585955492662189, + "learning_rate": 1.9058158764439844e-05, + "loss": 0.7988, + "step": 1521 + }, + { + "epoch": 0.8262757871878393, + "grad_norm": 5.784308925988881, + "learning_rate": 1.9056668219433595e-05, + "loss": 1.6078, + "step": 1522 + }, + { + "epoch": 0.8268186753528773, + "grad_norm": 5.621216253388429, + "learning_rate": 1.905517655428419e-05, + "loss": 1.389, + "step": 1523 + }, + { + "epoch": 0.8273615635179153, + "grad_norm": 5.514208208061458, + "learning_rate": 1.9053683769176115e-05, + "loss": 0.7612, + "step": 1524 + }, + { + "epoch": 0.8279044516829533, + "grad_norm": 4.6572185133859065, + "learning_rate": 1.9052189864294002e-05, + "loss": 0.744, + "step": 1525 + }, + { + "epoch": 0.8284473398479913, + "grad_norm": 5.239632270223703, + "learning_rate": 1.905069483982262e-05, + "loss": 0.939, + "step": 1526 + }, + { + "epoch": 0.8289902280130294, + "grad_norm": 3.7037650901217454, + "learning_rate": 1.9049198695946876e-05, + "loss": 1.0177, + "step": 1527 + }, + { + "epoch": 0.8295331161780674, + "grad_norm": 6.974869726679236, + "learning_rate": 1.9047701432851813e-05, + "loss": 1.2722, + "step": 1528 + }, + { + "epoch": 0.8300760043431054, + "grad_norm": 6.642878483620589, + "learning_rate": 1.904620305072262e-05, + "loss": 1.5369, + "step": 1529 + }, + { + "epoch": 0.8306188925081434, + "grad_norm": 7.99082645392899, + "learning_rate": 1.9044703549744616e-05, + "loss": 1.2245, + "step": 1530 + }, + { + "epoch": 0.8311617806731814, + "grad_norm": 6.5593948883008135, + "learning_rate": 1.904320293010326e-05, + "loss": 1.307, + "step": 1531 + }, + { + "epoch": 0.8317046688382194, + "grad_norm": 4.930764759519961, + "learning_rate": 1.9041701191984155e-05, + "loss": 0.9564, + "step": 1532 + }, + { + "epoch": 0.8322475570032574, + "grad_norm": 3.9621684331427773, + "learning_rate": 1.9040198335573033e-05, + "loss": 0.8153, + "step": 1533 + }, + { + "epoch": 0.8327904451682954, + "grad_norm": 4.859084711241092, + "learning_rate": 1.9038694361055774e-05, + "loss": 1.0967, + "step": 1534 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.9476278172327595, + "learning_rate": 1.903718926861839e-05, + "loss": 1.4743, + "step": 1535 + }, + { + "epoch": 0.8338762214983714, + "grad_norm": 4.312795088472635, + "learning_rate": 1.903568305844704e-05, + "loss": 0.8249, + "step": 1536 + }, + { + "epoch": 0.8344191096634094, + "grad_norm": 6.638027572281857, + "learning_rate": 1.9034175730728e-05, + "loss": 1.6353, + "step": 1537 + }, + { + "epoch": 0.8349619978284474, + "grad_norm": 5.709548912096025, + "learning_rate": 1.9032667285647714e-05, + "loss": 1.8512, + "step": 1538 + }, + { + "epoch": 0.8355048859934854, + "grad_norm": 5.098275426179771, + "learning_rate": 1.9031157723392738e-05, + "loss": 1.0348, + "step": 1539 + }, + { + "epoch": 0.8360477741585234, + "grad_norm": 5.05236269448673, + "learning_rate": 1.9029647044149783e-05, + "loss": 1.108, + "step": 1540 + }, + { + "epoch": 0.8365906623235614, + "grad_norm": 4.674708875536442, + "learning_rate": 1.9028135248105692e-05, + "loss": 1.0453, + "step": 1541 + }, + { + "epoch": 0.8371335504885994, + "grad_norm": 4.717321956389267, + "learning_rate": 1.902662233544744e-05, + "loss": 0.9042, + "step": 1542 + }, + { + "epoch": 0.8376764386536374, + "grad_norm": 6.217040367936212, + "learning_rate": 1.9025108306362158e-05, + "loss": 1.0762, + "step": 1543 + }, + { + "epoch": 0.8382193268186754, + "grad_norm": 5.338744288323046, + "learning_rate": 1.9023593161037094e-05, + "loss": 1.1631, + "step": 1544 + }, + { + "epoch": 0.8387622149837134, + "grad_norm": 4.8224874086687874, + "learning_rate": 1.9022076899659643e-05, + "loss": 1.3907, + "step": 1545 + }, + { + "epoch": 0.8393051031487514, + "grad_norm": 4.699800413179793, + "learning_rate": 1.9020559522417345e-05, + "loss": 0.7682, + "step": 1546 + }, + { + "epoch": 0.8398479913137894, + "grad_norm": 4.271439794022252, + "learning_rate": 1.9019041029497866e-05, + "loss": 0.8475, + "step": 1547 + }, + { + "epoch": 0.8403908794788274, + "grad_norm": 4.240562925498168, + "learning_rate": 1.9017521421089022e-05, + "loss": 0.8201, + "step": 1548 + }, + { + "epoch": 0.8409337676438654, + "grad_norm": 5.24718786534657, + "learning_rate": 1.9016000697378755e-05, + "loss": 1.0728, + "step": 1549 + }, + { + "epoch": 0.8414766558089034, + "grad_norm": 6.492220902503762, + "learning_rate": 1.9014478858555156e-05, + "loss": 0.9432, + "step": 1550 + }, + { + "epoch": 0.8420195439739414, + "grad_norm": 4.917811923935393, + "learning_rate": 1.9012955904806438e-05, + "loss": 1.1672, + "step": 1551 + }, + { + "epoch": 0.8425624321389794, + "grad_norm": 5.66709669173795, + "learning_rate": 1.9011431836320976e-05, + "loss": 1.5058, + "step": 1552 + }, + { + "epoch": 0.8431053203040174, + "grad_norm": 5.658634152413846, + "learning_rate": 1.9009906653287258e-05, + "loss": 1.3653, + "step": 1553 + }, + { + "epoch": 0.8436482084690554, + "grad_norm": 4.956250367213818, + "learning_rate": 1.9008380355893925e-05, + "loss": 0.8309, + "step": 1554 + }, + { + "epoch": 0.8441910966340934, + "grad_norm": 4.329497016180362, + "learning_rate": 1.9006852944329753e-05, + "loss": 1.1141, + "step": 1555 + }, + { + "epoch": 0.8447339847991314, + "grad_norm": 5.717864943326053, + "learning_rate": 1.9005324418783658e-05, + "loss": 1.3274, + "step": 1556 + }, + { + "epoch": 0.8452768729641694, + "grad_norm": 5.657300743895006, + "learning_rate": 1.900379477944468e-05, + "loss": 1.2432, + "step": 1557 + }, + { + "epoch": 0.8458197611292074, + "grad_norm": 5.325761071371062, + "learning_rate": 1.900226402650202e-05, + "loss": 1.3428, + "step": 1558 + }, + { + "epoch": 0.8463626492942454, + "grad_norm": 5.207304704126981, + "learning_rate": 1.9000732160144996e-05, + "loss": 1.455, + "step": 1559 + }, + { + "epoch": 0.8469055374592834, + "grad_norm": 5.248656371435267, + "learning_rate": 1.8999199180563074e-05, + "loss": 0.7851, + "step": 1560 + }, + { + "epoch": 0.8474484256243214, + "grad_norm": 6.92723307445887, + "learning_rate": 1.899766508794585e-05, + "loss": 1.5236, + "step": 1561 + }, + { + "epoch": 0.8479913137893594, + "grad_norm": 4.3434986726191935, + "learning_rate": 1.899612988248307e-05, + "loss": 0.9117, + "step": 1562 + }, + { + "epoch": 0.8485342019543974, + "grad_norm": 5.815800316925401, + "learning_rate": 1.8994593564364612e-05, + "loss": 1.0097, + "step": 1563 + }, + { + "epoch": 0.8490770901194354, + "grad_norm": 5.5494362406379905, + "learning_rate": 1.8993056133780484e-05, + "loss": 1.1829, + "step": 1564 + }, + { + "epoch": 0.8496199782844734, + "grad_norm": 4.323494913553237, + "learning_rate": 1.899151759092084e-05, + "loss": 0.9359, + "step": 1565 + }, + { + "epoch": 0.8501628664495114, + "grad_norm": 4.455911135073797, + "learning_rate": 1.898997793597597e-05, + "loss": 0.9354, + "step": 1566 + }, + { + "epoch": 0.8507057546145494, + "grad_norm": 6.684553981467422, + "learning_rate": 1.8988437169136302e-05, + "loss": 1.3693, + "step": 1567 + }, + { + "epoch": 0.8512486427795874, + "grad_norm": 6.341757024960091, + "learning_rate": 1.89868952905924e-05, + "loss": 0.9441, + "step": 1568 + }, + { + "epoch": 0.8517915309446255, + "grad_norm": 6.017605683279906, + "learning_rate": 1.8985352300534965e-05, + "loss": 1.2747, + "step": 1569 + }, + { + "epoch": 0.8523344191096635, + "grad_norm": 5.501152081837135, + "learning_rate": 1.8983808199154835e-05, + "loss": 0.8414, + "step": 1570 + }, + { + "epoch": 0.8528773072747015, + "grad_norm": 5.075164131477861, + "learning_rate": 1.8982262986642993e-05, + "loss": 1.124, + "step": 1571 + }, + { + "epoch": 0.8534201954397395, + "grad_norm": 5.7750932469486065, + "learning_rate": 1.8980716663190545e-05, + "loss": 1.3968, + "step": 1572 + }, + { + "epoch": 0.8539630836047775, + "grad_norm": 6.967872140444835, + "learning_rate": 1.897916922898875e-05, + "loss": 1.2395, + "step": 1573 + }, + { + "epoch": 0.8545059717698155, + "grad_norm": 5.037804725894728, + "learning_rate": 1.8977620684228994e-05, + "loss": 0.9741, + "step": 1574 + }, + { + "epoch": 0.8550488599348535, + "grad_norm": 4.586000206489282, + "learning_rate": 1.8976071029102802e-05, + "loss": 0.6757, + "step": 1575 + }, + { + "epoch": 0.8555917480998915, + "grad_norm": 4.686417605581938, + "learning_rate": 1.897452026380184e-05, + "loss": 0.8382, + "step": 1576 + }, + { + "epoch": 0.8561346362649295, + "grad_norm": 5.133601528061789, + "learning_rate": 1.8972968388517908e-05, + "loss": 1.2999, + "step": 1577 + }, + { + "epoch": 0.8566775244299675, + "grad_norm": 6.170796021401621, + "learning_rate": 1.8971415403442942e-05, + "loss": 1.1513, + "step": 1578 + }, + { + "epoch": 0.8572204125950055, + "grad_norm": 5.229050623246509, + "learning_rate": 1.8969861308769025e-05, + "loss": 0.9371, + "step": 1579 + }, + { + "epoch": 0.8577633007600435, + "grad_norm": 5.081229057143735, + "learning_rate": 1.8968306104688365e-05, + "loss": 0.9398, + "step": 1580 + }, + { + "epoch": 0.8583061889250815, + "grad_norm": 4.641876540865143, + "learning_rate": 1.896674979139331e-05, + "loss": 1.1401, + "step": 1581 + }, + { + "epoch": 0.8588490770901195, + "grad_norm": 4.149586485324207, + "learning_rate": 1.8965192369076356e-05, + "loss": 0.6768, + "step": 1582 + }, + { + "epoch": 0.8593919652551575, + "grad_norm": 6.422772444923944, + "learning_rate": 1.8963633837930114e-05, + "loss": 1.392, + "step": 1583 + }, + { + "epoch": 0.8599348534201955, + "grad_norm": 4.860635992388028, + "learning_rate": 1.8962074198147357e-05, + "loss": 0.9983, + "step": 1584 + }, + { + "epoch": 0.8604777415852335, + "grad_norm": 4.216877618358879, + "learning_rate": 1.8960513449920982e-05, + "loss": 0.6958, + "step": 1585 + }, + { + "epoch": 0.8610206297502715, + "grad_norm": 5.012340070254638, + "learning_rate": 1.8958951593444017e-05, + "loss": 0.9859, + "step": 1586 + }, + { + "epoch": 0.8615635179153095, + "grad_norm": 5.467321056155664, + "learning_rate": 1.8957388628909644e-05, + "loss": 1.3782, + "step": 1587 + }, + { + "epoch": 0.8621064060803475, + "grad_norm": 5.59589685839463, + "learning_rate": 1.8955824556511168e-05, + "loss": 0.8297, + "step": 1588 + }, + { + "epoch": 0.8626492942453855, + "grad_norm": 6.519624998298618, + "learning_rate": 1.895425937644204e-05, + "loss": 1.7264, + "step": 1589 + }, + { + "epoch": 0.8631921824104235, + "grad_norm": 6.6951270307269795, + "learning_rate": 1.8952693088895837e-05, + "loss": 1.2919, + "step": 1590 + }, + { + "epoch": 0.8637350705754615, + "grad_norm": 4.348741825803296, + "learning_rate": 1.895112569406629e-05, + "loss": 1.0062, + "step": 1591 + }, + { + "epoch": 0.8642779587404995, + "grad_norm": 5.0213162819553565, + "learning_rate": 1.8949557192147243e-05, + "loss": 1.2381, + "step": 1592 + }, + { + "epoch": 0.8648208469055375, + "grad_norm": 6.1637021381056885, + "learning_rate": 1.8947987583332705e-05, + "loss": 1.2561, + "step": 1593 + }, + { + "epoch": 0.8653637350705755, + "grad_norm": 6.638257635344733, + "learning_rate": 1.89464168678168e-05, + "loss": 0.9861, + "step": 1594 + }, + { + "epoch": 0.8659066232356135, + "grad_norm": 4.861237127098975, + "learning_rate": 1.89448450457938e-05, + "loss": 0.9407, + "step": 1595 + }, + { + "epoch": 0.8664495114006515, + "grad_norm": 6.4248430682281565, + "learning_rate": 1.894327211745811e-05, + "loss": 1.3079, + "step": 1596 + }, + { + "epoch": 0.8669923995656895, + "grad_norm": 5.712502208347676, + "learning_rate": 1.8941698083004265e-05, + "loss": 1.4422, + "step": 1597 + }, + { + "epoch": 0.8675352877307275, + "grad_norm": 5.634308307822415, + "learning_rate": 1.8940122942626957e-05, + "loss": 0.7869, + "step": 1598 + }, + { + "epoch": 0.8680781758957655, + "grad_norm": 7.839535783401759, + "learning_rate": 1.893854669652099e-05, + "loss": 1.3397, + "step": 1599 + }, + { + "epoch": 0.8686210640608035, + "grad_norm": 5.156814072355032, + "learning_rate": 1.8936969344881323e-05, + "loss": 1.2541, + "step": 1600 + }, + { + "epoch": 0.8691639522258415, + "grad_norm": 4.912293322601371, + "learning_rate": 1.8935390887903044e-05, + "loss": 0.8418, + "step": 1601 + }, + { + "epoch": 0.8697068403908795, + "grad_norm": 5.436641391280079, + "learning_rate": 1.8933811325781382e-05, + "loss": 1.4456, + "step": 1602 + }, + { + "epoch": 0.8702497285559175, + "grad_norm": 5.221596044730403, + "learning_rate": 1.8932230658711696e-05, + "loss": 0.9626, + "step": 1603 + }, + { + "epoch": 0.8707926167209555, + "grad_norm": 6.103653494209826, + "learning_rate": 1.8930648886889482e-05, + "loss": 1.2338, + "step": 1604 + }, + { + "epoch": 0.8713355048859935, + "grad_norm": 7.003639368124227, + "learning_rate": 1.8929066010510383e-05, + "loss": 1.2216, + "step": 1605 + }, + { + "epoch": 0.8718783930510315, + "grad_norm": 4.911905632269173, + "learning_rate": 1.8927482029770168e-05, + "loss": 0.9049, + "step": 1606 + }, + { + "epoch": 0.8724212812160695, + "grad_norm": 6.079178298711795, + "learning_rate": 1.8925896944864748e-05, + "loss": 1.6408, + "step": 1607 + }, + { + "epoch": 0.8729641693811075, + "grad_norm": 6.045568168649525, + "learning_rate": 1.892431075599017e-05, + "loss": 1.084, + "step": 1608 + }, + { + "epoch": 0.8735070575461455, + "grad_norm": 4.827633470123435, + "learning_rate": 1.892272346334261e-05, + "loss": 0.8121, + "step": 1609 + }, + { + "epoch": 0.8740499457111836, + "grad_norm": 5.314535878915851, + "learning_rate": 1.8921135067118396e-05, + "loss": 1.2657, + "step": 1610 + }, + { + "epoch": 0.8745928338762216, + "grad_norm": 4.786606110240068, + "learning_rate": 1.8919545567513976e-05, + "loss": 1.0534, + "step": 1611 + }, + { + "epoch": 0.8751357220412594, + "grad_norm": 4.075700855026623, + "learning_rate": 1.8917954964725948e-05, + "loss": 0.7757, + "step": 1612 + }, + { + "epoch": 0.8756786102062974, + "grad_norm": 6.286980562376595, + "learning_rate": 1.8916363258951033e-05, + "loss": 0.9782, + "step": 1613 + }, + { + "epoch": 0.8762214983713354, + "grad_norm": 6.745303990081763, + "learning_rate": 1.8914770450386102e-05, + "loss": 0.9488, + "step": 1614 + }, + { + "epoch": 0.8767643865363735, + "grad_norm": 6.2663007227218275, + "learning_rate": 1.8913176539228152e-05, + "loss": 1.9529, + "step": 1615 + }, + { + "epoch": 0.8773072747014115, + "grad_norm": 4.793153881320079, + "learning_rate": 1.8911581525674324e-05, + "loss": 1.0242, + "step": 1616 + }, + { + "epoch": 0.8778501628664495, + "grad_norm": 4.6844946413894, + "learning_rate": 1.890998540992189e-05, + "loss": 0.8565, + "step": 1617 + }, + { + "epoch": 0.8783930510314875, + "grad_norm": 3.993737542716379, + "learning_rate": 1.8908388192168256e-05, + "loss": 0.6107, + "step": 1618 + }, + { + "epoch": 0.8789359391965255, + "grad_norm": 6.099718972513945, + "learning_rate": 1.8906789872610977e-05, + "loss": 0.8043, + "step": 1619 + }, + { + "epoch": 0.8794788273615635, + "grad_norm": 7.113418049135383, + "learning_rate": 1.8905190451447726e-05, + "loss": 1.241, + "step": 1620 + }, + { + "epoch": 0.8800217155266015, + "grad_norm": 4.439254236407197, + "learning_rate": 1.8903589928876337e-05, + "loss": 1.0627, + "step": 1621 + }, + { + "epoch": 0.8805646036916395, + "grad_norm": 5.773948255627926, + "learning_rate": 1.8901988305094746e-05, + "loss": 1.3241, + "step": 1622 + }, + { + "epoch": 0.8811074918566775, + "grad_norm": 6.231870498005159, + "learning_rate": 1.890038558030106e-05, + "loss": 1.5241, + "step": 1623 + }, + { + "epoch": 0.8816503800217155, + "grad_norm": 5.888612961801449, + "learning_rate": 1.8898781754693495e-05, + "loss": 1.109, + "step": 1624 + }, + { + "epoch": 0.8821932681867535, + "grad_norm": 4.880160750019388, + "learning_rate": 1.8897176828470424e-05, + "loss": 1.0124, + "step": 1625 + }, + { + "epoch": 0.8827361563517915, + "grad_norm": 6.365487465081988, + "learning_rate": 1.889557080183034e-05, + "loss": 0.8999, + "step": 1626 + }, + { + "epoch": 0.8832790445168295, + "grad_norm": 4.461166628483842, + "learning_rate": 1.8893963674971883e-05, + "loss": 0.8762, + "step": 1627 + }, + { + "epoch": 0.8838219326818675, + "grad_norm": 5.675449614336963, + "learning_rate": 1.8892355448093825e-05, + "loss": 1.049, + "step": 1628 + }, + { + "epoch": 0.8843648208469055, + "grad_norm": 4.320721965444944, + "learning_rate": 1.8890746121395072e-05, + "loss": 1.1291, + "step": 1629 + }, + { + "epoch": 0.8849077090119435, + "grad_norm": 5.155835351166163, + "learning_rate": 1.8889135695074668e-05, + "loss": 1.1035, + "step": 1630 + }, + { + "epoch": 0.8854505971769815, + "grad_norm": 6.506050711845877, + "learning_rate": 1.8887524169331794e-05, + "loss": 2.1954, + "step": 1631 + }, + { + "epoch": 0.8859934853420195, + "grad_norm": 5.514989348863087, + "learning_rate": 1.8885911544365766e-05, + "loss": 1.0237, + "step": 1632 + }, + { + "epoch": 0.8865363735070575, + "grad_norm": 5.123449587076818, + "learning_rate": 1.8884297820376038e-05, + "loss": 0.9908, + "step": 1633 + }, + { + "epoch": 0.8870792616720955, + "grad_norm": 5.421671567850239, + "learning_rate": 1.8882682997562197e-05, + "loss": 1.7734, + "step": 1634 + }, + { + "epoch": 0.8876221498371335, + "grad_norm": 7.1012436118787186, + "learning_rate": 1.8881067076123963e-05, + "loss": 1.5488, + "step": 1635 + }, + { + "epoch": 0.8881650380021715, + "grad_norm": 5.328780061143714, + "learning_rate": 1.88794500562612e-05, + "loss": 1.3832, + "step": 1636 + }, + { + "epoch": 0.8887079261672095, + "grad_norm": 5.1503624666971595, + "learning_rate": 1.88778319381739e-05, + "loss": 0.8524, + "step": 1637 + }, + { + "epoch": 0.8892508143322475, + "grad_norm": 5.131553278195334, + "learning_rate": 1.88762127220622e-05, + "loss": 0.9769, + "step": 1638 + }, + { + "epoch": 0.8897937024972855, + "grad_norm": 3.1519686008526135, + "learning_rate": 1.8874592408126365e-05, + "loss": 0.472, + "step": 1639 + }, + { + "epoch": 0.8903365906623235, + "grad_norm": 4.164586140054025, + "learning_rate": 1.8872970996566794e-05, + "loss": 0.67, + "step": 1640 + }, + { + "epoch": 0.8908794788273615, + "grad_norm": 7.488659246842809, + "learning_rate": 1.8871348487584028e-05, + "loss": 1.3141, + "step": 1641 + }, + { + "epoch": 0.8914223669923995, + "grad_norm": 5.911298710361154, + "learning_rate": 1.8869724881378743e-05, + "loss": 1.084, + "step": 1642 + }, + { + "epoch": 0.8919652551574375, + "grad_norm": 5.977793625957034, + "learning_rate": 1.886810017815175e-05, + "loss": 0.9273, + "step": 1643 + }, + { + "epoch": 0.8925081433224755, + "grad_norm": 5.2773273690434515, + "learning_rate": 1.8866474378103993e-05, + "loss": 1.2594, + "step": 1644 + }, + { + "epoch": 0.8930510314875135, + "grad_norm": 8.049487951903771, + "learning_rate": 1.8864847481436554e-05, + "loss": 1.0104, + "step": 1645 + }, + { + "epoch": 0.8935939196525515, + "grad_norm": 5.428394606766542, + "learning_rate": 1.886321948835065e-05, + "loss": 0.827, + "step": 1646 + }, + { + "epoch": 0.8941368078175895, + "grad_norm": 5.687133062399138, + "learning_rate": 1.8861590399047635e-05, + "loss": 1.2759, + "step": 1647 + }, + { + "epoch": 0.8946796959826275, + "grad_norm": 5.037591954761082, + "learning_rate": 1.885996021372899e-05, + "loss": 0.8928, + "step": 1648 + }, + { + "epoch": 0.8952225841476655, + "grad_norm": 4.911533421640023, + "learning_rate": 1.8858328932596352e-05, + "loss": 0.705, + "step": 1649 + }, + { + "epoch": 0.8957654723127035, + "grad_norm": 6.054608860222289, + "learning_rate": 1.885669655585147e-05, + "loss": 1.296, + "step": 1650 + }, + { + "epoch": 0.8963083604777415, + "grad_norm": 6.221099773279408, + "learning_rate": 1.8855063083696244e-05, + "loss": 1.5253, + "step": 1651 + }, + { + "epoch": 0.8968512486427795, + "grad_norm": 5.0718329796444035, + "learning_rate": 1.8853428516332702e-05, + "loss": 1.1683, + "step": 1652 + }, + { + "epoch": 0.8973941368078175, + "grad_norm": 5.721985007773526, + "learning_rate": 1.8851792853963015e-05, + "loss": 1.1635, + "step": 1653 + }, + { + "epoch": 0.8979370249728555, + "grad_norm": 7.224045168649458, + "learning_rate": 1.8850156096789473e-05, + "loss": 1.0042, + "step": 1654 + }, + { + "epoch": 0.8984799131378935, + "grad_norm": 4.938883702947416, + "learning_rate": 1.8848518245014526e-05, + "loss": 0.7712, + "step": 1655 + }, + { + "epoch": 0.8990228013029316, + "grad_norm": 5.9261619850877, + "learning_rate": 1.8846879298840735e-05, + "loss": 1.4387, + "step": 1656 + }, + { + "epoch": 0.8995656894679696, + "grad_norm": 5.181216185736725, + "learning_rate": 1.8845239258470817e-05, + "loss": 0.9389, + "step": 1657 + }, + { + "epoch": 0.9001085776330076, + "grad_norm": 5.108249236166185, + "learning_rate": 1.8843598124107608e-05, + "loss": 1.0034, + "step": 1658 + }, + { + "epoch": 0.9006514657980456, + "grad_norm": 4.839799276438571, + "learning_rate": 1.8841955895954088e-05, + "loss": 0.8524, + "step": 1659 + }, + { + "epoch": 0.9011943539630836, + "grad_norm": 5.598018889964705, + "learning_rate": 1.8840312574213372e-05, + "loss": 1.0939, + "step": 1660 + }, + { + "epoch": 0.9017372421281216, + "grad_norm": 5.2496226470868725, + "learning_rate": 1.8838668159088707e-05, + "loss": 1.0812, + "step": 1661 + }, + { + "epoch": 0.9022801302931596, + "grad_norm": 5.2298567821168, + "learning_rate": 1.8837022650783477e-05, + "loss": 0.953, + "step": 1662 + }, + { + "epoch": 0.9028230184581976, + "grad_norm": 4.471282412199049, + "learning_rate": 1.88353760495012e-05, + "loss": 1.2652, + "step": 1663 + }, + { + "epoch": 0.9033659066232356, + "grad_norm": 5.892082884595572, + "learning_rate": 1.8833728355445534e-05, + "loss": 1.1165, + "step": 1664 + }, + { + "epoch": 0.9039087947882736, + "grad_norm": 5.49174731306936, + "learning_rate": 1.8832079568820268e-05, + "loss": 0.9262, + "step": 1665 + }, + { + "epoch": 0.9044516829533116, + "grad_norm": 5.55211658654164, + "learning_rate": 1.883042968982932e-05, + "loss": 1.3112, + "step": 1666 + }, + { + "epoch": 0.9049945711183496, + "grad_norm": 5.597392580982551, + "learning_rate": 1.8828778718676757e-05, + "loss": 0.8511, + "step": 1667 + }, + { + "epoch": 0.9055374592833876, + "grad_norm": 6.183127994440369, + "learning_rate": 1.8827126655566773e-05, + "loss": 1.2784, + "step": 1668 + }, + { + "epoch": 0.9060803474484256, + "grad_norm": 6.991796130593696, + "learning_rate": 1.882547350070369e-05, + "loss": 1.1703, + "step": 1669 + }, + { + "epoch": 0.9066232356134636, + "grad_norm": 5.7794859741652305, + "learning_rate": 1.8823819254291986e-05, + "loss": 1.0974, + "step": 1670 + }, + { + "epoch": 0.9071661237785016, + "grad_norm": 5.388338879165358, + "learning_rate": 1.8822163916536245e-05, + "loss": 1.2087, + "step": 1671 + }, + { + "epoch": 0.9077090119435396, + "grad_norm": 5.072420136637763, + "learning_rate": 1.8820507487641218e-05, + "loss": 0.9244, + "step": 1672 + }, + { + "epoch": 0.9082519001085776, + "grad_norm": 5.9571835605705115, + "learning_rate": 1.8818849967811762e-05, + "loss": 1.3299, + "step": 1673 + }, + { + "epoch": 0.9087947882736156, + "grad_norm": 5.706573186001289, + "learning_rate": 1.8817191357252892e-05, + "loss": 1.1059, + "step": 1674 + }, + { + "epoch": 0.9093376764386536, + "grad_norm": 5.4984127833712435, + "learning_rate": 1.8815531656169737e-05, + "loss": 1.1428, + "step": 1675 + }, + { + "epoch": 0.9098805646036916, + "grad_norm": 4.824306323799433, + "learning_rate": 1.8813870864767582e-05, + "loss": 0.7313, + "step": 1676 + }, + { + "epoch": 0.9104234527687296, + "grad_norm": 6.180183255961245, + "learning_rate": 1.8812208983251828e-05, + "loss": 1.0729, + "step": 1677 + }, + { + "epoch": 0.9109663409337676, + "grad_norm": 6.586598617269493, + "learning_rate": 1.8810546011828024e-05, + "loss": 1.5871, + "step": 1678 + }, + { + "epoch": 0.9115092290988056, + "grad_norm": 5.5611281281711875, + "learning_rate": 1.8808881950701845e-05, + "loss": 0.8133, + "step": 1679 + }, + { + "epoch": 0.9120521172638436, + "grad_norm": 5.214863992160379, + "learning_rate": 1.8807216800079108e-05, + "loss": 1.0437, + "step": 1680 + }, + { + "epoch": 0.9125950054288816, + "grad_norm": 6.631045200232836, + "learning_rate": 1.8805550560165763e-05, + "loss": 1.446, + "step": 1681 + }, + { + "epoch": 0.9131378935939196, + "grad_norm": 6.140822200691373, + "learning_rate": 1.8803883231167887e-05, + "loss": 1.8672, + "step": 1682 + }, + { + "epoch": 0.9136807817589576, + "grad_norm": 4.5606387908479995, + "learning_rate": 1.8802214813291708e-05, + "loss": 1.0662, + "step": 1683 + }, + { + "epoch": 0.9142236699239956, + "grad_norm": 4.166651706834997, + "learning_rate": 1.8800545306743567e-05, + "loss": 0.749, + "step": 1684 + }, + { + "epoch": 0.9147665580890336, + "grad_norm": 4.674238905131916, + "learning_rate": 1.8798874711729957e-05, + "loss": 1.1106, + "step": 1685 + }, + { + "epoch": 0.9153094462540716, + "grad_norm": 5.224600466137879, + "learning_rate": 1.8797203028457497e-05, + "loss": 0.9652, + "step": 1686 + }, + { + "epoch": 0.9158523344191096, + "grad_norm": 5.87706899393542, + "learning_rate": 1.8795530257132947e-05, + "loss": 1.2146, + "step": 1687 + }, + { + "epoch": 0.9163952225841476, + "grad_norm": 7.587853042260599, + "learning_rate": 1.87938563979632e-05, + "loss": 1.9051, + "step": 1688 + }, + { + "epoch": 0.9169381107491856, + "grad_norm": 4.988334377581133, + "learning_rate": 1.8792181451155275e-05, + "loss": 0.8502, + "step": 1689 + }, + { + "epoch": 0.9174809989142236, + "grad_norm": 5.454638378956608, + "learning_rate": 1.8790505416916338e-05, + "loss": 1.0493, + "step": 1690 + }, + { + "epoch": 0.9180238870792616, + "grad_norm": 5.95672873256444, + "learning_rate": 1.878882829545368e-05, + "loss": 1.1296, + "step": 1691 + }, + { + "epoch": 0.9185667752442996, + "grad_norm": 6.921816321230723, + "learning_rate": 1.8787150086974734e-05, + "loss": 1.6388, + "step": 1692 + }, + { + "epoch": 0.9191096634093376, + "grad_norm": 5.423998305707574, + "learning_rate": 1.878547079168706e-05, + "loss": 1.4865, + "step": 1693 + }, + { + "epoch": 0.9196525515743756, + "grad_norm": 4.479238264802495, + "learning_rate": 1.878379040979835e-05, + "loss": 1.0257, + "step": 1694 + }, + { + "epoch": 0.9201954397394136, + "grad_norm": 6.325821606507086, + "learning_rate": 1.8782108941516446e-05, + "loss": 1.1156, + "step": 1695 + }, + { + "epoch": 0.9207383279044516, + "grad_norm": 6.220712298120886, + "learning_rate": 1.8780426387049315e-05, + "loss": 1.375, + "step": 1696 + }, + { + "epoch": 0.9212812160694897, + "grad_norm": 5.813906943099404, + "learning_rate": 1.877874274660505e-05, + "loss": 1.1869, + "step": 1697 + }, + { + "epoch": 0.9218241042345277, + "grad_norm": 5.0160254446208965, + "learning_rate": 1.8777058020391893e-05, + "loss": 1.0443, + "step": 1698 + }, + { + "epoch": 0.9223669923995657, + "grad_norm": 4.273981514908416, + "learning_rate": 1.877537220861821e-05, + "loss": 1.0015, + "step": 1699 + }, + { + "epoch": 0.9229098805646037, + "grad_norm": 4.908796260576538, + "learning_rate": 1.8773685311492513e-05, + "loss": 1.0203, + "step": 1700 + }, + { + "epoch": 0.9234527687296417, + "grad_norm": 5.566794452728763, + "learning_rate": 1.8771997329223425e-05, + "loss": 1.4637, + "step": 1701 + }, + { + "epoch": 0.9239956568946797, + "grad_norm": 4.946565856604512, + "learning_rate": 1.8770308262019733e-05, + "loss": 0.8093, + "step": 1702 + }, + { + "epoch": 0.9245385450597177, + "grad_norm": 4.767566465605265, + "learning_rate": 1.8768618110090334e-05, + "loss": 0.5938, + "step": 1703 + }, + { + "epoch": 0.9250814332247557, + "grad_norm": 6.919371263621627, + "learning_rate": 1.8766926873644272e-05, + "loss": 0.9857, + "step": 1704 + }, + { + "epoch": 0.9256243213897937, + "grad_norm": 3.5297488420709944, + "learning_rate": 1.876523455289072e-05, + "loss": 0.8744, + "step": 1705 + }, + { + "epoch": 0.9261672095548317, + "grad_norm": 4.910386600622229, + "learning_rate": 1.8763541148038994e-05, + "loss": 0.9491, + "step": 1706 + }, + { + "epoch": 0.9267100977198697, + "grad_norm": 5.550892558363115, + "learning_rate": 1.876184665929853e-05, + "loss": 1.2793, + "step": 1707 + }, + { + "epoch": 0.9272529858849077, + "grad_norm": 4.470721855980947, + "learning_rate": 1.8760151086878905e-05, + "loss": 0.9706, + "step": 1708 + }, + { + "epoch": 0.9277958740499457, + "grad_norm": 7.1893201073258926, + "learning_rate": 1.8758454430989833e-05, + "loss": 1.2473, + "step": 1709 + }, + { + "epoch": 0.9283387622149837, + "grad_norm": 5.933304434057961, + "learning_rate": 1.875675669184116e-05, + "loss": 1.4917, + "step": 1710 + }, + { + "epoch": 0.9288816503800217, + "grad_norm": 5.386119916834561, + "learning_rate": 1.8755057869642857e-05, + "loss": 0.9417, + "step": 1711 + }, + { + "epoch": 0.9294245385450597, + "grad_norm": 6.121437468126703, + "learning_rate": 1.875335796460505e-05, + "loss": 1.203, + "step": 1712 + }, + { + "epoch": 0.9299674267100977, + "grad_norm": 5.34957970840911, + "learning_rate": 1.8751656976937974e-05, + "loss": 1.2557, + "step": 1713 + }, + { + "epoch": 0.9305103148751357, + "grad_norm": 7.404283054108375, + "learning_rate": 1.8749954906852023e-05, + "loss": 1.449, + "step": 1714 + }, + { + "epoch": 0.9310532030401737, + "grad_norm": 6.5511839319022585, + "learning_rate": 1.8748251754557696e-05, + "loss": 1.5548, + "step": 1715 + }, + { + "epoch": 0.9315960912052117, + "grad_norm": 5.0254557279034815, + "learning_rate": 1.8746547520265654e-05, + "loss": 0.9777, + "step": 1716 + }, + { + "epoch": 0.9321389793702497, + "grad_norm": 6.385252198006408, + "learning_rate": 1.874484220418667e-05, + "loss": 1.254, + "step": 1717 + }, + { + "epoch": 0.9326818675352877, + "grad_norm": 6.872281478717196, + "learning_rate": 1.874313580653167e-05, + "loss": 1.278, + "step": 1718 + }, + { + "epoch": 0.9332247557003257, + "grad_norm": 8.525842687889295, + "learning_rate": 1.8741428327511696e-05, + "loss": 1.3712, + "step": 1719 + }, + { + "epoch": 0.9337676438653637, + "grad_norm": 7.97420765579221, + "learning_rate": 1.8739719767337933e-05, + "loss": 1.7493, + "step": 1720 + }, + { + "epoch": 0.9343105320304017, + "grad_norm": 6.107109526814546, + "learning_rate": 1.8738010126221705e-05, + "loss": 0.9861, + "step": 1721 + }, + { + "epoch": 0.9348534201954397, + "grad_norm": 6.4666849584489645, + "learning_rate": 1.8736299404374453e-05, + "loss": 0.8041, + "step": 1722 + }, + { + "epoch": 0.9353963083604777, + "grad_norm": 7.489590627219774, + "learning_rate": 1.873458760200777e-05, + "loss": 1.2325, + "step": 1723 + }, + { + "epoch": 0.9359391965255157, + "grad_norm": 7.020182643385788, + "learning_rate": 1.8732874719333373e-05, + "loss": 1.1698, + "step": 1724 + }, + { + "epoch": 0.9364820846905537, + "grad_norm": 6.436449840675483, + "learning_rate": 1.873116075656311e-05, + "loss": 1.084, + "step": 1725 + }, + { + "epoch": 0.9370249728555917, + "grad_norm": 7.27100690341834, + "learning_rate": 1.872944571390897e-05, + "loss": 0.9793, + "step": 1726 + }, + { + "epoch": 0.9375678610206297, + "grad_norm": 6.337617737677888, + "learning_rate": 1.872772959158307e-05, + "loss": 1.0558, + "step": 1727 + }, + { + "epoch": 0.9381107491856677, + "grad_norm": 5.356493873446841, + "learning_rate": 1.8726012389797667e-05, + "loss": 1.1031, + "step": 1728 + }, + { + "epoch": 0.9386536373507057, + "grad_norm": 6.302965546816682, + "learning_rate": 1.8724294108765142e-05, + "loss": 1.1401, + "step": 1729 + }, + { + "epoch": 0.9391965255157437, + "grad_norm": 6.116650787967334, + "learning_rate": 1.872257474869802e-05, + "loss": 1.0848, + "step": 1730 + }, + { + "epoch": 0.9397394136807817, + "grad_norm": 4.3874746967683, + "learning_rate": 1.8720854309808948e-05, + "loss": 0.7463, + "step": 1731 + }, + { + "epoch": 0.9402823018458197, + "grad_norm": 5.00673534942558, + "learning_rate": 1.871913279231072e-05, + "loss": 1.2365, + "step": 1732 + }, + { + "epoch": 0.9408251900108577, + "grad_norm": 5.222845312743783, + "learning_rate": 1.871741019641625e-05, + "loss": 1.062, + "step": 1733 + }, + { + "epoch": 0.9413680781758957, + "grad_norm": 4.624403224726164, + "learning_rate": 1.871568652233859e-05, + "loss": 1.0155, + "step": 1734 + }, + { + "epoch": 0.9419109663409337, + "grad_norm": 5.386412889478532, + "learning_rate": 1.8713961770290936e-05, + "loss": 1.4316, + "step": 1735 + }, + { + "epoch": 0.9424538545059717, + "grad_norm": 6.0899860043030705, + "learning_rate": 1.87122359404866e-05, + "loss": 1.1372, + "step": 1736 + }, + { + "epoch": 0.9429967426710097, + "grad_norm": 5.306362358042622, + "learning_rate": 1.8710509033139037e-05, + "loss": 0.9353, + "step": 1737 + }, + { + "epoch": 0.9435396308360477, + "grad_norm": 6.479420699028343, + "learning_rate": 1.8708781048461832e-05, + "loss": 1.2435, + "step": 1738 + }, + { + "epoch": 0.9440825190010858, + "grad_norm": 6.205746500302448, + "learning_rate": 1.8707051986668712e-05, + "loss": 0.7872, + "step": 1739 + }, + { + "epoch": 0.9446254071661238, + "grad_norm": 4.157411034856641, + "learning_rate": 1.8705321847973523e-05, + "loss": 0.8004, + "step": 1740 + }, + { + "epoch": 0.9451682953311618, + "grad_norm": 5.5814599600578285, + "learning_rate": 1.8703590632590254e-05, + "loss": 0.7384, + "step": 1741 + }, + { + "epoch": 0.9457111834961998, + "grad_norm": 6.040759722497843, + "learning_rate": 1.8701858340733023e-05, + "loss": 1.5941, + "step": 1742 + }, + { + "epoch": 0.9462540716612378, + "grad_norm": 4.592040464655725, + "learning_rate": 1.8700124972616085e-05, + "loss": 1.0662, + "step": 1743 + }, + { + "epoch": 0.9467969598262758, + "grad_norm": 6.077319934267432, + "learning_rate": 1.8698390528453823e-05, + "loss": 1.3726, + "step": 1744 + }, + { + "epoch": 0.9473398479913138, + "grad_norm": 4.255180894541495, + "learning_rate": 1.869665500846076e-05, + "loss": 0.7001, + "step": 1745 + }, + { + "epoch": 0.9478827361563518, + "grad_norm": 6.390593673413393, + "learning_rate": 1.869491841285154e-05, + "loss": 1.2268, + "step": 1746 + }, + { + "epoch": 0.9484256243213898, + "grad_norm": 5.347324129941674, + "learning_rate": 1.8693180741840957e-05, + "loss": 1.2193, + "step": 1747 + }, + { + "epoch": 0.9489685124864278, + "grad_norm": 6.407261723862006, + "learning_rate": 1.8691441995643927e-05, + "loss": 0.868, + "step": 1748 + }, + { + "epoch": 0.9495114006514658, + "grad_norm": 4.400520878629861, + "learning_rate": 1.8689702174475496e-05, + "loss": 0.4679, + "step": 1749 + }, + { + "epoch": 0.9500542888165038, + "grad_norm": 5.6161548427455505, + "learning_rate": 1.8687961278550852e-05, + "loss": 0.8842, + "step": 1750 + }, + { + "epoch": 0.9505971769815418, + "grad_norm": 4.7163538240048, + "learning_rate": 1.8686219308085306e-05, + "loss": 0.9972, + "step": 1751 + }, + { + "epoch": 0.9511400651465798, + "grad_norm": 4.845854095585142, + "learning_rate": 1.8684476263294318e-05, + "loss": 1.1046, + "step": 1752 + }, + { + "epoch": 0.9516829533116178, + "grad_norm": 3.878984502983999, + "learning_rate": 1.8682732144393463e-05, + "loss": 0.7873, + "step": 1753 + }, + { + "epoch": 0.9522258414766558, + "grad_norm": 5.736472248757515, + "learning_rate": 1.8680986951598458e-05, + "loss": 1.2046, + "step": 1754 + }, + { + "epoch": 0.9527687296416938, + "grad_norm": 5.12988223940441, + "learning_rate": 1.867924068512515e-05, + "loss": 0.6293, + "step": 1755 + }, + { + "epoch": 0.9533116178067318, + "grad_norm": 7.822467968073818, + "learning_rate": 1.867749334518952e-05, + "loss": 1.3625, + "step": 1756 + }, + { + "epoch": 0.9538545059717698, + "grad_norm": 5.488388947029871, + "learning_rate": 1.8675744932007687e-05, + "loss": 1.078, + "step": 1757 + }, + { + "epoch": 0.9543973941368078, + "grad_norm": 4.631848438676548, + "learning_rate": 1.8673995445795894e-05, + "loss": 0.69, + "step": 1758 + }, + { + "epoch": 0.9549402823018458, + "grad_norm": 7.078171019993124, + "learning_rate": 1.8672244886770516e-05, + "loss": 1.1036, + "step": 1759 + }, + { + "epoch": 0.9554831704668838, + "grad_norm": 5.9259972832349215, + "learning_rate": 1.8670493255148073e-05, + "loss": 0.7919, + "step": 1760 + }, + { + "epoch": 0.9560260586319218, + "grad_norm": 5.846343833877514, + "learning_rate": 1.8668740551145205e-05, + "loss": 1.1653, + "step": 1761 + }, + { + "epoch": 0.9565689467969598, + "grad_norm": 5.559082125902064, + "learning_rate": 1.8666986774978685e-05, + "loss": 1.4214, + "step": 1762 + }, + { + "epoch": 0.9571118349619978, + "grad_norm": 6.511207981127819, + "learning_rate": 1.8665231926865433e-05, + "loss": 1.3552, + "step": 1763 + }, + { + "epoch": 0.9576547231270358, + "grad_norm": 5.728999474717542, + "learning_rate": 1.8663476007022482e-05, + "loss": 0.7861, + "step": 1764 + }, + { + "epoch": 0.9581976112920738, + "grad_norm": 6.152118564149309, + "learning_rate": 1.8661719015667016e-05, + "loss": 0.8679, + "step": 1765 + }, + { + "epoch": 0.9587404994571118, + "grad_norm": 6.77121947972117, + "learning_rate": 1.8659960953016334e-05, + "loss": 0.8103, + "step": 1766 + }, + { + "epoch": 0.9592833876221498, + "grad_norm": 7.822575173268641, + "learning_rate": 1.865820181928788e-05, + "loss": 1.2558, + "step": 1767 + }, + { + "epoch": 0.9598262757871878, + "grad_norm": 6.389073592671253, + "learning_rate": 1.8656441614699225e-05, + "loss": 1.0314, + "step": 1768 + }, + { + "epoch": 0.9603691639522258, + "grad_norm": 6.141689972496849, + "learning_rate": 1.8654680339468076e-05, + "loss": 1.2452, + "step": 1769 + }, + { + "epoch": 0.9609120521172638, + "grad_norm": 6.934092850250001, + "learning_rate": 1.8652917993812267e-05, + "loss": 1.1186, + "step": 1770 + }, + { + "epoch": 0.9614549402823018, + "grad_norm": 6.129320663785202, + "learning_rate": 1.865115457794977e-05, + "loss": 0.9076, + "step": 1771 + }, + { + "epoch": 0.9619978284473398, + "grad_norm": 7.008261896063745, + "learning_rate": 1.8649390092098693e-05, + "loss": 1.7152, + "step": 1772 + }, + { + "epoch": 0.9625407166123778, + "grad_norm": 6.36833090372477, + "learning_rate": 1.8647624536477255e-05, + "loss": 0.8087, + "step": 1773 + }, + { + "epoch": 0.9630836047774158, + "grad_norm": 4.7595903392838865, + "learning_rate": 1.8645857911303838e-05, + "loss": 0.6374, + "step": 1774 + }, + { + "epoch": 0.9636264929424538, + "grad_norm": 6.410325139984648, + "learning_rate": 1.8644090216796934e-05, + "loss": 1.3611, + "step": 1775 + }, + { + "epoch": 0.9641693811074918, + "grad_norm": 5.900212705071564, + "learning_rate": 1.8642321453175177e-05, + "loss": 0.891, + "step": 1776 + }, + { + "epoch": 0.9647122692725298, + "grad_norm": 6.90562183471938, + "learning_rate": 1.8640551620657326e-05, + "loss": 0.7899, + "step": 1777 + }, + { + "epoch": 0.9652551574375678, + "grad_norm": 6.476066490933977, + "learning_rate": 1.8638780719462278e-05, + "loss": 0.9614, + "step": 1778 + }, + { + "epoch": 0.9657980456026058, + "grad_norm": 6.659586852509603, + "learning_rate": 1.8637008749809065e-05, + "loss": 1.2419, + "step": 1779 + }, + { + "epoch": 0.9663409337676439, + "grad_norm": 6.587754652016677, + "learning_rate": 1.8635235711916847e-05, + "loss": 1.3664, + "step": 1780 + }, + { + "epoch": 0.9668838219326819, + "grad_norm": 6.558578755813245, + "learning_rate": 1.863346160600491e-05, + "loss": 1.1249, + "step": 1781 + }, + { + "epoch": 0.9674267100977199, + "grad_norm": 4.791301716076818, + "learning_rate": 1.8631686432292685e-05, + "loss": 0.6836, + "step": 1782 + }, + { + "epoch": 0.9679695982627579, + "grad_norm": 5.120778292537024, + "learning_rate": 1.862991019099972e-05, + "loss": 0.592, + "step": 1783 + }, + { + "epoch": 0.9685124864277959, + "grad_norm": 8.250755784410368, + "learning_rate": 1.8628132882345713e-05, + "loss": 1.2579, + "step": 1784 + }, + { + "epoch": 0.9690553745928339, + "grad_norm": 5.811263926579046, + "learning_rate": 1.862635450655048e-05, + "loss": 1.2935, + "step": 1785 + }, + { + "epoch": 0.9695982627578719, + "grad_norm": 6.042040633945999, + "learning_rate": 1.862457506383397e-05, + "loss": 0.8784, + "step": 1786 + }, + { + "epoch": 0.9701411509229099, + "grad_norm": 7.115011060082062, + "learning_rate": 1.8622794554416272e-05, + "loss": 1.0717, + "step": 1787 + }, + { + "epoch": 0.9706840390879479, + "grad_norm": 7.520775839533288, + "learning_rate": 1.8621012978517604e-05, + "loss": 1.3468, + "step": 1788 + }, + { + "epoch": 0.9712269272529859, + "grad_norm": 6.949921961354188, + "learning_rate": 1.8619230336358306e-05, + "loss": 1.5609, + "step": 1789 + }, + { + "epoch": 0.9717698154180239, + "grad_norm": 6.396218455762339, + "learning_rate": 1.8617446628158866e-05, + "loss": 1.1807, + "step": 1790 + }, + { + "epoch": 0.9723127035830619, + "grad_norm": 6.314794279037124, + "learning_rate": 1.861566185413989e-05, + "loss": 0.9255, + "step": 1791 + }, + { + "epoch": 0.9728555917480999, + "grad_norm": 5.393752442926871, + "learning_rate": 1.8613876014522128e-05, + "loss": 0.7926, + "step": 1792 + }, + { + "epoch": 0.9733984799131379, + "grad_norm": 5.908313693688933, + "learning_rate": 1.8612089109526453e-05, + "loss": 0.7984, + "step": 1793 + }, + { + "epoch": 0.9739413680781759, + "grad_norm": 4.6952509401002125, + "learning_rate": 1.8610301139373867e-05, + "loss": 1.0344, + "step": 1794 + }, + { + "epoch": 0.9744842562432139, + "grad_norm": 6.81835335105175, + "learning_rate": 1.8608512104285517e-05, + "loss": 1.1532, + "step": 1795 + }, + { + "epoch": 0.9750271444082519, + "grad_norm": 4.159411114912345, + "learning_rate": 1.860672200448267e-05, + "loss": 1.0665, + "step": 1796 + }, + { + "epoch": 0.9755700325732899, + "grad_norm": 4.911404446825004, + "learning_rate": 1.8604930840186726e-05, + "loss": 1.1784, + "step": 1797 + }, + { + "epoch": 0.9761129207383279, + "grad_norm": 5.399453855489055, + "learning_rate": 1.860313861161922e-05, + "loss": 0.8215, + "step": 1798 + }, + { + "epoch": 0.9766558089033659, + "grad_norm": 4.797137239148531, + "learning_rate": 1.860134531900182e-05, + "loss": 0.6405, + "step": 1799 + }, + { + "epoch": 0.9771986970684039, + "grad_norm": 3.937076316066272, + "learning_rate": 1.859955096255633e-05, + "loss": 0.5912, + "step": 1800 + }, + { + "epoch": 0.9777415852334419, + "grad_norm": 5.569110777302129, + "learning_rate": 1.859775554250466e-05, + "loss": 0.8089, + "step": 1801 + }, + { + "epoch": 0.9782844733984799, + "grad_norm": 6.364258901894282, + "learning_rate": 1.859595905906889e-05, + "loss": 1.5181, + "step": 1802 + }, + { + "epoch": 0.9788273615635179, + "grad_norm": 4.995224609073908, + "learning_rate": 1.85941615124712e-05, + "loss": 0.8573, + "step": 1803 + }, + { + "epoch": 0.9793702497285559, + "grad_norm": 5.041992740138782, + "learning_rate": 1.8592362902933918e-05, + "loss": 0.8982, + "step": 1804 + }, + { + "epoch": 0.9799131378935939, + "grad_norm": 5.6054574490953275, + "learning_rate": 1.8590563230679496e-05, + "loss": 0.9285, + "step": 1805 + }, + { + "epoch": 0.9804560260586319, + "grad_norm": 5.269503319178264, + "learning_rate": 1.8588762495930526e-05, + "loss": 1.0963, + "step": 1806 + }, + { + "epoch": 0.9809989142236699, + "grad_norm": 6.184787731593613, + "learning_rate": 1.8586960698909718e-05, + "loss": 1.4005, + "step": 1807 + }, + { + "epoch": 0.9815418023887079, + "grad_norm": 8.50824591479003, + "learning_rate": 1.858515783983993e-05, + "loss": 1.9607, + "step": 1808 + }, + { + "epoch": 0.9820846905537459, + "grad_norm": 5.399464732589782, + "learning_rate": 1.8583353918944134e-05, + "loss": 0.951, + "step": 1809 + }, + { + "epoch": 0.9826275787187839, + "grad_norm": 6.577634699737779, + "learning_rate": 1.8581548936445447e-05, + "loss": 1.5139, + "step": 1810 + }, + { + "epoch": 0.9831704668838219, + "grad_norm": 10.183489163585788, + "learning_rate": 1.8579742892567107e-05, + "loss": 2.0539, + "step": 1811 + }, + { + "epoch": 0.9837133550488599, + "grad_norm": 6.578605968943458, + "learning_rate": 1.8577935787532494e-05, + "loss": 1.1454, + "step": 1812 + }, + { + "epoch": 0.9842562432138979, + "grad_norm": 4.923285404215955, + "learning_rate": 1.8576127621565113e-05, + "loss": 0.8965, + "step": 1813 + }, + { + "epoch": 0.9847991313789359, + "grad_norm": 5.912445932400008, + "learning_rate": 1.85743183948886e-05, + "loss": 1.1311, + "step": 1814 + }, + { + "epoch": 0.9853420195439739, + "grad_norm": 4.5183231221646105, + "learning_rate": 1.8572508107726725e-05, + "loss": 0.7511, + "step": 1815 + }, + { + "epoch": 0.9858849077090119, + "grad_norm": 5.2218543332097935, + "learning_rate": 1.8570696760303378e-05, + "loss": 1.0063, + "step": 1816 + }, + { + "epoch": 0.9864277958740499, + "grad_norm": 7.215748803707923, + "learning_rate": 1.85688843528426e-05, + "loss": 1.3674, + "step": 1817 + }, + { + "epoch": 0.9869706840390879, + "grad_norm": 6.362354074709609, + "learning_rate": 1.8567070885568547e-05, + "loss": 1.1319, + "step": 1818 + }, + { + "epoch": 0.987513572204126, + "grad_norm": 4.680755706450828, + "learning_rate": 1.8565256358705513e-05, + "loss": 0.6613, + "step": 1819 + }, + { + "epoch": 0.988056460369164, + "grad_norm": 4.80241945107321, + "learning_rate": 1.8563440772477922e-05, + "loss": 0.8619, + "step": 1820 + }, + { + "epoch": 0.988599348534202, + "grad_norm": 5.858709563059296, + "learning_rate": 1.856162412711033e-05, + "loss": 0.9432, + "step": 1821 + }, + { + "epoch": 0.98914223669924, + "grad_norm": 6.478119299588854, + "learning_rate": 1.855980642282742e-05, + "loss": 1.1866, + "step": 1822 + }, + { + "epoch": 0.989685124864278, + "grad_norm": 7.081924554034699, + "learning_rate": 1.8557987659854006e-05, + "loss": 1.3657, + "step": 1823 + }, + { + "epoch": 0.990228013029316, + "grad_norm": 5.404410220733361, + "learning_rate": 1.855616783841504e-05, + "loss": 1.1051, + "step": 1824 + }, + { + "epoch": 0.990770901194354, + "grad_norm": 6.742207962837618, + "learning_rate": 1.8554346958735602e-05, + "loss": 1.4667, + "step": 1825 + }, + { + "epoch": 0.991313789359392, + "grad_norm": 6.164178667211635, + "learning_rate": 1.8552525021040895e-05, + "loss": 0.7861, + "step": 1826 + }, + { + "epoch": 0.99185667752443, + "grad_norm": 6.078017237490637, + "learning_rate": 1.8550702025556265e-05, + "loss": 1.3294, + "step": 1827 + }, + { + "epoch": 0.992399565689468, + "grad_norm": 4.7372949568909455, + "learning_rate": 1.8548877972507182e-05, + "loss": 0.9779, + "step": 1828 + }, + { + "epoch": 0.992942453854506, + "grad_norm": 5.25971277777658, + "learning_rate": 1.8547052862119247e-05, + "loss": 1.049, + "step": 1829 + }, + { + "epoch": 0.993485342019544, + "grad_norm": 4.950709970125391, + "learning_rate": 1.854522669461819e-05, + "loss": 0.6665, + "step": 1830 + }, + { + "epoch": 0.994028230184582, + "grad_norm": 5.856428194694733, + "learning_rate": 1.8543399470229876e-05, + "loss": 1.0954, + "step": 1831 + }, + { + "epoch": 0.99457111834962, + "grad_norm": 4.895286613591861, + "learning_rate": 1.85415711891803e-05, + "loss": 0.7069, + "step": 1832 + }, + { + "epoch": 0.995114006514658, + "grad_norm": 5.249228502887773, + "learning_rate": 1.8539741851695586e-05, + "loss": 1.0655, + "step": 1833 + }, + { + "epoch": 0.995656894679696, + "grad_norm": 4.855694799243912, + "learning_rate": 1.8537911458001988e-05, + "loss": 1.1872, + "step": 1834 + }, + { + "epoch": 0.996199782844734, + "grad_norm": 5.180821237913711, + "learning_rate": 1.8536080008325896e-05, + "loss": 0.8281, + "step": 1835 + }, + { + "epoch": 0.996742671009772, + "grad_norm": 5.073254295771654, + "learning_rate": 1.8534247502893823e-05, + "loss": 0.7002, + "step": 1836 + }, + { + "epoch": 0.99728555917481, + "grad_norm": 5.4022604097623255, + "learning_rate": 1.8532413941932416e-05, + "loss": 1.1793, + "step": 1837 + }, + { + "epoch": 0.997828447339848, + "grad_norm": 4.981761324197401, + "learning_rate": 1.8530579325668455e-05, + "loss": 0.8331, + "step": 1838 + }, + { + "epoch": 0.998371335504886, + "grad_norm": 7.325363733761658, + "learning_rate": 1.852874365432885e-05, + "loss": 1.3071, + "step": 1839 + }, + { + "epoch": 0.998914223669924, + "grad_norm": 7.5668251202705505, + "learning_rate": 1.852690692814063e-05, + "loss": 1.4955, + "step": 1840 + }, + { + "epoch": 0.999457111834962, + "grad_norm": 6.706936559053067, + "learning_rate": 1.8525069147330978e-05, + "loss": 1.0286, + "step": 1841 + }, + { + "epoch": 1.0, + "grad_norm": 6.338369283394405, + "learning_rate": 1.8523230312127183e-05, + "loss": 1.0509, + "step": 1842 + }, + { + "epoch": 1.000542888165038, + "grad_norm": 6.119502670416093, + "learning_rate": 1.8521390422756683e-05, + "loss": 1.0805, + "step": 1843 + }, + { + "epoch": 1.001085776330076, + "grad_norm": 5.074420967885639, + "learning_rate": 1.851954947944703e-05, + "loss": 1.0931, + "step": 1844 + }, + { + "epoch": 1.001628664495114, + "grad_norm": 8.194067497007946, + "learning_rate": 1.851770748242592e-05, + "loss": 2.0732, + "step": 1845 + }, + { + "epoch": 1.002171552660152, + "grad_norm": 4.389522929363591, + "learning_rate": 1.8515864431921177e-05, + "loss": 0.8684, + "step": 1846 + }, + { + "epoch": 1.00271444082519, + "grad_norm": 5.720373954211297, + "learning_rate": 1.8514020328160748e-05, + "loss": 1.0223, + "step": 1847 + }, + { + "epoch": 1.003257328990228, + "grad_norm": 5.163678744409467, + "learning_rate": 1.8512175171372713e-05, + "loss": 0.9611, + "step": 1848 + }, + { + "epoch": 1.003800217155266, + "grad_norm": 5.786056891749039, + "learning_rate": 1.8510328961785286e-05, + "loss": 1.3473, + "step": 1849 + }, + { + "epoch": 1.004343105320304, + "grad_norm": 5.9194250471944985, + "learning_rate": 1.850848169962681e-05, + "loss": 0.8218, + "step": 1850 + }, + { + "epoch": 1.004885993485342, + "grad_norm": 5.890241796399565, + "learning_rate": 1.850663338512576e-05, + "loss": 0.842, + "step": 1851 + }, + { + "epoch": 1.00542888165038, + "grad_norm": 6.939341519595904, + "learning_rate": 1.8504784018510732e-05, + "loss": 0.9263, + "step": 1852 + }, + { + "epoch": 1.005971769815418, + "grad_norm": 4.592014594714093, + "learning_rate": 1.850293360001046e-05, + "loss": 1.015, + "step": 1853 + }, + { + "epoch": 1.006514657980456, + "grad_norm": 4.575145990552055, + "learning_rate": 1.8501082129853816e-05, + "loss": 1.0505, + "step": 1854 + }, + { + "epoch": 1.007057546145494, + "grad_norm": 5.999218616951103, + "learning_rate": 1.849922960826978e-05, + "loss": 1.0945, + "step": 1855 + }, + { + "epoch": 1.007600434310532, + "grad_norm": 5.459227552994586, + "learning_rate": 1.8497376035487483e-05, + "loss": 1.0167, + "step": 1856 + }, + { + "epoch": 1.00814332247557, + "grad_norm": 4.785837605715454, + "learning_rate": 1.8495521411736173e-05, + "loss": 1.0559, + "step": 1857 + }, + { + "epoch": 1.008686210640608, + "grad_norm": 5.813312187108195, + "learning_rate": 1.8493665737245236e-05, + "loss": 1.3477, + "step": 1858 + }, + { + "epoch": 1.009229098805646, + "grad_norm": 6.512945450475729, + "learning_rate": 1.8491809012244182e-05, + "loss": 1.4416, + "step": 1859 + }, + { + "epoch": 1.009771986970684, + "grad_norm": 5.599962385405486, + "learning_rate": 1.8489951236962658e-05, + "loss": 0.8961, + "step": 1860 + }, + { + "epoch": 1.010314875135722, + "grad_norm": 4.863989016796961, + "learning_rate": 1.848809241163043e-05, + "loss": 0.639, + "step": 1861 + }, + { + "epoch": 1.01085776330076, + "grad_norm": 5.937471136238718, + "learning_rate": 1.848623253647741e-05, + "loss": 1.0548, + "step": 1862 + }, + { + "epoch": 1.011400651465798, + "grad_norm": 4.75172125273428, + "learning_rate": 1.8484371611733625e-05, + "loss": 0.6493, + "step": 1863 + }, + { + "epoch": 1.011943539630836, + "grad_norm": 5.867263430143688, + "learning_rate": 1.848250963762923e-05, + "loss": 0.8569, + "step": 1864 + }, + { + "epoch": 1.012486427795874, + "grad_norm": 5.932410925593062, + "learning_rate": 1.848064661439453e-05, + "loss": 0.8351, + "step": 1865 + }, + { + "epoch": 1.013029315960912, + "grad_norm": 5.8320858990674, + "learning_rate": 1.847878254225994e-05, + "loss": 0.8916, + "step": 1866 + }, + { + "epoch": 1.01357220412595, + "grad_norm": 5.836721791741835, + "learning_rate": 1.847691742145601e-05, + "loss": 1.1742, + "step": 1867 + }, + { + "epoch": 1.014115092290988, + "grad_norm": 6.806668722890977, + "learning_rate": 1.8475051252213423e-05, + "loss": 1.0714, + "step": 1868 + }, + { + "epoch": 1.014657980456026, + "grad_norm": 7.026347641082525, + "learning_rate": 1.8473184034762992e-05, + "loss": 1.4101, + "step": 1869 + }, + { + "epoch": 1.015200868621064, + "grad_norm": 5.9532414283206005, + "learning_rate": 1.8471315769335657e-05, + "loss": 0.7855, + "step": 1870 + }, + { + "epoch": 1.015743756786102, + "grad_norm": 5.7578706206451855, + "learning_rate": 1.846944645616248e-05, + "loss": 0.7644, + "step": 1871 + }, + { + "epoch": 1.01628664495114, + "grad_norm": 6.340541706204924, + "learning_rate": 1.846757609547467e-05, + "loss": 1.1607, + "step": 1872 + }, + { + "epoch": 1.016829533116178, + "grad_norm": 4.942991579941027, + "learning_rate": 1.8465704687503558e-05, + "loss": 0.7578, + "step": 1873 + }, + { + "epoch": 1.017372421281216, + "grad_norm": 6.42121491667965, + "learning_rate": 1.846383223248059e-05, + "loss": 0.9924, + "step": 1874 + }, + { + "epoch": 1.017915309446254, + "grad_norm": 6.252498990648505, + "learning_rate": 1.8461958730637368e-05, + "loss": 0.811, + "step": 1875 + }, + { + "epoch": 1.018458197611292, + "grad_norm": 5.941336187529878, + "learning_rate": 1.84600841822056e-05, + "loss": 0.772, + "step": 1876 + }, + { + "epoch": 1.01900108577633, + "grad_norm": 5.323895762793836, + "learning_rate": 1.845820858741714e-05, + "loss": 0.8812, + "step": 1877 + }, + { + "epoch": 1.019543973941368, + "grad_norm": 7.2574721612646105, + "learning_rate": 1.845633194650396e-05, + "loss": 1.2759, + "step": 1878 + }, + { + "epoch": 1.020086862106406, + "grad_norm": 4.307176114235153, + "learning_rate": 1.8454454259698165e-05, + "loss": 0.569, + "step": 1879 + }, + { + "epoch": 1.020629750271444, + "grad_norm": 6.504543473682945, + "learning_rate": 1.8452575527231997e-05, + "loss": 0.9355, + "step": 1880 + }, + { + "epoch": 1.021172638436482, + "grad_norm": 5.822065969327671, + "learning_rate": 1.8450695749337816e-05, + "loss": 1.0287, + "step": 1881 + }, + { + "epoch": 1.02171552660152, + "grad_norm": 8.90668319984316, + "learning_rate": 1.8448814926248112e-05, + "loss": 1.5348, + "step": 1882 + }, + { + "epoch": 1.022258414766558, + "grad_norm": 5.233217719773852, + "learning_rate": 1.844693305819552e-05, + "loss": 0.564, + "step": 1883 + }, + { + "epoch": 1.022801302931596, + "grad_norm": 5.806961114054563, + "learning_rate": 1.844505014541278e-05, + "loss": 0.9349, + "step": 1884 + }, + { + "epoch": 1.023344191096634, + "grad_norm": 6.239514346807307, + "learning_rate": 1.8443166188132777e-05, + "loss": 0.9677, + "step": 1885 + }, + { + "epoch": 1.023887079261672, + "grad_norm": 5.596402826374599, + "learning_rate": 1.8441281186588528e-05, + "loss": 0.695, + "step": 1886 + }, + { + "epoch": 1.02442996742671, + "grad_norm": 6.531429104200087, + "learning_rate": 1.8439395141013165e-05, + "loss": 1.1543, + "step": 1887 + }, + { + "epoch": 1.024972855591748, + "grad_norm": 5.129461327017003, + "learning_rate": 1.843750805163996e-05, + "loss": 1.2624, + "step": 1888 + }, + { + "epoch": 1.0255157437567861, + "grad_norm": 6.862993011679228, + "learning_rate": 1.8435619918702318e-05, + "loss": 1.2673, + "step": 1889 + }, + { + "epoch": 1.0260586319218241, + "grad_norm": 5.134671966877655, + "learning_rate": 1.8433730742433755e-05, + "loss": 0.6032, + "step": 1890 + }, + { + "epoch": 1.0266015200868621, + "grad_norm": 6.9199078267443195, + "learning_rate": 1.8431840523067932e-05, + "loss": 1.1255, + "step": 1891 + }, + { + "epoch": 1.0271444082519001, + "grad_norm": 5.808796921918892, + "learning_rate": 1.8429949260838635e-05, + "loss": 0.8488, + "step": 1892 + }, + { + "epoch": 1.0276872964169381, + "grad_norm": 7.102540931353532, + "learning_rate": 1.842805695597978e-05, + "loss": 1.0634, + "step": 1893 + }, + { + "epoch": 1.0282301845819761, + "grad_norm": 9.932212718124791, + "learning_rate": 1.8426163608725403e-05, + "loss": 1.3909, + "step": 1894 + }, + { + "epoch": 1.0287730727470141, + "grad_norm": 6.999752698080143, + "learning_rate": 1.8424269219309686e-05, + "loss": 1.1106, + "step": 1895 + }, + { + "epoch": 1.0293159609120521, + "grad_norm": 6.569873028691542, + "learning_rate": 1.842237378796693e-05, + "loss": 0.9455, + "step": 1896 + }, + { + "epoch": 1.0298588490770901, + "grad_norm": 6.418746090533204, + "learning_rate": 1.8420477314931554e-05, + "loss": 0.934, + "step": 1897 + }, + { + "epoch": 1.0304017372421281, + "grad_norm": 7.7235490740904815, + "learning_rate": 1.8418579800438125e-05, + "loss": 1.1582, + "step": 1898 + }, + { + "epoch": 1.0309446254071661, + "grad_norm": 5.4350436368749895, + "learning_rate": 1.841668124472133e-05, + "loss": 0.7726, + "step": 1899 + }, + { + "epoch": 1.0314875135722041, + "grad_norm": 5.34390520812136, + "learning_rate": 1.8414781648015983e-05, + "loss": 0.7497, + "step": 1900 + }, + { + "epoch": 1.0320304017372421, + "grad_norm": 6.563688266956639, + "learning_rate": 1.841288101055703e-05, + "loss": 1.3352, + "step": 1901 + }, + { + "epoch": 1.0325732899022801, + "grad_norm": 7.710283890330315, + "learning_rate": 1.841097933257955e-05, + "loss": 0.8804, + "step": 1902 + }, + { + "epoch": 1.0331161780673181, + "grad_norm": 5.31500431944041, + "learning_rate": 1.840907661431874e-05, + "loss": 0.8667, + "step": 1903 + }, + { + "epoch": 1.0336590662323561, + "grad_norm": 7.476982099026401, + "learning_rate": 1.840717285600993e-05, + "loss": 0.8317, + "step": 1904 + }, + { + "epoch": 1.0342019543973942, + "grad_norm": 5.1563433597221415, + "learning_rate": 1.840526805788858e-05, + "loss": 1.0108, + "step": 1905 + }, + { + "epoch": 1.0347448425624322, + "grad_norm": 6.411832164370409, + "learning_rate": 1.8403362220190284e-05, + "loss": 0.7778, + "step": 1906 + }, + { + "epoch": 1.0352877307274702, + "grad_norm": 6.688312152010462, + "learning_rate": 1.8401455343150757e-05, + "loss": 1.2438, + "step": 1907 + }, + { + "epoch": 1.0358306188925082, + "grad_norm": 6.412659783246144, + "learning_rate": 1.839954742700584e-05, + "loss": 0.8191, + "step": 1908 + }, + { + "epoch": 1.0363735070575462, + "grad_norm": 6.399618830323376, + "learning_rate": 1.839763847199151e-05, + "loss": 1.4485, + "step": 1909 + }, + { + "epoch": 1.0369163952225842, + "grad_norm": 8.041718562623554, + "learning_rate": 1.8395728478343873e-05, + "loss": 1.0043, + "step": 1910 + }, + { + "epoch": 1.0374592833876222, + "grad_norm": 6.4118854777282905, + "learning_rate": 1.8393817446299152e-05, + "loss": 0.9183, + "step": 1911 + }, + { + "epoch": 1.0380021715526602, + "grad_norm": 6.400834366436549, + "learning_rate": 1.8391905376093717e-05, + "loss": 1.1517, + "step": 1912 + }, + { + "epoch": 1.0385450597176982, + "grad_norm": 6.160364431172616, + "learning_rate": 1.8389992267964046e-05, + "loss": 0.5949, + "step": 1913 + }, + { + "epoch": 1.0390879478827362, + "grad_norm": 7.298564934931577, + "learning_rate": 1.8388078122146763e-05, + "loss": 1.0234, + "step": 1914 + }, + { + "epoch": 1.0396308360477742, + "grad_norm": 6.132796013684569, + "learning_rate": 1.838616293887861e-05, + "loss": 0.9417, + "step": 1915 + }, + { + "epoch": 1.0401737242128122, + "grad_norm": 8.083369310536504, + "learning_rate": 1.8384246718396458e-05, + "loss": 1.2617, + "step": 1916 + }, + { + "epoch": 1.0407166123778502, + "grad_norm": 5.492413210001038, + "learning_rate": 1.8382329460937306e-05, + "loss": 0.7906, + "step": 1917 + }, + { + "epoch": 1.0412595005428882, + "grad_norm": 5.936786319330656, + "learning_rate": 1.838041116673829e-05, + "loss": 0.8254, + "step": 1918 + }, + { + "epoch": 1.0418023887079262, + "grad_norm": 6.637068024744901, + "learning_rate": 1.8378491836036666e-05, + "loss": 1.1779, + "step": 1919 + }, + { + "epoch": 1.0423452768729642, + "grad_norm": 5.481354066687985, + "learning_rate": 1.8376571469069814e-05, + "loss": 0.9586, + "step": 1920 + }, + { + "epoch": 1.0428881650380022, + "grad_norm": 7.669719057080363, + "learning_rate": 1.8374650066075257e-05, + "loss": 1.2058, + "step": 1921 + }, + { + "epoch": 1.0434310532030402, + "grad_norm": 7.387160499244698, + "learning_rate": 1.8372727627290627e-05, + "loss": 1.3861, + "step": 1922 + }, + { + "epoch": 1.0439739413680782, + "grad_norm": 4.975886150156898, + "learning_rate": 1.8370804152953704e-05, + "loss": 0.5849, + "step": 1923 + }, + { + "epoch": 1.0445168295331162, + "grad_norm": 5.413290989228805, + "learning_rate": 1.8368879643302383e-05, + "loss": 0.8425, + "step": 1924 + }, + { + "epoch": 1.0450597176981542, + "grad_norm": 6.564098001301886, + "learning_rate": 1.836695409857469e-05, + "loss": 0.9658, + "step": 1925 + }, + { + "epoch": 1.0456026058631922, + "grad_norm": 8.050721283259312, + "learning_rate": 1.8365027519008774e-05, + "loss": 1.684, + "step": 1926 + }, + { + "epoch": 1.0461454940282302, + "grad_norm": 5.409117554986433, + "learning_rate": 1.836309990484293e-05, + "loss": 0.8247, + "step": 1927 + }, + { + "epoch": 1.0466883821932682, + "grad_norm": 5.108392155413988, + "learning_rate": 1.8361171256315555e-05, + "loss": 0.8244, + "step": 1928 + }, + { + "epoch": 1.0472312703583062, + "grad_norm": 6.806129539780182, + "learning_rate": 1.8359241573665194e-05, + "loss": 1.0678, + "step": 1929 + }, + { + "epoch": 1.0477741585233442, + "grad_norm": 4.43019983774416, + "learning_rate": 1.8357310857130514e-05, + "loss": 0.6213, + "step": 1930 + }, + { + "epoch": 1.0483170466883822, + "grad_norm": 6.194875830854703, + "learning_rate": 1.835537910695031e-05, + "loss": 0.8774, + "step": 1931 + }, + { + "epoch": 1.0488599348534202, + "grad_norm": 4.540451693265499, + "learning_rate": 1.8353446323363496e-05, + "loss": 0.8508, + "step": 1932 + }, + { + "epoch": 1.0494028230184582, + "grad_norm": 5.9491142341914225, + "learning_rate": 1.8351512506609133e-05, + "loss": 1.0139, + "step": 1933 + }, + { + "epoch": 1.0499457111834962, + "grad_norm": 6.436571178648067, + "learning_rate": 1.834957765692639e-05, + "loss": 0.9272, + "step": 1934 + }, + { + "epoch": 1.0504885993485342, + "grad_norm": 5.6538429161280535, + "learning_rate": 1.8347641774554573e-05, + "loss": 0.7834, + "step": 1935 + }, + { + "epoch": 1.0510314875135722, + "grad_norm": 6.481163094620501, + "learning_rate": 1.8345704859733123e-05, + "loss": 1.0916, + "step": 1936 + }, + { + "epoch": 1.0515743756786102, + "grad_norm": 6.165149195118496, + "learning_rate": 1.8343766912701588e-05, + "loss": 1.1054, + "step": 1937 + }, + { + "epoch": 1.0521172638436482, + "grad_norm": 7.778083794032891, + "learning_rate": 1.834182793369967e-05, + "loss": 1.297, + "step": 1938 + }, + { + "epoch": 1.0526601520086862, + "grad_norm": 8.871857523729659, + "learning_rate": 1.8339887922967176e-05, + "loss": 1.5749, + "step": 1939 + }, + { + "epoch": 1.0532030401737242, + "grad_norm": 6.563413115451867, + "learning_rate": 1.8337946880744047e-05, + "loss": 1.0271, + "step": 1940 + }, + { + "epoch": 1.0537459283387622, + "grad_norm": 6.5304770661428275, + "learning_rate": 1.833600480727036e-05, + "loss": 0.9236, + "step": 1941 + }, + { + "epoch": 1.0542888165038002, + "grad_norm": 6.802479270907137, + "learning_rate": 1.8334061702786317e-05, + "loss": 0.9076, + "step": 1942 + }, + { + "epoch": 1.0548317046688382, + "grad_norm": 5.92560525391674, + "learning_rate": 1.833211756753224e-05, + "loss": 0.9601, + "step": 1943 + }, + { + "epoch": 1.0553745928338762, + "grad_norm": 7.436926375817581, + "learning_rate": 1.8330172401748584e-05, + "loss": 1.1758, + "step": 1944 + }, + { + "epoch": 1.0559174809989142, + "grad_norm": 6.369742715679364, + "learning_rate": 1.8328226205675927e-05, + "loss": 0.8424, + "step": 1945 + }, + { + "epoch": 1.0564603691639523, + "grad_norm": 7.60632781901004, + "learning_rate": 1.8326278979554976e-05, + "loss": 1.7815, + "step": 1946 + }, + { + "epoch": 1.0570032573289903, + "grad_norm": 8.499242837541805, + "learning_rate": 1.8324330723626578e-05, + "loss": 1.1341, + "step": 1947 + }, + { + "epoch": 1.0575461454940283, + "grad_norm": 6.469082998862602, + "learning_rate": 1.8322381438131686e-05, + "loss": 0.7798, + "step": 1948 + }, + { + "epoch": 1.0580890336590663, + "grad_norm": 5.923622010018089, + "learning_rate": 1.8320431123311388e-05, + "loss": 0.7971, + "step": 1949 + }, + { + "epoch": 1.0586319218241043, + "grad_norm": 6.617488371294151, + "learning_rate": 1.8318479779406914e-05, + "loss": 1.1551, + "step": 1950 + }, + { + "epoch": 1.0591748099891423, + "grad_norm": 6.813827391520987, + "learning_rate": 1.8316527406659604e-05, + "loss": 1.0234, + "step": 1951 + }, + { + "epoch": 1.0597176981541803, + "grad_norm": 7.580867549521637, + "learning_rate": 1.831457400531093e-05, + "loss": 0.7896, + "step": 1952 + }, + { + "epoch": 1.0602605863192183, + "grad_norm": 5.975929968595557, + "learning_rate": 1.8312619575602486e-05, + "loss": 1.1248, + "step": 1953 + }, + { + "epoch": 1.0608034744842563, + "grad_norm": 6.2878616306091955, + "learning_rate": 1.831066411777601e-05, + "loss": 1.0833, + "step": 1954 + }, + { + "epoch": 1.0613463626492943, + "grad_norm": 6.284963948133933, + "learning_rate": 1.8308707632073345e-05, + "loss": 1.0773, + "step": 1955 + }, + { + "epoch": 1.0618892508143323, + "grad_norm": 4.009278799804313, + "learning_rate": 1.830675011873648e-05, + "loss": 0.4782, + "step": 1956 + }, + { + "epoch": 1.0624321389793703, + "grad_norm": 6.648653088879531, + "learning_rate": 1.8304791578007524e-05, + "loss": 0.9761, + "step": 1957 + }, + { + "epoch": 1.0629750271444083, + "grad_norm": 6.867797713608168, + "learning_rate": 1.830283201012871e-05, + "loss": 0.9315, + "step": 1958 + }, + { + "epoch": 1.0635179153094463, + "grad_norm": 5.989268423819992, + "learning_rate": 1.83008714153424e-05, + "loss": 0.8312, + "step": 1959 + }, + { + "epoch": 1.0640608034744843, + "grad_norm": 6.207518262073172, + "learning_rate": 1.8298909793891083e-05, + "loss": 1.1003, + "step": 1960 + }, + { + "epoch": 1.0646036916395223, + "grad_norm": 5.5435710943195815, + "learning_rate": 1.8296947146017373e-05, + "loss": 0.8012, + "step": 1961 + }, + { + "epoch": 1.0651465798045603, + "grad_norm": 8.000995164961113, + "learning_rate": 1.829498347196402e-05, + "loss": 0.9864, + "step": 1962 + }, + { + "epoch": 1.0656894679695983, + "grad_norm": 6.483090321755992, + "learning_rate": 1.829301877197389e-05, + "loss": 1.2823, + "step": 1963 + }, + { + "epoch": 1.0662323561346363, + "grad_norm": 7.147142303433649, + "learning_rate": 1.8291053046289985e-05, + "loss": 0.9185, + "step": 1964 + }, + { + "epoch": 1.0667752442996743, + "grad_norm": 6.000816189695904, + "learning_rate": 1.828908629515542e-05, + "loss": 1.0284, + "step": 1965 + }, + { + "epoch": 1.0673181324647123, + "grad_norm": 6.188793151434744, + "learning_rate": 1.8287118518813453e-05, + "loss": 0.7041, + "step": 1966 + }, + { + "epoch": 1.0678610206297503, + "grad_norm": 7.353722352349769, + "learning_rate": 1.828514971750746e-05, + "loss": 1.0707, + "step": 1967 + }, + { + "epoch": 1.0684039087947883, + "grad_norm": 7.553378116857428, + "learning_rate": 1.8283179891480944e-05, + "loss": 1.1426, + "step": 1968 + }, + { + "epoch": 1.0689467969598263, + "grad_norm": 7.134594743515001, + "learning_rate": 1.828120904097754e-05, + "loss": 0.9583, + "step": 1969 + }, + { + "epoch": 1.0694896851248643, + "grad_norm": 5.680617024729498, + "learning_rate": 1.8279237166241004e-05, + "loss": 0.9632, + "step": 1970 + }, + { + "epoch": 1.0700325732899023, + "grad_norm": 5.932910663063727, + "learning_rate": 1.8277264267515218e-05, + "loss": 0.9984, + "step": 1971 + }, + { + "epoch": 1.0705754614549403, + "grad_norm": 7.833159555663921, + "learning_rate": 1.8275290345044198e-05, + "loss": 0.8038, + "step": 1972 + }, + { + "epoch": 1.0711183496199783, + "grad_norm": 6.429988603781813, + "learning_rate": 1.8273315399072076e-05, + "loss": 0.9125, + "step": 1973 + }, + { + "epoch": 1.0716612377850163, + "grad_norm": 7.955595430340166, + "learning_rate": 1.827133942984312e-05, + "loss": 0.8697, + "step": 1974 + }, + { + "epoch": 1.0722041259500543, + "grad_norm": 5.983514211221058, + "learning_rate": 1.826936243760172e-05, + "loss": 0.8155, + "step": 1975 + }, + { + "epoch": 1.0727470141150923, + "grad_norm": 6.19913176090901, + "learning_rate": 1.8267384422592398e-05, + "loss": 1.3207, + "step": 1976 + }, + { + "epoch": 1.0732899022801303, + "grad_norm": 6.467400855860047, + "learning_rate": 1.8265405385059792e-05, + "loss": 1.1279, + "step": 1977 + }, + { + "epoch": 1.0738327904451683, + "grad_norm": 6.515150741776156, + "learning_rate": 1.8263425325248675e-05, + "loss": 1.372, + "step": 1978 + }, + { + "epoch": 1.0743756786102063, + "grad_norm": 6.181497786496646, + "learning_rate": 1.8261444243403945e-05, + "loss": 0.8571, + "step": 1979 + }, + { + "epoch": 1.0749185667752443, + "grad_norm": 7.637816753034266, + "learning_rate": 1.8259462139770624e-05, + "loss": 1.4861, + "step": 1980 + }, + { + "epoch": 1.0754614549402823, + "grad_norm": 7.0094523855293485, + "learning_rate": 1.825747901459386e-05, + "loss": 1.0113, + "step": 1981 + }, + { + "epoch": 1.0760043431053203, + "grad_norm": 7.6230550634556655, + "learning_rate": 1.8255494868118933e-05, + "loss": 1.3818, + "step": 1982 + }, + { + "epoch": 1.0765472312703583, + "grad_norm": 6.946643864799888, + "learning_rate": 1.8253509700591242e-05, + "loss": 1.1384, + "step": 1983 + }, + { + "epoch": 1.0770901194353963, + "grad_norm": 7.288235795183878, + "learning_rate": 1.825152351225632e-05, + "loss": 1.1288, + "step": 1984 + }, + { + "epoch": 1.0776330076004343, + "grad_norm": 8.872287367521855, + "learning_rate": 1.8249536303359816e-05, + "loss": 1.6016, + "step": 1985 + }, + { + "epoch": 1.0781758957654723, + "grad_norm": 5.75562973378046, + "learning_rate": 1.8247548074147515e-05, + "loss": 1.1803, + "step": 1986 + }, + { + "epoch": 1.0787187839305103, + "grad_norm": 6.377752945158632, + "learning_rate": 1.824555882486532e-05, + "loss": 0.6991, + "step": 1987 + }, + { + "epoch": 1.0792616720955484, + "grad_norm": 7.178965937765225, + "learning_rate": 1.8243568555759274e-05, + "loss": 0.988, + "step": 1988 + }, + { + "epoch": 1.0798045602605864, + "grad_norm": 6.0869015160234525, + "learning_rate": 1.824157726707553e-05, + "loss": 0.8587, + "step": 1989 + }, + { + "epoch": 1.0803474484256244, + "grad_norm": 5.9113747723637395, + "learning_rate": 1.823958495906037e-05, + "loss": 0.8815, + "step": 1990 + }, + { + "epoch": 1.0808903365906624, + "grad_norm": 7.84172234151508, + "learning_rate": 1.8237591631960218e-05, + "loss": 1.2621, + "step": 1991 + }, + { + "epoch": 1.0814332247557004, + "grad_norm": 7.188505727546616, + "learning_rate": 1.8235597286021597e-05, + "loss": 0.5991, + "step": 1992 + }, + { + "epoch": 1.0819761129207384, + "grad_norm": 6.358900329219305, + "learning_rate": 1.823360192149118e-05, + "loss": 1.3325, + "step": 1993 + }, + { + "epoch": 1.0825190010857764, + "grad_norm": 6.545296131842583, + "learning_rate": 1.8231605538615756e-05, + "loss": 1.1091, + "step": 1994 + }, + { + "epoch": 1.0830618892508144, + "grad_norm": 6.708966367161533, + "learning_rate": 1.8229608137642238e-05, + "loss": 1.1352, + "step": 1995 + }, + { + "epoch": 1.0836047774158524, + "grad_norm": 6.651776068492304, + "learning_rate": 1.822760971881767e-05, + "loss": 0.6818, + "step": 1996 + }, + { + "epoch": 1.0841476655808904, + "grad_norm": 4.745518059898734, + "learning_rate": 1.8225610282389222e-05, + "loss": 0.9403, + "step": 1997 + }, + { + "epoch": 1.0846905537459284, + "grad_norm": 6.526942367542791, + "learning_rate": 1.8223609828604184e-05, + "loss": 1.0543, + "step": 1998 + }, + { + "epoch": 1.0852334419109664, + "grad_norm": 5.331318824902818, + "learning_rate": 1.8221608357709973e-05, + "loss": 0.6626, + "step": 1999 + }, + { + "epoch": 1.0857763300760044, + "grad_norm": 6.640324946305343, + "learning_rate": 1.8219605869954134e-05, + "loss": 1.0166, + "step": 2000 + }, + { + "epoch": 1.0863192182410424, + "grad_norm": 5.684199608008352, + "learning_rate": 1.8217602365584352e-05, + "loss": 0.6796, + "step": 2001 + }, + { + "epoch": 1.0868621064060804, + "grad_norm": 5.429738239817468, + "learning_rate": 1.8215597844848403e-05, + "loss": 0.6783, + "step": 2002 + }, + { + "epoch": 1.0874049945711184, + "grad_norm": 6.1832795338939235, + "learning_rate": 1.821359230799422e-05, + "loss": 1.4244, + "step": 2003 + }, + { + "epoch": 1.0879478827361564, + "grad_norm": 7.118410329483546, + "learning_rate": 1.8211585755269852e-05, + "loss": 1.0274, + "step": 2004 + }, + { + "epoch": 1.0884907709011944, + "grad_norm": 5.5061218005078265, + "learning_rate": 1.820957818692347e-05, + "loss": 0.7143, + "step": 2005 + }, + { + "epoch": 1.0890336590662324, + "grad_norm": 5.518412565715972, + "learning_rate": 1.8207569603203373e-05, + "loss": 1.3158, + "step": 2006 + }, + { + "epoch": 1.0895765472312704, + "grad_norm": 5.47000403426222, + "learning_rate": 1.820556000435799e-05, + "loss": 1.0694, + "step": 2007 + }, + { + "epoch": 1.0901194353963084, + "grad_norm": 7.469961154104622, + "learning_rate": 1.820354939063586e-05, + "loss": 1.0702, + "step": 2008 + }, + { + "epoch": 1.0906623235613464, + "grad_norm": 5.889961089255285, + "learning_rate": 1.8201537762285674e-05, + "loss": 1.0598, + "step": 2009 + }, + { + "epoch": 1.0912052117263844, + "grad_norm": 6.659595389259375, + "learning_rate": 1.8199525119556226e-05, + "loss": 0.8456, + "step": 2010 + }, + { + "epoch": 1.0917480998914224, + "grad_norm": 6.328157257432415, + "learning_rate": 1.8197511462696443e-05, + "loss": 0.9763, + "step": 2011 + }, + { + "epoch": 1.0922909880564604, + "grad_norm": 5.916734142619132, + "learning_rate": 1.8195496791955373e-05, + "loss": 1.1194, + "step": 2012 + }, + { + "epoch": 1.0928338762214984, + "grad_norm": 5.904087624332602, + "learning_rate": 1.8193481107582203e-05, + "loss": 0.756, + "step": 2013 + }, + { + "epoch": 1.0933767643865364, + "grad_norm": 6.284745154848233, + "learning_rate": 1.8191464409826227e-05, + "loss": 0.7931, + "step": 2014 + }, + { + "epoch": 1.0939196525515744, + "grad_norm": 7.838306243079945, + "learning_rate": 1.8189446698936878e-05, + "loss": 1.3635, + "step": 2015 + }, + { + "epoch": 1.0944625407166124, + "grad_norm": 6.3405758352546995, + "learning_rate": 1.818742797516371e-05, + "loss": 0.5365, + "step": 2016 + }, + { + "epoch": 1.0950054288816504, + "grad_norm": 8.315910084823138, + "learning_rate": 1.8185408238756405e-05, + "loss": 1.6592, + "step": 2017 + }, + { + "epoch": 1.0955483170466884, + "grad_norm": 5.205053142209431, + "learning_rate": 1.8183387489964762e-05, + "loss": 0.6416, + "step": 2018 + }, + { + "epoch": 1.0960912052117264, + "grad_norm": 7.018373673519811, + "learning_rate": 1.8181365729038706e-05, + "loss": 1.0814, + "step": 2019 + }, + { + "epoch": 1.0966340933767644, + "grad_norm": 5.705860306884454, + "learning_rate": 1.8179342956228307e-05, + "loss": 0.5806, + "step": 2020 + }, + { + "epoch": 1.0971769815418024, + "grad_norm": 7.077662290720525, + "learning_rate": 1.8177319171783728e-05, + "loss": 0.8969, + "step": 2021 + }, + { + "epoch": 1.0977198697068404, + "grad_norm": 6.540901861387866, + "learning_rate": 1.8175294375955284e-05, + "loss": 0.7982, + "step": 2022 + }, + { + "epoch": 1.0982627578718784, + "grad_norm": 7.283213833491601, + "learning_rate": 1.81732685689934e-05, + "loss": 1.2659, + "step": 2023 + }, + { + "epoch": 1.0988056460369164, + "grad_norm": 7.016589270114781, + "learning_rate": 1.8171241751148633e-05, + "loss": 1.0011, + "step": 2024 + }, + { + "epoch": 1.0993485342019544, + "grad_norm": 5.724817250167612, + "learning_rate": 1.8169213922671666e-05, + "loss": 0.8967, + "step": 2025 + }, + { + "epoch": 1.0998914223669924, + "grad_norm": 7.431207902407107, + "learning_rate": 1.81671850838133e-05, + "loss": 0.9445, + "step": 2026 + }, + { + "epoch": 1.1004343105320304, + "grad_norm": 6.9152342645144955, + "learning_rate": 1.816515523482447e-05, + "loss": 0.8187, + "step": 2027 + }, + { + "epoch": 1.1009771986970684, + "grad_norm": 7.045028624610342, + "learning_rate": 1.816312437595622e-05, + "loss": 0.7323, + "step": 2028 + }, + { + "epoch": 1.1015200868621065, + "grad_norm": 6.63265573559803, + "learning_rate": 1.816109250745974e-05, + "loss": 0.6689, + "step": 2029 + }, + { + "epoch": 1.1020629750271445, + "grad_norm": 8.605675554379312, + "learning_rate": 1.8159059629586333e-05, + "loss": 1.3733, + "step": 2030 + }, + { + "epoch": 1.1026058631921825, + "grad_norm": 7.481808301387036, + "learning_rate": 1.8157025742587426e-05, + "loss": 0.9767, + "step": 2031 + }, + { + "epoch": 1.1031487513572205, + "grad_norm": 7.113214270917786, + "learning_rate": 1.8154990846714575e-05, + "loss": 0.6365, + "step": 2032 + }, + { + "epoch": 1.1036916395222585, + "grad_norm": 7.672152203903941, + "learning_rate": 1.8152954942219462e-05, + "loss": 1.3314, + "step": 2033 + }, + { + "epoch": 1.1042345276872965, + "grad_norm": 7.533257156973554, + "learning_rate": 1.8150918029353885e-05, + "loss": 0.967, + "step": 2034 + }, + { + "epoch": 1.1047774158523345, + "grad_norm": 6.469759228237502, + "learning_rate": 1.8148880108369775e-05, + "loss": 1.348, + "step": 2035 + }, + { + "epoch": 1.1053203040173725, + "grad_norm": 6.620692148711997, + "learning_rate": 1.8146841179519186e-05, + "loss": 0.7538, + "step": 2036 + }, + { + "epoch": 1.1058631921824105, + "grad_norm": 6.974705446587608, + "learning_rate": 1.8144801243054297e-05, + "loss": 1.3664, + "step": 2037 + }, + { + "epoch": 1.1064060803474485, + "grad_norm": 8.047700738112871, + "learning_rate": 1.8142760299227408e-05, + "loss": 1.0466, + "step": 2038 + }, + { + "epoch": 1.1069489685124865, + "grad_norm": 5.469412128245213, + "learning_rate": 1.814071834829095e-05, + "loss": 0.5464, + "step": 2039 + }, + { + "epoch": 1.1074918566775245, + "grad_norm": 5.732997692484992, + "learning_rate": 1.813867539049747e-05, + "loss": 0.7307, + "step": 2040 + }, + { + "epoch": 1.1080347448425625, + "grad_norm": 8.89935072196041, + "learning_rate": 1.8136631426099646e-05, + "loss": 1.8231, + "step": 2041 + }, + { + "epoch": 1.1085776330076005, + "grad_norm": 7.0036806828899465, + "learning_rate": 1.813458645535028e-05, + "loss": 1.063, + "step": 2042 + }, + { + "epoch": 1.1091205211726385, + "grad_norm": 6.013891951698058, + "learning_rate": 1.8132540478502297e-05, + "loss": 0.9836, + "step": 2043 + }, + { + "epoch": 1.1096634093376765, + "grad_norm": 4.234811309290956, + "learning_rate": 1.813049349580875e-05, + "loss": 0.3708, + "step": 2044 + }, + { + "epoch": 1.1102062975027145, + "grad_norm": 6.259459841046158, + "learning_rate": 1.8128445507522806e-05, + "loss": 0.8463, + "step": 2045 + }, + { + "epoch": 1.1107491856677525, + "grad_norm": 7.9568342619649775, + "learning_rate": 1.8126396513897764e-05, + "loss": 1.0563, + "step": 2046 + }, + { + "epoch": 1.1112920738327905, + "grad_norm": 6.937848428371049, + "learning_rate": 1.8124346515187056e-05, + "loss": 0.6359, + "step": 2047 + }, + { + "epoch": 1.1118349619978285, + "grad_norm": 7.672632972051955, + "learning_rate": 1.8122295511644218e-05, + "loss": 0.657, + "step": 2048 + }, + { + "epoch": 1.1123778501628665, + "grad_norm": 6.450214026534184, + "learning_rate": 1.8120243503522924e-05, + "loss": 0.7279, + "step": 2049 + }, + { + "epoch": 1.1129207383279045, + "grad_norm": 6.628216836060006, + "learning_rate": 1.8118190491076978e-05, + "loss": 0.5804, + "step": 2050 + }, + { + "epoch": 1.1134636264929425, + "grad_norm": 6.698074316145226, + "learning_rate": 1.8116136474560288e-05, + "loss": 1.011, + "step": 2051 + }, + { + "epoch": 1.1140065146579805, + "grad_norm": 5.4126983291571795, + "learning_rate": 1.8114081454226905e-05, + "loss": 0.8564, + "step": 2052 + }, + { + "epoch": 1.1145494028230185, + "grad_norm": 8.596797841329312, + "learning_rate": 1.8112025430331e-05, + "loss": 1.1924, + "step": 2053 + }, + { + "epoch": 1.1150922909880565, + "grad_norm": 7.416912053391948, + "learning_rate": 1.8109968403126856e-05, + "loss": 0.9894, + "step": 2054 + }, + { + "epoch": 1.1156351791530945, + "grad_norm": 7.434308789484601, + "learning_rate": 1.8107910372868898e-05, + "loss": 1.1544, + "step": 2055 + }, + { + "epoch": 1.1161780673181325, + "grad_norm": 8.325689788099726, + "learning_rate": 1.8105851339811663e-05, + "loss": 0.9994, + "step": 2056 + }, + { + "epoch": 1.1167209554831705, + "grad_norm": 8.384964960696278, + "learning_rate": 1.8103791304209813e-05, + "loss": 1.1768, + "step": 2057 + }, + { + "epoch": 1.1172638436482085, + "grad_norm": 7.044535285708613, + "learning_rate": 1.810173026631814e-05, + "loss": 1.1801, + "step": 2058 + }, + { + "epoch": 1.1178067318132465, + "grad_norm": 6.665810359686943, + "learning_rate": 1.8099668226391552e-05, + "loss": 0.7865, + "step": 2059 + }, + { + "epoch": 1.1183496199782845, + "grad_norm": 7.395697719004186, + "learning_rate": 1.8097605184685093e-05, + "loss": 1.1357, + "step": 2060 + }, + { + "epoch": 1.1188925081433225, + "grad_norm": 7.1495559268215185, + "learning_rate": 1.809554114145392e-05, + "loss": 1.1507, + "step": 2061 + }, + { + "epoch": 1.1194353963083605, + "grad_norm": 7.877063869804663, + "learning_rate": 1.8093476096953315e-05, + "loss": 1.634, + "step": 2062 + }, + { + "epoch": 1.1199782844733985, + "grad_norm": 6.528060830825324, + "learning_rate": 1.809141005143869e-05, + "loss": 0.697, + "step": 2063 + }, + { + "epoch": 1.1205211726384365, + "grad_norm": 6.5949576624011, + "learning_rate": 1.808934300516557e-05, + "loss": 1.1506, + "step": 2064 + }, + { + "epoch": 1.1210640608034745, + "grad_norm": 6.787348272435354, + "learning_rate": 1.8087274958389612e-05, + "loss": 0.9986, + "step": 2065 + }, + { + "epoch": 1.1216069489685125, + "grad_norm": 6.310955809912108, + "learning_rate": 1.8085205911366602e-05, + "loss": 0.89, + "step": 2066 + }, + { + "epoch": 1.1221498371335505, + "grad_norm": 6.8487503881448895, + "learning_rate": 1.8083135864352442e-05, + "loss": 1.1061, + "step": 2067 + }, + { + "epoch": 1.1226927252985885, + "grad_norm": 5.329494375939551, + "learning_rate": 1.808106481760315e-05, + "loss": 1.0011, + "step": 2068 + }, + { + "epoch": 1.1232356134636265, + "grad_norm": 6.286005313228098, + "learning_rate": 1.8078992771374886e-05, + "loss": 0.8399, + "step": 2069 + }, + { + "epoch": 1.1237785016286646, + "grad_norm": 6.324896503266866, + "learning_rate": 1.8076919725923917e-05, + "loss": 1.2716, + "step": 2070 + }, + { + "epoch": 1.1243213897937026, + "grad_norm": 5.663105266452551, + "learning_rate": 1.8074845681506644e-05, + "loss": 0.9943, + "step": 2071 + }, + { + "epoch": 1.1248642779587406, + "grad_norm": 5.175572461557621, + "learning_rate": 1.807277063837959e-05, + "loss": 0.6319, + "step": 2072 + }, + { + "epoch": 1.1254071661237786, + "grad_norm": 7.297487212372251, + "learning_rate": 1.8070694596799397e-05, + "loss": 0.7533, + "step": 2073 + }, + { + "epoch": 1.1259500542888166, + "grad_norm": 9.046578233072974, + "learning_rate": 1.806861755702283e-05, + "loss": 0.9991, + "step": 2074 + }, + { + "epoch": 1.1264929424538546, + "grad_norm": 4.516511292735119, + "learning_rate": 1.8066539519306786e-05, + "loss": 0.6393, + "step": 2075 + }, + { + "epoch": 1.1270358306188926, + "grad_norm": 5.16623460840236, + "learning_rate": 1.8064460483908283e-05, + "loss": 0.7083, + "step": 2076 + }, + { + "epoch": 1.1275787187839306, + "grad_norm": 10.7323945119629, + "learning_rate": 1.8062380451084445e-05, + "loss": 2.4193, + "step": 2077 + }, + { + "epoch": 1.1281216069489686, + "grad_norm": 6.438876475868993, + "learning_rate": 1.8060299421092554e-05, + "loss": 0.8468, + "step": 2078 + }, + { + "epoch": 1.1286644951140066, + "grad_norm": 8.437791617086905, + "learning_rate": 1.8058217394189976e-05, + "loss": 1.4895, + "step": 2079 + }, + { + "epoch": 1.1292073832790446, + "grad_norm": 5.237528459736948, + "learning_rate": 1.805613437063423e-05, + "loss": 0.643, + "step": 2080 + }, + { + "epoch": 1.1297502714440826, + "grad_norm": 6.010151293762264, + "learning_rate": 1.8054050350682947e-05, + "loss": 0.9001, + "step": 2081 + }, + { + "epoch": 1.1302931596091206, + "grad_norm": 4.816849804237718, + "learning_rate": 1.805196533459388e-05, + "loss": 0.6349, + "step": 2082 + }, + { + "epoch": 1.1308360477741586, + "grad_norm": 6.872496495387926, + "learning_rate": 1.8049879322624906e-05, + "loss": 0.8808, + "step": 2083 + }, + { + "epoch": 1.1313789359391966, + "grad_norm": 5.86067647581003, + "learning_rate": 1.804779231503403e-05, + "loss": 0.6811, + "step": 2084 + }, + { + "epoch": 1.1319218241042346, + "grad_norm": 6.073217718554074, + "learning_rate": 1.8045704312079376e-05, + "loss": 1.0086, + "step": 2085 + }, + { + "epoch": 1.1324647122692726, + "grad_norm": 7.519215180241719, + "learning_rate": 1.804361531401918e-05, + "loss": 0.8815, + "step": 2086 + }, + { + "epoch": 1.1330076004343106, + "grad_norm": 6.49012361077388, + "learning_rate": 1.8041525321111835e-05, + "loss": 1.0129, + "step": 2087 + }, + { + "epoch": 1.1335504885993486, + "grad_norm": 5.938773374319115, + "learning_rate": 1.8039434333615814e-05, + "loss": 0.8485, + "step": 2088 + }, + { + "epoch": 1.1340933767643866, + "grad_norm": 5.607636698742016, + "learning_rate": 1.8037342351789743e-05, + "loss": 0.6209, + "step": 2089 + }, + { + "epoch": 1.1346362649294246, + "grad_norm": 5.522935966563286, + "learning_rate": 1.803524937589236e-05, + "loss": 0.6576, + "step": 2090 + }, + { + "epoch": 1.1351791530944626, + "grad_norm": 8.130974168982206, + "learning_rate": 1.8033155406182533e-05, + "loss": 1.0676, + "step": 2091 + }, + { + "epoch": 1.1357220412595006, + "grad_norm": 6.027276275319545, + "learning_rate": 1.803106044291924e-05, + "loss": 0.9853, + "step": 2092 + }, + { + "epoch": 1.1362649294245386, + "grad_norm": 4.645985748184023, + "learning_rate": 1.8028964486361586e-05, + "loss": 0.7505, + "step": 2093 + }, + { + "epoch": 1.1368078175895766, + "grad_norm": 6.820894579082682, + "learning_rate": 1.8026867536768816e-05, + "loss": 0.9358, + "step": 2094 + }, + { + "epoch": 1.1373507057546146, + "grad_norm": 3.9530727606720872, + "learning_rate": 1.802476959440027e-05, + "loss": 0.7736, + "step": 2095 + }, + { + "epoch": 1.1378935939196526, + "grad_norm": 7.650275740883737, + "learning_rate": 1.8022670659515432e-05, + "loss": 1.4832, + "step": 2096 + }, + { + "epoch": 1.1384364820846906, + "grad_norm": 6.94341164115752, + "learning_rate": 1.80205707323739e-05, + "loss": 1.0336, + "step": 2097 + }, + { + "epoch": 1.1389793702497286, + "grad_norm": 5.608528053150328, + "learning_rate": 1.8018469813235403e-05, + "loss": 0.829, + "step": 2098 + }, + { + "epoch": 1.1395222584147666, + "grad_norm": 5.2191336182398995, + "learning_rate": 1.8016367902359776e-05, + "loss": 0.9708, + "step": 2099 + }, + { + "epoch": 1.1400651465798046, + "grad_norm": 6.773346810380977, + "learning_rate": 1.801426500000699e-05, + "loss": 0.6755, + "step": 2100 + }, + { + "epoch": 1.1406080347448426, + "grad_norm": 7.292992073718992, + "learning_rate": 1.8012161106437137e-05, + "loss": 0.8842, + "step": 2101 + }, + { + "epoch": 1.1411509229098806, + "grad_norm": 6.699434744454233, + "learning_rate": 1.8010056221910427e-05, + "loss": 1.0385, + "step": 2102 + }, + { + "epoch": 1.1416938110749186, + "grad_norm": 6.565299107509569, + "learning_rate": 1.8007950346687198e-05, + "loss": 0.6557, + "step": 2103 + }, + { + "epoch": 1.1422366992399566, + "grad_norm": 7.2741514068627025, + "learning_rate": 1.800584348102791e-05, + "loss": 1.0347, + "step": 2104 + }, + { + "epoch": 1.1427795874049946, + "grad_norm": 6.467241522225306, + "learning_rate": 1.800373562519314e-05, + "loss": 0.8026, + "step": 2105 + }, + { + "epoch": 1.1433224755700326, + "grad_norm": 6.109318081193457, + "learning_rate": 1.800162677944359e-05, + "loss": 0.9182, + "step": 2106 + }, + { + "epoch": 1.1438653637350706, + "grad_norm": 7.136347555199431, + "learning_rate": 1.7999516944040087e-05, + "loss": 1.2743, + "step": 2107 + }, + { + "epoch": 1.1444082519001086, + "grad_norm": 7.791284332349926, + "learning_rate": 1.7997406119243582e-05, + "loss": 0.9738, + "step": 2108 + }, + { + "epoch": 1.1449511400651466, + "grad_norm": 7.315957477694666, + "learning_rate": 1.7995294305315137e-05, + "loss": 1.0674, + "step": 2109 + }, + { + "epoch": 1.1454940282301846, + "grad_norm": 6.305984776426275, + "learning_rate": 1.7993181502515957e-05, + "loss": 0.8204, + "step": 2110 + }, + { + "epoch": 1.1460369163952226, + "grad_norm": 5.824114865699788, + "learning_rate": 1.7991067711107345e-05, + "loss": 0.7953, + "step": 2111 + }, + { + "epoch": 1.1465798045602607, + "grad_norm": 7.4284096250312786, + "learning_rate": 1.798895293135074e-05, + "loss": 0.7349, + "step": 2112 + }, + { + "epoch": 1.1471226927252987, + "grad_norm": 6.415949338549894, + "learning_rate": 1.798683716350771e-05, + "loss": 0.8346, + "step": 2113 + }, + { + "epoch": 1.1476655808903367, + "grad_norm": 6.360883520037482, + "learning_rate": 1.7984720407839925e-05, + "loss": 0.7966, + "step": 2114 + }, + { + "epoch": 1.1482084690553747, + "grad_norm": 6.703196393264975, + "learning_rate": 1.7982602664609198e-05, + "loss": 1.3341, + "step": 2115 + }, + { + "epoch": 1.1487513572204127, + "grad_norm": 8.115384707506347, + "learning_rate": 1.798048393407745e-05, + "loss": 1.0284, + "step": 2116 + }, + { + "epoch": 1.1492942453854507, + "grad_norm": 8.519166176194911, + "learning_rate": 1.797836421650673e-05, + "loss": 1.4075, + "step": 2117 + }, + { + "epoch": 1.1498371335504887, + "grad_norm": 7.120345871314324, + "learning_rate": 1.7976243512159207e-05, + "loss": 1.1933, + "step": 2118 + }, + { + "epoch": 1.1503800217155267, + "grad_norm": 5.887743000051609, + "learning_rate": 1.7974121821297178e-05, + "loss": 0.9143, + "step": 2119 + }, + { + "epoch": 1.1509229098805647, + "grad_norm": 6.264158167802545, + "learning_rate": 1.797199914418305e-05, + "loss": 0.805, + "step": 2120 + }, + { + "epoch": 1.1514657980456027, + "grad_norm": 7.747345696360016, + "learning_rate": 1.7969875481079363e-05, + "loss": 1.1374, + "step": 2121 + }, + { + "epoch": 1.1520086862106407, + "grad_norm": 5.932381639003466, + "learning_rate": 1.7967750832248774e-05, + "loss": 0.7801, + "step": 2122 + }, + { + "epoch": 1.1525515743756787, + "grad_norm": 5.646019326027894, + "learning_rate": 1.7965625197954064e-05, + "loss": 0.5053, + "step": 2123 + }, + { + "epoch": 1.1530944625407167, + "grad_norm": 6.085604242778934, + "learning_rate": 1.7963498578458135e-05, + "loss": 0.6711, + "step": 2124 + }, + { + "epoch": 1.1536373507057547, + "grad_norm": 6.390132028289734, + "learning_rate": 1.7961370974024008e-05, + "loss": 0.9897, + "step": 2125 + }, + { + "epoch": 1.1541802388707927, + "grad_norm": 5.052558465006266, + "learning_rate": 1.795924238491483e-05, + "loss": 0.8742, + "step": 2126 + }, + { + "epoch": 1.1547231270358307, + "grad_norm": 5.86713723099216, + "learning_rate": 1.7957112811393867e-05, + "loss": 0.965, + "step": 2127 + }, + { + "epoch": 1.1552660152008687, + "grad_norm": 6.092033075473227, + "learning_rate": 1.7954982253724515e-05, + "loss": 1.104, + "step": 2128 + }, + { + "epoch": 1.1558089033659067, + "grad_norm": 7.051795930424219, + "learning_rate": 1.7952850712170278e-05, + "loss": 1.0918, + "step": 2129 + }, + { + "epoch": 1.1563517915309447, + "grad_norm": 5.613114568974867, + "learning_rate": 1.7950718186994788e-05, + "loss": 0.6698, + "step": 2130 + }, + { + "epoch": 1.1568946796959827, + "grad_norm": 5.533439816776461, + "learning_rate": 1.7948584678461802e-05, + "loss": 1.4395, + "step": 2131 + }, + { + "epoch": 1.1574375678610207, + "grad_norm": 7.73758559868109, + "learning_rate": 1.7946450186835195e-05, + "loss": 1.0806, + "step": 2132 + }, + { + "epoch": 1.1579804560260587, + "grad_norm": 5.982378084297589, + "learning_rate": 1.7944314712378966e-05, + "loss": 1.0428, + "step": 2133 + }, + { + "epoch": 1.1585233441910967, + "grad_norm": 5.604421383798244, + "learning_rate": 1.7942178255357227e-05, + "loss": 0.9411, + "step": 2134 + }, + { + "epoch": 1.1590662323561347, + "grad_norm": 6.5637069585312435, + "learning_rate": 1.794004081603423e-05, + "loss": 0.8285, + "step": 2135 + }, + { + "epoch": 1.1596091205211727, + "grad_norm": 5.567535422744373, + "learning_rate": 1.7937902394674326e-05, + "loss": 1.0185, + "step": 2136 + }, + { + "epoch": 1.1601520086862107, + "grad_norm": 7.898198443147329, + "learning_rate": 1.7935762991542004e-05, + "loss": 0.8878, + "step": 2137 + }, + { + "epoch": 1.1606948968512487, + "grad_norm": 5.12407458777131, + "learning_rate": 1.7933622606901865e-05, + "loss": 0.9174, + "step": 2138 + }, + { + "epoch": 1.1612377850162867, + "grad_norm": 4.480165604953252, + "learning_rate": 1.7931481241018642e-05, + "loss": 0.9424, + "step": 2139 + }, + { + "epoch": 1.1617806731813247, + "grad_norm": 7.6708161194386495, + "learning_rate": 1.7929338894157173e-05, + "loss": 0.9875, + "step": 2140 + }, + { + "epoch": 1.1623235613463627, + "grad_norm": 9.749916559657342, + "learning_rate": 1.7927195566582435e-05, + "loss": 1.7168, + "step": 2141 + }, + { + "epoch": 1.1628664495114007, + "grad_norm": 5.50355689991458, + "learning_rate": 1.7925051258559516e-05, + "loss": 0.7957, + "step": 2142 + }, + { + "epoch": 1.1634093376764387, + "grad_norm": 5.910752363178181, + "learning_rate": 1.7922905970353627e-05, + "loss": 1.0525, + "step": 2143 + }, + { + "epoch": 1.1639522258414767, + "grad_norm": 6.523454927461212, + "learning_rate": 1.7920759702230098e-05, + "loss": 0.9938, + "step": 2144 + }, + { + "epoch": 1.1644951140065147, + "grad_norm": 6.963057538162332, + "learning_rate": 1.7918612454454387e-05, + "loss": 0.8677, + "step": 2145 + }, + { + "epoch": 1.1650380021715527, + "grad_norm": 6.55011455398606, + "learning_rate": 1.7916464227292067e-05, + "loss": 0.7268, + "step": 2146 + }, + { + "epoch": 1.1655808903365907, + "grad_norm": 6.47957363507043, + "learning_rate": 1.7914315021008836e-05, + "loss": 0.7449, + "step": 2147 + }, + { + "epoch": 1.1661237785016287, + "grad_norm": 6.574809242111077, + "learning_rate": 1.7912164835870506e-05, + "loss": 0.6999, + "step": 2148 + }, + { + "epoch": 1.1666666666666667, + "grad_norm": 8.307297080129953, + "learning_rate": 1.791001367214302e-05, + "loss": 1.0402, + "step": 2149 + }, + { + "epoch": 1.1672095548317047, + "grad_norm": 8.01469917888599, + "learning_rate": 1.7907861530092438e-05, + "loss": 1.1024, + "step": 2150 + }, + { + "epoch": 1.1677524429967427, + "grad_norm": 7.336381076899337, + "learning_rate": 1.7905708409984938e-05, + "loss": 0.9112, + "step": 2151 + }, + { + "epoch": 1.1682953311617807, + "grad_norm": 7.872105686379811, + "learning_rate": 1.7903554312086823e-05, + "loss": 0.9002, + "step": 2152 + }, + { + "epoch": 1.1688382193268188, + "grad_norm": 5.842761146463113, + "learning_rate": 1.7901399236664514e-05, + "loss": 0.9457, + "step": 2153 + }, + { + "epoch": 1.1693811074918568, + "grad_norm": 5.8100183378475005, + "learning_rate": 1.7899243183984552e-05, + "loss": 0.9675, + "step": 2154 + }, + { + "epoch": 1.1699239956568948, + "grad_norm": 5.208689262592845, + "learning_rate": 1.7897086154313604e-05, + "loss": 0.7562, + "step": 2155 + }, + { + "epoch": 1.1704668838219328, + "grad_norm": 7.695958654055186, + "learning_rate": 1.7894928147918458e-05, + "loss": 0.9995, + "step": 2156 + }, + { + "epoch": 1.1710097719869708, + "grad_norm": 5.8108200610880525, + "learning_rate": 1.789276916506601e-05, + "loss": 0.8179, + "step": 2157 + }, + { + "epoch": 1.1715526601520088, + "grad_norm": 5.776078861908261, + "learning_rate": 1.7890609206023298e-05, + "loss": 0.6612, + "step": 2158 + }, + { + "epoch": 1.1720955483170468, + "grad_norm": 6.458218841014727, + "learning_rate": 1.7888448271057463e-05, + "loss": 0.8419, + "step": 2159 + }, + { + "epoch": 1.1726384364820848, + "grad_norm": 7.236506499419343, + "learning_rate": 1.7886286360435772e-05, + "loss": 0.6336, + "step": 2160 + }, + { + "epoch": 1.1731813246471228, + "grad_norm": 7.20552304926134, + "learning_rate": 1.7884123474425614e-05, + "loss": 0.9157, + "step": 2161 + }, + { + "epoch": 1.1737242128121608, + "grad_norm": 8.090329675463652, + "learning_rate": 1.78819596132945e-05, + "loss": 1.6894, + "step": 2162 + }, + { + "epoch": 1.1742671009771988, + "grad_norm": 7.500510561892478, + "learning_rate": 1.787979477731006e-05, + "loss": 0.9469, + "step": 2163 + }, + { + "epoch": 1.1748099891422368, + "grad_norm": 7.6126459328076415, + "learning_rate": 1.7877628966740044e-05, + "loss": 0.9179, + "step": 2164 + }, + { + "epoch": 1.1753528773072748, + "grad_norm": 7.065751172246334, + "learning_rate": 1.787546218185232e-05, + "loss": 0.9649, + "step": 2165 + }, + { + "epoch": 1.1758957654723128, + "grad_norm": 6.923487399525626, + "learning_rate": 1.787329442291488e-05, + "loss": 1.0791, + "step": 2166 + }, + { + "epoch": 1.1764386536373508, + "grad_norm": 7.614292041669914, + "learning_rate": 1.7871125690195843e-05, + "loss": 1.2367, + "step": 2167 + }, + { + "epoch": 1.1769815418023888, + "grad_norm": 8.560468932577576, + "learning_rate": 1.7868955983963434e-05, + "loss": 1.0876, + "step": 2168 + }, + { + "epoch": 1.1775244299674268, + "grad_norm": 6.209836424501219, + "learning_rate": 1.786678530448601e-05, + "loss": 0.6251, + "step": 2169 + }, + { + "epoch": 1.1780673181324648, + "grad_norm": 5.926645140708837, + "learning_rate": 1.7864613652032035e-05, + "loss": 0.5939, + "step": 2170 + }, + { + "epoch": 1.1786102062975028, + "grad_norm": 5.460593849347875, + "learning_rate": 1.7862441026870114e-05, + "loss": 0.8769, + "step": 2171 + }, + { + "epoch": 1.1791530944625408, + "grad_norm": 8.063368903057139, + "learning_rate": 1.7860267429268954e-05, + "loss": 1.1492, + "step": 2172 + }, + { + "epoch": 1.1796959826275788, + "grad_norm": 6.246398264149539, + "learning_rate": 1.7858092859497392e-05, + "loss": 0.9146, + "step": 2173 + }, + { + "epoch": 1.1802388707926168, + "grad_norm": 9.483282452267977, + "learning_rate": 1.7855917317824383e-05, + "loss": 1.3714, + "step": 2174 + }, + { + "epoch": 1.1807817589576548, + "grad_norm": 7.377757446359213, + "learning_rate": 1.7853740804519e-05, + "loss": 0.7914, + "step": 2175 + }, + { + "epoch": 1.1813246471226928, + "grad_norm": 6.555194448443568, + "learning_rate": 1.7851563319850435e-05, + "loss": 1.0025, + "step": 2176 + }, + { + "epoch": 1.1818675352877308, + "grad_norm": 5.968820734002013, + "learning_rate": 1.7849384864088005e-05, + "loss": 1.0449, + "step": 2177 + }, + { + "epoch": 1.1824104234527688, + "grad_norm": 8.050731398373367, + "learning_rate": 1.7847205437501145e-05, + "loss": 1.2944, + "step": 2178 + }, + { + "epoch": 1.1829533116178068, + "grad_norm": 6.479093872062473, + "learning_rate": 1.784502504035941e-05, + "loss": 0.8312, + "step": 2179 + }, + { + "epoch": 1.1834961997828448, + "grad_norm": 9.645172625032446, + "learning_rate": 1.7842843672932473e-05, + "loss": 1.4877, + "step": 2180 + }, + { + "epoch": 1.1840390879478828, + "grad_norm": 7.811721210097083, + "learning_rate": 1.7840661335490133e-05, + "loss": 1.0861, + "step": 2181 + }, + { + "epoch": 1.1845819761129208, + "grad_norm": 5.939133632197409, + "learning_rate": 1.7838478028302303e-05, + "loss": 0.7359, + "step": 2182 + }, + { + "epoch": 1.1851248642779588, + "grad_norm": 7.121597254982603, + "learning_rate": 1.7836293751639017e-05, + "loss": 0.626, + "step": 2183 + }, + { + "epoch": 1.1856677524429968, + "grad_norm": 7.225734005859657, + "learning_rate": 1.783410850577043e-05, + "loss": 1.0651, + "step": 2184 + }, + { + "epoch": 1.1862106406080348, + "grad_norm": 6.3733995711762965, + "learning_rate": 1.783192229096682e-05, + "loss": 1.0399, + "step": 2185 + }, + { + "epoch": 1.1867535287730728, + "grad_norm": 7.215840642904109, + "learning_rate": 1.7829735107498576e-05, + "loss": 1.0958, + "step": 2186 + }, + { + "epoch": 1.1872964169381108, + "grad_norm": 7.2349637750528135, + "learning_rate": 1.7827546955636216e-05, + "loss": 0.8005, + "step": 2187 + }, + { + "epoch": 1.1878393051031488, + "grad_norm": 7.592927799049675, + "learning_rate": 1.7825357835650376e-05, + "loss": 1.0981, + "step": 2188 + }, + { + "epoch": 1.1883821932681868, + "grad_norm": 8.010034188435748, + "learning_rate": 1.7823167747811805e-05, + "loss": 1.3088, + "step": 2189 + }, + { + "epoch": 1.1889250814332248, + "grad_norm": 5.674359355487917, + "learning_rate": 1.7820976692391377e-05, + "loss": 0.8725, + "step": 2190 + }, + { + "epoch": 1.1894679695982628, + "grad_norm": 5.236415159647034, + "learning_rate": 1.781878466966009e-05, + "loss": 0.7162, + "step": 2191 + }, + { + "epoch": 1.1900108577633008, + "grad_norm": 5.677268305787393, + "learning_rate": 1.781659167988905e-05, + "loss": 0.7101, + "step": 2192 + }, + { + "epoch": 1.1905537459283388, + "grad_norm": 6.370730170332413, + "learning_rate": 1.7814397723349496e-05, + "loss": 0.9087, + "step": 2193 + }, + { + "epoch": 1.1910966340933768, + "grad_norm": 6.045624314512159, + "learning_rate": 1.7812202800312776e-05, + "loss": 1.1714, + "step": 2194 + }, + { + "epoch": 1.1916395222584149, + "grad_norm": 10.798495850925244, + "learning_rate": 1.7810006911050366e-05, + "loss": 1.1997, + "step": 2195 + }, + { + "epoch": 1.1921824104234529, + "grad_norm": 5.89010386817935, + "learning_rate": 1.780781005583385e-05, + "loss": 1.0702, + "step": 2196 + }, + { + "epoch": 1.1927252985884909, + "grad_norm": 6.09421937705502, + "learning_rate": 1.7805612234934946e-05, + "loss": 0.825, + "step": 2197 + }, + { + "epoch": 1.1932681867535289, + "grad_norm": 5.582107046729241, + "learning_rate": 1.780341344862548e-05, + "loss": 0.8548, + "step": 2198 + }, + { + "epoch": 1.1938110749185669, + "grad_norm": 7.598179521038406, + "learning_rate": 1.78012136971774e-05, + "loss": 1.0015, + "step": 2199 + }, + { + "epoch": 1.1943539630836049, + "grad_norm": 7.4068666684121585, + "learning_rate": 1.7799012980862777e-05, + "loss": 1.0645, + "step": 2200 + }, + { + "epoch": 1.1948968512486429, + "grad_norm": 6.087565282375561, + "learning_rate": 1.7796811299953796e-05, + "loss": 0.9169, + "step": 2201 + }, + { + "epoch": 1.1954397394136809, + "grad_norm": 6.440325481735819, + "learning_rate": 1.7794608654722772e-05, + "loss": 1.0707, + "step": 2202 + }, + { + "epoch": 1.1959826275787189, + "grad_norm": 7.562903613167714, + "learning_rate": 1.7792405045442125e-05, + "loss": 1.0086, + "step": 2203 + }, + { + "epoch": 1.1965255157437569, + "grad_norm": 6.2377346284888, + "learning_rate": 1.77902004723844e-05, + "loss": 0.6111, + "step": 2204 + }, + { + "epoch": 1.1970684039087949, + "grad_norm": 5.978299124706355, + "learning_rate": 1.7787994935822268e-05, + "loss": 0.5782, + "step": 2205 + }, + { + "epoch": 1.1976112920738329, + "grad_norm": 5.9766366336106715, + "learning_rate": 1.778578843602851e-05, + "loss": 0.8409, + "step": 2206 + }, + { + "epoch": 1.1981541802388709, + "grad_norm": 7.700010929432555, + "learning_rate": 1.7783580973276027e-05, + "loss": 1.091, + "step": 2207 + }, + { + "epoch": 1.1986970684039089, + "grad_norm": 8.043903647823342, + "learning_rate": 1.778137254783785e-05, + "loss": 1.2894, + "step": 2208 + }, + { + "epoch": 1.1992399565689469, + "grad_norm": 8.709529168595635, + "learning_rate": 1.777916315998711e-05, + "loss": 1.5601, + "step": 2209 + }, + { + "epoch": 1.1997828447339849, + "grad_norm": 6.234520197542074, + "learning_rate": 1.7776952809997073e-05, + "loss": 1.0414, + "step": 2210 + }, + { + "epoch": 1.200325732899023, + "grad_norm": 5.8108372705767, + "learning_rate": 1.7774741498141116e-05, + "loss": 0.8811, + "step": 2211 + }, + { + "epoch": 1.200868621064061, + "grad_norm": 6.423388338643525, + "learning_rate": 1.7772529224692744e-05, + "loss": 0.8371, + "step": 2212 + }, + { + "epoch": 1.201411509229099, + "grad_norm": 7.740996825838731, + "learning_rate": 1.7770315989925565e-05, + "loss": 0.9993, + "step": 2213 + }, + { + "epoch": 1.201954397394137, + "grad_norm": 6.33257469290407, + "learning_rate": 1.7768101794113323e-05, + "loss": 0.8563, + "step": 2214 + }, + { + "epoch": 1.202497285559175, + "grad_norm": 8.126954888789392, + "learning_rate": 1.776588663752987e-05, + "loss": 0.9479, + "step": 2215 + }, + { + "epoch": 1.203040173724213, + "grad_norm": 7.781450888823224, + "learning_rate": 1.7763670520449178e-05, + "loss": 1.0559, + "step": 2216 + }, + { + "epoch": 1.203583061889251, + "grad_norm": 6.615763983931515, + "learning_rate": 1.7761453443145348e-05, + "loss": 0.7463, + "step": 2217 + }, + { + "epoch": 1.204125950054289, + "grad_norm": 10.193394757123674, + "learning_rate": 1.7759235405892584e-05, + "loss": 1.3597, + "step": 2218 + }, + { + "epoch": 1.204668838219327, + "grad_norm": 7.365896137130193, + "learning_rate": 1.7757016408965217e-05, + "loss": 0.5822, + "step": 2219 + }, + { + "epoch": 1.205211726384365, + "grad_norm": 8.530190774406401, + "learning_rate": 1.77547964526377e-05, + "loss": 1.0075, + "step": 2220 + }, + { + "epoch": 1.205754614549403, + "grad_norm": 7.143571088845206, + "learning_rate": 1.7752575537184597e-05, + "loss": 0.9395, + "step": 2221 + }, + { + "epoch": 1.206297502714441, + "grad_norm": 7.790375269151535, + "learning_rate": 1.7750353662880595e-05, + "loss": 1.0556, + "step": 2222 + }, + { + "epoch": 1.206840390879479, + "grad_norm": 6.5033471136457255, + "learning_rate": 1.77481308300005e-05, + "loss": 0.7464, + "step": 2223 + }, + { + "epoch": 1.207383279044517, + "grad_norm": 7.06026002468489, + "learning_rate": 1.774590703881924e-05, + "loss": 0.6315, + "step": 2224 + }, + { + "epoch": 1.207926167209555, + "grad_norm": 8.773109114641178, + "learning_rate": 1.774368228961185e-05, + "loss": 0.9884, + "step": 2225 + }, + { + "epoch": 1.208469055374593, + "grad_norm": 7.343364313617542, + "learning_rate": 1.774145658265349e-05, + "loss": 1.1542, + "step": 2226 + }, + { + "epoch": 1.209011943539631, + "grad_norm": 5.953950803365371, + "learning_rate": 1.773922991821944e-05, + "loss": 0.8485, + "step": 2227 + }, + { + "epoch": 1.209554831704669, + "grad_norm": 8.905484324028157, + "learning_rate": 1.7737002296585105e-05, + "loss": 1.0171, + "step": 2228 + }, + { + "epoch": 1.210097719869707, + "grad_norm": 7.158929117808072, + "learning_rate": 1.7734773718025992e-05, + "loss": 0.6134, + "step": 2229 + }, + { + "epoch": 1.210640608034745, + "grad_norm": 8.342029051568577, + "learning_rate": 1.7732544182817737e-05, + "loss": 0.9775, + "step": 2230 + }, + { + "epoch": 1.211183496199783, + "grad_norm": 5.587278578636139, + "learning_rate": 1.7730313691236098e-05, + "loss": 0.9395, + "step": 2231 + }, + { + "epoch": 1.211726384364821, + "grad_norm": 6.1948883535007475, + "learning_rate": 1.7728082243556936e-05, + "loss": 0.9977, + "step": 2232 + }, + { + "epoch": 1.212269272529859, + "grad_norm": 6.095377323439518, + "learning_rate": 1.7725849840056248e-05, + "loss": 0.9141, + "step": 2233 + }, + { + "epoch": 1.212812160694897, + "grad_norm": 8.71522968216411, + "learning_rate": 1.7723616481010137e-05, + "loss": 1.4356, + "step": 2234 + }, + { + "epoch": 1.213355048859935, + "grad_norm": 5.416863970522961, + "learning_rate": 1.7721382166694834e-05, + "loss": 0.6594, + "step": 2235 + }, + { + "epoch": 1.213897937024973, + "grad_norm": 8.046753742934493, + "learning_rate": 1.7719146897386674e-05, + "loss": 1.1187, + "step": 2236 + }, + { + "epoch": 1.214440825190011, + "grad_norm": 6.625446441699582, + "learning_rate": 1.7716910673362123e-05, + "loss": 0.8731, + "step": 2237 + }, + { + "epoch": 1.214983713355049, + "grad_norm": 6.47539429072121, + "learning_rate": 1.7714673494897767e-05, + "loss": 0.9473, + "step": 2238 + }, + { + "epoch": 1.215526601520087, + "grad_norm": 7.458434209050833, + "learning_rate": 1.7712435362270292e-05, + "loss": 0.7844, + "step": 2239 + }, + { + "epoch": 1.216069489685125, + "grad_norm": 6.935847404573923, + "learning_rate": 1.7710196275756524e-05, + "loss": 0.8431, + "step": 2240 + }, + { + "epoch": 1.216612377850163, + "grad_norm": 8.015017144687471, + "learning_rate": 1.770795623563339e-05, + "loss": 1.2329, + "step": 2241 + }, + { + "epoch": 1.217155266015201, + "grad_norm": 6.914919733505818, + "learning_rate": 1.7705715242177944e-05, + "loss": 1.1335, + "step": 2242 + }, + { + "epoch": 1.217698154180239, + "grad_norm": 5.900061304870694, + "learning_rate": 1.7703473295667354e-05, + "loss": 0.7453, + "step": 2243 + }, + { + "epoch": 1.218241042345277, + "grad_norm": 4.959614765477869, + "learning_rate": 1.770123039637891e-05, + "loss": 0.6093, + "step": 2244 + }, + { + "epoch": 1.2187839305103148, + "grad_norm": 8.038465432918807, + "learning_rate": 1.769898654459002e-05, + "loss": 1.0029, + "step": 2245 + }, + { + "epoch": 1.219326818675353, + "grad_norm": 8.08635156220498, + "learning_rate": 1.7696741740578204e-05, + "loss": 1.0839, + "step": 2246 + }, + { + "epoch": 1.2198697068403908, + "grad_norm": 7.985335036181301, + "learning_rate": 1.7694495984621097e-05, + "loss": 1.262, + "step": 2247 + }, + { + "epoch": 1.220412595005429, + "grad_norm": 8.31037377117703, + "learning_rate": 1.769224927699647e-05, + "loss": 1.0265, + "step": 2248 + }, + { + "epoch": 1.2209554831704668, + "grad_norm": 6.241752286447289, + "learning_rate": 1.769000161798219e-05, + "loss": 0.6179, + "step": 2249 + }, + { + "epoch": 1.221498371335505, + "grad_norm": 6.767956890107702, + "learning_rate": 1.7687753007856253e-05, + "loss": 0.9897, + "step": 2250 + }, + { + "epoch": 1.2220412595005428, + "grad_norm": 7.119621854687504, + "learning_rate": 1.7685503446896772e-05, + "loss": 1.0529, + "step": 2251 + }, + { + "epoch": 1.222584147665581, + "grad_norm": 6.37786350129091, + "learning_rate": 1.7683252935381976e-05, + "loss": 0.8575, + "step": 2252 + }, + { + "epoch": 1.2231270358306188, + "grad_norm": 9.280258739154695, + "learning_rate": 1.768100147359021e-05, + "loss": 0.959, + "step": 2253 + }, + { + "epoch": 1.223669923995657, + "grad_norm": 7.172898793975323, + "learning_rate": 1.7678749061799942e-05, + "loss": 1.3745, + "step": 2254 + }, + { + "epoch": 1.2242128121606948, + "grad_norm": 5.3297508420409825, + "learning_rate": 1.7676495700289753e-05, + "loss": 0.6697, + "step": 2255 + }, + { + "epoch": 1.224755700325733, + "grad_norm": 6.518884308754971, + "learning_rate": 1.767424138933834e-05, + "loss": 0.9854, + "step": 2256 + }, + { + "epoch": 1.2252985884907708, + "grad_norm": 7.939498291049385, + "learning_rate": 1.767198612922452e-05, + "loss": 0.8918, + "step": 2257 + }, + { + "epoch": 1.225841476655809, + "grad_norm": 6.309602610605377, + "learning_rate": 1.7669729920227226e-05, + "loss": 0.9939, + "step": 2258 + }, + { + "epoch": 1.2263843648208468, + "grad_norm": 7.462197154770194, + "learning_rate": 1.7667472762625518e-05, + "loss": 1.1299, + "step": 2259 + }, + { + "epoch": 1.226927252985885, + "grad_norm": 6.700031938849371, + "learning_rate": 1.7665214656698555e-05, + "loss": 0.8655, + "step": 2260 + }, + { + "epoch": 1.2274701411509228, + "grad_norm": 6.72124660471967, + "learning_rate": 1.7662955602725627e-05, + "loss": 0.7383, + "step": 2261 + }, + { + "epoch": 1.228013029315961, + "grad_norm": 6.071348581517431, + "learning_rate": 1.7660695600986137e-05, + "loss": 0.8571, + "step": 2262 + }, + { + "epoch": 1.2285559174809988, + "grad_norm": 5.914184858865564, + "learning_rate": 1.7658434651759605e-05, + "loss": 0.7494, + "step": 2263 + }, + { + "epoch": 1.229098805646037, + "grad_norm": 12.685614201985125, + "learning_rate": 1.7656172755325675e-05, + "loss": 1.608, + "step": 2264 + }, + { + "epoch": 1.2296416938110748, + "grad_norm": 7.001237249554534, + "learning_rate": 1.765390991196409e-05, + "loss": 1.0437, + "step": 2265 + }, + { + "epoch": 1.230184581976113, + "grad_norm": 8.232094967642231, + "learning_rate": 1.765164612195473e-05, + "loss": 0.8574, + "step": 2266 + }, + { + "epoch": 1.2307274701411508, + "grad_norm": 6.367083948330669, + "learning_rate": 1.7649381385577582e-05, + "loss": 0.7106, + "step": 2267 + }, + { + "epoch": 1.231270358306189, + "grad_norm": 6.2062819181758195, + "learning_rate": 1.7647115703112756e-05, + "loss": 0.6838, + "step": 2268 + }, + { + "epoch": 1.2318132464712268, + "grad_norm": 6.896096048797596, + "learning_rate": 1.7644849074840472e-05, + "loss": 0.9595, + "step": 2269 + }, + { + "epoch": 1.232356134636265, + "grad_norm": 5.915145487902616, + "learning_rate": 1.7642581501041067e-05, + "loss": 0.5741, + "step": 2270 + }, + { + "epoch": 1.2328990228013028, + "grad_norm": 5.966871119524657, + "learning_rate": 1.7640312981995004e-05, + "loss": 0.8686, + "step": 2271 + }, + { + "epoch": 1.233441910966341, + "grad_norm": 6.835680775363237, + "learning_rate": 1.7638043517982854e-05, + "loss": 0.9532, + "step": 2272 + }, + { + "epoch": 1.2339847991313788, + "grad_norm": 8.228029492129801, + "learning_rate": 1.7635773109285306e-05, + "loss": 1.0834, + "step": 2273 + }, + { + "epoch": 1.234527687296417, + "grad_norm": 8.272927259623488, + "learning_rate": 1.7633501756183174e-05, + "loss": 1.2591, + "step": 2274 + }, + { + "epoch": 1.2350705754614548, + "grad_norm": 6.755748928056399, + "learning_rate": 1.7631229458957377e-05, + "loss": 0.997, + "step": 2275 + }, + { + "epoch": 1.235613463626493, + "grad_norm": 7.335027842642818, + "learning_rate": 1.762895621788896e-05, + "loss": 0.8695, + "step": 2276 + }, + { + "epoch": 1.2361563517915308, + "grad_norm": 5.237971036796901, + "learning_rate": 1.7626682033259077e-05, + "loss": 0.6846, + "step": 2277 + }, + { + "epoch": 1.236699239956569, + "grad_norm": 6.333455514649705, + "learning_rate": 1.7624406905349004e-05, + "loss": 0.7532, + "step": 2278 + }, + { + "epoch": 1.2372421281216068, + "grad_norm": 6.255556726847127, + "learning_rate": 1.7622130834440136e-05, + "loss": 0.7801, + "step": 2279 + }, + { + "epoch": 1.237785016286645, + "grad_norm": 5.602832960919912, + "learning_rate": 1.761985382081398e-05, + "loss": 0.5105, + "step": 2280 + }, + { + "epoch": 1.2383279044516828, + "grad_norm": 5.814619917758832, + "learning_rate": 1.7617575864752155e-05, + "loss": 0.9034, + "step": 2281 + }, + { + "epoch": 1.238870792616721, + "grad_norm": 6.743402665368447, + "learning_rate": 1.7615296966536407e-05, + "loss": 0.7713, + "step": 2282 + }, + { + "epoch": 1.2394136807817588, + "grad_norm": 7.176469994889228, + "learning_rate": 1.7613017126448597e-05, + "loss": 0.957, + "step": 2283 + }, + { + "epoch": 1.239956568946797, + "grad_norm": 8.5427155590311, + "learning_rate": 1.7610736344770693e-05, + "loss": 0.8789, + "step": 2284 + }, + { + "epoch": 1.2404994571118348, + "grad_norm": 6.674205629131359, + "learning_rate": 1.7608454621784786e-05, + "loss": 1.0239, + "step": 2285 + }, + { + "epoch": 1.241042345276873, + "grad_norm": 7.687894188099568, + "learning_rate": 1.7606171957773088e-05, + "loss": 1.3164, + "step": 2286 + }, + { + "epoch": 1.2415852334419109, + "grad_norm": 6.331613648212601, + "learning_rate": 1.760388835301792e-05, + "loss": 0.8047, + "step": 2287 + }, + { + "epoch": 1.242128121606949, + "grad_norm": 7.0776252458357725, + "learning_rate": 1.760160380780172e-05, + "loss": 1.1893, + "step": 2288 + }, + { + "epoch": 1.2426710097719869, + "grad_norm": 5.434346705496786, + "learning_rate": 1.7599318322407047e-05, + "loss": 0.763, + "step": 2289 + }, + { + "epoch": 1.243213897937025, + "grad_norm": 6.0271106768536775, + "learning_rate": 1.759703189711657e-05, + "loss": 1.1129, + "step": 2290 + }, + { + "epoch": 1.2437567861020629, + "grad_norm": 7.286292908201965, + "learning_rate": 1.7594744532213088e-05, + "loss": 1.1136, + "step": 2291 + }, + { + "epoch": 1.244299674267101, + "grad_norm": 6.685086643433707, + "learning_rate": 1.7592456227979488e-05, + "loss": 0.724, + "step": 2292 + }, + { + "epoch": 1.2448425624321389, + "grad_norm": 7.184174802374185, + "learning_rate": 1.7590166984698807e-05, + "loss": 1.1471, + "step": 2293 + }, + { + "epoch": 1.245385450597177, + "grad_norm": 7.787563610771557, + "learning_rate": 1.7587876802654176e-05, + "loss": 1.01, + "step": 2294 + }, + { + "epoch": 1.2459283387622149, + "grad_norm": 7.036440767647583, + "learning_rate": 1.7585585682128848e-05, + "loss": 0.8612, + "step": 2295 + }, + { + "epoch": 1.246471226927253, + "grad_norm": 6.6372154433448705, + "learning_rate": 1.7583293623406192e-05, + "loss": 1.3394, + "step": 2296 + }, + { + "epoch": 1.2470141150922909, + "grad_norm": 7.707986168491473, + "learning_rate": 1.7581000626769697e-05, + "loss": 0.8688, + "step": 2297 + }, + { + "epoch": 1.247557003257329, + "grad_norm": 7.862359368033938, + "learning_rate": 1.757870669250296e-05, + "loss": 1.1999, + "step": 2298 + }, + { + "epoch": 1.2480998914223669, + "grad_norm": 7.942280078814929, + "learning_rate": 1.75764118208897e-05, + "loss": 1.241, + "step": 2299 + }, + { + "epoch": 1.248642779587405, + "grad_norm": 7.258825765708901, + "learning_rate": 1.7574116012213753e-05, + "loss": 0.8336, + "step": 2300 + }, + { + "epoch": 1.2491856677524429, + "grad_norm": 7.0149158036864065, + "learning_rate": 1.7571819266759066e-05, + "loss": 0.8637, + "step": 2301 + }, + { + "epoch": 1.249728555917481, + "grad_norm": 7.069351546856077, + "learning_rate": 1.7569521584809703e-05, + "loss": 0.8146, + "step": 2302 + }, + { + "epoch": 1.250271444082519, + "grad_norm": 4.6424312647112425, + "learning_rate": 1.756722296664985e-05, + "loss": 0.5056, + "step": 2303 + }, + { + "epoch": 1.2508143322475571, + "grad_norm": 7.857412677197525, + "learning_rate": 1.7564923412563797e-05, + "loss": 1.3072, + "step": 2304 + }, + { + "epoch": 1.251357220412595, + "grad_norm": 6.081845076053873, + "learning_rate": 1.7562622922835962e-05, + "loss": 1.0212, + "step": 2305 + }, + { + "epoch": 1.2519001085776331, + "grad_norm": 5.244697868381976, + "learning_rate": 1.7560321497750867e-05, + "loss": 0.9677, + "step": 2306 + }, + { + "epoch": 1.252442996742671, + "grad_norm": 11.027915321544773, + "learning_rate": 1.7558019137593162e-05, + "loss": 0.8984, + "step": 2307 + }, + { + "epoch": 1.2529858849077091, + "grad_norm": 5.955816485468965, + "learning_rate": 1.7555715842647603e-05, + "loss": 0.6282, + "step": 2308 + }, + { + "epoch": 1.253528773072747, + "grad_norm": 8.925991341226974, + "learning_rate": 1.755341161319907e-05, + "loss": 1.2121, + "step": 2309 + }, + { + "epoch": 1.2540716612377851, + "grad_norm": 7.947406679428915, + "learning_rate": 1.7551106449532547e-05, + "loss": 1.0545, + "step": 2310 + }, + { + "epoch": 1.254614549402823, + "grad_norm": 7.329950391674132, + "learning_rate": 1.7548800351933147e-05, + "loss": 1.0585, + "step": 2311 + }, + { + "epoch": 1.2551574375678611, + "grad_norm": 5.0931347466751955, + "learning_rate": 1.7546493320686084e-05, + "loss": 0.5533, + "step": 2312 + }, + { + "epoch": 1.255700325732899, + "grad_norm": 7.125279667319334, + "learning_rate": 1.7544185356076703e-05, + "loss": 1.1011, + "step": 2313 + }, + { + "epoch": 1.2562432138979371, + "grad_norm": 8.060697991979964, + "learning_rate": 1.7541876458390453e-05, + "loss": 0.8732, + "step": 2314 + }, + { + "epoch": 1.256786102062975, + "grad_norm": 9.732797197135078, + "learning_rate": 1.75395666279129e-05, + "loss": 1.5663, + "step": 2315 + }, + { + "epoch": 1.2573289902280131, + "grad_norm": 5.3673655570709435, + "learning_rate": 1.7537255864929733e-05, + "loss": 0.7057, + "step": 2316 + }, + { + "epoch": 1.257871878393051, + "grad_norm": 5.997561584131186, + "learning_rate": 1.7534944169726752e-05, + "loss": 0.8526, + "step": 2317 + }, + { + "epoch": 1.2584147665580891, + "grad_norm": 7.236766116786611, + "learning_rate": 1.753263154258986e-05, + "loss": 0.8921, + "step": 2318 + }, + { + "epoch": 1.258957654723127, + "grad_norm": 7.663917187467161, + "learning_rate": 1.7530317983805096e-05, + "loss": 0.9309, + "step": 2319 + }, + { + "epoch": 1.2595005428881652, + "grad_norm": 5.704193869412544, + "learning_rate": 1.75280034936586e-05, + "loss": 0.6291, + "step": 2320 + }, + { + "epoch": 1.260043431053203, + "grad_norm": 7.365455665485428, + "learning_rate": 1.7525688072436636e-05, + "loss": 0.9946, + "step": 2321 + }, + { + "epoch": 1.2605863192182412, + "grad_norm": 4.8267670121648205, + "learning_rate": 1.752337172042558e-05, + "loss": 0.8254, + "step": 2322 + }, + { + "epoch": 1.261129207383279, + "grad_norm": 7.231116660749922, + "learning_rate": 1.752105443791192e-05, + "loss": 1.3835, + "step": 2323 + }, + { + "epoch": 1.2616720955483172, + "grad_norm": 6.02849395913157, + "learning_rate": 1.7518736225182253e-05, + "loss": 0.6229, + "step": 2324 + }, + { + "epoch": 1.262214983713355, + "grad_norm": 6.458823304206971, + "learning_rate": 1.7516417082523314e-05, + "loss": 0.7707, + "step": 2325 + }, + { + "epoch": 1.2627578718783932, + "grad_norm": 7.163598438660781, + "learning_rate": 1.7514097010221927e-05, + "loss": 1.0007, + "step": 2326 + }, + { + "epoch": 1.263300760043431, + "grad_norm": 6.607394766457111, + "learning_rate": 1.7511776008565043e-05, + "loss": 1.0289, + "step": 2327 + }, + { + "epoch": 1.2638436482084692, + "grad_norm": 8.20658895752514, + "learning_rate": 1.7509454077839736e-05, + "loss": 0.7553, + "step": 2328 + }, + { + "epoch": 1.264386536373507, + "grad_norm": 8.963409997263136, + "learning_rate": 1.750713121833318e-05, + "loss": 0.9558, + "step": 2329 + }, + { + "epoch": 1.2649294245385452, + "grad_norm": 6.463632425962391, + "learning_rate": 1.7504807430332668e-05, + "loss": 0.6553, + "step": 2330 + }, + { + "epoch": 1.265472312703583, + "grad_norm": 6.936173778617146, + "learning_rate": 1.7502482714125612e-05, + "loss": 0.7861, + "step": 2331 + }, + { + "epoch": 1.2660152008686212, + "grad_norm": 6.975529559460763, + "learning_rate": 1.7500157069999536e-05, + "loss": 0.665, + "step": 2332 + }, + { + "epoch": 1.266558089033659, + "grad_norm": 6.768480922885786, + "learning_rate": 1.7497830498242082e-05, + "loss": 1.0163, + "step": 2333 + }, + { + "epoch": 1.2671009771986972, + "grad_norm": 6.357421024100016, + "learning_rate": 1.7495502999141004e-05, + "loss": 0.5897, + "step": 2334 + }, + { + "epoch": 1.267643865363735, + "grad_norm": 7.82539792030223, + "learning_rate": 1.7493174572984168e-05, + "loss": 1.0509, + "step": 2335 + }, + { + "epoch": 1.2681867535287732, + "grad_norm": 7.6523935004894215, + "learning_rate": 1.7490845220059554e-05, + "loss": 0.7603, + "step": 2336 + }, + { + "epoch": 1.268729641693811, + "grad_norm": 6.677360283026087, + "learning_rate": 1.7488514940655267e-05, + "loss": 1.1797, + "step": 2337 + }, + { + "epoch": 1.2692725298588492, + "grad_norm": 7.837606738097467, + "learning_rate": 1.7486183735059517e-05, + "loss": 1.1995, + "step": 2338 + }, + { + "epoch": 1.269815418023887, + "grad_norm": 8.913928805014676, + "learning_rate": 1.7483851603560634e-05, + "loss": 1.6007, + "step": 2339 + }, + { + "epoch": 1.2703583061889252, + "grad_norm": 8.467363918664116, + "learning_rate": 1.7481518546447058e-05, + "loss": 0.9515, + "step": 2340 + }, + { + "epoch": 1.270901194353963, + "grad_norm": 6.758882762498143, + "learning_rate": 1.747918456400734e-05, + "loss": 0.7935, + "step": 2341 + }, + { + "epoch": 1.2714440825190012, + "grad_norm": 5.455764070482765, + "learning_rate": 1.7476849656530158e-05, + "loss": 0.4623, + "step": 2342 + }, + { + "epoch": 1.271986970684039, + "grad_norm": 8.417334291043444, + "learning_rate": 1.7474513824304294e-05, + "loss": 1.1563, + "step": 2343 + }, + { + "epoch": 1.2725298588490772, + "grad_norm": 5.61084844590966, + "learning_rate": 1.7472177067618646e-05, + "loss": 0.8566, + "step": 2344 + }, + { + "epoch": 1.273072747014115, + "grad_norm": 7.726478476188633, + "learning_rate": 1.746983938676223e-05, + "loss": 1.0568, + "step": 2345 + }, + { + "epoch": 1.2736156351791532, + "grad_norm": 8.882013629978706, + "learning_rate": 1.7467500782024177e-05, + "loss": 1.0921, + "step": 2346 + }, + { + "epoch": 1.274158523344191, + "grad_norm": 6.976574380776494, + "learning_rate": 1.746516125369372e-05, + "loss": 1.1222, + "step": 2347 + }, + { + "epoch": 1.2747014115092292, + "grad_norm": 7.641347465211616, + "learning_rate": 1.7462820802060224e-05, + "loss": 1.0672, + "step": 2348 + }, + { + "epoch": 1.275244299674267, + "grad_norm": 6.473243122755322, + "learning_rate": 1.7460479427413156e-05, + "loss": 0.9019, + "step": 2349 + }, + { + "epoch": 1.2757871878393052, + "grad_norm": 6.494958002593054, + "learning_rate": 1.7458137130042105e-05, + "loss": 0.7862, + "step": 2350 + }, + { + "epoch": 1.276330076004343, + "grad_norm": 6.54459648554074, + "learning_rate": 1.7455793910236764e-05, + "loss": 0.9572, + "step": 2351 + }, + { + "epoch": 1.2768729641693812, + "grad_norm": 6.249218254368162, + "learning_rate": 1.7453449768286952e-05, + "loss": 0.9415, + "step": 2352 + }, + { + "epoch": 1.277415852334419, + "grad_norm": 6.335956299458767, + "learning_rate": 1.7451104704482592e-05, + "loss": 0.7647, + "step": 2353 + }, + { + "epoch": 1.2779587404994572, + "grad_norm": 6.296362619304004, + "learning_rate": 1.744875871911373e-05, + "loss": 0.6293, + "step": 2354 + }, + { + "epoch": 1.278501628664495, + "grad_norm": 6.629983181103369, + "learning_rate": 1.7446411812470516e-05, + "loss": 0.7816, + "step": 2355 + }, + { + "epoch": 1.2790445168295332, + "grad_norm": 6.921460108918273, + "learning_rate": 1.744406398484322e-05, + "loss": 0.6373, + "step": 2356 + }, + { + "epoch": 1.279587404994571, + "grad_norm": 9.4421099603369, + "learning_rate": 1.7441715236522228e-05, + "loss": 1.5291, + "step": 2357 + }, + { + "epoch": 1.2801302931596092, + "grad_norm": 7.838644313955947, + "learning_rate": 1.7439365567798035e-05, + "loss": 0.7934, + "step": 2358 + }, + { + "epoch": 1.280673181324647, + "grad_norm": 5.7659414273416, + "learning_rate": 1.743701497896125e-05, + "loss": 0.8358, + "step": 2359 + }, + { + "epoch": 1.2812160694896852, + "grad_norm": 4.731198780322981, + "learning_rate": 1.7434663470302602e-05, + "loss": 0.7357, + "step": 2360 + }, + { + "epoch": 1.281758957654723, + "grad_norm": 8.993227122901354, + "learning_rate": 1.7432311042112926e-05, + "loss": 1.0679, + "step": 2361 + }, + { + "epoch": 1.2823018458197613, + "grad_norm": 7.193753575449645, + "learning_rate": 1.7429957694683175e-05, + "loss": 0.9257, + "step": 2362 + }, + { + "epoch": 1.282844733984799, + "grad_norm": 7.1027186942584635, + "learning_rate": 1.7427603428304416e-05, + "loss": 0.64, + "step": 2363 + }, + { + "epoch": 1.2833876221498373, + "grad_norm": 9.441739719572269, + "learning_rate": 1.7425248243267824e-05, + "loss": 1.3236, + "step": 2364 + }, + { + "epoch": 1.283930510314875, + "grad_norm": 6.478965109994284, + "learning_rate": 1.7422892139864696e-05, + "loss": 0.7726, + "step": 2365 + }, + { + "epoch": 1.2844733984799133, + "grad_norm": 6.372951649068815, + "learning_rate": 1.742053511838644e-05, + "loss": 0.7785, + "step": 2366 + }, + { + "epoch": 1.285016286644951, + "grad_norm": 10.052356612127076, + "learning_rate": 1.7418177179124574e-05, + "loss": 1.7877, + "step": 2367 + }, + { + "epoch": 1.2855591748099893, + "grad_norm": 5.24633801429042, + "learning_rate": 1.741581832237073e-05, + "loss": 0.4195, + "step": 2368 + }, + { + "epoch": 1.286102062975027, + "grad_norm": 7.958362658884473, + "learning_rate": 1.7413458548416656e-05, + "loss": 1.1782, + "step": 2369 + }, + { + "epoch": 1.2866449511400653, + "grad_norm": 7.923031881378813, + "learning_rate": 1.7411097857554216e-05, + "loss": 1.0861, + "step": 2370 + }, + { + "epoch": 1.287187839305103, + "grad_norm": 6.53099588563486, + "learning_rate": 1.7408736250075378e-05, + "loss": 0.8717, + "step": 2371 + }, + { + "epoch": 1.2877307274701413, + "grad_norm": 7.364911313605863, + "learning_rate": 1.740637372627224e-05, + "loss": 0.9341, + "step": 2372 + }, + { + "epoch": 1.288273615635179, + "grad_norm": 8.083622859612968, + "learning_rate": 1.740401028643699e-05, + "loss": 1.1654, + "step": 2373 + }, + { + "epoch": 1.2888165038002173, + "grad_norm": 6.561315990238234, + "learning_rate": 1.7401645930861944e-05, + "loss": 0.8088, + "step": 2374 + }, + { + "epoch": 1.289359391965255, + "grad_norm": 6.926193447339149, + "learning_rate": 1.739928065983954e-05, + "loss": 0.8335, + "step": 2375 + }, + { + "epoch": 1.2899022801302933, + "grad_norm": 7.876730774519387, + "learning_rate": 1.7396914473662306e-05, + "loss": 0.9969, + "step": 2376 + }, + { + "epoch": 1.290445168295331, + "grad_norm": 6.513645878383447, + "learning_rate": 1.73945473726229e-05, + "loss": 0.9302, + "step": 2377 + }, + { + "epoch": 1.2909880564603693, + "grad_norm": 6.441562626276137, + "learning_rate": 1.73921793570141e-05, + "loss": 1.0318, + "step": 2378 + }, + { + "epoch": 1.291530944625407, + "grad_norm": 5.226443422910506, + "learning_rate": 1.7389810427128766e-05, + "loss": 0.6142, + "step": 2379 + }, + { + "epoch": 1.2920738327904453, + "grad_norm": 7.436131751002477, + "learning_rate": 1.7387440583259906e-05, + "loss": 0.9907, + "step": 2380 + }, + { + "epoch": 1.292616720955483, + "grad_norm": 6.335395136223753, + "learning_rate": 1.7385069825700615e-05, + "loss": 0.8431, + "step": 2381 + }, + { + "epoch": 1.2931596091205213, + "grad_norm": 8.93013417703781, + "learning_rate": 1.738269815474412e-05, + "loss": 1.3162, + "step": 2382 + }, + { + "epoch": 1.293702497285559, + "grad_norm": 6.616000665685319, + "learning_rate": 1.7380325570683754e-05, + "loss": 0.81, + "step": 2383 + }, + { + "epoch": 1.2942453854505973, + "grad_norm": 9.126678309827314, + "learning_rate": 1.737795207381296e-05, + "loss": 1.4976, + "step": 2384 + }, + { + "epoch": 1.294788273615635, + "grad_norm": 9.726273416685123, + "learning_rate": 1.737557766442529e-05, + "loss": 1.0284, + "step": 2385 + }, + { + "epoch": 1.2953311617806733, + "grad_norm": 6.207571191669611, + "learning_rate": 1.737320234281442e-05, + "loss": 0.9409, + "step": 2386 + }, + { + "epoch": 1.295874049945711, + "grad_norm": 6.150632184961992, + "learning_rate": 1.7370826109274136e-05, + "loss": 0.7558, + "step": 2387 + }, + { + "epoch": 1.2964169381107493, + "grad_norm": 6.555861715157727, + "learning_rate": 1.736844896409833e-05, + "loss": 1.4217, + "step": 2388 + }, + { + "epoch": 1.296959826275787, + "grad_norm": 7.291607947091272, + "learning_rate": 1.7366070907581015e-05, + "loss": 0.9969, + "step": 2389 + }, + { + "epoch": 1.2975027144408253, + "grad_norm": 7.003873752430508, + "learning_rate": 1.7363691940016307e-05, + "loss": 0.9569, + "step": 2390 + }, + { + "epoch": 1.298045602605863, + "grad_norm": 6.355975398274633, + "learning_rate": 1.7361312061698444e-05, + "loss": 0.9244, + "step": 2391 + }, + { + "epoch": 1.2985884907709013, + "grad_norm": 5.093818720596728, + "learning_rate": 1.7358931272921773e-05, + "loss": 0.5581, + "step": 2392 + }, + { + "epoch": 1.299131378935939, + "grad_norm": 4.532047411652702, + "learning_rate": 1.7356549573980753e-05, + "loss": 0.5153, + "step": 2393 + }, + { + "epoch": 1.2996742671009773, + "grad_norm": 7.378796483754393, + "learning_rate": 1.735416696516996e-05, + "loss": 0.9631, + "step": 2394 + }, + { + "epoch": 1.3002171552660151, + "grad_norm": 6.186232096125507, + "learning_rate": 1.7351783446784075e-05, + "loss": 0.7876, + "step": 2395 + }, + { + "epoch": 1.3007600434310533, + "grad_norm": 7.851387526494335, + "learning_rate": 1.7349399019117897e-05, + "loss": 0.8983, + "step": 2396 + }, + { + "epoch": 1.3013029315960911, + "grad_norm": 7.275506442499215, + "learning_rate": 1.7347013682466335e-05, + "loss": 0.8334, + "step": 2397 + }, + { + "epoch": 1.3018458197611293, + "grad_norm": 6.264072188984264, + "learning_rate": 1.7344627437124407e-05, + "loss": 0.6585, + "step": 2398 + }, + { + "epoch": 1.3023887079261671, + "grad_norm": 7.537073303434296, + "learning_rate": 1.734224028338726e-05, + "loss": 1.1033, + "step": 2399 + }, + { + "epoch": 1.3029315960912053, + "grad_norm": 6.850281208830108, + "learning_rate": 1.7339852221550126e-05, + "loss": 0.8133, + "step": 2400 + }, + { + "epoch": 1.3034744842562431, + "grad_norm": 8.089656198012262, + "learning_rate": 1.7337463251908374e-05, + "loss": 1.1342, + "step": 2401 + }, + { + "epoch": 1.3040173724212814, + "grad_norm": 6.2919884156176105, + "learning_rate": 1.733507337475747e-05, + "loss": 0.6533, + "step": 2402 + }, + { + "epoch": 1.3045602605863191, + "grad_norm": 5.114953151143873, + "learning_rate": 1.7332682590393008e-05, + "loss": 0.7451, + "step": 2403 + }, + { + "epoch": 1.3051031487513574, + "grad_norm": 6.18470588828327, + "learning_rate": 1.733029089911067e-05, + "loss": 1.1299, + "step": 2404 + }, + { + "epoch": 1.3056460369163951, + "grad_norm": 5.950129534147113, + "learning_rate": 1.7327898301206273e-05, + "loss": 0.7329, + "step": 2405 + }, + { + "epoch": 1.3061889250814334, + "grad_norm": 6.293661255114301, + "learning_rate": 1.7325504796975732e-05, + "loss": 0.7535, + "step": 2406 + }, + { + "epoch": 1.3067318132464711, + "grad_norm": 7.682262802579054, + "learning_rate": 1.732311038671509e-05, + "loss": 1.2399, + "step": 2407 + }, + { + "epoch": 1.3072747014115094, + "grad_norm": 7.314299514485556, + "learning_rate": 1.7320715070720478e-05, + "loss": 0.9298, + "step": 2408 + }, + { + "epoch": 1.3078175895765471, + "grad_norm": 8.921126356976833, + "learning_rate": 1.7318318849288158e-05, + "loss": 1.4254, + "step": 2409 + }, + { + "epoch": 1.3083604777415854, + "grad_norm": 7.060288459365394, + "learning_rate": 1.7315921722714503e-05, + "loss": 0.9428, + "step": 2410 + }, + { + "epoch": 1.3089033659066232, + "grad_norm": 8.568251970472561, + "learning_rate": 1.7313523691295988e-05, + "loss": 0.9011, + "step": 2411 + }, + { + "epoch": 1.3094462540716614, + "grad_norm": 6.049363733159396, + "learning_rate": 1.7311124755329206e-05, + "loss": 1.0718, + "step": 2412 + }, + { + "epoch": 1.3099891422366992, + "grad_norm": 8.287809242433424, + "learning_rate": 1.7308724915110864e-05, + "loss": 0.9137, + "step": 2413 + }, + { + "epoch": 1.3105320304017374, + "grad_norm": 8.436233795197621, + "learning_rate": 1.7306324170937774e-05, + "loss": 1.2427, + "step": 2414 + }, + { + "epoch": 1.3110749185667752, + "grad_norm": 7.035710404708695, + "learning_rate": 1.7303922523106863e-05, + "loss": 0.8206, + "step": 2415 + }, + { + "epoch": 1.3116178067318134, + "grad_norm": 5.9728457186472275, + "learning_rate": 1.730151997191518e-05, + "loss": 0.7307, + "step": 2416 + }, + { + "epoch": 1.3121606948968512, + "grad_norm": 6.296494925379207, + "learning_rate": 1.7299116517659865e-05, + "loss": 0.9808, + "step": 2417 + }, + { + "epoch": 1.3127035830618892, + "grad_norm": 7.672949954931739, + "learning_rate": 1.7296712160638192e-05, + "loss": 1.0289, + "step": 2418 + }, + { + "epoch": 1.3132464712269272, + "grad_norm": 5.958844838187463, + "learning_rate": 1.7294306901147525e-05, + "loss": 0.9657, + "step": 2419 + }, + { + "epoch": 1.3137893593919652, + "grad_norm": 6.5874736925476745, + "learning_rate": 1.7291900739485356e-05, + "loss": 0.8701, + "step": 2420 + }, + { + "epoch": 1.3143322475570032, + "grad_norm": 8.273467490572564, + "learning_rate": 1.7289493675949282e-05, + "loss": 0.9353, + "step": 2421 + }, + { + "epoch": 1.3148751357220412, + "grad_norm": 8.115143215098843, + "learning_rate": 1.7287085710837013e-05, + "loss": 0.6241, + "step": 2422 + }, + { + "epoch": 1.3154180238870792, + "grad_norm": 7.004703854660018, + "learning_rate": 1.7284676844446368e-05, + "loss": 0.801, + "step": 2423 + }, + { + "epoch": 1.3159609120521172, + "grad_norm": 6.039976470078682, + "learning_rate": 1.728226707707528e-05, + "loss": 1.085, + "step": 2424 + }, + { + "epoch": 1.3165038002171552, + "grad_norm": 5.318299059101813, + "learning_rate": 1.72798564090218e-05, + "loss": 0.4945, + "step": 2425 + }, + { + "epoch": 1.3170466883821932, + "grad_norm": 6.095578475774351, + "learning_rate": 1.727744484058407e-05, + "loss": 0.7696, + "step": 2426 + }, + { + "epoch": 1.3175895765472312, + "grad_norm": 6.4886080927006695, + "learning_rate": 1.7275032372060368e-05, + "loss": 0.5364, + "step": 2427 + }, + { + "epoch": 1.3181324647122692, + "grad_norm": 8.681669485105683, + "learning_rate": 1.7272619003749066e-05, + "loss": 0.9132, + "step": 2428 + }, + { + "epoch": 1.3186753528773072, + "grad_norm": 6.819144122945338, + "learning_rate": 1.7270204735948653e-05, + "loss": 0.7867, + "step": 2429 + }, + { + "epoch": 1.3192182410423452, + "grad_norm": 6.65338033222052, + "learning_rate": 1.7267789568957734e-05, + "loss": 0.7236, + "step": 2430 + }, + { + "epoch": 1.3197611292073832, + "grad_norm": 6.622862530496735, + "learning_rate": 1.7265373503075014e-05, + "loss": 0.9038, + "step": 2431 + }, + { + "epoch": 1.3203040173724212, + "grad_norm": 6.284216416307323, + "learning_rate": 1.7262956538599323e-05, + "loss": 0.677, + "step": 2432 + }, + { + "epoch": 1.3208469055374592, + "grad_norm": 9.150803536423062, + "learning_rate": 1.7260538675829593e-05, + "loss": 1.7953, + "step": 2433 + }, + { + "epoch": 1.3213897937024972, + "grad_norm": 9.236250071710021, + "learning_rate": 1.7258119915064867e-05, + "loss": 0.8568, + "step": 2434 + }, + { + "epoch": 1.3219326818675352, + "grad_norm": 7.061272817814159, + "learning_rate": 1.72557002566043e-05, + "loss": 0.9355, + "step": 2435 + }, + { + "epoch": 1.3224755700325732, + "grad_norm": 6.786187626307571, + "learning_rate": 1.7253279700747164e-05, + "loss": 0.7707, + "step": 2436 + }, + { + "epoch": 1.3230184581976112, + "grad_norm": 5.592273858747665, + "learning_rate": 1.725085824779283e-05, + "loss": 0.5932, + "step": 2437 + }, + { + "epoch": 1.3235613463626492, + "grad_norm": 8.137496643840409, + "learning_rate": 1.72484358980408e-05, + "loss": 1.0244, + "step": 2438 + }, + { + "epoch": 1.3241042345276872, + "grad_norm": 5.981122138792478, + "learning_rate": 1.724601265179066e-05, + "loss": 0.9807, + "step": 2439 + }, + { + "epoch": 1.3246471226927252, + "grad_norm": 6.912027532729551, + "learning_rate": 1.7243588509342127e-05, + "loss": 0.8673, + "step": 2440 + }, + { + "epoch": 1.3251900108577632, + "grad_norm": 5.836398731450734, + "learning_rate": 1.7241163470995024e-05, + "loss": 0.5738, + "step": 2441 + }, + { + "epoch": 1.3257328990228012, + "grad_norm": 5.0689326588185795, + "learning_rate": 1.7238737537049283e-05, + "loss": 0.4449, + "step": 2442 + }, + { + "epoch": 1.3262757871878392, + "grad_norm": 6.628337387569613, + "learning_rate": 1.7236310707804943e-05, + "loss": 0.7975, + "step": 2443 + }, + { + "epoch": 1.3268186753528772, + "grad_norm": 9.359654484357105, + "learning_rate": 1.7233882983562168e-05, + "loss": 0.8996, + "step": 2444 + }, + { + "epoch": 1.3273615635179152, + "grad_norm": 5.979606493030731, + "learning_rate": 1.723145436462121e-05, + "loss": 0.732, + "step": 2445 + }, + { + "epoch": 1.3279044516829532, + "grad_norm": 6.393853922089353, + "learning_rate": 1.7229024851282453e-05, + "loss": 0.9713, + "step": 2446 + }, + { + "epoch": 1.3284473398479912, + "grad_norm": 9.783386290725945, + "learning_rate": 1.722659444384638e-05, + "loss": 1.2991, + "step": 2447 + }, + { + "epoch": 1.3289902280130292, + "grad_norm": 8.471557152983001, + "learning_rate": 1.722416314261359e-05, + "loss": 0.8517, + "step": 2448 + }, + { + "epoch": 1.3295331161780672, + "grad_norm": 7.20996107804807, + "learning_rate": 1.7221730947884793e-05, + "loss": 1.0892, + "step": 2449 + }, + { + "epoch": 1.3300760043431052, + "grad_norm": 9.97079093211024, + "learning_rate": 1.7219297859960796e-05, + "loss": 0.8866, + "step": 2450 + }, + { + "epoch": 1.3306188925081432, + "grad_norm": 5.88899239355285, + "learning_rate": 1.7216863879142536e-05, + "loss": 0.7286, + "step": 2451 + }, + { + "epoch": 1.3311617806731812, + "grad_norm": 8.317968838525326, + "learning_rate": 1.7214429005731054e-05, + "loss": 0.9374, + "step": 2452 + }, + { + "epoch": 1.3317046688382193, + "grad_norm": 8.278980033154179, + "learning_rate": 1.721199324002749e-05, + "loss": 1.2977, + "step": 2453 + }, + { + "epoch": 1.3322475570032573, + "grad_norm": 7.50073001311901, + "learning_rate": 1.7209556582333106e-05, + "loss": 0.9479, + "step": 2454 + }, + { + "epoch": 1.3327904451682953, + "grad_norm": 9.34140520001984, + "learning_rate": 1.720711903294928e-05, + "loss": 1.1834, + "step": 2455 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 6.155163014555799, + "learning_rate": 1.720468059217748e-05, + "loss": 1.0674, + "step": 2456 + }, + { + "epoch": 1.3338762214983713, + "grad_norm": 8.207152122618293, + "learning_rate": 1.7202241260319305e-05, + "loss": 1.1623, + "step": 2457 + }, + { + "epoch": 1.3344191096634093, + "grad_norm": 8.477127177193351, + "learning_rate": 1.719980103767645e-05, + "loss": 1.0812, + "step": 2458 + }, + { + "epoch": 1.3349619978284473, + "grad_norm": 9.352009366934986, + "learning_rate": 1.7197359924550726e-05, + "loss": 0.9593, + "step": 2459 + }, + { + "epoch": 1.3355048859934853, + "grad_norm": 6.377220189803284, + "learning_rate": 1.719491792124406e-05, + "loss": 0.8569, + "step": 2460 + }, + { + "epoch": 1.3360477741585233, + "grad_norm": 10.372722210417326, + "learning_rate": 1.7192475028058475e-05, + "loss": 0.8778, + "step": 2461 + }, + { + "epoch": 1.3365906623235613, + "grad_norm": 6.943692981514271, + "learning_rate": 1.7190031245296118e-05, + "loss": 0.8143, + "step": 2462 + }, + { + "epoch": 1.3371335504885993, + "grad_norm": 8.53588364120296, + "learning_rate": 1.7187586573259237e-05, + "loss": 1.2608, + "step": 2463 + }, + { + "epoch": 1.3376764386536373, + "grad_norm": 6.85497949710551, + "learning_rate": 1.7185141012250195e-05, + "loss": 0.621, + "step": 2464 + }, + { + "epoch": 1.3382193268186753, + "grad_norm": 9.575890345813113, + "learning_rate": 1.7182694562571458e-05, + "loss": 1.5686, + "step": 2465 + }, + { + "epoch": 1.3387622149837133, + "grad_norm": 8.637844905126164, + "learning_rate": 1.718024722452561e-05, + "loss": 0.8287, + "step": 2466 + }, + { + "epoch": 1.3393051031487513, + "grad_norm": 7.464168837468054, + "learning_rate": 1.7177798998415344e-05, + "loss": 0.9837, + "step": 2467 + }, + { + "epoch": 1.3398479913137893, + "grad_norm": 9.27927006463804, + "learning_rate": 1.7175349884543458e-05, + "loss": 1.3479, + "step": 2468 + }, + { + "epoch": 1.3403908794788273, + "grad_norm": 6.9150180643503765, + "learning_rate": 1.717289988321286e-05, + "loss": 0.7731, + "step": 2469 + }, + { + "epoch": 1.3409337676438653, + "grad_norm": 8.264286643294954, + "learning_rate": 1.7170448994726574e-05, + "loss": 1.1252, + "step": 2470 + }, + { + "epoch": 1.3414766558089033, + "grad_norm": 6.1769106952218715, + "learning_rate": 1.7167997219387728e-05, + "loss": 0.6077, + "step": 2471 + }, + { + "epoch": 1.3420195439739413, + "grad_norm": 6.00046554758586, + "learning_rate": 1.7165544557499563e-05, + "loss": 0.9807, + "step": 2472 + }, + { + "epoch": 1.3425624321389793, + "grad_norm": 5.684026968717126, + "learning_rate": 1.7163091009365427e-05, + "loss": 0.8755, + "step": 2473 + }, + { + "epoch": 1.3431053203040173, + "grad_norm": 8.910694398982464, + "learning_rate": 1.7160636575288777e-05, + "loss": 1.1781, + "step": 2474 + }, + { + "epoch": 1.3436482084690553, + "grad_norm": 6.090059050660724, + "learning_rate": 1.7158181255573184e-05, + "loss": 0.8757, + "step": 2475 + }, + { + "epoch": 1.3441910966340933, + "grad_norm": 5.3693699493967575, + "learning_rate": 1.7155725050522325e-05, + "loss": 0.4111, + "step": 2476 + }, + { + "epoch": 1.3447339847991313, + "grad_norm": 6.4883376045754675, + "learning_rate": 1.7153267960439987e-05, + "loss": 1.1667, + "step": 2477 + }, + { + "epoch": 1.3452768729641693, + "grad_norm": 6.9649527868325585, + "learning_rate": 1.7150809985630065e-05, + "loss": 1.2642, + "step": 2478 + }, + { + "epoch": 1.3458197611292073, + "grad_norm": 7.458967124353787, + "learning_rate": 1.714835112639657e-05, + "loss": 1.1198, + "step": 2479 + }, + { + "epoch": 1.3463626492942453, + "grad_norm": 7.4820540779231015, + "learning_rate": 1.7145891383043613e-05, + "loss": 1.069, + "step": 2480 + }, + { + "epoch": 1.3469055374592833, + "grad_norm": 6.8442083181868165, + "learning_rate": 1.7143430755875422e-05, + "loss": 1.1187, + "step": 2481 + }, + { + "epoch": 1.3474484256243213, + "grad_norm": 8.389465670773259, + "learning_rate": 1.7140969245196332e-05, + "loss": 1.216, + "step": 2482 + }, + { + "epoch": 1.3479913137893593, + "grad_norm": 8.099481746156716, + "learning_rate": 1.713850685131078e-05, + "loss": 1.1365, + "step": 2483 + }, + { + "epoch": 1.3485342019543973, + "grad_norm": 7.005684437567773, + "learning_rate": 1.713604357452333e-05, + "loss": 0.7665, + "step": 2484 + }, + { + "epoch": 1.3490770901194353, + "grad_norm": 6.032904961736961, + "learning_rate": 1.7133579415138634e-05, + "loss": 0.7527, + "step": 2485 + }, + { + "epoch": 1.3496199782844733, + "grad_norm": 7.644359802639278, + "learning_rate": 1.7131114373461466e-05, + "loss": 0.7293, + "step": 2486 + }, + { + "epoch": 1.3501628664495113, + "grad_norm": 9.681466854129468, + "learning_rate": 1.7128648449796706e-05, + "loss": 1.0718, + "step": 2487 + }, + { + "epoch": 1.3507057546145493, + "grad_norm": 7.703820713192501, + "learning_rate": 1.7126181644449348e-05, + "loss": 1.0049, + "step": 2488 + }, + { + "epoch": 1.3512486427795873, + "grad_norm": 8.313943023532504, + "learning_rate": 1.7123713957724482e-05, + "loss": 1.1647, + "step": 2489 + }, + { + "epoch": 1.3517915309446253, + "grad_norm": 6.869318843676937, + "learning_rate": 1.7121245389927326e-05, + "loss": 0.7056, + "step": 2490 + }, + { + "epoch": 1.3523344191096633, + "grad_norm": 8.354366894281007, + "learning_rate": 1.7118775941363186e-05, + "loss": 1.4107, + "step": 2491 + }, + { + "epoch": 1.3528773072747013, + "grad_norm": 7.594996874862703, + "learning_rate": 1.7116305612337493e-05, + "loss": 1.3947, + "step": 2492 + }, + { + "epoch": 1.3534201954397393, + "grad_norm": 5.912640331169303, + "learning_rate": 1.7113834403155782e-05, + "loss": 0.678, + "step": 2493 + }, + { + "epoch": 1.3539630836047774, + "grad_norm": 8.603089873660597, + "learning_rate": 1.7111362314123693e-05, + "loss": 1.1846, + "step": 2494 + }, + { + "epoch": 1.3545059717698154, + "grad_norm": 5.899587064063632, + "learning_rate": 1.710888934554698e-05, + "loss": 0.7241, + "step": 2495 + }, + { + "epoch": 1.3550488599348534, + "grad_norm": 5.1366064982711315, + "learning_rate": 1.7106415497731502e-05, + "loss": 0.6714, + "step": 2496 + }, + { + "epoch": 1.3555917480998914, + "grad_norm": 6.55484035818926, + "learning_rate": 1.710394077098323e-05, + "loss": 1.0163, + "step": 2497 + }, + { + "epoch": 1.3561346362649294, + "grad_norm": 7.060748735507699, + "learning_rate": 1.710146516560824e-05, + "loss": 0.8407, + "step": 2498 + }, + { + "epoch": 1.3566775244299674, + "grad_norm": 7.488417378829399, + "learning_rate": 1.709898868191272e-05, + "loss": 1.0936, + "step": 2499 + }, + { + "epoch": 1.3572204125950054, + "grad_norm": 5.616560879457259, + "learning_rate": 1.7096511320202965e-05, + "loss": 0.6962, + "step": 2500 + }, + { + "epoch": 1.3577633007600434, + "grad_norm": 6.5457002772857775, + "learning_rate": 1.7094033080785384e-05, + "loss": 0.6338, + "step": 2501 + }, + { + "epoch": 1.3583061889250814, + "grad_norm": 5.788382651374094, + "learning_rate": 1.709155396396648e-05, + "loss": 0.9251, + "step": 2502 + }, + { + "epoch": 1.3588490770901194, + "grad_norm": 5.94278314769549, + "learning_rate": 1.7089073970052883e-05, + "loss": 0.5738, + "step": 2503 + }, + { + "epoch": 1.3593919652551574, + "grad_norm": 8.987615230603502, + "learning_rate": 1.7086593099351318e-05, + "loss": 1.1561, + "step": 2504 + }, + { + "epoch": 1.3599348534201954, + "grad_norm": 6.25509019167934, + "learning_rate": 1.7084111352168627e-05, + "loss": 0.5584, + "step": 2505 + }, + { + "epoch": 1.3604777415852334, + "grad_norm": 6.42626784615777, + "learning_rate": 1.708162872881175e-05, + "loss": 0.6385, + "step": 2506 + }, + { + "epoch": 1.3610206297502714, + "grad_norm": 7.342680004921898, + "learning_rate": 1.707914522958775e-05, + "loss": 1.0729, + "step": 2507 + }, + { + "epoch": 1.3615635179153094, + "grad_norm": 6.858864236878158, + "learning_rate": 1.707666085480378e-05, + "loss": 0.8641, + "step": 2508 + }, + { + "epoch": 1.3621064060803474, + "grad_norm": 6.176457712037886, + "learning_rate": 1.707417560476712e-05, + "loss": 0.9472, + "step": 2509 + }, + { + "epoch": 1.3626492942453854, + "grad_norm": 7.346954574146698, + "learning_rate": 1.7071689479785145e-05, + "loss": 1.0253, + "step": 2510 + }, + { + "epoch": 1.3631921824104234, + "grad_norm": 4.9331066104603085, + "learning_rate": 1.7069202480165344e-05, + "loss": 0.7307, + "step": 2511 + }, + { + "epoch": 1.3637350705754614, + "grad_norm": 8.316711829847868, + "learning_rate": 1.7066714606215316e-05, + "loss": 0.8078, + "step": 2512 + }, + { + "epoch": 1.3642779587404994, + "grad_norm": 8.079444940270628, + "learning_rate": 1.706422585824276e-05, + "loss": 1.2922, + "step": 2513 + }, + { + "epoch": 1.3648208469055374, + "grad_norm": 7.503741374806181, + "learning_rate": 1.7061736236555494e-05, + "loss": 0.7943, + "step": 2514 + }, + { + "epoch": 1.3653637350705754, + "grad_norm": 6.4023293864360635, + "learning_rate": 1.7059245741461435e-05, + "loss": 1.0111, + "step": 2515 + }, + { + "epoch": 1.3659066232356134, + "grad_norm": 5.7325920409358595, + "learning_rate": 1.705675437326861e-05, + "loss": 1.1201, + "step": 2516 + }, + { + "epoch": 1.3664495114006514, + "grad_norm": 8.452660578831315, + "learning_rate": 1.705426213228516e-05, + "loss": 0.8818, + "step": 2517 + }, + { + "epoch": 1.3669923995656894, + "grad_norm": 7.8301031574049755, + "learning_rate": 1.7051769018819327e-05, + "loss": 0.8803, + "step": 2518 + }, + { + "epoch": 1.3675352877307274, + "grad_norm": 10.77827779179992, + "learning_rate": 1.704927503317946e-05, + "loss": 1.6667, + "step": 2519 + }, + { + "epoch": 1.3680781758957654, + "grad_norm": 4.579120603442138, + "learning_rate": 1.7046780175674023e-05, + "loss": 0.6101, + "step": 2520 + }, + { + "epoch": 1.3686210640608034, + "grad_norm": 7.443733277623912, + "learning_rate": 1.7044284446611584e-05, + "loss": 0.8174, + "step": 2521 + }, + { + "epoch": 1.3691639522258414, + "grad_norm": 10.860855146774309, + "learning_rate": 1.7041787846300817e-05, + "loss": 1.2849, + "step": 2522 + }, + { + "epoch": 1.3697068403908794, + "grad_norm": 9.311283837684355, + "learning_rate": 1.7039290375050503e-05, + "loss": 1.4923, + "step": 2523 + }, + { + "epoch": 1.3702497285559174, + "grad_norm": 7.171226901795553, + "learning_rate": 1.703679203316954e-05, + "loss": 0.9493, + "step": 2524 + }, + { + "epoch": 1.3707926167209554, + "grad_norm": 8.658053683953712, + "learning_rate": 1.7034292820966923e-05, + "loss": 1.1366, + "step": 2525 + }, + { + "epoch": 1.3713355048859934, + "grad_norm": 6.400723985129287, + "learning_rate": 1.703179273875176e-05, + "loss": 1.0484, + "step": 2526 + }, + { + "epoch": 1.3718783930510314, + "grad_norm": 6.2215780206112274, + "learning_rate": 1.702929178683326e-05, + "loss": 1.1058, + "step": 2527 + }, + { + "epoch": 1.3724212812160694, + "grad_norm": 7.548758658264992, + "learning_rate": 1.7026789965520755e-05, + "loss": 1.094, + "step": 2528 + }, + { + "epoch": 1.3729641693811074, + "grad_norm": 7.250992484079327, + "learning_rate": 1.7024287275123664e-05, + "loss": 0.8424, + "step": 2529 + }, + { + "epoch": 1.3735070575461454, + "grad_norm": 6.664109303085317, + "learning_rate": 1.702178371595153e-05, + "loss": 1.0498, + "step": 2530 + }, + { + "epoch": 1.3740499457111834, + "grad_norm": 6.379180216401721, + "learning_rate": 1.7019279288313994e-05, + "loss": 0.9038, + "step": 2531 + }, + { + "epoch": 1.3745928338762214, + "grad_norm": 7.524913368583668, + "learning_rate": 1.7016773992520806e-05, + "loss": 0.8787, + "step": 2532 + }, + { + "epoch": 1.3751357220412594, + "grad_norm": 5.722778911929788, + "learning_rate": 1.701426782888183e-05, + "loss": 0.8199, + "step": 2533 + }, + { + "epoch": 1.3756786102062974, + "grad_norm": 6.26381521799836, + "learning_rate": 1.701176079770703e-05, + "loss": 0.6989, + "step": 2534 + }, + { + "epoch": 1.3762214983713354, + "grad_norm": 7.2300935887481765, + "learning_rate": 1.700925289930648e-05, + "loss": 1.094, + "step": 2535 + }, + { + "epoch": 1.3767643865363735, + "grad_norm": 7.385614391862118, + "learning_rate": 1.7006744133990358e-05, + "loss": 0.7753, + "step": 2536 + }, + { + "epoch": 1.3773072747014115, + "grad_norm": 5.52011787904346, + "learning_rate": 1.7004234502068952e-05, + "loss": 0.9232, + "step": 2537 + }, + { + "epoch": 1.3778501628664495, + "grad_norm": 6.523028453950784, + "learning_rate": 1.7001724003852668e-05, + "loss": 1.1181, + "step": 2538 + }, + { + "epoch": 1.3783930510314875, + "grad_norm": 6.334090281128193, + "learning_rate": 1.6999212639651995e-05, + "loss": 0.9384, + "step": 2539 + }, + { + "epoch": 1.3789359391965255, + "grad_norm": 7.3687722051890265, + "learning_rate": 1.6996700409777548e-05, + "loss": 0.9678, + "step": 2540 + }, + { + "epoch": 1.3794788273615635, + "grad_norm": 9.49113126498042, + "learning_rate": 1.6994187314540042e-05, + "loss": 0.7875, + "step": 2541 + }, + { + "epoch": 1.3800217155266015, + "grad_norm": 4.885613246464195, + "learning_rate": 1.69916733542503e-05, + "loss": 0.4488, + "step": 2542 + }, + { + "epoch": 1.3805646036916395, + "grad_norm": 7.362521236541499, + "learning_rate": 1.6989158529219262e-05, + "loss": 0.9808, + "step": 2543 + }, + { + "epoch": 1.3811074918566775, + "grad_norm": 8.6488909724118, + "learning_rate": 1.6986642839757953e-05, + "loss": 1.0599, + "step": 2544 + }, + { + "epoch": 1.3816503800217155, + "grad_norm": 6.054541159551389, + "learning_rate": 1.698412628617752e-05, + "loss": 0.8252, + "step": 2545 + }, + { + "epoch": 1.3821932681867535, + "grad_norm": 7.77360294326381, + "learning_rate": 1.698160886878922e-05, + "loss": 1.0718, + "step": 2546 + }, + { + "epoch": 1.3827361563517915, + "grad_norm": 9.493714306510151, + "learning_rate": 1.697909058790441e-05, + "loss": 1.2784, + "step": 2547 + }, + { + "epoch": 1.3832790445168295, + "grad_norm": 7.7307988823000455, + "learning_rate": 1.6976571443834555e-05, + "loss": 1.0308, + "step": 2548 + }, + { + "epoch": 1.3838219326818675, + "grad_norm": 7.149534093427325, + "learning_rate": 1.697405143689122e-05, + "loss": 0.9702, + "step": 2549 + }, + { + "epoch": 1.3843648208469055, + "grad_norm": 6.16619254042592, + "learning_rate": 1.6971530567386087e-05, + "loss": 0.6922, + "step": 2550 + }, + { + "epoch": 1.3849077090119435, + "grad_norm": 6.555255328428672, + "learning_rate": 1.6969008835630947e-05, + "loss": 0.8771, + "step": 2551 + }, + { + "epoch": 1.3854505971769815, + "grad_norm": 7.652524555180243, + "learning_rate": 1.6966486241937685e-05, + "loss": 1.0635, + "step": 2552 + }, + { + "epoch": 1.3859934853420195, + "grad_norm": 9.675328775672968, + "learning_rate": 1.69639627866183e-05, + "loss": 1.1616, + "step": 2553 + }, + { + "epoch": 1.3865363735070575, + "grad_norm": 6.5667057730439256, + "learning_rate": 1.69614384699849e-05, + "loss": 0.7181, + "step": 2554 + }, + { + "epoch": 1.3870792616720955, + "grad_norm": 7.374476144001139, + "learning_rate": 1.6958913292349698e-05, + "loss": 1.123, + "step": 2555 + }, + { + "epoch": 1.3876221498371335, + "grad_norm": 8.873033944832025, + "learning_rate": 1.6956387254025005e-05, + "loss": 1.2651, + "step": 2556 + }, + { + "epoch": 1.3881650380021715, + "grad_norm": 6.877978468812787, + "learning_rate": 1.695386035532325e-05, + "loss": 1.0039, + "step": 2557 + }, + { + "epoch": 1.3887079261672095, + "grad_norm": 4.746919583421645, + "learning_rate": 1.6951332596556966e-05, + "loss": 0.7222, + "step": 2558 + }, + { + "epoch": 1.3892508143322475, + "grad_norm": 5.538693020619686, + "learning_rate": 1.6948803978038787e-05, + "loss": 0.8463, + "step": 2559 + }, + { + "epoch": 1.3897937024972855, + "grad_norm": 6.636448326582328, + "learning_rate": 1.6946274500081455e-05, + "loss": 0.7594, + "step": 2560 + }, + { + "epoch": 1.3903365906623235, + "grad_norm": 8.74658695496399, + "learning_rate": 1.6943744162997825e-05, + "loss": 1.224, + "step": 2561 + }, + { + "epoch": 1.3908794788273615, + "grad_norm": 7.305739264312674, + "learning_rate": 1.694121296710085e-05, + "loss": 1.0755, + "step": 2562 + }, + { + "epoch": 1.3914223669923995, + "grad_norm": 4.6475508332537165, + "learning_rate": 1.693868091270359e-05, + "loss": 0.6127, + "step": 2563 + }, + { + "epoch": 1.3919652551574375, + "grad_norm": 8.484116710765257, + "learning_rate": 1.6936148000119218e-05, + "loss": 1.029, + "step": 2564 + }, + { + "epoch": 1.3925081433224755, + "grad_norm": 6.532276109161297, + "learning_rate": 1.6933614229661008e-05, + "loss": 1.0517, + "step": 2565 + }, + { + "epoch": 1.3930510314875135, + "grad_norm": 5.43723324019315, + "learning_rate": 1.6931079601642336e-05, + "loss": 0.6915, + "step": 2566 + }, + { + "epoch": 1.3935939196525515, + "grad_norm": 7.562306350008547, + "learning_rate": 1.69285441163767e-05, + "loss": 0.5726, + "step": 2567 + }, + { + "epoch": 1.3941368078175895, + "grad_norm": 8.500602954507755, + "learning_rate": 1.6926007774177678e-05, + "loss": 1.0297, + "step": 2568 + }, + { + "epoch": 1.3946796959826275, + "grad_norm": 6.432151204429844, + "learning_rate": 1.6923470575358977e-05, + "loss": 1.0016, + "step": 2569 + }, + { + "epoch": 1.3952225841476655, + "grad_norm": 9.552257208599151, + "learning_rate": 1.6920932520234402e-05, + "loss": 1.0814, + "step": 2570 + }, + { + "epoch": 1.3957654723127035, + "grad_norm": 8.075918082294125, + "learning_rate": 1.6918393609117863e-05, + "loss": 1.1699, + "step": 2571 + }, + { + "epoch": 1.3963083604777415, + "grad_norm": 6.296577538864887, + "learning_rate": 1.6915853842323373e-05, + "loss": 1.0211, + "step": 2572 + }, + { + "epoch": 1.3968512486427795, + "grad_norm": 7.785873884786678, + "learning_rate": 1.691331322016506e-05, + "loss": 0.9225, + "step": 2573 + }, + { + "epoch": 1.3973941368078175, + "grad_norm": 6.768130802749829, + "learning_rate": 1.691077174295715e-05, + "loss": 0.8092, + "step": 2574 + }, + { + "epoch": 1.3979370249728555, + "grad_norm": 5.566346022945414, + "learning_rate": 1.6908229411013977e-05, + "loss": 0.8312, + "step": 2575 + }, + { + "epoch": 1.3984799131378935, + "grad_norm": 7.2860384645139815, + "learning_rate": 1.6905686224649978e-05, + "loss": 0.9357, + "step": 2576 + }, + { + "epoch": 1.3990228013029316, + "grad_norm": 7.072468614016554, + "learning_rate": 1.6903142184179704e-05, + "loss": 0.7362, + "step": 2577 + }, + { + "epoch": 1.3995656894679696, + "grad_norm": 6.421225685489366, + "learning_rate": 1.6900597289917803e-05, + "loss": 0.6828, + "step": 2578 + }, + { + "epoch": 1.4001085776330076, + "grad_norm": 8.167489592155093, + "learning_rate": 1.689805154217903e-05, + "loss": 1.1237, + "step": 2579 + }, + { + "epoch": 1.4006514657980456, + "grad_norm": 11.676833397587332, + "learning_rate": 1.6895504941278246e-05, + "loss": 0.9532, + "step": 2580 + }, + { + "epoch": 1.4011943539630836, + "grad_norm": 6.04293054204543, + "learning_rate": 1.689295748753042e-05, + "loss": 0.6541, + "step": 2581 + }, + { + "epoch": 1.4017372421281216, + "grad_norm": 8.432023560966908, + "learning_rate": 1.6890409181250632e-05, + "loss": 1.0167, + "step": 2582 + }, + { + "epoch": 1.4022801302931596, + "grad_norm": 7.494983940960417, + "learning_rate": 1.688786002275405e-05, + "loss": 0.8913, + "step": 2583 + }, + { + "epoch": 1.4028230184581976, + "grad_norm": 6.475912046078606, + "learning_rate": 1.6885310012355964e-05, + "loss": 0.8599, + "step": 2584 + }, + { + "epoch": 1.4033659066232356, + "grad_norm": 5.145369710998311, + "learning_rate": 1.6882759150371765e-05, + "loss": 0.8079, + "step": 2585 + }, + { + "epoch": 1.4039087947882736, + "grad_norm": 6.1855027756804155, + "learning_rate": 1.688020743711694e-05, + "loss": 0.8046, + "step": 2586 + }, + { + "epoch": 1.4044516829533116, + "grad_norm": 8.409570707749053, + "learning_rate": 1.68776548729071e-05, + "loss": 0.75, + "step": 2587 + }, + { + "epoch": 1.4049945711183496, + "grad_norm": 6.150666283054011, + "learning_rate": 1.687510145805794e-05, + "loss": 0.8339, + "step": 2588 + }, + { + "epoch": 1.4055374592833876, + "grad_norm": 5.4082205122977625, + "learning_rate": 1.6872547192885272e-05, + "loss": 0.7878, + "step": 2589 + }, + { + "epoch": 1.4060803474484256, + "grad_norm": 7.011727228130495, + "learning_rate": 1.686999207770502e-05, + "loss": 0.6415, + "step": 2590 + }, + { + "epoch": 1.4066232356134636, + "grad_norm": 7.384524419917869, + "learning_rate": 1.6867436112833193e-05, + "loss": 0.8489, + "step": 2591 + }, + { + "epoch": 1.4071661237785016, + "grad_norm": 6.846659366372164, + "learning_rate": 1.6864879298585925e-05, + "loss": 1.128, + "step": 2592 + }, + { + "epoch": 1.4077090119435396, + "grad_norm": 6.142948047083128, + "learning_rate": 1.6862321635279444e-05, + "loss": 0.7886, + "step": 2593 + }, + { + "epoch": 1.4082519001085776, + "grad_norm": 10.23557881641297, + "learning_rate": 1.6859763123230086e-05, + "loss": 0.8323, + "step": 2594 + }, + { + "epoch": 1.4087947882736156, + "grad_norm": 6.995135696669204, + "learning_rate": 1.6857203762754294e-05, + "loss": 1.0283, + "step": 2595 + }, + { + "epoch": 1.4093376764386536, + "grad_norm": 7.709014920487576, + "learning_rate": 1.685464355416861e-05, + "loss": 0.8586, + "step": 2596 + }, + { + "epoch": 1.4098805646036916, + "grad_norm": 8.153794053960306, + "learning_rate": 1.6852082497789684e-05, + "loss": 1.1797, + "step": 2597 + }, + { + "epoch": 1.4104234527687296, + "grad_norm": 7.721286185320884, + "learning_rate": 1.684952059393428e-05, + "loss": 0.9196, + "step": 2598 + }, + { + "epoch": 1.4109663409337676, + "grad_norm": 8.024349951964256, + "learning_rate": 1.684695784291925e-05, + "loss": 0.7769, + "step": 2599 + }, + { + "epoch": 1.4115092290988056, + "grad_norm": 8.135395328882872, + "learning_rate": 1.684439424506156e-05, + "loss": 1.1642, + "step": 2600 + }, + { + "epoch": 1.4120521172638436, + "grad_norm": 6.057801451822921, + "learning_rate": 1.684182980067828e-05, + "loss": 0.7352, + "step": 2601 + }, + { + "epoch": 1.4125950054288816, + "grad_norm": 7.808755828672067, + "learning_rate": 1.683926451008659e-05, + "loss": 0.9788, + "step": 2602 + }, + { + "epoch": 1.4131378935939196, + "grad_norm": 7.426450515083944, + "learning_rate": 1.6836698373603765e-05, + "loss": 0.7696, + "step": 2603 + }, + { + "epoch": 1.4136807817589576, + "grad_norm": 9.382792907740685, + "learning_rate": 1.6834131391547187e-05, + "loss": 1.0872, + "step": 2604 + }, + { + "epoch": 1.4142236699239956, + "grad_norm": 6.171700145892427, + "learning_rate": 1.6831563564234347e-05, + "loss": 0.7233, + "step": 2605 + }, + { + "epoch": 1.4147665580890336, + "grad_norm": 11.270930151197227, + "learning_rate": 1.682899489198284e-05, + "loss": 0.9739, + "step": 2606 + }, + { + "epoch": 1.4153094462540716, + "grad_norm": 9.050544540368543, + "learning_rate": 1.6826425375110357e-05, + "loss": 0.9436, + "step": 2607 + }, + { + "epoch": 1.4158523344191096, + "grad_norm": 6.969930349987723, + "learning_rate": 1.6823855013934705e-05, + "loss": 0.7632, + "step": 2608 + }, + { + "epoch": 1.4163952225841476, + "grad_norm": 5.791480821515424, + "learning_rate": 1.682128380877379e-05, + "loss": 0.8202, + "step": 2609 + }, + { + "epoch": 1.4169381107491856, + "grad_norm": 9.905757175497309, + "learning_rate": 1.6818711759945623e-05, + "loss": 1.2161, + "step": 2610 + }, + { + "epoch": 1.4174809989142236, + "grad_norm": 6.082684402096193, + "learning_rate": 1.6816138867768318e-05, + "loss": 0.7245, + "step": 2611 + }, + { + "epoch": 1.4180238870792616, + "grad_norm": 10.15499139340414, + "learning_rate": 1.6813565132560092e-05, + "loss": 1.1514, + "step": 2612 + }, + { + "epoch": 1.4185667752442996, + "grad_norm": 6.978513243855861, + "learning_rate": 1.6810990554639276e-05, + "loss": 0.5331, + "step": 2613 + }, + { + "epoch": 1.4191096634093376, + "grad_norm": 8.037739371135425, + "learning_rate": 1.6808415134324288e-05, + "loss": 0.7089, + "step": 2614 + }, + { + "epoch": 1.4196525515743756, + "grad_norm": 6.925791024229076, + "learning_rate": 1.6805838871933664e-05, + "loss": 0.8981, + "step": 2615 + }, + { + "epoch": 1.4201954397394136, + "grad_norm": 6.352165521713976, + "learning_rate": 1.6803261767786048e-05, + "loss": 0.5644, + "step": 2616 + }, + { + "epoch": 1.4207383279044516, + "grad_norm": 10.267497627922179, + "learning_rate": 1.680068382220017e-05, + "loss": 0.9607, + "step": 2617 + }, + { + "epoch": 1.4212812160694897, + "grad_norm": 7.696468630335302, + "learning_rate": 1.679810503549488e-05, + "loss": 0.729, + "step": 2618 + }, + { + "epoch": 1.4218241042345277, + "grad_norm": 6.868981955235214, + "learning_rate": 1.679552540798912e-05, + "loss": 0.9152, + "step": 2619 + }, + { + "epoch": 1.4223669923995657, + "grad_norm": 8.909790381587262, + "learning_rate": 1.6792944940001952e-05, + "loss": 0.9335, + "step": 2620 + }, + { + "epoch": 1.4229098805646037, + "grad_norm": 9.62415635421133, + "learning_rate": 1.6790363631852524e-05, + "loss": 1.1012, + "step": 2621 + }, + { + "epoch": 1.4234527687296417, + "grad_norm": 7.307385111665146, + "learning_rate": 1.67877814838601e-05, + "loss": 0.6963, + "step": 2622 + }, + { + "epoch": 1.4239956568946797, + "grad_norm": 9.018642559605246, + "learning_rate": 1.678519849634405e-05, + "loss": 1.0612, + "step": 2623 + }, + { + "epoch": 1.4245385450597177, + "grad_norm": 6.929244675286548, + "learning_rate": 1.6782614669623827e-05, + "loss": 0.9462, + "step": 2624 + }, + { + "epoch": 1.4250814332247557, + "grad_norm": 8.577419731292203, + "learning_rate": 1.6780030004019016e-05, + "loss": 1.112, + "step": 2625 + }, + { + "epoch": 1.4256243213897937, + "grad_norm": 7.125694710248634, + "learning_rate": 1.677744449984929e-05, + "loss": 0.851, + "step": 2626 + }, + { + "epoch": 1.4261672095548317, + "grad_norm": 9.488635353984538, + "learning_rate": 1.6774858157434425e-05, + "loss": 1.3064, + "step": 2627 + }, + { + "epoch": 1.4267100977198697, + "grad_norm": 7.500971938576216, + "learning_rate": 1.6772270977094307e-05, + "loss": 1.3675, + "step": 2628 + }, + { + "epoch": 1.4272529858849077, + "grad_norm": 6.979855088071626, + "learning_rate": 1.676968295914892e-05, + "loss": 0.8896, + "step": 2629 + }, + { + "epoch": 1.4277958740499457, + "grad_norm": 7.795843207695432, + "learning_rate": 1.6767094103918357e-05, + "loss": 0.7758, + "step": 2630 + }, + { + "epoch": 1.4283387622149837, + "grad_norm": 6.131884468994572, + "learning_rate": 1.6764504411722806e-05, + "loss": 0.7744, + "step": 2631 + }, + { + "epoch": 1.4288816503800217, + "grad_norm": 10.33693974618067, + "learning_rate": 1.676191388288257e-05, + "loss": 1.5364, + "step": 2632 + }, + { + "epoch": 1.4294245385450597, + "grad_norm": 10.005426546410897, + "learning_rate": 1.6759322517718048e-05, + "loss": 0.9713, + "step": 2633 + }, + { + "epoch": 1.4299674267100977, + "grad_norm": 9.480236119835816, + "learning_rate": 1.6756730316549745e-05, + "loss": 1.2573, + "step": 2634 + }, + { + "epoch": 1.4305103148751357, + "grad_norm": 10.874101086577593, + "learning_rate": 1.675413727969827e-05, + "loss": 1.0308, + "step": 2635 + }, + { + "epoch": 1.4310532030401737, + "grad_norm": 5.34984499382639, + "learning_rate": 1.675154340748433e-05, + "loss": 0.6013, + "step": 2636 + }, + { + "epoch": 1.4315960912052117, + "grad_norm": 7.464386702227026, + "learning_rate": 1.674894870022874e-05, + "loss": 0.9918, + "step": 2637 + }, + { + "epoch": 1.4321389793702497, + "grad_norm": 8.280480153626241, + "learning_rate": 1.674635315825242e-05, + "loss": 0.8745, + "step": 2638 + }, + { + "epoch": 1.4326818675352877, + "grad_norm": 6.169575731939732, + "learning_rate": 1.6743756781876385e-05, + "loss": 0.6984, + "step": 2639 + }, + { + "epoch": 1.4332247557003257, + "grad_norm": 5.390920772379431, + "learning_rate": 1.6741159571421768e-05, + "loss": 0.8886, + "step": 2640 + }, + { + "epoch": 1.4337676438653637, + "grad_norm": 7.003336357340687, + "learning_rate": 1.6738561527209792e-05, + "loss": 1.0335, + "step": 2641 + }, + { + "epoch": 1.4343105320304017, + "grad_norm": 7.722166856340297, + "learning_rate": 1.6735962649561784e-05, + "loss": 0.9804, + "step": 2642 + }, + { + "epoch": 1.4348534201954397, + "grad_norm": 7.328242752900802, + "learning_rate": 1.673336293879918e-05, + "loss": 0.8675, + "step": 2643 + }, + { + "epoch": 1.4353963083604777, + "grad_norm": 7.344577484351048, + "learning_rate": 1.6730762395243515e-05, + "loss": 1.2518, + "step": 2644 + }, + { + "epoch": 1.4359391965255157, + "grad_norm": 7.560893408391802, + "learning_rate": 1.6728161019216433e-05, + "loss": 1.0196, + "step": 2645 + }, + { + "epoch": 1.4364820846905537, + "grad_norm": 7.509821569775923, + "learning_rate": 1.6725558811039674e-05, + "loss": 1.1112, + "step": 2646 + }, + { + "epoch": 1.4370249728555917, + "grad_norm": 7.672412677248201, + "learning_rate": 1.672295577103508e-05, + "loss": 0.9075, + "step": 2647 + }, + { + "epoch": 1.4375678610206297, + "grad_norm": 8.562389123203614, + "learning_rate": 1.67203518995246e-05, + "loss": 1.2123, + "step": 2648 + }, + { + "epoch": 1.4381107491856677, + "grad_norm": 8.172008773566091, + "learning_rate": 1.671774719683029e-05, + "loss": 0.8064, + "step": 2649 + }, + { + "epoch": 1.4386536373507057, + "grad_norm": 7.5145061670971405, + "learning_rate": 1.6715141663274297e-05, + "loss": 1.3513, + "step": 2650 + }, + { + "epoch": 1.4391965255157437, + "grad_norm": 5.675981880534199, + "learning_rate": 1.6712535299178883e-05, + "loss": 0.5029, + "step": 2651 + }, + { + "epoch": 1.4397394136807817, + "grad_norm": 8.195158801834715, + "learning_rate": 1.6709928104866403e-05, + "loss": 0.9414, + "step": 2652 + }, + { + "epoch": 1.4402823018458197, + "grad_norm": 7.374510239535514, + "learning_rate": 1.6707320080659322e-05, + "loss": 0.8103, + "step": 2653 + }, + { + "epoch": 1.4408251900108577, + "grad_norm": 5.703232221762084, + "learning_rate": 1.6704711226880204e-05, + "loss": 1.1646, + "step": 2654 + }, + { + "epoch": 1.4413680781758957, + "grad_norm": 6.981749095787225, + "learning_rate": 1.6702101543851714e-05, + "loss": 0.7174, + "step": 2655 + }, + { + "epoch": 1.4419109663409337, + "grad_norm": 6.456866833246808, + "learning_rate": 1.6699491031896625e-05, + "loss": 0.9795, + "step": 2656 + }, + { + "epoch": 1.4424538545059717, + "grad_norm": 10.193228844116039, + "learning_rate": 1.6696879691337807e-05, + "loss": 1.1288, + "step": 2657 + }, + { + "epoch": 1.4429967426710097, + "grad_norm": 9.920534268729181, + "learning_rate": 1.6694267522498237e-05, + "loss": 0.8379, + "step": 2658 + }, + { + "epoch": 1.4435396308360477, + "grad_norm": 6.153776792682404, + "learning_rate": 1.669165452570099e-05, + "loss": 0.8818, + "step": 2659 + }, + { + "epoch": 1.4440825190010858, + "grad_norm": 5.650871346064418, + "learning_rate": 1.6689040701269245e-05, + "loss": 0.732, + "step": 2660 + }, + { + "epoch": 1.4446254071661238, + "grad_norm": 7.829649689315363, + "learning_rate": 1.668642604952629e-05, + "loss": 0.8119, + "step": 2661 + }, + { + "epoch": 1.4451682953311618, + "grad_norm": 5.959350361834891, + "learning_rate": 1.6683810570795498e-05, + "loss": 0.6829, + "step": 2662 + }, + { + "epoch": 1.4457111834961998, + "grad_norm": 8.76687690585514, + "learning_rate": 1.6681194265400365e-05, + "loss": 1.4888, + "step": 2663 + }, + { + "epoch": 1.4462540716612378, + "grad_norm": 6.129763847493397, + "learning_rate": 1.6678577133664476e-05, + "loss": 0.8372, + "step": 2664 + }, + { + "epoch": 1.4467969598262758, + "grad_norm": 6.529235674005995, + "learning_rate": 1.6675959175911527e-05, + "loss": 0.9655, + "step": 2665 + }, + { + "epoch": 1.4473398479913138, + "grad_norm": 7.741547370588863, + "learning_rate": 1.6673340392465304e-05, + "loss": 1.0772, + "step": 2666 + }, + { + "epoch": 1.4478827361563518, + "grad_norm": 6.674042204787818, + "learning_rate": 1.6670720783649706e-05, + "loss": 0.9681, + "step": 2667 + }, + { + "epoch": 1.4484256243213898, + "grad_norm": 7.920578917847642, + "learning_rate": 1.666810034978873e-05, + "loss": 1.1068, + "step": 2668 + }, + { + "epoch": 1.4489685124864278, + "grad_norm": 7.249023283604718, + "learning_rate": 1.6665479091206476e-05, + "loss": 1.1025, + "step": 2669 + }, + { + "epoch": 1.4495114006514658, + "grad_norm": 6.407638163624444, + "learning_rate": 1.6662857008227145e-05, + "loss": 0.8236, + "step": 2670 + }, + { + "epoch": 1.4500542888165038, + "grad_norm": 7.289336395940614, + "learning_rate": 1.6660234101175036e-05, + "loss": 0.9386, + "step": 2671 + }, + { + "epoch": 1.4505971769815418, + "grad_norm": 6.973959594036107, + "learning_rate": 1.665761037037456e-05, + "loss": 0.8435, + "step": 2672 + }, + { + "epoch": 1.4511400651465798, + "grad_norm": 6.971928597160326, + "learning_rate": 1.665498581615023e-05, + "loss": 0.9065, + "step": 2673 + }, + { + "epoch": 1.4516829533116178, + "grad_norm": 6.1608850694063735, + "learning_rate": 1.665236043882664e-05, + "loss": 0.9289, + "step": 2674 + }, + { + "epoch": 1.4522258414766558, + "grad_norm": 6.62873967155635, + "learning_rate": 1.6649734238728512e-05, + "loss": 0.7797, + "step": 2675 + }, + { + "epoch": 1.4527687296416938, + "grad_norm": 7.089586063125632, + "learning_rate": 1.6647107216180655e-05, + "loss": 0.6306, + "step": 2676 + }, + { + "epoch": 1.4533116178067318, + "grad_norm": 10.986990416571043, + "learning_rate": 1.6644479371507985e-05, + "loss": 0.8764, + "step": 2677 + }, + { + "epoch": 1.4538545059717698, + "grad_norm": 5.865852703001031, + "learning_rate": 1.664185070503551e-05, + "loss": 0.9106, + "step": 2678 + }, + { + "epoch": 1.4543973941368078, + "grad_norm": 7.303323841843, + "learning_rate": 1.663922121708836e-05, + "loss": 0.8188, + "step": 2679 + }, + { + "epoch": 1.4549402823018458, + "grad_norm": 5.715280365147904, + "learning_rate": 1.663659090799175e-05, + "loss": 1.0573, + "step": 2680 + }, + { + "epoch": 1.4554831704668838, + "grad_norm": 7.084313061900368, + "learning_rate": 1.6633959778070992e-05, + "loss": 1.0749, + "step": 2681 + }, + { + "epoch": 1.4560260586319218, + "grad_norm": 7.902311581251679, + "learning_rate": 1.6631327827651524e-05, + "loss": 0.8677, + "step": 2682 + }, + { + "epoch": 1.4565689467969598, + "grad_norm": 8.1594384957164, + "learning_rate": 1.6628695057058855e-05, + "loss": 0.9999, + "step": 2683 + }, + { + "epoch": 1.4571118349619978, + "grad_norm": 5.580141751242927, + "learning_rate": 1.6626061466618623e-05, + "loss": 0.6387, + "step": 2684 + }, + { + "epoch": 1.4576547231270358, + "grad_norm": 8.145754034329466, + "learning_rate": 1.6623427056656544e-05, + "loss": 0.8984, + "step": 2685 + }, + { + "epoch": 1.4581976112920738, + "grad_norm": 6.453637820501138, + "learning_rate": 1.6620791827498454e-05, + "loss": 0.6922, + "step": 2686 + }, + { + "epoch": 1.4587404994571118, + "grad_norm": 7.3511834666776075, + "learning_rate": 1.6618155779470275e-05, + "loss": 0.8462, + "step": 2687 + }, + { + "epoch": 1.4592833876221498, + "grad_norm": 7.962874909840381, + "learning_rate": 1.6615518912898043e-05, + "loss": 0.9639, + "step": 2688 + }, + { + "epoch": 1.4598262757871878, + "grad_norm": 6.975602073560851, + "learning_rate": 1.6612881228107886e-05, + "loss": 0.9364, + "step": 2689 + }, + { + "epoch": 1.4603691639522258, + "grad_norm": 6.916712660881488, + "learning_rate": 1.6610242725426044e-05, + "loss": 0.8033, + "step": 2690 + }, + { + "epoch": 1.4609120521172638, + "grad_norm": 6.268208953592866, + "learning_rate": 1.6607603405178842e-05, + "loss": 0.7542, + "step": 2691 + }, + { + "epoch": 1.4614549402823018, + "grad_norm": 9.047138051843492, + "learning_rate": 1.660496326769272e-05, + "loss": 1.5956, + "step": 2692 + }, + { + "epoch": 1.4619978284473398, + "grad_norm": 7.228556514923675, + "learning_rate": 1.6602322313294216e-05, + "loss": 0.8461, + "step": 2693 + }, + { + "epoch": 1.4625407166123778, + "grad_norm": 6.195288408674038, + "learning_rate": 1.659968054230997e-05, + "loss": 0.9851, + "step": 2694 + }, + { + "epoch": 1.4630836047774158, + "grad_norm": 7.413116367868534, + "learning_rate": 1.6597037955066713e-05, + "loss": 0.9693, + "step": 2695 + }, + { + "epoch": 1.4636264929424538, + "grad_norm": 7.014031390984585, + "learning_rate": 1.6594394551891288e-05, + "loss": 1.0539, + "step": 2696 + }, + { + "epoch": 1.4641693811074918, + "grad_norm": 6.8196013668965225, + "learning_rate": 1.6591750333110634e-05, + "loss": 0.712, + "step": 2697 + }, + { + "epoch": 1.4647122692725298, + "grad_norm": 8.160690900976675, + "learning_rate": 1.658910529905179e-05, + "loss": 0.6551, + "step": 2698 + }, + { + "epoch": 1.4652551574375678, + "grad_norm": 5.174705288640052, + "learning_rate": 1.6586459450041906e-05, + "loss": 0.8285, + "step": 2699 + }, + { + "epoch": 1.4657980456026058, + "grad_norm": 7.029482771999071, + "learning_rate": 1.6583812786408216e-05, + "loss": 0.923, + "step": 2700 + }, + { + "epoch": 1.4663409337676439, + "grad_norm": 7.8877107275923946, + "learning_rate": 1.658116530847807e-05, + "loss": 1.1915, + "step": 2701 + }, + { + "epoch": 1.4668838219326819, + "grad_norm": 9.269305925663183, + "learning_rate": 1.657851701657891e-05, + "loss": 1.098, + "step": 2702 + }, + { + "epoch": 1.4674267100977199, + "grad_norm": 10.478546362437942, + "learning_rate": 1.657586791103828e-05, + "loss": 1.2416, + "step": 2703 + }, + { + "epoch": 1.4679695982627579, + "grad_norm": 8.207786471433751, + "learning_rate": 1.6573217992183826e-05, + "loss": 0.8658, + "step": 2704 + }, + { + "epoch": 1.4685124864277959, + "grad_norm": 7.862118410069405, + "learning_rate": 1.6570567260343294e-05, + "loss": 0.8714, + "step": 2705 + }, + { + "epoch": 1.4690553745928339, + "grad_norm": 7.744957075186954, + "learning_rate": 1.6567915715844534e-05, + "loss": 0.9374, + "step": 2706 + }, + { + "epoch": 1.4695982627578719, + "grad_norm": 6.97424297060321, + "learning_rate": 1.6565263359015488e-05, + "loss": 0.7183, + "step": 2707 + }, + { + "epoch": 1.4701411509229099, + "grad_norm": 7.349900880491086, + "learning_rate": 1.6562610190184206e-05, + "loss": 0.6066, + "step": 2708 + }, + { + "epoch": 1.4706840390879479, + "grad_norm": 6.982564531951699, + "learning_rate": 1.655995620967884e-05, + "loss": 1.2218, + "step": 2709 + }, + { + "epoch": 1.4712269272529859, + "grad_norm": 7.572476514419708, + "learning_rate": 1.6557301417827632e-05, + "loss": 0.89, + "step": 2710 + }, + { + "epoch": 1.4717698154180239, + "grad_norm": 8.654751087453503, + "learning_rate": 1.6554645814958932e-05, + "loss": 0.7188, + "step": 2711 + }, + { + "epoch": 1.4723127035830619, + "grad_norm": 6.282711052010393, + "learning_rate": 1.6551989401401196e-05, + "loss": 0.7586, + "step": 2712 + }, + { + "epoch": 1.4728555917480999, + "grad_norm": 8.732262593936317, + "learning_rate": 1.6549332177482966e-05, + "loss": 1.0178, + "step": 2713 + }, + { + "epoch": 1.4733984799131379, + "grad_norm": 6.989470691687101, + "learning_rate": 1.6546674143532895e-05, + "loss": 0.6981, + "step": 2714 + }, + { + "epoch": 1.4739413680781759, + "grad_norm": 7.179372740296565, + "learning_rate": 1.6544015299879734e-05, + "loss": 0.7617, + "step": 2715 + }, + { + "epoch": 1.4744842562432139, + "grad_norm": 9.117294639273645, + "learning_rate": 1.6541355646852327e-05, + "loss": 0.8822, + "step": 2716 + }, + { + "epoch": 1.475027144408252, + "grad_norm": 7.875727317127435, + "learning_rate": 1.653869518477963e-05, + "loss": 0.9149, + "step": 2717 + }, + { + "epoch": 1.47557003257329, + "grad_norm": 7.926910105501724, + "learning_rate": 1.6536033913990687e-05, + "loss": 0.8797, + "step": 2718 + }, + { + "epoch": 1.476112920738328, + "grad_norm": 5.14878192977047, + "learning_rate": 1.6533371834814657e-05, + "loss": 0.4491, + "step": 2719 + }, + { + "epoch": 1.476655808903366, + "grad_norm": 5.327596982546826, + "learning_rate": 1.6530708947580785e-05, + "loss": 0.4335, + "step": 2720 + }, + { + "epoch": 1.477198697068404, + "grad_norm": 9.158336314582998, + "learning_rate": 1.6528045252618423e-05, + "loss": 1.1208, + "step": 2721 + }, + { + "epoch": 1.477741585233442, + "grad_norm": 10.630652314970348, + "learning_rate": 1.6525380750257022e-05, + "loss": 1.097, + "step": 2722 + }, + { + "epoch": 1.47828447339848, + "grad_norm": 8.97220182564547, + "learning_rate": 1.652271544082613e-05, + "loss": 0.7838, + "step": 2723 + }, + { + "epoch": 1.478827361563518, + "grad_norm": 10.609251261516084, + "learning_rate": 1.652004932465539e-05, + "loss": 0.8861, + "step": 2724 + }, + { + "epoch": 1.479370249728556, + "grad_norm": 7.303801635012744, + "learning_rate": 1.6517382402074563e-05, + "loss": 0.5437, + "step": 2725 + }, + { + "epoch": 1.479913137893594, + "grad_norm": 11.770584674671793, + "learning_rate": 1.651471467341349e-05, + "loss": 1.4947, + "step": 2726 + }, + { + "epoch": 1.480456026058632, + "grad_norm": 6.824299484966605, + "learning_rate": 1.6512046139002128e-05, + "loss": 0.7144, + "step": 2727 + }, + { + "epoch": 1.48099891422367, + "grad_norm": 7.595224046166877, + "learning_rate": 1.650937679917052e-05, + "loss": 1.0259, + "step": 2728 + }, + { + "epoch": 1.481541802388708, + "grad_norm": 8.979875171863954, + "learning_rate": 1.6506706654248813e-05, + "loss": 1.0951, + "step": 2729 + }, + { + "epoch": 1.482084690553746, + "grad_norm": 7.700209331252355, + "learning_rate": 1.650403570456726e-05, + "loss": 1.0228, + "step": 2730 + }, + { + "epoch": 1.482627578718784, + "grad_norm": 9.090566362046857, + "learning_rate": 1.65013639504562e-05, + "loss": 1.0199, + "step": 2731 + }, + { + "epoch": 1.483170466883822, + "grad_norm": 6.644791741009562, + "learning_rate": 1.6498691392246088e-05, + "loss": 0.5478, + "step": 2732 + }, + { + "epoch": 1.48371335504886, + "grad_norm": 7.067718277939746, + "learning_rate": 1.6496018030267467e-05, + "loss": 0.864, + "step": 2733 + }, + { + "epoch": 1.484256243213898, + "grad_norm": 8.017737461642762, + "learning_rate": 1.6493343864850984e-05, + "loss": 1.1688, + "step": 2734 + }, + { + "epoch": 1.484799131378936, + "grad_norm": 6.479930584435578, + "learning_rate": 1.6490668896327382e-05, + "loss": 0.7227, + "step": 2735 + }, + { + "epoch": 1.485342019543974, + "grad_norm": 8.392289588009671, + "learning_rate": 1.6487993125027504e-05, + "loss": 0.7213, + "step": 2736 + }, + { + "epoch": 1.485884907709012, + "grad_norm": 7.817891248616158, + "learning_rate": 1.64853165512823e-05, + "loss": 0.7521, + "step": 2737 + }, + { + "epoch": 1.48642779587405, + "grad_norm": 8.824885297850575, + "learning_rate": 1.6482639175422804e-05, + "loss": 1.016, + "step": 2738 + }, + { + "epoch": 1.486970684039088, + "grad_norm": 11.48655873033237, + "learning_rate": 1.6479960997780165e-05, + "loss": 0.9814, + "step": 2739 + }, + { + "epoch": 1.487513572204126, + "grad_norm": 10.88221555712033, + "learning_rate": 1.6477282018685628e-05, + "loss": 1.464, + "step": 2740 + }, + { + "epoch": 1.488056460369164, + "grad_norm": 8.701247869952772, + "learning_rate": 1.6474602238470524e-05, + "loss": 0.918, + "step": 2741 + }, + { + "epoch": 1.488599348534202, + "grad_norm": 6.4327269926453505, + "learning_rate": 1.6471921657466294e-05, + "loss": 0.7949, + "step": 2742 + }, + { + "epoch": 1.48914223669924, + "grad_norm": 7.592315420975657, + "learning_rate": 1.6469240276004477e-05, + "loss": 0.9485, + "step": 2743 + }, + { + "epoch": 1.489685124864278, + "grad_norm": 7.682698616258038, + "learning_rate": 1.6466558094416717e-05, + "loss": 0.8439, + "step": 2744 + }, + { + "epoch": 1.490228013029316, + "grad_norm": 6.127626205001148, + "learning_rate": 1.6463875113034743e-05, + "loss": 0.8652, + "step": 2745 + }, + { + "epoch": 1.490770901194354, + "grad_norm": 9.65248236696942, + "learning_rate": 1.6461191332190397e-05, + "loss": 1.6278, + "step": 2746 + }, + { + "epoch": 1.491313789359392, + "grad_norm": 8.46309710838681, + "learning_rate": 1.6458506752215603e-05, + "loss": 1.5697, + "step": 2747 + }, + { + "epoch": 1.49185667752443, + "grad_norm": 6.578253971678976, + "learning_rate": 1.6455821373442407e-05, + "loss": 0.8836, + "step": 2748 + }, + { + "epoch": 1.492399565689468, + "grad_norm": 5.927853369454149, + "learning_rate": 1.645313519620293e-05, + "loss": 0.7182, + "step": 2749 + }, + { + "epoch": 1.492942453854506, + "grad_norm": 7.725891960915116, + "learning_rate": 1.645044822082941e-05, + "loss": 1.1794, + "step": 2750 + }, + { + "epoch": 1.493485342019544, + "grad_norm": 8.975220195142995, + "learning_rate": 1.644776044765417e-05, + "loss": 0.8909, + "step": 2751 + }, + { + "epoch": 1.494028230184582, + "grad_norm": 7.34643792387778, + "learning_rate": 1.6445071877009643e-05, + "loss": 1.0047, + "step": 2752 + }, + { + "epoch": 1.49457111834962, + "grad_norm": 7.561064198594243, + "learning_rate": 1.6442382509228355e-05, + "loss": 0.7533, + "step": 2753 + }, + { + "epoch": 1.495114006514658, + "grad_norm": 9.327085390245353, + "learning_rate": 1.6439692344642933e-05, + "loss": 0.8048, + "step": 2754 + }, + { + "epoch": 1.495656894679696, + "grad_norm": 8.662401438640156, + "learning_rate": 1.6437001383586095e-05, + "loss": 1.2513, + "step": 2755 + }, + { + "epoch": 1.496199782844734, + "grad_norm": 7.683989985812991, + "learning_rate": 1.6434309626390667e-05, + "loss": 0.8005, + "step": 2756 + }, + { + "epoch": 1.496742671009772, + "grad_norm": 6.295413547214335, + "learning_rate": 1.6431617073389574e-05, + "loss": 0.8778, + "step": 2757 + }, + { + "epoch": 1.49728555917481, + "grad_norm": 10.390849259618786, + "learning_rate": 1.6428923724915825e-05, + "loss": 1.6234, + "step": 2758 + }, + { + "epoch": 1.497828447339848, + "grad_norm": 6.815996692584369, + "learning_rate": 1.6426229581302545e-05, + "loss": 0.8142, + "step": 2759 + }, + { + "epoch": 1.498371335504886, + "grad_norm": 7.203388917066917, + "learning_rate": 1.642353464288295e-05, + "loss": 1.1273, + "step": 2760 + }, + { + "epoch": 1.498914223669924, + "grad_norm": 6.45381693158517, + "learning_rate": 1.6420838909990356e-05, + "loss": 0.7989, + "step": 2761 + }, + { + "epoch": 1.499457111834962, + "grad_norm": 7.135095705156932, + "learning_rate": 1.6418142382958167e-05, + "loss": 0.6753, + "step": 2762 + }, + { + "epoch": 1.5, + "grad_norm": 8.263302638252904, + "learning_rate": 1.64154450621199e-05, + "loss": 1.0357, + "step": 2763 + }, + { + "epoch": 1.500542888165038, + "grad_norm": 8.19537793615059, + "learning_rate": 1.6412746947809165e-05, + "loss": 0.8135, + "step": 2764 + }, + { + "epoch": 1.501085776330076, + "grad_norm": 5.084112201229359, + "learning_rate": 1.6410048040359665e-05, + "loss": 0.3466, + "step": 2765 + }, + { + "epoch": 1.501628664495114, + "grad_norm": 6.621661501175659, + "learning_rate": 1.6407348340105208e-05, + "loss": 1.1468, + "step": 2766 + }, + { + "epoch": 1.502171552660152, + "grad_norm": 7.5339066223692495, + "learning_rate": 1.6404647847379696e-05, + "loss": 0.8808, + "step": 2767 + }, + { + "epoch": 1.50271444082519, + "grad_norm": 6.501647846893153, + "learning_rate": 1.6401946562517134e-05, + "loss": 0.7589, + "step": 2768 + }, + { + "epoch": 1.503257328990228, + "grad_norm": 8.29356168583173, + "learning_rate": 1.6399244485851614e-05, + "loss": 1.3969, + "step": 2769 + }, + { + "epoch": 1.503800217155266, + "grad_norm": 6.518146966456392, + "learning_rate": 1.6396541617717337e-05, + "loss": 0.7951, + "step": 2770 + }, + { + "epoch": 1.504343105320304, + "grad_norm": 6.51134736256057, + "learning_rate": 1.63938379584486e-05, + "loss": 0.9073, + "step": 2771 + }, + { + "epoch": 1.504885993485342, + "grad_norm": 6.682459044312825, + "learning_rate": 1.6391133508379797e-05, + "loss": 0.733, + "step": 2772 + }, + { + "epoch": 1.50542888165038, + "grad_norm": 6.390687464891663, + "learning_rate": 1.638842826784541e-05, + "loss": 0.6934, + "step": 2773 + }, + { + "epoch": 1.505971769815418, + "grad_norm": 8.130197298030478, + "learning_rate": 1.6385722237180038e-05, + "loss": 0.8567, + "step": 2774 + }, + { + "epoch": 1.506514657980456, + "grad_norm": 7.959636836831885, + "learning_rate": 1.6383015416718356e-05, + "loss": 0.9926, + "step": 2775 + }, + { + "epoch": 1.507057546145494, + "grad_norm": 7.732366564125716, + "learning_rate": 1.638030780679516e-05, + "loss": 1.2228, + "step": 2776 + }, + { + "epoch": 1.507600434310532, + "grad_norm": 4.633651796670195, + "learning_rate": 1.6377599407745324e-05, + "loss": 0.4633, + "step": 2777 + }, + { + "epoch": 1.50814332247557, + "grad_norm": 7.016748827947926, + "learning_rate": 1.6374890219903828e-05, + "loss": 0.9593, + "step": 2778 + }, + { + "epoch": 1.508686210640608, + "grad_norm": 8.428067394567327, + "learning_rate": 1.637218024360575e-05, + "loss": 1.0493, + "step": 2779 + }, + { + "epoch": 1.509229098805646, + "grad_norm": 8.297567044395613, + "learning_rate": 1.6369469479186266e-05, + "loss": 1.2335, + "step": 2780 + }, + { + "epoch": 1.509771986970684, + "grad_norm": 5.8755950768313205, + "learning_rate": 1.6366757926980643e-05, + "loss": 0.6556, + "step": 2781 + }, + { + "epoch": 1.510314875135722, + "grad_norm": 6.837923725489964, + "learning_rate": 1.6364045587324254e-05, + "loss": 0.7031, + "step": 2782 + }, + { + "epoch": 1.51085776330076, + "grad_norm": 6.718777411432131, + "learning_rate": 1.6361332460552565e-05, + "loss": 0.8982, + "step": 2783 + }, + { + "epoch": 1.511400651465798, + "grad_norm": 8.788667184000067, + "learning_rate": 1.6358618547001137e-05, + "loss": 0.8548, + "step": 2784 + }, + { + "epoch": 1.511943539630836, + "grad_norm": 6.559897969720367, + "learning_rate": 1.635590384700563e-05, + "loss": 0.9627, + "step": 2785 + }, + { + "epoch": 1.512486427795874, + "grad_norm": 6.538968751920423, + "learning_rate": 1.635318836090181e-05, + "loss": 1.0731, + "step": 2786 + }, + { + "epoch": 1.513029315960912, + "grad_norm": 8.149646454407613, + "learning_rate": 1.6350472089025523e-05, + "loss": 0.9898, + "step": 2787 + }, + { + "epoch": 1.51357220412595, + "grad_norm": 8.544281989077668, + "learning_rate": 1.6347755031712734e-05, + "loss": 1.1125, + "step": 2788 + }, + { + "epoch": 1.514115092290988, + "grad_norm": 6.930736082842491, + "learning_rate": 1.634503718929948e-05, + "loss": 0.7617, + "step": 2789 + }, + { + "epoch": 1.514657980456026, + "grad_norm": 8.527931500643358, + "learning_rate": 1.6342318562121916e-05, + "loss": 1.2179, + "step": 2790 + }, + { + "epoch": 1.515200868621064, + "grad_norm": 7.509610271006853, + "learning_rate": 1.6339599150516283e-05, + "loss": 0.908, + "step": 2791 + }, + { + "epoch": 1.515743756786102, + "grad_norm": 6.0902820346277045, + "learning_rate": 1.6336878954818926e-05, + "loss": 0.829, + "step": 2792 + }, + { + "epoch": 1.51628664495114, + "grad_norm": 7.3274444131512935, + "learning_rate": 1.6334157975366278e-05, + "loss": 0.9667, + "step": 2793 + }, + { + "epoch": 1.516829533116178, + "grad_norm": 7.753641450951333, + "learning_rate": 1.633143621249488e-05, + "loss": 0.7961, + "step": 2794 + }, + { + "epoch": 1.517372421281216, + "grad_norm": 14.786266283108317, + "learning_rate": 1.6328713666541357e-05, + "loss": 1.4119, + "step": 2795 + }, + { + "epoch": 1.517915309446254, + "grad_norm": 6.1388168060810875, + "learning_rate": 1.632599033784244e-05, + "loss": 0.7726, + "step": 2796 + }, + { + "epoch": 1.518458197611292, + "grad_norm": 9.134190897524304, + "learning_rate": 1.632326622673496e-05, + "loss": 1.208, + "step": 2797 + }, + { + "epoch": 1.51900108577633, + "grad_norm": 6.898277731717493, + "learning_rate": 1.632054133355583e-05, + "loss": 0.6929, + "step": 2798 + }, + { + "epoch": 1.519543973941368, + "grad_norm": 6.90513711625334, + "learning_rate": 1.631781565864208e-05, + "loss": 0.7279, + "step": 2799 + }, + { + "epoch": 1.520086862106406, + "grad_norm": 8.332173251791465, + "learning_rate": 1.6315089202330817e-05, + "loss": 0.987, + "step": 2800 + }, + { + "epoch": 1.520629750271444, + "grad_norm": 6.783078863499027, + "learning_rate": 1.631236196495926e-05, + "loss": 0.6793, + "step": 2801 + }, + { + "epoch": 1.521172638436482, + "grad_norm": 7.841204229725534, + "learning_rate": 1.6309633946864712e-05, + "loss": 0.832, + "step": 2802 + }, + { + "epoch": 1.52171552660152, + "grad_norm": 6.4931904342190405, + "learning_rate": 1.630690514838458e-05, + "loss": 0.6662, + "step": 2803 + }, + { + "epoch": 1.522258414766558, + "grad_norm": 8.50187086749661, + "learning_rate": 1.6304175569856368e-05, + "loss": 0.8359, + "step": 2804 + }, + { + "epoch": 1.522801302931596, + "grad_norm": 6.196174007418195, + "learning_rate": 1.6301445211617676e-05, + "loss": 0.8437, + "step": 2805 + }, + { + "epoch": 1.523344191096634, + "grad_norm": 8.232047070633076, + "learning_rate": 1.6298714074006196e-05, + "loss": 0.9134, + "step": 2806 + }, + { + "epoch": 1.523887079261672, + "grad_norm": 6.594428796101092, + "learning_rate": 1.629598215735972e-05, + "loss": 0.6703, + "step": 2807 + }, + { + "epoch": 1.52442996742671, + "grad_norm": 7.634546429927066, + "learning_rate": 1.629324946201614e-05, + "loss": 0.8208, + "step": 2808 + }, + { + "epoch": 1.524972855591748, + "grad_norm": 8.225245989940067, + "learning_rate": 1.6290515988313432e-05, + "loss": 1.0002, + "step": 2809 + }, + { + "epoch": 1.5255157437567861, + "grad_norm": 8.334512060802671, + "learning_rate": 1.628778173658968e-05, + "loss": 0.8883, + "step": 2810 + }, + { + "epoch": 1.5260586319218241, + "grad_norm": 7.733244019009635, + "learning_rate": 1.6285046707183068e-05, + "loss": 0.7155, + "step": 2811 + }, + { + "epoch": 1.5266015200868621, + "grad_norm": 9.607081900344873, + "learning_rate": 1.628231090043186e-05, + "loss": 0.8321, + "step": 2812 + }, + { + "epoch": 1.5271444082519001, + "grad_norm": 8.390046905361096, + "learning_rate": 1.6279574316674426e-05, + "loss": 0.8818, + "step": 2813 + }, + { + "epoch": 1.5276872964169381, + "grad_norm": 7.608191285481891, + "learning_rate": 1.6276836956249235e-05, + "loss": 0.7382, + "step": 2814 + }, + { + "epoch": 1.5282301845819761, + "grad_norm": 8.998269112751723, + "learning_rate": 1.6274098819494844e-05, + "loss": 0.9285, + "step": 2815 + }, + { + "epoch": 1.5287730727470141, + "grad_norm": 5.877268783539307, + "learning_rate": 1.627135990674991e-05, + "loss": 0.8434, + "step": 2816 + }, + { + "epoch": 1.5293159609120521, + "grad_norm": 6.956730702284698, + "learning_rate": 1.6268620218353188e-05, + "loss": 0.5463, + "step": 2817 + }, + { + "epoch": 1.5298588490770901, + "grad_norm": 6.516430093047641, + "learning_rate": 1.626587975464353e-05, + "loss": 0.4886, + "step": 2818 + }, + { + "epoch": 1.5304017372421281, + "grad_norm": 6.751333342823053, + "learning_rate": 1.626313851595987e-05, + "loss": 0.6279, + "step": 2819 + }, + { + "epoch": 1.5309446254071661, + "grad_norm": 8.120808986494415, + "learning_rate": 1.6260396502641264e-05, + "loss": 0.8192, + "step": 2820 + }, + { + "epoch": 1.5314875135722041, + "grad_norm": 7.570228095805488, + "learning_rate": 1.6257653715026837e-05, + "loss": 0.5813, + "step": 2821 + }, + { + "epoch": 1.5320304017372421, + "grad_norm": 8.436330907807918, + "learning_rate": 1.625491015345583e-05, + "loss": 0.9115, + "step": 2822 + }, + { + "epoch": 1.5325732899022801, + "grad_norm": 8.384048301051982, + "learning_rate": 1.6252165818267564e-05, + "loss": 1.0131, + "step": 2823 + }, + { + "epoch": 1.5331161780673181, + "grad_norm": 6.665532827172479, + "learning_rate": 1.6249420709801462e-05, + "loss": 0.5454, + "step": 2824 + }, + { + "epoch": 1.5336590662323561, + "grad_norm": 7.096056294730765, + "learning_rate": 1.624667482839705e-05, + "loss": 1.043, + "step": 2825 + }, + { + "epoch": 1.5342019543973942, + "grad_norm": 5.384161272714602, + "learning_rate": 1.6243928174393935e-05, + "loss": 0.8345, + "step": 2826 + }, + { + "epoch": 1.5347448425624322, + "grad_norm": 8.398527018753486, + "learning_rate": 1.6241180748131834e-05, + "loss": 0.6911, + "step": 2827 + }, + { + "epoch": 1.5352877307274702, + "grad_norm": 6.880924484125559, + "learning_rate": 1.6238432549950552e-05, + "loss": 0.6039, + "step": 2828 + }, + { + "epoch": 1.5358306188925082, + "grad_norm": 9.23669884568822, + "learning_rate": 1.623568358018999e-05, + "loss": 1.1145, + "step": 2829 + }, + { + "epoch": 1.5363735070575462, + "grad_norm": 9.838376476241027, + "learning_rate": 1.6232933839190146e-05, + "loss": 1.0435, + "step": 2830 + }, + { + "epoch": 1.5369163952225842, + "grad_norm": 8.64475205930121, + "learning_rate": 1.6230183327291108e-05, + "loss": 1.0886, + "step": 2831 + }, + { + "epoch": 1.5374592833876222, + "grad_norm": 8.446847258487077, + "learning_rate": 1.6227432044833072e-05, + "loss": 0.8922, + "step": 2832 + }, + { + "epoch": 1.5380021715526602, + "grad_norm": 7.373724851347831, + "learning_rate": 1.622467999215631e-05, + "loss": 0.7823, + "step": 2833 + }, + { + "epoch": 1.5385450597176982, + "grad_norm": 9.661226436708493, + "learning_rate": 1.622192716960121e-05, + "loss": 1.2651, + "step": 2834 + }, + { + "epoch": 1.5390879478827362, + "grad_norm": 9.316115050127708, + "learning_rate": 1.6219173577508237e-05, + "loss": 0.897, + "step": 2835 + }, + { + "epoch": 1.5396308360477742, + "grad_norm": 12.011440258586251, + "learning_rate": 1.621641921621797e-05, + "loss": 0.868, + "step": 2836 + }, + { + "epoch": 1.5401737242128122, + "grad_norm": 7.737580132404622, + "learning_rate": 1.6213664086071058e-05, + "loss": 1.0727, + "step": 2837 + }, + { + "epoch": 1.5407166123778502, + "grad_norm": 6.696935187657473, + "learning_rate": 1.6210908187408275e-05, + "loss": 0.7404, + "step": 2838 + }, + { + "epoch": 1.5412595005428882, + "grad_norm": 6.460846604904407, + "learning_rate": 1.6208151520570465e-05, + "loss": 0.5892, + "step": 2839 + }, + { + "epoch": 1.5418023887079262, + "grad_norm": 8.244871763303658, + "learning_rate": 1.6205394085898586e-05, + "loss": 1.0692, + "step": 2840 + }, + { + "epoch": 1.5423452768729642, + "grad_norm": 6.140143281113326, + "learning_rate": 1.620263588373367e-05, + "loss": 0.7648, + "step": 2841 + }, + { + "epoch": 1.5428881650380022, + "grad_norm": 6.3369705303866235, + "learning_rate": 1.619987691441687e-05, + "loss": 0.6665, + "step": 2842 + }, + { + "epoch": 1.5434310532030402, + "grad_norm": 8.864356355873657, + "learning_rate": 1.6197117178289405e-05, + "loss": 0.9855, + "step": 2843 + }, + { + "epoch": 1.5439739413680782, + "grad_norm": 8.487808912288623, + "learning_rate": 1.6194356675692614e-05, + "loss": 1.3813, + "step": 2844 + }, + { + "epoch": 1.5445168295331162, + "grad_norm": 6.5782618837581435, + "learning_rate": 1.619159540696792e-05, + "loss": 1.1941, + "step": 2845 + }, + { + "epoch": 1.5450597176981542, + "grad_norm": 8.668580644259926, + "learning_rate": 1.6188833372456833e-05, + "loss": 1.1044, + "step": 2846 + }, + { + "epoch": 1.5456026058631922, + "grad_norm": 8.677506532765584, + "learning_rate": 1.6186070572500972e-05, + "loss": 1.3984, + "step": 2847 + }, + { + "epoch": 1.5461454940282302, + "grad_norm": 6.822422879140709, + "learning_rate": 1.6183307007442046e-05, + "loss": 0.6903, + "step": 2848 + }, + { + "epoch": 1.5466883821932682, + "grad_norm": 8.480828664812217, + "learning_rate": 1.6180542677621852e-05, + "loss": 1.2466, + "step": 2849 + }, + { + "epoch": 1.5472312703583062, + "grad_norm": 7.804300661419004, + "learning_rate": 1.617777758338229e-05, + "loss": 0.8833, + "step": 2850 + }, + { + "epoch": 1.5477741585233442, + "grad_norm": 6.418809144902363, + "learning_rate": 1.617501172506535e-05, + "loss": 0.9174, + "step": 2851 + }, + { + "epoch": 1.5483170466883822, + "grad_norm": 6.627083812958508, + "learning_rate": 1.617224510301312e-05, + "loss": 0.7867, + "step": 2852 + }, + { + "epoch": 1.5488599348534202, + "grad_norm": 7.792114190076133, + "learning_rate": 1.616947771756778e-05, + "loss": 0.6242, + "step": 2853 + }, + { + "epoch": 1.5494028230184582, + "grad_norm": 6.965108246318713, + "learning_rate": 1.6166709569071598e-05, + "loss": 0.5844, + "step": 2854 + }, + { + "epoch": 1.5499457111834962, + "grad_norm": 6.9966373703206655, + "learning_rate": 1.616394065786695e-05, + "loss": 1.1688, + "step": 2855 + }, + { + "epoch": 1.5504885993485342, + "grad_norm": 8.466603441198636, + "learning_rate": 1.6161170984296298e-05, + "loss": 0.9506, + "step": 2856 + }, + { + "epoch": 1.5510314875135722, + "grad_norm": 6.361127259105106, + "learning_rate": 1.61584005487022e-05, + "loss": 0.5573, + "step": 2857 + }, + { + "epoch": 1.5515743756786102, + "grad_norm": 6.137955961198825, + "learning_rate": 1.6155629351427306e-05, + "loss": 1.0474, + "step": 2858 + }, + { + "epoch": 1.5521172638436482, + "grad_norm": 6.262211837195136, + "learning_rate": 1.6152857392814367e-05, + "loss": 0.6786, + "step": 2859 + }, + { + "epoch": 1.5526601520086862, + "grad_norm": 9.3980797837583, + "learning_rate": 1.6150084673206214e-05, + "loss": 1.2893, + "step": 2860 + }, + { + "epoch": 1.5532030401737242, + "grad_norm": 9.600537200283373, + "learning_rate": 1.614731119294579e-05, + "loss": 0.8668, + "step": 2861 + }, + { + "epoch": 1.5537459283387622, + "grad_norm": 8.65304941049683, + "learning_rate": 1.614453695237612e-05, + "loss": 1.0377, + "step": 2862 + }, + { + "epoch": 1.5542888165038002, + "grad_norm": 6.075747510405264, + "learning_rate": 1.6141761951840327e-05, + "loss": 0.8203, + "step": 2863 + }, + { + "epoch": 1.5548317046688382, + "grad_norm": 6.578770531031146, + "learning_rate": 1.6138986191681626e-05, + "loss": 0.6808, + "step": 2864 + }, + { + "epoch": 1.5553745928338762, + "grad_norm": 9.206515322081769, + "learning_rate": 1.6136209672243332e-05, + "loss": 0.9111, + "step": 2865 + }, + { + "epoch": 1.5559174809989142, + "grad_norm": 8.829723041453656, + "learning_rate": 1.613343239386884e-05, + "loss": 1.7392, + "step": 2866 + }, + { + "epoch": 1.5564603691639523, + "grad_norm": 9.489367814866748, + "learning_rate": 1.613065435690166e-05, + "loss": 0.9186, + "step": 2867 + }, + { + "epoch": 1.5570032573289903, + "grad_norm": 7.7196736231335175, + "learning_rate": 1.6127875561685376e-05, + "loss": 0.769, + "step": 2868 + }, + { + "epoch": 1.5575461454940283, + "grad_norm": 7.668511006453606, + "learning_rate": 1.6125096008563677e-05, + "loss": 0.7572, + "step": 2869 + }, + { + "epoch": 1.5580890336590663, + "grad_norm": 8.31551289254719, + "learning_rate": 1.6122315697880343e-05, + "loss": 1.0657, + "step": 2870 + }, + { + "epoch": 1.5586319218241043, + "grad_norm": 11.828690633927348, + "learning_rate": 1.6119534629979244e-05, + "loss": 1.0095, + "step": 2871 + }, + { + "epoch": 1.5591748099891423, + "grad_norm": 7.137823181609677, + "learning_rate": 1.611675280520435e-05, + "loss": 0.7246, + "step": 2872 + }, + { + "epoch": 1.5597176981541803, + "grad_norm": 5.051313327297567, + "learning_rate": 1.611397022389972e-05, + "loss": 0.5905, + "step": 2873 + }, + { + "epoch": 1.5602605863192183, + "grad_norm": 8.165805991671068, + "learning_rate": 1.6111186886409504e-05, + "loss": 1.0678, + "step": 2874 + }, + { + "epoch": 1.5608034744842563, + "grad_norm": 11.597649604311457, + "learning_rate": 1.6108402793077957e-05, + "loss": 1.4836, + "step": 2875 + }, + { + "epoch": 1.5613463626492943, + "grad_norm": 7.953203184026488, + "learning_rate": 1.610561794424942e-05, + "loss": 0.9216, + "step": 2876 + }, + { + "epoch": 1.5618892508143323, + "grad_norm": 7.397930684647979, + "learning_rate": 1.6102832340268322e-05, + "loss": 0.6688, + "step": 2877 + }, + { + "epoch": 1.5624321389793703, + "grad_norm": 8.999084750858366, + "learning_rate": 1.6100045981479195e-05, + "loss": 1.254, + "step": 2878 + }, + { + "epoch": 1.5629750271444083, + "grad_norm": 7.445594920489225, + "learning_rate": 1.6097258868226658e-05, + "loss": 0.6406, + "step": 2879 + }, + { + "epoch": 1.5635179153094463, + "grad_norm": 10.469355169581702, + "learning_rate": 1.609447100085543e-05, + "loss": 0.7928, + "step": 2880 + }, + { + "epoch": 1.5640608034744843, + "grad_norm": 8.853634937458189, + "learning_rate": 1.6091682379710313e-05, + "loss": 1.0788, + "step": 2881 + }, + { + "epoch": 1.5646036916395223, + "grad_norm": 7.909620213759866, + "learning_rate": 1.6088893005136206e-05, + "loss": 1.011, + "step": 2882 + }, + { + "epoch": 1.5651465798045603, + "grad_norm": 10.379252728708625, + "learning_rate": 1.6086102877478117e-05, + "loss": 0.8675, + "step": 2883 + }, + { + "epoch": 1.5656894679695983, + "grad_norm": 8.143888785034273, + "learning_rate": 1.6083311997081116e-05, + "loss": 0.8935, + "step": 2884 + }, + { + "epoch": 1.5662323561346363, + "grad_norm": 8.577520797942114, + "learning_rate": 1.6080520364290396e-05, + "loss": 0.6763, + "step": 2885 + }, + { + "epoch": 1.5667752442996743, + "grad_norm": 9.226094142154105, + "learning_rate": 1.6077727979451228e-05, + "loss": 0.6679, + "step": 2886 + }, + { + "epoch": 1.5673181324647123, + "grad_norm": 11.197536431412058, + "learning_rate": 1.607493484290897e-05, + "loss": 1.0525, + "step": 2887 + }, + { + "epoch": 1.5678610206297503, + "grad_norm": 8.43012225245798, + "learning_rate": 1.6072140955009093e-05, + "loss": 1.0317, + "step": 2888 + }, + { + "epoch": 1.5684039087947883, + "grad_norm": 6.69358812320454, + "learning_rate": 1.606934631609715e-05, + "loss": 0.7703, + "step": 2889 + }, + { + "epoch": 1.5689467969598263, + "grad_norm": 9.36327291463899, + "learning_rate": 1.6066550926518776e-05, + "loss": 1.0844, + "step": 2890 + }, + { + "epoch": 1.5694896851248643, + "grad_norm": 8.40518653624583, + "learning_rate": 1.6063754786619716e-05, + "loss": 0.6863, + "step": 2891 + }, + { + "epoch": 1.5700325732899023, + "grad_norm": 9.552486022227878, + "learning_rate": 1.60609578967458e-05, + "loss": 1.0342, + "step": 2892 + }, + { + "epoch": 1.5705754614549403, + "grad_norm": 6.76445711513327, + "learning_rate": 1.6058160257242953e-05, + "loss": 0.8777, + "step": 2893 + }, + { + "epoch": 1.5711183496199783, + "grad_norm": 9.040602694092291, + "learning_rate": 1.6055361868457188e-05, + "loss": 0.9634, + "step": 2894 + }, + { + "epoch": 1.5716612377850163, + "grad_norm": 7.00278619777266, + "learning_rate": 1.6052562730734614e-05, + "loss": 0.8995, + "step": 2895 + }, + { + "epoch": 1.5722041259500543, + "grad_norm": 8.533327622490077, + "learning_rate": 1.604976284442144e-05, + "loss": 1.0896, + "step": 2896 + }, + { + "epoch": 1.5727470141150923, + "grad_norm": 9.895218231842426, + "learning_rate": 1.6046962209863953e-05, + "loss": 0.9309, + "step": 2897 + }, + { + "epoch": 1.5732899022801303, + "grad_norm": 5.945823958044379, + "learning_rate": 1.604416082740854e-05, + "loss": 0.6512, + "step": 2898 + }, + { + "epoch": 1.5738327904451683, + "grad_norm": 10.822429137471252, + "learning_rate": 1.6041358697401687e-05, + "loss": 1.0744, + "step": 2899 + }, + { + "epoch": 1.5743756786102063, + "grad_norm": 7.728147258511819, + "learning_rate": 1.603855582018996e-05, + "loss": 0.9566, + "step": 2900 + }, + { + "epoch": 1.5749185667752443, + "grad_norm": 8.643748761880973, + "learning_rate": 1.603575219612003e-05, + "loss": 0.9188, + "step": 2901 + }, + { + "epoch": 1.5754614549402823, + "grad_norm": 9.357572097357403, + "learning_rate": 1.603294782553864e-05, + "loss": 0.9045, + "step": 2902 + }, + { + "epoch": 1.5760043431053203, + "grad_norm": 6.76207345266047, + "learning_rate": 1.6030142708792653e-05, + "loss": 0.7658, + "step": 2903 + }, + { + "epoch": 1.5765472312703583, + "grad_norm": 6.47603683733189, + "learning_rate": 1.6027336846229005e-05, + "loss": 0.5406, + "step": 2904 + }, + { + "epoch": 1.5770901194353963, + "grad_norm": 6.030033648190992, + "learning_rate": 1.602453023819473e-05, + "loss": 0.5141, + "step": 2905 + }, + { + "epoch": 1.5776330076004343, + "grad_norm": 8.344821553681937, + "learning_rate": 1.6021722885036954e-05, + "loss": 1.255, + "step": 2906 + }, + { + "epoch": 1.5781758957654723, + "grad_norm": 7.2392259468937885, + "learning_rate": 1.601891478710289e-05, + "loss": 0.8358, + "step": 2907 + }, + { + "epoch": 1.5787187839305103, + "grad_norm": 8.011046574978268, + "learning_rate": 1.6016105944739856e-05, + "loss": 1.0316, + "step": 2908 + }, + { + "epoch": 1.5792616720955484, + "grad_norm": 6.609140115647256, + "learning_rate": 1.601329635829525e-05, + "loss": 0.7324, + "step": 2909 + }, + { + "epoch": 1.5798045602605864, + "grad_norm": 6.672864395642082, + "learning_rate": 1.6010486028116568e-05, + "loss": 0.6386, + "step": 2910 + }, + { + "epoch": 1.5803474484256244, + "grad_norm": 7.270447246601273, + "learning_rate": 1.600767495455139e-05, + "loss": 0.6699, + "step": 2911 + }, + { + "epoch": 1.5808903365906624, + "grad_norm": 9.7186017734922, + "learning_rate": 1.6004863137947405e-05, + "loss": 0.9604, + "step": 2912 + }, + { + "epoch": 1.5814332247557004, + "grad_norm": 7.18305646407285, + "learning_rate": 1.6002050578652374e-05, + "loss": 0.7286, + "step": 2913 + }, + { + "epoch": 1.5819761129207384, + "grad_norm": 8.171408196684215, + "learning_rate": 1.5999237277014162e-05, + "loss": 0.9121, + "step": 2914 + }, + { + "epoch": 1.5825190010857764, + "grad_norm": 7.710673421134947, + "learning_rate": 1.599642323338072e-05, + "loss": 0.7287, + "step": 2915 + }, + { + "epoch": 1.5830618892508144, + "grad_norm": 8.264934044661224, + "learning_rate": 1.5993608448100095e-05, + "loss": 0.6537, + "step": 2916 + }, + { + "epoch": 1.5836047774158524, + "grad_norm": 7.874914961028911, + "learning_rate": 1.599079292152043e-05, + "loss": 0.6511, + "step": 2917 + }, + { + "epoch": 1.5841476655808904, + "grad_norm": 7.399378371785218, + "learning_rate": 1.5987976653989945e-05, + "loss": 0.6236, + "step": 2918 + }, + { + "epoch": 1.5846905537459284, + "grad_norm": 7.837764322959507, + "learning_rate": 1.5985159645856966e-05, + "loss": 0.8184, + "step": 2919 + }, + { + "epoch": 1.5852334419109664, + "grad_norm": 8.64272103945922, + "learning_rate": 1.5982341897469903e-05, + "loss": 0.9182, + "step": 2920 + }, + { + "epoch": 1.5857763300760044, + "grad_norm": 9.994158036669248, + "learning_rate": 1.5979523409177254e-05, + "loss": 1.3466, + "step": 2921 + }, + { + "epoch": 1.5863192182410424, + "grad_norm": 6.671268632879107, + "learning_rate": 1.5976704181327626e-05, + "loss": 0.7506, + "step": 2922 + }, + { + "epoch": 1.5868621064060804, + "grad_norm": 8.161895609948308, + "learning_rate": 1.59738842142697e-05, + "loss": 0.9683, + "step": 2923 + }, + { + "epoch": 1.5874049945711184, + "grad_norm": 7.236619510875819, + "learning_rate": 1.597106350835225e-05, + "loss": 0.798, + "step": 2924 + }, + { + "epoch": 1.5879478827361564, + "grad_norm": 10.263209688443808, + "learning_rate": 1.5968242063924152e-05, + "loss": 1.0799, + "step": 2925 + }, + { + "epoch": 1.5884907709011944, + "grad_norm": 9.433308260958457, + "learning_rate": 1.596541988133436e-05, + "loss": 0.8755, + "step": 2926 + }, + { + "epoch": 1.5890336590662324, + "grad_norm": 9.48995441422385, + "learning_rate": 1.5962596960931927e-05, + "loss": 1.4712, + "step": 2927 + }, + { + "epoch": 1.5895765472312704, + "grad_norm": 7.317297215230638, + "learning_rate": 1.5959773303066005e-05, + "loss": 1.0421, + "step": 2928 + }, + { + "epoch": 1.5901194353963084, + "grad_norm": 7.491426794362114, + "learning_rate": 1.595694890808582e-05, + "loss": 0.9468, + "step": 2929 + }, + { + "epoch": 1.5906623235613464, + "grad_norm": 7.320440852081977, + "learning_rate": 1.5954123776340702e-05, + "loss": 0.7459, + "step": 2930 + }, + { + "epoch": 1.5912052117263844, + "grad_norm": 8.334158307527314, + "learning_rate": 1.5951297908180062e-05, + "loss": 1.3182, + "step": 2931 + }, + { + "epoch": 1.5917480998914224, + "grad_norm": 7.02613629768153, + "learning_rate": 1.5948471303953418e-05, + "loss": 0.919, + "step": 2932 + }, + { + "epoch": 1.5922909880564604, + "grad_norm": 6.904443487786306, + "learning_rate": 1.594564396401036e-05, + "loss": 0.5737, + "step": 2933 + }, + { + "epoch": 1.5928338762214984, + "grad_norm": 8.041576521446402, + "learning_rate": 1.594281588870058e-05, + "loss": 0.7191, + "step": 2934 + }, + { + "epoch": 1.5933767643865364, + "grad_norm": 8.557472225459021, + "learning_rate": 1.5939987078373856e-05, + "loss": 0.8936, + "step": 2935 + }, + { + "epoch": 1.5939196525515744, + "grad_norm": 10.051674589973098, + "learning_rate": 1.5937157533380065e-05, + "loss": 1.1777, + "step": 2936 + }, + { + "epoch": 1.5944625407166124, + "grad_norm": 7.1580018473739235, + "learning_rate": 1.5934327254069167e-05, + "loss": 0.8628, + "step": 2937 + }, + { + "epoch": 1.5950054288816504, + "grad_norm": 8.689338133459966, + "learning_rate": 1.593149624079122e-05, + "loss": 1.1877, + "step": 2938 + }, + { + "epoch": 1.5955483170466884, + "grad_norm": 8.40904043481009, + "learning_rate": 1.5928664493896364e-05, + "loss": 0.9815, + "step": 2939 + }, + { + "epoch": 1.5960912052117264, + "grad_norm": 7.778305354024487, + "learning_rate": 1.5925832013734832e-05, + "loss": 0.8282, + "step": 2940 + }, + { + "epoch": 1.5966340933767644, + "grad_norm": 7.074281482828989, + "learning_rate": 1.5922998800656956e-05, + "loss": 0.8347, + "step": 2941 + }, + { + "epoch": 1.5971769815418024, + "grad_norm": 9.188840060350078, + "learning_rate": 1.5920164855013145e-05, + "loss": 0.8566, + "step": 2942 + }, + { + "epoch": 1.5977198697068404, + "grad_norm": 7.021116417649334, + "learning_rate": 1.591733017715391e-05, + "loss": 0.8367, + "step": 2943 + }, + { + "epoch": 1.5982627578718784, + "grad_norm": 8.039041768789302, + "learning_rate": 1.5914494767429846e-05, + "loss": 0.9889, + "step": 2944 + }, + { + "epoch": 1.5988056460369164, + "grad_norm": 6.252701385804125, + "learning_rate": 1.5911658626191645e-05, + "loss": 1.0478, + "step": 2945 + }, + { + "epoch": 1.5993485342019544, + "grad_norm": 8.223409826369243, + "learning_rate": 1.5908821753790083e-05, + "loss": 1.2458, + "step": 2946 + }, + { + "epoch": 1.5998914223669924, + "grad_norm": 8.972690403999096, + "learning_rate": 1.590598415057603e-05, + "loss": 0.8471, + "step": 2947 + }, + { + "epoch": 1.6004343105320304, + "grad_norm": 8.163032020074416, + "learning_rate": 1.5903145816900445e-05, + "loss": 0.8213, + "step": 2948 + }, + { + "epoch": 1.6009771986970684, + "grad_norm": 5.667868413023322, + "learning_rate": 1.5900306753114375e-05, + "loss": 0.8426, + "step": 2949 + }, + { + "epoch": 1.6015200868621065, + "grad_norm": 6.960744978365833, + "learning_rate": 1.5897466959568967e-05, + "loss": 0.7224, + "step": 2950 + }, + { + "epoch": 1.6020629750271445, + "grad_norm": 7.648210267430006, + "learning_rate": 1.589462643661544e-05, + "loss": 0.7451, + "step": 2951 + }, + { + "epoch": 1.6026058631921825, + "grad_norm": 5.9568596208122875, + "learning_rate": 1.5891785184605123e-05, + "loss": 0.5252, + "step": 2952 + }, + { + "epoch": 1.6031487513572205, + "grad_norm": 6.41676218105273, + "learning_rate": 1.5888943203889427e-05, + "loss": 0.6502, + "step": 2953 + }, + { + "epoch": 1.6036916395222585, + "grad_norm": 7.704871607753619, + "learning_rate": 1.5886100494819846e-05, + "loss": 1.1762, + "step": 2954 + }, + { + "epoch": 1.6042345276872965, + "grad_norm": 7.656219757403813, + "learning_rate": 1.5883257057747975e-05, + "loss": 0.8076, + "step": 2955 + }, + { + "epoch": 1.6047774158523345, + "grad_norm": 6.897030921112048, + "learning_rate": 1.58804128930255e-05, + "loss": 0.883, + "step": 2956 + }, + { + "epoch": 1.6053203040173725, + "grad_norm": 6.718724904599124, + "learning_rate": 1.5877568001004182e-05, + "loss": 0.8533, + "step": 2957 + }, + { + "epoch": 1.6058631921824105, + "grad_norm": 8.458861423777194, + "learning_rate": 1.5874722382035887e-05, + "loss": 1.2017, + "step": 2958 + }, + { + "epoch": 1.6064060803474485, + "grad_norm": 9.16622725543938, + "learning_rate": 1.5871876036472565e-05, + "loss": 1.0074, + "step": 2959 + }, + { + "epoch": 1.6069489685124865, + "grad_norm": 7.053786350768386, + "learning_rate": 1.5869028964666254e-05, + "loss": 0.8193, + "step": 2960 + }, + { + "epoch": 1.6074918566775245, + "grad_norm": 7.670976634760575, + "learning_rate": 1.5866181166969088e-05, + "loss": 0.8331, + "step": 2961 + }, + { + "epoch": 1.6080347448425625, + "grad_norm": 9.942186867470006, + "learning_rate": 1.586333264373329e-05, + "loss": 0.8672, + "step": 2962 + }, + { + "epoch": 1.6085776330076005, + "grad_norm": 7.744463789862724, + "learning_rate": 1.586048339531116e-05, + "loss": 0.8859, + "step": 2963 + }, + { + "epoch": 1.6091205211726385, + "grad_norm": 7.181553947846623, + "learning_rate": 1.5857633422055104e-05, + "loss": 1.0376, + "step": 2964 + }, + { + "epoch": 1.6096634093376765, + "grad_norm": 7.184586780377298, + "learning_rate": 1.5854782724317616e-05, + "loss": 0.694, + "step": 2965 + }, + { + "epoch": 1.6102062975027145, + "grad_norm": 6.9154974167720855, + "learning_rate": 1.5851931302451262e-05, + "loss": 0.8573, + "step": 2966 + }, + { + "epoch": 1.6107491856677525, + "grad_norm": 5.056944401416877, + "learning_rate": 1.5849079156808726e-05, + "loss": 0.4972, + "step": 2967 + }, + { + "epoch": 1.6112920738327905, + "grad_norm": 6.98635366168859, + "learning_rate": 1.584622628774275e-05, + "loss": 0.9002, + "step": 2968 + }, + { + "epoch": 1.6118349619978285, + "grad_norm": 9.22418255248469, + "learning_rate": 1.5843372695606196e-05, + "loss": 1.4143, + "step": 2969 + }, + { + "epoch": 1.6123778501628665, + "grad_norm": 9.49766654897752, + "learning_rate": 1.584051838075199e-05, + "loss": 1.0145, + "step": 2970 + }, + { + "epoch": 1.6129207383279045, + "grad_norm": 9.681135461557052, + "learning_rate": 1.5837663343533166e-05, + "loss": 0.8841, + "step": 2971 + }, + { + "epoch": 1.6134636264929425, + "grad_norm": 6.00661489729409, + "learning_rate": 1.583480758430283e-05, + "loss": 0.677, + "step": 2972 + }, + { + "epoch": 1.6140065146579805, + "grad_norm": 7.843885336692957, + "learning_rate": 1.5831951103414194e-05, + "loss": 0.4729, + "step": 2973 + }, + { + "epoch": 1.6145494028230185, + "grad_norm": 9.391971866087287, + "learning_rate": 1.5829093901220557e-05, + "loss": 1.0833, + "step": 2974 + }, + { + "epoch": 1.6150922909880565, + "grad_norm": 7.202123614778415, + "learning_rate": 1.582623597807529e-05, + "loss": 0.9755, + "step": 2975 + }, + { + "epoch": 1.6156351791530945, + "grad_norm": 8.050976157508988, + "learning_rate": 1.5823377334331875e-05, + "loss": 0.8981, + "step": 2976 + }, + { + "epoch": 1.6161780673181325, + "grad_norm": 6.409986462897389, + "learning_rate": 1.5820517970343867e-05, + "loss": 0.3994, + "step": 2977 + }, + { + "epoch": 1.6167209554831705, + "grad_norm": 6.001133896600598, + "learning_rate": 1.581765788646492e-05, + "loss": 0.4701, + "step": 2978 + }, + { + "epoch": 1.6172638436482085, + "grad_norm": 12.858193971885504, + "learning_rate": 1.581479708304878e-05, + "loss": 1.4327, + "step": 2979 + }, + { + "epoch": 1.6178067318132465, + "grad_norm": 9.559044098763156, + "learning_rate": 1.5811935560449262e-05, + "loss": 0.9058, + "step": 2980 + }, + { + "epoch": 1.6183496199782845, + "grad_norm": 8.041859101089647, + "learning_rate": 1.5809073319020293e-05, + "loss": 0.6251, + "step": 2981 + }, + { + "epoch": 1.6188925081433225, + "grad_norm": 7.413714357196275, + "learning_rate": 1.580621035911588e-05, + "loss": 0.7887, + "step": 2982 + }, + { + "epoch": 1.6194353963083605, + "grad_norm": 11.941805543277479, + "learning_rate": 1.5803346681090113e-05, + "loss": 1.2547, + "step": 2983 + }, + { + "epoch": 1.6199782844733985, + "grad_norm": 5.728003242952451, + "learning_rate": 1.580048228529718e-05, + "loss": 0.4941, + "step": 2984 + }, + { + "epoch": 1.6205211726384365, + "grad_norm": 8.67398306592389, + "learning_rate": 1.5797617172091354e-05, + "loss": 0.7758, + "step": 2985 + }, + { + "epoch": 1.6210640608034745, + "grad_norm": 6.198163559825324, + "learning_rate": 1.5794751341826996e-05, + "loss": 0.4601, + "step": 2986 + }, + { + "epoch": 1.6216069489685125, + "grad_norm": 8.257469675141675, + "learning_rate": 1.5791884794858557e-05, + "loss": 0.8679, + "step": 2987 + }, + { + "epoch": 1.6221498371335505, + "grad_norm": 12.953868836673164, + "learning_rate": 1.5789017531540575e-05, + "loss": 1.572, + "step": 2988 + }, + { + "epoch": 1.6226927252985885, + "grad_norm": 9.232767459675694, + "learning_rate": 1.5786149552227682e-05, + "loss": 0.7511, + "step": 2989 + }, + { + "epoch": 1.6232356134636265, + "grad_norm": 7.465499623540017, + "learning_rate": 1.5783280857274586e-05, + "loss": 0.5754, + "step": 2990 + }, + { + "epoch": 1.6237785016286646, + "grad_norm": 5.087952971939767, + "learning_rate": 1.5780411447036097e-05, + "loss": 0.532, + "step": 2991 + }, + { + "epoch": 1.6243213897937026, + "grad_norm": 11.064786965172743, + "learning_rate": 1.577754132186711e-05, + "loss": 0.7195, + "step": 2992 + }, + { + "epoch": 1.6248642779587406, + "grad_norm": 10.268274657097892, + "learning_rate": 1.57746704821226e-05, + "loss": 1.1258, + "step": 2993 + }, + { + "epoch": 1.6254071661237783, + "grad_norm": 7.650078703788778, + "learning_rate": 1.5771798928157645e-05, + "loss": 0.5949, + "step": 2994 + }, + { + "epoch": 1.6259500542888166, + "grad_norm": 9.330939436907839, + "learning_rate": 1.5768926660327396e-05, + "loss": 0.9243, + "step": 2995 + }, + { + "epoch": 1.6264929424538543, + "grad_norm": 11.566902464205237, + "learning_rate": 1.576605367898711e-05, + "loss": 1.5588, + "step": 2996 + }, + { + "epoch": 1.6270358306188926, + "grad_norm": 9.91827217636312, + "learning_rate": 1.576317998449211e-05, + "loss": 1.1418, + "step": 2997 + }, + { + "epoch": 1.6275787187839303, + "grad_norm": 5.842910545051031, + "learning_rate": 1.5760305577197824e-05, + "loss": 0.5064, + "step": 2998 + }, + { + "epoch": 1.6281216069489686, + "grad_norm": 7.20637691224012, + "learning_rate": 1.5757430457459765e-05, + "loss": 0.8129, + "step": 2999 + }, + { + "epoch": 1.6286644951140063, + "grad_norm": 7.5338669003432965, + "learning_rate": 1.5754554625633535e-05, + "loss": 0.7244, + "step": 3000 + }, + { + "epoch": 1.6292073832790446, + "grad_norm": 7.489311206471124, + "learning_rate": 1.5751678082074813e-05, + "loss": 1.1409, + "step": 3001 + }, + { + "epoch": 1.6297502714440824, + "grad_norm": 7.26299222123934, + "learning_rate": 1.574880082713938e-05, + "loss": 0.7061, + "step": 3002 + }, + { + "epoch": 1.6302931596091206, + "grad_norm": 6.292246912680897, + "learning_rate": 1.5745922861183095e-05, + "loss": 0.6383, + "step": 3003 + }, + { + "epoch": 1.6308360477741584, + "grad_norm": 7.483086031387996, + "learning_rate": 1.574304418456192e-05, + "loss": 0.8194, + "step": 3004 + }, + { + "epoch": 1.6313789359391966, + "grad_norm": 8.062451557552306, + "learning_rate": 1.5740164797631882e-05, + "loss": 0.9474, + "step": 3005 + }, + { + "epoch": 1.6319218241042344, + "grad_norm": 7.62904254588456, + "learning_rate": 1.5737284700749116e-05, + "loss": 1.2324, + "step": 3006 + }, + { + "epoch": 1.6324647122692726, + "grad_norm": 8.533361938490982, + "learning_rate": 1.573440389426983e-05, + "loss": 0.6182, + "step": 3007 + }, + { + "epoch": 1.6330076004343104, + "grad_norm": 7.0451736765954855, + "learning_rate": 1.5731522378550337e-05, + "loss": 0.7464, + "step": 3008 + }, + { + "epoch": 1.6335504885993486, + "grad_norm": 8.121336222871497, + "learning_rate": 1.572864015394702e-05, + "loss": 1.1087, + "step": 3009 + }, + { + "epoch": 1.6340933767643864, + "grad_norm": 7.2112540180194955, + "learning_rate": 1.5725757220816356e-05, + "loss": 0.7756, + "step": 3010 + }, + { + "epoch": 1.6346362649294246, + "grad_norm": 8.092755806422884, + "learning_rate": 1.5722873579514915e-05, + "loss": 1.1184, + "step": 3011 + }, + { + "epoch": 1.6351791530944624, + "grad_norm": 7.485157971844339, + "learning_rate": 1.5719989230399347e-05, + "loss": 0.6839, + "step": 3012 + }, + { + "epoch": 1.6357220412595006, + "grad_norm": 8.516980380341119, + "learning_rate": 1.5717104173826397e-05, + "loss": 0.948, + "step": 3013 + }, + { + "epoch": 1.6362649294245384, + "grad_norm": 7.602547351124631, + "learning_rate": 1.571421841015289e-05, + "loss": 0.8479, + "step": 3014 + }, + { + "epoch": 1.6368078175895766, + "grad_norm": 7.0470960880026245, + "learning_rate": 1.5711331939735744e-05, + "loss": 0.9014, + "step": 3015 + }, + { + "epoch": 1.6373507057546144, + "grad_norm": 6.760810209806174, + "learning_rate": 1.570844476293196e-05, + "loss": 0.9316, + "step": 3016 + }, + { + "epoch": 1.6378935939196526, + "grad_norm": 8.02053578699558, + "learning_rate": 1.570555688009863e-05, + "loss": 0.7837, + "step": 3017 + }, + { + "epoch": 1.6384364820846904, + "grad_norm": 4.624578610614632, + "learning_rate": 1.5702668291592936e-05, + "loss": 0.4706, + "step": 3018 + }, + { + "epoch": 1.6389793702497286, + "grad_norm": 8.011712265192347, + "learning_rate": 1.569977899777213e-05, + "loss": 1.0338, + "step": 3019 + }, + { + "epoch": 1.6395222584147664, + "grad_norm": 7.773921575637797, + "learning_rate": 1.569688899899358e-05, + "loss": 0.7512, + "step": 3020 + }, + { + "epoch": 1.6400651465798046, + "grad_norm": 12.19433508686472, + "learning_rate": 1.569399829561472e-05, + "loss": 1.3502, + "step": 3021 + }, + { + "epoch": 1.6406080347448424, + "grad_norm": 11.319107865931453, + "learning_rate": 1.569110688799307e-05, + "loss": 0.9996, + "step": 3022 + }, + { + "epoch": 1.6411509229098806, + "grad_norm": 7.1911299711947585, + "learning_rate": 1.5688214776486255e-05, + "loss": 0.9037, + "step": 3023 + }, + { + "epoch": 1.6416938110749184, + "grad_norm": 8.326843836043436, + "learning_rate": 1.5685321961451968e-05, + "loss": 1.1179, + "step": 3024 + }, + { + "epoch": 1.6422366992399566, + "grad_norm": 8.257534801333424, + "learning_rate": 1.5682428443248002e-05, + "loss": 0.7197, + "step": 3025 + }, + { + "epoch": 1.6427795874049944, + "grad_norm": 6.669062512754087, + "learning_rate": 1.567953422223223e-05, + "loss": 0.6292, + "step": 3026 + }, + { + "epoch": 1.6433224755700326, + "grad_norm": 7.99477616152165, + "learning_rate": 1.567663929876261e-05, + "loss": 0.926, + "step": 3027 + }, + { + "epoch": 1.6438653637350704, + "grad_norm": 5.3717184947082615, + "learning_rate": 1.56737436731972e-05, + "loss": 0.5378, + "step": 3028 + }, + { + "epoch": 1.6444082519001086, + "grad_norm": 6.522826762176589, + "learning_rate": 1.5670847345894125e-05, + "loss": 0.7393, + "step": 3029 + }, + { + "epoch": 1.6449511400651464, + "grad_norm": 9.503371738808127, + "learning_rate": 1.5667950317211612e-05, + "loss": 1.2727, + "step": 3030 + }, + { + "epoch": 1.6454940282301846, + "grad_norm": 8.208077816222845, + "learning_rate": 1.5665052587507974e-05, + "loss": 0.7032, + "step": 3031 + }, + { + "epoch": 1.6460369163952224, + "grad_norm": 9.101101018529372, + "learning_rate": 1.56621541571416e-05, + "loss": 0.6983, + "step": 3032 + }, + { + "epoch": 1.6465798045602607, + "grad_norm": 7.254188682380924, + "learning_rate": 1.565925502647098e-05, + "loss": 0.7196, + "step": 3033 + }, + { + "epoch": 1.6471226927252984, + "grad_norm": 7.837018641144827, + "learning_rate": 1.5656355195854676e-05, + "loss": 1.2035, + "step": 3034 + }, + { + "epoch": 1.6476655808903367, + "grad_norm": 8.949191823273495, + "learning_rate": 1.5653454665651344e-05, + "loss": 1.0169, + "step": 3035 + }, + { + "epoch": 1.6482084690553744, + "grad_norm": 8.166317934132502, + "learning_rate": 1.5650553436219732e-05, + "loss": 0.672, + "step": 3036 + }, + { + "epoch": 1.6487513572204127, + "grad_norm": 6.443983188038021, + "learning_rate": 1.564765150791866e-05, + "loss": 0.5552, + "step": 3037 + }, + { + "epoch": 1.6492942453854504, + "grad_norm": 10.403060348162544, + "learning_rate": 1.5644748881107057e-05, + "loss": 1.0709, + "step": 3038 + }, + { + "epoch": 1.6498371335504887, + "grad_norm": 8.319421686790951, + "learning_rate": 1.564184555614391e-05, + "loss": 0.7923, + "step": 3039 + }, + { + "epoch": 1.6503800217155264, + "grad_norm": 9.65474819234395, + "learning_rate": 1.5638941533388318e-05, + "loss": 0.8519, + "step": 3040 + }, + { + "epoch": 1.6509229098805647, + "grad_norm": 8.635959327913778, + "learning_rate": 1.5636036813199445e-05, + "loss": 0.8447, + "step": 3041 + }, + { + "epoch": 1.6514657980456025, + "grad_norm": 7.674155625456285, + "learning_rate": 1.563313139593656e-05, + "loss": 0.7205, + "step": 3042 + }, + { + "epoch": 1.6520086862106407, + "grad_norm": 7.249478120616782, + "learning_rate": 1.5630225281959003e-05, + "loss": 0.6632, + "step": 3043 + }, + { + "epoch": 1.6525515743756785, + "grad_norm": 8.90384172612564, + "learning_rate": 1.5627318471626208e-05, + "loss": 0.8243, + "step": 3044 + }, + { + "epoch": 1.6530944625407167, + "grad_norm": 8.274462585020828, + "learning_rate": 1.5624410965297703e-05, + "loss": 0.7381, + "step": 3045 + }, + { + "epoch": 1.6536373507057545, + "grad_norm": 7.266215459792981, + "learning_rate": 1.562150276333308e-05, + "loss": 0.4989, + "step": 3046 + }, + { + "epoch": 1.6541802388707927, + "grad_norm": 4.816150056433193, + "learning_rate": 1.5618593866092036e-05, + "loss": 0.5382, + "step": 3047 + }, + { + "epoch": 1.6547231270358305, + "grad_norm": 6.835417235425346, + "learning_rate": 1.561568427393435e-05, + "loss": 1.044, + "step": 3048 + }, + { + "epoch": 1.6552660152008687, + "grad_norm": 7.2539344971377435, + "learning_rate": 1.5612773987219885e-05, + "loss": 0.5901, + "step": 3049 + }, + { + "epoch": 1.6558089033659065, + "grad_norm": 11.003646663954257, + "learning_rate": 1.5609863006308586e-05, + "loss": 0.9407, + "step": 3050 + }, + { + "epoch": 1.6563517915309447, + "grad_norm": 8.268114443948154, + "learning_rate": 1.560695133156049e-05, + "loss": 1.0164, + "step": 3051 + }, + { + "epoch": 1.6568946796959825, + "grad_norm": 7.824779774533612, + "learning_rate": 1.5604038963335716e-05, + "loss": 0.7111, + "step": 3052 + }, + { + "epoch": 1.6574375678610207, + "grad_norm": 8.542948361290183, + "learning_rate": 1.560112590199447e-05, + "loss": 0.6501, + "step": 3053 + }, + { + "epoch": 1.6579804560260585, + "grad_norm": 7.85824626614946, + "learning_rate": 1.5598212147897047e-05, + "loss": 0.752, + "step": 3054 + }, + { + "epoch": 1.6585233441910967, + "grad_norm": 6.588621135611242, + "learning_rate": 1.559529770140382e-05, + "loss": 0.507, + "step": 3055 + }, + { + "epoch": 1.6590662323561345, + "grad_norm": 6.418764525989286, + "learning_rate": 1.559238256287526e-05, + "loss": 0.5947, + "step": 3056 + }, + { + "epoch": 1.6596091205211727, + "grad_norm": 7.681071515192806, + "learning_rate": 1.5589466732671913e-05, + "loss": 0.8854, + "step": 3057 + }, + { + "epoch": 1.6601520086862105, + "grad_norm": 5.918230957225734, + "learning_rate": 1.558655021115441e-05, + "loss": 0.5546, + "step": 3058 + }, + { + "epoch": 1.6606948968512487, + "grad_norm": 10.669803193569846, + "learning_rate": 1.5583632998683475e-05, + "loss": 1.1548, + "step": 3059 + }, + { + "epoch": 1.6612377850162865, + "grad_norm": 9.271579004183549, + "learning_rate": 1.558071509561991e-05, + "loss": 0.8049, + "step": 3060 + }, + { + "epoch": 1.6617806731813247, + "grad_norm": 8.337085337830533, + "learning_rate": 1.557779650232461e-05, + "loss": 0.954, + "step": 3061 + }, + { + "epoch": 1.6623235613463625, + "grad_norm": 11.145075141347627, + "learning_rate": 1.5574877219158543e-05, + "loss": 1.1592, + "step": 3062 + }, + { + "epoch": 1.6628664495114007, + "grad_norm": 13.953049521423079, + "learning_rate": 1.557195724648278e-05, + "loss": 1.5699, + "step": 3063 + }, + { + "epoch": 1.6634093376764385, + "grad_norm": 8.654584332416588, + "learning_rate": 1.5569036584658466e-05, + "loss": 1.0185, + "step": 3064 + }, + { + "epoch": 1.6639522258414767, + "grad_norm": 7.520260829640502, + "learning_rate": 1.556611523404683e-05, + "loss": 0.7898, + "step": 3065 + }, + { + "epoch": 1.6644951140065145, + "grad_norm": 6.7373805901465875, + "learning_rate": 1.5563193195009188e-05, + "loss": 0.8392, + "step": 3066 + }, + { + "epoch": 1.6650380021715527, + "grad_norm": 6.732598633253056, + "learning_rate": 1.556027046790695e-05, + "loss": 0.6672, + "step": 3067 + }, + { + "epoch": 1.6655808903365905, + "grad_norm": 9.217725100287737, + "learning_rate": 1.55573470531016e-05, + "loss": 0.9874, + "step": 3068 + }, + { + "epoch": 1.6661237785016287, + "grad_norm": 6.963043776606882, + "learning_rate": 1.5554422950954706e-05, + "loss": 0.812, + "step": 3069 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 6.449074540999346, + "learning_rate": 1.5551498161827928e-05, + "loss": 0.5035, + "step": 3070 + }, + { + "epoch": 1.6672095548317047, + "grad_norm": 7.793693453975787, + "learning_rate": 1.554857268608301e-05, + "loss": 0.7974, + "step": 3071 + }, + { + "epoch": 1.6677524429967425, + "grad_norm": 10.124707886106192, + "learning_rate": 1.554564652408178e-05, + "loss": 1.4946, + "step": 3072 + }, + { + "epoch": 1.6682953311617807, + "grad_norm": 5.481775138355068, + "learning_rate": 1.5542719676186153e-05, + "loss": 0.502, + "step": 3073 + }, + { + "epoch": 1.6688382193268185, + "grad_norm": 6.055974601989459, + "learning_rate": 1.5539792142758114e-05, + "loss": 0.7463, + "step": 3074 + }, + { + "epoch": 1.6693811074918568, + "grad_norm": 7.452121005302326, + "learning_rate": 1.5536863924159762e-05, + "loss": 0.7282, + "step": 3075 + }, + { + "epoch": 1.6699239956568945, + "grad_norm": 8.991453230181806, + "learning_rate": 1.5533935020753252e-05, + "loss": 0.9327, + "step": 3076 + }, + { + "epoch": 1.6704668838219328, + "grad_norm": 8.138196256677595, + "learning_rate": 1.5531005432900838e-05, + "loss": 0.658, + "step": 3077 + }, + { + "epoch": 1.6710097719869705, + "grad_norm": 6.674313540160648, + "learning_rate": 1.552807516096486e-05, + "loss": 0.569, + "step": 3078 + }, + { + "epoch": 1.6715526601520088, + "grad_norm": 7.2303438904479105, + "learning_rate": 1.5525144205307734e-05, + "loss": 0.5186, + "step": 3079 + }, + { + "epoch": 1.6720955483170465, + "grad_norm": 6.872945037718312, + "learning_rate": 1.5522212566291966e-05, + "loss": 0.7282, + "step": 3080 + }, + { + "epoch": 1.6726384364820848, + "grad_norm": 7.913543579080126, + "learning_rate": 1.5519280244280146e-05, + "loss": 0.7015, + "step": 3081 + }, + { + "epoch": 1.6731813246471225, + "grad_norm": 6.144062932600708, + "learning_rate": 1.551634723963495e-05, + "loss": 0.7829, + "step": 3082 + }, + { + "epoch": 1.6737242128121608, + "grad_norm": 7.111349951384809, + "learning_rate": 1.5513413552719137e-05, + "loss": 0.6904, + "step": 3083 + }, + { + "epoch": 1.6742671009771986, + "grad_norm": 10.453360973997349, + "learning_rate": 1.5510479183895545e-05, + "loss": 1.4584, + "step": 3084 + }, + { + "epoch": 1.6748099891422368, + "grad_norm": 13.976555611119428, + "learning_rate": 1.5507544133527108e-05, + "loss": 1.542, + "step": 3085 + }, + { + "epoch": 1.6753528773072746, + "grad_norm": 5.894489690173581, + "learning_rate": 1.5504608401976835e-05, + "loss": 0.6792, + "step": 3086 + }, + { + "epoch": 1.6758957654723128, + "grad_norm": 8.937162315762803, + "learning_rate": 1.550167198960782e-05, + "loss": 0.8243, + "step": 3087 + }, + { + "epoch": 1.6764386536373506, + "grad_norm": 9.787723535687237, + "learning_rate": 1.549873489678325e-05, + "loss": 1.0071, + "step": 3088 + }, + { + "epoch": 1.6769815418023888, + "grad_norm": 6.0371827348978755, + "learning_rate": 1.5495797123866378e-05, + "loss": 0.6407, + "step": 3089 + }, + { + "epoch": 1.6775244299674266, + "grad_norm": 5.997673025227573, + "learning_rate": 1.5492858671220563e-05, + "loss": 0.7129, + "step": 3090 + }, + { + "epoch": 1.6780673181324648, + "grad_norm": 11.08813931574187, + "learning_rate": 1.5489919539209232e-05, + "loss": 0.8435, + "step": 3091 + }, + { + "epoch": 1.6786102062975026, + "grad_norm": 8.730951161091724, + "learning_rate": 1.54869797281959e-05, + "loss": 0.9159, + "step": 3092 + }, + { + "epoch": 1.6791530944625408, + "grad_norm": 7.293119927853279, + "learning_rate": 1.5484039238544173e-05, + "loss": 1.1001, + "step": 3093 + }, + { + "epoch": 1.6796959826275786, + "grad_norm": 7.091810627268002, + "learning_rate": 1.5481098070617734e-05, + "loss": 0.747, + "step": 3094 + }, + { + "epoch": 1.6802388707926168, + "grad_norm": 9.079124871383598, + "learning_rate": 1.547815622478035e-05, + "loss": 1.2292, + "step": 3095 + }, + { + "epoch": 1.6807817589576546, + "grad_norm": 7.238161650182196, + "learning_rate": 1.5475213701395867e-05, + "loss": 0.6657, + "step": 3096 + }, + { + "epoch": 1.6813246471226928, + "grad_norm": 9.385885763398004, + "learning_rate": 1.5472270500828236e-05, + "loss": 1.305, + "step": 3097 + }, + { + "epoch": 1.6818675352877306, + "grad_norm": 8.98902696621929, + "learning_rate": 1.5469326623441463e-05, + "loss": 0.8641, + "step": 3098 + }, + { + "epoch": 1.6824104234527688, + "grad_norm": 6.519773416264212, + "learning_rate": 1.5466382069599656e-05, + "loss": 0.4567, + "step": 3099 + }, + { + "epoch": 1.6829533116178066, + "grad_norm": 6.775917495491388, + "learning_rate": 1.5463436839667007e-05, + "loss": 0.6874, + "step": 3100 + }, + { + "epoch": 1.6834961997828448, + "grad_norm": 7.82569199258335, + "learning_rate": 1.5460490934007776e-05, + "loss": 0.819, + "step": 3101 + }, + { + "epoch": 1.6840390879478826, + "grad_norm": 6.885216490352139, + "learning_rate": 1.5457544352986326e-05, + "loss": 0.7891, + "step": 3102 + }, + { + "epoch": 1.6845819761129208, + "grad_norm": 6.733247427994312, + "learning_rate": 1.5454597096967093e-05, + "loss": 0.6664, + "step": 3103 + }, + { + "epoch": 1.6851248642779586, + "grad_norm": 6.96769902934235, + "learning_rate": 1.5451649166314598e-05, + "loss": 0.6578, + "step": 3104 + }, + { + "epoch": 1.6856677524429968, + "grad_norm": 6.941713803652074, + "learning_rate": 1.5448700561393444e-05, + "loss": 0.5855, + "step": 3105 + }, + { + "epoch": 1.6862106406080346, + "grad_norm": 8.203921934361416, + "learning_rate": 1.5445751282568324e-05, + "loss": 0.9843, + "step": 3106 + }, + { + "epoch": 1.6867535287730728, + "grad_norm": 8.822160140984163, + "learning_rate": 1.5442801330204004e-05, + "loss": 1.1003, + "step": 3107 + }, + { + "epoch": 1.6872964169381106, + "grad_norm": 8.330851912949297, + "learning_rate": 1.5439850704665338e-05, + "loss": 1.0992, + "step": 3108 + }, + { + "epoch": 1.6878393051031488, + "grad_norm": 6.552171568470756, + "learning_rate": 1.543689940631727e-05, + "loss": 0.5434, + "step": 3109 + }, + { + "epoch": 1.6883821932681866, + "grad_norm": 5.772049074992556, + "learning_rate": 1.5433947435524822e-05, + "loss": 0.7515, + "step": 3110 + }, + { + "epoch": 1.6889250814332248, + "grad_norm": 9.575440058135785, + "learning_rate": 1.543099479265309e-05, + "loss": 0.9691, + "step": 3111 + }, + { + "epoch": 1.6894679695982626, + "grad_norm": 6.008162314259628, + "learning_rate": 1.5428041478067263e-05, + "loss": 0.6137, + "step": 3112 + }, + { + "epoch": 1.6900108577633008, + "grad_norm": 8.993171311918207, + "learning_rate": 1.542508749213262e-05, + "loss": 0.7608, + "step": 3113 + }, + { + "epoch": 1.6905537459283386, + "grad_norm": 8.042583538721042, + "learning_rate": 1.542213283521451e-05, + "loss": 0.8188, + "step": 3114 + }, + { + "epoch": 1.6910966340933768, + "grad_norm": 9.290959640283983, + "learning_rate": 1.541917750767837e-05, + "loss": 1.1737, + "step": 3115 + }, + { + "epoch": 1.6916395222584146, + "grad_norm": 6.3883435583657535, + "learning_rate": 1.5416221509889718e-05, + "loss": 0.4376, + "step": 3116 + }, + { + "epoch": 1.6921824104234529, + "grad_norm": 8.943225081508276, + "learning_rate": 1.541326484221416e-05, + "loss": 0.9983, + "step": 3117 + }, + { + "epoch": 1.6927252985884906, + "grad_norm": 7.583355336491315, + "learning_rate": 1.541030750501737e-05, + "loss": 0.6078, + "step": 3118 + }, + { + "epoch": 1.6932681867535289, + "grad_norm": 7.318011543277784, + "learning_rate": 1.5407349498665133e-05, + "loss": 0.745, + "step": 3119 + }, + { + "epoch": 1.6938110749185666, + "grad_norm": 6.212465291912028, + "learning_rate": 1.5404390823523287e-05, + "loss": 0.4739, + "step": 3120 + }, + { + "epoch": 1.6943539630836049, + "grad_norm": 8.223585888775384, + "learning_rate": 1.5401431479957775e-05, + "loss": 1.1819, + "step": 3121 + }, + { + "epoch": 1.6948968512486426, + "grad_norm": 7.957232407413796, + "learning_rate": 1.5398471468334605e-05, + "loss": 0.7071, + "step": 3122 + }, + { + "epoch": 1.6954397394136809, + "grad_norm": 5.52597775233848, + "learning_rate": 1.5395510789019884e-05, + "loss": 0.6171, + "step": 3123 + }, + { + "epoch": 1.6959826275787186, + "grad_norm": 8.015216509333372, + "learning_rate": 1.5392549442379785e-05, + "loss": 0.9051, + "step": 3124 + }, + { + "epoch": 1.6965255157437569, + "grad_norm": 8.330470177699917, + "learning_rate": 1.538958742878058e-05, + "loss": 0.5037, + "step": 3125 + }, + { + "epoch": 1.6970684039087947, + "grad_norm": 6.1063862074204245, + "learning_rate": 1.538662474858861e-05, + "loss": 0.6509, + "step": 3126 + }, + { + "epoch": 1.6976112920738329, + "grad_norm": 8.11378323370098, + "learning_rate": 1.5383661402170308e-05, + "loss": 1.0356, + "step": 3127 + }, + { + "epoch": 1.6981541802388707, + "grad_norm": 5.995828293111682, + "learning_rate": 1.5380697389892185e-05, + "loss": 0.5297, + "step": 3128 + }, + { + "epoch": 1.6986970684039089, + "grad_norm": 6.962108385144619, + "learning_rate": 1.537773271212083e-05, + "loss": 0.8711, + "step": 3129 + }, + { + "epoch": 1.6992399565689467, + "grad_norm": 8.69611973474616, + "learning_rate": 1.5374767369222922e-05, + "loss": 0.911, + "step": 3130 + }, + { + "epoch": 1.6997828447339849, + "grad_norm": 11.047832664292853, + "learning_rate": 1.5371801361565223e-05, + "loss": 0.8499, + "step": 3131 + }, + { + "epoch": 1.7003257328990227, + "grad_norm": 10.490928296052365, + "learning_rate": 1.5368834689514568e-05, + "loss": 0.7617, + "step": 3132 + }, + { + "epoch": 1.700868621064061, + "grad_norm": 8.723324926351989, + "learning_rate": 1.536586735343788e-05, + "loss": 0.7158, + "step": 3133 + }, + { + "epoch": 1.7014115092290987, + "grad_norm": 7.780646168464086, + "learning_rate": 1.536289935370217e-05, + "loss": 0.6452, + "step": 3134 + }, + { + "epoch": 1.701954397394137, + "grad_norm": 8.037748179528911, + "learning_rate": 1.5359930690674518e-05, + "loss": 0.8368, + "step": 3135 + }, + { + "epoch": 1.7024972855591747, + "grad_norm": 7.539822669872962, + "learning_rate": 1.5356961364722096e-05, + "loss": 0.7224, + "step": 3136 + }, + { + "epoch": 1.703040173724213, + "grad_norm": 10.929660828744602, + "learning_rate": 1.5353991376212155e-05, + "loss": 1.2459, + "step": 3137 + }, + { + "epoch": 1.7035830618892507, + "grad_norm": 11.118929306326729, + "learning_rate": 1.5351020725512028e-05, + "loss": 1.2769, + "step": 3138 + }, + { + "epoch": 1.704125950054289, + "grad_norm": 8.820097629927263, + "learning_rate": 1.534804941298913e-05, + "loss": 1.0722, + "step": 3139 + }, + { + "epoch": 1.7046688382193267, + "grad_norm": 8.79991079654506, + "learning_rate": 1.5345077439010956e-05, + "loss": 0.7596, + "step": 3140 + }, + { + "epoch": 1.705211726384365, + "grad_norm": 7.5009896900263096, + "learning_rate": 1.5342104803945087e-05, + "loss": 0.6457, + "step": 3141 + }, + { + "epoch": 1.7057546145494027, + "grad_norm": 9.190121352339093, + "learning_rate": 1.533913150815918e-05, + "loss": 0.7152, + "step": 3142 + }, + { + "epoch": 1.706297502714441, + "grad_norm": 7.209026897095485, + "learning_rate": 1.5336157552020977e-05, + "loss": 0.534, + "step": 3143 + }, + { + "epoch": 1.7068403908794787, + "grad_norm": 8.752247270964865, + "learning_rate": 1.5333182935898306e-05, + "loss": 1.0953, + "step": 3144 + }, + { + "epoch": 1.707383279044517, + "grad_norm": 8.714563086288697, + "learning_rate": 1.5330207660159068e-05, + "loss": 0.8016, + "step": 3145 + }, + { + "epoch": 1.7079261672095547, + "grad_norm": 9.540320185420228, + "learning_rate": 1.5327231725171255e-05, + "loss": 1.2679, + "step": 3146 + }, + { + "epoch": 1.708469055374593, + "grad_norm": 8.091263210809633, + "learning_rate": 1.532425513130293e-05, + "loss": 0.7521, + "step": 3147 + }, + { + "epoch": 1.7090119435396307, + "grad_norm": 10.062104841108379, + "learning_rate": 1.5321277878922246e-05, + "loss": 1.0696, + "step": 3148 + }, + { + "epoch": 1.709554831704669, + "grad_norm": 8.184716628635337, + "learning_rate": 1.531829996839743e-05, + "loss": 0.643, + "step": 3149 + }, + { + "epoch": 1.7100977198697067, + "grad_norm": 11.130045309860368, + "learning_rate": 1.53153214000968e-05, + "loss": 1.2565, + "step": 3150 + }, + { + "epoch": 1.710640608034745, + "grad_norm": 8.856470164658607, + "learning_rate": 1.5312342174388746e-05, + "loss": 0.8376, + "step": 3151 + }, + { + "epoch": 1.7111834961997827, + "grad_norm": 11.958863336707164, + "learning_rate": 1.5309362291641747e-05, + "loss": 0.978, + "step": 3152 + }, + { + "epoch": 1.711726384364821, + "grad_norm": 10.369392130776264, + "learning_rate": 1.5306381752224357e-05, + "loss": 1.1822, + "step": 3153 + }, + { + "epoch": 1.7122692725298587, + "grad_norm": 7.346196971786586, + "learning_rate": 1.5303400556505213e-05, + "loss": 0.9284, + "step": 3154 + }, + { + "epoch": 1.712812160694897, + "grad_norm": 6.652184209108613, + "learning_rate": 1.5300418704853042e-05, + "loss": 0.5787, + "step": 3155 + }, + { + "epoch": 1.7133550488599347, + "grad_norm": 9.745597378440513, + "learning_rate": 1.5297436197636634e-05, + "loss": 1.0194, + "step": 3156 + }, + { + "epoch": 1.713897937024973, + "grad_norm": 6.17200933356452, + "learning_rate": 1.5294453035224874e-05, + "loss": 0.6315, + "step": 3157 + }, + { + "epoch": 1.7144408251900107, + "grad_norm": 8.373044157038668, + "learning_rate": 1.5291469217986724e-05, + "loss": 0.9127, + "step": 3158 + }, + { + "epoch": 1.714983713355049, + "grad_norm": 9.811623309823384, + "learning_rate": 1.5288484746291227e-05, + "loss": 0.7791, + "step": 3159 + }, + { + "epoch": 1.7155266015200867, + "grad_norm": 9.264521529810194, + "learning_rate": 1.5285499620507513e-05, + "loss": 0.6764, + "step": 3160 + }, + { + "epoch": 1.716069489685125, + "grad_norm": 8.688562646594697, + "learning_rate": 1.5282513841004777e-05, + "loss": 1.0202, + "step": 3161 + }, + { + "epoch": 1.7166123778501627, + "grad_norm": 8.522452395448054, + "learning_rate": 1.527952740815231e-05, + "loss": 1.0043, + "step": 3162 + }, + { + "epoch": 1.717155266015201, + "grad_norm": 6.330008909603884, + "learning_rate": 1.527654032231948e-05, + "loss": 0.6651, + "step": 3163 + }, + { + "epoch": 1.7176981541802387, + "grad_norm": 9.314495623946685, + "learning_rate": 1.5273552583875736e-05, + "loss": 0.8312, + "step": 3164 + }, + { + "epoch": 1.718241042345277, + "grad_norm": 9.157480690732486, + "learning_rate": 1.52705641931906e-05, + "loss": 1.0017, + "step": 3165 + }, + { + "epoch": 1.7187839305103148, + "grad_norm": 8.300251884517943, + "learning_rate": 1.5267575150633687e-05, + "loss": 0.9929, + "step": 3166 + }, + { + "epoch": 1.719326818675353, + "grad_norm": 8.328644452020212, + "learning_rate": 1.5264585456574684e-05, + "loss": 0.8105, + "step": 3167 + }, + { + "epoch": 1.7198697068403908, + "grad_norm": 9.499775808882656, + "learning_rate": 1.526159511138336e-05, + "loss": 0.8593, + "step": 3168 + }, + { + "epoch": 1.720412595005429, + "grad_norm": 7.684886832360556, + "learning_rate": 1.5258604115429567e-05, + "loss": 0.5919, + "step": 3169 + }, + { + "epoch": 1.7209554831704668, + "grad_norm": 8.666869336691208, + "learning_rate": 1.5255612469083239e-05, + "loss": 0.6537, + "step": 3170 + }, + { + "epoch": 1.721498371335505, + "grad_norm": 11.871617471990923, + "learning_rate": 1.5252620172714378e-05, + "loss": 1.3826, + "step": 3171 + }, + { + "epoch": 1.7220412595005428, + "grad_norm": 6.088995555337644, + "learning_rate": 1.5249627226693089e-05, + "loss": 0.5085, + "step": 3172 + }, + { + "epoch": 1.722584147665581, + "grad_norm": 9.255760547122362, + "learning_rate": 1.5246633631389536e-05, + "loss": 0.8823, + "step": 3173 + }, + { + "epoch": 1.7231270358306188, + "grad_norm": 9.91965748862537, + "learning_rate": 1.5243639387173974e-05, + "loss": 0.8752, + "step": 3174 + }, + { + "epoch": 1.723669923995657, + "grad_norm": 6.397561552165108, + "learning_rate": 1.5240644494416734e-05, + "loss": 0.6673, + "step": 3175 + }, + { + "epoch": 1.7242128121606948, + "grad_norm": 7.066817871358342, + "learning_rate": 1.523764895348823e-05, + "loss": 1.041, + "step": 3176 + }, + { + "epoch": 1.724755700325733, + "grad_norm": 8.282830238008364, + "learning_rate": 1.5234652764758959e-05, + "loss": 0.9521, + "step": 3177 + }, + { + "epoch": 1.7252985884907708, + "grad_norm": 7.611547937947443, + "learning_rate": 1.523165592859949e-05, + "loss": 0.7182, + "step": 3178 + }, + { + "epoch": 1.725841476655809, + "grad_norm": 8.068698205768472, + "learning_rate": 1.5228658445380475e-05, + "loss": 0.9078, + "step": 3179 + }, + { + "epoch": 1.7263843648208468, + "grad_norm": 8.731772184014808, + "learning_rate": 1.5225660315472652e-05, + "loss": 1.2148, + "step": 3180 + }, + { + "epoch": 1.726927252985885, + "grad_norm": 7.785982141134894, + "learning_rate": 1.5222661539246832e-05, + "loss": 0.8653, + "step": 3181 + }, + { + "epoch": 1.7274701411509228, + "grad_norm": 7.6267653051373285, + "learning_rate": 1.521966211707391e-05, + "loss": 0.7454, + "step": 3182 + }, + { + "epoch": 1.728013029315961, + "grad_norm": 7.440711922406893, + "learning_rate": 1.521666204932486e-05, + "loss": 0.6576, + "step": 3183 + }, + { + "epoch": 1.7285559174809988, + "grad_norm": 7.69035147644333, + "learning_rate": 1.521366133637073e-05, + "loss": 0.8269, + "step": 3184 + }, + { + "epoch": 1.729098805646037, + "grad_norm": 8.387844627258719, + "learning_rate": 1.5210659978582662e-05, + "loss": 1.1049, + "step": 3185 + }, + { + "epoch": 1.7296416938110748, + "grad_norm": 9.08085381726355, + "learning_rate": 1.5207657976331862e-05, + "loss": 0.97, + "step": 3186 + }, + { + "epoch": 1.730184581976113, + "grad_norm": 7.621613071814458, + "learning_rate": 1.5204655329989617e-05, + "loss": 0.7113, + "step": 3187 + }, + { + "epoch": 1.7307274701411508, + "grad_norm": 6.743838306403499, + "learning_rate": 1.5201652039927313e-05, + "loss": 0.7307, + "step": 3188 + }, + { + "epoch": 1.731270358306189, + "grad_norm": 9.602658786805504, + "learning_rate": 1.5198648106516392e-05, + "loss": 1.0515, + "step": 3189 + }, + { + "epoch": 1.7318132464712268, + "grad_norm": 8.153765987898234, + "learning_rate": 1.5195643530128387e-05, + "loss": 1.0879, + "step": 3190 + }, + { + "epoch": 1.732356134636265, + "grad_norm": 9.745244927474419, + "learning_rate": 1.519263831113491e-05, + "loss": 0.9542, + "step": 3191 + }, + { + "epoch": 1.7328990228013028, + "grad_norm": 8.908425729214914, + "learning_rate": 1.5189632449907654e-05, + "loss": 0.8936, + "step": 3192 + }, + { + "epoch": 1.733441910966341, + "grad_norm": 7.409065356665995, + "learning_rate": 1.5186625946818382e-05, + "loss": 0.7645, + "step": 3193 + }, + { + "epoch": 1.7339847991313788, + "grad_norm": 7.958784484679066, + "learning_rate": 1.5183618802238949e-05, + "loss": 0.8512, + "step": 3194 + }, + { + "epoch": 1.734527687296417, + "grad_norm": 5.808082449761269, + "learning_rate": 1.5180611016541278e-05, + "loss": 0.7241, + "step": 3195 + }, + { + "epoch": 1.7350705754614548, + "grad_norm": 8.88976614296984, + "learning_rate": 1.5177602590097382e-05, + "loss": 1.0506, + "step": 3196 + }, + { + "epoch": 1.735613463626493, + "grad_norm": 8.197925091983134, + "learning_rate": 1.5174593523279346e-05, + "loss": 0.7959, + "step": 3197 + }, + { + "epoch": 1.7361563517915308, + "grad_norm": 10.2752921216035, + "learning_rate": 1.5171583816459334e-05, + "loss": 1.2859, + "step": 3198 + }, + { + "epoch": 1.736699239956569, + "grad_norm": 9.43498245507836, + "learning_rate": 1.5168573470009596e-05, + "loss": 1.0459, + "step": 3199 + }, + { + "epoch": 1.7372421281216068, + "grad_norm": 9.033995794822129, + "learning_rate": 1.516556248430245e-05, + "loss": 1.0003, + "step": 3200 + }, + { + "epoch": 1.737785016286645, + "grad_norm": 9.577836032406582, + "learning_rate": 1.5162550859710306e-05, + "loss": 1.2394, + "step": 3201 + }, + { + "epoch": 1.7383279044516828, + "grad_norm": 5.2063230201693775, + "learning_rate": 1.5159538596605642e-05, + "loss": 0.3625, + "step": 3202 + }, + { + "epoch": 1.738870792616721, + "grad_norm": 7.741298457702819, + "learning_rate": 1.5156525695361022e-05, + "loss": 0.6744, + "step": 3203 + }, + { + "epoch": 1.7394136807817588, + "grad_norm": 8.718382461609222, + "learning_rate": 1.5153512156349086e-05, + "loss": 1.0423, + "step": 3204 + }, + { + "epoch": 1.739956568946797, + "grad_norm": 6.706656901306365, + "learning_rate": 1.5150497979942556e-05, + "loss": 0.5121, + "step": 3205 + }, + { + "epoch": 1.7404994571118348, + "grad_norm": 9.934397785927217, + "learning_rate": 1.5147483166514225e-05, + "loss": 1.5269, + "step": 3206 + }, + { + "epoch": 1.741042345276873, + "grad_norm": 8.069224944733923, + "learning_rate": 1.514446771643697e-05, + "loss": 0.7801, + "step": 3207 + }, + { + "epoch": 1.7415852334419109, + "grad_norm": 9.193161021474104, + "learning_rate": 1.514145163008375e-05, + "loss": 1.2144, + "step": 3208 + }, + { + "epoch": 1.742128121606949, + "grad_norm": 6.5647281089901, + "learning_rate": 1.5138434907827599e-05, + "loss": 0.6692, + "step": 3209 + }, + { + "epoch": 1.7426710097719869, + "grad_norm": 9.387942500806554, + "learning_rate": 1.513541755004163e-05, + "loss": 1.1655, + "step": 3210 + }, + { + "epoch": 1.743213897937025, + "grad_norm": 6.936398185140981, + "learning_rate": 1.5132399557099031e-05, + "loss": 0.7761, + "step": 3211 + }, + { + "epoch": 1.7437567861020629, + "grad_norm": 8.699280554503527, + "learning_rate": 1.5129380929373076e-05, + "loss": 1.2891, + "step": 3212 + }, + { + "epoch": 1.744299674267101, + "grad_norm": 11.932192326740894, + "learning_rate": 1.5126361667237116e-05, + "loss": 0.9868, + "step": 3213 + }, + { + "epoch": 1.7448425624321389, + "grad_norm": 8.456694582402795, + "learning_rate": 1.512334177106457e-05, + "loss": 1.199, + "step": 3214 + }, + { + "epoch": 1.745385450597177, + "grad_norm": 8.284979161281417, + "learning_rate": 1.512032124122895e-05, + "loss": 0.658, + "step": 3215 + }, + { + "epoch": 1.7459283387622149, + "grad_norm": 8.490755188379255, + "learning_rate": 1.5117300078103841e-05, + "loss": 0.6974, + "step": 3216 + }, + { + "epoch": 1.746471226927253, + "grad_norm": 14.621603534181931, + "learning_rate": 1.5114278282062898e-05, + "loss": 1.4386, + "step": 3217 + }, + { + "epoch": 1.7470141150922909, + "grad_norm": 8.35853855400652, + "learning_rate": 1.511125585347987e-05, + "loss": 0.7991, + "step": 3218 + }, + { + "epoch": 1.747557003257329, + "grad_norm": 10.14192231306983, + "learning_rate": 1.5108232792728567e-05, + "loss": 0.9942, + "step": 3219 + }, + { + "epoch": 1.7480998914223669, + "grad_norm": 6.003113555943506, + "learning_rate": 1.5105209100182893e-05, + "loss": 0.5614, + "step": 3220 + }, + { + "epoch": 1.748642779587405, + "grad_norm": 7.420823995167257, + "learning_rate": 1.5102184776216824e-05, + "loss": 0.6918, + "step": 3221 + }, + { + "epoch": 1.7491856677524429, + "grad_norm": 7.3724702610256, + "learning_rate": 1.5099159821204406e-05, + "loss": 0.815, + "step": 3222 + }, + { + "epoch": 1.749728555917481, + "grad_norm": 8.539056209113738, + "learning_rate": 1.509613423551978e-05, + "loss": 1.0859, + "step": 3223 + }, + { + "epoch": 1.750271444082519, + "grad_norm": 10.518635534148203, + "learning_rate": 1.5093108019537147e-05, + "loss": 1.1873, + "step": 3224 + }, + { + "epoch": 1.7508143322475571, + "grad_norm": 8.049010262043616, + "learning_rate": 1.5090081173630795e-05, + "loss": 0.971, + "step": 3225 + }, + { + "epoch": 1.751357220412595, + "grad_norm": 7.874724964543097, + "learning_rate": 1.5087053698175092e-05, + "loss": 1.06, + "step": 3226 + }, + { + "epoch": 1.7519001085776331, + "grad_norm": 8.018813791724389, + "learning_rate": 1.508402559354448e-05, + "loss": 0.7192, + "step": 3227 + }, + { + "epoch": 1.752442996742671, + "grad_norm": 11.748757549078345, + "learning_rate": 1.508099686011348e-05, + "loss": 1.4476, + "step": 3228 + }, + { + "epoch": 1.7529858849077091, + "grad_norm": 7.885344513181394, + "learning_rate": 1.507796749825669e-05, + "loss": 0.7493, + "step": 3229 + }, + { + "epoch": 1.753528773072747, + "grad_norm": 6.808213775155858, + "learning_rate": 1.507493750834879e-05, + "loss": 0.6646, + "step": 3230 + }, + { + "epoch": 1.7540716612377851, + "grad_norm": 6.304489454276083, + "learning_rate": 1.5071906890764527e-05, + "loss": 0.7375, + "step": 3231 + }, + { + "epoch": 1.754614549402823, + "grad_norm": 8.217636823407403, + "learning_rate": 1.5068875645878739e-05, + "loss": 1.0069, + "step": 3232 + }, + { + "epoch": 1.7551574375678611, + "grad_norm": 10.706707510999644, + "learning_rate": 1.5065843774066329e-05, + "loss": 0.689, + "step": 3233 + }, + { + "epoch": 1.755700325732899, + "grad_norm": 7.600175930285094, + "learning_rate": 1.5062811275702291e-05, + "loss": 0.8645, + "step": 3234 + }, + { + "epoch": 1.7562432138979371, + "grad_norm": 8.463001355349205, + "learning_rate": 1.5059778151161684e-05, + "loss": 0.7185, + "step": 3235 + }, + { + "epoch": 1.756786102062975, + "grad_norm": 5.834156873881953, + "learning_rate": 1.5056744400819651e-05, + "loss": 0.6486, + "step": 3236 + }, + { + "epoch": 1.7573289902280131, + "grad_norm": 7.152665898997485, + "learning_rate": 1.5053710025051411e-05, + "loss": 0.9116, + "step": 3237 + }, + { + "epoch": 1.757871878393051, + "grad_norm": 8.037482464894001, + "learning_rate": 1.5050675024232262e-05, + "loss": 0.8231, + "step": 3238 + }, + { + "epoch": 1.7584147665580891, + "grad_norm": 10.416359782719013, + "learning_rate": 1.5047639398737573e-05, + "loss": 1.2555, + "step": 3239 + }, + { + "epoch": 1.758957654723127, + "grad_norm": 6.789336798229921, + "learning_rate": 1.50446031489428e-05, + "loss": 0.7295, + "step": 3240 + }, + { + "epoch": 1.7595005428881652, + "grad_norm": 9.927450801451293, + "learning_rate": 1.5041566275223472e-05, + "loss": 1.5417, + "step": 3241 + }, + { + "epoch": 1.760043431053203, + "grad_norm": 7.040644062299268, + "learning_rate": 1.5038528777955188e-05, + "loss": 0.4495, + "step": 3242 + }, + { + "epoch": 1.7605863192182412, + "grad_norm": 8.848733486454524, + "learning_rate": 1.5035490657513638e-05, + "loss": 1.0123, + "step": 3243 + }, + { + "epoch": 1.761129207383279, + "grad_norm": 8.109736914130819, + "learning_rate": 1.503245191427458e-05, + "loss": 0.8984, + "step": 3244 + }, + { + "epoch": 1.7616720955483172, + "grad_norm": 7.210785707597705, + "learning_rate": 1.5029412548613845e-05, + "loss": 0.6947, + "step": 3245 + }, + { + "epoch": 1.762214983713355, + "grad_norm": 10.368251730052483, + "learning_rate": 1.502637256090735e-05, + "loss": 0.8413, + "step": 3246 + }, + { + "epoch": 1.7627578718783932, + "grad_norm": 6.770955437649876, + "learning_rate": 1.5023331951531086e-05, + "loss": 0.6386, + "step": 3247 + }, + { + "epoch": 1.763300760043431, + "grad_norm": 10.818384833705863, + "learning_rate": 1.5020290720861122e-05, + "loss": 1.2544, + "step": 3248 + }, + { + "epoch": 1.7638436482084692, + "grad_norm": 7.9104385792951195, + "learning_rate": 1.50172488692736e-05, + "loss": 0.6538, + "step": 3249 + }, + { + "epoch": 1.764386536373507, + "grad_norm": 9.565689306609565, + "learning_rate": 1.5014206397144742e-05, + "loss": 1.1232, + "step": 3250 + }, + { + "epoch": 1.7649294245385452, + "grad_norm": 11.651444088862487, + "learning_rate": 1.5011163304850844e-05, + "loss": 1.2054, + "step": 3251 + }, + { + "epoch": 1.765472312703583, + "grad_norm": 10.326063989708704, + "learning_rate": 1.500811959276828e-05, + "loss": 0.9441, + "step": 3252 + }, + { + "epoch": 1.7660152008686212, + "grad_norm": 8.0529225996154, + "learning_rate": 1.5005075261273506e-05, + "loss": 0.9431, + "step": 3253 + }, + { + "epoch": 1.766558089033659, + "grad_norm": 7.391183931496241, + "learning_rate": 1.500203031074305e-05, + "loss": 0.8619, + "step": 3254 + }, + { + "epoch": 1.7671009771986972, + "grad_norm": 6.892445188811786, + "learning_rate": 1.4998984741553508e-05, + "loss": 0.7039, + "step": 3255 + }, + { + "epoch": 1.767643865363735, + "grad_norm": 6.541771639151763, + "learning_rate": 1.4995938554081568e-05, + "loss": 0.6359, + "step": 3256 + }, + { + "epoch": 1.7681867535287732, + "grad_norm": 7.690176361123399, + "learning_rate": 1.4992891748703985e-05, + "loss": 0.8998, + "step": 3257 + }, + { + "epoch": 1.768729641693811, + "grad_norm": 6.497301879748706, + "learning_rate": 1.498984432579759e-05, + "loss": 0.5743, + "step": 3258 + }, + { + "epoch": 1.7692725298588492, + "grad_norm": 6.656723951695686, + "learning_rate": 1.4986796285739298e-05, + "loss": 0.7002, + "step": 3259 + }, + { + "epoch": 1.769815418023887, + "grad_norm": 7.9129397206513685, + "learning_rate": 1.4983747628906095e-05, + "loss": 0.9239, + "step": 3260 + }, + { + "epoch": 1.7703583061889252, + "grad_norm": 4.544544755639526, + "learning_rate": 1.4980698355675043e-05, + "loss": 0.3431, + "step": 3261 + }, + { + "epoch": 1.770901194353963, + "grad_norm": 8.243743590575315, + "learning_rate": 1.4977648466423278e-05, + "loss": 1.1386, + "step": 3262 + }, + { + "epoch": 1.7714440825190012, + "grad_norm": 9.404041237838763, + "learning_rate": 1.4974597961528021e-05, + "loss": 0.9087, + "step": 3263 + }, + { + "epoch": 1.771986970684039, + "grad_norm": 8.534670509056685, + "learning_rate": 1.4971546841366556e-05, + "loss": 0.8769, + "step": 3264 + }, + { + "epoch": 1.7725298588490772, + "grad_norm": 9.442139570892468, + "learning_rate": 1.4968495106316254e-05, + "loss": 0.6942, + "step": 3265 + }, + { + "epoch": 1.773072747014115, + "grad_norm": 8.449365139837813, + "learning_rate": 1.496544275675456e-05, + "loss": 1.0743, + "step": 3266 + }, + { + "epoch": 1.7736156351791532, + "grad_norm": 9.461545898435599, + "learning_rate": 1.4962389793058994e-05, + "loss": 0.7473, + "step": 3267 + }, + { + "epoch": 1.774158523344191, + "grad_norm": 11.871672085414737, + "learning_rate": 1.4959336215607147e-05, + "loss": 1.5433, + "step": 3268 + }, + { + "epoch": 1.7747014115092292, + "grad_norm": 6.417686676489462, + "learning_rate": 1.4956282024776693e-05, + "loss": 0.5006, + "step": 3269 + }, + { + "epoch": 1.775244299674267, + "grad_norm": 10.49161190534651, + "learning_rate": 1.4953227220945381e-05, + "loss": 0.9918, + "step": 3270 + }, + { + "epoch": 1.7757871878393052, + "grad_norm": 9.298743913365943, + "learning_rate": 1.4950171804491031e-05, + "loss": 1.3149, + "step": 3271 + }, + { + "epoch": 1.776330076004343, + "grad_norm": 7.775473503246929, + "learning_rate": 1.4947115775791541e-05, + "loss": 0.6873, + "step": 3272 + }, + { + "epoch": 1.7768729641693812, + "grad_norm": 6.75591368497042, + "learning_rate": 1.4944059135224891e-05, + "loss": 0.7799, + "step": 3273 + }, + { + "epoch": 1.777415852334419, + "grad_norm": 8.234180005996857, + "learning_rate": 1.4941001883169124e-05, + "loss": 1.0393, + "step": 3274 + }, + { + "epoch": 1.7779587404994572, + "grad_norm": 8.419005181659063, + "learning_rate": 1.4937944020002371e-05, + "loss": 0.6798, + "step": 3275 + }, + { + "epoch": 1.778501628664495, + "grad_norm": 7.0982600255490205, + "learning_rate": 1.4934885546102833e-05, + "loss": 0.4936, + "step": 3276 + }, + { + "epoch": 1.7790445168295332, + "grad_norm": 8.404626943584326, + "learning_rate": 1.4931826461848785e-05, + "loss": 1.0507, + "step": 3277 + }, + { + "epoch": 1.779587404994571, + "grad_norm": 5.861882830901275, + "learning_rate": 1.4928766767618576e-05, + "loss": 0.6785, + "step": 3278 + }, + { + "epoch": 1.7801302931596092, + "grad_norm": 8.114055102204162, + "learning_rate": 1.4925706463790642e-05, + "loss": 0.765, + "step": 3279 + }, + { + "epoch": 1.780673181324647, + "grad_norm": 10.033269670592238, + "learning_rate": 1.492264555074348e-05, + "loss": 0.9163, + "step": 3280 + }, + { + "epoch": 1.7812160694896852, + "grad_norm": 7.040755628619449, + "learning_rate": 1.4919584028855671e-05, + "loss": 0.5962, + "step": 3281 + }, + { + "epoch": 1.781758957654723, + "grad_norm": 12.093987616559675, + "learning_rate": 1.4916521898505872e-05, + "loss": 0.8999, + "step": 3282 + }, + { + "epoch": 1.7823018458197613, + "grad_norm": 9.400200287574112, + "learning_rate": 1.4913459160072805e-05, + "loss": 1.0203, + "step": 3283 + }, + { + "epoch": 1.782844733984799, + "grad_norm": 8.605701710844945, + "learning_rate": 1.4910395813935279e-05, + "loss": 0.7874, + "step": 3284 + }, + { + "epoch": 1.7833876221498373, + "grad_norm": 10.279849472300688, + "learning_rate": 1.4907331860472174e-05, + "loss": 1.4306, + "step": 3285 + }, + { + "epoch": 1.783930510314875, + "grad_norm": 10.582881297295359, + "learning_rate": 1.4904267300062443e-05, + "loss": 0.6536, + "step": 3286 + }, + { + "epoch": 1.7844733984799133, + "grad_norm": 6.960885594125151, + "learning_rate": 1.4901202133085115e-05, + "loss": 0.7256, + "step": 3287 + }, + { + "epoch": 1.785016286644951, + "grad_norm": 9.088589913670967, + "learning_rate": 1.4898136359919298e-05, + "loss": 0.9299, + "step": 3288 + }, + { + "epoch": 1.7855591748099893, + "grad_norm": 8.922355254191798, + "learning_rate": 1.4895069980944168e-05, + "loss": 0.8159, + "step": 3289 + }, + { + "epoch": 1.786102062975027, + "grad_norm": 9.459478678151763, + "learning_rate": 1.4892002996538982e-05, + "loss": 1.132, + "step": 3290 + }, + { + "epoch": 1.7866449511400653, + "grad_norm": 9.746458814865415, + "learning_rate": 1.4888935407083065e-05, + "loss": 0.9874, + "step": 3291 + }, + { + "epoch": 1.787187839305103, + "grad_norm": 8.13065950431162, + "learning_rate": 1.4885867212955827e-05, + "loss": 0.8013, + "step": 3292 + }, + { + "epoch": 1.7877307274701413, + "grad_norm": 8.469962208509646, + "learning_rate": 1.4882798414536749e-05, + "loss": 1.0408, + "step": 3293 + }, + { + "epoch": 1.788273615635179, + "grad_norm": 9.018264603768202, + "learning_rate": 1.4879729012205379e-05, + "loss": 0.8324, + "step": 3294 + }, + { + "epoch": 1.7888165038002173, + "grad_norm": 8.934092671110381, + "learning_rate": 1.4876659006341347e-05, + "loss": 0.8842, + "step": 3295 + }, + { + "epoch": 1.789359391965255, + "grad_norm": 7.612464888658943, + "learning_rate": 1.4873588397324356e-05, + "loss": 0.9356, + "step": 3296 + }, + { + "epoch": 1.7899022801302933, + "grad_norm": 7.303452266512586, + "learning_rate": 1.4870517185534184e-05, + "loss": 1.0291, + "step": 3297 + }, + { + "epoch": 1.790445168295331, + "grad_norm": 9.805172223903986, + "learning_rate": 1.4867445371350687e-05, + "loss": 1.0243, + "step": 3298 + }, + { + "epoch": 1.7909880564603693, + "grad_norm": 7.742224062288237, + "learning_rate": 1.4864372955153785e-05, + "loss": 0.7173, + "step": 3299 + }, + { + "epoch": 1.791530944625407, + "grad_norm": 9.174714938761847, + "learning_rate": 1.4861299937323485e-05, + "loss": 0.9721, + "step": 3300 + }, + { + "epoch": 1.7920738327904453, + "grad_norm": 8.116980821679304, + "learning_rate": 1.4858226318239864e-05, + "loss": 0.7776, + "step": 3301 + }, + { + "epoch": 1.792616720955483, + "grad_norm": 5.885386861657401, + "learning_rate": 1.4855152098283068e-05, + "loss": 0.3833, + "step": 3302 + }, + { + "epoch": 1.7931596091205213, + "grad_norm": 6.804177263707199, + "learning_rate": 1.4852077277833319e-05, + "loss": 0.7972, + "step": 3303 + }, + { + "epoch": 1.793702497285559, + "grad_norm": 8.874804659745028, + "learning_rate": 1.4849001857270922e-05, + "loss": 1.063, + "step": 3304 + }, + { + "epoch": 1.7942453854505973, + "grad_norm": 10.231899453600786, + "learning_rate": 1.4845925836976247e-05, + "loss": 1.6617, + "step": 3305 + }, + { + "epoch": 1.794788273615635, + "grad_norm": 9.067079935451718, + "learning_rate": 1.4842849217329742e-05, + "loss": 1.1062, + "step": 3306 + }, + { + "epoch": 1.7953311617806733, + "grad_norm": 9.5261315220339, + "learning_rate": 1.4839771998711925e-05, + "loss": 1.5991, + "step": 3307 + }, + { + "epoch": 1.795874049945711, + "grad_norm": 9.53442504374803, + "learning_rate": 1.4836694181503397e-05, + "loss": 0.9267, + "step": 3308 + }, + { + "epoch": 1.7964169381107493, + "grad_norm": 8.231497735162955, + "learning_rate": 1.4833615766084821e-05, + "loss": 0.9802, + "step": 3309 + }, + { + "epoch": 1.796959826275787, + "grad_norm": 7.2171054452276815, + "learning_rate": 1.4830536752836944e-05, + "loss": 0.6116, + "step": 3310 + }, + { + "epoch": 1.7975027144408253, + "grad_norm": 8.412637332813022, + "learning_rate": 1.4827457142140583e-05, + "loss": 0.7482, + "step": 3311 + }, + { + "epoch": 1.798045602605863, + "grad_norm": 6.24239912297862, + "learning_rate": 1.482437693437663e-05, + "loss": 0.4333, + "step": 3312 + }, + { + "epoch": 1.7985884907709013, + "grad_norm": 9.706212881828902, + "learning_rate": 1.4821296129926049e-05, + "loss": 1.3862, + "step": 3313 + }, + { + "epoch": 1.799131378935939, + "grad_norm": 8.275124350006633, + "learning_rate": 1.4818214729169878e-05, + "loss": 0.812, + "step": 3314 + }, + { + "epoch": 1.7996742671009773, + "grad_norm": 7.0810692617964275, + "learning_rate": 1.4815132732489228e-05, + "loss": 0.8919, + "step": 3315 + }, + { + "epoch": 1.8002171552660151, + "grad_norm": 7.225811411306636, + "learning_rate": 1.4812050140265288e-05, + "loss": 0.7512, + "step": 3316 + }, + { + "epoch": 1.8007600434310533, + "grad_norm": 6.458874811050634, + "learning_rate": 1.4808966952879318e-05, + "loss": 0.6914, + "step": 3317 + }, + { + "epoch": 1.8013029315960911, + "grad_norm": 6.537680447766501, + "learning_rate": 1.4805883170712653e-05, + "loss": 0.5511, + "step": 3318 + }, + { + "epoch": 1.8018458197611293, + "grad_norm": 8.023777979063889, + "learning_rate": 1.4802798794146695e-05, + "loss": 0.6814, + "step": 3319 + }, + { + "epoch": 1.8023887079261671, + "grad_norm": 13.062126505855263, + "learning_rate": 1.4799713823562932e-05, + "loss": 1.0526, + "step": 3320 + }, + { + "epoch": 1.8029315960912053, + "grad_norm": 8.173027381817493, + "learning_rate": 1.479662825934291e-05, + "loss": 0.8939, + "step": 3321 + }, + { + "epoch": 1.8034744842562431, + "grad_norm": 7.465360921787222, + "learning_rate": 1.4793542101868261e-05, + "loss": 0.5965, + "step": 3322 + }, + { + "epoch": 1.8040173724212814, + "grad_norm": 6.5098081549027835, + "learning_rate": 1.4790455351520684e-05, + "loss": 0.5963, + "step": 3323 + }, + { + "epoch": 1.8045602605863191, + "grad_norm": 11.071700718231854, + "learning_rate": 1.4787368008681956e-05, + "loss": 0.9728, + "step": 3324 + }, + { + "epoch": 1.8051031487513574, + "grad_norm": 9.251168051579464, + "learning_rate": 1.4784280073733924e-05, + "loss": 0.8816, + "step": 3325 + }, + { + "epoch": 1.8056460369163951, + "grad_norm": 7.876824650963893, + "learning_rate": 1.4781191547058505e-05, + "loss": 1.0365, + "step": 3326 + }, + { + "epoch": 1.8061889250814334, + "grad_norm": 7.984254709859244, + "learning_rate": 1.4778102429037696e-05, + "loss": 0.7252, + "step": 3327 + }, + { + "epoch": 1.8067318132464711, + "grad_norm": 8.162664097932598, + "learning_rate": 1.4775012720053563e-05, + "loss": 0.8297, + "step": 3328 + }, + { + "epoch": 1.8072747014115094, + "grad_norm": 6.631733087479191, + "learning_rate": 1.4771922420488245e-05, + "loss": 0.6962, + "step": 3329 + }, + { + "epoch": 1.8078175895765471, + "grad_norm": 8.80250228426254, + "learning_rate": 1.476883153072396e-05, + "loss": 1.0576, + "step": 3330 + }, + { + "epoch": 1.8083604777415854, + "grad_norm": 10.637063152950418, + "learning_rate": 1.476574005114299e-05, + "loss": 0.8866, + "step": 3331 + }, + { + "epoch": 1.8089033659066232, + "grad_norm": 9.58332512337787, + "learning_rate": 1.4762647982127696e-05, + "loss": 1.6148, + "step": 3332 + }, + { + "epoch": 1.8094462540716614, + "grad_norm": 7.826567769007031, + "learning_rate": 1.4759555324060508e-05, + "loss": 0.9256, + "step": 3333 + }, + { + "epoch": 1.8099891422366992, + "grad_norm": 9.445307614493151, + "learning_rate": 1.4756462077323931e-05, + "loss": 1.1389, + "step": 3334 + }, + { + "epoch": 1.8105320304017374, + "grad_norm": 7.504505604214132, + "learning_rate": 1.4753368242300543e-05, + "loss": 1.1333, + "step": 3335 + }, + { + "epoch": 1.8110749185667752, + "grad_norm": 6.996150668976886, + "learning_rate": 1.4750273819372996e-05, + "loss": 0.6721, + "step": 3336 + }, + { + "epoch": 1.8116178067318134, + "grad_norm": 7.593185686378431, + "learning_rate": 1.4747178808924012e-05, + "loss": 0.6856, + "step": 3337 + }, + { + "epoch": 1.8121606948968512, + "grad_norm": 7.277948846492979, + "learning_rate": 1.474408321133639e-05, + "loss": 0.5487, + "step": 3338 + }, + { + "epoch": 1.8127035830618894, + "grad_norm": 9.500801329454344, + "learning_rate": 1.474098702699299e-05, + "loss": 1.063, + "step": 3339 + }, + { + "epoch": 1.8132464712269272, + "grad_norm": 7.732920511283457, + "learning_rate": 1.4737890256276763e-05, + "loss": 0.8875, + "step": 3340 + }, + { + "epoch": 1.8137893593919654, + "grad_norm": 9.56546309643246, + "learning_rate": 1.4734792899570716e-05, + "loss": 0.7724, + "step": 3341 + }, + { + "epoch": 1.8143322475570032, + "grad_norm": 7.735046028758256, + "learning_rate": 1.4731694957257938e-05, + "loss": 0.8498, + "step": 3342 + }, + { + "epoch": 1.8148751357220414, + "grad_norm": 9.077413763001829, + "learning_rate": 1.4728596429721587e-05, + "loss": 1.1604, + "step": 3343 + }, + { + "epoch": 1.8154180238870792, + "grad_norm": 8.33530084886159, + "learning_rate": 1.4725497317344892e-05, + "loss": 1.0126, + "step": 3344 + }, + { + "epoch": 1.8159609120521174, + "grad_norm": 11.938749223535789, + "learning_rate": 1.4722397620511158e-05, + "loss": 0.9538, + "step": 3345 + }, + { + "epoch": 1.8165038002171552, + "grad_norm": 4.946988727080576, + "learning_rate": 1.471929733960376e-05, + "loss": 0.3007, + "step": 3346 + }, + { + "epoch": 1.8170466883821934, + "grad_norm": 6.934881579248824, + "learning_rate": 1.4716196475006145e-05, + "loss": 0.5286, + "step": 3347 + }, + { + "epoch": 1.8175895765472312, + "grad_norm": 12.504257989709076, + "learning_rate": 1.4713095027101833e-05, + "loss": 0.9948, + "step": 3348 + }, + { + "epoch": 1.8181324647122694, + "grad_norm": 6.299717892798866, + "learning_rate": 1.4709992996274416e-05, + "loss": 0.6503, + "step": 3349 + }, + { + "epoch": 1.8186753528773072, + "grad_norm": 8.48347618555589, + "learning_rate": 1.4706890382907562e-05, + "loss": 0.7195, + "step": 3350 + }, + { + "epoch": 1.8192182410423454, + "grad_norm": 7.992565784639017, + "learning_rate": 1.4703787187385002e-05, + "loss": 0.7349, + "step": 3351 + }, + { + "epoch": 1.8197611292073832, + "grad_norm": 7.567197193038979, + "learning_rate": 1.4700683410090546e-05, + "loss": 0.4719, + "step": 3352 + }, + { + "epoch": 1.8203040173724214, + "grad_norm": 7.276389893333727, + "learning_rate": 1.4697579051408073e-05, + "loss": 0.93, + "step": 3353 + }, + { + "epoch": 1.8208469055374592, + "grad_norm": 8.13251629417801, + "learning_rate": 1.4694474111721537e-05, + "loss": 0.9865, + "step": 3354 + }, + { + "epoch": 1.8213897937024974, + "grad_norm": 6.788010338042944, + "learning_rate": 1.469136859141496e-05, + "loss": 0.5671, + "step": 3355 + }, + { + "epoch": 1.8219326818675352, + "grad_norm": 6.83194614525194, + "learning_rate": 1.4688262490872438e-05, + "loss": 0.5486, + "step": 3356 + }, + { + "epoch": 1.8224755700325734, + "grad_norm": 12.201391760355081, + "learning_rate": 1.4685155810478143e-05, + "loss": 0.9178, + "step": 3357 + }, + { + "epoch": 1.8230184581976112, + "grad_norm": 7.955977866907339, + "learning_rate": 1.4682048550616306e-05, + "loss": 0.9671, + "step": 3358 + }, + { + "epoch": 1.8235613463626494, + "grad_norm": 9.699881747283202, + "learning_rate": 1.4678940711671245e-05, + "loss": 0.7699, + "step": 3359 + }, + { + "epoch": 1.8241042345276872, + "grad_norm": 7.703855587772515, + "learning_rate": 1.4675832294027339e-05, + "loss": 0.6471, + "step": 3360 + }, + { + "epoch": 1.8246471226927254, + "grad_norm": 11.230217201340418, + "learning_rate": 1.4672723298069043e-05, + "loss": 1.1256, + "step": 3361 + }, + { + "epoch": 1.8251900108577632, + "grad_norm": 11.964156332533788, + "learning_rate": 1.4669613724180883e-05, + "loss": 1.1829, + "step": 3362 + }, + { + "epoch": 1.8257328990228014, + "grad_norm": 8.221701868139569, + "learning_rate": 1.4666503572747452e-05, + "loss": 0.6004, + "step": 3363 + }, + { + "epoch": 1.8262757871878392, + "grad_norm": 6.179895131215982, + "learning_rate": 1.4663392844153426e-05, + "loss": 0.6923, + "step": 3364 + }, + { + "epoch": 1.8268186753528775, + "grad_norm": 8.333959237389427, + "learning_rate": 1.4660281538783538e-05, + "loss": 0.8785, + "step": 3365 + }, + { + "epoch": 1.8273615635179152, + "grad_norm": 9.742872979143373, + "learning_rate": 1.4657169657022603e-05, + "loss": 1.0129, + "step": 3366 + }, + { + "epoch": 1.8279044516829535, + "grad_norm": 8.019435507225802, + "learning_rate": 1.46540571992555e-05, + "loss": 1.0615, + "step": 3367 + }, + { + "epoch": 1.8284473398479912, + "grad_norm": 5.872559756186675, + "learning_rate": 1.4650944165867187e-05, + "loss": 0.6391, + "step": 3368 + }, + { + "epoch": 1.8289902280130295, + "grad_norm": 7.878073300625923, + "learning_rate": 1.4647830557242685e-05, + "loss": 1.0317, + "step": 3369 + }, + { + "epoch": 1.8295331161780672, + "grad_norm": 10.290400619980408, + "learning_rate": 1.4644716373767096e-05, + "loss": 1.3108, + "step": 3370 + }, + { + "epoch": 1.8300760043431055, + "grad_norm": 7.655417729944032, + "learning_rate": 1.4641601615825582e-05, + "loss": 0.8434, + "step": 3371 + }, + { + "epoch": 1.8306188925081432, + "grad_norm": 8.894875963607664, + "learning_rate": 1.4638486283803384e-05, + "loss": 1.4973, + "step": 3372 + }, + { + "epoch": 1.8311617806731815, + "grad_norm": 7.70346849256546, + "learning_rate": 1.4635370378085809e-05, + "loss": 0.9202, + "step": 3373 + }, + { + "epoch": 1.8317046688382193, + "grad_norm": 7.428334777443758, + "learning_rate": 1.4632253899058233e-05, + "loss": 1.005, + "step": 3374 + }, + { + "epoch": 1.8322475570032575, + "grad_norm": 6.905061230055496, + "learning_rate": 1.4629136847106118e-05, + "loss": 0.7045, + "step": 3375 + }, + { + "epoch": 1.8327904451682953, + "grad_norm": 9.092059147160219, + "learning_rate": 1.4626019222614977e-05, + "loss": 1.1478, + "step": 3376 + }, + { + "epoch": 1.8333333333333335, + "grad_norm": 8.171417716573657, + "learning_rate": 1.4622901025970406e-05, + "loss": 0.8098, + "step": 3377 + }, + { + "epoch": 1.8338762214983713, + "grad_norm": 8.935159598593566, + "learning_rate": 1.461978225755807e-05, + "loss": 0.9121, + "step": 3378 + }, + { + "epoch": 1.8344191096634095, + "grad_norm": 9.059802045574934, + "learning_rate": 1.46166629177637e-05, + "loss": 0.952, + "step": 3379 + }, + { + "epoch": 1.8349619978284473, + "grad_norm": 6.404641837192718, + "learning_rate": 1.4613543006973103e-05, + "loss": 0.7178, + "step": 3380 + }, + { + "epoch": 1.8355048859934855, + "grad_norm": 7.714640842582964, + "learning_rate": 1.4610422525572155e-05, + "loss": 0.7053, + "step": 3381 + }, + { + "epoch": 1.8360477741585233, + "grad_norm": 9.232659815202982, + "learning_rate": 1.46073014739468e-05, + "loss": 0.7707, + "step": 3382 + }, + { + "epoch": 1.8365906623235615, + "grad_norm": 7.597273066110282, + "learning_rate": 1.4604179852483056e-05, + "loss": 0.9593, + "step": 3383 + }, + { + "epoch": 1.8371335504885993, + "grad_norm": 7.246167686662127, + "learning_rate": 1.460105766156701e-05, + "loss": 0.8489, + "step": 3384 + }, + { + "epoch": 1.8376764386536375, + "grad_norm": 9.724005322694314, + "learning_rate": 1.4597934901584818e-05, + "loss": 1.4838, + "step": 3385 + }, + { + "epoch": 1.8382193268186753, + "grad_norm": 5.623806840885622, + "learning_rate": 1.459481157292271e-05, + "loss": 0.4218, + "step": 3386 + }, + { + "epoch": 1.8387622149837135, + "grad_norm": 10.505590233471066, + "learning_rate": 1.4591687675966984e-05, + "loss": 1.3634, + "step": 3387 + }, + { + "epoch": 1.8393051031487513, + "grad_norm": 7.225825527188378, + "learning_rate": 1.458856321110401e-05, + "loss": 1.0878, + "step": 3388 + }, + { + "epoch": 1.8398479913137895, + "grad_norm": 6.92346369662784, + "learning_rate": 1.4585438178720221e-05, + "loss": 0.8579, + "step": 3389 + }, + { + "epoch": 1.8403908794788273, + "grad_norm": 6.78903317106508, + "learning_rate": 1.4582312579202134e-05, + "loss": 0.9986, + "step": 3390 + }, + { + "epoch": 1.8409337676438655, + "grad_norm": 7.493115074862596, + "learning_rate": 1.4579186412936323e-05, + "loss": 0.6403, + "step": 3391 + }, + { + "epoch": 1.8414766558089033, + "grad_norm": 7.292140639462925, + "learning_rate": 1.457605968030944e-05, + "loss": 0.7705, + "step": 3392 + }, + { + "epoch": 1.8420195439739415, + "grad_norm": 8.413656799764597, + "learning_rate": 1.4572932381708198e-05, + "loss": 0.7362, + "step": 3393 + }, + { + "epoch": 1.8425624321389793, + "grad_norm": 8.092404119538978, + "learning_rate": 1.4569804517519392e-05, + "loss": 0.8419, + "step": 3394 + }, + { + "epoch": 1.8431053203040175, + "grad_norm": 8.746713556642208, + "learning_rate": 1.456667608812988e-05, + "loss": 0.8992, + "step": 3395 + }, + { + "epoch": 1.8436482084690553, + "grad_norm": 9.315011907817567, + "learning_rate": 1.4563547093926595e-05, + "loss": 1.1811, + "step": 3396 + }, + { + "epoch": 1.8441910966340935, + "grad_norm": 7.921580843874425, + "learning_rate": 1.4560417535296529e-05, + "loss": 0.9896, + "step": 3397 + }, + { + "epoch": 1.8447339847991313, + "grad_norm": 5.324625495868631, + "learning_rate": 1.4557287412626755e-05, + "loss": 0.5757, + "step": 3398 + }, + { + "epoch": 1.8452768729641695, + "grad_norm": 7.061326556489923, + "learning_rate": 1.4554156726304411e-05, + "loss": 0.7016, + "step": 3399 + }, + { + "epoch": 1.8458197611292073, + "grad_norm": 7.833004564597817, + "learning_rate": 1.4551025476716704e-05, + "loss": 0.6112, + "step": 3400 + }, + { + "epoch": 1.8463626492942455, + "grad_norm": 6.762019290951184, + "learning_rate": 1.4547893664250912e-05, + "loss": 0.6197, + "step": 3401 + }, + { + "epoch": 1.8469055374592833, + "grad_norm": 8.02135944193365, + "learning_rate": 1.4544761289294384e-05, + "loss": 0.7665, + "step": 3402 + }, + { + "epoch": 1.8474484256243215, + "grad_norm": 9.145981041987138, + "learning_rate": 1.4541628352234538e-05, + "loss": 1.4491, + "step": 3403 + }, + { + "epoch": 1.8479913137893593, + "grad_norm": 10.127072377919164, + "learning_rate": 1.4538494853458858e-05, + "loss": 1.1929, + "step": 3404 + }, + { + "epoch": 1.8485342019543975, + "grad_norm": 9.041637256167727, + "learning_rate": 1.45353607933549e-05, + "loss": 0.8471, + "step": 3405 + }, + { + "epoch": 1.8490770901194353, + "grad_norm": 7.08038124220055, + "learning_rate": 1.453222617231029e-05, + "loss": 0.5152, + "step": 3406 + }, + { + "epoch": 1.8496199782844736, + "grad_norm": 8.460108981189723, + "learning_rate": 1.4529090990712726e-05, + "loss": 0.9495, + "step": 3407 + }, + { + "epoch": 1.8501628664495113, + "grad_norm": 8.939702203617836, + "learning_rate": 1.452595524894997e-05, + "loss": 1.0959, + "step": 3408 + }, + { + "epoch": 1.8507057546145496, + "grad_norm": 7.5616663713048595, + "learning_rate": 1.4522818947409855e-05, + "loss": 1.0391, + "step": 3409 + }, + { + "epoch": 1.8512486427795873, + "grad_norm": 8.75531450227851, + "learning_rate": 1.4519682086480287e-05, + "loss": 0.7298, + "step": 3410 + }, + { + "epoch": 1.8517915309446256, + "grad_norm": 7.003753183958329, + "learning_rate": 1.4516544666549233e-05, + "loss": 0.4788, + "step": 3411 + }, + { + "epoch": 1.8523344191096633, + "grad_norm": 10.952441700747523, + "learning_rate": 1.4513406688004734e-05, + "loss": 1.0903, + "step": 3412 + }, + { + "epoch": 1.8528773072747016, + "grad_norm": 8.61536651212902, + "learning_rate": 1.4510268151234903e-05, + "loss": 0.6924, + "step": 3413 + }, + { + "epoch": 1.8534201954397393, + "grad_norm": 9.237257682023538, + "learning_rate": 1.4507129056627922e-05, + "loss": 0.8686, + "step": 3414 + }, + { + "epoch": 1.8539630836047776, + "grad_norm": 7.792081331119547, + "learning_rate": 1.4503989404572031e-05, + "loss": 0.9818, + "step": 3415 + }, + { + "epoch": 1.8545059717698154, + "grad_norm": 8.622336937881878, + "learning_rate": 1.4500849195455557e-05, + "loss": 0.9444, + "step": 3416 + }, + { + "epoch": 1.8550488599348536, + "grad_norm": 7.607437519589793, + "learning_rate": 1.4497708429666882e-05, + "loss": 0.846, + "step": 3417 + }, + { + "epoch": 1.8555917480998914, + "grad_norm": 8.358779281223866, + "learning_rate": 1.4494567107594457e-05, + "loss": 0.7125, + "step": 3418 + }, + { + "epoch": 1.8561346362649296, + "grad_norm": 6.8176708905901044, + "learning_rate": 1.449142522962681e-05, + "loss": 0.8921, + "step": 3419 + }, + { + "epoch": 1.8566775244299674, + "grad_norm": 8.931750092418099, + "learning_rate": 1.448828279615253e-05, + "loss": 0.7989, + "step": 3420 + }, + { + "epoch": 1.8572204125950056, + "grad_norm": 7.481066786036915, + "learning_rate": 1.4485139807560281e-05, + "loss": 0.6642, + "step": 3421 + }, + { + "epoch": 1.8577633007600434, + "grad_norm": 7.460924029721227, + "learning_rate": 1.4481996264238796e-05, + "loss": 0.6509, + "step": 3422 + }, + { + "epoch": 1.8583061889250816, + "grad_norm": 7.224784834236484, + "learning_rate": 1.4478852166576869e-05, + "loss": 0.7244, + "step": 3423 + }, + { + "epoch": 1.8588490770901194, + "grad_norm": 5.912102755785491, + "learning_rate": 1.4475707514963367e-05, + "loss": 0.5713, + "step": 3424 + }, + { + "epoch": 1.8593919652551576, + "grad_norm": 6.994884495015334, + "learning_rate": 1.4472562309787224e-05, + "loss": 0.4875, + "step": 3425 + }, + { + "epoch": 1.8599348534201954, + "grad_norm": 6.38845062016643, + "learning_rate": 1.4469416551437445e-05, + "loss": 0.4708, + "step": 3426 + }, + { + "epoch": 1.8604777415852336, + "grad_norm": 11.250341421225603, + "learning_rate": 1.4466270240303109e-05, + "loss": 1.4338, + "step": 3427 + }, + { + "epoch": 1.8610206297502714, + "grad_norm": 9.828853213376544, + "learning_rate": 1.4463123376773348e-05, + "loss": 0.7307, + "step": 3428 + }, + { + "epoch": 1.8615635179153096, + "grad_norm": 7.96498248613748, + "learning_rate": 1.4459975961237375e-05, + "loss": 0.6727, + "step": 3429 + }, + { + "epoch": 1.8621064060803474, + "grad_norm": 9.736354023562699, + "learning_rate": 1.4456827994084473e-05, + "loss": 0.8091, + "step": 3430 + }, + { + "epoch": 1.8626492942453856, + "grad_norm": 10.317215804231997, + "learning_rate": 1.4453679475703974e-05, + "loss": 1.0997, + "step": 3431 + }, + { + "epoch": 1.8631921824104234, + "grad_norm": 10.00211430988322, + "learning_rate": 1.4450530406485301e-05, + "loss": 1.4007, + "step": 3432 + }, + { + "epoch": 1.8637350705754616, + "grad_norm": 10.665872978467544, + "learning_rate": 1.4447380786817934e-05, + "loss": 0.9221, + "step": 3433 + }, + { + "epoch": 1.8642779587404994, + "grad_norm": 8.8489121199226, + "learning_rate": 1.4444230617091424e-05, + "loss": 0.7745, + "step": 3434 + }, + { + "epoch": 1.8648208469055376, + "grad_norm": 7.083707820329958, + "learning_rate": 1.444107989769539e-05, + "loss": 0.7152, + "step": 3435 + }, + { + "epoch": 1.8653637350705754, + "grad_norm": 7.496522090143239, + "learning_rate": 1.4437928629019515e-05, + "loss": 0.5901, + "step": 3436 + }, + { + "epoch": 1.8659066232356136, + "grad_norm": 10.576310319834935, + "learning_rate": 1.4434776811453552e-05, + "loss": 0.7752, + "step": 3437 + }, + { + "epoch": 1.8664495114006514, + "grad_norm": 8.686730112665717, + "learning_rate": 1.4431624445387327e-05, + "loss": 0.8748, + "step": 3438 + }, + { + "epoch": 1.8669923995656896, + "grad_norm": 9.435491336462407, + "learning_rate": 1.4428471531210726e-05, + "loss": 1.1954, + "step": 3439 + }, + { + "epoch": 1.8675352877307274, + "grad_norm": 9.012159097359817, + "learning_rate": 1.4425318069313712e-05, + "loss": 0.7353, + "step": 3440 + }, + { + "epoch": 1.8680781758957656, + "grad_norm": 9.617409218513528, + "learning_rate": 1.4422164060086304e-05, + "loss": 0.6611, + "step": 3441 + }, + { + "epoch": 1.8686210640608034, + "grad_norm": 9.612516235298202, + "learning_rate": 1.4419009503918598e-05, + "loss": 0.8101, + "step": 3442 + }, + { + "epoch": 1.8691639522258416, + "grad_norm": 6.542519560792006, + "learning_rate": 1.4415854401200751e-05, + "loss": 0.473, + "step": 3443 + }, + { + "epoch": 1.8697068403908794, + "grad_norm": 10.696166776054064, + "learning_rate": 1.4412698752322998e-05, + "loss": 1.0001, + "step": 3444 + }, + { + "epoch": 1.8702497285559176, + "grad_norm": 10.087499378994824, + "learning_rate": 1.4409542557675625e-05, + "loss": 0.9323, + "step": 3445 + }, + { + "epoch": 1.8707926167209554, + "grad_norm": 7.829085764895678, + "learning_rate": 1.4406385817649002e-05, + "loss": 0.8624, + "step": 3446 + }, + { + "epoch": 1.8713355048859937, + "grad_norm": 11.554825371147222, + "learning_rate": 1.4403228532633561e-05, + "loss": 0.7735, + "step": 3447 + }, + { + "epoch": 1.8718783930510314, + "grad_norm": 9.43390979006707, + "learning_rate": 1.4400070703019797e-05, + "loss": 0.9128, + "step": 3448 + }, + { + "epoch": 1.8724212812160697, + "grad_norm": 11.851717520188132, + "learning_rate": 1.4396912329198276e-05, + "loss": 1.2379, + "step": 3449 + }, + { + "epoch": 1.8729641693811074, + "grad_norm": 8.408242915866566, + "learning_rate": 1.4393753411559632e-05, + "loss": 0.7549, + "step": 3450 + }, + { + "epoch": 1.8735070575461457, + "grad_norm": 8.144856314123112, + "learning_rate": 1.4390593950494557e-05, + "loss": 0.7281, + "step": 3451 + }, + { + "epoch": 1.8740499457111834, + "grad_norm": 8.202846690293642, + "learning_rate": 1.438743394639383e-05, + "loss": 0.6773, + "step": 3452 + }, + { + "epoch": 1.8745928338762217, + "grad_norm": 9.423562458268494, + "learning_rate": 1.438427339964828e-05, + "loss": 1.1, + "step": 3453 + }, + { + "epoch": 1.8751357220412594, + "grad_norm": 7.196483568403804, + "learning_rate": 1.4381112310648806e-05, + "loss": 0.5809, + "step": 3454 + }, + { + "epoch": 1.8756786102062974, + "grad_norm": 7.26770042201738, + "learning_rate": 1.4377950679786382e-05, + "loss": 0.8053, + "step": 3455 + }, + { + "epoch": 1.8762214983713354, + "grad_norm": 9.472969211327861, + "learning_rate": 1.4374788507452039e-05, + "loss": 0.7998, + "step": 3456 + }, + { + "epoch": 1.8767643865363735, + "grad_norm": 9.514618174868874, + "learning_rate": 1.437162579403688e-05, + "loss": 0.9572, + "step": 3457 + }, + { + "epoch": 1.8773072747014115, + "grad_norm": 9.363034597703797, + "learning_rate": 1.4368462539932073e-05, + "loss": 0.8528, + "step": 3458 + }, + { + "epoch": 1.8778501628664495, + "grad_norm": 7.113386807438533, + "learning_rate": 1.4365298745528863e-05, + "loss": 0.733, + "step": 3459 + }, + { + "epoch": 1.8783930510314875, + "grad_norm": 9.09654372206602, + "learning_rate": 1.4362134411218545e-05, + "loss": 0.7779, + "step": 3460 + }, + { + "epoch": 1.8789359391965255, + "grad_norm": 7.449242636613164, + "learning_rate": 1.4358969537392489e-05, + "loss": 0.899, + "step": 3461 + }, + { + "epoch": 1.8794788273615635, + "grad_norm": 10.755480560042127, + "learning_rate": 1.4355804124442133e-05, + "loss": 1.099, + "step": 3462 + }, + { + "epoch": 1.8800217155266015, + "grad_norm": 9.091218073930861, + "learning_rate": 1.4352638172758984e-05, + "loss": 0.7489, + "step": 3463 + }, + { + "epoch": 1.8805646036916395, + "grad_norm": 9.701260384390375, + "learning_rate": 1.4349471682734603e-05, + "loss": 0.8495, + "step": 3464 + }, + { + "epoch": 1.8811074918566775, + "grad_norm": 9.008809447502522, + "learning_rate": 1.4346304654760637e-05, + "loss": 0.8984, + "step": 3465 + }, + { + "epoch": 1.8816503800217155, + "grad_norm": 8.176404536031864, + "learning_rate": 1.4343137089228783e-05, + "loss": 0.8278, + "step": 3466 + }, + { + "epoch": 1.8821932681867535, + "grad_norm": 9.197307290475337, + "learning_rate": 1.4339968986530813e-05, + "loss": 0.7142, + "step": 3467 + }, + { + "epoch": 1.8827361563517915, + "grad_norm": 5.505228455971892, + "learning_rate": 1.433680034705856e-05, + "loss": 0.4204, + "step": 3468 + }, + { + "epoch": 1.8832790445168295, + "grad_norm": 9.014964705534542, + "learning_rate": 1.433363117120393e-05, + "loss": 0.8988, + "step": 3469 + }, + { + "epoch": 1.8838219326818675, + "grad_norm": 7.290687267145188, + "learning_rate": 1.4330461459358889e-05, + "loss": 0.7683, + "step": 3470 + }, + { + "epoch": 1.8843648208469055, + "grad_norm": 10.834879009670003, + "learning_rate": 1.4327291211915473e-05, + "loss": 1.0446, + "step": 3471 + }, + { + "epoch": 1.8849077090119435, + "grad_norm": 8.240452054600638, + "learning_rate": 1.4324120429265781e-05, + "loss": 0.893, + "step": 3472 + }, + { + "epoch": 1.8854505971769815, + "grad_norm": 8.17919491677486, + "learning_rate": 1.4320949111801987e-05, + "loss": 0.9846, + "step": 3473 + }, + { + "epoch": 1.8859934853420195, + "grad_norm": 7.693481880154859, + "learning_rate": 1.431777725991632e-05, + "loss": 0.6054, + "step": 3474 + }, + { + "epoch": 1.8865363735070575, + "grad_norm": 7.929110152467141, + "learning_rate": 1.4314604874001076e-05, + "loss": 0.7299, + "step": 3475 + }, + { + "epoch": 1.8870792616720955, + "grad_norm": 9.650029930461764, + "learning_rate": 1.4311431954448629e-05, + "loss": 1.0855, + "step": 3476 + }, + { + "epoch": 1.8876221498371335, + "grad_norm": 8.487868906333745, + "learning_rate": 1.4308258501651404e-05, + "loss": 0.7117, + "step": 3477 + }, + { + "epoch": 1.8881650380021715, + "grad_norm": 7.111157851383161, + "learning_rate": 1.43050845160019e-05, + "loss": 0.6477, + "step": 3478 + }, + { + "epoch": 1.8887079261672095, + "grad_norm": 8.2908796123436, + "learning_rate": 1.4301909997892684e-05, + "loss": 0.9819, + "step": 3479 + }, + { + "epoch": 1.8892508143322475, + "grad_norm": 10.322900044501562, + "learning_rate": 1.4298734947716385e-05, + "loss": 0.9368, + "step": 3480 + }, + { + "epoch": 1.8897937024972855, + "grad_norm": 6.902927389625658, + "learning_rate": 1.4295559365865694e-05, + "loss": 0.6201, + "step": 3481 + }, + { + "epoch": 1.8903365906623235, + "grad_norm": 8.909574060437844, + "learning_rate": 1.4292383252733375e-05, + "loss": 0.9393, + "step": 3482 + }, + { + "epoch": 1.8908794788273615, + "grad_norm": 7.643883869480783, + "learning_rate": 1.4289206608712252e-05, + "loss": 0.9349, + "step": 3483 + }, + { + "epoch": 1.8914223669923995, + "grad_norm": 7.103928943926094, + "learning_rate": 1.428602943419522e-05, + "loss": 0.6511, + "step": 3484 + }, + { + "epoch": 1.8919652551574375, + "grad_norm": 8.843983286039231, + "learning_rate": 1.428285172957524e-05, + "loss": 0.6958, + "step": 3485 + }, + { + "epoch": 1.8925081433224755, + "grad_norm": 5.044738479793456, + "learning_rate": 1.427967349524533e-05, + "loss": 0.59, + "step": 3486 + }, + { + "epoch": 1.8930510314875135, + "grad_norm": 6.9998626206641195, + "learning_rate": 1.4276494731598582e-05, + "loss": 0.5973, + "step": 3487 + }, + { + "epoch": 1.8935939196525515, + "grad_norm": 8.316471360458495, + "learning_rate": 1.4273315439028151e-05, + "loss": 0.8941, + "step": 3488 + }, + { + "epoch": 1.8941368078175895, + "grad_norm": 12.713883241844957, + "learning_rate": 1.4270135617927254e-05, + "loss": 1.4921, + "step": 3489 + }, + { + "epoch": 1.8946796959826275, + "grad_norm": 10.043253133675226, + "learning_rate": 1.426695526868918e-05, + "loss": 1.0242, + "step": 3490 + }, + { + "epoch": 1.8952225841476655, + "grad_norm": 11.142761788047633, + "learning_rate": 1.4263774391707274e-05, + "loss": 1.265, + "step": 3491 + }, + { + "epoch": 1.8957654723127035, + "grad_norm": 7.12450280478247, + "learning_rate": 1.4260592987374962e-05, + "loss": 0.484, + "step": 3492 + }, + { + "epoch": 1.8963083604777415, + "grad_norm": 7.017938163026924, + "learning_rate": 1.4257411056085712e-05, + "loss": 0.6549, + "step": 3493 + }, + { + "epoch": 1.8968512486427795, + "grad_norm": 7.643680244898419, + "learning_rate": 1.4254228598233082e-05, + "loss": 0.614, + "step": 3494 + }, + { + "epoch": 1.8973941368078175, + "grad_norm": 9.484247292353862, + "learning_rate": 1.4251045614210678e-05, + "loss": 0.9667, + "step": 3495 + }, + { + "epoch": 1.8979370249728555, + "grad_norm": 10.846455465412243, + "learning_rate": 1.4247862104412175e-05, + "loss": 1.4425, + "step": 3496 + }, + { + "epoch": 1.8984799131378935, + "grad_norm": 7.211983265810251, + "learning_rate": 1.4244678069231319e-05, + "loss": 0.7508, + "step": 3497 + }, + { + "epoch": 1.8990228013029316, + "grad_norm": 8.283019454355765, + "learning_rate": 1.4241493509061912e-05, + "loss": 0.9747, + "step": 3498 + }, + { + "epoch": 1.8995656894679696, + "grad_norm": 6.678762476220025, + "learning_rate": 1.4238308424297833e-05, + "loss": 0.4702, + "step": 3499 + }, + { + "epoch": 1.9001085776330076, + "grad_norm": 8.750314729983545, + "learning_rate": 1.423512281533301e-05, + "loss": 0.7636, + "step": 3500 + }, + { + "epoch": 1.9006514657980456, + "grad_norm": 7.304249105438763, + "learning_rate": 1.4231936682561446e-05, + "loss": 0.7198, + "step": 3501 + }, + { + "epoch": 1.9011943539630836, + "grad_norm": 7.676960806676722, + "learning_rate": 1.4228750026377212e-05, + "loss": 0.8639, + "step": 3502 + }, + { + "epoch": 1.9017372421281216, + "grad_norm": 10.327339388528987, + "learning_rate": 1.4225562847174431e-05, + "loss": 0.9386, + "step": 3503 + }, + { + "epoch": 1.9022801302931596, + "grad_norm": 8.828430779713369, + "learning_rate": 1.4222375145347304e-05, + "loss": 1.0821, + "step": 3504 + }, + { + "epoch": 1.9028230184581976, + "grad_norm": 8.099300889556567, + "learning_rate": 1.421918692129009e-05, + "loss": 0.6251, + "step": 3505 + }, + { + "epoch": 1.9033659066232356, + "grad_norm": 9.197832742387652, + "learning_rate": 1.4215998175397115e-05, + "loss": 1.1794, + "step": 3506 + }, + { + "epoch": 1.9039087947882736, + "grad_norm": 9.305579433655693, + "learning_rate": 1.4212808908062763e-05, + "loss": 1.3447, + "step": 3507 + }, + { + "epoch": 1.9044516829533116, + "grad_norm": 6.897677914233707, + "learning_rate": 1.4209619119681497e-05, + "loss": 0.5178, + "step": 3508 + }, + { + "epoch": 1.9049945711183496, + "grad_norm": 9.359746343598495, + "learning_rate": 1.420642881064782e-05, + "loss": 0.985, + "step": 3509 + }, + { + "epoch": 1.9055374592833876, + "grad_norm": 5.855418324834113, + "learning_rate": 1.4203237981356331e-05, + "loss": 0.5144, + "step": 3510 + }, + { + "epoch": 1.9060803474484256, + "grad_norm": 7.408991833255532, + "learning_rate": 1.4200046632201665e-05, + "loss": 0.7358, + "step": 3511 + }, + { + "epoch": 1.9066232356134636, + "grad_norm": 9.024544122531433, + "learning_rate": 1.419685476357854e-05, + "loss": 0.778, + "step": 3512 + }, + { + "epoch": 1.9071661237785016, + "grad_norm": 7.874409137592207, + "learning_rate": 1.419366237588173e-05, + "loss": 0.8703, + "step": 3513 + }, + { + "epoch": 1.9077090119435396, + "grad_norm": 8.221027989379657, + "learning_rate": 1.4190469469506073e-05, + "loss": 0.7436, + "step": 3514 + }, + { + "epoch": 1.9082519001085776, + "grad_norm": 10.797659549408706, + "learning_rate": 1.4187276044846473e-05, + "loss": 1.0635, + "step": 3515 + }, + { + "epoch": 1.9087947882736156, + "grad_norm": 9.718910477411724, + "learning_rate": 1.4184082102297896e-05, + "loss": 0.8802, + "step": 3516 + }, + { + "epoch": 1.9093376764386536, + "grad_norm": 7.365463419520298, + "learning_rate": 1.4180887642255376e-05, + "loss": 0.5919, + "step": 3517 + }, + { + "epoch": 1.9098805646036916, + "grad_norm": 6.487376521266865, + "learning_rate": 1.4177692665114014e-05, + "loss": 0.7098, + "step": 3518 + }, + { + "epoch": 1.9104234527687296, + "grad_norm": 8.328541010595174, + "learning_rate": 1.4174497171268962e-05, + "loss": 0.6124, + "step": 3519 + }, + { + "epoch": 1.9109663409337676, + "grad_norm": 7.888322073346588, + "learning_rate": 1.4171301161115447e-05, + "loss": 0.827, + "step": 3520 + }, + { + "epoch": 1.9115092290988056, + "grad_norm": 10.407728423373431, + "learning_rate": 1.4168104635048756e-05, + "loss": 1.0511, + "step": 3521 + }, + { + "epoch": 1.9120521172638436, + "grad_norm": 8.154984984105669, + "learning_rate": 1.4164907593464239e-05, + "loss": 0.8156, + "step": 3522 + }, + { + "epoch": 1.9125950054288816, + "grad_norm": 9.142522776167024, + "learning_rate": 1.4161710036757314e-05, + "loss": 0.9046, + "step": 3523 + }, + { + "epoch": 1.9131378935939196, + "grad_norm": 8.610209396948733, + "learning_rate": 1.415851196532346e-05, + "loss": 0.5673, + "step": 3524 + }, + { + "epoch": 1.9136807817589576, + "grad_norm": 8.072698611315632, + "learning_rate": 1.415531337955822e-05, + "loss": 0.7375, + "step": 3525 + }, + { + "epoch": 1.9142236699239956, + "grad_norm": 9.878888535422037, + "learning_rate": 1.4152114279857197e-05, + "loss": 0.9677, + "step": 3526 + }, + { + "epoch": 1.9147665580890336, + "grad_norm": 8.524331697287977, + "learning_rate": 1.4148914666616062e-05, + "loss": 0.7617, + "step": 3527 + }, + { + "epoch": 1.9153094462540716, + "grad_norm": 8.04939882692512, + "learning_rate": 1.4145714540230549e-05, + "loss": 0.8767, + "step": 3528 + }, + { + "epoch": 1.9158523344191096, + "grad_norm": 10.69782500673993, + "learning_rate": 1.4142513901096453e-05, + "loss": 0.6357, + "step": 3529 + }, + { + "epoch": 1.9163952225841476, + "grad_norm": 7.279779735048647, + "learning_rate": 1.4139312749609637e-05, + "loss": 0.6474, + "step": 3530 + }, + { + "epoch": 1.9169381107491856, + "grad_norm": 7.290241222497209, + "learning_rate": 1.4136111086166024e-05, + "loss": 0.6929, + "step": 3531 + }, + { + "epoch": 1.9174809989142236, + "grad_norm": 8.3613074350218, + "learning_rate": 1.4132908911161598e-05, + "loss": 0.8199, + "step": 3532 + }, + { + "epoch": 1.9180238870792616, + "grad_norm": 9.015830183265928, + "learning_rate": 1.4129706224992413e-05, + "loss": 0.9811, + "step": 3533 + }, + { + "epoch": 1.9185667752442996, + "grad_norm": 9.621458967761185, + "learning_rate": 1.4126503028054579e-05, + "loss": 0.7498, + "step": 3534 + }, + { + "epoch": 1.9191096634093376, + "grad_norm": 11.105255757566027, + "learning_rate": 1.412329932074427e-05, + "loss": 0.7587, + "step": 3535 + }, + { + "epoch": 1.9196525515743756, + "grad_norm": 8.599145695443788, + "learning_rate": 1.4120095103457734e-05, + "loss": 0.934, + "step": 3536 + }, + { + "epoch": 1.9201954397394136, + "grad_norm": 7.727889210926642, + "learning_rate": 1.4116890376591268e-05, + "loss": 0.8305, + "step": 3537 + }, + { + "epoch": 1.9207383279044516, + "grad_norm": 12.029529213228301, + "learning_rate": 1.4113685140541242e-05, + "loss": 1.494, + "step": 3538 + }, + { + "epoch": 1.9212812160694897, + "grad_norm": 7.404838471492053, + "learning_rate": 1.411047939570408e-05, + "loss": 0.5168, + "step": 3539 + }, + { + "epoch": 1.9218241042345277, + "grad_norm": 7.540024472508733, + "learning_rate": 1.4107273142476272e-05, + "loss": 0.7162, + "step": 3540 + }, + { + "epoch": 1.9223669923995657, + "grad_norm": 7.885769945154571, + "learning_rate": 1.4104066381254378e-05, + "loss": 0.5964, + "step": 3541 + }, + { + "epoch": 1.9229098805646037, + "grad_norm": 8.287652724014645, + "learning_rate": 1.4100859112435013e-05, + "loss": 0.6175, + "step": 3542 + }, + { + "epoch": 1.9234527687296417, + "grad_norm": 8.335133777216624, + "learning_rate": 1.4097651336414857e-05, + "loss": 0.7163, + "step": 3543 + }, + { + "epoch": 1.9239956568946797, + "grad_norm": 7.765701865468595, + "learning_rate": 1.4094443053590652e-05, + "loss": 0.4534, + "step": 3544 + }, + { + "epoch": 1.9245385450597177, + "grad_norm": 7.716709219714201, + "learning_rate": 1.4091234264359206e-05, + "loss": 0.4425, + "step": 3545 + }, + { + "epoch": 1.9250814332247557, + "grad_norm": 11.979255681165771, + "learning_rate": 1.4088024969117387e-05, + "loss": 1.0249, + "step": 3546 + }, + { + "epoch": 1.9256243213897937, + "grad_norm": 9.424753754574079, + "learning_rate": 1.4084815168262123e-05, + "loss": 0.9663, + "step": 3547 + }, + { + "epoch": 1.9261672095548317, + "grad_norm": 10.101408847774994, + "learning_rate": 1.4081604862190407e-05, + "loss": 0.8136, + "step": 3548 + }, + { + "epoch": 1.9267100977198697, + "grad_norm": 11.755751498543427, + "learning_rate": 1.4078394051299298e-05, + "loss": 0.9451, + "step": 3549 + }, + { + "epoch": 1.9272529858849077, + "grad_norm": 12.251136923575924, + "learning_rate": 1.4075182735985913e-05, + "loss": 1.5716, + "step": 3550 + }, + { + "epoch": 1.9277958740499457, + "grad_norm": 10.087922707091717, + "learning_rate": 1.4071970916647432e-05, + "loss": 0.7026, + "step": 3551 + }, + { + "epoch": 1.9283387622149837, + "grad_norm": 8.507872704310849, + "learning_rate": 1.40687585936811e-05, + "loss": 0.7016, + "step": 3552 + }, + { + "epoch": 1.9288816503800217, + "grad_norm": 8.98138340766033, + "learning_rate": 1.4065545767484218e-05, + "loss": 0.6834, + "step": 3553 + }, + { + "epoch": 1.9294245385450597, + "grad_norm": 7.1085330003672045, + "learning_rate": 1.4062332438454156e-05, + "loss": 0.6813, + "step": 3554 + }, + { + "epoch": 1.9299674267100977, + "grad_norm": 13.499346110936985, + "learning_rate": 1.4059118606988345e-05, + "loss": 1.0068, + "step": 3555 + }, + { + "epoch": 1.9305103148751357, + "grad_norm": 7.698407375447584, + "learning_rate": 1.4055904273484275e-05, + "loss": 0.8418, + "step": 3556 + }, + { + "epoch": 1.9310532030401737, + "grad_norm": 8.424294783097983, + "learning_rate": 1.40526894383395e-05, + "loss": 0.8087, + "step": 3557 + }, + { + "epoch": 1.9315960912052117, + "grad_norm": 8.962780671623113, + "learning_rate": 1.4049474101951639e-05, + "loss": 0.8644, + "step": 3558 + }, + { + "epoch": 1.9321389793702497, + "grad_norm": 8.580581507909985, + "learning_rate": 1.4046258264718363e-05, + "loss": 1.0325, + "step": 3559 + }, + { + "epoch": 1.9326818675352877, + "grad_norm": 10.417878574145533, + "learning_rate": 1.4043041927037418e-05, + "loss": 0.7705, + "step": 3560 + }, + { + "epoch": 1.9332247557003257, + "grad_norm": 8.250628073154727, + "learning_rate": 1.4039825089306603e-05, + "loss": 0.794, + "step": 3561 + }, + { + "epoch": 1.9337676438653637, + "grad_norm": 8.228275233396532, + "learning_rate": 1.4036607751923783e-05, + "loss": 0.8759, + "step": 3562 + }, + { + "epoch": 1.9343105320304017, + "grad_norm": 7.760702739983814, + "learning_rate": 1.4033389915286884e-05, + "loss": 0.6458, + "step": 3563 + }, + { + "epoch": 1.9348534201954397, + "grad_norm": 6.508974748032192, + "learning_rate": 1.403017157979389e-05, + "loss": 0.6348, + "step": 3564 + }, + { + "epoch": 1.9353963083604777, + "grad_norm": 11.570824985629717, + "learning_rate": 1.4026952745842853e-05, + "loss": 1.0935, + "step": 3565 + }, + { + "epoch": 1.9359391965255157, + "grad_norm": 9.901046454739335, + "learning_rate": 1.4023733413831881e-05, + "loss": 0.7458, + "step": 3566 + }, + { + "epoch": 1.9364820846905537, + "grad_norm": 9.103951624002182, + "learning_rate": 1.402051358415915e-05, + "loss": 0.8253, + "step": 3567 + }, + { + "epoch": 1.9370249728555917, + "grad_norm": 9.293062225974824, + "learning_rate": 1.4017293257222887e-05, + "loss": 0.5801, + "step": 3568 + }, + { + "epoch": 1.9375678610206297, + "grad_norm": 7.180133735690273, + "learning_rate": 1.4014072433421398e-05, + "loss": 0.6063, + "step": 3569 + }, + { + "epoch": 1.9381107491856677, + "grad_norm": 9.028614205338394, + "learning_rate": 1.4010851113153028e-05, + "loss": 0.9412, + "step": 3570 + }, + { + "epoch": 1.9386536373507057, + "grad_norm": 9.51192022384654, + "learning_rate": 1.4007629296816202e-05, + "loss": 0.6459, + "step": 3571 + }, + { + "epoch": 1.9391965255157437, + "grad_norm": 9.883078787028563, + "learning_rate": 1.4004406984809396e-05, + "loss": 1.0324, + "step": 3572 + }, + { + "epoch": 1.9397394136807817, + "grad_norm": 7.433318187249177, + "learning_rate": 1.4001184177531154e-05, + "loss": 0.6152, + "step": 3573 + }, + { + "epoch": 1.9402823018458197, + "grad_norm": 10.165800890508075, + "learning_rate": 1.399796087538007e-05, + "loss": 0.8751, + "step": 3574 + }, + { + "epoch": 1.9408251900108577, + "grad_norm": 7.668138219396951, + "learning_rate": 1.3994737078754819e-05, + "loss": 0.43, + "step": 3575 + }, + { + "epoch": 1.9413680781758957, + "grad_norm": 8.929591410847367, + "learning_rate": 1.3991512788054115e-05, + "loss": 0.6173, + "step": 3576 + }, + { + "epoch": 1.9419109663409337, + "grad_norm": 8.069899029370944, + "learning_rate": 1.3988288003676755e-05, + "loss": 0.6782, + "step": 3577 + }, + { + "epoch": 1.9424538545059717, + "grad_norm": 7.635775018643444, + "learning_rate": 1.3985062726021574e-05, + "loss": 0.7503, + "step": 3578 + }, + { + "epoch": 1.9429967426710097, + "grad_norm": 11.649932328973133, + "learning_rate": 1.3981836955487485e-05, + "loss": 0.9154, + "step": 3579 + }, + { + "epoch": 1.9435396308360477, + "grad_norm": 9.11865472044189, + "learning_rate": 1.397861069247345e-05, + "loss": 0.8799, + "step": 3580 + }, + { + "epoch": 1.9440825190010858, + "grad_norm": 7.628563459317163, + "learning_rate": 1.3975383937378508e-05, + "loss": 0.7199, + "step": 3581 + }, + { + "epoch": 1.9446254071661238, + "grad_norm": 10.640127088045624, + "learning_rate": 1.3972156690601747e-05, + "loss": 0.7885, + "step": 3582 + }, + { + "epoch": 1.9451682953311618, + "grad_norm": 10.192406733466123, + "learning_rate": 1.3968928952542313e-05, + "loss": 0.6036, + "step": 3583 + }, + { + "epoch": 1.9457111834961998, + "grad_norm": 12.020287450972278, + "learning_rate": 1.396570072359942e-05, + "loss": 0.9477, + "step": 3584 + }, + { + "epoch": 1.9462540716612378, + "grad_norm": 8.855072111194623, + "learning_rate": 1.3962472004172343e-05, + "loss": 0.4384, + "step": 3585 + }, + { + "epoch": 1.9467969598262758, + "grad_norm": 7.173236338345466, + "learning_rate": 1.3959242794660412e-05, + "loss": 0.4969, + "step": 3586 + }, + { + "epoch": 1.9473398479913138, + "grad_norm": 6.505005692892076, + "learning_rate": 1.3956013095463024e-05, + "loss": 0.4585, + "step": 3587 + }, + { + "epoch": 1.9478827361563518, + "grad_norm": 7.041517208688192, + "learning_rate": 1.395278290697963e-05, + "loss": 0.5439, + "step": 3588 + }, + { + "epoch": 1.9484256243213898, + "grad_norm": 9.730465359550896, + "learning_rate": 1.3949552229609746e-05, + "loss": 0.8865, + "step": 3589 + }, + { + "epoch": 1.9489685124864278, + "grad_norm": 7.73678668778, + "learning_rate": 1.3946321063752948e-05, + "loss": 0.5094, + "step": 3590 + }, + { + "epoch": 1.9495114006514658, + "grad_norm": 7.095011334397053, + "learning_rate": 1.3943089409808872e-05, + "loss": 0.5224, + "step": 3591 + }, + { + "epoch": 1.9500542888165038, + "grad_norm": 8.086388468165458, + "learning_rate": 1.393985726817721e-05, + "loss": 0.7798, + "step": 3592 + }, + { + "epoch": 1.9505971769815418, + "grad_norm": 10.262604564335476, + "learning_rate": 1.3936624639257726e-05, + "loss": 1.2595, + "step": 3593 + }, + { + "epoch": 1.9511400651465798, + "grad_norm": 8.535907967530573, + "learning_rate": 1.393339152345023e-05, + "loss": 0.8205, + "step": 3594 + }, + { + "epoch": 1.9516829533116178, + "grad_norm": 7.864477340620307, + "learning_rate": 1.3930157921154601e-05, + "loss": 0.5967, + "step": 3595 + }, + { + "epoch": 1.9522258414766558, + "grad_norm": 6.739275201436143, + "learning_rate": 1.392692383277078e-05, + "loss": 0.5836, + "step": 3596 + }, + { + "epoch": 1.9527687296416938, + "grad_norm": 9.533970605538912, + "learning_rate": 1.392368925869876e-05, + "loss": 0.5761, + "step": 3597 + }, + { + "epoch": 1.9533116178067318, + "grad_norm": 11.376343280359167, + "learning_rate": 1.3920454199338598e-05, + "loss": 0.7596, + "step": 3598 + }, + { + "epoch": 1.9538545059717698, + "grad_norm": 7.280453649201749, + "learning_rate": 1.391721865509041e-05, + "loss": 0.4949, + "step": 3599 + }, + { + "epoch": 1.9543973941368078, + "grad_norm": 7.734190007799969, + "learning_rate": 1.391398262635438e-05, + "loss": 0.9177, + "step": 3600 + }, + { + "epoch": 1.9549402823018458, + "grad_norm": 13.106124045444247, + "learning_rate": 1.3910746113530738e-05, + "loss": 1.1884, + "step": 3601 + }, + { + "epoch": 1.9554831704668838, + "grad_norm": 11.868912051178873, + "learning_rate": 1.3907509117019783e-05, + "loss": 0.9245, + "step": 3602 + }, + { + "epoch": 1.9560260586319218, + "grad_norm": 10.307498071834987, + "learning_rate": 1.3904271637221876e-05, + "loss": 1.0802, + "step": 3603 + }, + { + "epoch": 1.9565689467969598, + "grad_norm": 8.308697157532139, + "learning_rate": 1.390103367453743e-05, + "loss": 0.4979, + "step": 3604 + }, + { + "epoch": 1.9571118349619978, + "grad_norm": 8.447604525372762, + "learning_rate": 1.3897795229366919e-05, + "loss": 0.8942, + "step": 3605 + }, + { + "epoch": 1.9576547231270358, + "grad_norm": 7.864013969756676, + "learning_rate": 1.3894556302110883e-05, + "loss": 0.642, + "step": 3606 + }, + { + "epoch": 1.9581976112920738, + "grad_norm": 7.903822544650853, + "learning_rate": 1.389131689316992e-05, + "loss": 0.8251, + "step": 3607 + }, + { + "epoch": 1.9587404994571118, + "grad_norm": 9.744275276467558, + "learning_rate": 1.3888077002944678e-05, + "loss": 1.0052, + "step": 3608 + }, + { + "epoch": 1.9592833876221498, + "grad_norm": 7.731164092689308, + "learning_rate": 1.3884836631835877e-05, + "loss": 0.6883, + "step": 3609 + }, + { + "epoch": 1.9598262757871878, + "grad_norm": 8.849291385162568, + "learning_rate": 1.3881595780244288e-05, + "loss": 0.9985, + "step": 3610 + }, + { + "epoch": 1.9603691639522258, + "grad_norm": 9.531929458494986, + "learning_rate": 1.3878354448570748e-05, + "loss": 0.7643, + "step": 3611 + }, + { + "epoch": 1.9609120521172638, + "grad_norm": 10.32443732773501, + "learning_rate": 1.3875112637216145e-05, + "loss": 0.848, + "step": 3612 + }, + { + "epoch": 1.9614549402823018, + "grad_norm": 7.417155617136626, + "learning_rate": 1.387187034658144e-05, + "loss": 0.5682, + "step": 3613 + }, + { + "epoch": 1.9619978284473398, + "grad_norm": 14.072831276291689, + "learning_rate": 1.386862757706764e-05, + "loss": 0.8279, + "step": 3614 + }, + { + "epoch": 1.9625407166123778, + "grad_norm": 7.764451928370161, + "learning_rate": 1.3865384329075812e-05, + "loss": 0.6687, + "step": 3615 + }, + { + "epoch": 1.9630836047774158, + "grad_norm": 8.686094567618806, + "learning_rate": 1.3862140603007095e-05, + "loss": 1.1034, + "step": 3616 + }, + { + "epoch": 1.9636264929424538, + "grad_norm": 10.753407337858437, + "learning_rate": 1.3858896399262669e-05, + "loss": 1.1692, + "step": 3617 + }, + { + "epoch": 1.9641693811074918, + "grad_norm": 9.096005648867607, + "learning_rate": 1.3855651718243786e-05, + "loss": 0.6191, + "step": 3618 + }, + { + "epoch": 1.9647122692725298, + "grad_norm": 10.387647049254962, + "learning_rate": 1.3852406560351752e-05, + "loss": 1.3573, + "step": 3619 + }, + { + "epoch": 1.9652551574375678, + "grad_norm": 7.173958455599766, + "learning_rate": 1.3849160925987936e-05, + "loss": 0.8106, + "step": 3620 + }, + { + "epoch": 1.9657980456026058, + "grad_norm": 9.434903281870886, + "learning_rate": 1.3845914815553765e-05, + "loss": 0.9581, + "step": 3621 + }, + { + "epoch": 1.9663409337676439, + "grad_norm": 7.671221453670278, + "learning_rate": 1.3842668229450717e-05, + "loss": 0.6989, + "step": 3622 + }, + { + "epoch": 1.9668838219326819, + "grad_norm": 7.360534692664803, + "learning_rate": 1.3839421168080338e-05, + "loss": 0.4912, + "step": 3623 + }, + { + "epoch": 1.9674267100977199, + "grad_norm": 6.853527063108637, + "learning_rate": 1.3836173631844231e-05, + "loss": 0.7485, + "step": 3624 + }, + { + "epoch": 1.9679695982627579, + "grad_norm": 8.619792011062271, + "learning_rate": 1.3832925621144057e-05, + "loss": 0.712, + "step": 3625 + }, + { + "epoch": 1.9685124864277959, + "grad_norm": 8.062990147577946, + "learning_rate": 1.3829677136381532e-05, + "loss": 0.7126, + "step": 3626 + }, + { + "epoch": 1.9690553745928339, + "grad_norm": 6.35400375341867, + "learning_rate": 1.3826428177958433e-05, + "loss": 0.5262, + "step": 3627 + }, + { + "epoch": 1.9695982627578719, + "grad_norm": 7.322594621519839, + "learning_rate": 1.3823178746276603e-05, + "loss": 0.6644, + "step": 3628 + }, + { + "epoch": 1.9701411509229099, + "grad_norm": 9.976906946653425, + "learning_rate": 1.3819928841737929e-05, + "loss": 0.7411, + "step": 3629 + }, + { + "epoch": 1.9706840390879479, + "grad_norm": 11.052922199516026, + "learning_rate": 1.3816678464744368e-05, + "loss": 0.9998, + "step": 3630 + }, + { + "epoch": 1.9712269272529859, + "grad_norm": 7.695988647358978, + "learning_rate": 1.381342761569793e-05, + "loss": 0.6805, + "step": 3631 + }, + { + "epoch": 1.9717698154180239, + "grad_norm": 10.387946538947885, + "learning_rate": 1.381017629500069e-05, + "loss": 0.8954, + "step": 3632 + }, + { + "epoch": 1.9723127035830619, + "grad_norm": 8.215553841008688, + "learning_rate": 1.380692450305477e-05, + "loss": 1.0488, + "step": 3633 + }, + { + "epoch": 1.9728555917480999, + "grad_norm": 8.550898170705977, + "learning_rate": 1.3803672240262364e-05, + "loss": 0.9301, + "step": 3634 + }, + { + "epoch": 1.9733984799131379, + "grad_norm": 9.978410770263547, + "learning_rate": 1.380041950702571e-05, + "loss": 0.9725, + "step": 3635 + }, + { + "epoch": 1.9739413680781759, + "grad_norm": 10.211160158432088, + "learning_rate": 1.3797166303747119e-05, + "loss": 1.1789, + "step": 3636 + }, + { + "epoch": 1.9744842562432139, + "grad_norm": 7.47091491709439, + "learning_rate": 1.3793912630828942e-05, + "loss": 1.2311, + "step": 3637 + }, + { + "epoch": 1.975027144408252, + "grad_norm": 9.966534074490724, + "learning_rate": 1.3790658488673607e-05, + "loss": 1.0057, + "step": 3638 + }, + { + "epoch": 1.97557003257329, + "grad_norm": 9.341517091004942, + "learning_rate": 1.378740387768359e-05, + "loss": 0.7708, + "step": 3639 + }, + { + "epoch": 1.976112920738328, + "grad_norm": 10.066755071795829, + "learning_rate": 1.3784148798261422e-05, + "loss": 1.3146, + "step": 3640 + }, + { + "epoch": 1.976655808903366, + "grad_norm": 8.306959980081418, + "learning_rate": 1.3780893250809705e-05, + "loss": 0.6926, + "step": 3641 + }, + { + "epoch": 1.977198697068404, + "grad_norm": 5.980548702977229, + "learning_rate": 1.377763723573108e-05, + "loss": 0.6776, + "step": 3642 + }, + { + "epoch": 1.977741585233442, + "grad_norm": 10.796897781225315, + "learning_rate": 1.3774380753428263e-05, + "loss": 1.0422, + "step": 3643 + }, + { + "epoch": 1.97828447339848, + "grad_norm": 9.650854393967004, + "learning_rate": 1.3771123804304018e-05, + "loss": 1.2566, + "step": 3644 + }, + { + "epoch": 1.978827361563518, + "grad_norm": 6.718663747121732, + "learning_rate": 1.3767866388761168e-05, + "loss": 0.5018, + "step": 3645 + }, + { + "epoch": 1.979370249728556, + "grad_norm": 8.345918705516796, + "learning_rate": 1.3764608507202604e-05, + "loss": 0.8501, + "step": 3646 + }, + { + "epoch": 1.979913137893594, + "grad_norm": 11.065999609001787, + "learning_rate": 1.3761350160031258e-05, + "loss": 1.2627, + "step": 3647 + }, + { + "epoch": 1.980456026058632, + "grad_norm": 8.567448073304528, + "learning_rate": 1.3758091347650126e-05, + "loss": 0.4477, + "step": 3648 + }, + { + "epoch": 1.98099891422367, + "grad_norm": 11.930236219703469, + "learning_rate": 1.3754832070462269e-05, + "loss": 1.5386, + "step": 3649 + }, + { + "epoch": 1.981541802388708, + "grad_norm": 7.334841489157127, + "learning_rate": 1.3751572328870797e-05, + "loss": 0.5759, + "step": 3650 + }, + { + "epoch": 1.982084690553746, + "grad_norm": 6.392865090419794, + "learning_rate": 1.3748312123278879e-05, + "loss": 0.6078, + "step": 3651 + }, + { + "epoch": 1.982627578718784, + "grad_norm": 8.598388892264492, + "learning_rate": 1.3745051454089744e-05, + "loss": 0.7682, + "step": 3652 + }, + { + "epoch": 1.983170466883822, + "grad_norm": 9.310542957266444, + "learning_rate": 1.3741790321706678e-05, + "loss": 0.9833, + "step": 3653 + }, + { + "epoch": 1.98371335504886, + "grad_norm": 8.917570196452148, + "learning_rate": 1.3738528726533021e-05, + "loss": 0.5915, + "step": 3654 + }, + { + "epoch": 1.984256243213898, + "grad_norm": 9.503202436394638, + "learning_rate": 1.3735266668972174e-05, + "loss": 0.9347, + "step": 3655 + }, + { + "epoch": 1.984799131378936, + "grad_norm": 9.187024367480415, + "learning_rate": 1.3732004149427592e-05, + "loss": 0.6669, + "step": 3656 + }, + { + "epoch": 1.985342019543974, + "grad_norm": 10.497812893691144, + "learning_rate": 1.3728741168302785e-05, + "loss": 1.0663, + "step": 3657 + }, + { + "epoch": 1.985884907709012, + "grad_norm": 9.646030688042543, + "learning_rate": 1.3725477726001332e-05, + "loss": 0.9913, + "step": 3658 + }, + { + "epoch": 1.98642779587405, + "grad_norm": 9.475431343889321, + "learning_rate": 1.3722213822926855e-05, + "loss": 0.8377, + "step": 3659 + }, + { + "epoch": 1.986970684039088, + "grad_norm": 9.449591675048769, + "learning_rate": 1.371894945948304e-05, + "loss": 0.8763, + "step": 3660 + }, + { + "epoch": 1.987513572204126, + "grad_norm": 7.3204142422253025, + "learning_rate": 1.3715684636073628e-05, + "loss": 0.7384, + "step": 3661 + }, + { + "epoch": 1.988056460369164, + "grad_norm": 10.339298014193695, + "learning_rate": 1.371241935310242e-05, + "loss": 1.0206, + "step": 3662 + }, + { + "epoch": 1.988599348534202, + "grad_norm": 7.083819102334954, + "learning_rate": 1.3709153610973266e-05, + "loss": 0.6905, + "step": 3663 + }, + { + "epoch": 1.98914223669924, + "grad_norm": 6.318629816327554, + "learning_rate": 1.3705887410090085e-05, + "loss": 0.7946, + "step": 3664 + }, + { + "epoch": 1.989685124864278, + "grad_norm": 7.109392074745231, + "learning_rate": 1.3702620750856843e-05, + "loss": 0.7125, + "step": 3665 + }, + { + "epoch": 1.990228013029316, + "grad_norm": 6.7832401368423545, + "learning_rate": 1.3699353633677565e-05, + "loss": 0.6091, + "step": 3666 + }, + { + "epoch": 1.990770901194354, + "grad_norm": 9.301471508167843, + "learning_rate": 1.3696086058956333e-05, + "loss": 0.9335, + "step": 3667 + }, + { + "epoch": 1.991313789359392, + "grad_norm": 8.167349191349723, + "learning_rate": 1.3692818027097288e-05, + "loss": 0.7148, + "step": 3668 + }, + { + "epoch": 1.99185667752443, + "grad_norm": 7.582331201930941, + "learning_rate": 1.3689549538504622e-05, + "loss": 0.5952, + "step": 3669 + }, + { + "epoch": 1.992399565689468, + "grad_norm": 8.743965167551632, + "learning_rate": 1.3686280593582588e-05, + "loss": 0.9268, + "step": 3670 + }, + { + "epoch": 1.992942453854506, + "grad_norm": 10.291738284427778, + "learning_rate": 1.3683011192735496e-05, + "loss": 0.9934, + "step": 3671 + }, + { + "epoch": 1.993485342019544, + "grad_norm": 9.808936601876956, + "learning_rate": 1.3679741336367711e-05, + "loss": 1.0506, + "step": 3672 + }, + { + "epoch": 1.994028230184582, + "grad_norm": 10.002691567142412, + "learning_rate": 1.3676471024883654e-05, + "loss": 1.1439, + "step": 3673 + }, + { + "epoch": 1.99457111834962, + "grad_norm": 7.558103827510508, + "learning_rate": 1.36732002586878e-05, + "loss": 1.0059, + "step": 3674 + }, + { + "epoch": 1.995114006514658, + "grad_norm": 7.958890420546252, + "learning_rate": 1.3669929038184684e-05, + "loss": 1.0301, + "step": 3675 + }, + { + "epoch": 1.995656894679696, + "grad_norm": 8.042682701022132, + "learning_rate": 1.3666657363778895e-05, + "loss": 0.6881, + "step": 3676 + }, + { + "epoch": 1.996199782844734, + "grad_norm": 7.54271762136155, + "learning_rate": 1.366338523587508e-05, + "loss": 0.5227, + "step": 3677 + }, + { + "epoch": 1.996742671009772, + "grad_norm": 9.060117135659972, + "learning_rate": 1.3660112654877939e-05, + "loss": 0.9847, + "step": 3678 + }, + { + "epoch": 1.99728555917481, + "grad_norm": 11.00270077841202, + "learning_rate": 1.3656839621192233e-05, + "loss": 1.0918, + "step": 3679 + }, + { + "epoch": 1.997828447339848, + "grad_norm": 9.156434005073692, + "learning_rate": 1.3653566135222774e-05, + "loss": 0.8104, + "step": 3680 + }, + { + "epoch": 1.998371335504886, + "grad_norm": 8.405413514743108, + "learning_rate": 1.3650292197374433e-05, + "loss": 0.9414, + "step": 3681 + }, + { + "epoch": 1.998914223669924, + "grad_norm": 7.9594907003419495, + "learning_rate": 1.3647017808052135e-05, + "loss": 0.9257, + "step": 3682 + }, + { + "epoch": 1.999457111834962, + "grad_norm": 9.821777823228977, + "learning_rate": 1.3643742967660859e-05, + "loss": 0.8787, + "step": 3683 + }, + { + "epoch": 2.0, + "grad_norm": 8.766507481527489, + "learning_rate": 1.3640467676605648e-05, + "loss": 0.8472, + "step": 3684 + }, + { + "epoch": 2.000542888165038, + "grad_norm": 9.493942149196295, + "learning_rate": 1.3637191935291596e-05, + "loss": 0.8147, + "step": 3685 + }, + { + "epoch": 2.001085776330076, + "grad_norm": 9.782638508357579, + "learning_rate": 1.3633915744123844e-05, + "loss": 0.8788, + "step": 3686 + }, + { + "epoch": 2.001628664495114, + "grad_norm": 8.46210638839138, + "learning_rate": 1.3630639103507604e-05, + "loss": 1.078, + "step": 3687 + }, + { + "epoch": 2.002171552660152, + "grad_norm": 7.345658753018104, + "learning_rate": 1.3627362013848134e-05, + "loss": 0.6838, + "step": 3688 + }, + { + "epoch": 2.00271444082519, + "grad_norm": 7.329182601936008, + "learning_rate": 1.3624084475550743e-05, + "loss": 0.7249, + "step": 3689 + }, + { + "epoch": 2.003257328990228, + "grad_norm": 9.89510466995638, + "learning_rate": 1.3620806489020813e-05, + "loss": 0.8018, + "step": 3690 + }, + { + "epoch": 2.003800217155266, + "grad_norm": 6.836942603575798, + "learning_rate": 1.3617528054663764e-05, + "loss": 0.6645, + "step": 3691 + }, + { + "epoch": 2.004343105320304, + "grad_norm": 8.842045073299985, + "learning_rate": 1.3614249172885081e-05, + "loss": 0.5143, + "step": 3692 + }, + { + "epoch": 2.004885993485342, + "grad_norm": 7.671200527248645, + "learning_rate": 1.36109698440903e-05, + "loss": 0.8038, + "step": 3693 + }, + { + "epoch": 2.00542888165038, + "grad_norm": 9.544464428119456, + "learning_rate": 1.3607690068685013e-05, + "loss": 0.6396, + "step": 3694 + }, + { + "epoch": 2.005971769815418, + "grad_norm": 8.537202930183216, + "learning_rate": 1.3604409847074868e-05, + "loss": 0.7104, + "step": 3695 + }, + { + "epoch": 2.006514657980456, + "grad_norm": 7.121091005738386, + "learning_rate": 1.3601129179665572e-05, + "loss": 0.4337, + "step": 3696 + }, + { + "epoch": 2.007057546145494, + "grad_norm": 9.740494539210806, + "learning_rate": 1.3597848066862875e-05, + "loss": 0.9114, + "step": 3697 + }, + { + "epoch": 2.007600434310532, + "grad_norm": 8.100951122902112, + "learning_rate": 1.3594566509072599e-05, + "loss": 0.722, + "step": 3698 + }, + { + "epoch": 2.00814332247557, + "grad_norm": 6.959347625774498, + "learning_rate": 1.3591284506700606e-05, + "loss": 0.5429, + "step": 3699 + }, + { + "epoch": 2.008686210640608, + "grad_norm": 9.993493206941414, + "learning_rate": 1.3588002060152822e-05, + "loss": 0.8251, + "step": 3700 + }, + { + "epoch": 2.009229098805646, + "grad_norm": 9.966526343412045, + "learning_rate": 1.3584719169835226e-05, + "loss": 0.4847, + "step": 3701 + }, + { + "epoch": 2.009771986970684, + "grad_norm": 7.888503709599375, + "learning_rate": 1.3581435836153847e-05, + "loss": 0.7941, + "step": 3702 + }, + { + "epoch": 2.010314875135722, + "grad_norm": 9.057972693252164, + "learning_rate": 1.3578152059514778e-05, + "loss": 0.8963, + "step": 3703 + }, + { + "epoch": 2.01085776330076, + "grad_norm": 10.717910066603423, + "learning_rate": 1.3574867840324157e-05, + "loss": 0.6485, + "step": 3704 + }, + { + "epoch": 2.011400651465798, + "grad_norm": 7.257931643777971, + "learning_rate": 1.3571583178988188e-05, + "loss": 0.4569, + "step": 3705 + }, + { + "epoch": 2.011943539630836, + "grad_norm": 16.546104792345954, + "learning_rate": 1.3568298075913119e-05, + "loss": 1.0473, + "step": 3706 + }, + { + "epoch": 2.012486427795874, + "grad_norm": 9.882729561546991, + "learning_rate": 1.3565012531505252e-05, + "loss": 0.9729, + "step": 3707 + }, + { + "epoch": 2.013029315960912, + "grad_norm": 7.060037163975289, + "learning_rate": 1.3561726546170956e-05, + "loss": 0.5935, + "step": 3708 + }, + { + "epoch": 2.01357220412595, + "grad_norm": 7.081117948590212, + "learning_rate": 1.3558440120316644e-05, + "loss": 0.6676, + "step": 3709 + }, + { + "epoch": 2.014115092290988, + "grad_norm": 10.048606586375149, + "learning_rate": 1.3555153254348788e-05, + "loss": 0.6307, + "step": 3710 + }, + { + "epoch": 2.014657980456026, + "grad_norm": 7.720688031890385, + "learning_rate": 1.3551865948673912e-05, + "loss": 0.7882, + "step": 3711 + }, + { + "epoch": 2.015200868621064, + "grad_norm": 7.644495073412038, + "learning_rate": 1.3548578203698592e-05, + "loss": 0.8904, + "step": 3712 + }, + { + "epoch": 2.015743756786102, + "grad_norm": 10.13216917020047, + "learning_rate": 1.3545290019829466e-05, + "loss": 0.7832, + "step": 3713 + }, + { + "epoch": 2.01628664495114, + "grad_norm": 6.321686691568546, + "learning_rate": 1.3542001397473219e-05, + "loss": 0.5682, + "step": 3714 + }, + { + "epoch": 2.016829533116178, + "grad_norm": 7.966895819853924, + "learning_rate": 1.3538712337036594e-05, + "loss": 0.7704, + "step": 3715 + }, + { + "epoch": 2.017372421281216, + "grad_norm": 8.827713158118389, + "learning_rate": 1.3535422838926389e-05, + "loss": 0.6741, + "step": 3716 + }, + { + "epoch": 2.017915309446254, + "grad_norm": 7.404288938912365, + "learning_rate": 1.3532132903549453e-05, + "loss": 0.5325, + "step": 3717 + }, + { + "epoch": 2.018458197611292, + "grad_norm": 9.869324924245625, + "learning_rate": 1.352884253131269e-05, + "loss": 0.7326, + "step": 3718 + }, + { + "epoch": 2.01900108577633, + "grad_norm": 7.824783976445888, + "learning_rate": 1.3525551722623056e-05, + "loss": 0.4683, + "step": 3719 + }, + { + "epoch": 2.019543973941368, + "grad_norm": 10.516375854488775, + "learning_rate": 1.3522260477887566e-05, + "loss": 0.839, + "step": 3720 + }, + { + "epoch": 2.020086862106406, + "grad_norm": 10.698678918246005, + "learning_rate": 1.3518968797513288e-05, + "loss": 1.0513, + "step": 3721 + }, + { + "epoch": 2.020629750271444, + "grad_norm": 9.430167728870957, + "learning_rate": 1.351567668190734e-05, + "loss": 0.6696, + "step": 3722 + }, + { + "epoch": 2.021172638436482, + "grad_norm": 9.619270238605017, + "learning_rate": 1.3512384131476897e-05, + "loss": 0.4608, + "step": 3723 + }, + { + "epoch": 2.02171552660152, + "grad_norm": 6.9328636261250525, + "learning_rate": 1.350909114662919e-05, + "loss": 0.3952, + "step": 3724 + }, + { + "epoch": 2.022258414766558, + "grad_norm": 11.694555068035188, + "learning_rate": 1.3505797727771493e-05, + "loss": 0.7526, + "step": 3725 + }, + { + "epoch": 2.022801302931596, + "grad_norm": 7.417427417334792, + "learning_rate": 1.3502503875311149e-05, + "loss": 0.5239, + "step": 3726 + }, + { + "epoch": 2.023344191096634, + "grad_norm": 10.99931861711161, + "learning_rate": 1.349920958965554e-05, + "loss": 0.5504, + "step": 3727 + }, + { + "epoch": 2.023887079261672, + "grad_norm": 9.418315087755559, + "learning_rate": 1.3495914871212113e-05, + "loss": 0.6791, + "step": 3728 + }, + { + "epoch": 2.02442996742671, + "grad_norm": 12.589472190346669, + "learning_rate": 1.3492619720388363e-05, + "loss": 0.815, + "step": 3729 + }, + { + "epoch": 2.024972855591748, + "grad_norm": 6.60691672927324, + "learning_rate": 1.348932413759184e-05, + "loss": 0.4112, + "step": 3730 + }, + { + "epoch": 2.025515743756786, + "grad_norm": 10.970405529985833, + "learning_rate": 1.3486028123230145e-05, + "loss": 0.6164, + "step": 3731 + }, + { + "epoch": 2.026058631921824, + "grad_norm": 9.419473243249211, + "learning_rate": 1.3482731677710938e-05, + "loss": 0.829, + "step": 3732 + }, + { + "epoch": 2.026601520086862, + "grad_norm": 8.809487998193466, + "learning_rate": 1.3479434801441925e-05, + "loss": 0.6386, + "step": 3733 + }, + { + "epoch": 2.0271444082519, + "grad_norm": 8.586869223296738, + "learning_rate": 1.3476137494830872e-05, + "loss": 0.4586, + "step": 3734 + }, + { + "epoch": 2.027687296416938, + "grad_norm": 10.792985384487755, + "learning_rate": 1.3472839758285595e-05, + "loss": 0.68, + "step": 3735 + }, + { + "epoch": 2.028230184581976, + "grad_norm": 11.494014730485675, + "learning_rate": 1.346954159221396e-05, + "loss": 0.9465, + "step": 3736 + }, + { + "epoch": 2.028773072747014, + "grad_norm": 9.047474116723855, + "learning_rate": 1.3466242997023891e-05, + "loss": 0.637, + "step": 3737 + }, + { + "epoch": 2.029315960912052, + "grad_norm": 7.32591412143244, + "learning_rate": 1.3462943973123362e-05, + "loss": 0.5071, + "step": 3738 + }, + { + "epoch": 2.02985884907709, + "grad_norm": 8.462579835722188, + "learning_rate": 1.3459644520920405e-05, + "loss": 0.5568, + "step": 3739 + }, + { + "epoch": 2.030401737242128, + "grad_norm": 11.667085873745426, + "learning_rate": 1.34563446408231e-05, + "loss": 0.579, + "step": 3740 + }, + { + "epoch": 2.030944625407166, + "grad_norm": 6.778347309469529, + "learning_rate": 1.3453044333239577e-05, + "loss": 0.3399, + "step": 3741 + }, + { + "epoch": 2.031487513572204, + "grad_norm": 7.280157509234148, + "learning_rate": 1.3449743598578033e-05, + "loss": 0.4236, + "step": 3742 + }, + { + "epoch": 2.032030401737242, + "grad_norm": 10.902366947251025, + "learning_rate": 1.34464424372467e-05, + "loss": 0.7984, + "step": 3743 + }, + { + "epoch": 2.03257328990228, + "grad_norm": 7.729938792603154, + "learning_rate": 1.344314084965388e-05, + "loss": 0.5115, + "step": 3744 + }, + { + "epoch": 2.033116178067318, + "grad_norm": 9.034690624481534, + "learning_rate": 1.3439838836207905e-05, + "loss": 0.3528, + "step": 3745 + }, + { + "epoch": 2.033659066232356, + "grad_norm": 10.813371264470662, + "learning_rate": 1.3436536397317183e-05, + "loss": 0.4972, + "step": 3746 + }, + { + "epoch": 2.034201954397394, + "grad_norm": 12.116428074423327, + "learning_rate": 1.343323353339016e-05, + "loss": 0.9414, + "step": 3747 + }, + { + "epoch": 2.034744842562432, + "grad_norm": 11.714995187138708, + "learning_rate": 1.3429930244835343e-05, + "loss": 0.7842, + "step": 3748 + }, + { + "epoch": 2.03528773072747, + "grad_norm": 9.956102297397003, + "learning_rate": 1.3426626532061287e-05, + "loss": 0.7304, + "step": 3749 + }, + { + "epoch": 2.035830618892508, + "grad_norm": 10.670014393137768, + "learning_rate": 1.34233223954766e-05, + "loss": 0.746, + "step": 3750 + }, + { + "epoch": 2.036373507057546, + "grad_norm": 10.266437034084714, + "learning_rate": 1.3420017835489945e-05, + "loss": 1.0871, + "step": 3751 + }, + { + "epoch": 2.036916395222584, + "grad_norm": 11.75190588784515, + "learning_rate": 1.3416712852510033e-05, + "loss": 0.7606, + "step": 3752 + }, + { + "epoch": 2.037459283387622, + "grad_norm": 11.54288048263274, + "learning_rate": 1.3413407446945627e-05, + "loss": 0.5887, + "step": 3753 + }, + { + "epoch": 2.03800217155266, + "grad_norm": 18.002644323614724, + "learning_rate": 1.3410101619205552e-05, + "loss": 1.0593, + "step": 3754 + }, + { + "epoch": 2.038545059717698, + "grad_norm": 8.97759172274721, + "learning_rate": 1.3406795369698671e-05, + "loss": 0.7194, + "step": 3755 + }, + { + "epoch": 2.039087947882736, + "grad_norm": 8.793507308104033, + "learning_rate": 1.3403488698833912e-05, + "loss": 0.4891, + "step": 3756 + }, + { + "epoch": 2.039630836047774, + "grad_norm": 9.263725202719241, + "learning_rate": 1.3400181607020243e-05, + "loss": 0.8258, + "step": 3757 + }, + { + "epoch": 2.040173724212812, + "grad_norm": 10.568214203997627, + "learning_rate": 1.3396874094666694e-05, + "loss": 0.5841, + "step": 3758 + }, + { + "epoch": 2.04071661237785, + "grad_norm": 10.237711379932081, + "learning_rate": 1.3393566162182346e-05, + "loss": 0.6901, + "step": 3759 + }, + { + "epoch": 2.041259500542888, + "grad_norm": 11.264856346485661, + "learning_rate": 1.3390257809976322e-05, + "loss": 0.8029, + "step": 3760 + }, + { + "epoch": 2.041802388707926, + "grad_norm": 7.504173750209935, + "learning_rate": 1.3386949038457813e-05, + "loss": 0.3739, + "step": 3761 + }, + { + "epoch": 2.042345276872964, + "grad_norm": 8.099570741006275, + "learning_rate": 1.3383639848036044e-05, + "loss": 0.7457, + "step": 3762 + }, + { + "epoch": 2.042888165038002, + "grad_norm": 9.87563463161881, + "learning_rate": 1.3380330239120313e-05, + "loss": 0.632, + "step": 3763 + }, + { + "epoch": 2.04343105320304, + "grad_norm": 9.304965814712034, + "learning_rate": 1.3377020212119946e-05, + "loss": 0.657, + "step": 3764 + }, + { + "epoch": 2.043973941368078, + "grad_norm": 7.738549848691211, + "learning_rate": 1.3373709767444339e-05, + "loss": 0.4198, + "step": 3765 + }, + { + "epoch": 2.044516829533116, + "grad_norm": 11.805824611419341, + "learning_rate": 1.3370398905502928e-05, + "loss": 0.8173, + "step": 3766 + }, + { + "epoch": 2.045059717698154, + "grad_norm": 10.425380865754251, + "learning_rate": 1.3367087626705211e-05, + "loss": 0.7149, + "step": 3767 + }, + { + "epoch": 2.045602605863192, + "grad_norm": 11.884153232685467, + "learning_rate": 1.336377593146073e-05, + "loss": 0.723, + "step": 3768 + }, + { + "epoch": 2.04614549402823, + "grad_norm": 9.663581488842127, + "learning_rate": 1.336046382017908e-05, + "loss": 0.6457, + "step": 3769 + }, + { + "epoch": 2.046688382193268, + "grad_norm": 9.184284700313228, + "learning_rate": 1.335715129326991e-05, + "loss": 0.6645, + "step": 3770 + }, + { + "epoch": 2.047231270358306, + "grad_norm": 7.398657595562234, + "learning_rate": 1.3353838351142915e-05, + "loss": 0.7288, + "step": 3771 + }, + { + "epoch": 2.047774158523344, + "grad_norm": 10.416946087073072, + "learning_rate": 1.335052499420785e-05, + "loss": 0.4974, + "step": 3772 + }, + { + "epoch": 2.048317046688382, + "grad_norm": 8.676282075018099, + "learning_rate": 1.3347211222874514e-05, + "loss": 0.7128, + "step": 3773 + }, + { + "epoch": 2.04885993485342, + "grad_norm": 10.339910800835435, + "learning_rate": 1.3343897037552758e-05, + "loss": 0.6674, + "step": 3774 + }, + { + "epoch": 2.049402823018458, + "grad_norm": 10.976308424624598, + "learning_rate": 1.3340582438652488e-05, + "loss": 0.7683, + "step": 3775 + }, + { + "epoch": 2.049945711183496, + "grad_norm": 11.874329810000683, + "learning_rate": 1.3337267426583658e-05, + "loss": 1.2199, + "step": 3776 + }, + { + "epoch": 2.050488599348534, + "grad_norm": 9.306473243935873, + "learning_rate": 1.3333952001756272e-05, + "loss": 0.8681, + "step": 3777 + }, + { + "epoch": 2.0510314875135722, + "grad_norm": 9.083116427983501, + "learning_rate": 1.333063616458039e-05, + "loss": 0.7512, + "step": 3778 + }, + { + "epoch": 2.05157437567861, + "grad_norm": 5.762711234835306, + "learning_rate": 1.3327319915466119e-05, + "loss": 0.3384, + "step": 3779 + }, + { + "epoch": 2.0521172638436482, + "grad_norm": 7.431272084519616, + "learning_rate": 1.332400325482362e-05, + "loss": 0.5188, + "step": 3780 + }, + { + "epoch": 2.052660152008686, + "grad_norm": 5.736965263179954, + "learning_rate": 1.3320686183063096e-05, + "loss": 0.3476, + "step": 3781 + }, + { + "epoch": 2.0532030401737242, + "grad_norm": 7.130870597882685, + "learning_rate": 1.3317368700594815e-05, + "loss": 0.7929, + "step": 3782 + }, + { + "epoch": 2.053745928338762, + "grad_norm": 7.713657404988891, + "learning_rate": 1.3314050807829088e-05, + "loss": 0.422, + "step": 3783 + }, + { + "epoch": 2.0542888165038002, + "grad_norm": 6.383542293487279, + "learning_rate": 1.3310732505176276e-05, + "loss": 0.568, + "step": 3784 + }, + { + "epoch": 2.054831704668838, + "grad_norm": 12.326747003957852, + "learning_rate": 1.3307413793046787e-05, + "loss": 0.8061, + "step": 3785 + }, + { + "epoch": 2.0553745928338762, + "grad_norm": 10.668098962086818, + "learning_rate": 1.330409467185109e-05, + "loss": 1.4344, + "step": 3786 + }, + { + "epoch": 2.055917480998914, + "grad_norm": 9.844537720560512, + "learning_rate": 1.3300775141999698e-05, + "loss": 1.0089, + "step": 3787 + }, + { + "epoch": 2.0564603691639523, + "grad_norm": 9.714672703692647, + "learning_rate": 1.3297455203903176e-05, + "loss": 0.6069, + "step": 3788 + }, + { + "epoch": 2.05700325732899, + "grad_norm": 10.27882170416886, + "learning_rate": 1.3294134857972139e-05, + "loss": 0.7279, + "step": 3789 + }, + { + "epoch": 2.0575461454940283, + "grad_norm": 10.99914755541959, + "learning_rate": 1.3290814104617253e-05, + "loss": 0.6787, + "step": 3790 + }, + { + "epoch": 2.058089033659066, + "grad_norm": 13.627823006472678, + "learning_rate": 1.3287492944249234e-05, + "loss": 0.9052, + "step": 3791 + }, + { + "epoch": 2.0586319218241043, + "grad_norm": 7.734783593910536, + "learning_rate": 1.3284171377278849e-05, + "loss": 0.7141, + "step": 3792 + }, + { + "epoch": 2.059174809989142, + "grad_norm": 9.247896244835857, + "learning_rate": 1.3280849404116913e-05, + "loss": 0.6071, + "step": 3793 + }, + { + "epoch": 2.0597176981541803, + "grad_norm": 9.004013607687236, + "learning_rate": 1.3277527025174295e-05, + "loss": 0.5887, + "step": 3794 + }, + { + "epoch": 2.060260586319218, + "grad_norm": 9.62029412051341, + "learning_rate": 1.3274204240861908e-05, + "loss": 0.8847, + "step": 3795 + }, + { + "epoch": 2.0608034744842563, + "grad_norm": 7.500348953035017, + "learning_rate": 1.3270881051590725e-05, + "loss": 0.4769, + "step": 3796 + }, + { + "epoch": 2.061346362649294, + "grad_norm": 10.153144580789224, + "learning_rate": 1.326755745777176e-05, + "loss": 0.645, + "step": 3797 + }, + { + "epoch": 2.0618892508143323, + "grad_norm": 6.813155627292939, + "learning_rate": 1.326423345981608e-05, + "loss": 0.6711, + "step": 3798 + }, + { + "epoch": 2.06243213897937, + "grad_norm": 9.362898780009592, + "learning_rate": 1.32609090581348e-05, + "loss": 0.5464, + "step": 3799 + }, + { + "epoch": 2.0629750271444083, + "grad_norm": 8.540638734313639, + "learning_rate": 1.3257584253139096e-05, + "loss": 0.6523, + "step": 3800 + }, + { + "epoch": 2.063517915309446, + "grad_norm": 9.000753471311711, + "learning_rate": 1.3254259045240176e-05, + "loss": 0.5362, + "step": 3801 + }, + { + "epoch": 2.0640608034744843, + "grad_norm": 10.275669325762056, + "learning_rate": 1.3250933434849316e-05, + "loss": 0.6061, + "step": 3802 + }, + { + "epoch": 2.064603691639522, + "grad_norm": 11.615294601299743, + "learning_rate": 1.3247607422377823e-05, + "loss": 0.673, + "step": 3803 + }, + { + "epoch": 2.0651465798045603, + "grad_norm": 10.316289286330115, + "learning_rate": 1.324428100823707e-05, + "loss": 0.7997, + "step": 3804 + }, + { + "epoch": 2.065689467969598, + "grad_norm": 14.555740784089037, + "learning_rate": 1.3240954192838472e-05, + "loss": 0.9957, + "step": 3805 + }, + { + "epoch": 2.0662323561346363, + "grad_norm": 7.674873455902871, + "learning_rate": 1.323762697659349e-05, + "loss": 0.4771, + "step": 3806 + }, + { + "epoch": 2.066775244299674, + "grad_norm": 8.530555441214027, + "learning_rate": 1.3234299359913647e-05, + "loss": 0.8666, + "step": 3807 + }, + { + "epoch": 2.0673181324647123, + "grad_norm": 6.463731480876421, + "learning_rate": 1.3230971343210503e-05, + "loss": 0.5742, + "step": 3808 + }, + { + "epoch": 2.06786102062975, + "grad_norm": 8.074017879990592, + "learning_rate": 1.3227642926895676e-05, + "loss": 0.3637, + "step": 3809 + }, + { + "epoch": 2.0684039087947883, + "grad_norm": 9.989752857060338, + "learning_rate": 1.3224314111380828e-05, + "loss": 0.7389, + "step": 3810 + }, + { + "epoch": 2.068946796959826, + "grad_norm": 12.892065053280843, + "learning_rate": 1.3220984897077669e-05, + "loss": 1.2959, + "step": 3811 + }, + { + "epoch": 2.0694896851248643, + "grad_norm": 7.497956491794191, + "learning_rate": 1.3217655284397965e-05, + "loss": 0.4384, + "step": 3812 + }, + { + "epoch": 2.070032573289902, + "grad_norm": 14.719876212901973, + "learning_rate": 1.3214325273753528e-05, + "loss": 1.3054, + "step": 3813 + }, + { + "epoch": 2.0705754614549403, + "grad_norm": 10.767865599283661, + "learning_rate": 1.3210994865556219e-05, + "loss": 0.9961, + "step": 3814 + }, + { + "epoch": 2.071118349619978, + "grad_norm": 9.198818655810543, + "learning_rate": 1.3207664060217946e-05, + "loss": 0.5916, + "step": 3815 + }, + { + "epoch": 2.0716612377850163, + "grad_norm": 9.598428080121797, + "learning_rate": 1.320433285815067e-05, + "loss": 0.7907, + "step": 3816 + }, + { + "epoch": 2.072204125950054, + "grad_norm": 9.48389213525687, + "learning_rate": 1.32010012597664e-05, + "loss": 0.8385, + "step": 3817 + }, + { + "epoch": 2.0727470141150923, + "grad_norm": 7.562927002528845, + "learning_rate": 1.3197669265477191e-05, + "loss": 0.6748, + "step": 3818 + }, + { + "epoch": 2.07328990228013, + "grad_norm": 7.5241368226931185, + "learning_rate": 1.319433687569515e-05, + "loss": 0.5605, + "step": 3819 + }, + { + "epoch": 2.0738327904451683, + "grad_norm": 7.952800933557159, + "learning_rate": 1.3191004090832436e-05, + "loss": 0.6624, + "step": 3820 + }, + { + "epoch": 2.074375678610206, + "grad_norm": 8.02814530035842, + "learning_rate": 1.318767091130125e-05, + "loss": 0.488, + "step": 3821 + }, + { + "epoch": 2.0749185667752443, + "grad_norm": 7.000417608550192, + "learning_rate": 1.3184337337513849e-05, + "loss": 0.7194, + "step": 3822 + }, + { + "epoch": 2.075461454940282, + "grad_norm": 8.965861432524546, + "learning_rate": 1.3181003369882527e-05, + "loss": 0.4823, + "step": 3823 + }, + { + "epoch": 2.0760043431053203, + "grad_norm": 8.717885349674289, + "learning_rate": 1.3177669008819635e-05, + "loss": 0.6482, + "step": 3824 + }, + { + "epoch": 2.076547231270358, + "grad_norm": 9.217767101253152, + "learning_rate": 1.317433425473758e-05, + "loss": 0.5276, + "step": 3825 + }, + { + "epoch": 2.0770901194353963, + "grad_norm": 8.84305940258517, + "learning_rate": 1.3170999108048804e-05, + "loss": 0.6579, + "step": 3826 + }, + { + "epoch": 2.077633007600434, + "grad_norm": 7.501762934992334, + "learning_rate": 1.3167663569165803e-05, + "loss": 0.6633, + "step": 3827 + }, + { + "epoch": 2.0781758957654723, + "grad_norm": 7.815925996775423, + "learning_rate": 1.3164327638501126e-05, + "loss": 0.6881, + "step": 3828 + }, + { + "epoch": 2.07871878393051, + "grad_norm": 11.789396888417095, + "learning_rate": 1.3160991316467362e-05, + "loss": 0.5761, + "step": 3829 + }, + { + "epoch": 2.0792616720955484, + "grad_norm": 8.450719528350264, + "learning_rate": 1.3157654603477155e-05, + "loss": 0.7263, + "step": 3830 + }, + { + "epoch": 2.079804560260586, + "grad_norm": 9.544428310813693, + "learning_rate": 1.315431749994319e-05, + "loss": 0.6408, + "step": 3831 + }, + { + "epoch": 2.0803474484256244, + "grad_norm": 7.454950880630089, + "learning_rate": 1.3150980006278219e-05, + "loss": 0.4425, + "step": 3832 + }, + { + "epoch": 2.080890336590662, + "grad_norm": 11.462923993924347, + "learning_rate": 1.314764212289501e-05, + "loss": 0.5349, + "step": 3833 + }, + { + "epoch": 2.0814332247557004, + "grad_norm": 8.900231872199269, + "learning_rate": 1.314430385020641e-05, + "loss": 0.8022, + "step": 3834 + }, + { + "epoch": 2.081976112920738, + "grad_norm": 7.035528648373333, + "learning_rate": 1.3140965188625299e-05, + "loss": 0.6513, + "step": 3835 + }, + { + "epoch": 2.0825190010857764, + "grad_norm": 12.327229999200162, + "learning_rate": 1.3137626138564606e-05, + "loss": 1.0623, + "step": 3836 + }, + { + "epoch": 2.083061889250814, + "grad_norm": 8.942692654829559, + "learning_rate": 1.3134286700437308e-05, + "loss": 0.4689, + "step": 3837 + }, + { + "epoch": 2.0836047774158524, + "grad_norm": 11.092455380734007, + "learning_rate": 1.313094687465644e-05, + "loss": 0.6808, + "step": 3838 + }, + { + "epoch": 2.08414766558089, + "grad_norm": 8.918018917563646, + "learning_rate": 1.3127606661635075e-05, + "loss": 0.523, + "step": 3839 + }, + { + "epoch": 2.0846905537459284, + "grad_norm": 13.878014157853947, + "learning_rate": 1.312426606178633e-05, + "loss": 1.0231, + "step": 3840 + }, + { + "epoch": 2.085233441910966, + "grad_norm": 10.432596711564004, + "learning_rate": 1.3120925075523379e-05, + "loss": 0.8011, + "step": 3841 + }, + { + "epoch": 2.0857763300760044, + "grad_norm": 8.126214858686065, + "learning_rate": 1.3117583703259445e-05, + "loss": 0.5272, + "step": 3842 + }, + { + "epoch": 2.086319218241042, + "grad_norm": 9.596541824389169, + "learning_rate": 1.3114241945407783e-05, + "loss": 0.6585, + "step": 3843 + }, + { + "epoch": 2.0868621064060804, + "grad_norm": 10.438282441705661, + "learning_rate": 1.3110899802381718e-05, + "loss": 0.8474, + "step": 3844 + }, + { + "epoch": 2.087404994571118, + "grad_norm": 11.98868629313521, + "learning_rate": 1.3107557274594607e-05, + "loss": 1.0492, + "step": 3845 + }, + { + "epoch": 2.0879478827361564, + "grad_norm": 9.734289423556795, + "learning_rate": 1.310421436245986e-05, + "loss": 0.6573, + "step": 3846 + }, + { + "epoch": 2.088490770901194, + "grad_norm": 10.196913105458549, + "learning_rate": 1.310087106639093e-05, + "loss": 0.8964, + "step": 3847 + }, + { + "epoch": 2.0890336590662324, + "grad_norm": 9.155222462665467, + "learning_rate": 1.3097527386801327e-05, + "loss": 0.6668, + "step": 3848 + }, + { + "epoch": 2.08957654723127, + "grad_norm": 8.512351137441192, + "learning_rate": 1.3094183324104602e-05, + "loss": 0.6084, + "step": 3849 + }, + { + "epoch": 2.0901194353963084, + "grad_norm": 7.62701351090082, + "learning_rate": 1.3090838878714349e-05, + "loss": 0.4199, + "step": 3850 + }, + { + "epoch": 2.090662323561346, + "grad_norm": 8.57862015121728, + "learning_rate": 1.3087494051044218e-05, + "loss": 0.3694, + "step": 3851 + }, + { + "epoch": 2.0912052117263844, + "grad_norm": 11.407572249271688, + "learning_rate": 1.3084148841507904e-05, + "loss": 0.8861, + "step": 3852 + }, + { + "epoch": 2.091748099891422, + "grad_norm": 9.297247466948287, + "learning_rate": 1.3080803250519142e-05, + "loss": 0.8432, + "step": 3853 + }, + { + "epoch": 2.0922909880564604, + "grad_norm": 7.909660106539073, + "learning_rate": 1.3077457278491728e-05, + "loss": 0.5103, + "step": 3854 + }, + { + "epoch": 2.092833876221498, + "grad_norm": 11.013011688454522, + "learning_rate": 1.3074110925839491e-05, + "loss": 1.0117, + "step": 3855 + }, + { + "epoch": 2.0933767643865364, + "grad_norm": 10.249008356283914, + "learning_rate": 1.3070764192976315e-05, + "loss": 0.5248, + "step": 3856 + }, + { + "epoch": 2.093919652551574, + "grad_norm": 8.240438304204865, + "learning_rate": 1.306741708031613e-05, + "loss": 0.6811, + "step": 3857 + }, + { + "epoch": 2.0944625407166124, + "grad_norm": 9.542880797186081, + "learning_rate": 1.3064069588272913e-05, + "loss": 0.4746, + "step": 3858 + }, + { + "epoch": 2.09500542888165, + "grad_norm": 9.251214264026915, + "learning_rate": 1.3060721717260685e-05, + "loss": 0.6405, + "step": 3859 + }, + { + "epoch": 2.0955483170466884, + "grad_norm": 12.366199541130733, + "learning_rate": 1.3057373467693515e-05, + "loss": 0.8378, + "step": 3860 + }, + { + "epoch": 2.096091205211726, + "grad_norm": 10.566968498265949, + "learning_rate": 1.3054024839985526e-05, + "loss": 0.5975, + "step": 3861 + }, + { + "epoch": 2.0966340933767644, + "grad_norm": 9.58134406953771, + "learning_rate": 1.3050675834550872e-05, + "loss": 0.6846, + "step": 3862 + }, + { + "epoch": 2.097176981541802, + "grad_norm": 8.437576527492594, + "learning_rate": 1.3047326451803772e-05, + "loss": 0.6301, + "step": 3863 + }, + { + "epoch": 2.0977198697068404, + "grad_norm": 8.68957421381677, + "learning_rate": 1.304397669215848e-05, + "loss": 0.6121, + "step": 3864 + }, + { + "epoch": 2.098262757871878, + "grad_norm": 12.281098753982102, + "learning_rate": 1.30406265560293e-05, + "loss": 1.174, + "step": 3865 + }, + { + "epoch": 2.0988056460369164, + "grad_norm": 12.64390360725308, + "learning_rate": 1.303727604383058e-05, + "loss": 0.727, + "step": 3866 + }, + { + "epoch": 2.099348534201954, + "grad_norm": 8.982715547555616, + "learning_rate": 1.3033925155976718e-05, + "loss": 0.4946, + "step": 3867 + }, + { + "epoch": 2.0998914223669924, + "grad_norm": 8.77924533690129, + "learning_rate": 1.303057389288216e-05, + "loss": 0.5027, + "step": 3868 + }, + { + "epoch": 2.1004343105320302, + "grad_norm": 12.246059108799978, + "learning_rate": 1.302722225496139e-05, + "loss": 0.6522, + "step": 3869 + }, + { + "epoch": 2.1009771986970684, + "grad_norm": 13.95610327170378, + "learning_rate": 1.3023870242628944e-05, + "loss": 1.043, + "step": 3870 + }, + { + "epoch": 2.1015200868621062, + "grad_norm": 10.355142745873641, + "learning_rate": 1.3020517856299413e-05, + "loss": 0.9167, + "step": 3871 + }, + { + "epoch": 2.1020629750271445, + "grad_norm": 8.43332484970558, + "learning_rate": 1.3017165096387419e-05, + "loss": 0.5952, + "step": 3872 + }, + { + "epoch": 2.1026058631921822, + "grad_norm": 9.224204311836482, + "learning_rate": 1.3013811963307634e-05, + "loss": 0.5631, + "step": 3873 + }, + { + "epoch": 2.1031487513572205, + "grad_norm": 9.719137418718985, + "learning_rate": 1.3010458457474784e-05, + "loss": 0.9293, + "step": 3874 + }, + { + "epoch": 2.1036916395222582, + "grad_norm": 8.234100299546002, + "learning_rate": 1.3007104579303636e-05, + "loss": 0.7041, + "step": 3875 + }, + { + "epoch": 2.1042345276872965, + "grad_norm": 11.0180135357339, + "learning_rate": 1.3003750329208995e-05, + "loss": 0.6068, + "step": 3876 + }, + { + "epoch": 2.1047774158523342, + "grad_norm": 8.790070650912016, + "learning_rate": 1.300039570760573e-05, + "loss": 0.4956, + "step": 3877 + }, + { + "epoch": 2.1053203040173725, + "grad_norm": 8.43754180724615, + "learning_rate": 1.2997040714908742e-05, + "loss": 0.3965, + "step": 3878 + }, + { + "epoch": 2.1058631921824102, + "grad_norm": 11.815520632897917, + "learning_rate": 1.299368535153298e-05, + "loss": 0.5963, + "step": 3879 + }, + { + "epoch": 2.1064060803474485, + "grad_norm": 14.168765746771772, + "learning_rate": 1.2990329617893445e-05, + "loss": 0.8111, + "step": 3880 + }, + { + "epoch": 2.1069489685124863, + "grad_norm": 5.706418305089391, + "learning_rate": 1.2986973514405176e-05, + "loss": 0.41, + "step": 3881 + }, + { + "epoch": 2.1074918566775245, + "grad_norm": 10.870432547178469, + "learning_rate": 1.2983617041483261e-05, + "loss": 0.7805, + "step": 3882 + }, + { + "epoch": 2.1080347448425623, + "grad_norm": 9.440648088759033, + "learning_rate": 1.2980260199542838e-05, + "loss": 0.6047, + "step": 3883 + }, + { + "epoch": 2.1085776330076005, + "grad_norm": 12.145529985557056, + "learning_rate": 1.2976902988999081e-05, + "loss": 0.8366, + "step": 3884 + }, + { + "epoch": 2.1091205211726383, + "grad_norm": 9.17886258135476, + "learning_rate": 1.2973545410267218e-05, + "loss": 0.6402, + "step": 3885 + }, + { + "epoch": 2.1096634093376765, + "grad_norm": 9.613840291682566, + "learning_rate": 1.297018746376252e-05, + "loss": 0.8175, + "step": 3886 + }, + { + "epoch": 2.1102062975027143, + "grad_norm": 12.898243756363112, + "learning_rate": 1.2966829149900304e-05, + "loss": 0.8514, + "step": 3887 + }, + { + "epoch": 2.1107491856677525, + "grad_norm": 8.825284004643853, + "learning_rate": 1.2963470469095928e-05, + "loss": 0.4508, + "step": 3888 + }, + { + "epoch": 2.1112920738327903, + "grad_norm": 9.953568896257451, + "learning_rate": 1.2960111421764803e-05, + "loss": 0.5657, + "step": 3889 + }, + { + "epoch": 2.1118349619978285, + "grad_norm": 8.988870757227582, + "learning_rate": 1.2956752008322378e-05, + "loss": 0.7242, + "step": 3890 + }, + { + "epoch": 2.1123778501628663, + "grad_norm": 9.68770492429417, + "learning_rate": 1.2953392229184156e-05, + "loss": 0.9403, + "step": 3891 + }, + { + "epoch": 2.1129207383279045, + "grad_norm": 11.903522492661367, + "learning_rate": 1.2950032084765674e-05, + "loss": 0.8743, + "step": 3892 + }, + { + "epoch": 2.1134636264929423, + "grad_norm": 12.24198259147124, + "learning_rate": 1.294667157548252e-05, + "loss": 0.9861, + "step": 3893 + }, + { + "epoch": 2.1140065146579805, + "grad_norm": 13.105368254945933, + "learning_rate": 1.2943310701750331e-05, + "loss": 0.8815, + "step": 3894 + }, + { + "epoch": 2.1145494028230183, + "grad_norm": 9.322992215038694, + "learning_rate": 1.2939949463984782e-05, + "loss": 0.4986, + "step": 3895 + }, + { + "epoch": 2.1150922909880565, + "grad_norm": 7.59330739725767, + "learning_rate": 1.29365878626016e-05, + "loss": 0.4343, + "step": 3896 + }, + { + "epoch": 2.1156351791530943, + "grad_norm": 9.782670203398764, + "learning_rate": 1.293322589801655e-05, + "loss": 0.4843, + "step": 3897 + }, + { + "epoch": 2.1161780673181325, + "grad_norm": 7.130500958222279, + "learning_rate": 1.2929863570645446e-05, + "loss": 0.3784, + "step": 3898 + }, + { + "epoch": 2.1167209554831703, + "grad_norm": 7.336707110452157, + "learning_rate": 1.2926500880904147e-05, + "loss": 0.4938, + "step": 3899 + }, + { + "epoch": 2.1172638436482085, + "grad_norm": 7.047478722817777, + "learning_rate": 1.2923137829208555e-05, + "loss": 0.3002, + "step": 3900 + }, + { + "epoch": 2.1178067318132463, + "grad_norm": 9.533093539252569, + "learning_rate": 1.2919774415974616e-05, + "loss": 0.711, + "step": 3901 + }, + { + "epoch": 2.1183496199782845, + "grad_norm": 8.468755965185016, + "learning_rate": 1.2916410641618324e-05, + "loss": 0.5491, + "step": 3902 + }, + { + "epoch": 2.1188925081433223, + "grad_norm": 10.351769959990103, + "learning_rate": 1.2913046506555715e-05, + "loss": 0.6269, + "step": 3903 + }, + { + "epoch": 2.1194353963083605, + "grad_norm": 10.2212813809041, + "learning_rate": 1.2909682011202875e-05, + "loss": 0.7249, + "step": 3904 + }, + { + "epoch": 2.1199782844733983, + "grad_norm": 9.344205468199693, + "learning_rate": 1.2906317155975922e-05, + "loss": 0.6277, + "step": 3905 + }, + { + "epoch": 2.1205211726384365, + "grad_norm": 8.704990474328149, + "learning_rate": 1.2902951941291035e-05, + "loss": 0.6317, + "step": 3906 + }, + { + "epoch": 2.1210640608034743, + "grad_norm": 11.644285794605782, + "learning_rate": 1.2899586367564422e-05, + "loss": 0.8585, + "step": 3907 + }, + { + "epoch": 2.1216069489685125, + "grad_norm": 8.462034849045626, + "learning_rate": 1.2896220435212347e-05, + "loss": 0.5174, + "step": 3908 + }, + { + "epoch": 2.1221498371335503, + "grad_norm": 10.023149647380059, + "learning_rate": 1.2892854144651112e-05, + "loss": 0.6044, + "step": 3909 + }, + { + "epoch": 2.1226927252985885, + "grad_norm": 9.437366874691593, + "learning_rate": 1.2889487496297068e-05, + "loss": 0.5081, + "step": 3910 + }, + { + "epoch": 2.1232356134636263, + "grad_norm": 11.06235798573597, + "learning_rate": 1.2886120490566607e-05, + "loss": 0.7385, + "step": 3911 + }, + { + "epoch": 2.1237785016286646, + "grad_norm": 9.45153797141243, + "learning_rate": 1.288275312787616e-05, + "loss": 0.6165, + "step": 3912 + }, + { + "epoch": 2.1243213897937023, + "grad_norm": 8.907808417684745, + "learning_rate": 1.2879385408642214e-05, + "loss": 0.3956, + "step": 3913 + }, + { + "epoch": 2.1248642779587406, + "grad_norm": 12.022235830361943, + "learning_rate": 1.2876017333281288e-05, + "loss": 0.7492, + "step": 3914 + }, + { + "epoch": 2.1254071661237783, + "grad_norm": 10.18229463165741, + "learning_rate": 1.2872648902209957e-05, + "loss": 1.0188, + "step": 3915 + }, + { + "epoch": 2.1259500542888166, + "grad_norm": 8.068860605172112, + "learning_rate": 1.2869280115844831e-05, + "loss": 0.4258, + "step": 3916 + }, + { + "epoch": 2.1264929424538543, + "grad_norm": 7.300141006448243, + "learning_rate": 1.2865910974602568e-05, + "loss": 0.414, + "step": 3917 + }, + { + "epoch": 2.1270358306188926, + "grad_norm": 9.447874959421453, + "learning_rate": 1.2862541478899868e-05, + "loss": 0.6323, + "step": 3918 + }, + { + "epoch": 2.1275787187839303, + "grad_norm": 11.845175678539553, + "learning_rate": 1.2859171629153476e-05, + "loss": 0.8539, + "step": 3919 + }, + { + "epoch": 2.1281216069489686, + "grad_norm": 8.29306868561595, + "learning_rate": 1.2855801425780179e-05, + "loss": 0.7127, + "step": 3920 + }, + { + "epoch": 2.1286644951140063, + "grad_norm": 8.040765832896788, + "learning_rate": 1.285243086919681e-05, + "loss": 0.495, + "step": 3921 + }, + { + "epoch": 2.1292073832790446, + "grad_norm": 13.042470321623425, + "learning_rate": 1.2849059959820248e-05, + "loss": 0.616, + "step": 3922 + }, + { + "epoch": 2.1297502714440824, + "grad_norm": 8.585317997471494, + "learning_rate": 1.2845688698067406e-05, + "loss": 0.5599, + "step": 3923 + }, + { + "epoch": 2.1302931596091206, + "grad_norm": 9.356896896219936, + "learning_rate": 1.2842317084355251e-05, + "loss": 0.605, + "step": 3924 + }, + { + "epoch": 2.1308360477741584, + "grad_norm": 17.057557558601207, + "learning_rate": 1.283894511910079e-05, + "loss": 0.9396, + "step": 3925 + }, + { + "epoch": 2.1313789359391966, + "grad_norm": 10.659876178353805, + "learning_rate": 1.2835572802721072e-05, + "loss": 0.765, + "step": 3926 + }, + { + "epoch": 2.1319218241042344, + "grad_norm": 9.903600273604221, + "learning_rate": 1.2832200135633191e-05, + "loss": 0.4745, + "step": 3927 + }, + { + "epoch": 2.1324647122692726, + "grad_norm": 12.075983241730247, + "learning_rate": 1.2828827118254279e-05, + "loss": 0.8613, + "step": 3928 + }, + { + "epoch": 2.1330076004343104, + "grad_norm": 10.13381566988187, + "learning_rate": 1.2825453751001526e-05, + "loss": 0.7342, + "step": 3929 + }, + { + "epoch": 2.1335504885993486, + "grad_norm": 12.671225163451464, + "learning_rate": 1.282208003429215e-05, + "loss": 0.7765, + "step": 3930 + }, + { + "epoch": 2.1340933767643864, + "grad_norm": 10.503393267034513, + "learning_rate": 1.2818705968543417e-05, + "loss": 0.7042, + "step": 3931 + }, + { + "epoch": 2.1346362649294246, + "grad_norm": 7.8720699552641635, + "learning_rate": 1.2815331554172634e-05, + "loss": 0.4393, + "step": 3932 + }, + { + "epoch": 2.1351791530944624, + "grad_norm": 14.057115653429742, + "learning_rate": 1.281195679159716e-05, + "loss": 0.8676, + "step": 3933 + }, + { + "epoch": 2.1357220412595006, + "grad_norm": 12.006814229853843, + "learning_rate": 1.2808581681234387e-05, + "loss": 0.8764, + "step": 3934 + }, + { + "epoch": 2.1362649294245384, + "grad_norm": 11.511042628784308, + "learning_rate": 1.2805206223501756e-05, + "loss": 0.6034, + "step": 3935 + }, + { + "epoch": 2.1368078175895766, + "grad_norm": 6.922027915468186, + "learning_rate": 1.2801830418816749e-05, + "loss": 0.4303, + "step": 3936 + }, + { + "epoch": 2.1373507057546144, + "grad_norm": 8.583211919241508, + "learning_rate": 1.2798454267596892e-05, + "loss": 0.6059, + "step": 3937 + }, + { + "epoch": 2.1378935939196526, + "grad_norm": 10.400540758946061, + "learning_rate": 1.2795077770259749e-05, + "loss": 0.7327, + "step": 3938 + }, + { + "epoch": 2.1384364820846904, + "grad_norm": 11.253848098269156, + "learning_rate": 1.2791700927222932e-05, + "loss": 0.7247, + "step": 3939 + }, + { + "epoch": 2.1389793702497286, + "grad_norm": 10.650957738366092, + "learning_rate": 1.2788323738904098e-05, + "loss": 0.8589, + "step": 3940 + }, + { + "epoch": 2.1395222584147664, + "grad_norm": 7.284559079353519, + "learning_rate": 1.2784946205720936e-05, + "loss": 0.3361, + "step": 3941 + }, + { + "epoch": 2.1400651465798046, + "grad_norm": 10.532878125436458, + "learning_rate": 1.2781568328091192e-05, + "loss": 0.8466, + "step": 3942 + }, + { + "epoch": 2.1406080347448424, + "grad_norm": 16.5717179039462, + "learning_rate": 1.2778190106432643e-05, + "loss": 1.1453, + "step": 3943 + }, + { + "epoch": 2.1411509229098806, + "grad_norm": 8.591001506273539, + "learning_rate": 1.2774811541163114e-05, + "loss": 0.5379, + "step": 3944 + }, + { + "epoch": 2.1416938110749184, + "grad_norm": 11.173192436567485, + "learning_rate": 1.2771432632700471e-05, + "loss": 1.0715, + "step": 3945 + }, + { + "epoch": 2.1422366992399566, + "grad_norm": 11.381968938636014, + "learning_rate": 1.2768053381462625e-05, + "loss": 0.8903, + "step": 3946 + }, + { + "epoch": 2.1427795874049944, + "grad_norm": 9.847481766271917, + "learning_rate": 1.2764673787867519e-05, + "loss": 0.5874, + "step": 3947 + }, + { + "epoch": 2.1433224755700326, + "grad_norm": 9.183549346287741, + "learning_rate": 1.2761293852333156e-05, + "loss": 0.6745, + "step": 3948 + }, + { + "epoch": 2.1438653637350704, + "grad_norm": 9.610890909648008, + "learning_rate": 1.2757913575277572e-05, + "loss": 0.9781, + "step": 3949 + }, + { + "epoch": 2.1444082519001086, + "grad_norm": 9.814885217738354, + "learning_rate": 1.275453295711884e-05, + "loss": 0.9977, + "step": 3950 + }, + { + "epoch": 2.1449511400651464, + "grad_norm": 10.984041832783662, + "learning_rate": 1.275115199827508e-05, + "loss": 0.7822, + "step": 3951 + }, + { + "epoch": 2.1454940282301846, + "grad_norm": 9.634528803708669, + "learning_rate": 1.2747770699164457e-05, + "loss": 1.1719, + "step": 3952 + }, + { + "epoch": 2.1460369163952224, + "grad_norm": 9.767304548901404, + "learning_rate": 1.2744389060205173e-05, + "loss": 0.6284, + "step": 3953 + }, + { + "epoch": 2.1465798045602607, + "grad_norm": 10.373642481495464, + "learning_rate": 1.2741007081815478e-05, + "loss": 0.6742, + "step": 3954 + }, + { + "epoch": 2.1471226927252984, + "grad_norm": 10.390187789918086, + "learning_rate": 1.2737624764413659e-05, + "loss": 0.6647, + "step": 3955 + }, + { + "epoch": 2.1476655808903367, + "grad_norm": 10.617840626931814, + "learning_rate": 1.2734242108418044e-05, + "loss": 0.877, + "step": 3956 + }, + { + "epoch": 2.1482084690553744, + "grad_norm": 9.898468473630299, + "learning_rate": 1.273085911424701e-05, + "loss": 0.7335, + "step": 3957 + }, + { + "epoch": 2.1487513572204127, + "grad_norm": 10.409230729626815, + "learning_rate": 1.2727475782318966e-05, + "loss": 0.7169, + "step": 3958 + }, + { + "epoch": 2.1492942453854504, + "grad_norm": 10.425915655294723, + "learning_rate": 1.2724092113052372e-05, + "loss": 0.6578, + "step": 3959 + }, + { + "epoch": 2.1498371335504887, + "grad_norm": 6.3375473265246995, + "learning_rate": 1.2720708106865722e-05, + "loss": 0.3471, + "step": 3960 + }, + { + "epoch": 2.1503800217155264, + "grad_norm": 9.396081485808933, + "learning_rate": 1.2717323764177559e-05, + "loss": 0.5232, + "step": 3961 + }, + { + "epoch": 2.1509229098805647, + "grad_norm": 13.419504298012779, + "learning_rate": 1.2713939085406461e-05, + "loss": 1.8096, + "step": 3962 + }, + { + "epoch": 2.1514657980456025, + "grad_norm": 9.635837991132963, + "learning_rate": 1.2710554070971053e-05, + "loss": 0.5949, + "step": 3963 + }, + { + "epoch": 2.1520086862106407, + "grad_norm": 10.111061783908927, + "learning_rate": 1.2707168721289995e-05, + "loss": 0.8731, + "step": 3964 + }, + { + "epoch": 2.1525515743756785, + "grad_norm": 8.256139809775819, + "learning_rate": 1.2703783036781998e-05, + "loss": 0.726, + "step": 3965 + }, + { + "epoch": 2.1530944625407167, + "grad_norm": 8.385511442711055, + "learning_rate": 1.2700397017865802e-05, + "loss": 0.6209, + "step": 3966 + }, + { + "epoch": 2.1536373507057545, + "grad_norm": 8.556891915186654, + "learning_rate": 1.2697010664960201e-05, + "loss": 0.6746, + "step": 3967 + }, + { + "epoch": 2.1541802388707927, + "grad_norm": 8.219611855039698, + "learning_rate": 1.2693623978484022e-05, + "loss": 0.7542, + "step": 3968 + }, + { + "epoch": 2.1547231270358305, + "grad_norm": 9.425031325511556, + "learning_rate": 1.2690236958856139e-05, + "loss": 0.6735, + "step": 3969 + }, + { + "epoch": 2.1552660152008687, + "grad_norm": 9.21346190912602, + "learning_rate": 1.2686849606495461e-05, + "loss": 0.5311, + "step": 3970 + }, + { + "epoch": 2.1558089033659065, + "grad_norm": 12.261438418537729, + "learning_rate": 1.268346192182094e-05, + "loss": 0.8556, + "step": 3971 + }, + { + "epoch": 2.1563517915309447, + "grad_norm": 8.630907790930149, + "learning_rate": 1.268007390525157e-05, + "loss": 0.7789, + "step": 3972 + }, + { + "epoch": 2.1568946796959825, + "grad_norm": 10.654367663061162, + "learning_rate": 1.267668555720639e-05, + "loss": 1.1708, + "step": 3973 + }, + { + "epoch": 2.1574375678610207, + "grad_norm": 15.113883332023352, + "learning_rate": 1.2673296878104472e-05, + "loss": 1.2273, + "step": 3974 + }, + { + "epoch": 2.1579804560260585, + "grad_norm": 10.205014286438582, + "learning_rate": 1.266990786836494e-05, + "loss": 1.2711, + "step": 3975 + }, + { + "epoch": 2.1585233441910967, + "grad_norm": 8.32550465973343, + "learning_rate": 1.2666518528406944e-05, + "loss": 0.6387, + "step": 3976 + }, + { + "epoch": 2.1590662323561345, + "grad_norm": 8.80250087611585, + "learning_rate": 1.2663128858649686e-05, + "loss": 0.5669, + "step": 3977 + }, + { + "epoch": 2.1596091205211727, + "grad_norm": 7.977255679001254, + "learning_rate": 1.2659738859512411e-05, + "loss": 0.4443, + "step": 3978 + }, + { + "epoch": 2.1601520086862105, + "grad_norm": 6.086622373481402, + "learning_rate": 1.265634853141439e-05, + "loss": 0.3029, + "step": 3979 + }, + { + "epoch": 2.1606948968512487, + "grad_norm": 16.46377636281654, + "learning_rate": 1.2652957874774953e-05, + "loss": 1.09, + "step": 3980 + }, + { + "epoch": 2.1612377850162865, + "grad_norm": 10.805039141935024, + "learning_rate": 1.2649566890013456e-05, + "loss": 0.672, + "step": 3981 + }, + { + "epoch": 2.1617806731813247, + "grad_norm": 10.731407400272554, + "learning_rate": 1.2646175577549303e-05, + "loss": 0.6618, + "step": 3982 + }, + { + "epoch": 2.1623235613463625, + "grad_norm": 9.572301578766334, + "learning_rate": 1.2642783937801937e-05, + "loss": 0.6608, + "step": 3983 + }, + { + "epoch": 2.1628664495114007, + "grad_norm": 8.415534879772189, + "learning_rate": 1.2639391971190842e-05, + "loss": 0.5314, + "step": 3984 + }, + { + "epoch": 2.1634093376764385, + "grad_norm": 8.290688286613944, + "learning_rate": 1.263599967813554e-05, + "loss": 0.6736, + "step": 3985 + }, + { + "epoch": 2.1639522258414767, + "grad_norm": 8.92112434155926, + "learning_rate": 1.2632607059055597e-05, + "loss": 0.5572, + "step": 3986 + }, + { + "epoch": 2.1644951140065145, + "grad_norm": 9.82202342503724, + "learning_rate": 1.2629214114370618e-05, + "loss": 0.7838, + "step": 3987 + }, + { + "epoch": 2.1650380021715527, + "grad_norm": 7.20619661733841, + "learning_rate": 1.262582084450025e-05, + "loss": 0.5577, + "step": 3988 + }, + { + "epoch": 2.1655808903365905, + "grad_norm": 8.640482408220107, + "learning_rate": 1.2622427249864172e-05, + "loss": 0.819, + "step": 3989 + }, + { + "epoch": 2.1661237785016287, + "grad_norm": 9.245400712044432, + "learning_rate": 1.2619033330882114e-05, + "loss": 0.6009, + "step": 3990 + }, + { + "epoch": 2.1666666666666665, + "grad_norm": 8.891547209594252, + "learning_rate": 1.2615639087973835e-05, + "loss": 0.6085, + "step": 3991 + }, + { + "epoch": 2.1672095548317047, + "grad_norm": 12.196798951360252, + "learning_rate": 1.2612244521559148e-05, + "loss": 0.745, + "step": 3992 + }, + { + "epoch": 2.1677524429967425, + "grad_norm": 9.610928616001752, + "learning_rate": 1.2608849632057895e-05, + "loss": 0.7709, + "step": 3993 + }, + { + "epoch": 2.1682953311617807, + "grad_norm": 9.454740121591225, + "learning_rate": 1.2605454419889962e-05, + "loss": 0.6455, + "step": 3994 + }, + { + "epoch": 2.1688382193268185, + "grad_norm": 13.245987794421069, + "learning_rate": 1.2602058885475273e-05, + "loss": 0.7104, + "step": 3995 + }, + { + "epoch": 2.1693811074918568, + "grad_norm": 10.51068899063252, + "learning_rate": 1.2598663029233794e-05, + "loss": 0.5747, + "step": 3996 + }, + { + "epoch": 2.1699239956568945, + "grad_norm": 10.589584490390159, + "learning_rate": 1.2595266851585532e-05, + "loss": 0.7878, + "step": 3997 + }, + { + "epoch": 2.1704668838219328, + "grad_norm": 7.818930212111502, + "learning_rate": 1.2591870352950528e-05, + "loss": 0.5179, + "step": 3998 + }, + { + "epoch": 2.1710097719869705, + "grad_norm": 8.222376749362372, + "learning_rate": 1.2588473533748868e-05, + "loss": 0.4722, + "step": 3999 + }, + { + "epoch": 2.1715526601520088, + "grad_norm": 12.451645756942625, + "learning_rate": 1.2585076394400675e-05, + "loss": 0.7742, + "step": 4000 + }, + { + "epoch": 2.1720955483170465, + "grad_norm": 11.33368393512842, + "learning_rate": 1.2581678935326117e-05, + "loss": 0.3945, + "step": 4001 + }, + { + "epoch": 2.1726384364820848, + "grad_norm": 10.914311466768305, + "learning_rate": 1.2578281156945389e-05, + "loss": 0.6345, + "step": 4002 + }, + { + "epoch": 2.1731813246471225, + "grad_norm": 12.867661375339019, + "learning_rate": 1.2574883059678742e-05, + "loss": 1.1569, + "step": 4003 + }, + { + "epoch": 2.1737242128121608, + "grad_norm": 11.4420969393128, + "learning_rate": 1.2571484643946452e-05, + "loss": 0.4927, + "step": 4004 + }, + { + "epoch": 2.1742671009771986, + "grad_norm": 10.234468502706436, + "learning_rate": 1.2568085910168842e-05, + "loss": 0.463, + "step": 4005 + }, + { + "epoch": 2.1748099891422368, + "grad_norm": 7.649699643480241, + "learning_rate": 1.2564686858766275e-05, + "loss": 0.3981, + "step": 4006 + }, + { + "epoch": 2.1753528773072746, + "grad_norm": 9.503268181536553, + "learning_rate": 1.2561287490159151e-05, + "loss": 0.4162, + "step": 4007 + }, + { + "epoch": 2.175895765472313, + "grad_norm": 8.81726258724681, + "learning_rate": 1.2557887804767907e-05, + "loss": 0.92, + "step": 4008 + }, + { + "epoch": 2.1764386536373506, + "grad_norm": 9.625823227183314, + "learning_rate": 1.2554487803013027e-05, + "loss": 0.7527, + "step": 4009 + }, + { + "epoch": 2.176981541802389, + "grad_norm": 11.452435192265499, + "learning_rate": 1.2551087485315017e-05, + "loss": 0.8611, + "step": 4010 + }, + { + "epoch": 2.1775244299674266, + "grad_norm": 9.123700040373722, + "learning_rate": 1.2547686852094445e-05, + "loss": 0.475, + "step": 4011 + }, + { + "epoch": 2.178067318132465, + "grad_norm": 9.306092992521913, + "learning_rate": 1.2544285903771902e-05, + "loss": 0.6063, + "step": 4012 + }, + { + "epoch": 2.1786102062975026, + "grad_norm": 14.283804972037375, + "learning_rate": 1.2540884640768022e-05, + "loss": 0.9142, + "step": 4013 + }, + { + "epoch": 2.179153094462541, + "grad_norm": 14.299740055929474, + "learning_rate": 1.2537483063503483e-05, + "loss": 0.8124, + "step": 4014 + }, + { + "epoch": 2.1796959826275786, + "grad_norm": 9.637221226214642, + "learning_rate": 1.2534081172398993e-05, + "loss": 0.6278, + "step": 4015 + }, + { + "epoch": 2.180238870792617, + "grad_norm": 6.674062415604907, + "learning_rate": 1.2530678967875304e-05, + "loss": 0.5078, + "step": 4016 + }, + { + "epoch": 2.1807817589576546, + "grad_norm": 11.68250334366375, + "learning_rate": 1.252727645035321e-05, + "loss": 0.7393, + "step": 4017 + }, + { + "epoch": 2.181324647122693, + "grad_norm": 9.642668530719417, + "learning_rate": 1.2523873620253535e-05, + "loss": 0.6218, + "step": 4018 + }, + { + "epoch": 2.1818675352877306, + "grad_norm": 9.955103488106618, + "learning_rate": 1.2520470477997146e-05, + "loss": 0.899, + "step": 4019 + }, + { + "epoch": 2.182410423452769, + "grad_norm": 13.184397199493825, + "learning_rate": 1.2517067024004955e-05, + "loss": 0.8437, + "step": 4020 + }, + { + "epoch": 2.1829533116178066, + "grad_norm": 11.142111068408965, + "learning_rate": 1.2513663258697901e-05, + "loss": 0.6064, + "step": 4021 + }, + { + "epoch": 2.183496199782845, + "grad_norm": 9.336503366824324, + "learning_rate": 1.251025918249697e-05, + "loss": 0.6649, + "step": 4022 + }, + { + "epoch": 2.1840390879478826, + "grad_norm": 11.35476662379509, + "learning_rate": 1.250685479582318e-05, + "loss": 0.8573, + "step": 4023 + }, + { + "epoch": 2.184581976112921, + "grad_norm": 9.049806197407708, + "learning_rate": 1.2503450099097594e-05, + "loss": 0.3839, + "step": 4024 + }, + { + "epoch": 2.1851248642779586, + "grad_norm": 12.739046308083323, + "learning_rate": 1.250004509274131e-05, + "loss": 0.9344, + "step": 4025 + }, + { + "epoch": 2.185667752442997, + "grad_norm": 7.539200963039068, + "learning_rate": 1.2496639777175465e-05, + "loss": 0.3691, + "step": 4026 + }, + { + "epoch": 2.1862106406080346, + "grad_norm": 10.659038181144831, + "learning_rate": 1.2493234152821234e-05, + "loss": 0.9462, + "step": 4027 + }, + { + "epoch": 2.186753528773073, + "grad_norm": 8.047186696422612, + "learning_rate": 1.2489828220099831e-05, + "loss": 0.4974, + "step": 4028 + }, + { + "epoch": 2.1872964169381106, + "grad_norm": 11.048564474051437, + "learning_rate": 1.2486421979432503e-05, + "loss": 0.6221, + "step": 4029 + }, + { + "epoch": 2.187839305103149, + "grad_norm": 9.808559304060044, + "learning_rate": 1.2483015431240542e-05, + "loss": 0.5221, + "step": 4030 + }, + { + "epoch": 2.1883821932681866, + "grad_norm": 7.720481766274275, + "learning_rate": 1.2479608575945274e-05, + "loss": 0.5248, + "step": 4031 + }, + { + "epoch": 2.188925081433225, + "grad_norm": 7.791803772948659, + "learning_rate": 1.2476201413968068e-05, + "loss": 0.5313, + "step": 4032 + }, + { + "epoch": 2.1894679695982626, + "grad_norm": 8.280464830936529, + "learning_rate": 1.2472793945730323e-05, + "loss": 0.5155, + "step": 4033 + }, + { + "epoch": 2.190010857763301, + "grad_norm": 6.500045291722395, + "learning_rate": 1.2469386171653483e-05, + "loss": 0.6057, + "step": 4034 + }, + { + "epoch": 2.1905537459283386, + "grad_norm": 6.735518823264981, + "learning_rate": 1.2465978092159025e-05, + "loss": 0.5207, + "step": 4035 + }, + { + "epoch": 2.191096634093377, + "grad_norm": 12.977966385021237, + "learning_rate": 1.2462569707668468e-05, + "loss": 1.0472, + "step": 4036 + }, + { + "epoch": 2.1916395222584146, + "grad_norm": 12.805295393279883, + "learning_rate": 1.245916101860336e-05, + "loss": 0.7486, + "step": 4037 + }, + { + "epoch": 2.192182410423453, + "grad_norm": 11.04531935760714, + "learning_rate": 1.2455752025385304e-05, + "loss": 0.6107, + "step": 4038 + }, + { + "epoch": 2.1927252985884906, + "grad_norm": 12.12096270238078, + "learning_rate": 1.245234272843592e-05, + "loss": 0.6264, + "step": 4039 + }, + { + "epoch": 2.193268186753529, + "grad_norm": 14.202650747934204, + "learning_rate": 1.244893312817688e-05, + "loss": 0.7965, + "step": 4040 + }, + { + "epoch": 2.1938110749185666, + "grad_norm": 10.409453866661186, + "learning_rate": 1.2445523225029887e-05, + "loss": 0.6012, + "step": 4041 + }, + { + "epoch": 2.194353963083605, + "grad_norm": 8.68329891299571, + "learning_rate": 1.2442113019416683e-05, + "loss": 0.4393, + "step": 4042 + }, + { + "epoch": 2.1948968512486426, + "grad_norm": 11.77894618341781, + "learning_rate": 1.2438702511759049e-05, + "loss": 0.7838, + "step": 4043 + }, + { + "epoch": 2.195439739413681, + "grad_norm": 8.31374036003515, + "learning_rate": 1.2435291702478802e-05, + "loss": 0.6043, + "step": 4044 + }, + { + "epoch": 2.1959826275787186, + "grad_norm": 10.219603175661236, + "learning_rate": 1.2431880591997799e-05, + "loss": 0.9292, + "step": 4045 + }, + { + "epoch": 2.196525515743757, + "grad_norm": 10.662478372576002, + "learning_rate": 1.2428469180737923e-05, + "loss": 0.5951, + "step": 4046 + }, + { + "epoch": 2.1970684039087947, + "grad_norm": 15.414471746340716, + "learning_rate": 1.2425057469121113e-05, + "loss": 1.0091, + "step": 4047 + }, + { + "epoch": 2.197611292073833, + "grad_norm": 8.553095532374297, + "learning_rate": 1.242164545756933e-05, + "loss": 0.4494, + "step": 4048 + }, + { + "epoch": 2.1981541802388707, + "grad_norm": 10.520590758759942, + "learning_rate": 1.2418233146504575e-05, + "loss": 0.5277, + "step": 4049 + }, + { + "epoch": 2.198697068403909, + "grad_norm": 10.792986797196212, + "learning_rate": 1.241482053634889e-05, + "loss": 0.7036, + "step": 4050 + }, + { + "epoch": 2.1992399565689467, + "grad_norm": 10.425002068049416, + "learning_rate": 1.2411407627524354e-05, + "loss": 0.8283, + "step": 4051 + }, + { + "epoch": 2.199782844733985, + "grad_norm": 10.421667203553737, + "learning_rate": 1.240799442045308e-05, + "loss": 0.8689, + "step": 4052 + }, + { + "epoch": 2.2003257328990227, + "grad_norm": 9.866413466645723, + "learning_rate": 1.2404580915557217e-05, + "loss": 0.517, + "step": 4053 + }, + { + "epoch": 2.200868621064061, + "grad_norm": 9.806755744401848, + "learning_rate": 1.2401167113258954e-05, + "loss": 0.7176, + "step": 4054 + }, + { + "epoch": 2.2014115092290987, + "grad_norm": 12.103200009209074, + "learning_rate": 1.2397753013980516e-05, + "loss": 0.8563, + "step": 4055 + }, + { + "epoch": 2.201954397394137, + "grad_norm": 9.310089002310463, + "learning_rate": 1.2394338618144162e-05, + "loss": 0.5478, + "step": 4056 + }, + { + "epoch": 2.2024972855591747, + "grad_norm": 8.147499111803365, + "learning_rate": 1.2390923926172194e-05, + "loss": 0.4616, + "step": 4057 + }, + { + "epoch": 2.203040173724213, + "grad_norm": 10.363627124726912, + "learning_rate": 1.2387508938486945e-05, + "loss": 0.6039, + "step": 4058 + }, + { + "epoch": 2.2035830618892507, + "grad_norm": 11.09580557172493, + "learning_rate": 1.2384093655510785e-05, + "loss": 0.6772, + "step": 4059 + }, + { + "epoch": 2.204125950054289, + "grad_norm": 10.762880100606296, + "learning_rate": 1.238067807766612e-05, + "loss": 0.8104, + "step": 4060 + }, + { + "epoch": 2.2046688382193267, + "grad_norm": 10.590539714326667, + "learning_rate": 1.2377262205375398e-05, + "loss": 0.6766, + "step": 4061 + }, + { + "epoch": 2.205211726384365, + "grad_norm": 10.060689580140192, + "learning_rate": 1.2373846039061095e-05, + "loss": 1.0282, + "step": 4062 + }, + { + "epoch": 2.2057546145494027, + "grad_norm": 7.612672548635284, + "learning_rate": 1.237042957914573e-05, + "loss": 0.449, + "step": 4063 + }, + { + "epoch": 2.206297502714441, + "grad_norm": 12.239451929254052, + "learning_rate": 1.2367012826051861e-05, + "loss": 0.5385, + "step": 4064 + }, + { + "epoch": 2.2068403908794787, + "grad_norm": 14.199356139948698, + "learning_rate": 1.236359578020207e-05, + "loss": 0.7867, + "step": 4065 + }, + { + "epoch": 2.207383279044517, + "grad_norm": 8.357180617953833, + "learning_rate": 1.2360178442018989e-05, + "loss": 0.6803, + "step": 4066 + }, + { + "epoch": 2.2079261672095547, + "grad_norm": 9.203584379654275, + "learning_rate": 1.2356760811925277e-05, + "loss": 0.677, + "step": 4067 + }, + { + "epoch": 2.208469055374593, + "grad_norm": 8.063224350440471, + "learning_rate": 1.2353342890343626e-05, + "loss": 0.644, + "step": 4068 + }, + { + "epoch": 2.2090119435396307, + "grad_norm": 9.352085551954278, + "learning_rate": 1.2349924677696781e-05, + "loss": 0.9888, + "step": 4069 + }, + { + "epoch": 2.209554831704669, + "grad_norm": 10.54696840423902, + "learning_rate": 1.2346506174407505e-05, + "loss": 0.5895, + "step": 4070 + }, + { + "epoch": 2.2100977198697067, + "grad_norm": 8.568183164428444, + "learning_rate": 1.2343087380898604e-05, + "loss": 0.6287, + "step": 4071 + }, + { + "epoch": 2.210640608034745, + "grad_norm": 11.191021916852304, + "learning_rate": 1.2339668297592924e-05, + "loss": 0.8216, + "step": 4072 + }, + { + "epoch": 2.2111834961997827, + "grad_norm": 7.67795379285754, + "learning_rate": 1.233624892491334e-05, + "loss": 0.423, + "step": 4073 + }, + { + "epoch": 2.211726384364821, + "grad_norm": 8.108373133947762, + "learning_rate": 1.2332829263282764e-05, + "loss": 0.6548, + "step": 4074 + }, + { + "epoch": 2.2122692725298587, + "grad_norm": 9.0962841262343, + "learning_rate": 1.2329409313124146e-05, + "loss": 0.5506, + "step": 4075 + }, + { + "epoch": 2.212812160694897, + "grad_norm": 8.85949097192624, + "learning_rate": 1.2325989074860472e-05, + "loss": 0.5258, + "step": 4076 + }, + { + "epoch": 2.2133550488599347, + "grad_norm": 7.051914022367066, + "learning_rate": 1.2322568548914764e-05, + "loss": 0.5242, + "step": 4077 + }, + { + "epoch": 2.213897937024973, + "grad_norm": 8.287257693111446, + "learning_rate": 1.2319147735710076e-05, + "loss": 0.4769, + "step": 4078 + }, + { + "epoch": 2.2144408251900107, + "grad_norm": 9.473240892632637, + "learning_rate": 1.2315726635669498e-05, + "loss": 1.0853, + "step": 4079 + }, + { + "epoch": 2.214983713355049, + "grad_norm": 10.460874830183508, + "learning_rate": 1.231230524921616e-05, + "loss": 0.8907, + "step": 4080 + }, + { + "epoch": 2.2155266015200867, + "grad_norm": 9.17618869023913, + "learning_rate": 1.2308883576773223e-05, + "loss": 0.6355, + "step": 4081 + }, + { + "epoch": 2.216069489685125, + "grad_norm": 10.389188046047401, + "learning_rate": 1.2305461618763886e-05, + "loss": 0.9455, + "step": 4082 + }, + { + "epoch": 2.2166123778501627, + "grad_norm": 8.41220072964392, + "learning_rate": 1.230203937561138e-05, + "loss": 0.5574, + "step": 4083 + }, + { + "epoch": 2.217155266015201, + "grad_norm": 8.193062738430084, + "learning_rate": 1.2298616847738978e-05, + "loss": 0.548, + "step": 4084 + }, + { + "epoch": 2.2176981541802387, + "grad_norm": 8.55082220874788, + "learning_rate": 1.2295194035569979e-05, + "loss": 0.6087, + "step": 4085 + }, + { + "epoch": 2.218241042345277, + "grad_norm": 10.482752217531827, + "learning_rate": 1.2291770939527725e-05, + "loss": 0.5404, + "step": 4086 + }, + { + "epoch": 2.2187839305103148, + "grad_norm": 9.786686689665721, + "learning_rate": 1.2288347560035587e-05, + "loss": 0.6795, + "step": 4087 + }, + { + "epoch": 2.219326818675353, + "grad_norm": 9.696178175990593, + "learning_rate": 1.2284923897516978e-05, + "loss": 1.2474, + "step": 4088 + }, + { + "epoch": 2.2198697068403908, + "grad_norm": 9.135935376096977, + "learning_rate": 1.2281499952395336e-05, + "loss": 0.6746, + "step": 4089 + }, + { + "epoch": 2.220412595005429, + "grad_norm": 10.341327194770198, + "learning_rate": 1.2278075725094147e-05, + "loss": 0.6501, + "step": 4090 + }, + { + "epoch": 2.2209554831704668, + "grad_norm": 11.750926749777816, + "learning_rate": 1.2274651216036921e-05, + "loss": 0.6421, + "step": 4091 + }, + { + "epoch": 2.221498371335505, + "grad_norm": 15.264249369051182, + "learning_rate": 1.2271226425647207e-05, + "loss": 0.9053, + "step": 4092 + }, + { + "epoch": 2.2220412595005428, + "grad_norm": 10.996946312108877, + "learning_rate": 1.2267801354348589e-05, + "loss": 0.6776, + "step": 4093 + }, + { + "epoch": 2.222584147665581, + "grad_norm": 9.01852590664764, + "learning_rate": 1.2264376002564687e-05, + "loss": 0.7342, + "step": 4094 + }, + { + "epoch": 2.2231270358306188, + "grad_norm": 9.776717004980412, + "learning_rate": 1.2260950370719146e-05, + "loss": 0.7353, + "step": 4095 + }, + { + "epoch": 2.223669923995657, + "grad_norm": 12.928031483014212, + "learning_rate": 1.2257524459235666e-05, + "loss": 0.8649, + "step": 4096 + }, + { + "epoch": 2.2242128121606948, + "grad_norm": 9.549545586005916, + "learning_rate": 1.2254098268537962e-05, + "loss": 0.8056, + "step": 4097 + }, + { + "epoch": 2.224755700325733, + "grad_norm": 10.729492841694832, + "learning_rate": 1.2250671799049791e-05, + "loss": 0.5817, + "step": 4098 + }, + { + "epoch": 2.225298588490771, + "grad_norm": 14.13032585380651, + "learning_rate": 1.2247245051194944e-05, + "loss": 0.5998, + "step": 4099 + }, + { + "epoch": 2.225841476655809, + "grad_norm": 9.143101156244217, + "learning_rate": 1.2243818025397247e-05, + "loss": 0.5633, + "step": 4100 + }, + { + "epoch": 2.226384364820847, + "grad_norm": 9.010721857216959, + "learning_rate": 1.224039072208056e-05, + "loss": 0.5855, + "step": 4101 + }, + { + "epoch": 2.226927252985885, + "grad_norm": 11.257457693848512, + "learning_rate": 1.2236963141668778e-05, + "loss": 0.5672, + "step": 4102 + }, + { + "epoch": 2.227470141150923, + "grad_norm": 9.238231248800124, + "learning_rate": 1.2233535284585831e-05, + "loss": 0.4633, + "step": 4103 + }, + { + "epoch": 2.228013029315961, + "grad_norm": 8.93957445803232, + "learning_rate": 1.223010715125568e-05, + "loss": 0.8019, + "step": 4104 + }, + { + "epoch": 2.228555917480999, + "grad_norm": 10.869791394117428, + "learning_rate": 1.2226678742102322e-05, + "loss": 0.4235, + "step": 4105 + }, + { + "epoch": 2.229098805646037, + "grad_norm": 11.71960282550904, + "learning_rate": 1.2223250057549789e-05, + "loss": 0.8825, + "step": 4106 + }, + { + "epoch": 2.229641693811075, + "grad_norm": 9.133591574750724, + "learning_rate": 1.2219821098022146e-05, + "loss": 0.456, + "step": 4107 + }, + { + "epoch": 2.230184581976113, + "grad_norm": 7.713253727965175, + "learning_rate": 1.2216391863943492e-05, + "loss": 0.407, + "step": 4108 + }, + { + "epoch": 2.230727470141151, + "grad_norm": 8.351720726424757, + "learning_rate": 1.2212962355737959e-05, + "loss": 0.6003, + "step": 4109 + }, + { + "epoch": 2.231270358306189, + "grad_norm": 12.012727761361187, + "learning_rate": 1.2209532573829714e-05, + "loss": 0.731, + "step": 4110 + }, + { + "epoch": 2.231813246471227, + "grad_norm": 12.902339895008394, + "learning_rate": 1.220610251864296e-05, + "loss": 0.9873, + "step": 4111 + }, + { + "epoch": 2.232356134636265, + "grad_norm": 8.935941303438996, + "learning_rate": 1.220267219060193e-05, + "loss": 0.7549, + "step": 4112 + }, + { + "epoch": 2.232899022801303, + "grad_norm": 9.69521468611751, + "learning_rate": 1.2199241590130892e-05, + "loss": 0.6099, + "step": 4113 + }, + { + "epoch": 2.233441910966341, + "grad_norm": 7.437030925496973, + "learning_rate": 1.2195810717654148e-05, + "loss": 0.4743, + "step": 4114 + }, + { + "epoch": 2.233984799131379, + "grad_norm": 7.435823614079823, + "learning_rate": 1.2192379573596036e-05, + "loss": 0.5019, + "step": 4115 + }, + { + "epoch": 2.234527687296417, + "grad_norm": 7.563032718583443, + "learning_rate": 1.2188948158380927e-05, + "loss": 0.7176, + "step": 4116 + }, + { + "epoch": 2.235070575461455, + "grad_norm": 10.449779515276235, + "learning_rate": 1.2185516472433218e-05, + "loss": 0.7921, + "step": 4117 + }, + { + "epoch": 2.235613463626493, + "grad_norm": 9.088046167756731, + "learning_rate": 1.2182084516177351e-05, + "loss": 0.811, + "step": 4118 + }, + { + "epoch": 2.236156351791531, + "grad_norm": 9.664269830378425, + "learning_rate": 1.217865229003779e-05, + "loss": 0.8024, + "step": 4119 + }, + { + "epoch": 2.236699239956569, + "grad_norm": 8.055354430405984, + "learning_rate": 1.2175219794439044e-05, + "loss": 0.5747, + "step": 4120 + }, + { + "epoch": 2.237242128121607, + "grad_norm": 6.428342674450092, + "learning_rate": 1.2171787029805646e-05, + "loss": 0.5673, + "step": 4121 + }, + { + "epoch": 2.237785016286645, + "grad_norm": 7.502074776894542, + "learning_rate": 1.2168353996562167e-05, + "loss": 0.566, + "step": 4122 + }, + { + "epoch": 2.238327904451683, + "grad_norm": 10.380110541434405, + "learning_rate": 1.2164920695133212e-05, + "loss": 1.0609, + "step": 4123 + }, + { + "epoch": 2.238870792616721, + "grad_norm": 13.199317832444493, + "learning_rate": 1.2161487125943413e-05, + "loss": 0.7662, + "step": 4124 + }, + { + "epoch": 2.239413680781759, + "grad_norm": 9.067564902950002, + "learning_rate": 1.2158053289417439e-05, + "loss": 0.6325, + "step": 4125 + }, + { + "epoch": 2.239956568946797, + "grad_norm": 8.074724230493178, + "learning_rate": 1.215461918598e-05, + "loss": 0.6854, + "step": 4126 + }, + { + "epoch": 2.240499457111835, + "grad_norm": 7.916509466850638, + "learning_rate": 1.2151184816055822e-05, + "loss": 0.3852, + "step": 4127 + }, + { + "epoch": 2.241042345276873, + "grad_norm": 8.647089691838135, + "learning_rate": 1.214775018006968e-05, + "loss": 0.6291, + "step": 4128 + }, + { + "epoch": 2.241585233441911, + "grad_norm": 9.238872667328287, + "learning_rate": 1.214431527844637e-05, + "loss": 0.6117, + "step": 4129 + }, + { + "epoch": 2.242128121606949, + "grad_norm": 12.537117099920934, + "learning_rate": 1.214088011161073e-05, + "loss": 0.745, + "step": 4130 + }, + { + "epoch": 2.242671009771987, + "grad_norm": 9.99529796933279, + "learning_rate": 1.2137444679987627e-05, + "loss": 0.612, + "step": 4131 + }, + { + "epoch": 2.243213897937025, + "grad_norm": 6.673816908475189, + "learning_rate": 1.2134008984001959e-05, + "loss": 0.3219, + "step": 4132 + }, + { + "epoch": 2.243756786102063, + "grad_norm": 13.936566760747976, + "learning_rate": 1.2130573024078656e-05, + "loss": 0.8689, + "step": 4133 + }, + { + "epoch": 2.244299674267101, + "grad_norm": 7.800578548383095, + "learning_rate": 1.2127136800642689e-05, + "loss": 0.3804, + "step": 4134 + }, + { + "epoch": 2.244842562432139, + "grad_norm": 9.283785411998215, + "learning_rate": 1.2123700314119052e-05, + "loss": 0.699, + "step": 4135 + }, + { + "epoch": 2.245385450597177, + "grad_norm": 8.613596910195628, + "learning_rate": 1.2120263564932775e-05, + "loss": 0.7155, + "step": 4136 + }, + { + "epoch": 2.245928338762215, + "grad_norm": 8.522662049603209, + "learning_rate": 1.2116826553508923e-05, + "loss": 0.5705, + "step": 4137 + }, + { + "epoch": 2.246471226927253, + "grad_norm": 9.023598184433123, + "learning_rate": 1.2113389280272587e-05, + "loss": 0.4392, + "step": 4138 + }, + { + "epoch": 2.247014115092291, + "grad_norm": 10.959276458237868, + "learning_rate": 1.2109951745648894e-05, + "loss": 0.7827, + "step": 4139 + }, + { + "epoch": 2.247557003257329, + "grad_norm": 11.660193747264707, + "learning_rate": 1.210651395006301e-05, + "loss": 0.9824, + "step": 4140 + }, + { + "epoch": 2.248099891422367, + "grad_norm": 18.121219443467986, + "learning_rate": 1.2103075893940122e-05, + "loss": 1.2339, + "step": 4141 + }, + { + "epoch": 2.248642779587405, + "grad_norm": 15.059306090767514, + "learning_rate": 1.2099637577705457e-05, + "loss": 0.7434, + "step": 4142 + }, + { + "epoch": 2.249185667752443, + "grad_norm": 10.999940106944315, + "learning_rate": 1.2096199001784268e-05, + "loss": 0.6422, + "step": 4143 + }, + { + "epoch": 2.249728555917481, + "grad_norm": 6.704480729179984, + "learning_rate": 1.2092760166601848e-05, + "loss": 0.4829, + "step": 4144 + }, + { + "epoch": 2.250271444082519, + "grad_norm": 11.849496386091012, + "learning_rate": 1.2089321072583512e-05, + "loss": 0.8969, + "step": 4145 + }, + { + "epoch": 2.250814332247557, + "grad_norm": 8.596596068581539, + "learning_rate": 1.2085881720154617e-05, + "loss": 0.6148, + "step": 4146 + }, + { + "epoch": 2.251357220412595, + "grad_norm": 7.50181808267645, + "learning_rate": 1.2082442109740548e-05, + "loss": 0.5499, + "step": 4147 + }, + { + "epoch": 2.251900108577633, + "grad_norm": 10.178327453067736, + "learning_rate": 1.207900224176672e-05, + "loss": 0.5582, + "step": 4148 + }, + { + "epoch": 2.252442996742671, + "grad_norm": 10.74128468394649, + "learning_rate": 1.2075562116658581e-05, + "loss": 0.8023, + "step": 4149 + }, + { + "epoch": 2.252985884907709, + "grad_norm": 10.195557872174708, + "learning_rate": 1.207212173484161e-05, + "loss": 0.6092, + "step": 4150 + }, + { + "epoch": 2.253528773072747, + "grad_norm": 9.843275181764184, + "learning_rate": 1.2068681096741322e-05, + "loss": 0.8451, + "step": 4151 + }, + { + "epoch": 2.254071661237785, + "grad_norm": 7.859122806337064, + "learning_rate": 1.2065240202783259e-05, + "loss": 0.6158, + "step": 4152 + }, + { + "epoch": 2.254614549402823, + "grad_norm": 11.4233860908698, + "learning_rate": 1.2061799053392994e-05, + "loss": 0.6552, + "step": 4153 + }, + { + "epoch": 2.255157437567861, + "grad_norm": 7.339673441521232, + "learning_rate": 1.2058357648996138e-05, + "loss": 0.5103, + "step": 4154 + }, + { + "epoch": 2.255700325732899, + "grad_norm": 11.603888679814519, + "learning_rate": 1.2054915990018326e-05, + "loss": 0.8117, + "step": 4155 + }, + { + "epoch": 2.256243213897937, + "grad_norm": 9.757430713637703, + "learning_rate": 1.2051474076885232e-05, + "loss": 0.6942, + "step": 4156 + }, + { + "epoch": 2.256786102062975, + "grad_norm": 10.286384353132442, + "learning_rate": 1.2048031910022553e-05, + "loss": 0.8958, + "step": 4157 + }, + { + "epoch": 2.257328990228013, + "grad_norm": 7.6945320206947185, + "learning_rate": 1.2044589489856023e-05, + "loss": 0.4508, + "step": 4158 + }, + { + "epoch": 2.257871878393051, + "grad_norm": 9.546205961324182, + "learning_rate": 1.2041146816811406e-05, + "loss": 0.6889, + "step": 4159 + }, + { + "epoch": 2.258414766558089, + "grad_norm": 6.440627468414342, + "learning_rate": 1.2037703891314498e-05, + "loss": 0.5146, + "step": 4160 + }, + { + "epoch": 2.258957654723127, + "grad_norm": 8.769926030437105, + "learning_rate": 1.2034260713791124e-05, + "loss": 0.539, + "step": 4161 + }, + { + "epoch": 2.259500542888165, + "grad_norm": 8.385311572142552, + "learning_rate": 1.2030817284667145e-05, + "loss": 0.6471, + "step": 4162 + }, + { + "epoch": 2.260043431053203, + "grad_norm": 10.687000132395355, + "learning_rate": 1.2027373604368446e-05, + "loss": 0.4047, + "step": 4163 + }, + { + "epoch": 2.260586319218241, + "grad_norm": 9.22667613364525, + "learning_rate": 1.2023929673320952e-05, + "loss": 0.6039, + "step": 4164 + }, + { + "epoch": 2.261129207383279, + "grad_norm": 8.822764152626888, + "learning_rate": 1.2020485491950607e-05, + "loss": 0.5333, + "step": 4165 + }, + { + "epoch": 2.261672095548317, + "grad_norm": 8.505680622817092, + "learning_rate": 1.2017041060683394e-05, + "loss": 0.6679, + "step": 4166 + }, + { + "epoch": 2.262214983713355, + "grad_norm": 12.980914094830798, + "learning_rate": 1.2013596379945331e-05, + "loss": 0.7265, + "step": 4167 + }, + { + "epoch": 2.262757871878393, + "grad_norm": 12.278007785084847, + "learning_rate": 1.201015145016246e-05, + "loss": 0.7175, + "step": 4168 + }, + { + "epoch": 2.263300760043431, + "grad_norm": 8.871985541095814, + "learning_rate": 1.2006706271760851e-05, + "loss": 0.4937, + "step": 4169 + }, + { + "epoch": 2.263843648208469, + "grad_norm": 10.070308144714824, + "learning_rate": 1.2003260845166613e-05, + "loss": 0.5994, + "step": 4170 + }, + { + "epoch": 2.264386536373507, + "grad_norm": 8.846376302708403, + "learning_rate": 1.1999815170805882e-05, + "loss": 0.5795, + "step": 4171 + }, + { + "epoch": 2.264929424538545, + "grad_norm": 11.984661783852376, + "learning_rate": 1.199636924910482e-05, + "loss": 1.1718, + "step": 4172 + }, + { + "epoch": 2.265472312703583, + "grad_norm": 10.035244756765978, + "learning_rate": 1.1992923080489629e-05, + "loss": 0.8101, + "step": 4173 + }, + { + "epoch": 2.266015200868621, + "grad_norm": 7.429999917349929, + "learning_rate": 1.1989476665386537e-05, + "loss": 0.4121, + "step": 4174 + }, + { + "epoch": 2.266558089033659, + "grad_norm": 8.464944728590373, + "learning_rate": 1.1986030004221802e-05, + "loss": 0.5553, + "step": 4175 + }, + { + "epoch": 2.267100977198697, + "grad_norm": 11.414355640267063, + "learning_rate": 1.1982583097421707e-05, + "loss": 0.8705, + "step": 4176 + }, + { + "epoch": 2.267643865363735, + "grad_norm": 10.094383752442985, + "learning_rate": 1.1979135945412576e-05, + "loss": 0.5512, + "step": 4177 + }, + { + "epoch": 2.268186753528773, + "grad_norm": 10.131657295101125, + "learning_rate": 1.1975688548620756e-05, + "loss": 0.5678, + "step": 4178 + }, + { + "epoch": 2.268729641693811, + "grad_norm": 11.675191701181866, + "learning_rate": 1.197224090747263e-05, + "loss": 0.5967, + "step": 4179 + }, + { + "epoch": 2.269272529858849, + "grad_norm": 9.879570041096912, + "learning_rate": 1.1968793022394603e-05, + "loss": 1.0245, + "step": 4180 + }, + { + "epoch": 2.269815418023887, + "grad_norm": 12.255119190489571, + "learning_rate": 1.1965344893813117e-05, + "loss": 0.9963, + "step": 4181 + }, + { + "epoch": 2.270358306188925, + "grad_norm": 8.870271458097042, + "learning_rate": 1.1961896522154643e-05, + "loss": 0.3658, + "step": 4182 + }, + { + "epoch": 2.270901194353963, + "grad_norm": 9.50086592138889, + "learning_rate": 1.1958447907845684e-05, + "loss": 0.588, + "step": 4183 + }, + { + "epoch": 2.271444082519001, + "grad_norm": 8.92062913157286, + "learning_rate": 1.1954999051312762e-05, + "loss": 0.6038, + "step": 4184 + }, + { + "epoch": 2.271986970684039, + "grad_norm": 11.511359098427171, + "learning_rate": 1.1951549952982442e-05, + "loss": 0.7837, + "step": 4185 + }, + { + "epoch": 2.272529858849077, + "grad_norm": 7.7230189119336075, + "learning_rate": 1.1948100613281315e-05, + "loss": 0.5235, + "step": 4186 + }, + { + "epoch": 2.273072747014115, + "grad_norm": 8.85303744288786, + "learning_rate": 1.1944651032635997e-05, + "loss": 0.5194, + "step": 4187 + }, + { + "epoch": 2.273615635179153, + "grad_norm": 9.931552012214697, + "learning_rate": 1.1941201211473142e-05, + "loss": 0.6419, + "step": 4188 + }, + { + "epoch": 2.274158523344191, + "grad_norm": 11.635637385284054, + "learning_rate": 1.1937751150219428e-05, + "loss": 0.7948, + "step": 4189 + }, + { + "epoch": 2.274701411509229, + "grad_norm": 7.783494952264655, + "learning_rate": 1.1934300849301561e-05, + "loss": 0.568, + "step": 4190 + }, + { + "epoch": 2.275244299674267, + "grad_norm": 7.979443655042872, + "learning_rate": 1.1930850309146283e-05, + "loss": 0.531, + "step": 4191 + }, + { + "epoch": 2.2757871878393052, + "grad_norm": 12.081321074807914, + "learning_rate": 1.1927399530180359e-05, + "loss": 0.9917, + "step": 4192 + }, + { + "epoch": 2.276330076004343, + "grad_norm": 9.376910406413142, + "learning_rate": 1.1923948512830595e-05, + "loss": 0.6587, + "step": 4193 + }, + { + "epoch": 2.2768729641693812, + "grad_norm": 9.576589478578342, + "learning_rate": 1.1920497257523809e-05, + "loss": 0.5703, + "step": 4194 + }, + { + "epoch": 2.277415852334419, + "grad_norm": 9.129784891039465, + "learning_rate": 1.1917045764686863e-05, + "loss": 0.668, + "step": 4195 + }, + { + "epoch": 2.2779587404994572, + "grad_norm": 11.723388678968943, + "learning_rate": 1.191359403474664e-05, + "loss": 0.6293, + "step": 4196 + }, + { + "epoch": 2.278501628664495, + "grad_norm": 10.662032565840782, + "learning_rate": 1.1910142068130058e-05, + "loss": 1.0251, + "step": 4197 + }, + { + "epoch": 2.2790445168295332, + "grad_norm": 5.976705265828664, + "learning_rate": 1.1906689865264058e-05, + "loss": 0.3892, + "step": 4198 + }, + { + "epoch": 2.279587404994571, + "grad_norm": 9.77428172632025, + "learning_rate": 1.190323742657562e-05, + "loss": 0.8023, + "step": 4199 + }, + { + "epoch": 2.2801302931596092, + "grad_norm": 9.701826347676784, + "learning_rate": 1.1899784752491743e-05, + "loss": 0.6668, + "step": 4200 + }, + { + "epoch": 2.280673181324647, + "grad_norm": 10.939483572398276, + "learning_rate": 1.1896331843439459e-05, + "loss": 0.8751, + "step": 4201 + }, + { + "epoch": 2.2812160694896852, + "grad_norm": 12.656192710489524, + "learning_rate": 1.1892878699845832e-05, + "loss": 0.7874, + "step": 4202 + }, + { + "epoch": 2.281758957654723, + "grad_norm": 8.821175958313784, + "learning_rate": 1.1889425322137948e-05, + "loss": 0.7599, + "step": 4203 + }, + { + "epoch": 2.2823018458197613, + "grad_norm": 11.535196460357659, + "learning_rate": 1.188597171074293e-05, + "loss": 0.7448, + "step": 4204 + }, + { + "epoch": 2.282844733984799, + "grad_norm": 9.221866003570257, + "learning_rate": 1.1882517866087926e-05, + "loss": 0.7176, + "step": 4205 + }, + { + "epoch": 2.2833876221498373, + "grad_norm": 10.791249990779047, + "learning_rate": 1.1879063788600113e-05, + "loss": 0.6905, + "step": 4206 + }, + { + "epoch": 2.283930510314875, + "grad_norm": 11.953403018554845, + "learning_rate": 1.1875609478706694e-05, + "loss": 0.7006, + "step": 4207 + }, + { + "epoch": 2.2844733984799133, + "grad_norm": 10.830659589867402, + "learning_rate": 1.1872154936834905e-05, + "loss": 0.6872, + "step": 4208 + }, + { + "epoch": 2.285016286644951, + "grad_norm": 10.8147041656128, + "learning_rate": 1.186870016341201e-05, + "loss": 0.6429, + "step": 4209 + }, + { + "epoch": 2.2855591748099893, + "grad_norm": 10.70508214004579, + "learning_rate": 1.18652451588653e-05, + "loss": 0.9063, + "step": 4210 + }, + { + "epoch": 2.286102062975027, + "grad_norm": 8.333893553672915, + "learning_rate": 1.1861789923622099e-05, + "loss": 0.5245, + "step": 4211 + }, + { + "epoch": 2.2866449511400653, + "grad_norm": 9.62415991156623, + "learning_rate": 1.1858334458109747e-05, + "loss": 0.734, + "step": 4212 + }, + { + "epoch": 2.287187839305103, + "grad_norm": 15.893278058161348, + "learning_rate": 1.1854878762755633e-05, + "loss": 0.898, + "step": 4213 + }, + { + "epoch": 2.2877307274701413, + "grad_norm": 8.12705480823925, + "learning_rate": 1.1851422837987159e-05, + "loss": 0.5586, + "step": 4214 + }, + { + "epoch": 2.288273615635179, + "grad_norm": 19.440766463719637, + "learning_rate": 1.1847966684231754e-05, + "loss": 1.0309, + "step": 4215 + }, + { + "epoch": 2.2888165038002173, + "grad_norm": 10.590183829770972, + "learning_rate": 1.1844510301916883e-05, + "loss": 0.8545, + "step": 4216 + }, + { + "epoch": 2.289359391965255, + "grad_norm": 9.825370158957117, + "learning_rate": 1.184105369147004e-05, + "loss": 0.7426, + "step": 4217 + }, + { + "epoch": 2.2899022801302933, + "grad_norm": 10.341219416998992, + "learning_rate": 1.1837596853318742e-05, + "loss": 0.8584, + "step": 4218 + }, + { + "epoch": 2.290445168295331, + "grad_norm": 11.121322184848584, + "learning_rate": 1.1834139787890538e-05, + "loss": 0.6598, + "step": 4219 + }, + { + "epoch": 2.2909880564603693, + "grad_norm": 9.600608599739235, + "learning_rate": 1.1830682495613e-05, + "loss": 0.5885, + "step": 4220 + }, + { + "epoch": 2.291530944625407, + "grad_norm": 10.517374576113433, + "learning_rate": 1.1827224976913732e-05, + "loss": 0.751, + "step": 4221 + }, + { + "epoch": 2.2920738327904453, + "grad_norm": 9.666522143402908, + "learning_rate": 1.1823767232220369e-05, + "loss": 0.7602, + "step": 4222 + }, + { + "epoch": 2.292616720955483, + "grad_norm": 9.390811834354404, + "learning_rate": 1.1820309261960567e-05, + "loss": 0.6187, + "step": 4223 + }, + { + "epoch": 2.2931596091205213, + "grad_norm": 8.772653626817414, + "learning_rate": 1.1816851066562017e-05, + "loss": 0.5221, + "step": 4224 + }, + { + "epoch": 2.293702497285559, + "grad_norm": 11.957278593216312, + "learning_rate": 1.1813392646452429e-05, + "loss": 0.9231, + "step": 4225 + }, + { + "epoch": 2.2942453854505973, + "grad_norm": 9.58191568158383, + "learning_rate": 1.1809934002059547e-05, + "loss": 0.6747, + "step": 4226 + }, + { + "epoch": 2.294788273615635, + "grad_norm": 8.242154702381846, + "learning_rate": 1.1806475133811146e-05, + "loss": 0.4053, + "step": 4227 + }, + { + "epoch": 2.2953311617806733, + "grad_norm": 8.83768826569311, + "learning_rate": 1.1803016042135021e-05, + "loss": 0.4177, + "step": 4228 + }, + { + "epoch": 2.295874049945711, + "grad_norm": 7.496802017426867, + "learning_rate": 1.1799556727458996e-05, + "loss": 0.6402, + "step": 4229 + }, + { + "epoch": 2.2964169381107493, + "grad_norm": 10.15598151517949, + "learning_rate": 1.1796097190210927e-05, + "loss": 0.6301, + "step": 4230 + }, + { + "epoch": 2.296959826275787, + "grad_norm": 9.100517578174307, + "learning_rate": 1.1792637430818695e-05, + "loss": 0.6059, + "step": 4231 + }, + { + "epoch": 2.2975027144408253, + "grad_norm": 9.341766570492698, + "learning_rate": 1.178917744971021e-05, + "loss": 0.7462, + "step": 4232 + }, + { + "epoch": 2.298045602605863, + "grad_norm": 9.790098160117006, + "learning_rate": 1.1785717247313407e-05, + "loss": 0.7422, + "step": 4233 + }, + { + "epoch": 2.2985884907709013, + "grad_norm": 11.785065190423747, + "learning_rate": 1.1782256824056249e-05, + "loss": 0.8544, + "step": 4234 + }, + { + "epoch": 2.299131378935939, + "grad_norm": 11.165339576639699, + "learning_rate": 1.1778796180366722e-05, + "loss": 0.6492, + "step": 4235 + }, + { + "epoch": 2.2996742671009773, + "grad_norm": 9.026810304485423, + "learning_rate": 1.1775335316672854e-05, + "loss": 0.751, + "step": 4236 + }, + { + "epoch": 2.300217155266015, + "grad_norm": 10.085898624762818, + "learning_rate": 1.1771874233402682e-05, + "loss": 0.6052, + "step": 4237 + }, + { + "epoch": 2.3007600434310533, + "grad_norm": 9.302471978496024, + "learning_rate": 1.176841293098428e-05, + "loss": 0.7269, + "step": 4238 + }, + { + "epoch": 2.301302931596091, + "grad_norm": 11.406933536282311, + "learning_rate": 1.176495140984575e-05, + "loss": 0.7287, + "step": 4239 + }, + { + "epoch": 2.3018458197611293, + "grad_norm": 8.172755818720033, + "learning_rate": 1.1761489670415218e-05, + "loss": 0.4692, + "step": 4240 + }, + { + "epoch": 2.302388707926167, + "grad_norm": 11.729275499093248, + "learning_rate": 1.1758027713120833e-05, + "loss": 0.8301, + "step": 4241 + }, + { + "epoch": 2.3029315960912053, + "grad_norm": 7.31457707003035, + "learning_rate": 1.1754565538390782e-05, + "loss": 0.5372, + "step": 4242 + }, + { + "epoch": 2.303474484256243, + "grad_norm": 10.15517896282683, + "learning_rate": 1.1751103146653265e-05, + "loss": 0.8744, + "step": 4243 + }, + { + "epoch": 2.3040173724212814, + "grad_norm": 7.780368081203029, + "learning_rate": 1.1747640538336525e-05, + "loss": 0.4743, + "step": 4244 + }, + { + "epoch": 2.304560260586319, + "grad_norm": 7.814816774582812, + "learning_rate": 1.1744177713868814e-05, + "loss": 0.628, + "step": 4245 + }, + { + "epoch": 2.3051031487513574, + "grad_norm": 8.531245398152228, + "learning_rate": 1.1740714673678425e-05, + "loss": 0.4419, + "step": 4246 + }, + { + "epoch": 2.305646036916395, + "grad_norm": 8.091385303458143, + "learning_rate": 1.1737251418193669e-05, + "loss": 0.6249, + "step": 4247 + }, + { + "epoch": 2.3061889250814334, + "grad_norm": 10.253051982681455, + "learning_rate": 1.1733787947842892e-05, + "loss": 0.8892, + "step": 4248 + }, + { + "epoch": 2.306731813246471, + "grad_norm": 8.95406502670429, + "learning_rate": 1.1730324263054453e-05, + "loss": 0.7452, + "step": 4249 + }, + { + "epoch": 2.3072747014115094, + "grad_norm": 10.491159901998326, + "learning_rate": 1.1726860364256756e-05, + "loss": 0.4995, + "step": 4250 + }, + { + "epoch": 2.307817589576547, + "grad_norm": 9.241135791509704, + "learning_rate": 1.1723396251878214e-05, + "loss": 0.7167, + "step": 4251 + }, + { + "epoch": 2.3083604777415854, + "grad_norm": 9.466621171091207, + "learning_rate": 1.1719931926347279e-05, + "loss": 0.5849, + "step": 4252 + }, + { + "epoch": 2.308903365906623, + "grad_norm": 10.933415330757905, + "learning_rate": 1.1716467388092423e-05, + "loss": 1.0411, + "step": 4253 + }, + { + "epoch": 2.3094462540716614, + "grad_norm": 10.009779873858768, + "learning_rate": 1.1713002637542146e-05, + "loss": 0.5456, + "step": 4254 + }, + { + "epoch": 2.309989142236699, + "grad_norm": 11.678038797864954, + "learning_rate": 1.1709537675124966e-05, + "loss": 0.5537, + "step": 4255 + }, + { + "epoch": 2.3105320304017374, + "grad_norm": 11.578035258644414, + "learning_rate": 1.1706072501269448e-05, + "loss": 0.7804, + "step": 4256 + }, + { + "epoch": 2.311074918566775, + "grad_norm": 8.678307181037598, + "learning_rate": 1.1702607116404163e-05, + "loss": 0.5378, + "step": 4257 + }, + { + "epoch": 2.3116178067318134, + "grad_norm": 8.487670920129334, + "learning_rate": 1.1699141520957716e-05, + "loss": 0.558, + "step": 4258 + }, + { + "epoch": 2.312160694896851, + "grad_norm": 10.92371973791016, + "learning_rate": 1.1695675715358737e-05, + "loss": 0.6615, + "step": 4259 + }, + { + "epoch": 2.3127035830618894, + "grad_norm": 9.360985487621843, + "learning_rate": 1.169220970003588e-05, + "loss": 0.5896, + "step": 4260 + }, + { + "epoch": 2.313246471226927, + "grad_norm": 10.382914656511591, + "learning_rate": 1.1688743475417836e-05, + "loss": 0.8869, + "step": 4261 + }, + { + "epoch": 2.3137893593919654, + "grad_norm": 10.27408683638607, + "learning_rate": 1.1685277041933303e-05, + "loss": 1.0929, + "step": 4262 + }, + { + "epoch": 2.314332247557003, + "grad_norm": 10.738574683790112, + "learning_rate": 1.168181040001102e-05, + "loss": 0.845, + "step": 4263 + }, + { + "epoch": 2.3148751357220414, + "grad_norm": 11.783082834204276, + "learning_rate": 1.1678343550079751e-05, + "loss": 0.8358, + "step": 4264 + }, + { + "epoch": 2.315418023887079, + "grad_norm": 9.277577078134165, + "learning_rate": 1.1674876492568272e-05, + "loss": 0.4774, + "step": 4265 + }, + { + "epoch": 2.3159609120521174, + "grad_norm": 10.232983577395474, + "learning_rate": 1.1671409227905401e-05, + "loss": 0.6358, + "step": 4266 + }, + { + "epoch": 2.316503800217155, + "grad_norm": 7.971516257815676, + "learning_rate": 1.1667941756519973e-05, + "loss": 0.5724, + "step": 4267 + }, + { + "epoch": 2.3170466883821934, + "grad_norm": 6.657326238165826, + "learning_rate": 1.1664474078840848e-05, + "loss": 0.3377, + "step": 4268 + }, + { + "epoch": 2.317589576547231, + "grad_norm": 9.831416988328497, + "learning_rate": 1.1661006195296918e-05, + "loss": 0.7115, + "step": 4269 + }, + { + "epoch": 2.3181324647122694, + "grad_norm": 7.2473297381695225, + "learning_rate": 1.1657538106317094e-05, + "loss": 0.4828, + "step": 4270 + }, + { + "epoch": 2.318675352877307, + "grad_norm": 11.474212650892378, + "learning_rate": 1.1654069812330315e-05, + "loss": 0.66, + "step": 4271 + }, + { + "epoch": 2.3192182410423454, + "grad_norm": 9.354650178037069, + "learning_rate": 1.1650601313765547e-05, + "loss": 0.6651, + "step": 4272 + }, + { + "epoch": 2.319761129207383, + "grad_norm": 8.945160340560053, + "learning_rate": 1.1647132611051779e-05, + "loss": 0.6422, + "step": 4273 + }, + { + "epoch": 2.3203040173724214, + "grad_norm": 8.790681463696455, + "learning_rate": 1.164366370461802e-05, + "loss": 0.6387, + "step": 4274 + }, + { + "epoch": 2.320846905537459, + "grad_norm": 8.950872212457588, + "learning_rate": 1.1640194594893317e-05, + "loss": 0.4455, + "step": 4275 + }, + { + "epoch": 2.3213897937024974, + "grad_norm": 9.879748686678086, + "learning_rate": 1.1636725282306733e-05, + "loss": 0.5539, + "step": 4276 + }, + { + "epoch": 2.321932681867535, + "grad_norm": 8.351885716181597, + "learning_rate": 1.1633255767287356e-05, + "loss": 0.4273, + "step": 4277 + }, + { + "epoch": 2.3224755700325734, + "grad_norm": 12.165689696691171, + "learning_rate": 1.1629786050264304e-05, + "loss": 0.6877, + "step": 4278 + }, + { + "epoch": 2.323018458197611, + "grad_norm": 10.70545900286579, + "learning_rate": 1.1626316131666714e-05, + "loss": 0.6667, + "step": 4279 + }, + { + "epoch": 2.3235613463626494, + "grad_norm": 7.521776321190906, + "learning_rate": 1.1622846011923756e-05, + "loss": 0.6538, + "step": 4280 + }, + { + "epoch": 2.324104234527687, + "grad_norm": 10.429691195494106, + "learning_rate": 1.1619375691464614e-05, + "loss": 0.7918, + "step": 4281 + }, + { + "epoch": 2.3246471226927254, + "grad_norm": 13.40241359482942, + "learning_rate": 1.1615905170718507e-05, + "loss": 0.8492, + "step": 4282 + }, + { + "epoch": 2.3251900108577632, + "grad_norm": 9.180074174181755, + "learning_rate": 1.1612434450114674e-05, + "loss": 0.9197, + "step": 4283 + }, + { + "epoch": 2.3257328990228014, + "grad_norm": 8.662388876238891, + "learning_rate": 1.1608963530082375e-05, + "loss": 0.4911, + "step": 4284 + }, + { + "epoch": 2.3262757871878392, + "grad_norm": 12.915716152146215, + "learning_rate": 1.1605492411050907e-05, + "loss": 1.2342, + "step": 4285 + }, + { + "epoch": 2.3268186753528775, + "grad_norm": 15.064512853420336, + "learning_rate": 1.1602021093449577e-05, + "loss": 0.9441, + "step": 4286 + }, + { + "epoch": 2.3273615635179152, + "grad_norm": 8.152078540931088, + "learning_rate": 1.159854957770772e-05, + "loss": 0.4407, + "step": 4287 + }, + { + "epoch": 2.3279044516829535, + "grad_norm": 13.440334362031807, + "learning_rate": 1.1595077864254708e-05, + "loss": 1.0186, + "step": 4288 + }, + { + "epoch": 2.3284473398479912, + "grad_norm": 14.960404464650807, + "learning_rate": 1.1591605953519925e-05, + "loss": 0.854, + "step": 4289 + }, + { + "epoch": 2.3289902280130295, + "grad_norm": 10.871106066518651, + "learning_rate": 1.1588133845932779e-05, + "loss": 0.8522, + "step": 4290 + }, + { + "epoch": 2.3295331161780672, + "grad_norm": 9.218798486570465, + "learning_rate": 1.1584661541922706e-05, + "loss": 0.5529, + "step": 4291 + }, + { + "epoch": 2.3300760043431055, + "grad_norm": 9.644924391242567, + "learning_rate": 1.158118904191917e-05, + "loss": 0.619, + "step": 4292 + }, + { + "epoch": 2.3306188925081432, + "grad_norm": 7.573505575520159, + "learning_rate": 1.1577716346351655e-05, + "loss": 0.5931, + "step": 4293 + }, + { + "epoch": 2.3311617806731815, + "grad_norm": 11.909869183875244, + "learning_rate": 1.1574243455649666e-05, + "loss": 0.8922, + "step": 4294 + }, + { + "epoch": 2.3317046688382193, + "grad_norm": 13.963856423636752, + "learning_rate": 1.1570770370242737e-05, + "loss": 0.9243, + "step": 4295 + }, + { + "epoch": 2.3322475570032575, + "grad_norm": 8.732100079298183, + "learning_rate": 1.1567297090560425e-05, + "loss": 0.5075, + "step": 4296 + }, + { + "epoch": 2.3327904451682953, + "grad_norm": 8.823180159543773, + "learning_rate": 1.1563823617032313e-05, + "loss": 0.5471, + "step": 4297 + }, + { + "epoch": 2.3333333333333335, + "grad_norm": 8.566225161651246, + "learning_rate": 1.1560349950088001e-05, + "loss": 0.5719, + "step": 4298 + }, + { + "epoch": 2.3338762214983713, + "grad_norm": 10.549198011052681, + "learning_rate": 1.155687609015712e-05, + "loss": 0.4775, + "step": 4299 + }, + { + "epoch": 2.3344191096634095, + "grad_norm": 9.79276879932382, + "learning_rate": 1.1553402037669326e-05, + "loss": 0.5429, + "step": 4300 + }, + { + "epoch": 2.3349619978284473, + "grad_norm": 12.065070131397079, + "learning_rate": 1.154992779305429e-05, + "loss": 0.9434, + "step": 4301 + }, + { + "epoch": 2.3355048859934855, + "grad_norm": 11.007790011549671, + "learning_rate": 1.1546453356741714e-05, + "loss": 0.6199, + "step": 4302 + }, + { + "epoch": 2.3360477741585233, + "grad_norm": 12.466762411503309, + "learning_rate": 1.1542978729161326e-05, + "loss": 0.7941, + "step": 4303 + }, + { + "epoch": 2.3365906623235615, + "grad_norm": 11.55197462123976, + "learning_rate": 1.1539503910742866e-05, + "loss": 0.6766, + "step": 4304 + }, + { + "epoch": 2.3371335504885993, + "grad_norm": 8.462143561329682, + "learning_rate": 1.153602890191611e-05, + "loss": 0.5183, + "step": 4305 + }, + { + "epoch": 2.3376764386536375, + "grad_norm": 13.818285051654057, + "learning_rate": 1.1532553703110852e-05, + "loss": 0.8363, + "step": 4306 + }, + { + "epoch": 2.3382193268186753, + "grad_norm": 8.676120343621662, + "learning_rate": 1.1529078314756905e-05, + "loss": 0.6976, + "step": 4307 + }, + { + "epoch": 2.3387622149837135, + "grad_norm": 10.024443385726268, + "learning_rate": 1.1525602737284119e-05, + "loss": 0.4879, + "step": 4308 + }, + { + "epoch": 2.3393051031487513, + "grad_norm": 11.965265440206402, + "learning_rate": 1.1522126971122355e-05, + "loss": 0.8655, + "step": 4309 + }, + { + "epoch": 2.3398479913137895, + "grad_norm": 9.503916256652449, + "learning_rate": 1.15186510167015e-05, + "loss": 0.8966, + "step": 4310 + }, + { + "epoch": 2.3403908794788273, + "grad_norm": 11.942859003506952, + "learning_rate": 1.1515174874451466e-05, + "loss": 0.7769, + "step": 4311 + }, + { + "epoch": 2.3409337676438655, + "grad_norm": 9.50234633554061, + "learning_rate": 1.1511698544802193e-05, + "loss": 0.552, + "step": 4312 + }, + { + "epoch": 2.3414766558089033, + "grad_norm": 8.495987313736284, + "learning_rate": 1.1508222028183632e-05, + "loss": 0.8566, + "step": 4313 + }, + { + "epoch": 2.3420195439739415, + "grad_norm": 11.854978708479615, + "learning_rate": 1.1504745325025766e-05, + "loss": 0.6736, + "step": 4314 + }, + { + "epoch": 2.3425624321389793, + "grad_norm": 11.169148481350696, + "learning_rate": 1.1501268435758602e-05, + "loss": 0.8075, + "step": 4315 + }, + { + "epoch": 2.3431053203040175, + "grad_norm": 11.983506203130691, + "learning_rate": 1.1497791360812163e-05, + "loss": 0.6407, + "step": 4316 + }, + { + "epoch": 2.3436482084690553, + "grad_norm": 8.507852970222672, + "learning_rate": 1.1494314100616502e-05, + "loss": 0.6247, + "step": 4317 + }, + { + "epoch": 2.3441910966340935, + "grad_norm": 19.964294777139003, + "learning_rate": 1.1490836655601692e-05, + "loss": 0.9219, + "step": 4318 + }, + { + "epoch": 2.3447339847991313, + "grad_norm": 8.202074828225685, + "learning_rate": 1.1487359026197828e-05, + "loss": 0.4887, + "step": 4319 + }, + { + "epoch": 2.3452768729641695, + "grad_norm": 10.22404099258637, + "learning_rate": 1.1483881212835029e-05, + "loss": 0.4028, + "step": 4320 + }, + { + "epoch": 2.3458197611292073, + "grad_norm": 8.30587529218613, + "learning_rate": 1.1480403215943436e-05, + "loss": 0.4047, + "step": 4321 + }, + { + "epoch": 2.3463626492942455, + "grad_norm": 7.121924947407594, + "learning_rate": 1.1476925035953217e-05, + "loss": 0.4922, + "step": 4322 + }, + { + "epoch": 2.3469055374592833, + "grad_norm": 10.58793763221932, + "learning_rate": 1.1473446673294552e-05, + "loss": 0.9691, + "step": 4323 + }, + { + "epoch": 2.3474484256243215, + "grad_norm": 11.51770882137354, + "learning_rate": 1.1469968128397656e-05, + "loss": 0.6713, + "step": 4324 + }, + { + "epoch": 2.3479913137893593, + "grad_norm": 9.960125063324698, + "learning_rate": 1.146648940169276e-05, + "loss": 0.8266, + "step": 4325 + }, + { + "epoch": 2.3485342019543975, + "grad_norm": 14.574982915069748, + "learning_rate": 1.1463010493610114e-05, + "loss": 1.4129, + "step": 4326 + }, + { + "epoch": 2.3490770901194353, + "grad_norm": 7.9866843222052, + "learning_rate": 1.145953140458e-05, + "loss": 0.6368, + "step": 4327 + }, + { + "epoch": 2.3496199782844736, + "grad_norm": 14.879386143441858, + "learning_rate": 1.1456052135032717e-05, + "loss": 0.579, + "step": 4328 + }, + { + "epoch": 2.3501628664495113, + "grad_norm": 9.951154243493171, + "learning_rate": 1.1452572685398586e-05, + "loss": 0.6989, + "step": 4329 + }, + { + "epoch": 2.3507057546145496, + "grad_norm": 8.242511521892395, + "learning_rate": 1.144909305610795e-05, + "loss": 0.5213, + "step": 4330 + }, + { + "epoch": 2.3512486427795873, + "grad_norm": 10.259159837692009, + "learning_rate": 1.1445613247591174e-05, + "loss": 0.9849, + "step": 4331 + }, + { + "epoch": 2.3517915309446256, + "grad_norm": 7.97479592417325, + "learning_rate": 1.1442133260278648e-05, + "loss": 0.5867, + "step": 4332 + }, + { + "epoch": 2.3523344191096633, + "grad_norm": 8.493125656661944, + "learning_rate": 1.143865309460078e-05, + "loss": 0.9095, + "step": 4333 + }, + { + "epoch": 2.3528773072747016, + "grad_norm": 12.49067508129187, + "learning_rate": 1.1435172750988009e-05, + "loss": 0.8484, + "step": 4334 + }, + { + "epoch": 2.3534201954397393, + "grad_norm": 11.134649233778442, + "learning_rate": 1.1431692229870782e-05, + "loss": 0.8422, + "step": 4335 + }, + { + "epoch": 2.3539630836047776, + "grad_norm": 6.99064506687852, + "learning_rate": 1.1428211531679576e-05, + "loss": 0.5388, + "step": 4336 + }, + { + "epoch": 2.3545059717698154, + "grad_norm": 7.798255243454909, + "learning_rate": 1.1424730656844895e-05, + "loss": 0.593, + "step": 4337 + }, + { + "epoch": 2.3550488599348536, + "grad_norm": 8.178641043550261, + "learning_rate": 1.1421249605797255e-05, + "loss": 0.4778, + "step": 4338 + }, + { + "epoch": 2.3555917480998914, + "grad_norm": 8.917024765042555, + "learning_rate": 1.1417768378967196e-05, + "loss": 0.5782, + "step": 4339 + }, + { + "epoch": 2.3561346362649296, + "grad_norm": 9.742472427719518, + "learning_rate": 1.1414286976785286e-05, + "loss": 0.8709, + "step": 4340 + }, + { + "epoch": 2.3566775244299674, + "grad_norm": 10.08457903035175, + "learning_rate": 1.141080539968211e-05, + "loss": 0.5582, + "step": 4341 + }, + { + "epoch": 2.3572204125950056, + "grad_norm": 10.74958408516018, + "learning_rate": 1.1407323648088275e-05, + "loss": 0.7498, + "step": 4342 + }, + { + "epoch": 2.3577633007600434, + "grad_norm": 12.356830179446728, + "learning_rate": 1.1403841722434405e-05, + "loss": 0.9357, + "step": 4343 + }, + { + "epoch": 2.3583061889250816, + "grad_norm": 8.220220873503408, + "learning_rate": 1.1400359623151153e-05, + "loss": 0.5554, + "step": 4344 + }, + { + "epoch": 2.3588490770901194, + "grad_norm": 10.916598567149965, + "learning_rate": 1.1396877350669194e-05, + "loss": 0.8239, + "step": 4345 + }, + { + "epoch": 2.3593919652551576, + "grad_norm": 10.51236366107464, + "learning_rate": 1.1393394905419215e-05, + "loss": 0.5945, + "step": 4346 + }, + { + "epoch": 2.3599348534201954, + "grad_norm": 10.534153676139686, + "learning_rate": 1.1389912287831935e-05, + "loss": 0.575, + "step": 4347 + }, + { + "epoch": 2.3604777415852336, + "grad_norm": 10.06758846687261, + "learning_rate": 1.1386429498338088e-05, + "loss": 0.702, + "step": 4348 + }, + { + "epoch": 2.3610206297502714, + "grad_norm": 9.282335539362101, + "learning_rate": 1.1382946537368432e-05, + "loss": 0.5645, + "step": 4349 + }, + { + "epoch": 2.3615635179153096, + "grad_norm": 8.137108078430895, + "learning_rate": 1.1379463405353741e-05, + "loss": 0.5057, + "step": 4350 + }, + { + "epoch": 2.3621064060803474, + "grad_norm": 10.651520296934317, + "learning_rate": 1.137598010272482e-05, + "loss": 0.762, + "step": 4351 + }, + { + "epoch": 2.3626492942453856, + "grad_norm": 10.001083727438939, + "learning_rate": 1.1372496629912486e-05, + "loss": 0.7255, + "step": 4352 + }, + { + "epoch": 2.3631921824104234, + "grad_norm": 12.180927670825834, + "learning_rate": 1.1369012987347582e-05, + "loss": 0.7295, + "step": 4353 + }, + { + "epoch": 2.3637350705754616, + "grad_norm": 10.697210588401258, + "learning_rate": 1.136552917546097e-05, + "loss": 0.5159, + "step": 4354 + }, + { + "epoch": 2.3642779587404994, + "grad_norm": 13.551389197850602, + "learning_rate": 1.1362045194683534e-05, + "loss": 0.9265, + "step": 4355 + }, + { + "epoch": 2.3648208469055376, + "grad_norm": 12.38857946888456, + "learning_rate": 1.1358561045446175e-05, + "loss": 0.7577, + "step": 4356 + }, + { + "epoch": 2.3653637350705754, + "grad_norm": 10.138108604403497, + "learning_rate": 1.1355076728179824e-05, + "loss": 0.5328, + "step": 4357 + }, + { + "epoch": 2.3659066232356136, + "grad_norm": 9.823785539850283, + "learning_rate": 1.1351592243315416e-05, + "loss": 0.6116, + "step": 4358 + }, + { + "epoch": 2.3664495114006514, + "grad_norm": 8.238880239030014, + "learning_rate": 1.1348107591283932e-05, + "loss": 0.3863, + "step": 4359 + }, + { + "epoch": 2.3669923995656896, + "grad_norm": 11.34031175123055, + "learning_rate": 1.1344622772516352e-05, + "loss": 0.9601, + "step": 4360 + }, + { + "epoch": 2.3675352877307274, + "grad_norm": 11.270983899241404, + "learning_rate": 1.1341137787443683e-05, + "loss": 0.7598, + "step": 4361 + }, + { + "epoch": 2.3680781758957656, + "grad_norm": 10.649130568340542, + "learning_rate": 1.1337652636496959e-05, + "loss": 0.6547, + "step": 4362 + }, + { + "epoch": 2.3686210640608034, + "grad_norm": 9.462979596192575, + "learning_rate": 1.1334167320107224e-05, + "loss": 0.5247, + "step": 4363 + }, + { + "epoch": 2.3691639522258416, + "grad_norm": 11.496313370817235, + "learning_rate": 1.1330681838705544e-05, + "loss": 0.5465, + "step": 4364 + }, + { + "epoch": 2.3697068403908794, + "grad_norm": 10.125563730301518, + "learning_rate": 1.1327196192723016e-05, + "loss": 0.5687, + "step": 4365 + }, + { + "epoch": 2.3702497285559176, + "grad_norm": 14.511657264928447, + "learning_rate": 1.1323710382590747e-05, + "loss": 0.7274, + "step": 4366 + }, + { + "epoch": 2.3707926167209554, + "grad_norm": 10.08856219742556, + "learning_rate": 1.1320224408739868e-05, + "loss": 0.4633, + "step": 4367 + }, + { + "epoch": 2.3713355048859937, + "grad_norm": 9.943874647753937, + "learning_rate": 1.131673827160153e-05, + "loss": 0.7042, + "step": 4368 + }, + { + "epoch": 2.3718783930510314, + "grad_norm": 10.550654748070723, + "learning_rate": 1.1313251971606906e-05, + "loss": 0.6336, + "step": 4369 + }, + { + "epoch": 2.3724212812160697, + "grad_norm": 14.84394771322204, + "learning_rate": 1.1309765509187182e-05, + "loss": 0.6591, + "step": 4370 + }, + { + "epoch": 2.3729641693811074, + "grad_norm": 10.962774929698096, + "learning_rate": 1.1306278884773576e-05, + "loss": 0.9648, + "step": 4371 + }, + { + "epoch": 2.3735070575461457, + "grad_norm": 9.063486875880555, + "learning_rate": 1.1302792098797312e-05, + "loss": 0.4618, + "step": 4372 + }, + { + "epoch": 2.3740499457111834, + "grad_norm": 10.458070080982525, + "learning_rate": 1.1299305151689648e-05, + "loss": 0.6216, + "step": 4373 + }, + { + "epoch": 2.3745928338762217, + "grad_norm": 10.041421519621146, + "learning_rate": 1.1295818043881852e-05, + "loss": 0.6642, + "step": 4374 + }, + { + "epoch": 2.3751357220412594, + "grad_norm": 8.817266669804289, + "learning_rate": 1.1292330775805214e-05, + "loss": 0.4725, + "step": 4375 + }, + { + "epoch": 2.3756786102062977, + "grad_norm": 12.823614653687773, + "learning_rate": 1.1288843347891045e-05, + "loss": 0.553, + "step": 4376 + }, + { + "epoch": 2.3762214983713354, + "grad_norm": 9.281557160359569, + "learning_rate": 1.1285355760570677e-05, + "loss": 0.6436, + "step": 4377 + }, + { + "epoch": 2.3767643865363737, + "grad_norm": 13.234297666234115, + "learning_rate": 1.1281868014275458e-05, + "loss": 0.7665, + "step": 4378 + }, + { + "epoch": 2.3773072747014115, + "grad_norm": 13.013112559574942, + "learning_rate": 1.1278380109436764e-05, + "loss": 0.9087, + "step": 4379 + }, + { + "epoch": 2.3778501628664497, + "grad_norm": 9.413433369666237, + "learning_rate": 1.1274892046485979e-05, + "loss": 0.5248, + "step": 4380 + }, + { + "epoch": 2.3783930510314875, + "grad_norm": 10.865873661981059, + "learning_rate": 1.1271403825854516e-05, + "loss": 0.7398, + "step": 4381 + }, + { + "epoch": 2.3789359391965257, + "grad_norm": 8.516596816721957, + "learning_rate": 1.12679154479738e-05, + "loss": 0.4652, + "step": 4382 + }, + { + "epoch": 2.3794788273615635, + "grad_norm": 7.9464803374429085, + "learning_rate": 1.1264426913275277e-05, + "loss": 0.6952, + "step": 4383 + }, + { + "epoch": 2.3800217155266017, + "grad_norm": 9.171877282099603, + "learning_rate": 1.1260938222190423e-05, + "loss": 0.7534, + "step": 4384 + }, + { + "epoch": 2.3805646036916395, + "grad_norm": 10.833526250938174, + "learning_rate": 1.1257449375150717e-05, + "loss": 0.7966, + "step": 4385 + }, + { + "epoch": 2.3811074918566777, + "grad_norm": 7.087015821603363, + "learning_rate": 1.1253960372587669e-05, + "loss": 0.5145, + "step": 4386 + }, + { + "epoch": 2.3816503800217155, + "grad_norm": 9.355325600419727, + "learning_rate": 1.1250471214932805e-05, + "loss": 0.6084, + "step": 4387 + }, + { + "epoch": 2.3821932681867537, + "grad_norm": 11.679312988029317, + "learning_rate": 1.1246981902617666e-05, + "loss": 0.7734, + "step": 4388 + }, + { + "epoch": 2.3827361563517915, + "grad_norm": 11.696985760230557, + "learning_rate": 1.124349243607382e-05, + "loss": 0.7843, + "step": 4389 + }, + { + "epoch": 2.3832790445168297, + "grad_norm": 11.045261072625825, + "learning_rate": 1.1240002815732847e-05, + "loss": 0.8102, + "step": 4390 + }, + { + "epoch": 2.3838219326818675, + "grad_norm": 13.853716237422205, + "learning_rate": 1.123651304202635e-05, + "loss": 0.6548, + "step": 4391 + }, + { + "epoch": 2.3843648208469057, + "grad_norm": 9.256284103603257, + "learning_rate": 1.1233023115385949e-05, + "loss": 0.5769, + "step": 4392 + }, + { + "epoch": 2.3849077090119435, + "grad_norm": 11.719289709502613, + "learning_rate": 1.1229533036243283e-05, + "loss": 0.9445, + "step": 4393 + }, + { + "epoch": 2.3854505971769817, + "grad_norm": 13.181528127455582, + "learning_rate": 1.1226042805030014e-05, + "loss": 0.78, + "step": 4394 + }, + { + "epoch": 2.3859934853420195, + "grad_norm": 10.610212858388932, + "learning_rate": 1.1222552422177813e-05, + "loss": 0.9325, + "step": 4395 + }, + { + "epoch": 2.3865363735070577, + "grad_norm": 8.750069163521957, + "learning_rate": 1.1219061888118381e-05, + "loss": 0.5576, + "step": 4396 + }, + { + "epoch": 2.3870792616720955, + "grad_norm": 11.697241415039246, + "learning_rate": 1.121557120328343e-05, + "loss": 0.6042, + "step": 4397 + }, + { + "epoch": 2.3876221498371337, + "grad_norm": 14.150909202475516, + "learning_rate": 1.1212080368104699e-05, + "loss": 0.8888, + "step": 4398 + }, + { + "epoch": 2.3881650380021715, + "grad_norm": 8.45338495407854, + "learning_rate": 1.1208589383013934e-05, + "loss": 0.5394, + "step": 4399 + }, + { + "epoch": 2.3887079261672097, + "grad_norm": 8.581855639244504, + "learning_rate": 1.120509824844291e-05, + "loss": 0.4231, + "step": 4400 + }, + { + "epoch": 2.3892508143322475, + "grad_norm": 9.723553293714806, + "learning_rate": 1.1201606964823415e-05, + "loss": 0.5184, + "step": 4401 + }, + { + "epoch": 2.3897937024972857, + "grad_norm": 11.437338247112033, + "learning_rate": 1.1198115532587254e-05, + "loss": 0.9033, + "step": 4402 + }, + { + "epoch": 2.3903365906623235, + "grad_norm": 11.185644591472458, + "learning_rate": 1.1194623952166252e-05, + "loss": 0.7607, + "step": 4403 + }, + { + "epoch": 2.3908794788273617, + "grad_norm": 11.125764861054968, + "learning_rate": 1.1191132223992258e-05, + "loss": 0.7844, + "step": 4404 + }, + { + "epoch": 2.3914223669923995, + "grad_norm": 10.915426846928037, + "learning_rate": 1.1187640348497131e-05, + "loss": 0.5862, + "step": 4405 + }, + { + "epoch": 2.3919652551574377, + "grad_norm": 8.032417531111763, + "learning_rate": 1.1184148326112753e-05, + "loss": 0.5992, + "step": 4406 + }, + { + "epoch": 2.3925081433224755, + "grad_norm": 10.322050440175012, + "learning_rate": 1.1180656157271026e-05, + "loss": 0.6275, + "step": 4407 + }, + { + "epoch": 2.3930510314875137, + "grad_norm": 5.756218508975568, + "learning_rate": 1.117716384240386e-05, + "loss": 0.3605, + "step": 4408 + }, + { + "epoch": 2.3935939196525515, + "grad_norm": 13.072320066260373, + "learning_rate": 1.1173671381943197e-05, + "loss": 1.1771, + "step": 4409 + }, + { + "epoch": 2.3941368078175898, + "grad_norm": 8.390483377503248, + "learning_rate": 1.1170178776320982e-05, + "loss": 0.484, + "step": 4410 + }, + { + "epoch": 2.3946796959826275, + "grad_norm": 9.168734569918955, + "learning_rate": 1.1166686025969196e-05, + "loss": 0.4637, + "step": 4411 + }, + { + "epoch": 2.3952225841476658, + "grad_norm": 11.945726197992448, + "learning_rate": 1.116319313131982e-05, + "loss": 0.6501, + "step": 4412 + }, + { + "epoch": 2.3957654723127035, + "grad_norm": 9.925140956274975, + "learning_rate": 1.1159700092804866e-05, + "loss": 0.5583, + "step": 4413 + }, + { + "epoch": 2.3963083604777418, + "grad_norm": 10.0377125619045, + "learning_rate": 1.1156206910856353e-05, + "loss": 0.5783, + "step": 4414 + }, + { + "epoch": 2.3968512486427795, + "grad_norm": 13.878494510161136, + "learning_rate": 1.115271358590633e-05, + "loss": 0.7029, + "step": 4415 + }, + { + "epoch": 2.3973941368078178, + "grad_norm": 8.414419089776398, + "learning_rate": 1.114922011838685e-05, + "loss": 0.4398, + "step": 4416 + }, + { + "epoch": 2.3979370249728555, + "grad_norm": 10.455474671185716, + "learning_rate": 1.1145726508729994e-05, + "loss": 0.5462, + "step": 4417 + }, + { + "epoch": 2.3984799131378938, + "grad_norm": 8.80305307358664, + "learning_rate": 1.1142232757367861e-05, + "loss": 0.3885, + "step": 4418 + }, + { + "epoch": 2.3990228013029316, + "grad_norm": 9.381551586854757, + "learning_rate": 1.1138738864732559e-05, + "loss": 0.6646, + "step": 4419 + }, + { + "epoch": 2.3995656894679698, + "grad_norm": 9.409317053539013, + "learning_rate": 1.113524483125622e-05, + "loss": 0.7772, + "step": 4420 + }, + { + "epoch": 2.4001085776330076, + "grad_norm": 7.414798924114551, + "learning_rate": 1.113175065737099e-05, + "loss": 0.4508, + "step": 4421 + }, + { + "epoch": 2.400651465798046, + "grad_norm": 10.931216387643186, + "learning_rate": 1.1128256343509032e-05, + "loss": 0.6434, + "step": 4422 + }, + { + "epoch": 2.4011943539630836, + "grad_norm": 9.811349462515302, + "learning_rate": 1.1124761890102535e-05, + "loss": 0.5104, + "step": 4423 + }, + { + "epoch": 2.401737242128122, + "grad_norm": 10.856233458965448, + "learning_rate": 1.1121267297583694e-05, + "loss": 0.6535, + "step": 4424 + }, + { + "epoch": 2.4022801302931596, + "grad_norm": 10.2119482085421, + "learning_rate": 1.1117772566384725e-05, + "loss": 0.593, + "step": 4425 + }, + { + "epoch": 2.402823018458198, + "grad_norm": 11.327455944037217, + "learning_rate": 1.1114277696937865e-05, + "loss": 0.9199, + "step": 4426 + }, + { + "epoch": 2.4033659066232356, + "grad_norm": 10.64004734338448, + "learning_rate": 1.1110782689675365e-05, + "loss": 0.793, + "step": 4427 + }, + { + "epoch": 2.403908794788274, + "grad_norm": 5.71867532259649, + "learning_rate": 1.110728754502949e-05, + "loss": 0.2726, + "step": 4428 + }, + { + "epoch": 2.4044516829533116, + "grad_norm": 10.33820400695821, + "learning_rate": 1.1103792263432525e-05, + "loss": 0.787, + "step": 4429 + }, + { + "epoch": 2.40499457111835, + "grad_norm": 10.02381386485252, + "learning_rate": 1.1100296845316777e-05, + "loss": 0.7701, + "step": 4430 + }, + { + "epoch": 2.4055374592833876, + "grad_norm": 12.111852512562214, + "learning_rate": 1.1096801291114561e-05, + "loss": 0.6123, + "step": 4431 + }, + { + "epoch": 2.406080347448426, + "grad_norm": 8.432777542688768, + "learning_rate": 1.1093305601258213e-05, + "loss": 0.4817, + "step": 4432 + }, + { + "epoch": 2.4066232356134636, + "grad_norm": 12.648769211741241, + "learning_rate": 1.1089809776180084e-05, + "loss": 0.7328, + "step": 4433 + }, + { + "epoch": 2.407166123778502, + "grad_norm": 9.260727895387426, + "learning_rate": 1.1086313816312546e-05, + "loss": 0.2951, + "step": 4434 + }, + { + "epoch": 2.4077090119435396, + "grad_norm": 7.801968230318713, + "learning_rate": 1.1082817722087981e-05, + "loss": 0.6204, + "step": 4435 + }, + { + "epoch": 2.408251900108578, + "grad_norm": 18.28104514211005, + "learning_rate": 1.10793214939388e-05, + "loss": 1.2717, + "step": 4436 + }, + { + "epoch": 2.4087947882736156, + "grad_norm": 8.074911353186037, + "learning_rate": 1.1075825132297414e-05, + "loss": 0.391, + "step": 4437 + }, + { + "epoch": 2.409337676438654, + "grad_norm": 10.223302145592744, + "learning_rate": 1.1072328637596259e-05, + "loss": 0.8764, + "step": 4438 + }, + { + "epoch": 2.4098805646036916, + "grad_norm": 8.909676311441572, + "learning_rate": 1.1068832010267792e-05, + "loss": 0.5866, + "step": 4439 + }, + { + "epoch": 2.41042345276873, + "grad_norm": 8.270875830864977, + "learning_rate": 1.1065335250744478e-05, + "loss": 0.4935, + "step": 4440 + }, + { + "epoch": 2.4109663409337676, + "grad_norm": 8.125694150069458, + "learning_rate": 1.1061838359458801e-05, + "loss": 0.5855, + "step": 4441 + }, + { + "epoch": 2.411509229098806, + "grad_norm": 9.00158598137614, + "learning_rate": 1.1058341336843264e-05, + "loss": 0.4464, + "step": 4442 + }, + { + "epoch": 2.4120521172638436, + "grad_norm": 10.568261973334824, + "learning_rate": 1.1054844183330385e-05, + "loss": 0.844, + "step": 4443 + }, + { + "epoch": 2.412595005428882, + "grad_norm": 10.29648068473729, + "learning_rate": 1.1051346899352693e-05, + "loss": 0.7083, + "step": 4444 + }, + { + "epoch": 2.4131378935939196, + "grad_norm": 10.658053637486233, + "learning_rate": 1.1047849485342744e-05, + "loss": 0.917, + "step": 4445 + }, + { + "epoch": 2.413680781758958, + "grad_norm": 10.85927844636501, + "learning_rate": 1.1044351941733097e-05, + "loss": 0.7737, + "step": 4446 + }, + { + "epoch": 2.4142236699239956, + "grad_norm": 13.105614857394716, + "learning_rate": 1.1040854268956343e-05, + "loss": 0.7626, + "step": 4447 + }, + { + "epoch": 2.414766558089034, + "grad_norm": 12.088913830608213, + "learning_rate": 1.1037356467445071e-05, + "loss": 0.7439, + "step": 4448 + }, + { + "epoch": 2.4153094462540716, + "grad_norm": 13.524031933483466, + "learning_rate": 1.1033858537631894e-05, + "loss": 0.6172, + "step": 4449 + }, + { + "epoch": 2.41585233441911, + "grad_norm": 12.25037908572634, + "learning_rate": 1.1030360479949452e-05, + "loss": 0.7959, + "step": 4450 + }, + { + "epoch": 2.4163952225841476, + "grad_norm": 10.26766943217781, + "learning_rate": 1.102686229483038e-05, + "loss": 0.8928, + "step": 4451 + }, + { + "epoch": 2.416938110749186, + "grad_norm": 10.66247714207056, + "learning_rate": 1.1023363982707345e-05, + "loss": 0.8498, + "step": 4452 + }, + { + "epoch": 2.4174809989142236, + "grad_norm": 9.37397925289073, + "learning_rate": 1.101986554401302e-05, + "loss": 0.5633, + "step": 4453 + }, + { + "epoch": 2.418023887079262, + "grad_norm": 8.656197915403546, + "learning_rate": 1.10163669791801e-05, + "loss": 0.6403, + "step": 4454 + }, + { + "epoch": 2.4185667752442996, + "grad_norm": 10.199878069425685, + "learning_rate": 1.101286828864129e-05, + "loss": 0.7139, + "step": 4455 + }, + { + "epoch": 2.419109663409338, + "grad_norm": 8.176125775602177, + "learning_rate": 1.1009369472829316e-05, + "loss": 0.5322, + "step": 4456 + }, + { + "epoch": 2.4196525515743756, + "grad_norm": 9.441655350085792, + "learning_rate": 1.1005870532176918e-05, + "loss": 0.8574, + "step": 4457 + }, + { + "epoch": 2.420195439739414, + "grad_norm": 10.487844747641686, + "learning_rate": 1.100237146711685e-05, + "loss": 0.7399, + "step": 4458 + }, + { + "epoch": 2.4207383279044516, + "grad_norm": 8.456961156540515, + "learning_rate": 1.0998872278081883e-05, + "loss": 0.5795, + "step": 4459 + }, + { + "epoch": 2.42128121606949, + "grad_norm": 10.247575624097415, + "learning_rate": 1.0995372965504798e-05, + "loss": 0.7048, + "step": 4460 + }, + { + "epoch": 2.4218241042345277, + "grad_norm": 8.127194271974078, + "learning_rate": 1.09918735298184e-05, + "loss": 0.6339, + "step": 4461 + }, + { + "epoch": 2.422366992399566, + "grad_norm": 11.080151666019892, + "learning_rate": 1.0988373971455502e-05, + "loss": 0.6, + "step": 4462 + }, + { + "epoch": 2.4229098805646037, + "grad_norm": 9.249791609753641, + "learning_rate": 1.0984874290848935e-05, + "loss": 0.5823, + "step": 4463 + }, + { + "epoch": 2.423452768729642, + "grad_norm": 8.673439204574109, + "learning_rate": 1.0981374488431546e-05, + "loss": 0.4486, + "step": 4464 + }, + { + "epoch": 2.4239956568946797, + "grad_norm": 8.530441262511525, + "learning_rate": 1.0977874564636199e-05, + "loss": 0.4685, + "step": 4465 + }, + { + "epoch": 2.424538545059718, + "grad_norm": 11.975099667402777, + "learning_rate": 1.0974374519895765e-05, + "loss": 0.5608, + "step": 4466 + }, + { + "epoch": 2.4250814332247557, + "grad_norm": 12.28650683165383, + "learning_rate": 1.097087435464314e-05, + "loss": 0.6428, + "step": 4467 + }, + { + "epoch": 2.425624321389794, + "grad_norm": 11.443904338187624, + "learning_rate": 1.0967374069311224e-05, + "loss": 0.5093, + "step": 4468 + }, + { + "epoch": 2.4261672095548317, + "grad_norm": 12.11642038793139, + "learning_rate": 1.0963873664332944e-05, + "loss": 0.7573, + "step": 4469 + }, + { + "epoch": 2.42671009771987, + "grad_norm": 12.716855768902285, + "learning_rate": 1.0960373140141232e-05, + "loss": 1.0742, + "step": 4470 + }, + { + "epoch": 2.4272529858849077, + "grad_norm": 8.060615842022353, + "learning_rate": 1.095687249716904e-05, + "loss": 0.5299, + "step": 4471 + }, + { + "epoch": 2.427795874049946, + "grad_norm": 10.663178608260651, + "learning_rate": 1.0953371735849329e-05, + "loss": 0.6575, + "step": 4472 + }, + { + "epoch": 2.4283387622149837, + "grad_norm": 11.883215571850693, + "learning_rate": 1.0949870856615088e-05, + "loss": 0.7516, + "step": 4473 + }, + { + "epoch": 2.428881650380022, + "grad_norm": 8.725361323128087, + "learning_rate": 1.0946369859899299e-05, + "loss": 0.5543, + "step": 4474 + }, + { + "epoch": 2.4294245385450597, + "grad_norm": 7.964297132373328, + "learning_rate": 1.0942868746134983e-05, + "loss": 0.3799, + "step": 4475 + }, + { + "epoch": 2.429967426710098, + "grad_norm": 11.129935043446988, + "learning_rate": 1.0939367515755155e-05, + "loss": 0.5274, + "step": 4476 + }, + { + "epoch": 2.4305103148751357, + "grad_norm": 8.84506268400313, + "learning_rate": 1.0935866169192858e-05, + "loss": 0.6314, + "step": 4477 + }, + { + "epoch": 2.431053203040174, + "grad_norm": 9.898346545614544, + "learning_rate": 1.0932364706881138e-05, + "loss": 0.4733, + "step": 4478 + }, + { + "epoch": 2.4315960912052117, + "grad_norm": 12.708528067939865, + "learning_rate": 1.092886312925307e-05, + "loss": 1.1591, + "step": 4479 + }, + { + "epoch": 2.43213897937025, + "grad_norm": 8.852913949596426, + "learning_rate": 1.0925361436741726e-05, + "loss": 0.4934, + "step": 4480 + }, + { + "epoch": 2.4326818675352877, + "grad_norm": 9.976618553528468, + "learning_rate": 1.0921859629780205e-05, + "loss": 0.5726, + "step": 4481 + }, + { + "epoch": 2.433224755700326, + "grad_norm": 10.591185923462703, + "learning_rate": 1.0918357708801617e-05, + "loss": 0.9099, + "step": 4482 + }, + { + "epoch": 2.4337676438653637, + "grad_norm": 9.782319605062739, + "learning_rate": 1.0914855674239083e-05, + "loss": 0.7016, + "step": 4483 + }, + { + "epoch": 2.434310532030402, + "grad_norm": 7.962337485760134, + "learning_rate": 1.0911353526525741e-05, + "loss": 0.3899, + "step": 4484 + }, + { + "epoch": 2.4348534201954397, + "grad_norm": 10.926665567713723, + "learning_rate": 1.0907851266094742e-05, + "loss": 1.0416, + "step": 4485 + }, + { + "epoch": 2.435396308360478, + "grad_norm": 10.774706653267238, + "learning_rate": 1.0904348893379252e-05, + "loss": 0.8539, + "step": 4486 + }, + { + "epoch": 2.4359391965255157, + "grad_norm": 9.962550119554402, + "learning_rate": 1.0900846408812449e-05, + "loss": 0.6202, + "step": 4487 + }, + { + "epoch": 2.436482084690554, + "grad_norm": 7.425101710467455, + "learning_rate": 1.0897343812827528e-05, + "loss": 0.4138, + "step": 4488 + }, + { + "epoch": 2.4370249728555917, + "grad_norm": 11.159990654051901, + "learning_rate": 1.0893841105857695e-05, + "loss": 0.7927, + "step": 4489 + }, + { + "epoch": 2.4375678610206295, + "grad_norm": 8.584412414000917, + "learning_rate": 1.0890338288336168e-05, + "loss": 0.4596, + "step": 4490 + }, + { + "epoch": 2.4381107491856677, + "grad_norm": 13.182969196780135, + "learning_rate": 1.0886835360696182e-05, + "loss": 0.7052, + "step": 4491 + }, + { + "epoch": 2.438653637350706, + "grad_norm": 12.643144928400368, + "learning_rate": 1.0883332323370986e-05, + "loss": 0.7273, + "step": 4492 + }, + { + "epoch": 2.4391965255157437, + "grad_norm": 11.758017246786887, + "learning_rate": 1.087982917679384e-05, + "loss": 0.6389, + "step": 4493 + }, + { + "epoch": 2.4397394136807815, + "grad_norm": 10.332831381965816, + "learning_rate": 1.0876325921398016e-05, + "loss": 0.49, + "step": 4494 + }, + { + "epoch": 2.4402823018458197, + "grad_norm": 12.79909493300538, + "learning_rate": 1.0872822557616811e-05, + "loss": 1.027, + "step": 4495 + }, + { + "epoch": 2.440825190010858, + "grad_norm": 10.225359261113429, + "learning_rate": 1.0869319085883518e-05, + "loss": 0.6319, + "step": 4496 + }, + { + "epoch": 2.4413680781758957, + "grad_norm": 10.600390533226815, + "learning_rate": 1.0865815506631456e-05, + "loss": 0.5513, + "step": 4497 + }, + { + "epoch": 2.4419109663409335, + "grad_norm": 9.069709981087057, + "learning_rate": 1.0862311820293951e-05, + "loss": 0.4358, + "step": 4498 + }, + { + "epoch": 2.4424538545059717, + "grad_norm": 10.228930326876846, + "learning_rate": 1.0858808027304346e-05, + "loss": 1.0149, + "step": 4499 + }, + { + "epoch": 2.44299674267101, + "grad_norm": 6.981580424693794, + "learning_rate": 1.0855304128095997e-05, + "loss": 0.3823, + "step": 4500 + }, + { + "epoch": 2.4435396308360477, + "grad_norm": 12.347328486710586, + "learning_rate": 1.085180012310227e-05, + "loss": 0.8202, + "step": 4501 + }, + { + "epoch": 2.4440825190010855, + "grad_norm": 7.794475194787575, + "learning_rate": 1.0848296012756545e-05, + "loss": 0.4972, + "step": 4502 + }, + { + "epoch": 2.4446254071661238, + "grad_norm": 8.640639888495457, + "learning_rate": 1.0844791797492218e-05, + "loss": 0.4174, + "step": 4503 + }, + { + "epoch": 2.445168295331162, + "grad_norm": 9.363787480425108, + "learning_rate": 1.0841287477742695e-05, + "loss": 0.5783, + "step": 4504 + }, + { + "epoch": 2.4457111834961998, + "grad_norm": 9.455569437628846, + "learning_rate": 1.0837783053941396e-05, + "loss": 0.5611, + "step": 4505 + }, + { + "epoch": 2.4462540716612375, + "grad_norm": 7.587727857256139, + "learning_rate": 1.083427852652175e-05, + "loss": 0.4129, + "step": 4506 + }, + { + "epoch": 2.4467969598262758, + "grad_norm": 11.579317957056839, + "learning_rate": 1.083077389591721e-05, + "loss": 0.6628, + "step": 4507 + }, + { + "epoch": 2.447339847991314, + "grad_norm": 9.130000079525795, + "learning_rate": 1.0827269162561229e-05, + "loss": 0.4626, + "step": 4508 + }, + { + "epoch": 2.4478827361563518, + "grad_norm": 8.791608822057356, + "learning_rate": 1.0823764326887283e-05, + "loss": 0.7441, + "step": 4509 + }, + { + "epoch": 2.4484256243213895, + "grad_norm": 8.698379661142646, + "learning_rate": 1.0820259389328852e-05, + "loss": 0.4924, + "step": 4510 + }, + { + "epoch": 2.4489685124864278, + "grad_norm": 12.811407852860862, + "learning_rate": 1.0816754350319427e-05, + "loss": 0.9829, + "step": 4511 + }, + { + "epoch": 2.449511400651466, + "grad_norm": 11.588653893201888, + "learning_rate": 1.0813249210292525e-05, + "loss": 0.7218, + "step": 4512 + }, + { + "epoch": 2.450054288816504, + "grad_norm": 12.334703169006005, + "learning_rate": 1.0809743969681665e-05, + "loss": 0.8737, + "step": 4513 + }, + { + "epoch": 2.4505971769815416, + "grad_norm": 12.583313163331177, + "learning_rate": 1.080623862892038e-05, + "loss": 0.7226, + "step": 4514 + }, + { + "epoch": 2.45114006514658, + "grad_norm": 12.11103490873811, + "learning_rate": 1.0802733188442216e-05, + "loss": 0.7867, + "step": 4515 + }, + { + "epoch": 2.451682953311618, + "grad_norm": 11.222788743446143, + "learning_rate": 1.0799227648680732e-05, + "loss": 0.5457, + "step": 4516 + }, + { + "epoch": 2.452225841476656, + "grad_norm": 7.685503695913846, + "learning_rate": 1.07957220100695e-05, + "loss": 0.6483, + "step": 4517 + }, + { + "epoch": 2.4527687296416936, + "grad_norm": 8.407197016108878, + "learning_rate": 1.0792216273042099e-05, + "loss": 0.4895, + "step": 4518 + }, + { + "epoch": 2.453311617806732, + "grad_norm": 9.426708870733849, + "learning_rate": 1.0788710438032131e-05, + "loss": 1.0719, + "step": 4519 + }, + { + "epoch": 2.45385450597177, + "grad_norm": 13.638918388289357, + "learning_rate": 1.0785204505473197e-05, + "loss": 0.8828, + "step": 4520 + }, + { + "epoch": 2.454397394136808, + "grad_norm": 12.478693804182608, + "learning_rate": 1.0781698475798917e-05, + "loss": 0.6527, + "step": 4521 + }, + { + "epoch": 2.4549402823018456, + "grad_norm": 10.20424250017949, + "learning_rate": 1.0778192349442923e-05, + "loss": 0.9303, + "step": 4522 + }, + { + "epoch": 2.455483170466884, + "grad_norm": 12.175885097025716, + "learning_rate": 1.0774686126838863e-05, + "loss": 0.6391, + "step": 4523 + }, + { + "epoch": 2.456026058631922, + "grad_norm": 10.568732307533892, + "learning_rate": 1.0771179808420385e-05, + "loss": 0.5524, + "step": 4524 + }, + { + "epoch": 2.45656894679696, + "grad_norm": 7.314447097275954, + "learning_rate": 1.0767673394621159e-05, + "loss": 0.4976, + "step": 4525 + }, + { + "epoch": 2.4571118349619976, + "grad_norm": 11.397811881243696, + "learning_rate": 1.0764166885874862e-05, + "loss": 0.4918, + "step": 4526 + }, + { + "epoch": 2.457654723127036, + "grad_norm": 9.691957338181856, + "learning_rate": 1.0760660282615191e-05, + "loss": 0.6725, + "step": 4527 + }, + { + "epoch": 2.458197611292074, + "grad_norm": 7.510379855172465, + "learning_rate": 1.0757153585275844e-05, + "loss": 0.4684, + "step": 4528 + }, + { + "epoch": 2.458740499457112, + "grad_norm": 7.751570484202345, + "learning_rate": 1.0753646794290534e-05, + "loss": 0.5495, + "step": 4529 + }, + { + "epoch": 2.4592833876221496, + "grad_norm": 7.390182612219503, + "learning_rate": 1.0750139910092987e-05, + "loss": 0.4538, + "step": 4530 + }, + { + "epoch": 2.459826275787188, + "grad_norm": 10.180132351193016, + "learning_rate": 1.074663293311694e-05, + "loss": 0.5579, + "step": 4531 + }, + { + "epoch": 2.460369163952226, + "grad_norm": 8.909847677295557, + "learning_rate": 1.074312586379614e-05, + "loss": 0.5023, + "step": 4532 + }, + { + "epoch": 2.460912052117264, + "grad_norm": 10.091297535262147, + "learning_rate": 1.073961870256435e-05, + "loss": 0.7068, + "step": 4533 + }, + { + "epoch": 2.4614549402823016, + "grad_norm": 11.778234555569338, + "learning_rate": 1.0736111449855341e-05, + "loss": 0.7687, + "step": 4534 + }, + { + "epoch": 2.46199782844734, + "grad_norm": 11.871444812620142, + "learning_rate": 1.0732604106102895e-05, + "loss": 0.8333, + "step": 4535 + }, + { + "epoch": 2.462540716612378, + "grad_norm": 6.164790402663993, + "learning_rate": 1.0729096671740806e-05, + "loss": 0.3601, + "step": 4536 + }, + { + "epoch": 2.463083604777416, + "grad_norm": 9.524060445013603, + "learning_rate": 1.0725589147202877e-05, + "loss": 0.6859, + "step": 4537 + }, + { + "epoch": 2.4636264929424536, + "grad_norm": 11.266171306725775, + "learning_rate": 1.0722081532922925e-05, + "loss": 0.7459, + "step": 4538 + }, + { + "epoch": 2.464169381107492, + "grad_norm": 8.945835373858307, + "learning_rate": 1.0718573829334782e-05, + "loss": 0.3787, + "step": 4539 + }, + { + "epoch": 2.46471226927253, + "grad_norm": 6.838806611638338, + "learning_rate": 1.071506603687228e-05, + "loss": 0.4995, + "step": 4540 + }, + { + "epoch": 2.465255157437568, + "grad_norm": 11.246624755560871, + "learning_rate": 1.071155815596927e-05, + "loss": 0.8699, + "step": 4541 + }, + { + "epoch": 2.4657980456026056, + "grad_norm": 7.324342918236285, + "learning_rate": 1.0708050187059615e-05, + "loss": 0.4778, + "step": 4542 + }, + { + "epoch": 2.466340933767644, + "grad_norm": 10.063475303955434, + "learning_rate": 1.0704542130577185e-05, + "loss": 0.5774, + "step": 4543 + }, + { + "epoch": 2.466883821932682, + "grad_norm": 12.644825932835714, + "learning_rate": 1.070103398695586e-05, + "loss": 0.8158, + "step": 4544 + }, + { + "epoch": 2.46742671009772, + "grad_norm": 10.332286780133, + "learning_rate": 1.0697525756629531e-05, + "loss": 0.6518, + "step": 4545 + }, + { + "epoch": 2.4679695982627576, + "grad_norm": 9.98899797230979, + "learning_rate": 1.069401744003211e-05, + "loss": 0.6167, + "step": 4546 + }, + { + "epoch": 2.468512486427796, + "grad_norm": 13.445015438013797, + "learning_rate": 1.0690509037597502e-05, + "loss": 0.8229, + "step": 4547 + }, + { + "epoch": 2.469055374592834, + "grad_norm": 13.637667335089724, + "learning_rate": 1.0687000549759641e-05, + "loss": 0.7186, + "step": 4548 + }, + { + "epoch": 2.469598262757872, + "grad_norm": 10.021629994296763, + "learning_rate": 1.0683491976952454e-05, + "loss": 0.6209, + "step": 4549 + }, + { + "epoch": 2.4701411509229096, + "grad_norm": 10.510874896443969, + "learning_rate": 1.067998331960989e-05, + "loss": 0.9009, + "step": 4550 + }, + { + "epoch": 2.470684039087948, + "grad_norm": 9.53841539843291, + "learning_rate": 1.0676474578165903e-05, + "loss": 0.7187, + "step": 4551 + }, + { + "epoch": 2.471226927252986, + "grad_norm": 11.693787891306421, + "learning_rate": 1.0672965753054466e-05, + "loss": 0.5568, + "step": 4552 + }, + { + "epoch": 2.471769815418024, + "grad_norm": 11.523811853237843, + "learning_rate": 1.0669456844709549e-05, + "loss": 0.8465, + "step": 4553 + }, + { + "epoch": 2.4723127035830617, + "grad_norm": 9.165552336797202, + "learning_rate": 1.0665947853565144e-05, + "loss": 0.7148, + "step": 4554 + }, + { + "epoch": 2.4728555917481, + "grad_norm": 10.390907422337587, + "learning_rate": 1.0662438780055247e-05, + "loss": 0.5271, + "step": 4555 + }, + { + "epoch": 2.473398479913138, + "grad_norm": 10.514660404488053, + "learning_rate": 1.0658929624613866e-05, + "loss": 0.7972, + "step": 4556 + }, + { + "epoch": 2.473941368078176, + "grad_norm": 8.804604897853023, + "learning_rate": 1.065542038767502e-05, + "loss": 0.4391, + "step": 4557 + }, + { + "epoch": 2.4744842562432137, + "grad_norm": 12.58202482779834, + "learning_rate": 1.0651911069672733e-05, + "loss": 0.601, + "step": 4558 + }, + { + "epoch": 2.475027144408252, + "grad_norm": 8.510520469826693, + "learning_rate": 1.0648401671041048e-05, + "loss": 0.6638, + "step": 4559 + }, + { + "epoch": 2.47557003257329, + "grad_norm": 10.132731434552191, + "learning_rate": 1.0644892192214012e-05, + "loss": 0.547, + "step": 4560 + }, + { + "epoch": 2.476112920738328, + "grad_norm": 9.486912806484277, + "learning_rate": 1.0641382633625678e-05, + "loss": 0.6903, + "step": 4561 + }, + { + "epoch": 2.4766558089033657, + "grad_norm": 8.656529900389968, + "learning_rate": 1.0637872995710122e-05, + "loss": 0.6445, + "step": 4562 + }, + { + "epoch": 2.477198697068404, + "grad_norm": 8.377471382601675, + "learning_rate": 1.0634363278901413e-05, + "loss": 0.475, + "step": 4563 + }, + { + "epoch": 2.477741585233442, + "grad_norm": 11.941789514685023, + "learning_rate": 1.0630853483633643e-05, + "loss": 0.8705, + "step": 4564 + }, + { + "epoch": 2.47828447339848, + "grad_norm": 8.551632836218364, + "learning_rate": 1.0627343610340912e-05, + "loss": 0.3577, + "step": 4565 + }, + { + "epoch": 2.4788273615635177, + "grad_norm": 10.220220103336683, + "learning_rate": 1.0623833659457322e-05, + "loss": 0.7002, + "step": 4566 + }, + { + "epoch": 2.479370249728556, + "grad_norm": 10.783849224421678, + "learning_rate": 1.0620323631416994e-05, + "loss": 0.6969, + "step": 4567 + }, + { + "epoch": 2.479913137893594, + "grad_norm": 9.018283325771, + "learning_rate": 1.0616813526654047e-05, + "loss": 0.6844, + "step": 4568 + }, + { + "epoch": 2.480456026058632, + "grad_norm": 9.570551013210757, + "learning_rate": 1.061330334560262e-05, + "loss": 0.5063, + "step": 4569 + }, + { + "epoch": 2.4809989142236697, + "grad_norm": 10.106289919073877, + "learning_rate": 1.0609793088696856e-05, + "loss": 0.6131, + "step": 4570 + }, + { + "epoch": 2.481541802388708, + "grad_norm": 10.842801269455231, + "learning_rate": 1.060628275637091e-05, + "loss": 0.6637, + "step": 4571 + }, + { + "epoch": 2.482084690553746, + "grad_norm": 11.953941591792438, + "learning_rate": 1.0602772349058952e-05, + "loss": 0.7631, + "step": 4572 + }, + { + "epoch": 2.482627578718784, + "grad_norm": 8.194683405468492, + "learning_rate": 1.0599261867195146e-05, + "loss": 0.5936, + "step": 4573 + }, + { + "epoch": 2.4831704668838217, + "grad_norm": 10.216659522745752, + "learning_rate": 1.0595751311213678e-05, + "loss": 0.7017, + "step": 4574 + }, + { + "epoch": 2.48371335504886, + "grad_norm": 11.0782985653065, + "learning_rate": 1.0592240681548738e-05, + "loss": 0.7285, + "step": 4575 + }, + { + "epoch": 2.484256243213898, + "grad_norm": 7.740854250685081, + "learning_rate": 1.0588729978634527e-05, + "loss": 0.4544, + "step": 4576 + }, + { + "epoch": 2.484799131378936, + "grad_norm": 6.9863990778515745, + "learning_rate": 1.0585219202905255e-05, + "loss": 0.2831, + "step": 4577 + }, + { + "epoch": 2.4853420195439737, + "grad_norm": 9.687327462597597, + "learning_rate": 1.0581708354795136e-05, + "loss": 0.5028, + "step": 4578 + }, + { + "epoch": 2.485884907709012, + "grad_norm": 8.009180785791076, + "learning_rate": 1.0578197434738405e-05, + "loss": 0.342, + "step": 4579 + }, + { + "epoch": 2.48642779587405, + "grad_norm": 8.101662777133065, + "learning_rate": 1.057468644316929e-05, + "loss": 0.4146, + "step": 4580 + }, + { + "epoch": 2.486970684039088, + "grad_norm": 9.370936142076715, + "learning_rate": 1.0571175380522041e-05, + "loss": 0.6006, + "step": 4581 + }, + { + "epoch": 2.4875135722041257, + "grad_norm": 11.164770907813724, + "learning_rate": 1.0567664247230911e-05, + "loss": 0.6787, + "step": 4582 + }, + { + "epoch": 2.488056460369164, + "grad_norm": 12.573930877103525, + "learning_rate": 1.056415304373016e-05, + "loss": 0.8051, + "step": 4583 + }, + { + "epoch": 2.488599348534202, + "grad_norm": 8.395060833285095, + "learning_rate": 1.0560641770454063e-05, + "loss": 0.5596, + "step": 4584 + }, + { + "epoch": 2.48914223669924, + "grad_norm": 11.015478890741145, + "learning_rate": 1.0557130427836899e-05, + "loss": 0.5442, + "step": 4585 + }, + { + "epoch": 2.4896851248642777, + "grad_norm": 9.420432906299153, + "learning_rate": 1.0553619016312955e-05, + "loss": 0.4917, + "step": 4586 + }, + { + "epoch": 2.490228013029316, + "grad_norm": 11.026358328990858, + "learning_rate": 1.0550107536316532e-05, + "loss": 0.4291, + "step": 4587 + }, + { + "epoch": 2.490770901194354, + "grad_norm": 16.545112067555284, + "learning_rate": 1.0546595988281928e-05, + "loss": 0.9417, + "step": 4588 + }, + { + "epoch": 2.491313789359392, + "grad_norm": 7.252757680637398, + "learning_rate": 1.0543084372643462e-05, + "loss": 0.7908, + "step": 4589 + }, + { + "epoch": 2.4918566775244297, + "grad_norm": 7.598562688636616, + "learning_rate": 1.0539572689835455e-05, + "loss": 0.3229, + "step": 4590 + }, + { + "epoch": 2.492399565689468, + "grad_norm": 12.449529704914275, + "learning_rate": 1.0536060940292238e-05, + "loss": 0.7373, + "step": 4591 + }, + { + "epoch": 2.492942453854506, + "grad_norm": 11.706159828307792, + "learning_rate": 1.0532549124448151e-05, + "loss": 0.6525, + "step": 4592 + }, + { + "epoch": 2.493485342019544, + "grad_norm": 8.000677691767947, + "learning_rate": 1.0529037242737538e-05, + "loss": 0.3613, + "step": 4593 + }, + { + "epoch": 2.4940282301845818, + "grad_norm": 9.529463393752437, + "learning_rate": 1.0525525295594755e-05, + "loss": 0.5921, + "step": 4594 + }, + { + "epoch": 2.49457111834962, + "grad_norm": 11.71769497271457, + "learning_rate": 1.0522013283454169e-05, + "loss": 0.5507, + "step": 4595 + }, + { + "epoch": 2.495114006514658, + "grad_norm": 8.546908524296436, + "learning_rate": 1.0518501206750145e-05, + "loss": 0.4564, + "step": 4596 + }, + { + "epoch": 2.495656894679696, + "grad_norm": 14.968617219210655, + "learning_rate": 1.0514989065917068e-05, + "loss": 0.5593, + "step": 4597 + }, + { + "epoch": 2.4961997828447338, + "grad_norm": 13.0682164313941, + "learning_rate": 1.051147686138932e-05, + "loss": 0.8812, + "step": 4598 + }, + { + "epoch": 2.496742671009772, + "grad_norm": 15.01014358585534, + "learning_rate": 1.05079645936013e-05, + "loss": 1.0756, + "step": 4599 + }, + { + "epoch": 2.49728555917481, + "grad_norm": 7.598240677419264, + "learning_rate": 1.0504452262987407e-05, + "loss": 0.3936, + "step": 4600 + }, + { + "epoch": 2.497828447339848, + "grad_norm": 10.39263281879832, + "learning_rate": 1.0500939869982058e-05, + "loss": 0.4944, + "step": 4601 + }, + { + "epoch": 2.4983713355048858, + "grad_norm": 12.104880157441473, + "learning_rate": 1.0497427415019665e-05, + "loss": 0.7498, + "step": 4602 + }, + { + "epoch": 2.498914223669924, + "grad_norm": 9.18522035093318, + "learning_rate": 1.0493914898534656e-05, + "loss": 0.6921, + "step": 4603 + }, + { + "epoch": 2.499457111834962, + "grad_norm": 15.812637613525052, + "learning_rate": 1.0490402320961466e-05, + "loss": 0.937, + "step": 4604 + }, + { + "epoch": 2.5, + "grad_norm": 11.400921058050775, + "learning_rate": 1.0486889682734535e-05, + "loss": 0.6453, + "step": 4605 + }, + { + "epoch": 2.500542888165038, + "grad_norm": 8.270129268701748, + "learning_rate": 1.0483376984288314e-05, + "loss": 0.3921, + "step": 4606 + }, + { + "epoch": 2.501085776330076, + "grad_norm": 15.017103371329522, + "learning_rate": 1.0479864226057258e-05, + "loss": 1.0397, + "step": 4607 + }, + { + "epoch": 2.5016286644951142, + "grad_norm": 8.346121650934089, + "learning_rate": 1.047635140847583e-05, + "loss": 0.3872, + "step": 4608 + }, + { + "epoch": 2.502171552660152, + "grad_norm": 14.17791420098929, + "learning_rate": 1.0472838531978495e-05, + "loss": 0.6932, + "step": 4609 + }, + { + "epoch": 2.50271444082519, + "grad_norm": 12.940166215438651, + "learning_rate": 1.046932559699974e-05, + "loss": 0.7045, + "step": 4610 + }, + { + "epoch": 2.503257328990228, + "grad_norm": 8.154813733233013, + "learning_rate": 1.0465812603974048e-05, + "loss": 0.4225, + "step": 4611 + }, + { + "epoch": 2.5038002171552662, + "grad_norm": 13.356750387494248, + "learning_rate": 1.0462299553335911e-05, + "loss": 0.7951, + "step": 4612 + }, + { + "epoch": 2.504343105320304, + "grad_norm": 8.683404141286427, + "learning_rate": 1.0458786445519827e-05, + "loss": 0.5871, + "step": 4613 + }, + { + "epoch": 2.504885993485342, + "grad_norm": 5.750758875120564, + "learning_rate": 1.0455273280960306e-05, + "loss": 0.3815, + "step": 4614 + }, + { + "epoch": 2.50542888165038, + "grad_norm": 9.328766968136149, + "learning_rate": 1.0451760060091857e-05, + "loss": 0.6451, + "step": 4615 + }, + { + "epoch": 2.5059717698154182, + "grad_norm": 9.704349389930364, + "learning_rate": 1.0448246783349002e-05, + "loss": 0.5136, + "step": 4616 + }, + { + "epoch": 2.506514657980456, + "grad_norm": 10.438938086465326, + "learning_rate": 1.0444733451166277e-05, + "loss": 0.6278, + "step": 4617 + }, + { + "epoch": 2.507057546145494, + "grad_norm": 11.019760442005925, + "learning_rate": 1.0441220063978208e-05, + "loss": 0.7477, + "step": 4618 + }, + { + "epoch": 2.507600434310532, + "grad_norm": 11.738067010881544, + "learning_rate": 1.0437706622219337e-05, + "loss": 0.5768, + "step": 4619 + }, + { + "epoch": 2.5081433224755703, + "grad_norm": 9.601242936349621, + "learning_rate": 1.0434193126324215e-05, + "loss": 0.4236, + "step": 4620 + }, + { + "epoch": 2.508686210640608, + "grad_norm": 8.041097216565952, + "learning_rate": 1.0430679576727392e-05, + "loss": 0.494, + "step": 4621 + }, + { + "epoch": 2.509229098805646, + "grad_norm": 11.144286090694623, + "learning_rate": 1.0427165973863434e-05, + "loss": 0.7882, + "step": 4622 + }, + { + "epoch": 2.509771986970684, + "grad_norm": 14.369854225910128, + "learning_rate": 1.0423652318166905e-05, + "loss": 0.7112, + "step": 4623 + }, + { + "epoch": 2.5103148751357223, + "grad_norm": 9.688103222900846, + "learning_rate": 1.0420138610072386e-05, + "loss": 0.4564, + "step": 4624 + }, + { + "epoch": 2.51085776330076, + "grad_norm": 8.2914031160658, + "learning_rate": 1.0416624850014454e-05, + "loss": 0.4106, + "step": 4625 + }, + { + "epoch": 2.511400651465798, + "grad_norm": 13.145683230031937, + "learning_rate": 1.0413111038427696e-05, + "loss": 0.905, + "step": 4626 + }, + { + "epoch": 2.511943539630836, + "grad_norm": 10.544391990984453, + "learning_rate": 1.0409597175746709e-05, + "loss": 0.5791, + "step": 4627 + }, + { + "epoch": 2.5124864277958743, + "grad_norm": 8.830789770145405, + "learning_rate": 1.0406083262406084e-05, + "loss": 0.4806, + "step": 4628 + }, + { + "epoch": 2.513029315960912, + "grad_norm": 9.186020814962909, + "learning_rate": 1.0402569298840437e-05, + "loss": 0.7227, + "step": 4629 + }, + { + "epoch": 2.51357220412595, + "grad_norm": 8.395511535015437, + "learning_rate": 1.0399055285484378e-05, + "loss": 0.5372, + "step": 4630 + }, + { + "epoch": 2.514115092290988, + "grad_norm": 9.969741133621088, + "learning_rate": 1.0395541222772524e-05, + "loss": 0.5151, + "step": 4631 + }, + { + "epoch": 2.5146579804560263, + "grad_norm": 11.64674593889149, + "learning_rate": 1.0392027111139505e-05, + "loss": 0.6159, + "step": 4632 + }, + { + "epoch": 2.515200868621064, + "grad_norm": 12.194306814356107, + "learning_rate": 1.0388512951019945e-05, + "loss": 0.6977, + "step": 4633 + }, + { + "epoch": 2.515743756786102, + "grad_norm": 9.063328798639105, + "learning_rate": 1.0384998742848485e-05, + "loss": 0.6518, + "step": 4634 + }, + { + "epoch": 2.51628664495114, + "grad_norm": 8.709419510130173, + "learning_rate": 1.0381484487059764e-05, + "loss": 0.5571, + "step": 4635 + }, + { + "epoch": 2.5168295331161783, + "grad_norm": 10.875178380627338, + "learning_rate": 1.0377970184088438e-05, + "loss": 0.8384, + "step": 4636 + }, + { + "epoch": 2.517372421281216, + "grad_norm": 8.42478873312803, + "learning_rate": 1.0374455834369155e-05, + "loss": 0.6209, + "step": 4637 + }, + { + "epoch": 2.517915309446254, + "grad_norm": 13.875461781038107, + "learning_rate": 1.0370941438336577e-05, + "loss": 0.8787, + "step": 4638 + }, + { + "epoch": 2.518458197611292, + "grad_norm": 11.372446219871874, + "learning_rate": 1.036742699642537e-05, + "loss": 0.569, + "step": 4639 + }, + { + "epoch": 2.5190010857763303, + "grad_norm": 10.725098253853016, + "learning_rate": 1.036391250907021e-05, + "loss": 0.8431, + "step": 4640 + }, + { + "epoch": 2.519543973941368, + "grad_norm": 8.268642654673096, + "learning_rate": 1.0360397976705767e-05, + "loss": 0.5833, + "step": 4641 + }, + { + "epoch": 2.520086862106406, + "grad_norm": 10.560802722537995, + "learning_rate": 1.0356883399766731e-05, + "loss": 0.7495, + "step": 4642 + }, + { + "epoch": 2.520629750271444, + "grad_norm": 11.789754393456334, + "learning_rate": 1.0353368778687784e-05, + "loss": 0.644, + "step": 4643 + }, + { + "epoch": 2.5211726384364823, + "grad_norm": 10.236138559776574, + "learning_rate": 1.0349854113903625e-05, + "loss": 0.6544, + "step": 4644 + }, + { + "epoch": 2.52171552660152, + "grad_norm": 8.552964039446971, + "learning_rate": 1.0346339405848953e-05, + "loss": 0.5093, + "step": 4645 + }, + { + "epoch": 2.522258414766558, + "grad_norm": 8.301860195209034, + "learning_rate": 1.0342824654958471e-05, + "loss": 0.6285, + "step": 4646 + }, + { + "epoch": 2.522801302931596, + "grad_norm": 9.238558119037732, + "learning_rate": 1.0339309861666889e-05, + "loss": 0.6519, + "step": 4647 + }, + { + "epoch": 2.5233441910966343, + "grad_norm": 9.840812219132841, + "learning_rate": 1.0335795026408922e-05, + "loss": 0.5874, + "step": 4648 + }, + { + "epoch": 2.523887079261672, + "grad_norm": 8.108461441633077, + "learning_rate": 1.0332280149619291e-05, + "loss": 0.5793, + "step": 4649 + }, + { + "epoch": 2.52442996742671, + "grad_norm": 9.446694020377457, + "learning_rate": 1.032876523173272e-05, + "loss": 0.607, + "step": 4650 + }, + { + "epoch": 2.524972855591748, + "grad_norm": 11.88681485670853, + "learning_rate": 1.0325250273183944e-05, + "loss": 1.0745, + "step": 4651 + }, + { + "epoch": 2.5255157437567863, + "grad_norm": 8.347193432264572, + "learning_rate": 1.0321735274407694e-05, + "loss": 0.8431, + "step": 4652 + }, + { + "epoch": 2.526058631921824, + "grad_norm": 9.639244837832804, + "learning_rate": 1.0318220235838715e-05, + "loss": 0.7827, + "step": 4653 + }, + { + "epoch": 2.526601520086862, + "grad_norm": 12.239750504764533, + "learning_rate": 1.0314705157911744e-05, + "loss": 0.6735, + "step": 4654 + }, + { + "epoch": 2.5271444082519, + "grad_norm": 15.826940901953726, + "learning_rate": 1.0311190041061543e-05, + "loss": 0.9182, + "step": 4655 + }, + { + "epoch": 2.5276872964169383, + "grad_norm": 11.015508331511823, + "learning_rate": 1.0307674885722863e-05, + "loss": 0.7834, + "step": 4656 + }, + { + "epoch": 2.528230184581976, + "grad_norm": 11.163753697057714, + "learning_rate": 1.030415969233046e-05, + "loss": 0.6613, + "step": 4657 + }, + { + "epoch": 2.528773072747014, + "grad_norm": 8.747908439433298, + "learning_rate": 1.03006444613191e-05, + "loss": 0.6974, + "step": 4658 + }, + { + "epoch": 2.529315960912052, + "grad_norm": 8.370456762868022, + "learning_rate": 1.0297129193123555e-05, + "loss": 0.5343, + "step": 4659 + }, + { + "epoch": 2.5298588490770904, + "grad_norm": 13.46061984772866, + "learning_rate": 1.0293613888178597e-05, + "loss": 1.3054, + "step": 4660 + }, + { + "epoch": 2.530401737242128, + "grad_norm": 10.948007938187232, + "learning_rate": 1.0290098546919007e-05, + "loss": 0.7028, + "step": 4661 + }, + { + "epoch": 2.530944625407166, + "grad_norm": 8.833095994203507, + "learning_rate": 1.0286583169779568e-05, + "loss": 0.7663, + "step": 4662 + }, + { + "epoch": 2.531487513572204, + "grad_norm": 10.283493007620729, + "learning_rate": 1.0283067757195062e-05, + "loss": 0.7804, + "step": 4663 + }, + { + "epoch": 2.5320304017372424, + "grad_norm": 13.55508996688301, + "learning_rate": 1.0279552309600288e-05, + "loss": 0.656, + "step": 4664 + }, + { + "epoch": 2.53257328990228, + "grad_norm": 8.738397704085456, + "learning_rate": 1.0276036827430036e-05, + "loss": 0.4902, + "step": 4665 + }, + { + "epoch": 2.533116178067318, + "grad_norm": 10.260011622402263, + "learning_rate": 1.0272521311119113e-05, + "loss": 0.4798, + "step": 4666 + }, + { + "epoch": 2.533659066232356, + "grad_norm": 10.743076748133591, + "learning_rate": 1.0269005761102315e-05, + "loss": 0.9671, + "step": 4667 + }, + { + "epoch": 2.5342019543973944, + "grad_norm": 11.846945314251228, + "learning_rate": 1.0265490177814458e-05, + "loss": 0.7168, + "step": 4668 + }, + { + "epoch": 2.534744842562432, + "grad_norm": 9.129703791341635, + "learning_rate": 1.0261974561690356e-05, + "loss": 0.6008, + "step": 4669 + }, + { + "epoch": 2.53528773072747, + "grad_norm": 8.464074733631888, + "learning_rate": 1.025845891316482e-05, + "loss": 0.4327, + "step": 4670 + }, + { + "epoch": 2.535830618892508, + "grad_norm": 10.894552922917367, + "learning_rate": 1.0254943232672671e-05, + "loss": 0.5627, + "step": 4671 + }, + { + "epoch": 2.5363735070575464, + "grad_norm": 12.365454725760225, + "learning_rate": 1.025142752064874e-05, + "loss": 0.606, + "step": 4672 + }, + { + "epoch": 2.536916395222584, + "grad_norm": 10.066952200044973, + "learning_rate": 1.0247911777527852e-05, + "loss": 0.7503, + "step": 4673 + }, + { + "epoch": 2.537459283387622, + "grad_norm": 8.386860314534095, + "learning_rate": 1.0244396003744836e-05, + "loss": 0.7663, + "step": 4674 + }, + { + "epoch": 2.53800217155266, + "grad_norm": 9.309667750090998, + "learning_rate": 1.024088019973454e-05, + "loss": 0.5658, + "step": 4675 + }, + { + "epoch": 2.5385450597176984, + "grad_norm": 13.530278392237364, + "learning_rate": 1.0237364365931794e-05, + "loss": 0.799, + "step": 4676 + }, + { + "epoch": 2.539087947882736, + "grad_norm": 11.344034257453236, + "learning_rate": 1.0233848502771444e-05, + "loss": 0.5931, + "step": 4677 + }, + { + "epoch": 2.539630836047774, + "grad_norm": 9.26197757297852, + "learning_rate": 1.0230332610688338e-05, + "loss": 0.5419, + "step": 4678 + }, + { + "epoch": 2.540173724212812, + "grad_norm": 12.620486559012823, + "learning_rate": 1.022681669011733e-05, + "loss": 0.6746, + "step": 4679 + }, + { + "epoch": 2.5407166123778504, + "grad_norm": 13.640092752737832, + "learning_rate": 1.0223300741493268e-05, + "loss": 0.8806, + "step": 4680 + }, + { + "epoch": 2.541259500542888, + "grad_norm": 10.435820181571753, + "learning_rate": 1.0219784765251018e-05, + "loss": 0.6876, + "step": 4681 + }, + { + "epoch": 2.541802388707926, + "grad_norm": 11.624678093656978, + "learning_rate": 1.0216268761825436e-05, + "loss": 0.7377, + "step": 4682 + }, + { + "epoch": 2.542345276872964, + "grad_norm": 10.361224010713974, + "learning_rate": 1.021275273165139e-05, + "loss": 0.5911, + "step": 4683 + }, + { + "epoch": 2.5428881650380024, + "grad_norm": 6.869744257762391, + "learning_rate": 1.0209236675163747e-05, + "loss": 0.4204, + "step": 4684 + }, + { + "epoch": 2.54343105320304, + "grad_norm": 16.790885412103638, + "learning_rate": 1.020572059279738e-05, + "loss": 0.8837, + "step": 4685 + }, + { + "epoch": 2.543973941368078, + "grad_norm": 9.226978433564653, + "learning_rate": 1.0202204484987159e-05, + "loss": 0.4686, + "step": 4686 + }, + { + "epoch": 2.544516829533116, + "grad_norm": 7.391909229375343, + "learning_rate": 1.019868835216797e-05, + "loss": 0.6996, + "step": 4687 + }, + { + "epoch": 2.5450597176981544, + "grad_norm": 9.103089555285077, + "learning_rate": 1.0195172194774684e-05, + "loss": 0.7243, + "step": 4688 + }, + { + "epoch": 2.545602605863192, + "grad_norm": 7.27509790187114, + "learning_rate": 1.0191656013242192e-05, + "loss": 0.4612, + "step": 4689 + }, + { + "epoch": 2.54614549402823, + "grad_norm": 6.939999239823949, + "learning_rate": 1.0188139808005379e-05, + "loss": 0.3483, + "step": 4690 + }, + { + "epoch": 2.546688382193268, + "grad_norm": 9.036119044447904, + "learning_rate": 1.0184623579499133e-05, + "loss": 0.5822, + "step": 4691 + }, + { + "epoch": 2.5472312703583064, + "grad_norm": 15.325233133671158, + "learning_rate": 1.018110732815835e-05, + "loss": 0.9375, + "step": 4692 + }, + { + "epoch": 2.547774158523344, + "grad_norm": 13.335480222530155, + "learning_rate": 1.017759105441792e-05, + "loss": 0.8165, + "step": 4693 + }, + { + "epoch": 2.548317046688382, + "grad_norm": 10.097959894991517, + "learning_rate": 1.0174074758712751e-05, + "loss": 0.7773, + "step": 4694 + }, + { + "epoch": 2.54885993485342, + "grad_norm": 9.7484700487623, + "learning_rate": 1.0170558441477738e-05, + "loss": 0.7805, + "step": 4695 + }, + { + "epoch": 2.5494028230184584, + "grad_norm": 9.687849482731272, + "learning_rate": 1.0167042103147784e-05, + "loss": 0.6675, + "step": 4696 + }, + { + "epoch": 2.549945711183496, + "grad_norm": 12.578144763595915, + "learning_rate": 1.0163525744157796e-05, + "loss": 0.7757, + "step": 4697 + }, + { + "epoch": 2.550488599348534, + "grad_norm": 9.03070780679602, + "learning_rate": 1.0160009364942684e-05, + "loss": 0.5321, + "step": 4698 + }, + { + "epoch": 2.5510314875135722, + "grad_norm": 10.632448946778698, + "learning_rate": 1.0156492965937357e-05, + "loss": 0.5589, + "step": 4699 + }, + { + "epoch": 2.5515743756786105, + "grad_norm": 10.053170332804227, + "learning_rate": 1.015297654757673e-05, + "loss": 0.4974, + "step": 4700 + }, + { + "epoch": 2.5521172638436482, + "grad_norm": 9.833516973903698, + "learning_rate": 1.0149460110295722e-05, + "loss": 0.6022, + "step": 4701 + }, + { + "epoch": 2.552660152008686, + "grad_norm": 16.704821665027783, + "learning_rate": 1.0145943654529248e-05, + "loss": 0.776, + "step": 4702 + }, + { + "epoch": 2.5532030401737242, + "grad_norm": 12.589948540180313, + "learning_rate": 1.0142427180712231e-05, + "loss": 0.8549, + "step": 4703 + }, + { + "epoch": 2.5537459283387625, + "grad_norm": 10.414828597120934, + "learning_rate": 1.0138910689279596e-05, + "loss": 0.7209, + "step": 4704 + }, + { + "epoch": 2.5542888165038002, + "grad_norm": 10.158714848928406, + "learning_rate": 1.0135394180666261e-05, + "loss": 0.6835, + "step": 4705 + }, + { + "epoch": 2.554831704668838, + "grad_norm": 8.255885711809917, + "learning_rate": 1.013187765530716e-05, + "loss": 0.4542, + "step": 4706 + }, + { + "epoch": 2.5553745928338762, + "grad_norm": 10.386471672795576, + "learning_rate": 1.012836111363722e-05, + "loss": 0.9478, + "step": 4707 + }, + { + "epoch": 2.5559174809989145, + "grad_norm": 11.828466617801858, + "learning_rate": 1.0124844556091374e-05, + "loss": 0.7975, + "step": 4708 + }, + { + "epoch": 2.5564603691639523, + "grad_norm": 10.742130491540307, + "learning_rate": 1.0121327983104555e-05, + "loss": 0.9712, + "step": 4709 + }, + { + "epoch": 2.55700325732899, + "grad_norm": 13.314461670532042, + "learning_rate": 1.0117811395111695e-05, + "loss": 0.9351, + "step": 4710 + }, + { + "epoch": 2.5575461454940283, + "grad_norm": 10.15371491897671, + "learning_rate": 1.0114294792547733e-05, + "loss": 0.6056, + "step": 4711 + }, + { + "epoch": 2.5580890336590665, + "grad_norm": 8.317356068473407, + "learning_rate": 1.011077817584761e-05, + "loss": 0.6411, + "step": 4712 + }, + { + "epoch": 2.5586319218241043, + "grad_norm": 11.215419343504372, + "learning_rate": 1.0107261545446267e-05, + "loss": 0.5411, + "step": 4713 + }, + { + "epoch": 2.559174809989142, + "grad_norm": 10.343171151673559, + "learning_rate": 1.0103744901778645e-05, + "loss": 0.5395, + "step": 4714 + }, + { + "epoch": 2.5597176981541803, + "grad_norm": 14.089833480392985, + "learning_rate": 1.0100228245279688e-05, + "loss": 0.9855, + "step": 4715 + }, + { + "epoch": 2.5602605863192185, + "grad_norm": 14.836069456421422, + "learning_rate": 1.0096711576384344e-05, + "loss": 0.8219, + "step": 4716 + }, + { + "epoch": 2.5608034744842563, + "grad_norm": 16.933747543735457, + "learning_rate": 1.0093194895527556e-05, + "loss": 0.9341, + "step": 4717 + }, + { + "epoch": 2.561346362649294, + "grad_norm": 13.574017237717102, + "learning_rate": 1.008967820314427e-05, + "loss": 1.1551, + "step": 4718 + }, + { + "epoch": 2.5618892508143323, + "grad_norm": 8.877246747020687, + "learning_rate": 1.0086161499669447e-05, + "loss": 0.6341, + "step": 4719 + }, + { + "epoch": 2.5624321389793705, + "grad_norm": 9.391996392121968, + "learning_rate": 1.0082644785538031e-05, + "loss": 0.5285, + "step": 4720 + }, + { + "epoch": 2.5629750271444083, + "grad_norm": 10.876272607307573, + "learning_rate": 1.0079128061184977e-05, + "loss": 0.5678, + "step": 4721 + }, + { + "epoch": 2.563517915309446, + "grad_norm": 12.516236640792583, + "learning_rate": 1.0075611327045236e-05, + "loss": 0.8382, + "step": 4722 + }, + { + "epoch": 2.5640608034744843, + "grad_norm": 6.395711230996481, + "learning_rate": 1.0072094583553769e-05, + "loss": 0.3785, + "step": 4723 + }, + { + "epoch": 2.5646036916395225, + "grad_norm": 8.671727273867468, + "learning_rate": 1.0068577831145526e-05, + "loss": 0.5039, + "step": 4724 + }, + { + "epoch": 2.5651465798045603, + "grad_norm": 12.77911107984442, + "learning_rate": 1.0065061070255469e-05, + "loss": 0.7647, + "step": 4725 + }, + { + "epoch": 2.565689467969598, + "grad_norm": 9.108473588917152, + "learning_rate": 1.0061544301318556e-05, + "loss": 0.798, + "step": 4726 + }, + { + "epoch": 2.5662323561346363, + "grad_norm": 8.776410877174934, + "learning_rate": 1.0058027524769742e-05, + "loss": 0.6388, + "step": 4727 + }, + { + "epoch": 2.5667752442996745, + "grad_norm": 10.392346204038525, + "learning_rate": 1.0054510741043995e-05, + "loss": 0.8178, + "step": 4728 + }, + { + "epoch": 2.5673181324647123, + "grad_norm": 8.611264791773579, + "learning_rate": 1.005099395057627e-05, + "loss": 0.5022, + "step": 4729 + }, + { + "epoch": 2.56786102062975, + "grad_norm": 8.228466341722912, + "learning_rate": 1.0047477153801534e-05, + "loss": 0.6572, + "step": 4730 + }, + { + "epoch": 2.5684039087947883, + "grad_norm": 8.410912812963819, + "learning_rate": 1.0043960351154745e-05, + "loss": 0.4526, + "step": 4731 + }, + { + "epoch": 2.5689467969598265, + "grad_norm": 8.24917306612946, + "learning_rate": 1.0040443543070872e-05, + "loss": 0.5492, + "step": 4732 + }, + { + "epoch": 2.5694896851248643, + "grad_norm": 7.578877723490418, + "learning_rate": 1.0036926729984878e-05, + "loss": 0.5482, + "step": 4733 + }, + { + "epoch": 2.570032573289902, + "grad_norm": 8.464719447395414, + "learning_rate": 1.003340991233173e-05, + "loss": 0.5236, + "step": 4734 + }, + { + "epoch": 2.5705754614549403, + "grad_norm": 10.297009632330814, + "learning_rate": 1.0029893090546385e-05, + "loss": 0.6747, + "step": 4735 + }, + { + "epoch": 2.5711183496199785, + "grad_norm": 9.895183624345862, + "learning_rate": 1.0026376265063818e-05, + "loss": 0.7036, + "step": 4736 + }, + { + "epoch": 2.5716612377850163, + "grad_norm": 16.034084078355107, + "learning_rate": 1.0022859436318988e-05, + "loss": 0.8583, + "step": 4737 + }, + { + "epoch": 2.572204125950054, + "grad_norm": 10.881395091863821, + "learning_rate": 1.0019342604746871e-05, + "loss": 0.7214, + "step": 4738 + }, + { + "epoch": 2.5727470141150923, + "grad_norm": 9.051906164481618, + "learning_rate": 1.0015825770782428e-05, + "loss": 0.36, + "step": 4739 + }, + { + "epoch": 2.5732899022801305, + "grad_norm": 11.71081741433651, + "learning_rate": 1.001230893486063e-05, + "loss": 0.8467, + "step": 4740 + }, + { + "epoch": 2.5738327904451683, + "grad_norm": 8.4727788666991, + "learning_rate": 1.0008792097416442e-05, + "loss": 0.5312, + "step": 4741 + }, + { + "epoch": 2.574375678610206, + "grad_norm": 10.308791039753308, + "learning_rate": 1.0005275258884835e-05, + "loss": 0.7132, + "step": 4742 + }, + { + "epoch": 2.5749185667752443, + "grad_norm": 13.591326014197278, + "learning_rate": 1.0001758419700773e-05, + "loss": 0.7087, + "step": 4743 + }, + { + "epoch": 2.5754614549402826, + "grad_norm": 9.610601406060551, + "learning_rate": 9.998241580299229e-06, + "loss": 0.6708, + "step": 4744 + }, + { + "epoch": 2.5760043431053203, + "grad_norm": 10.04740706892737, + "learning_rate": 9.994724741115169e-06, + "loss": 0.6644, + "step": 4745 + }, + { + "epoch": 2.576547231270358, + "grad_norm": 7.760395046863199, + "learning_rate": 9.991207902583558e-06, + "loss": 0.4361, + "step": 4746 + }, + { + "epoch": 2.5770901194353963, + "grad_norm": 9.097495186854097, + "learning_rate": 9.987691065139373e-06, + "loss": 0.5613, + "step": 4747 + }, + { + "epoch": 2.5776330076004346, + "grad_norm": 9.584167359092984, + "learning_rate": 9.984174229217572e-06, + "loss": 0.6408, + "step": 4748 + }, + { + "epoch": 2.5781758957654723, + "grad_norm": 10.329246788617224, + "learning_rate": 9.980657395253132e-06, + "loss": 0.6776, + "step": 4749 + }, + { + "epoch": 2.57871878393051, + "grad_norm": 10.000264678013584, + "learning_rate": 9.977140563681015e-06, + "loss": 0.7489, + "step": 4750 + }, + { + "epoch": 2.5792616720955484, + "grad_norm": 6.80240569067885, + "learning_rate": 9.973623734936185e-06, + "loss": 0.4639, + "step": 4751 + }, + { + "epoch": 2.5798045602605866, + "grad_norm": 7.578175941165148, + "learning_rate": 9.97010690945362e-06, + "loss": 0.38, + "step": 4752 + }, + { + "epoch": 2.5803474484256244, + "grad_norm": 12.48681779678407, + "learning_rate": 9.966590087668274e-06, + "loss": 0.8046, + "step": 4753 + }, + { + "epoch": 2.580890336590662, + "grad_norm": 13.31142746407252, + "learning_rate": 9.963073270015126e-06, + "loss": 0.7671, + "step": 4754 + }, + { + "epoch": 2.5814332247557004, + "grad_norm": 8.987640426318515, + "learning_rate": 9.959556456929128e-06, + "loss": 0.4691, + "step": 4755 + }, + { + "epoch": 2.5819761129207386, + "grad_norm": 9.551368066054666, + "learning_rate": 9.956039648845257e-06, + "loss": 0.6207, + "step": 4756 + }, + { + "epoch": 2.5825190010857764, + "grad_norm": 17.060876787385865, + "learning_rate": 9.952522846198468e-06, + "loss": 0.9276, + "step": 4757 + }, + { + "epoch": 2.583061889250814, + "grad_norm": 9.760247190255836, + "learning_rate": 9.949006049423731e-06, + "loss": 0.539, + "step": 4758 + }, + { + "epoch": 2.5836047774158524, + "grad_norm": 14.929770896239804, + "learning_rate": 9.94548925895601e-06, + "loss": 1.0896, + "step": 4759 + }, + { + "epoch": 2.5841476655808906, + "grad_norm": 14.902014520576888, + "learning_rate": 9.94197247523026e-06, + "loss": 1.0207, + "step": 4760 + }, + { + "epoch": 2.5846905537459284, + "grad_norm": 9.630889650448959, + "learning_rate": 9.93845569868145e-06, + "loss": 0.6322, + "step": 4761 + }, + { + "epoch": 2.585233441910966, + "grad_norm": 6.534627620002312, + "learning_rate": 9.934938929744535e-06, + "loss": 0.2627, + "step": 4762 + }, + { + "epoch": 2.5857763300760044, + "grad_norm": 10.701147725996508, + "learning_rate": 9.931422168854476e-06, + "loss": 0.6789, + "step": 4763 + }, + { + "epoch": 2.5863192182410426, + "grad_norm": 9.539331065006966, + "learning_rate": 9.927905416446233e-06, + "loss": 0.7109, + "step": 4764 + }, + { + "epoch": 2.5868621064060804, + "grad_norm": 9.979058002114003, + "learning_rate": 9.924388672954766e-06, + "loss": 0.6497, + "step": 4765 + }, + { + "epoch": 2.587404994571118, + "grad_norm": 9.638452627150874, + "learning_rate": 9.920871938815024e-06, + "loss": 0.6695, + "step": 4766 + }, + { + "epoch": 2.5879478827361564, + "grad_norm": 12.326422327522844, + "learning_rate": 9.91735521446197e-06, + "loss": 0.6354, + "step": 4767 + }, + { + "epoch": 2.5884907709011946, + "grad_norm": 8.571541287165285, + "learning_rate": 9.913838500330553e-06, + "loss": 0.3939, + "step": 4768 + }, + { + "epoch": 2.5890336590662324, + "grad_norm": 9.709310867338898, + "learning_rate": 9.910321796855732e-06, + "loss": 0.5866, + "step": 4769 + }, + { + "epoch": 2.58957654723127, + "grad_norm": 8.48642736575235, + "learning_rate": 9.90680510447245e-06, + "loss": 0.8167, + "step": 4770 + }, + { + "epoch": 2.5901194353963084, + "grad_norm": 9.153780615106163, + "learning_rate": 9.90328842361566e-06, + "loss": 0.6551, + "step": 4771 + }, + { + "epoch": 2.5906623235613466, + "grad_norm": 9.331431330636146, + "learning_rate": 9.899771754720315e-06, + "loss": 0.598, + "step": 4772 + }, + { + "epoch": 2.5912052117263844, + "grad_norm": 10.71941498109364, + "learning_rate": 9.896255098221357e-06, + "loss": 0.6996, + "step": 4773 + }, + { + "epoch": 2.591748099891422, + "grad_norm": 11.415708336876241, + "learning_rate": 9.892738454553736e-06, + "loss": 0.8247, + "step": 4774 + }, + { + "epoch": 2.5922909880564604, + "grad_norm": 9.526809613609492, + "learning_rate": 9.889221824152391e-06, + "loss": 0.5088, + "step": 4775 + }, + { + "epoch": 2.5928338762214986, + "grad_norm": 8.911159845652058, + "learning_rate": 9.885705207452268e-06, + "loss": 0.3377, + "step": 4776 + }, + { + "epoch": 2.5933767643865364, + "grad_norm": 13.451504414556261, + "learning_rate": 9.882188604888307e-06, + "loss": 0.9247, + "step": 4777 + }, + { + "epoch": 2.593919652551574, + "grad_norm": 10.251509263282227, + "learning_rate": 9.87867201689545e-06, + "loss": 0.4384, + "step": 4778 + }, + { + "epoch": 2.5944625407166124, + "grad_norm": 9.955424126921221, + "learning_rate": 9.875155443908631e-06, + "loss": 0.6318, + "step": 4779 + }, + { + "epoch": 2.5950054288816506, + "grad_norm": 8.584868222198555, + "learning_rate": 9.871638886362782e-06, + "loss": 0.5148, + "step": 4780 + }, + { + "epoch": 2.5955483170466884, + "grad_norm": 16.858447088677025, + "learning_rate": 9.868122344692846e-06, + "loss": 1.2564, + "step": 4781 + }, + { + "epoch": 2.596091205211726, + "grad_norm": 9.7339544501354, + "learning_rate": 9.86460581933374e-06, + "loss": 0.7158, + "step": 4782 + }, + { + "epoch": 2.5966340933767644, + "grad_norm": 11.03551156903236, + "learning_rate": 9.861089310720409e-06, + "loss": 0.5136, + "step": 4783 + }, + { + "epoch": 2.5971769815418027, + "grad_norm": 11.173351784755749, + "learning_rate": 9.857572819287768e-06, + "loss": 0.6439, + "step": 4784 + }, + { + "epoch": 2.5977198697068404, + "grad_norm": 14.318584084756125, + "learning_rate": 9.854056345470754e-06, + "loss": 1.2827, + "step": 4785 + }, + { + "epoch": 2.598262757871878, + "grad_norm": 11.132635136137834, + "learning_rate": 9.850539889704278e-06, + "loss": 0.499, + "step": 4786 + }, + { + "epoch": 2.5988056460369164, + "grad_norm": 11.710993594625656, + "learning_rate": 9.847023452423272e-06, + "loss": 0.466, + "step": 4787 + }, + { + "epoch": 2.5993485342019547, + "grad_norm": 12.05431003500401, + "learning_rate": 9.843507034062646e-06, + "loss": 1.002, + "step": 4788 + }, + { + "epoch": 2.5998914223669924, + "grad_norm": 12.651344078635795, + "learning_rate": 9.83999063505732e-06, + "loss": 0.6906, + "step": 4789 + }, + { + "epoch": 2.6004343105320302, + "grad_norm": 11.592009871908612, + "learning_rate": 9.83647425584221e-06, + "loss": 0.908, + "step": 4790 + }, + { + "epoch": 2.6009771986970684, + "grad_norm": 11.164823794029719, + "learning_rate": 9.83295789685222e-06, + "loss": 1.1006, + "step": 4791 + }, + { + "epoch": 2.6015200868621067, + "grad_norm": 18.55453736068596, + "learning_rate": 9.829441558522267e-06, + "loss": 0.6774, + "step": 4792 + }, + { + "epoch": 2.6020629750271445, + "grad_norm": 9.004957503789628, + "learning_rate": 9.825925241287249e-06, + "loss": 0.612, + "step": 4793 + }, + { + "epoch": 2.6026058631921822, + "grad_norm": 15.331324393349687, + "learning_rate": 9.822408945582081e-06, + "loss": 0.588, + "step": 4794 + }, + { + "epoch": 2.6031487513572205, + "grad_norm": 11.30698173094918, + "learning_rate": 9.818892671841653e-06, + "loss": 0.8413, + "step": 4795 + }, + { + "epoch": 2.6036916395222587, + "grad_norm": 10.970237049386885, + "learning_rate": 9.815376420500869e-06, + "loss": 0.5687, + "step": 4796 + }, + { + "epoch": 2.6042345276872965, + "grad_norm": 12.062898313733237, + "learning_rate": 9.811860191994623e-06, + "loss": 0.88, + "step": 4797 + }, + { + "epoch": 2.6047774158523342, + "grad_norm": 13.760783812579094, + "learning_rate": 9.808343986757811e-06, + "loss": 0.7319, + "step": 4798 + }, + { + "epoch": 2.6053203040173725, + "grad_norm": 10.142158775926303, + "learning_rate": 9.80482780522532e-06, + "loss": 0.5404, + "step": 4799 + }, + { + "epoch": 2.6058631921824107, + "grad_norm": 14.235611160734313, + "learning_rate": 9.801311647832035e-06, + "loss": 0.6849, + "step": 4800 + }, + { + "epoch": 2.6064060803474485, + "grad_norm": 14.33484497964673, + "learning_rate": 9.797795515012843e-06, + "loss": 1.0677, + "step": 4801 + }, + { + "epoch": 2.6069489685124863, + "grad_norm": 12.238534475065178, + "learning_rate": 9.794279407202624e-06, + "loss": 0.8279, + "step": 4802 + }, + { + "epoch": 2.6074918566775245, + "grad_norm": 13.51950414574984, + "learning_rate": 9.790763324836256e-06, + "loss": 0.8372, + "step": 4803 + }, + { + "epoch": 2.6080347448425627, + "grad_norm": 9.360760584245645, + "learning_rate": 9.78724726834861e-06, + "loss": 0.666, + "step": 4804 + }, + { + "epoch": 2.6085776330076005, + "grad_norm": 10.793352407977682, + "learning_rate": 9.783731238174566e-06, + "loss": 0.8308, + "step": 4805 + }, + { + "epoch": 2.6091205211726383, + "grad_norm": 13.862491939748782, + "learning_rate": 9.780215234748984e-06, + "loss": 0.8018, + "step": 4806 + }, + { + "epoch": 2.6096634093376765, + "grad_norm": 8.019652931555422, + "learning_rate": 9.776699258506734e-06, + "loss": 0.3801, + "step": 4807 + }, + { + "epoch": 2.6102062975027147, + "grad_norm": 10.161487802712038, + "learning_rate": 9.773183309882675e-06, + "loss": 0.6723, + "step": 4808 + }, + { + "epoch": 2.6107491856677525, + "grad_norm": 9.64703000959591, + "learning_rate": 9.769667389311664e-06, + "loss": 0.6294, + "step": 4809 + }, + { + "epoch": 2.6112920738327903, + "grad_norm": 9.792927662736604, + "learning_rate": 9.766151497228561e-06, + "loss": 0.4596, + "step": 4810 + }, + { + "epoch": 2.6118349619978285, + "grad_norm": 10.336511067045754, + "learning_rate": 9.76263563406821e-06, + "loss": 0.6494, + "step": 4811 + }, + { + "epoch": 2.6123778501628667, + "grad_norm": 12.121823172890695, + "learning_rate": 9.759119800265464e-06, + "loss": 1.0394, + "step": 4812 + }, + { + "epoch": 2.6129207383279045, + "grad_norm": 10.891668605270631, + "learning_rate": 9.755603996255164e-06, + "loss": 0.6212, + "step": 4813 + }, + { + "epoch": 2.6134636264929423, + "grad_norm": 9.599793914114507, + "learning_rate": 9.752088222472151e-06, + "loss": 0.6182, + "step": 4814 + }, + { + "epoch": 2.6140065146579805, + "grad_norm": 10.448317247236231, + "learning_rate": 9.748572479351263e-06, + "loss": 0.6779, + "step": 4815 + }, + { + "epoch": 2.6145494028230187, + "grad_norm": 12.703205416876392, + "learning_rate": 9.74505676732733e-06, + "loss": 0.8093, + "step": 4816 + }, + { + "epoch": 2.6150922909880565, + "grad_norm": 10.14294888595903, + "learning_rate": 9.741541086835182e-06, + "loss": 0.6493, + "step": 4817 + }, + { + "epoch": 2.6156351791530943, + "grad_norm": 11.455318908063175, + "learning_rate": 9.738025438309649e-06, + "loss": 0.9375, + "step": 4818 + }, + { + "epoch": 2.6161780673181325, + "grad_norm": 8.90202557317791, + "learning_rate": 9.734509822185545e-06, + "loss": 0.5978, + "step": 4819 + }, + { + "epoch": 2.6167209554831707, + "grad_norm": 10.38484629007874, + "learning_rate": 9.730994238897687e-06, + "loss": 0.9772, + "step": 4820 + }, + { + "epoch": 2.6172638436482085, + "grad_norm": 10.650781833239714, + "learning_rate": 9.72747868888089e-06, + "loss": 0.7354, + "step": 4821 + }, + { + "epoch": 2.6178067318132463, + "grad_norm": 11.462972783418381, + "learning_rate": 9.723963172569964e-06, + "loss": 0.8748, + "step": 4822 + }, + { + "epoch": 2.6183496199782845, + "grad_norm": 7.4804704893307505, + "learning_rate": 9.720447690399716e-06, + "loss": 0.4853, + "step": 4823 + }, + { + "epoch": 2.6188925081433228, + "grad_norm": 8.032622743372611, + "learning_rate": 9.716932242804938e-06, + "loss": 0.6484, + "step": 4824 + }, + { + "epoch": 2.6194353963083605, + "grad_norm": 10.141962204854094, + "learning_rate": 9.713416830220436e-06, + "loss": 0.669, + "step": 4825 + }, + { + "epoch": 2.6199782844733983, + "grad_norm": 11.48618310533414, + "learning_rate": 9.709901453080993e-06, + "loss": 0.6279, + "step": 4826 + }, + { + "epoch": 2.6205211726384365, + "grad_norm": 9.789449860925128, + "learning_rate": 9.706386111821406e-06, + "loss": 0.6072, + "step": 4827 + }, + { + "epoch": 2.6210640608034748, + "grad_norm": 8.10222084311329, + "learning_rate": 9.702870806876448e-06, + "loss": 0.6476, + "step": 4828 + }, + { + "epoch": 2.6216069489685125, + "grad_norm": 10.30951278757575, + "learning_rate": 9.699355538680902e-06, + "loss": 0.4691, + "step": 4829 + }, + { + "epoch": 2.6221498371335503, + "grad_norm": 10.960835467770405, + "learning_rate": 9.695840307669547e-06, + "loss": 0.7962, + "step": 4830 + }, + { + "epoch": 2.6226927252985885, + "grad_norm": 6.691408537531251, + "learning_rate": 9.69232511427714e-06, + "loss": 0.3644, + "step": 4831 + }, + { + "epoch": 2.6232356134636268, + "grad_norm": 7.019904615334962, + "learning_rate": 9.68880995893846e-06, + "loss": 0.4594, + "step": 4832 + }, + { + "epoch": 2.6237785016286646, + "grad_norm": 11.780644240657415, + "learning_rate": 9.685294842088256e-06, + "loss": 1.0625, + "step": 4833 + }, + { + "epoch": 2.6243213897937023, + "grad_norm": 7.34603375318287, + "learning_rate": 9.681779764161289e-06, + "loss": 0.3502, + "step": 4834 + }, + { + "epoch": 2.6248642779587406, + "grad_norm": 9.787656886442962, + "learning_rate": 9.678264725592306e-06, + "loss": 0.5054, + "step": 4835 + }, + { + "epoch": 2.6254071661237783, + "grad_norm": 9.687484513116981, + "learning_rate": 9.674749726816058e-06, + "loss": 0.8678, + "step": 4836 + }, + { + "epoch": 2.6259500542888166, + "grad_norm": 7.620468702427845, + "learning_rate": 9.671234768267284e-06, + "loss": 0.4672, + "step": 4837 + }, + { + "epoch": 2.6264929424538543, + "grad_norm": 14.051025442281276, + "learning_rate": 9.667719850380712e-06, + "loss": 0.6462, + "step": 4838 + }, + { + "epoch": 2.6270358306188926, + "grad_norm": 11.579748119704858, + "learning_rate": 9.664204973591081e-06, + "loss": 0.8856, + "step": 4839 + }, + { + "epoch": 2.6275787187839303, + "grad_norm": 8.348102354642592, + "learning_rate": 9.660690138333114e-06, + "loss": 0.5623, + "step": 4840 + }, + { + "epoch": 2.6281216069489686, + "grad_norm": 9.351950984929493, + "learning_rate": 9.657175345041532e-06, + "loss": 0.5809, + "step": 4841 + }, + { + "epoch": 2.6286644951140063, + "grad_norm": 12.421448876984805, + "learning_rate": 9.653660594151047e-06, + "loss": 0.7831, + "step": 4842 + }, + { + "epoch": 2.6292073832790446, + "grad_norm": 11.981212578199576, + "learning_rate": 9.650145886096376e-06, + "loss": 0.8767, + "step": 4843 + }, + { + "epoch": 2.6297502714440824, + "grad_norm": 10.659578351567276, + "learning_rate": 9.646631221312216e-06, + "loss": 0.7313, + "step": 4844 + }, + { + "epoch": 2.6302931596091206, + "grad_norm": 12.480207661204222, + "learning_rate": 9.643116600233274e-06, + "loss": 0.7412, + "step": 4845 + }, + { + "epoch": 2.6308360477741584, + "grad_norm": 9.131072461148838, + "learning_rate": 9.639602023294233e-06, + "loss": 0.5518, + "step": 4846 + }, + { + "epoch": 2.6313789359391966, + "grad_norm": 7.489089256780018, + "learning_rate": 9.636087490929793e-06, + "loss": 0.4321, + "step": 4847 + }, + { + "epoch": 2.6319218241042344, + "grad_norm": 12.12610683915495, + "learning_rate": 9.632573003574634e-06, + "loss": 0.8265, + "step": 4848 + }, + { + "epoch": 2.6324647122692726, + "grad_norm": 7.965586818160769, + "learning_rate": 9.629058561663426e-06, + "loss": 0.373, + "step": 4849 + }, + { + "epoch": 2.6330076004343104, + "grad_norm": 9.348866579305472, + "learning_rate": 9.62554416563085e-06, + "loss": 0.4274, + "step": 4850 + }, + { + "epoch": 2.6335504885993486, + "grad_norm": 10.13050086693858, + "learning_rate": 9.622029815911566e-06, + "loss": 0.5276, + "step": 4851 + }, + { + "epoch": 2.6340933767643864, + "grad_norm": 10.19180911105006, + "learning_rate": 9.61851551294024e-06, + "loss": 0.5871, + "step": 4852 + }, + { + "epoch": 2.6346362649294246, + "grad_norm": 11.184621464328854, + "learning_rate": 9.61500125715152e-06, + "loss": 0.6047, + "step": 4853 + }, + { + "epoch": 2.6351791530944624, + "grad_norm": 14.484355041392742, + "learning_rate": 9.611487048980058e-06, + "loss": 0.8628, + "step": 4854 + }, + { + "epoch": 2.6357220412595006, + "grad_norm": 13.070986825901516, + "learning_rate": 9.607972888860497e-06, + "loss": 0.7793, + "step": 4855 + }, + { + "epoch": 2.6362649294245384, + "grad_norm": 9.640708751060075, + "learning_rate": 9.604458777227477e-06, + "loss": 0.6016, + "step": 4856 + }, + { + "epoch": 2.6368078175895766, + "grad_norm": 11.05523237140924, + "learning_rate": 9.600944714515627e-06, + "loss": 0.5999, + "step": 4857 + }, + { + "epoch": 2.6373507057546144, + "grad_norm": 13.240640169530616, + "learning_rate": 9.597430701159565e-06, + "loss": 0.6779, + "step": 4858 + }, + { + "epoch": 2.6378935939196526, + "grad_norm": 10.878893821343203, + "learning_rate": 9.593916737593919e-06, + "loss": 0.6119, + "step": 4859 + }, + { + "epoch": 2.6384364820846904, + "grad_norm": 10.820526634571284, + "learning_rate": 9.590402824253295e-06, + "loss": 0.839, + "step": 4860 + }, + { + "epoch": 2.6389793702497286, + "grad_norm": 12.938750622065529, + "learning_rate": 9.586888961572307e-06, + "loss": 0.517, + "step": 4861 + }, + { + "epoch": 2.6395222584147664, + "grad_norm": 9.725430374953143, + "learning_rate": 9.583375149985547e-06, + "loss": 0.7604, + "step": 4862 + }, + { + "epoch": 2.6400651465798046, + "grad_norm": 11.840312760690757, + "learning_rate": 9.579861389927615e-06, + "loss": 0.7614, + "step": 4863 + }, + { + "epoch": 2.6406080347448424, + "grad_norm": 11.047185391206618, + "learning_rate": 9.576347681833093e-06, + "loss": 0.6351, + "step": 4864 + }, + { + "epoch": 2.6411509229098806, + "grad_norm": 9.986349266550256, + "learning_rate": 9.572834026136571e-06, + "loss": 0.5549, + "step": 4865 + }, + { + "epoch": 2.6416938110749184, + "grad_norm": 14.455847384105025, + "learning_rate": 9.56932042327261e-06, + "loss": 0.8206, + "step": 4866 + }, + { + "epoch": 2.6422366992399566, + "grad_norm": 10.377988559117977, + "learning_rate": 9.56580687367579e-06, + "loss": 0.5275, + "step": 4867 + }, + { + "epoch": 2.6427795874049944, + "grad_norm": 10.71910613137521, + "learning_rate": 9.562293377780668e-06, + "loss": 0.6958, + "step": 4868 + }, + { + "epoch": 2.6433224755700326, + "grad_norm": 8.635819291033194, + "learning_rate": 9.558779936021795e-06, + "loss": 0.3946, + "step": 4869 + }, + { + "epoch": 2.6438653637350704, + "grad_norm": 16.60173101533996, + "learning_rate": 9.555266548833728e-06, + "loss": 1.066, + "step": 4870 + }, + { + "epoch": 2.6444082519001086, + "grad_norm": 14.726517961222527, + "learning_rate": 9.551753216650998e-06, + "loss": 1.1356, + "step": 4871 + }, + { + "epoch": 2.6449511400651464, + "grad_norm": 11.123184991281361, + "learning_rate": 9.548239939908146e-06, + "loss": 0.757, + "step": 4872 + }, + { + "epoch": 2.6454940282301846, + "grad_norm": 11.666285426731331, + "learning_rate": 9.544726719039699e-06, + "loss": 0.7109, + "step": 4873 + }, + { + "epoch": 2.6460369163952224, + "grad_norm": 9.657892010966115, + "learning_rate": 9.541213554480176e-06, + "loss": 0.6191, + "step": 4874 + }, + { + "epoch": 2.6465798045602607, + "grad_norm": 9.984832056257446, + "learning_rate": 9.53770044666409e-06, + "loss": 0.8649, + "step": 4875 + }, + { + "epoch": 2.6471226927252984, + "grad_norm": 9.143852173958573, + "learning_rate": 9.534187396025955e-06, + "loss": 0.5282, + "step": 4876 + }, + { + "epoch": 2.6476655808903367, + "grad_norm": 8.129441981336987, + "learning_rate": 9.530674403000266e-06, + "loss": 0.3783, + "step": 4877 + }, + { + "epoch": 2.6482084690553744, + "grad_norm": 10.147175407433437, + "learning_rate": 9.527161468021508e-06, + "loss": 0.6676, + "step": 4878 + }, + { + "epoch": 2.6487513572204127, + "grad_norm": 11.015226101188722, + "learning_rate": 9.523648591524176e-06, + "loss": 0.5218, + "step": 4879 + }, + { + "epoch": 2.6492942453854504, + "grad_norm": 9.374140703922862, + "learning_rate": 9.520135773942743e-06, + "loss": 0.6754, + "step": 4880 + }, + { + "epoch": 2.6498371335504887, + "grad_norm": 8.84122429163583, + "learning_rate": 9.516623015711688e-06, + "loss": 0.5236, + "step": 4881 + }, + { + "epoch": 2.6503800217155264, + "grad_norm": 10.659047956662723, + "learning_rate": 9.513110317265463e-06, + "loss": 0.4742, + "step": 4882 + }, + { + "epoch": 2.6509229098805647, + "grad_norm": 8.798273983585764, + "learning_rate": 9.509597679038536e-06, + "loss": 0.4949, + "step": 4883 + }, + { + "epoch": 2.6514657980456025, + "grad_norm": 14.827321458575321, + "learning_rate": 9.506085101465344e-06, + "loss": 0.967, + "step": 4884 + }, + { + "epoch": 2.6520086862106407, + "grad_norm": 13.212543627120823, + "learning_rate": 9.502572584980338e-06, + "loss": 0.904, + "step": 4885 + }, + { + "epoch": 2.6525515743756785, + "grad_norm": 12.151700192108333, + "learning_rate": 9.499060130017947e-06, + "loss": 0.6512, + "step": 4886 + }, + { + "epoch": 2.6530944625407167, + "grad_norm": 9.051653631541132, + "learning_rate": 9.495547737012594e-06, + "loss": 0.5264, + "step": 4887 + }, + { + "epoch": 2.6536373507057545, + "grad_norm": 11.286333388065833, + "learning_rate": 9.492035406398706e-06, + "loss": 0.5864, + "step": 4888 + }, + { + "epoch": 2.6541802388707927, + "grad_norm": 8.303712468523708, + "learning_rate": 9.488523138610684e-06, + "loss": 0.5887, + "step": 4889 + }, + { + "epoch": 2.6547231270358305, + "grad_norm": 8.762643235318839, + "learning_rate": 9.485010934082939e-06, + "loss": 0.5079, + "step": 4890 + }, + { + "epoch": 2.6552660152008687, + "grad_norm": 13.211351989906888, + "learning_rate": 9.481498793249859e-06, + "loss": 0.8983, + "step": 4891 + }, + { + "epoch": 2.6558089033659065, + "grad_norm": 11.087558391384889, + "learning_rate": 9.477986716545834e-06, + "loss": 1.1671, + "step": 4892 + }, + { + "epoch": 2.6563517915309447, + "grad_norm": 6.825801048730381, + "learning_rate": 9.474474704405245e-06, + "loss": 0.3664, + "step": 4893 + }, + { + "epoch": 2.6568946796959825, + "grad_norm": 6.583440546159338, + "learning_rate": 9.470962757262465e-06, + "loss": 0.5464, + "step": 4894 + }, + { + "epoch": 2.6574375678610207, + "grad_norm": 8.987234989797157, + "learning_rate": 9.46745087555185e-06, + "loss": 0.5954, + "step": 4895 + }, + { + "epoch": 2.6579804560260585, + "grad_norm": 10.699367568639158, + "learning_rate": 9.463939059707763e-06, + "loss": 0.632, + "step": 4896 + }, + { + "epoch": 2.6585233441910967, + "grad_norm": 10.246298042329498, + "learning_rate": 9.460427310164548e-06, + "loss": 0.6766, + "step": 4897 + }, + { + "epoch": 2.6590662323561345, + "grad_norm": 12.300177238674, + "learning_rate": 9.456915627356542e-06, + "loss": 0.7277, + "step": 4898 + }, + { + "epoch": 2.6596091205211727, + "grad_norm": 9.893644041111745, + "learning_rate": 9.453404011718077e-06, + "loss": 0.644, + "step": 4899 + }, + { + "epoch": 2.6601520086862105, + "grad_norm": 10.80727748895235, + "learning_rate": 9.449892463683471e-06, + "loss": 0.5844, + "step": 4900 + }, + { + "epoch": 2.6606948968512487, + "grad_norm": 8.350836469669698, + "learning_rate": 9.446380983687046e-06, + "loss": 0.4745, + "step": 4901 + }, + { + "epoch": 2.6612377850162865, + "grad_norm": 10.697471057638213, + "learning_rate": 9.442869572163101e-06, + "loss": 0.5749, + "step": 4902 + }, + { + "epoch": 2.6617806731813247, + "grad_norm": 7.582465634030485, + "learning_rate": 9.43935822954594e-06, + "loss": 0.4481, + "step": 4903 + }, + { + "epoch": 2.6623235613463625, + "grad_norm": 10.01801356389796, + "learning_rate": 9.435846956269841e-06, + "loss": 0.5817, + "step": 4904 + }, + { + "epoch": 2.6628664495114007, + "grad_norm": 9.383471785643394, + "learning_rate": 9.432335752769092e-06, + "loss": 0.526, + "step": 4905 + }, + { + "epoch": 2.6634093376764385, + "grad_norm": 9.555536887097055, + "learning_rate": 9.428824619477964e-06, + "loss": 0.5056, + "step": 4906 + }, + { + "epoch": 2.6639522258414767, + "grad_norm": 7.636054626405326, + "learning_rate": 9.425313556830713e-06, + "loss": 0.4383, + "step": 4907 + }, + { + "epoch": 2.6644951140065145, + "grad_norm": 9.842725681691716, + "learning_rate": 9.421802565261602e-06, + "loss": 0.548, + "step": 4908 + }, + { + "epoch": 2.6650380021715527, + "grad_norm": 12.891942531084513, + "learning_rate": 9.418291645204865e-06, + "loss": 0.6489, + "step": 4909 + }, + { + "epoch": 2.6655808903365905, + "grad_norm": 15.298821348423894, + "learning_rate": 9.41478079709475e-06, + "loss": 0.8286, + "step": 4910 + }, + { + "epoch": 2.6661237785016287, + "grad_norm": 10.80698538838323, + "learning_rate": 9.411270021365475e-06, + "loss": 0.7704, + "step": 4911 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 10.837498285149955, + "learning_rate": 9.407759318451264e-06, + "loss": 0.5384, + "step": 4912 + }, + { + "epoch": 2.6672095548317047, + "grad_norm": 11.666848847579812, + "learning_rate": 9.404248688786322e-06, + "loss": 0.6402, + "step": 4913 + }, + { + "epoch": 2.6677524429967425, + "grad_norm": 12.038932116915525, + "learning_rate": 9.400738132804856e-06, + "loss": 0.559, + "step": 4914 + }, + { + "epoch": 2.6682953311617807, + "grad_norm": 13.411348889811451, + "learning_rate": 9.397227650941048e-06, + "loss": 0.9587, + "step": 4915 + }, + { + "epoch": 2.6688382193268185, + "grad_norm": 8.79354274091956, + "learning_rate": 9.393717243629091e-06, + "loss": 0.5627, + "step": 4916 + }, + { + "epoch": 2.6693811074918568, + "grad_norm": 8.956388294197298, + "learning_rate": 9.390206911303148e-06, + "loss": 0.5694, + "step": 4917 + }, + { + "epoch": 2.6699239956568945, + "grad_norm": 11.296789676569125, + "learning_rate": 9.386696654397384e-06, + "loss": 0.8356, + "step": 4918 + }, + { + "epoch": 2.6704668838219328, + "grad_norm": 8.778967068719545, + "learning_rate": 9.38318647334596e-06, + "loss": 0.5283, + "step": 4919 + }, + { + "epoch": 2.6710097719869705, + "grad_norm": 10.041165955923448, + "learning_rate": 9.379676368583011e-06, + "loss": 0.7492, + "step": 4920 + }, + { + "epoch": 2.6715526601520088, + "grad_norm": 7.152774382953927, + "learning_rate": 9.376166340542682e-06, + "loss": 0.4563, + "step": 4921 + }, + { + "epoch": 2.6720955483170465, + "grad_norm": 10.189040304576451, + "learning_rate": 9.37265638965909e-06, + "loss": 0.5869, + "step": 4922 + }, + { + "epoch": 2.6726384364820848, + "grad_norm": 8.47728562416046, + "learning_rate": 9.369146516366359e-06, + "loss": 0.6168, + "step": 4923 + }, + { + "epoch": 2.6731813246471225, + "grad_norm": 9.47866332824828, + "learning_rate": 9.365636721098588e-06, + "loss": 0.5233, + "step": 4924 + }, + { + "epoch": 2.6737242128121608, + "grad_norm": 11.264475439058582, + "learning_rate": 9.362127004289883e-06, + "loss": 0.693, + "step": 4925 + }, + { + "epoch": 2.6742671009771986, + "grad_norm": 11.495205201206721, + "learning_rate": 9.358617366374327e-06, + "loss": 0.587, + "step": 4926 + }, + { + "epoch": 2.6748099891422368, + "grad_norm": 10.081706772004408, + "learning_rate": 9.355107807785993e-06, + "loss": 0.65, + "step": 4927 + }, + { + "epoch": 2.6753528773072746, + "grad_norm": 8.912371373305913, + "learning_rate": 9.351598328958957e-06, + "loss": 0.6475, + "step": 4928 + }, + { + "epoch": 2.675895765472313, + "grad_norm": 8.434539074965906, + "learning_rate": 9.348088930327269e-06, + "loss": 0.4321, + "step": 4929 + }, + { + "epoch": 2.6764386536373506, + "grad_norm": 10.356652945785703, + "learning_rate": 9.344579612324984e-06, + "loss": 0.4243, + "step": 4930 + }, + { + "epoch": 2.676981541802389, + "grad_norm": 10.975361657654028, + "learning_rate": 9.341070375386134e-06, + "loss": 0.6042, + "step": 4931 + }, + { + "epoch": 2.6775244299674266, + "grad_norm": 10.912795228418068, + "learning_rate": 9.337561219944756e-06, + "loss": 0.6655, + "step": 4932 + }, + { + "epoch": 2.678067318132465, + "grad_norm": 9.852947274867201, + "learning_rate": 9.334052146434857e-06, + "loss": 0.7696, + "step": 4933 + }, + { + "epoch": 2.6786102062975026, + "grad_norm": 7.323342130507832, + "learning_rate": 9.330543155290454e-06, + "loss": 0.3453, + "step": 4934 + }, + { + "epoch": 2.679153094462541, + "grad_norm": 7.335485317101141, + "learning_rate": 9.32703424694554e-06, + "loss": 0.4291, + "step": 4935 + }, + { + "epoch": 2.6796959826275786, + "grad_norm": 12.700387251668076, + "learning_rate": 9.3235254218341e-06, + "loss": 0.7156, + "step": 4936 + }, + { + "epoch": 2.680238870792617, + "grad_norm": 10.828431765432423, + "learning_rate": 9.320016680390115e-06, + "loss": 0.7293, + "step": 4937 + }, + { + "epoch": 2.6807817589576546, + "grad_norm": 9.224454671275938, + "learning_rate": 9.316508023047548e-06, + "loss": 0.4157, + "step": 4938 + }, + { + "epoch": 2.681324647122693, + "grad_norm": 13.095737517917204, + "learning_rate": 9.312999450240362e-06, + "loss": 0.7928, + "step": 4939 + }, + { + "epoch": 2.6818675352877306, + "grad_norm": 10.338830489426515, + "learning_rate": 9.309490962402498e-06, + "loss": 0.6826, + "step": 4940 + }, + { + "epoch": 2.682410423452769, + "grad_norm": 11.650643384237476, + "learning_rate": 9.305982559967894e-06, + "loss": 0.6709, + "step": 4941 + }, + { + "epoch": 2.6829533116178066, + "grad_norm": 9.153121673948778, + "learning_rate": 9.302474243370469e-06, + "loss": 0.4998, + "step": 4942 + }, + { + "epoch": 2.683496199782845, + "grad_norm": 11.760747525125, + "learning_rate": 9.298966013044144e-06, + "loss": 1.0389, + "step": 4943 + }, + { + "epoch": 2.6840390879478826, + "grad_norm": 13.657927099528628, + "learning_rate": 9.295457869422818e-06, + "loss": 0.9807, + "step": 4944 + }, + { + "epoch": 2.684581976112921, + "grad_norm": 12.726310962899635, + "learning_rate": 9.291949812940387e-06, + "loss": 0.6495, + "step": 4945 + }, + { + "epoch": 2.6851248642779586, + "grad_norm": 11.439120086220523, + "learning_rate": 9.288441844030735e-06, + "loss": 0.6616, + "step": 4946 + }, + { + "epoch": 2.685667752442997, + "grad_norm": 10.384831951074661, + "learning_rate": 9.284933963127724e-06, + "loss": 0.536, + "step": 4947 + }, + { + "epoch": 2.6862106406080346, + "grad_norm": 8.803102808116847, + "learning_rate": 9.281426170665225e-06, + "loss": 0.6621, + "step": 4948 + }, + { + "epoch": 2.686753528773073, + "grad_norm": 14.516361098495818, + "learning_rate": 9.277918467077077e-06, + "loss": 1.0579, + "step": 4949 + }, + { + "epoch": 2.6872964169381106, + "grad_norm": 11.08498885907847, + "learning_rate": 9.274410852797126e-06, + "loss": 0.463, + "step": 4950 + }, + { + "epoch": 2.687839305103149, + "grad_norm": 10.827937520590183, + "learning_rate": 9.270903328259195e-06, + "loss": 0.5015, + "step": 4951 + }, + { + "epoch": 2.6883821932681866, + "grad_norm": 11.572829724801146, + "learning_rate": 9.267395893897107e-06, + "loss": 0.663, + "step": 4952 + }, + { + "epoch": 2.688925081433225, + "grad_norm": 9.3016488903525, + "learning_rate": 9.26388855014466e-06, + "loss": 0.6876, + "step": 4953 + }, + { + "epoch": 2.6894679695982626, + "grad_norm": 17.324535315422423, + "learning_rate": 9.260381297435652e-06, + "loss": 1.1741, + "step": 4954 + }, + { + "epoch": 2.690010857763301, + "grad_norm": 11.465003825461562, + "learning_rate": 9.256874136203864e-06, + "loss": 0.6867, + "step": 4955 + }, + { + "epoch": 2.6905537459283386, + "grad_norm": 8.846935307480056, + "learning_rate": 9.253367066883063e-06, + "loss": 0.4518, + "step": 4956 + }, + { + "epoch": 2.691096634093377, + "grad_norm": 11.007513424878196, + "learning_rate": 9.249860089907018e-06, + "loss": 0.7987, + "step": 4957 + }, + { + "epoch": 2.6916395222584146, + "grad_norm": 12.297659996678936, + "learning_rate": 9.246353205709468e-06, + "loss": 0.588, + "step": 4958 + }, + { + "epoch": 2.692182410423453, + "grad_norm": 10.549862062131211, + "learning_rate": 9.24284641472416e-06, + "loss": 0.737, + "step": 4959 + }, + { + "epoch": 2.6927252985884906, + "grad_norm": 10.712460113916158, + "learning_rate": 9.23933971738481e-06, + "loss": 0.6092, + "step": 4960 + }, + { + "epoch": 2.693268186753529, + "grad_norm": 14.104109743551868, + "learning_rate": 9.235833114125141e-06, + "loss": 1.008, + "step": 4961 + }, + { + "epoch": 2.6938110749185666, + "grad_norm": 7.755916952898445, + "learning_rate": 9.232326605378843e-06, + "loss": 0.4708, + "step": 4962 + }, + { + "epoch": 2.694353963083605, + "grad_norm": 11.677016459785008, + "learning_rate": 9.228820191579618e-06, + "loss": 0.7099, + "step": 4963 + }, + { + "epoch": 2.6948968512486426, + "grad_norm": 10.070884129580872, + "learning_rate": 9.225313873161139e-06, + "loss": 0.6546, + "step": 4964 + }, + { + "epoch": 2.695439739413681, + "grad_norm": 8.108710814711873, + "learning_rate": 9.221807650557078e-06, + "loss": 0.6102, + "step": 4965 + }, + { + "epoch": 2.6959826275787186, + "grad_norm": 10.279519502018594, + "learning_rate": 9.218301524201087e-06, + "loss": 0.5783, + "step": 4966 + }, + { + "epoch": 2.696525515743757, + "grad_norm": 11.882105089319197, + "learning_rate": 9.214795494526806e-06, + "loss": 1.1489, + "step": 4967 + }, + { + "epoch": 2.6970684039087947, + "grad_norm": 9.767493447568013, + "learning_rate": 9.211289561967872e-06, + "loss": 0.6051, + "step": 4968 + }, + { + "epoch": 2.697611292073833, + "grad_norm": 11.690415242614575, + "learning_rate": 9.207783726957903e-06, + "loss": 0.592, + "step": 4969 + }, + { + "epoch": 2.6981541802388707, + "grad_norm": 7.427254360689941, + "learning_rate": 9.204277989930502e-06, + "loss": 0.4672, + "step": 4970 + }, + { + "epoch": 2.698697068403909, + "grad_norm": 9.41305388818184, + "learning_rate": 9.200772351319266e-06, + "loss": 0.515, + "step": 4971 + }, + { + "epoch": 2.6992399565689467, + "grad_norm": 11.02585368985108, + "learning_rate": 9.197266811557787e-06, + "loss": 0.7867, + "step": 4972 + }, + { + "epoch": 2.699782844733985, + "grad_norm": 10.462642998204657, + "learning_rate": 9.193761371079622e-06, + "loss": 0.6689, + "step": 4973 + }, + { + "epoch": 2.7003257328990227, + "grad_norm": 9.964585703249949, + "learning_rate": 9.190256030318339e-06, + "loss": 0.6676, + "step": 4974 + }, + { + "epoch": 2.700868621064061, + "grad_norm": 12.383650152937822, + "learning_rate": 9.186750789707478e-06, + "loss": 0.9124, + "step": 4975 + }, + { + "epoch": 2.7014115092290987, + "grad_norm": 7.428651587888086, + "learning_rate": 9.183245649680574e-06, + "loss": 0.6241, + "step": 4976 + }, + { + "epoch": 2.701954397394137, + "grad_norm": 8.171702461332401, + "learning_rate": 9.179740610671155e-06, + "loss": 0.468, + "step": 4977 + }, + { + "epoch": 2.7024972855591747, + "grad_norm": 15.630545262276021, + "learning_rate": 9.176235673112719e-06, + "loss": 1.4177, + "step": 4978 + }, + { + "epoch": 2.703040173724213, + "grad_norm": 9.04535779717474, + "learning_rate": 9.172730837438774e-06, + "loss": 0.504, + "step": 4979 + }, + { + "epoch": 2.7035830618892507, + "grad_norm": 9.263908210238581, + "learning_rate": 9.169226104082792e-06, + "loss": 0.4043, + "step": 4980 + }, + { + "epoch": 2.704125950054289, + "grad_norm": 10.022418469090024, + "learning_rate": 9.165721473478253e-06, + "loss": 0.6372, + "step": 4981 + }, + { + "epoch": 2.7046688382193267, + "grad_norm": 12.924869609819282, + "learning_rate": 9.16221694605861e-06, + "loss": 0.9433, + "step": 4982 + }, + { + "epoch": 2.705211726384365, + "grad_norm": 8.79512920135937, + "learning_rate": 9.158712522257309e-06, + "loss": 0.4725, + "step": 4983 + }, + { + "epoch": 2.7057546145494027, + "grad_norm": 11.753381756239166, + "learning_rate": 9.155208202507789e-06, + "loss": 0.7395, + "step": 4984 + }, + { + "epoch": 2.706297502714441, + "grad_norm": 6.454949124720575, + "learning_rate": 9.151703987243459e-06, + "loss": 0.3108, + "step": 4985 + }, + { + "epoch": 2.7068403908794787, + "grad_norm": 15.244281036714051, + "learning_rate": 9.148199876897737e-06, + "loss": 0.802, + "step": 4986 + }, + { + "epoch": 2.707383279044517, + "grad_norm": 11.093235618680396, + "learning_rate": 9.144695871904005e-06, + "loss": 0.6252, + "step": 4987 + }, + { + "epoch": 2.7079261672095547, + "grad_norm": 10.703835311171614, + "learning_rate": 9.141191972695655e-06, + "loss": 0.6642, + "step": 4988 + }, + { + "epoch": 2.708469055374593, + "grad_norm": 11.327820449600674, + "learning_rate": 9.137688179706049e-06, + "loss": 1.1613, + "step": 4989 + }, + { + "epoch": 2.7090119435396307, + "grad_norm": 7.527861958807383, + "learning_rate": 9.134184493368548e-06, + "loss": 0.4991, + "step": 4990 + }, + { + "epoch": 2.709554831704669, + "grad_norm": 10.199776895453677, + "learning_rate": 9.130680914116484e-06, + "loss": 0.977, + "step": 4991 + }, + { + "epoch": 2.7100977198697067, + "grad_norm": 9.778998506495286, + "learning_rate": 9.127177442383192e-06, + "loss": 0.6729, + "step": 4992 + }, + { + "epoch": 2.710640608034745, + "grad_norm": 10.831437132716289, + "learning_rate": 9.123674078601984e-06, + "loss": 0.6252, + "step": 4993 + }, + { + "epoch": 2.7111834961997827, + "grad_norm": 9.722012246624114, + "learning_rate": 9.120170823206165e-06, + "loss": 0.6169, + "step": 4994 + }, + { + "epoch": 2.711726384364821, + "grad_norm": 9.696836781330795, + "learning_rate": 9.116667676629019e-06, + "loss": 0.7824, + "step": 4995 + }, + { + "epoch": 2.7122692725298587, + "grad_norm": 8.532030770830737, + "learning_rate": 9.11316463930382e-06, + "loss": 0.7649, + "step": 4996 + }, + { + "epoch": 2.712812160694897, + "grad_norm": 10.155581062168379, + "learning_rate": 9.109661711663837e-06, + "loss": 0.5581, + "step": 4997 + }, + { + "epoch": 2.7133550488599347, + "grad_norm": 11.913092610845165, + "learning_rate": 9.106158894142307e-06, + "loss": 0.6256, + "step": 4998 + }, + { + "epoch": 2.713897937024973, + "grad_norm": 10.326179923874385, + "learning_rate": 9.102656187172475e-06, + "loss": 0.8703, + "step": 4999 + }, + { + "epoch": 2.7144408251900107, + "grad_norm": 10.51344451860866, + "learning_rate": 9.099153591187553e-06, + "loss": 0.8281, + "step": 5000 + }, + { + "epoch": 2.714983713355049, + "grad_norm": 9.101512118782772, + "learning_rate": 9.09565110662075e-06, + "loss": 0.4127, + "step": 5001 + }, + { + "epoch": 2.7155266015200867, + "grad_norm": 14.912595390103256, + "learning_rate": 9.092148733905257e-06, + "loss": 0.828, + "step": 5002 + }, + { + "epoch": 2.716069489685125, + "grad_norm": 10.567445107841705, + "learning_rate": 9.088646473474262e-06, + "loss": 0.8542, + "step": 5003 + }, + { + "epoch": 2.7166123778501627, + "grad_norm": 11.727344925695448, + "learning_rate": 9.085144325760922e-06, + "loss": 0.7593, + "step": 5004 + }, + { + "epoch": 2.717155266015201, + "grad_norm": 10.964390265146204, + "learning_rate": 9.081642291198387e-06, + "loss": 0.696, + "step": 5005 + }, + { + "epoch": 2.7176981541802387, + "grad_norm": 8.594728435704301, + "learning_rate": 9.0781403702198e-06, + "loss": 0.3582, + "step": 5006 + }, + { + "epoch": 2.718241042345277, + "grad_norm": 10.662428429321805, + "learning_rate": 9.074638563258279e-06, + "loss": 0.6213, + "step": 5007 + }, + { + "epoch": 2.7187839305103148, + "grad_norm": 13.853700163509458, + "learning_rate": 9.071136870746934e-06, + "loss": 0.653, + "step": 5008 + }, + { + "epoch": 2.719326818675353, + "grad_norm": 9.421600748659138, + "learning_rate": 9.067635293118862e-06, + "loss": 0.4653, + "step": 5009 + }, + { + "epoch": 2.7198697068403908, + "grad_norm": 15.40532102462592, + "learning_rate": 9.064133830807147e-06, + "loss": 0.8502, + "step": 5010 + }, + { + "epoch": 2.720412595005429, + "grad_norm": 14.246891389640892, + "learning_rate": 9.060632484244845e-06, + "loss": 0.711, + "step": 5011 + }, + { + "epoch": 2.7209554831704668, + "grad_norm": 13.657358501095473, + "learning_rate": 9.057131253865022e-06, + "loss": 1.0586, + "step": 5012 + }, + { + "epoch": 2.721498371335505, + "grad_norm": 11.08278909637767, + "learning_rate": 9.053630140100701e-06, + "loss": 0.5658, + "step": 5013 + }, + { + "epoch": 2.7220412595005428, + "grad_norm": 11.922765678431421, + "learning_rate": 9.050129143384917e-06, + "loss": 0.6929, + "step": 5014 + }, + { + "epoch": 2.722584147665581, + "grad_norm": 10.93390692730677, + "learning_rate": 9.046628264150674e-06, + "loss": 1.417, + "step": 5015 + }, + { + "epoch": 2.7231270358306188, + "grad_norm": 12.86011863530373, + "learning_rate": 9.043127502830964e-06, + "loss": 1.0364, + "step": 5016 + }, + { + "epoch": 2.723669923995657, + "grad_norm": 7.848749292562499, + "learning_rate": 9.039626859858773e-06, + "loss": 0.6696, + "step": 5017 + }, + { + "epoch": 2.7242128121606948, + "grad_norm": 11.212923870670298, + "learning_rate": 9.036126335667059e-06, + "loss": 0.5751, + "step": 5018 + }, + { + "epoch": 2.724755700325733, + "grad_norm": 9.666644723437658, + "learning_rate": 9.032625930688781e-06, + "loss": 0.5699, + "step": 5019 + }, + { + "epoch": 2.725298588490771, + "grad_norm": 8.860936490743164, + "learning_rate": 9.029125645356864e-06, + "loss": 0.4262, + "step": 5020 + }, + { + "epoch": 2.725841476655809, + "grad_norm": 8.597880980516297, + "learning_rate": 9.025625480104238e-06, + "loss": 0.5646, + "step": 5021 + }, + { + "epoch": 2.726384364820847, + "grad_norm": 11.334362425216222, + "learning_rate": 9.022125435363803e-06, + "loss": 0.7424, + "step": 5022 + }, + { + "epoch": 2.726927252985885, + "grad_norm": 10.551777615550284, + "learning_rate": 9.018625511568456e-06, + "loss": 0.4961, + "step": 5023 + }, + { + "epoch": 2.727470141150923, + "grad_norm": 9.061655577456406, + "learning_rate": 9.015125709151069e-06, + "loss": 0.6466, + "step": 5024 + }, + { + "epoch": 2.728013029315961, + "grad_norm": 12.51368822693772, + "learning_rate": 9.011626028544502e-06, + "loss": 0.8461, + "step": 5025 + }, + { + "epoch": 2.728555917480999, + "grad_norm": 9.972383117450178, + "learning_rate": 9.008126470181605e-06, + "loss": 0.7941, + "step": 5026 + }, + { + "epoch": 2.729098805646037, + "grad_norm": 9.372092186313578, + "learning_rate": 9.004627034495204e-06, + "loss": 0.5291, + "step": 5027 + }, + { + "epoch": 2.729641693811075, + "grad_norm": 11.649348346812042, + "learning_rate": 9.00112772191812e-06, + "loss": 0.658, + "step": 5028 + }, + { + "epoch": 2.730184581976113, + "grad_norm": 9.4217070731629, + "learning_rate": 8.997628532883149e-06, + "loss": 0.7391, + "step": 5029 + }, + { + "epoch": 2.730727470141151, + "grad_norm": 9.294886195505088, + "learning_rate": 8.994129467823083e-06, + "loss": 0.7955, + "step": 5030 + }, + { + "epoch": 2.731270358306189, + "grad_norm": 9.046521616766688, + "learning_rate": 8.990630527170684e-06, + "loss": 0.6163, + "step": 5031 + }, + { + "epoch": 2.731813246471227, + "grad_norm": 13.078763070358596, + "learning_rate": 8.987131711358714e-06, + "loss": 0.6352, + "step": 5032 + }, + { + "epoch": 2.732356134636265, + "grad_norm": 8.372819066707777, + "learning_rate": 8.983633020819906e-06, + "loss": 0.5139, + "step": 5033 + }, + { + "epoch": 2.732899022801303, + "grad_norm": 7.908976790945395, + "learning_rate": 8.980134455986982e-06, + "loss": 0.4263, + "step": 5034 + }, + { + "epoch": 2.733441910966341, + "grad_norm": 9.508847688216756, + "learning_rate": 8.97663601729266e-06, + "loss": 0.614, + "step": 5035 + }, + { + "epoch": 2.733984799131379, + "grad_norm": 9.135340041370506, + "learning_rate": 8.973137705169621e-06, + "loss": 0.5803, + "step": 5036 + }, + { + "epoch": 2.734527687296417, + "grad_norm": 10.900299047123902, + "learning_rate": 8.969639520050553e-06, + "loss": 0.4833, + "step": 5037 + }, + { + "epoch": 2.735070575461455, + "grad_norm": 9.243890767443602, + "learning_rate": 8.966141462368106e-06, + "loss": 0.4405, + "step": 5038 + }, + { + "epoch": 2.735613463626493, + "grad_norm": 11.372121051541363, + "learning_rate": 8.962643532554934e-06, + "loss": 0.6618, + "step": 5039 + }, + { + "epoch": 2.736156351791531, + "grad_norm": 11.318259460716959, + "learning_rate": 8.95914573104366e-06, + "loss": 1.0367, + "step": 5040 + }, + { + "epoch": 2.736699239956569, + "grad_norm": 9.055918919260264, + "learning_rate": 8.955648058266904e-06, + "loss": 0.5463, + "step": 5041 + }, + { + "epoch": 2.737242128121607, + "grad_norm": 12.452665937187591, + "learning_rate": 8.952150514657258e-06, + "loss": 0.9772, + "step": 5042 + }, + { + "epoch": 2.737785016286645, + "grad_norm": 11.532813733829784, + "learning_rate": 8.94865310064731e-06, + "loss": 0.6372, + "step": 5043 + }, + { + "epoch": 2.738327904451683, + "grad_norm": 9.800145767151978, + "learning_rate": 8.945155816669622e-06, + "loss": 0.6384, + "step": 5044 + }, + { + "epoch": 2.738870792616721, + "grad_norm": 13.619886494303222, + "learning_rate": 8.94165866315674e-06, + "loss": 0.7295, + "step": 5045 + }, + { + "epoch": 2.739413680781759, + "grad_norm": 13.097331078127143, + "learning_rate": 8.938161640541202e-06, + "loss": 1.1075, + "step": 5046 + }, + { + "epoch": 2.739956568946797, + "grad_norm": 17.458162898454454, + "learning_rate": 8.934664749255524e-06, + "loss": 1.0174, + "step": 5047 + }, + { + "epoch": 2.740499457111835, + "grad_norm": 7.51493556495456, + "learning_rate": 8.931167989732212e-06, + "loss": 0.6407, + "step": 5048 + }, + { + "epoch": 2.741042345276873, + "grad_norm": 11.717405055058808, + "learning_rate": 8.927671362403741e-06, + "loss": 0.6899, + "step": 5049 + }, + { + "epoch": 2.741585233441911, + "grad_norm": 11.433736258508228, + "learning_rate": 8.924174867702591e-06, + "loss": 0.8481, + "step": 5050 + }, + { + "epoch": 2.742128121606949, + "grad_norm": 11.800039466500984, + "learning_rate": 8.920678506061202e-06, + "loss": 0.6044, + "step": 5051 + }, + { + "epoch": 2.742671009771987, + "grad_norm": 8.499003916350551, + "learning_rate": 8.91718227791202e-06, + "loss": 0.475, + "step": 5052 + }, + { + "epoch": 2.743213897937025, + "grad_norm": 10.586652955325345, + "learning_rate": 8.913686183687459e-06, + "loss": 0.5853, + "step": 5053 + }, + { + "epoch": 2.743756786102063, + "grad_norm": 14.066921486210767, + "learning_rate": 8.910190223819919e-06, + "loss": 0.8164, + "step": 5054 + }, + { + "epoch": 2.744299674267101, + "grad_norm": 9.77990149419258, + "learning_rate": 8.906694398741792e-06, + "loss": 0.6053, + "step": 5055 + }, + { + "epoch": 2.744842562432139, + "grad_norm": 15.328290145831401, + "learning_rate": 8.903198708885442e-06, + "loss": 1.3555, + "step": 5056 + }, + { + "epoch": 2.745385450597177, + "grad_norm": 11.626317894758118, + "learning_rate": 8.899703154683228e-06, + "loss": 0.6692, + "step": 5057 + }, + { + "epoch": 2.745928338762215, + "grad_norm": 11.566102899846406, + "learning_rate": 8.896207736567476e-06, + "loss": 0.6589, + "step": 5058 + }, + { + "epoch": 2.746471226927253, + "grad_norm": 9.521589192518187, + "learning_rate": 8.892712454970512e-06, + "loss": 0.7559, + "step": 5059 + }, + { + "epoch": 2.747014115092291, + "grad_norm": 8.719647496284242, + "learning_rate": 8.889217310324636e-06, + "loss": 0.4123, + "step": 5060 + }, + { + "epoch": 2.747557003257329, + "grad_norm": 6.709750992678039, + "learning_rate": 8.885722303062136e-06, + "loss": 0.3492, + "step": 5061 + }, + { + "epoch": 2.748099891422367, + "grad_norm": 11.273080989021693, + "learning_rate": 8.882227433615275e-06, + "loss": 0.7416, + "step": 5062 + }, + { + "epoch": 2.748642779587405, + "grad_norm": 13.864388014548679, + "learning_rate": 8.87873270241631e-06, + "loss": 0.8197, + "step": 5063 + }, + { + "epoch": 2.749185667752443, + "grad_norm": 9.492558149681301, + "learning_rate": 8.875238109897468e-06, + "loss": 0.6517, + "step": 5064 + }, + { + "epoch": 2.749728555917481, + "grad_norm": 10.634433093222732, + "learning_rate": 8.87174365649097e-06, + "loss": 0.5431, + "step": 5065 + }, + { + "epoch": 2.750271444082519, + "grad_norm": 8.469745230774542, + "learning_rate": 8.868249342629015e-06, + "loss": 0.5189, + "step": 5066 + }, + { + "epoch": 2.750814332247557, + "grad_norm": 8.182617911488233, + "learning_rate": 8.864755168743783e-06, + "loss": 0.4322, + "step": 5067 + }, + { + "epoch": 2.751357220412595, + "grad_norm": 7.6335690540689685, + "learning_rate": 8.861261135267444e-06, + "loss": 0.5268, + "step": 5068 + }, + { + "epoch": 2.751900108577633, + "grad_norm": 11.52493444073719, + "learning_rate": 8.85776724263214e-06, + "loss": 0.9757, + "step": 5069 + }, + { + "epoch": 2.752442996742671, + "grad_norm": 9.591556669015672, + "learning_rate": 8.854273491270008e-06, + "loss": 0.5527, + "step": 5070 + }, + { + "epoch": 2.752985884907709, + "grad_norm": 11.178077098634855, + "learning_rate": 8.850779881613151e-06, + "loss": 0.581, + "step": 5071 + }, + { + "epoch": 2.753528773072747, + "grad_norm": 8.921814933147093, + "learning_rate": 8.847286414093673e-06, + "loss": 0.5604, + "step": 5072 + }, + { + "epoch": 2.754071661237785, + "grad_norm": 11.887605361931898, + "learning_rate": 8.84379308914365e-06, + "loss": 0.6055, + "step": 5073 + }, + { + "epoch": 2.754614549402823, + "grad_norm": 14.12418845494581, + "learning_rate": 8.840299907195137e-06, + "loss": 0.9614, + "step": 5074 + }, + { + "epoch": 2.755157437567861, + "grad_norm": 10.094252933928091, + "learning_rate": 8.836806868680185e-06, + "loss": 0.6836, + "step": 5075 + }, + { + "epoch": 2.755700325732899, + "grad_norm": 10.048762466683991, + "learning_rate": 8.833313974030807e-06, + "loss": 0.5722, + "step": 5076 + }, + { + "epoch": 2.756243213897937, + "grad_norm": 12.51909069586841, + "learning_rate": 8.829821223679022e-06, + "loss": 1.2238, + "step": 5077 + }, + { + "epoch": 2.756786102062975, + "grad_norm": 9.50434691504725, + "learning_rate": 8.826328618056808e-06, + "loss": 0.6393, + "step": 5078 + }, + { + "epoch": 2.757328990228013, + "grad_norm": 10.737059961404412, + "learning_rate": 8.822836157596141e-06, + "loss": 0.7401, + "step": 5079 + }, + { + "epoch": 2.757871878393051, + "grad_norm": 9.16480609854369, + "learning_rate": 8.819343842728976e-06, + "loss": 0.5289, + "step": 5080 + }, + { + "epoch": 2.758414766558089, + "grad_norm": 10.849970138344764, + "learning_rate": 8.815851673887248e-06, + "loss": 0.6157, + "step": 5081 + }, + { + "epoch": 2.758957654723127, + "grad_norm": 8.609887529945185, + "learning_rate": 8.812359651502872e-06, + "loss": 0.5961, + "step": 5082 + }, + { + "epoch": 2.759500542888165, + "grad_norm": 10.765964781119916, + "learning_rate": 8.808867776007745e-06, + "loss": 1.1942, + "step": 5083 + }, + { + "epoch": 2.760043431053203, + "grad_norm": 10.466671227569803, + "learning_rate": 8.80537604783375e-06, + "loss": 0.7117, + "step": 5084 + }, + { + "epoch": 2.760586319218241, + "grad_norm": 10.079228354003783, + "learning_rate": 8.801884467412747e-06, + "loss": 0.4238, + "step": 5085 + }, + { + "epoch": 2.761129207383279, + "grad_norm": 11.422859396303908, + "learning_rate": 8.798393035176588e-06, + "loss": 0.8841, + "step": 5086 + }, + { + "epoch": 2.761672095548317, + "grad_norm": 9.736302153055926, + "learning_rate": 8.79490175155709e-06, + "loss": 0.6414, + "step": 5087 + }, + { + "epoch": 2.762214983713355, + "grad_norm": 8.513897139101958, + "learning_rate": 8.791410616986067e-06, + "loss": 0.3052, + "step": 5088 + }, + { + "epoch": 2.762757871878393, + "grad_norm": 10.101572440599828, + "learning_rate": 8.787919631895301e-06, + "loss": 0.5305, + "step": 5089 + }, + { + "epoch": 2.763300760043431, + "grad_norm": 12.17319240146357, + "learning_rate": 8.784428796716571e-06, + "loss": 0.6585, + "step": 5090 + }, + { + "epoch": 2.763843648208469, + "grad_norm": 10.930394499045825, + "learning_rate": 8.78093811188162e-06, + "loss": 0.5984, + "step": 5091 + }, + { + "epoch": 2.764386536373507, + "grad_norm": 9.550475415502788, + "learning_rate": 8.77744757782219e-06, + "loss": 0.5634, + "step": 5092 + }, + { + "epoch": 2.764929424538545, + "grad_norm": 12.890085749205962, + "learning_rate": 8.773957194969993e-06, + "loss": 0.9274, + "step": 5093 + }, + { + "epoch": 2.765472312703583, + "grad_norm": 12.38000962026181, + "learning_rate": 8.77046696375672e-06, + "loss": 0.6988, + "step": 5094 + }, + { + "epoch": 2.766015200868621, + "grad_norm": 9.523692226826551, + "learning_rate": 8.766976884614056e-06, + "loss": 0.4856, + "step": 5095 + }, + { + "epoch": 2.766558089033659, + "grad_norm": 9.26829606100349, + "learning_rate": 8.763486957973652e-06, + "loss": 0.6863, + "step": 5096 + }, + { + "epoch": 2.767100977198697, + "grad_norm": 11.227870751984687, + "learning_rate": 8.759997184267155e-06, + "loss": 0.6344, + "step": 5097 + }, + { + "epoch": 2.767643865363735, + "grad_norm": 11.653609442391305, + "learning_rate": 8.756507563926182e-06, + "loss": 0.5884, + "step": 5098 + }, + { + "epoch": 2.768186753528773, + "grad_norm": 11.59192980688709, + "learning_rate": 8.753018097382336e-06, + "loss": 0.8214, + "step": 5099 + }, + { + "epoch": 2.768729641693811, + "grad_norm": 8.668298021688548, + "learning_rate": 8.749528785067196e-06, + "loss": 0.6487, + "step": 5100 + }, + { + "epoch": 2.769272529858849, + "grad_norm": 5.456725122843202, + "learning_rate": 8.746039627412333e-06, + "loss": 0.2868, + "step": 5101 + }, + { + "epoch": 2.769815418023887, + "grad_norm": 12.29265212176352, + "learning_rate": 8.742550624849288e-06, + "loss": 0.512, + "step": 5102 + }, + { + "epoch": 2.770358306188925, + "grad_norm": 14.125650271042089, + "learning_rate": 8.73906177780958e-06, + "loss": 0.8903, + "step": 5103 + }, + { + "epoch": 2.770901194353963, + "grad_norm": 12.046023026774584, + "learning_rate": 8.735573086724725e-06, + "loss": 0.801, + "step": 5104 + }, + { + "epoch": 2.771444082519001, + "grad_norm": 10.153944591634119, + "learning_rate": 8.732084552026203e-06, + "loss": 0.6888, + "step": 5105 + }, + { + "epoch": 2.771986970684039, + "grad_norm": 10.581737293617769, + "learning_rate": 8.72859617414549e-06, + "loss": 0.7775, + "step": 5106 + }, + { + "epoch": 2.772529858849077, + "grad_norm": 8.304371234513637, + "learning_rate": 8.725107953514021e-06, + "loss": 0.4704, + "step": 5107 + }, + { + "epoch": 2.773072747014115, + "grad_norm": 12.579288268640527, + "learning_rate": 8.72161989056324e-06, + "loss": 0.5861, + "step": 5108 + }, + { + "epoch": 2.773615635179153, + "grad_norm": 10.68813978428288, + "learning_rate": 8.718131985724542e-06, + "loss": 0.6367, + "step": 5109 + }, + { + "epoch": 2.774158523344191, + "grad_norm": 8.921584649335454, + "learning_rate": 8.714644239429326e-06, + "loss": 0.5597, + "step": 5110 + }, + { + "epoch": 2.774701411509229, + "grad_norm": 10.473739328784887, + "learning_rate": 8.711156652108957e-06, + "loss": 0.8587, + "step": 5111 + }, + { + "epoch": 2.775244299674267, + "grad_norm": 10.905504725358476, + "learning_rate": 8.70766922419479e-06, + "loss": 0.5729, + "step": 5112 + }, + { + "epoch": 2.7757871878393052, + "grad_norm": 7.88310879226585, + "learning_rate": 8.704181956118153e-06, + "loss": 0.4998, + "step": 5113 + }, + { + "epoch": 2.776330076004343, + "grad_norm": 8.595679686197345, + "learning_rate": 8.700694848310354e-06, + "loss": 0.699, + "step": 5114 + }, + { + "epoch": 2.7768729641693812, + "grad_norm": 9.784504057993846, + "learning_rate": 8.697207901202691e-06, + "loss": 0.6509, + "step": 5115 + }, + { + "epoch": 2.777415852334419, + "grad_norm": 9.537890155200703, + "learning_rate": 8.693721115226427e-06, + "loss": 0.4484, + "step": 5116 + }, + { + "epoch": 2.7779587404994572, + "grad_norm": 10.110968424420188, + "learning_rate": 8.69023449081282e-06, + "loss": 0.6746, + "step": 5117 + }, + { + "epoch": 2.778501628664495, + "grad_norm": 14.837427287717714, + "learning_rate": 8.686748028393096e-06, + "loss": 1.3601, + "step": 5118 + }, + { + "epoch": 2.7790445168295332, + "grad_norm": 11.553555350476039, + "learning_rate": 8.683261728398472e-06, + "loss": 0.6054, + "step": 5119 + }, + { + "epoch": 2.779587404994571, + "grad_norm": 12.684609984344696, + "learning_rate": 8.679775591260132e-06, + "loss": 0.947, + "step": 5120 + }, + { + "epoch": 2.7801302931596092, + "grad_norm": 9.50144977278699, + "learning_rate": 8.676289617409256e-06, + "loss": 0.6401, + "step": 5121 + }, + { + "epoch": 2.780673181324647, + "grad_norm": 9.588585115620386, + "learning_rate": 8.672803807276988e-06, + "loss": 0.7214, + "step": 5122 + }, + { + "epoch": 2.7812160694896852, + "grad_norm": 14.597044499431716, + "learning_rate": 8.66931816129446e-06, + "loss": 1.0037, + "step": 5123 + }, + { + "epoch": 2.781758957654723, + "grad_norm": 10.552996924508232, + "learning_rate": 8.665832679892783e-06, + "loss": 0.5097, + "step": 5124 + }, + { + "epoch": 2.7823018458197613, + "grad_norm": 8.368460915764464, + "learning_rate": 8.662347363503043e-06, + "loss": 0.5289, + "step": 5125 + }, + { + "epoch": 2.782844733984799, + "grad_norm": 7.412774968306893, + "learning_rate": 8.658862212556318e-06, + "loss": 0.4117, + "step": 5126 + }, + { + "epoch": 2.7833876221498373, + "grad_norm": 7.965326482678092, + "learning_rate": 8.655377227483648e-06, + "loss": 0.4492, + "step": 5127 + }, + { + "epoch": 2.783930510314875, + "grad_norm": 10.16388699920939, + "learning_rate": 8.65189240871607e-06, + "loss": 0.6847, + "step": 5128 + }, + { + "epoch": 2.7844733984799133, + "grad_norm": 8.606167058304848, + "learning_rate": 8.648407756684582e-06, + "loss": 0.5952, + "step": 5129 + }, + { + "epoch": 2.785016286644951, + "grad_norm": 16.346182786192927, + "learning_rate": 8.64492327182018e-06, + "loss": 1.0626, + "step": 5130 + }, + { + "epoch": 2.7855591748099893, + "grad_norm": 9.397161435588243, + "learning_rate": 8.64143895455383e-06, + "loss": 0.6416, + "step": 5131 + }, + { + "epoch": 2.786102062975027, + "grad_norm": 10.215541582389136, + "learning_rate": 8.63795480531647e-06, + "loss": 0.5311, + "step": 5132 + }, + { + "epoch": 2.7866449511400653, + "grad_norm": 13.362954088182768, + "learning_rate": 8.634470824539035e-06, + "loss": 1.0249, + "step": 5133 + }, + { + "epoch": 2.787187839305103, + "grad_norm": 13.843076908870096, + "learning_rate": 8.630987012652421e-06, + "loss": 0.767, + "step": 5134 + }, + { + "epoch": 2.7877307274701413, + "grad_norm": 10.581731921492887, + "learning_rate": 8.627503370087519e-06, + "loss": 0.6027, + "step": 5135 + }, + { + "epoch": 2.788273615635179, + "grad_norm": 11.161057087658808, + "learning_rate": 8.624019897275184e-06, + "loss": 0.8193, + "step": 5136 + }, + { + "epoch": 2.7888165038002173, + "grad_norm": 7.972214509867566, + "learning_rate": 8.620536594646262e-06, + "loss": 0.4776, + "step": 5137 + }, + { + "epoch": 2.789359391965255, + "grad_norm": 12.57635432127455, + "learning_rate": 8.61705346263157e-06, + "loss": 0.7194, + "step": 5138 + }, + { + "epoch": 2.7899022801302933, + "grad_norm": 7.077115343788877, + "learning_rate": 8.613570501661915e-06, + "loss": 0.3571, + "step": 5139 + }, + { + "epoch": 2.790445168295331, + "grad_norm": 13.158625298591346, + "learning_rate": 8.610087712168065e-06, + "loss": 0.5872, + "step": 5140 + }, + { + "epoch": 2.7909880564603693, + "grad_norm": 8.652483474203775, + "learning_rate": 8.606605094580788e-06, + "loss": 0.624, + "step": 5141 + }, + { + "epoch": 2.791530944625407, + "grad_norm": 8.73276991998644, + "learning_rate": 8.603122649330811e-06, + "loss": 0.4358, + "step": 5142 + }, + { + "epoch": 2.7920738327904453, + "grad_norm": 12.749713905257105, + "learning_rate": 8.599640376848849e-06, + "loss": 0.7574, + "step": 5143 + }, + { + "epoch": 2.792616720955483, + "grad_norm": 12.054416741196698, + "learning_rate": 8.5961582775656e-06, + "loss": 0.886, + "step": 5144 + }, + { + "epoch": 2.7931596091205213, + "grad_norm": 10.958538797565106, + "learning_rate": 8.592676351911728e-06, + "loss": 0.7312, + "step": 5145 + }, + { + "epoch": 2.793702497285559, + "grad_norm": 10.124234819891349, + "learning_rate": 8.589194600317894e-06, + "loss": 0.5059, + "step": 5146 + }, + { + "epoch": 2.7942453854505973, + "grad_norm": 12.84934263063866, + "learning_rate": 8.585713023214715e-06, + "loss": 0.7016, + "step": 5147 + }, + { + "epoch": 2.794788273615635, + "grad_norm": 11.859965090765057, + "learning_rate": 8.582231621032807e-06, + "loss": 0.7158, + "step": 5148 + }, + { + "epoch": 2.7953311617806733, + "grad_norm": 11.57843340583913, + "learning_rate": 8.578750394202749e-06, + "loss": 0.7103, + "step": 5149 + }, + { + "epoch": 2.795874049945711, + "grad_norm": 7.143782607145388, + "learning_rate": 8.575269343155108e-06, + "loss": 0.5275, + "step": 5150 + }, + { + "epoch": 2.7964169381107493, + "grad_norm": 12.095108158973952, + "learning_rate": 8.571788468320427e-06, + "loss": 0.7308, + "step": 5151 + }, + { + "epoch": 2.796959826275787, + "grad_norm": 11.153004271537196, + "learning_rate": 8.568307770129223e-06, + "loss": 0.7576, + "step": 5152 + }, + { + "epoch": 2.7975027144408253, + "grad_norm": 8.326263454089252, + "learning_rate": 8.564827249011998e-06, + "loss": 0.408, + "step": 5153 + }, + { + "epoch": 2.798045602605863, + "grad_norm": 13.46902871500607, + "learning_rate": 8.561346905399221e-06, + "loss": 0.6624, + "step": 5154 + }, + { + "epoch": 2.7985884907709013, + "grad_norm": 8.925567749596054, + "learning_rate": 8.557866739721356e-06, + "loss": 0.4557, + "step": 5155 + }, + { + "epoch": 2.799131378935939, + "grad_norm": 11.700305884529763, + "learning_rate": 8.554386752408827e-06, + "loss": 0.6818, + "step": 5156 + }, + { + "epoch": 2.7996742671009773, + "grad_norm": 9.49767460620272, + "learning_rate": 8.550906943892054e-06, + "loss": 0.7256, + "step": 5157 + }, + { + "epoch": 2.800217155266015, + "grad_norm": 11.749650388906772, + "learning_rate": 8.547427314601416e-06, + "loss": 0.6933, + "step": 5158 + }, + { + "epoch": 2.8007600434310533, + "grad_norm": 11.830065328109, + "learning_rate": 8.543947864967286e-06, + "loss": 0.4282, + "step": 5159 + }, + { + "epoch": 2.801302931596091, + "grad_norm": 12.11100989733807, + "learning_rate": 8.540468595419999e-06, + "loss": 0.3617, + "step": 5160 + }, + { + "epoch": 2.8018458197611293, + "grad_norm": 9.923764364092053, + "learning_rate": 8.536989506389889e-06, + "loss": 0.6469, + "step": 5161 + }, + { + "epoch": 2.802388707926167, + "grad_norm": 10.555923377692814, + "learning_rate": 8.533510598307244e-06, + "loss": 0.9672, + "step": 5162 + }, + { + "epoch": 2.8029315960912053, + "grad_norm": 12.544866122081839, + "learning_rate": 8.530031871602345e-06, + "loss": 0.6885, + "step": 5163 + }, + { + "epoch": 2.803474484256243, + "grad_norm": 11.680102099467787, + "learning_rate": 8.526553326705452e-06, + "loss": 0.5571, + "step": 5164 + }, + { + "epoch": 2.8040173724212814, + "grad_norm": 11.769615311559438, + "learning_rate": 8.523074964046785e-06, + "loss": 0.6176, + "step": 5165 + }, + { + "epoch": 2.804560260586319, + "grad_norm": 10.249168095085986, + "learning_rate": 8.519596784056567e-06, + "loss": 0.7974, + "step": 5166 + }, + { + "epoch": 2.8051031487513574, + "grad_norm": 7.602995989085923, + "learning_rate": 8.516118787164973e-06, + "loss": 0.4132, + "step": 5167 + }, + { + "epoch": 2.805646036916395, + "grad_norm": 9.32078805379704, + "learning_rate": 8.512640973802175e-06, + "loss": 0.6492, + "step": 5168 + }, + { + "epoch": 2.8061889250814334, + "grad_norm": 13.80437475122657, + "learning_rate": 8.50916334439831e-06, + "loss": 0.6663, + "step": 5169 + }, + { + "epoch": 2.806731813246471, + "grad_norm": 9.121579447999185, + "learning_rate": 8.5056858993835e-06, + "loss": 0.4546, + "step": 5170 + }, + { + "epoch": 2.8072747014115094, + "grad_norm": 8.073186873408805, + "learning_rate": 8.502208639187842e-06, + "loss": 0.4467, + "step": 5171 + }, + { + "epoch": 2.807817589576547, + "grad_norm": 10.498644561860338, + "learning_rate": 8.498731564241403e-06, + "loss": 0.4907, + "step": 5172 + }, + { + "epoch": 2.8083604777415854, + "grad_norm": 9.027361205852865, + "learning_rate": 8.495254674974239e-06, + "loss": 0.4581, + "step": 5173 + }, + { + "epoch": 2.808903365906623, + "grad_norm": 9.692775527105594, + "learning_rate": 8.491777971816372e-06, + "loss": 0.5898, + "step": 5174 + }, + { + "epoch": 2.8094462540716614, + "grad_norm": 9.171182553280271, + "learning_rate": 8.48830145519781e-06, + "loss": 0.4129, + "step": 5175 + }, + { + "epoch": 2.809989142236699, + "grad_norm": 11.635096708230561, + "learning_rate": 8.484825125548532e-06, + "loss": 0.7632, + "step": 5176 + }, + { + "epoch": 2.8105320304017374, + "grad_norm": 12.18174333972076, + "learning_rate": 8.481348983298503e-06, + "loss": 0.6158, + "step": 5177 + }, + { + "epoch": 2.811074918566775, + "grad_norm": 8.880576943532684, + "learning_rate": 8.477873028877645e-06, + "loss": 0.374, + "step": 5178 + }, + { + "epoch": 2.8116178067318134, + "grad_norm": 11.735757011642823, + "learning_rate": 8.474397262715884e-06, + "loss": 0.6772, + "step": 5179 + }, + { + "epoch": 2.812160694896851, + "grad_norm": 9.50535973508058, + "learning_rate": 8.470921685243098e-06, + "loss": 0.6021, + "step": 5180 + }, + { + "epoch": 2.8127035830618894, + "grad_norm": 11.780091573431722, + "learning_rate": 8.467446296889151e-06, + "loss": 0.6944, + "step": 5181 + }, + { + "epoch": 2.813246471226927, + "grad_norm": 10.177875512109438, + "learning_rate": 8.463971098083896e-06, + "loss": 0.4312, + "step": 5182 + }, + { + "epoch": 2.8137893593919654, + "grad_norm": 16.588608629478436, + "learning_rate": 8.460496089257136e-06, + "loss": 0.818, + "step": 5183 + }, + { + "epoch": 2.814332247557003, + "grad_norm": 12.416295723031903, + "learning_rate": 8.45702127083868e-06, + "loss": 0.5531, + "step": 5184 + }, + { + "epoch": 2.8148751357220414, + "grad_norm": 13.657036134609482, + "learning_rate": 8.453546643258286e-06, + "loss": 0.7343, + "step": 5185 + }, + { + "epoch": 2.815418023887079, + "grad_norm": 9.833034529955514, + "learning_rate": 8.450072206945715e-06, + "loss": 0.6592, + "step": 5186 + }, + { + "epoch": 2.8159609120521174, + "grad_norm": 10.036112921946827, + "learning_rate": 8.446597962330675e-06, + "loss": 0.8127, + "step": 5187 + }, + { + "epoch": 2.816503800217155, + "grad_norm": 12.901123274951528, + "learning_rate": 8.443123909842882e-06, + "loss": 1.0507, + "step": 5188 + }, + { + "epoch": 2.8170466883821934, + "grad_norm": 10.692152694001132, + "learning_rate": 8.439650049911999e-06, + "loss": 0.5602, + "step": 5189 + }, + { + "epoch": 2.817589576547231, + "grad_norm": 19.731345695292617, + "learning_rate": 8.436176382967692e-06, + "loss": 1.0473, + "step": 5190 + }, + { + "epoch": 2.8181324647122694, + "grad_norm": 14.567179107047348, + "learning_rate": 8.432702909439579e-06, + "loss": 0.6359, + "step": 5191 + }, + { + "epoch": 2.818675352877307, + "grad_norm": 13.853200529175792, + "learning_rate": 8.429229629757266e-06, + "loss": 0.8277, + "step": 5192 + }, + { + "epoch": 2.8192182410423454, + "grad_norm": 9.811701731637578, + "learning_rate": 8.425756544350338e-06, + "loss": 0.533, + "step": 5193 + }, + { + "epoch": 2.819761129207383, + "grad_norm": 11.486229326022617, + "learning_rate": 8.422283653648348e-06, + "loss": 0.9078, + "step": 5194 + }, + { + "epoch": 2.8203040173724214, + "grad_norm": 11.086905618662822, + "learning_rate": 8.418810958080832e-06, + "loss": 0.7101, + "step": 5195 + }, + { + "epoch": 2.820846905537459, + "grad_norm": 8.065915758013317, + "learning_rate": 8.415338458077293e-06, + "loss": 0.5744, + "step": 5196 + }, + { + "epoch": 2.8213897937024974, + "grad_norm": 9.063098353131359, + "learning_rate": 8.411866154067224e-06, + "loss": 0.5008, + "step": 5197 + }, + { + "epoch": 2.821932681867535, + "grad_norm": 12.457981789036095, + "learning_rate": 8.408394046480077e-06, + "loss": 0.6377, + "step": 5198 + }, + { + "epoch": 2.8224755700325734, + "grad_norm": 10.806638216082748, + "learning_rate": 8.404922135745295e-06, + "loss": 0.6137, + "step": 5199 + }, + { + "epoch": 2.823018458197611, + "grad_norm": 10.78339089292669, + "learning_rate": 8.401450422292281e-06, + "loss": 0.7984, + "step": 5200 + }, + { + "epoch": 2.8235613463626494, + "grad_norm": 9.574355357184858, + "learning_rate": 8.397978906550429e-06, + "loss": 0.6307, + "step": 5201 + }, + { + "epoch": 2.824104234527687, + "grad_norm": 9.457238410125305, + "learning_rate": 8.3945075889491e-06, + "loss": 0.6782, + "step": 5202 + }, + { + "epoch": 2.8246471226927254, + "grad_norm": 15.132207636686802, + "learning_rate": 8.391036469917626e-06, + "loss": 0.7843, + "step": 5203 + }, + { + "epoch": 2.8251900108577632, + "grad_norm": 12.425847301998207, + "learning_rate": 8.387565549885331e-06, + "loss": 0.9953, + "step": 5204 + }, + { + "epoch": 2.8257328990228014, + "grad_norm": 9.948899666849417, + "learning_rate": 8.384094829281495e-06, + "loss": 0.6601, + "step": 5205 + }, + { + "epoch": 2.8262757871878392, + "grad_norm": 9.412819573011207, + "learning_rate": 8.38062430853539e-06, + "loss": 0.3747, + "step": 5206 + }, + { + "epoch": 2.8268186753528775, + "grad_norm": 8.137925073552552, + "learning_rate": 8.377153988076247e-06, + "loss": 0.3818, + "step": 5207 + }, + { + "epoch": 2.8273615635179152, + "grad_norm": 8.994244424095461, + "learning_rate": 8.373683868333287e-06, + "loss": 0.796, + "step": 5208 + }, + { + "epoch": 2.8279044516829535, + "grad_norm": 7.570203929319876, + "learning_rate": 8.370213949735696e-06, + "loss": 0.5587, + "step": 5209 + }, + { + "epoch": 2.8284473398479912, + "grad_norm": 13.134200586671845, + "learning_rate": 8.366744232712645e-06, + "loss": 0.8274, + "step": 5210 + }, + { + "epoch": 2.8289902280130295, + "grad_norm": 9.833443120274195, + "learning_rate": 8.363274717693272e-06, + "loss": 0.7984, + "step": 5211 + }, + { + "epoch": 2.8295331161780672, + "grad_norm": 11.623259569780537, + "learning_rate": 8.359805405106685e-06, + "loss": 0.7305, + "step": 5212 + }, + { + "epoch": 2.8300760043431055, + "grad_norm": 10.494299983717905, + "learning_rate": 8.356336295381981e-06, + "loss": 0.6299, + "step": 5213 + }, + { + "epoch": 2.8306188925081432, + "grad_norm": 7.662506144803812, + "learning_rate": 8.352867388948223e-06, + "loss": 0.4432, + "step": 5214 + }, + { + "epoch": 2.8311617806731815, + "grad_norm": 11.250218590721435, + "learning_rate": 8.349398686234455e-06, + "loss": 0.5805, + "step": 5215 + }, + { + "epoch": 2.8317046688382193, + "grad_norm": 12.63080713053447, + "learning_rate": 8.345930187669685e-06, + "loss": 0.7236, + "step": 5216 + }, + { + "epoch": 2.8322475570032575, + "grad_norm": 12.902831533172366, + "learning_rate": 8.342461893682908e-06, + "loss": 0.6462, + "step": 5217 + }, + { + "epoch": 2.8327904451682953, + "grad_norm": 8.440334577273788, + "learning_rate": 8.338993804703082e-06, + "loss": 0.6557, + "step": 5218 + }, + { + "epoch": 2.8333333333333335, + "grad_norm": 8.946600771683327, + "learning_rate": 8.335525921159155e-06, + "loss": 0.4964, + "step": 5219 + }, + { + "epoch": 2.8338762214983713, + "grad_norm": 11.427720200073038, + "learning_rate": 8.332058243480032e-06, + "loss": 0.5832, + "step": 5220 + }, + { + "epoch": 2.8344191096634095, + "grad_norm": 10.663320470599537, + "learning_rate": 8.328590772094602e-06, + "loss": 0.385, + "step": 5221 + }, + { + "epoch": 2.8349619978284473, + "grad_norm": 10.091874752226794, + "learning_rate": 8.325123507431732e-06, + "loss": 0.8347, + "step": 5222 + }, + { + "epoch": 2.8355048859934855, + "grad_norm": 11.844959588854952, + "learning_rate": 8.321656449920252e-06, + "loss": 0.5577, + "step": 5223 + }, + { + "epoch": 2.8360477741585233, + "grad_norm": 10.895554793526724, + "learning_rate": 8.318189599988982e-06, + "loss": 0.7961, + "step": 5224 + }, + { + "epoch": 2.8365906623235615, + "grad_norm": 15.457894886801157, + "learning_rate": 8.314722958066699e-06, + "loss": 0.9784, + "step": 5225 + }, + { + "epoch": 2.8371335504885993, + "grad_norm": 9.287014113188532, + "learning_rate": 8.311256524582167e-06, + "loss": 0.5947, + "step": 5226 + }, + { + "epoch": 2.8376764386536375, + "grad_norm": 12.953808801302275, + "learning_rate": 8.307790299964119e-06, + "loss": 0.6073, + "step": 5227 + }, + { + "epoch": 2.8382193268186753, + "grad_norm": 12.53202947557571, + "learning_rate": 8.304324284641268e-06, + "loss": 0.8199, + "step": 5228 + }, + { + "epoch": 2.8387622149837135, + "grad_norm": 10.62038819495681, + "learning_rate": 8.300858479042291e-06, + "loss": 0.488, + "step": 5229 + }, + { + "epoch": 2.8393051031487513, + "grad_norm": 10.914681492949684, + "learning_rate": 8.29739288359584e-06, + "loss": 0.6159, + "step": 5230 + }, + { + "epoch": 2.8398479913137895, + "grad_norm": 9.514892864934003, + "learning_rate": 8.293927498730556e-06, + "loss": 0.8757, + "step": 5231 + }, + { + "epoch": 2.8403908794788273, + "grad_norm": 10.004218198768447, + "learning_rate": 8.290462324875036e-06, + "loss": 0.4543, + "step": 5232 + }, + { + "epoch": 2.8409337676438655, + "grad_norm": 7.955044675492623, + "learning_rate": 8.286997362457859e-06, + "loss": 0.3227, + "step": 5233 + }, + { + "epoch": 2.8414766558089033, + "grad_norm": 5.674531662978047, + "learning_rate": 8.283532611907577e-06, + "loss": 0.2674, + "step": 5234 + }, + { + "epoch": 2.8420195439739415, + "grad_norm": 10.825015665199876, + "learning_rate": 8.280068073652723e-06, + "loss": 0.7373, + "step": 5235 + }, + { + "epoch": 2.8425624321389793, + "grad_norm": 11.937523350465815, + "learning_rate": 8.276603748121786e-06, + "loss": 0.6019, + "step": 5236 + }, + { + "epoch": 2.8431053203040175, + "grad_norm": 11.660316243586271, + "learning_rate": 8.273139635743249e-06, + "loss": 0.8063, + "step": 5237 + }, + { + "epoch": 2.8436482084690553, + "grad_norm": 10.978384303044228, + "learning_rate": 8.269675736945547e-06, + "loss": 0.6178, + "step": 5238 + }, + { + "epoch": 2.8441910966340935, + "grad_norm": 10.489917172678325, + "learning_rate": 8.266212052157113e-06, + "loss": 0.4382, + "step": 5239 + }, + { + "epoch": 2.8447339847991313, + "grad_norm": 9.430697797795169, + "learning_rate": 8.262748581806336e-06, + "loss": 0.4811, + "step": 5240 + }, + { + "epoch": 2.8452768729641695, + "grad_norm": 13.552717272826865, + "learning_rate": 8.259285326321579e-06, + "loss": 0.6866, + "step": 5241 + }, + { + "epoch": 2.8458197611292073, + "grad_norm": 12.883094545932988, + "learning_rate": 8.25582228613119e-06, + "loss": 0.9637, + "step": 5242 + }, + { + "epoch": 2.8463626492942455, + "grad_norm": 18.04770906924292, + "learning_rate": 8.25235946166348e-06, + "loss": 0.801, + "step": 5243 + }, + { + "epoch": 2.8469055374592833, + "grad_norm": 12.482559439953985, + "learning_rate": 8.248896853346738e-06, + "loss": 0.7968, + "step": 5244 + }, + { + "epoch": 2.8474484256243215, + "grad_norm": 10.728211626455765, + "learning_rate": 8.245434461609221e-06, + "loss": 0.5192, + "step": 5245 + }, + { + "epoch": 2.8479913137893593, + "grad_norm": 10.793312471383818, + "learning_rate": 8.241972286879168e-06, + "loss": 0.6121, + "step": 5246 + }, + { + "epoch": 2.8485342019543975, + "grad_norm": 12.822804552881774, + "learning_rate": 8.238510329584782e-06, + "loss": 0.7807, + "step": 5247 + }, + { + "epoch": 2.8490770901194353, + "grad_norm": 14.208687960885756, + "learning_rate": 8.23504859015425e-06, + "loss": 0.73, + "step": 5248 + }, + { + "epoch": 2.8496199782844736, + "grad_norm": 13.561299495458854, + "learning_rate": 8.231587069015723e-06, + "loss": 1.269, + "step": 5249 + }, + { + "epoch": 2.8501628664495113, + "grad_norm": 12.542912517352129, + "learning_rate": 8.22812576659732e-06, + "loss": 0.5758, + "step": 5250 + }, + { + "epoch": 2.8507057546145496, + "grad_norm": 9.184884000564415, + "learning_rate": 8.22466468332715e-06, + "loss": 0.5323, + "step": 5251 + }, + { + "epoch": 2.8512486427795873, + "grad_norm": 10.234457404263107, + "learning_rate": 8.22120381963328e-06, + "loss": 0.5455, + "step": 5252 + }, + { + "epoch": 2.8517915309446256, + "grad_norm": 13.327787107696029, + "learning_rate": 8.217743175943756e-06, + "loss": 0.6097, + "step": 5253 + }, + { + "epoch": 2.8523344191096633, + "grad_norm": 10.375306748712186, + "learning_rate": 8.214282752686595e-06, + "loss": 0.9622, + "step": 5254 + }, + { + "epoch": 2.8528773072747016, + "grad_norm": 10.445896680603816, + "learning_rate": 8.210822550289794e-06, + "loss": 0.6347, + "step": 5255 + }, + { + "epoch": 2.8534201954397393, + "grad_norm": 8.375912125960554, + "learning_rate": 8.207362569181305e-06, + "loss": 0.419, + "step": 5256 + }, + { + "epoch": 2.8539630836047776, + "grad_norm": 12.786819256474223, + "learning_rate": 8.203902809789078e-06, + "loss": 1.0155, + "step": 5257 + }, + { + "epoch": 2.8545059717698154, + "grad_norm": 9.916362057542294, + "learning_rate": 8.200443272541007e-06, + "loss": 0.5501, + "step": 5258 + }, + { + "epoch": 2.8550488599348536, + "grad_norm": 12.303081851739867, + "learning_rate": 8.196983957864984e-06, + "loss": 0.5965, + "step": 5259 + }, + { + "epoch": 2.8555917480998914, + "grad_norm": 11.992761200744544, + "learning_rate": 8.19352486618886e-06, + "loss": 0.4804, + "step": 5260 + }, + { + "epoch": 2.8561346362649296, + "grad_norm": 11.342631954988665, + "learning_rate": 8.190065997940455e-06, + "loss": 0.6241, + "step": 5261 + }, + { + "epoch": 2.8566775244299674, + "grad_norm": 14.06120317718869, + "learning_rate": 8.186607353547578e-06, + "loss": 1.3883, + "step": 5262 + }, + { + "epoch": 2.8572204125950056, + "grad_norm": 9.985876763230308, + "learning_rate": 8.183148933437988e-06, + "loss": 0.6843, + "step": 5263 + }, + { + "epoch": 2.8577633007600434, + "grad_norm": 13.658693433537799, + "learning_rate": 8.179690738039436e-06, + "loss": 1.027, + "step": 5264 + }, + { + "epoch": 2.8583061889250816, + "grad_norm": 13.070549937452688, + "learning_rate": 8.176232767779634e-06, + "loss": 0.9171, + "step": 5265 + }, + { + "epoch": 2.8588490770901194, + "grad_norm": 10.202999590557772, + "learning_rate": 8.17277502308627e-06, + "loss": 0.6412, + "step": 5266 + }, + { + "epoch": 2.8593919652551576, + "grad_norm": 11.358682840396053, + "learning_rate": 8.169317504387002e-06, + "loss": 0.8102, + "step": 5267 + }, + { + "epoch": 2.8599348534201954, + "grad_norm": 12.795930497437475, + "learning_rate": 8.165860212109467e-06, + "loss": 0.7885, + "step": 5268 + }, + { + "epoch": 2.8604777415852336, + "grad_norm": 12.540977748153804, + "learning_rate": 8.162403146681263e-06, + "loss": 0.453, + "step": 5269 + }, + { + "epoch": 2.8610206297502714, + "grad_norm": 9.835742357069565, + "learning_rate": 8.158946308529964e-06, + "loss": 0.553, + "step": 5270 + }, + { + "epoch": 2.8615635179153096, + "grad_norm": 8.768245847600197, + "learning_rate": 8.15548969808312e-06, + "loss": 0.6084, + "step": 5271 + }, + { + "epoch": 2.8621064060803474, + "grad_norm": 11.27517155767251, + "learning_rate": 8.152033315768248e-06, + "loss": 0.708, + "step": 5272 + }, + { + "epoch": 2.8626492942453856, + "grad_norm": 10.175542619256326, + "learning_rate": 8.148577162012848e-06, + "loss": 0.567, + "step": 5273 + }, + { + "epoch": 2.8631921824104234, + "grad_norm": 13.49575371946775, + "learning_rate": 8.145121237244367e-06, + "loss": 0.7178, + "step": 5274 + }, + { + "epoch": 2.8637350705754616, + "grad_norm": 12.848120779354371, + "learning_rate": 8.141665541890254e-06, + "loss": 0.7601, + "step": 5275 + }, + { + "epoch": 2.8642779587404994, + "grad_norm": 9.660459046348308, + "learning_rate": 8.138210076377903e-06, + "loss": 0.5867, + "step": 5276 + }, + { + "epoch": 2.8648208469055376, + "grad_norm": 12.014506663974018, + "learning_rate": 8.134754841134704e-06, + "loss": 0.6157, + "step": 5277 + }, + { + "epoch": 2.8653637350705754, + "grad_norm": 9.006817793845983, + "learning_rate": 8.131299836587994e-06, + "loss": 0.417, + "step": 5278 + }, + { + "epoch": 2.8659066232356136, + "grad_norm": 14.075391102924492, + "learning_rate": 8.127845063165097e-06, + "loss": 0.9453, + "step": 5279 + }, + { + "epoch": 2.8664495114006514, + "grad_norm": 11.384071845529691, + "learning_rate": 8.124390521293311e-06, + "loss": 0.757, + "step": 5280 + }, + { + "epoch": 2.8669923995656896, + "grad_norm": 11.412348591705651, + "learning_rate": 8.12093621139989e-06, + "loss": 0.7149, + "step": 5281 + }, + { + "epoch": 2.8675352877307274, + "grad_norm": 10.533279510688539, + "learning_rate": 8.117482133912078e-06, + "loss": 0.4981, + "step": 5282 + }, + { + "epoch": 2.8680781758957656, + "grad_norm": 10.800216526898991, + "learning_rate": 8.114028289257071e-06, + "loss": 0.6033, + "step": 5283 + }, + { + "epoch": 2.8686210640608034, + "grad_norm": 10.834462944873792, + "learning_rate": 8.110574677862054e-06, + "loss": 0.5554, + "step": 5284 + }, + { + "epoch": 2.8691639522258416, + "grad_norm": 13.02642177423764, + "learning_rate": 8.10712130015417e-06, + "loss": 0.7303, + "step": 5285 + }, + { + "epoch": 2.8697068403908794, + "grad_norm": 13.210484866681897, + "learning_rate": 8.103668156560543e-06, + "loss": 0.9579, + "step": 5286 + }, + { + "epoch": 2.8702497285559176, + "grad_norm": 7.866855822388311, + "learning_rate": 8.100215247508259e-06, + "loss": 0.5878, + "step": 5287 + }, + { + "epoch": 2.8707926167209554, + "grad_norm": 10.857793278838274, + "learning_rate": 8.096762573424384e-06, + "loss": 0.6734, + "step": 5288 + }, + { + "epoch": 2.8713355048859937, + "grad_norm": 11.519175454268337, + "learning_rate": 8.093310134735945e-06, + "loss": 0.7203, + "step": 5289 + }, + { + "epoch": 2.8718783930510314, + "grad_norm": 12.16192132511212, + "learning_rate": 8.089857931869947e-06, + "loss": 0.5887, + "step": 5290 + }, + { + "epoch": 2.8724212812160697, + "grad_norm": 8.545914428475232, + "learning_rate": 8.086405965253364e-06, + "loss": 0.677, + "step": 5291 + }, + { + "epoch": 2.8729641693811074, + "grad_norm": 12.463412130388328, + "learning_rate": 8.08295423531314e-06, + "loss": 0.8954, + "step": 5292 + }, + { + "epoch": 2.8735070575461457, + "grad_norm": 10.296132022290973, + "learning_rate": 8.079502742476195e-06, + "loss": 0.6582, + "step": 5293 + }, + { + "epoch": 2.8740499457111834, + "grad_norm": 9.170071716678386, + "learning_rate": 8.076051487169407e-06, + "loss": 0.5406, + "step": 5294 + }, + { + "epoch": 2.8745928338762217, + "grad_norm": 14.43641149914591, + "learning_rate": 8.072600469819643e-06, + "loss": 0.7459, + "step": 5295 + }, + { + "epoch": 2.8751357220412594, + "grad_norm": 9.705510058137687, + "learning_rate": 8.069149690853719e-06, + "loss": 0.6341, + "step": 5296 + }, + { + "epoch": 2.8756786102062977, + "grad_norm": 11.720382904656155, + "learning_rate": 8.065699150698442e-06, + "loss": 0.6016, + "step": 5297 + }, + { + "epoch": 2.8762214983713354, + "grad_norm": 14.838955909579008, + "learning_rate": 8.062248849780578e-06, + "loss": 0.8757, + "step": 5298 + }, + { + "epoch": 2.8767643865363732, + "grad_norm": 10.594386765695665, + "learning_rate": 8.05879878852686e-06, + "loss": 0.4864, + "step": 5299 + }, + { + "epoch": 2.8773072747014115, + "grad_norm": 9.83553313698661, + "learning_rate": 8.055348967364007e-06, + "loss": 0.5116, + "step": 5300 + }, + { + "epoch": 2.8778501628664497, + "grad_norm": 12.500251504139728, + "learning_rate": 8.05189938671869e-06, + "loss": 0.891, + "step": 5301 + }, + { + "epoch": 2.8783930510314875, + "grad_norm": 8.988571871064325, + "learning_rate": 8.048450047017563e-06, + "loss": 0.3699, + "step": 5302 + }, + { + "epoch": 2.8789359391965252, + "grad_norm": 10.893701604379793, + "learning_rate": 8.045000948687242e-06, + "loss": 0.5933, + "step": 5303 + }, + { + "epoch": 2.8794788273615635, + "grad_norm": 10.474072412748654, + "learning_rate": 8.041552092154321e-06, + "loss": 0.8787, + "step": 5304 + }, + { + "epoch": 2.8800217155266017, + "grad_norm": 8.555744151752272, + "learning_rate": 8.038103477845357e-06, + "loss": 0.3793, + "step": 5305 + }, + { + "epoch": 2.8805646036916395, + "grad_norm": 9.792363571958322, + "learning_rate": 8.034655106186884e-06, + "loss": 0.5347, + "step": 5306 + }, + { + "epoch": 2.8811074918566772, + "grad_norm": 8.810561156197549, + "learning_rate": 8.031206977605399e-06, + "loss": 0.5128, + "step": 5307 + }, + { + "epoch": 2.8816503800217155, + "grad_norm": 12.383361133227218, + "learning_rate": 8.027759092527374e-06, + "loss": 0.7834, + "step": 5308 + }, + { + "epoch": 2.8821932681867537, + "grad_norm": 9.51185389588206, + "learning_rate": 8.024311451379247e-06, + "loss": 0.4661, + "step": 5309 + }, + { + "epoch": 2.8827361563517915, + "grad_norm": 10.564107506597352, + "learning_rate": 8.020864054587426e-06, + "loss": 0.8199, + "step": 5310 + }, + { + "epoch": 2.8832790445168293, + "grad_norm": 14.579672874726727, + "learning_rate": 8.017416902578296e-06, + "loss": 1.1114, + "step": 5311 + }, + { + "epoch": 2.8838219326818675, + "grad_norm": 10.739540875321001, + "learning_rate": 8.013969995778201e-06, + "loss": 0.5384, + "step": 5312 + }, + { + "epoch": 2.8843648208469057, + "grad_norm": 11.955074827477672, + "learning_rate": 8.010523334613466e-06, + "loss": 0.8475, + "step": 5313 + }, + { + "epoch": 2.8849077090119435, + "grad_norm": 12.842414101367579, + "learning_rate": 8.007076919510371e-06, + "loss": 0.7096, + "step": 5314 + }, + { + "epoch": 2.8854505971769813, + "grad_norm": 9.17766217916239, + "learning_rate": 8.003630750895183e-06, + "loss": 0.4673, + "step": 5315 + }, + { + "epoch": 2.8859934853420195, + "grad_norm": 13.581449358552742, + "learning_rate": 8.000184829194121e-06, + "loss": 0.7501, + "step": 5316 + }, + { + "epoch": 2.8865363735070577, + "grad_norm": 6.157129192897698, + "learning_rate": 7.99673915483339e-06, + "loss": 0.2521, + "step": 5317 + }, + { + "epoch": 2.8870792616720955, + "grad_norm": 9.179432551374715, + "learning_rate": 7.993293728239154e-06, + "loss": 0.5682, + "step": 5318 + }, + { + "epoch": 2.8876221498371333, + "grad_norm": 8.458089104581813, + "learning_rate": 7.989848549837544e-06, + "loss": 0.7131, + "step": 5319 + }, + { + "epoch": 2.8881650380021715, + "grad_norm": 11.039294062728864, + "learning_rate": 7.986403620054674e-06, + "loss": 0.5421, + "step": 5320 + }, + { + "epoch": 2.8887079261672097, + "grad_norm": 8.296966992960693, + "learning_rate": 7.982958939316607e-06, + "loss": 0.5809, + "step": 5321 + }, + { + "epoch": 2.8892508143322475, + "grad_norm": 9.862357699114755, + "learning_rate": 7.979514508049398e-06, + "loss": 0.7253, + "step": 5322 + }, + { + "epoch": 2.8897937024972853, + "grad_norm": 10.60481802559805, + "learning_rate": 7.976070326679053e-06, + "loss": 0.5593, + "step": 5323 + }, + { + "epoch": 2.8903365906623235, + "grad_norm": 12.452210665582081, + "learning_rate": 7.972626395631556e-06, + "loss": 0.9987, + "step": 5324 + }, + { + "epoch": 2.8908794788273617, + "grad_norm": 8.161513982505882, + "learning_rate": 7.969182715332855e-06, + "loss": 0.5242, + "step": 5325 + }, + { + "epoch": 2.8914223669923995, + "grad_norm": 9.073053317988423, + "learning_rate": 7.965739286208878e-06, + "loss": 0.5015, + "step": 5326 + }, + { + "epoch": 2.8919652551574373, + "grad_norm": 8.632672314975157, + "learning_rate": 7.962296108685507e-06, + "loss": 0.4398, + "step": 5327 + }, + { + "epoch": 2.8925081433224755, + "grad_norm": 12.024378540292089, + "learning_rate": 7.958853183188597e-06, + "loss": 0.7876, + "step": 5328 + }, + { + "epoch": 2.8930510314875137, + "grad_norm": 9.943181911450774, + "learning_rate": 7.955410510143982e-06, + "loss": 0.6408, + "step": 5329 + }, + { + "epoch": 2.8935939196525515, + "grad_norm": 11.580662527673274, + "learning_rate": 7.951968089977449e-06, + "loss": 0.6484, + "step": 5330 + }, + { + "epoch": 2.8941368078175893, + "grad_norm": 10.477418271397871, + "learning_rate": 7.948525923114773e-06, + "loss": 0.7197, + "step": 5331 + }, + { + "epoch": 2.8946796959826275, + "grad_norm": 8.959473056544534, + "learning_rate": 7.945084009981674e-06, + "loss": 0.5344, + "step": 5332 + }, + { + "epoch": 2.8952225841476658, + "grad_norm": 7.339268198721023, + "learning_rate": 7.941642351003867e-06, + "loss": 0.4373, + "step": 5333 + }, + { + "epoch": 2.8957654723127035, + "grad_norm": 11.710220973199451, + "learning_rate": 7.938200946607008e-06, + "loss": 0.7354, + "step": 5334 + }, + { + "epoch": 2.8963083604777413, + "grad_norm": 10.548034804633037, + "learning_rate": 7.934759797216744e-06, + "loss": 1.0912, + "step": 5335 + }, + { + "epoch": 2.8968512486427795, + "grad_norm": 7.427458411514844, + "learning_rate": 7.93131890325868e-06, + "loss": 0.4515, + "step": 5336 + }, + { + "epoch": 2.8973941368078178, + "grad_norm": 7.5336162811756475, + "learning_rate": 7.927878265158391e-06, + "loss": 0.4106, + "step": 5337 + }, + { + "epoch": 2.8979370249728555, + "grad_norm": 9.845844831585431, + "learning_rate": 7.924437883341424e-06, + "loss": 0.394, + "step": 5338 + }, + { + "epoch": 2.8984799131378933, + "grad_norm": 11.91533381613983, + "learning_rate": 7.920997758233282e-06, + "loss": 0.5863, + "step": 5339 + }, + { + "epoch": 2.8990228013029316, + "grad_norm": 7.229828747210386, + "learning_rate": 7.917557890259457e-06, + "loss": 0.5663, + "step": 5340 + }, + { + "epoch": 2.8995656894679698, + "grad_norm": 11.378354320604084, + "learning_rate": 7.914118279845385e-06, + "loss": 0.7622, + "step": 5341 + }, + { + "epoch": 2.9001085776330076, + "grad_norm": 7.591103634044025, + "learning_rate": 7.91067892741649e-06, + "loss": 0.3933, + "step": 5342 + }, + { + "epoch": 2.9006514657980453, + "grad_norm": 12.040725731955204, + "learning_rate": 7.907239833398154e-06, + "loss": 0.7143, + "step": 5343 + }, + { + "epoch": 2.9011943539630836, + "grad_norm": 14.022224272834306, + "learning_rate": 7.903800998215735e-06, + "loss": 0.7389, + "step": 5344 + }, + { + "epoch": 2.901737242128122, + "grad_norm": 9.316170523577583, + "learning_rate": 7.900362422294545e-06, + "loss": 0.6061, + "step": 5345 + }, + { + "epoch": 2.9022801302931596, + "grad_norm": 8.20692587330132, + "learning_rate": 7.896924106059882e-06, + "loss": 0.4652, + "step": 5346 + }, + { + "epoch": 2.9028230184581973, + "grad_norm": 12.086544028526333, + "learning_rate": 7.893486049936993e-06, + "loss": 0.6579, + "step": 5347 + }, + { + "epoch": 2.9033659066232356, + "grad_norm": 11.562435941032952, + "learning_rate": 7.89004825435111e-06, + "loss": 0.6619, + "step": 5348 + }, + { + "epoch": 2.903908794788274, + "grad_norm": 9.535770499152292, + "learning_rate": 7.88661071972742e-06, + "loss": 0.4646, + "step": 5349 + }, + { + "epoch": 2.9044516829533116, + "grad_norm": 9.216711023311518, + "learning_rate": 7.88317344649108e-06, + "loss": 0.8006, + "step": 5350 + }, + { + "epoch": 2.9049945711183494, + "grad_norm": 9.577156917577227, + "learning_rate": 7.879736435067228e-06, + "loss": 0.7352, + "step": 5351 + }, + { + "epoch": 2.9055374592833876, + "grad_norm": 7.559386359727195, + "learning_rate": 7.87629968588095e-06, + "loss": 0.3841, + "step": 5352 + }, + { + "epoch": 2.906080347448426, + "grad_norm": 13.704096139083338, + "learning_rate": 7.872863199357315e-06, + "loss": 0.992, + "step": 5353 + }, + { + "epoch": 2.9066232356134636, + "grad_norm": 9.988539472899605, + "learning_rate": 7.869426975921344e-06, + "loss": 0.455, + "step": 5354 + }, + { + "epoch": 2.9071661237785014, + "grad_norm": 11.253082081483269, + "learning_rate": 7.865991015998043e-06, + "loss": 0.8226, + "step": 5355 + }, + { + "epoch": 2.9077090119435396, + "grad_norm": 10.526429101674665, + "learning_rate": 7.862555320012373e-06, + "loss": 0.5357, + "step": 5356 + }, + { + "epoch": 2.908251900108578, + "grad_norm": 9.646031093569851, + "learning_rate": 7.859119888389271e-06, + "loss": 0.4366, + "step": 5357 + }, + { + "epoch": 2.9087947882736156, + "grad_norm": 11.695549963620051, + "learning_rate": 7.855684721553633e-06, + "loss": 0.5834, + "step": 5358 + }, + { + "epoch": 2.9093376764386534, + "grad_norm": 13.315809341009361, + "learning_rate": 7.852249819930324e-06, + "loss": 1.0645, + "step": 5359 + }, + { + "epoch": 2.9098805646036916, + "grad_norm": 12.349221876871807, + "learning_rate": 7.848815183944183e-06, + "loss": 0.577, + "step": 5360 + }, + { + "epoch": 2.91042345276873, + "grad_norm": 16.63305773978975, + "learning_rate": 7.845380814020004e-06, + "loss": 0.9412, + "step": 5361 + }, + { + "epoch": 2.9109663409337676, + "grad_norm": 8.450548593743887, + "learning_rate": 7.841946710582563e-06, + "loss": 0.4955, + "step": 5362 + }, + { + "epoch": 2.9115092290988054, + "grad_norm": 11.465063371310775, + "learning_rate": 7.838512874056589e-06, + "loss": 0.6399, + "step": 5363 + }, + { + "epoch": 2.9120521172638436, + "grad_norm": 8.712777448735437, + "learning_rate": 7.835079304866793e-06, + "loss": 0.309, + "step": 5364 + }, + { + "epoch": 2.912595005428882, + "grad_norm": 9.44188587266759, + "learning_rate": 7.831646003437835e-06, + "loss": 0.435, + "step": 5365 + }, + { + "epoch": 2.9131378935939196, + "grad_norm": 12.094488103222616, + "learning_rate": 7.828212970194357e-06, + "loss": 0.7245, + "step": 5366 + }, + { + "epoch": 2.9136807817589574, + "grad_norm": 11.219679589625631, + "learning_rate": 7.82478020556096e-06, + "loss": 0.8716, + "step": 5367 + }, + { + "epoch": 2.9142236699239956, + "grad_norm": 10.091805149105353, + "learning_rate": 7.821347709962211e-06, + "loss": 0.553, + "step": 5368 + }, + { + "epoch": 2.914766558089034, + "grad_norm": 12.047544009676127, + "learning_rate": 7.817915483822654e-06, + "loss": 0.7283, + "step": 5369 + }, + { + "epoch": 2.9153094462540716, + "grad_norm": 10.891198338724942, + "learning_rate": 7.814483527566783e-06, + "loss": 0.6233, + "step": 5370 + }, + { + "epoch": 2.9158523344191094, + "grad_norm": 8.372829160348498, + "learning_rate": 7.811051841619077e-06, + "loss": 0.3781, + "step": 5371 + }, + { + "epoch": 2.9163952225841476, + "grad_norm": 7.8341866326850464, + "learning_rate": 7.807620426403964e-06, + "loss": 0.3455, + "step": 5372 + }, + { + "epoch": 2.916938110749186, + "grad_norm": 11.836979275353194, + "learning_rate": 7.804189282345855e-06, + "loss": 0.5573, + "step": 5373 + }, + { + "epoch": 2.9174809989142236, + "grad_norm": 8.195935708953526, + "learning_rate": 7.800758409869111e-06, + "loss": 0.4219, + "step": 5374 + }, + { + "epoch": 2.9180238870792614, + "grad_norm": 12.93514826751349, + "learning_rate": 7.797327809398074e-06, + "loss": 1.0471, + "step": 5375 + }, + { + "epoch": 2.9185667752442996, + "grad_norm": 9.384641490219252, + "learning_rate": 7.793897481357047e-06, + "loss": 0.5004, + "step": 5376 + }, + { + "epoch": 2.919109663409338, + "grad_norm": 10.40559875162228, + "learning_rate": 7.79046742617029e-06, + "loss": 0.5434, + "step": 5377 + }, + { + "epoch": 2.9196525515743756, + "grad_norm": 10.11339680443781, + "learning_rate": 7.787037644262048e-06, + "loss": 0.6003, + "step": 5378 + }, + { + "epoch": 2.9201954397394134, + "grad_norm": 12.37057409853331, + "learning_rate": 7.783608136056512e-06, + "loss": 0.685, + "step": 5379 + }, + { + "epoch": 2.9207383279044516, + "grad_norm": 11.589913705537441, + "learning_rate": 7.780178901977857e-06, + "loss": 0.7646, + "step": 5380 + }, + { + "epoch": 2.92128121606949, + "grad_norm": 12.462452174896484, + "learning_rate": 7.77674994245021e-06, + "loss": 0.6442, + "step": 5381 + }, + { + "epoch": 2.9218241042345277, + "grad_norm": 9.567483949166586, + "learning_rate": 7.77332125789768e-06, + "loss": 0.4645, + "step": 5382 + }, + { + "epoch": 2.9223669923995654, + "grad_norm": 11.079212789248007, + "learning_rate": 7.76989284874432e-06, + "loss": 0.7285, + "step": 5383 + }, + { + "epoch": 2.9229098805646037, + "grad_norm": 14.661322701022721, + "learning_rate": 7.76646471541417e-06, + "loss": 0.7481, + "step": 5384 + }, + { + "epoch": 2.923452768729642, + "grad_norm": 13.968723101600672, + "learning_rate": 7.763036858331222e-06, + "loss": 0.6632, + "step": 5385 + }, + { + "epoch": 2.9239956568946797, + "grad_norm": 12.145893524297147, + "learning_rate": 7.759609277919442e-06, + "loss": 0.8389, + "step": 5386 + }, + { + "epoch": 2.9245385450597174, + "grad_norm": 10.052846347427483, + "learning_rate": 7.756181974602757e-06, + "loss": 0.5351, + "step": 5387 + }, + { + "epoch": 2.9250814332247557, + "grad_norm": 11.030654536959574, + "learning_rate": 7.75275494880506e-06, + "loss": 0.6469, + "step": 5388 + }, + { + "epoch": 2.925624321389794, + "grad_norm": 13.670311382356902, + "learning_rate": 7.749328200950215e-06, + "loss": 0.984, + "step": 5389 + }, + { + "epoch": 2.9261672095548317, + "grad_norm": 10.129392847361466, + "learning_rate": 7.745901731462041e-06, + "loss": 0.8149, + "step": 5390 + }, + { + "epoch": 2.9267100977198695, + "grad_norm": 7.867905327161546, + "learning_rate": 7.742475540764339e-06, + "loss": 0.4957, + "step": 5391 + }, + { + "epoch": 2.9272529858849077, + "grad_norm": 8.467285225937513, + "learning_rate": 7.739049629280854e-06, + "loss": 0.579, + "step": 5392 + }, + { + "epoch": 2.927795874049946, + "grad_norm": 8.215885752516279, + "learning_rate": 7.735623997435318e-06, + "loss": 0.4073, + "step": 5393 + }, + { + "epoch": 2.9283387622149837, + "grad_norm": 8.51797242171681, + "learning_rate": 7.732198645651413e-06, + "loss": 0.63, + "step": 5394 + }, + { + "epoch": 2.9288816503800215, + "grad_norm": 15.898718458126094, + "learning_rate": 7.728773574352795e-06, + "loss": 0.7468, + "step": 5395 + }, + { + "epoch": 2.9294245385450597, + "grad_norm": 11.412421000549452, + "learning_rate": 7.725348783963084e-06, + "loss": 0.9522, + "step": 5396 + }, + { + "epoch": 2.929967426710098, + "grad_norm": 9.428473061604137, + "learning_rate": 7.721924274905855e-06, + "loss": 0.5388, + "step": 5397 + }, + { + "epoch": 2.9305103148751357, + "grad_norm": 10.083367258939811, + "learning_rate": 7.718500047604667e-06, + "loss": 0.6605, + "step": 5398 + }, + { + "epoch": 2.9310532030401735, + "grad_norm": 12.85805147473796, + "learning_rate": 7.715076102483026e-06, + "loss": 0.7407, + "step": 5399 + }, + { + "epoch": 2.9315960912052117, + "grad_norm": 9.866769374539361, + "learning_rate": 7.711652439964415e-06, + "loss": 0.6007, + "step": 5400 + }, + { + "epoch": 2.93213897937025, + "grad_norm": 12.367813318049626, + "learning_rate": 7.708229060472277e-06, + "loss": 0.5462, + "step": 5401 + }, + { + "epoch": 2.9326818675352877, + "grad_norm": 13.515093387837554, + "learning_rate": 7.704805964430023e-06, + "loss": 1.0205, + "step": 5402 + }, + { + "epoch": 2.9332247557003255, + "grad_norm": 10.667525494454107, + "learning_rate": 7.701383152261022e-06, + "loss": 0.7873, + "step": 5403 + }, + { + "epoch": 2.9337676438653637, + "grad_norm": 13.024983806475014, + "learning_rate": 7.697960624388621e-06, + "loss": 0.6388, + "step": 5404 + }, + { + "epoch": 2.934310532030402, + "grad_norm": 7.885758920671933, + "learning_rate": 7.694538381236115e-06, + "loss": 0.44, + "step": 5405 + }, + { + "epoch": 2.9348534201954397, + "grad_norm": 9.862140672970398, + "learning_rate": 7.69111642322678e-06, + "loss": 0.4065, + "step": 5406 + }, + { + "epoch": 2.9353963083604775, + "grad_norm": 11.605538896695489, + "learning_rate": 7.687694750783844e-06, + "loss": 0.6574, + "step": 5407 + }, + { + "epoch": 2.9359391965255157, + "grad_norm": 9.384935873492674, + "learning_rate": 7.684273364330505e-06, + "loss": 0.573, + "step": 5408 + }, + { + "epoch": 2.936482084690554, + "grad_norm": 7.506351724565119, + "learning_rate": 7.68085226428993e-06, + "loss": 0.4421, + "step": 5409 + }, + { + "epoch": 2.9370249728555917, + "grad_norm": 9.783370092521682, + "learning_rate": 7.677431451085238e-06, + "loss": 0.49, + "step": 5410 + }, + { + "epoch": 2.9375678610206295, + "grad_norm": 10.612737595311591, + "learning_rate": 7.674010925139533e-06, + "loss": 0.6855, + "step": 5411 + }, + { + "epoch": 2.9381107491856677, + "grad_norm": 10.77127188283207, + "learning_rate": 7.670590686875856e-06, + "loss": 0.4701, + "step": 5412 + }, + { + "epoch": 2.938653637350706, + "grad_norm": 9.009892989770995, + "learning_rate": 7.66717073671724e-06, + "loss": 0.3802, + "step": 5413 + }, + { + "epoch": 2.9391965255157437, + "grad_norm": 9.718396629861862, + "learning_rate": 7.663751075086662e-06, + "loss": 0.6433, + "step": 5414 + }, + { + "epoch": 2.9397394136807815, + "grad_norm": 8.544955706387817, + "learning_rate": 7.66033170240708e-06, + "loss": 0.3449, + "step": 5415 + }, + { + "epoch": 2.9402823018458197, + "grad_norm": 12.174166457717448, + "learning_rate": 7.656912619101401e-06, + "loss": 0.4685, + "step": 5416 + }, + { + "epoch": 2.940825190010858, + "grad_norm": 8.829471274310071, + "learning_rate": 7.6534938255925e-06, + "loss": 0.6213, + "step": 5417 + }, + { + "epoch": 2.9413680781758957, + "grad_norm": 11.089620150396728, + "learning_rate": 7.650075322303222e-06, + "loss": 0.5682, + "step": 5418 + }, + { + "epoch": 2.9419109663409335, + "grad_norm": 7.099888540087664, + "learning_rate": 7.646657109656375e-06, + "loss": 0.2937, + "step": 5419 + }, + { + "epoch": 2.9424538545059717, + "grad_norm": 9.8313977098716, + "learning_rate": 7.64323918807473e-06, + "loss": 0.6224, + "step": 5420 + }, + { + "epoch": 2.94299674267101, + "grad_norm": 14.892485492602605, + "learning_rate": 7.639821557981013e-06, + "loss": 1.0678, + "step": 5421 + }, + { + "epoch": 2.9435396308360477, + "grad_norm": 8.648910922847081, + "learning_rate": 7.636404219797931e-06, + "loss": 0.3605, + "step": 5422 + }, + { + "epoch": 2.9440825190010855, + "grad_norm": 12.093378946161085, + "learning_rate": 7.63298717394814e-06, + "loss": 0.9133, + "step": 5423 + }, + { + "epoch": 2.9446254071661238, + "grad_norm": 9.302461570715176, + "learning_rate": 7.629570420854271e-06, + "loss": 0.5305, + "step": 5424 + }, + { + "epoch": 2.945168295331162, + "grad_norm": 10.492242594764365, + "learning_rate": 7.626153960938909e-06, + "loss": 0.6007, + "step": 5425 + }, + { + "epoch": 2.9457111834961998, + "grad_norm": 11.212789857802488, + "learning_rate": 7.622737794624605e-06, + "loss": 0.4854, + "step": 5426 + }, + { + "epoch": 2.9462540716612375, + "grad_norm": 11.607518266622822, + "learning_rate": 7.619321922333884e-06, + "loss": 0.5543, + "step": 5427 + }, + { + "epoch": 2.9467969598262758, + "grad_norm": 11.268089998182159, + "learning_rate": 7.615906344489219e-06, + "loss": 0.6998, + "step": 5428 + }, + { + "epoch": 2.947339847991314, + "grad_norm": 12.383131636281371, + "learning_rate": 7.61249106151306e-06, + "loss": 0.8179, + "step": 5429 + }, + { + "epoch": 2.9478827361563518, + "grad_norm": 11.603914807740104, + "learning_rate": 7.609076073827808e-06, + "loss": 0.9006, + "step": 5430 + }, + { + "epoch": 2.9484256243213895, + "grad_norm": 9.674307468707637, + "learning_rate": 7.6056613818558415e-06, + "loss": 0.6638, + "step": 5431 + }, + { + "epoch": 2.9489685124864278, + "grad_norm": 13.055177059547185, + "learning_rate": 7.602246986019487e-06, + "loss": 0.9642, + "step": 5432 + }, + { + "epoch": 2.949511400651466, + "grad_norm": 11.00057075160912, + "learning_rate": 7.598832886741049e-06, + "loss": 0.4843, + "step": 5433 + }, + { + "epoch": 2.950054288816504, + "grad_norm": 11.15214171419394, + "learning_rate": 7.5954190844427834e-06, + "loss": 0.6722, + "step": 5434 + }, + { + "epoch": 2.9505971769815416, + "grad_norm": 8.91145682769725, + "learning_rate": 7.592005579546924e-06, + "loss": 0.6258, + "step": 5435 + }, + { + "epoch": 2.95114006514658, + "grad_norm": 12.685207726915342, + "learning_rate": 7.588592372475651e-06, + "loss": 0.8481, + "step": 5436 + }, + { + "epoch": 2.951682953311618, + "grad_norm": 13.10934163141668, + "learning_rate": 7.585179463651112e-06, + "loss": 0.6655, + "step": 5437 + }, + { + "epoch": 2.952225841476656, + "grad_norm": 10.993826657178786, + "learning_rate": 7.5817668534954295e-06, + "loss": 0.5687, + "step": 5438 + }, + { + "epoch": 2.9527687296416936, + "grad_norm": 8.424743777393193, + "learning_rate": 7.578354542430673e-06, + "loss": 0.5485, + "step": 5439 + }, + { + "epoch": 2.953311617806732, + "grad_norm": 14.730970588587084, + "learning_rate": 7.574942530878891e-06, + "loss": 0.8759, + "step": 5440 + }, + { + "epoch": 2.95385450597177, + "grad_norm": 8.667781572678388, + "learning_rate": 7.571530819262077e-06, + "loss": 0.3704, + "step": 5441 + }, + { + "epoch": 2.954397394136808, + "grad_norm": 6.739834901590041, + "learning_rate": 7.568119408002206e-06, + "loss": 0.3619, + "step": 5442 + }, + { + "epoch": 2.9549402823018456, + "grad_norm": 5.9221212094491795, + "learning_rate": 7.564708297521197e-06, + "loss": 0.3177, + "step": 5443 + }, + { + "epoch": 2.955483170466884, + "grad_norm": 9.008687743468037, + "learning_rate": 7.561297488240953e-06, + "loss": 0.5756, + "step": 5444 + }, + { + "epoch": 2.956026058631922, + "grad_norm": 10.49570531884228, + "learning_rate": 7.55788698058332e-06, + "loss": 0.609, + "step": 5445 + }, + { + "epoch": 2.95656894679696, + "grad_norm": 11.053313913817204, + "learning_rate": 7.554476774970115e-06, + "loss": 0.6893, + "step": 5446 + }, + { + "epoch": 2.9571118349619976, + "grad_norm": 10.747800237741414, + "learning_rate": 7.551066871823125e-06, + "loss": 0.6361, + "step": 5447 + }, + { + "epoch": 2.957654723127036, + "grad_norm": 12.158436457559148, + "learning_rate": 7.547657271564083e-06, + "loss": 0.6427, + "step": 5448 + }, + { + "epoch": 2.958197611292074, + "grad_norm": 11.021084054397756, + "learning_rate": 7.544247974614701e-06, + "loss": 0.4399, + "step": 5449 + }, + { + "epoch": 2.958740499457112, + "grad_norm": 8.337942068125022, + "learning_rate": 7.540838981396641e-06, + "loss": 0.3781, + "step": 5450 + }, + { + "epoch": 2.9592833876221496, + "grad_norm": 9.86638573791189, + "learning_rate": 7.537430292331536e-06, + "loss": 0.4257, + "step": 5451 + }, + { + "epoch": 2.959826275787188, + "grad_norm": 11.512553822962575, + "learning_rate": 7.534021907840975e-06, + "loss": 0.64, + "step": 5452 + }, + { + "epoch": 2.960369163952226, + "grad_norm": 10.540338627981921, + "learning_rate": 7.530613828346519e-06, + "loss": 0.6669, + "step": 5453 + }, + { + "epoch": 2.960912052117264, + "grad_norm": 12.092382555989715, + "learning_rate": 7.5272060542696765e-06, + "loss": 0.6346, + "step": 5454 + }, + { + "epoch": 2.9614549402823016, + "grad_norm": 11.429168527550765, + "learning_rate": 7.5237985860319344e-06, + "loss": 0.7454, + "step": 5455 + }, + { + "epoch": 2.96199782844734, + "grad_norm": 10.331080624952307, + "learning_rate": 7.52039142405473e-06, + "loss": 0.5472, + "step": 5456 + }, + { + "epoch": 2.962540716612378, + "grad_norm": 16.484391160691512, + "learning_rate": 7.516984568759461e-06, + "loss": 1.3039, + "step": 5457 + }, + { + "epoch": 2.963083604777416, + "grad_norm": 12.853254391289898, + "learning_rate": 7.5135780205675e-06, + "loss": 0.6846, + "step": 5458 + }, + { + "epoch": 2.9636264929424536, + "grad_norm": 12.062646480747443, + "learning_rate": 7.510171779900171e-06, + "loss": 0.5656, + "step": 5459 + }, + { + "epoch": 2.964169381107492, + "grad_norm": 8.21924446991787, + "learning_rate": 7.506765847178768e-06, + "loss": 0.3401, + "step": 5460 + }, + { + "epoch": 2.96471226927253, + "grad_norm": 11.116098821110214, + "learning_rate": 7.503360222824535e-06, + "loss": 0.4931, + "step": 5461 + }, + { + "epoch": 2.965255157437568, + "grad_norm": 9.759573410672228, + "learning_rate": 7.499954907258693e-06, + "loss": 0.5363, + "step": 5462 + }, + { + "epoch": 2.9657980456026056, + "grad_norm": 15.195018271421278, + "learning_rate": 7.496549900902408e-06, + "loss": 0.5349, + "step": 5463 + }, + { + "epoch": 2.966340933767644, + "grad_norm": 11.555768004386108, + "learning_rate": 7.493145204176823e-06, + "loss": 0.8273, + "step": 5464 + }, + { + "epoch": 2.966883821932682, + "grad_norm": 12.516961882115927, + "learning_rate": 7.4897408175030366e-06, + "loss": 0.7826, + "step": 5465 + }, + { + "epoch": 2.96742671009772, + "grad_norm": 11.67450620684471, + "learning_rate": 7.486336741302103e-06, + "loss": 0.5689, + "step": 5466 + }, + { + "epoch": 2.9679695982627576, + "grad_norm": 13.075852566862391, + "learning_rate": 7.48293297599505e-06, + "loss": 0.8355, + "step": 5467 + }, + { + "epoch": 2.968512486427796, + "grad_norm": 8.943584587537828, + "learning_rate": 7.479529522002855e-06, + "loss": 0.3921, + "step": 5468 + }, + { + "epoch": 2.969055374592834, + "grad_norm": 10.91179454178022, + "learning_rate": 7.476126379746471e-06, + "loss": 0.7276, + "step": 5469 + }, + { + "epoch": 2.969598262757872, + "grad_norm": 11.79182407811329, + "learning_rate": 7.472723549646793e-06, + "loss": 1.012, + "step": 5470 + }, + { + "epoch": 2.9701411509229096, + "grad_norm": 12.476871332438998, + "learning_rate": 7.469321032124697e-06, + "loss": 0.7994, + "step": 5471 + }, + { + "epoch": 2.970684039087948, + "grad_norm": 14.350758882559228, + "learning_rate": 7.465918827601008e-06, + "loss": 0.8477, + "step": 5472 + }, + { + "epoch": 2.971226927252986, + "grad_norm": 13.79963097595528, + "learning_rate": 7.46251693649652e-06, + "loss": 1.0897, + "step": 5473 + }, + { + "epoch": 2.971769815418024, + "grad_norm": 9.061867840853571, + "learning_rate": 7.459115359231977e-06, + "loss": 0.2972, + "step": 5474 + }, + { + "epoch": 2.9723127035830617, + "grad_norm": 11.193898477701971, + "learning_rate": 7.455714096228102e-06, + "loss": 0.9238, + "step": 5475 + }, + { + "epoch": 2.9728555917481, + "grad_norm": 13.156649778942683, + "learning_rate": 7.452313147905559e-06, + "loss": 0.9945, + "step": 5476 + }, + { + "epoch": 2.973398479913138, + "grad_norm": 10.673119691620487, + "learning_rate": 7.448912514684985e-06, + "loss": 0.6296, + "step": 5477 + }, + { + "epoch": 2.973941368078176, + "grad_norm": 13.208883224921669, + "learning_rate": 7.44551219698698e-06, + "loss": 0.5018, + "step": 5478 + }, + { + "epoch": 2.9744842562432137, + "grad_norm": 10.477928943580013, + "learning_rate": 7.442112195232093e-06, + "loss": 0.4782, + "step": 5479 + }, + { + "epoch": 2.975027144408252, + "grad_norm": 10.052094162546567, + "learning_rate": 7.438712509840852e-06, + "loss": 0.5661, + "step": 5480 + }, + { + "epoch": 2.97557003257329, + "grad_norm": 14.921523553916886, + "learning_rate": 7.435313141233724e-06, + "loss": 1.3182, + "step": 5481 + }, + { + "epoch": 2.976112920738328, + "grad_norm": 9.3079131640488, + "learning_rate": 7.43191408983116e-06, + "loss": 0.5672, + "step": 5482 + }, + { + "epoch": 2.9766558089033657, + "grad_norm": 8.76037752811841, + "learning_rate": 7.428515356053551e-06, + "loss": 0.6026, + "step": 5483 + }, + { + "epoch": 2.977198697068404, + "grad_norm": 11.037798492427186, + "learning_rate": 7.425116940321262e-06, + "loss": 0.7053, + "step": 5484 + }, + { + "epoch": 2.977741585233442, + "grad_norm": 12.329608171901635, + "learning_rate": 7.421718843054615e-06, + "loss": 0.6576, + "step": 5485 + }, + { + "epoch": 2.97828447339848, + "grad_norm": 14.664205153544405, + "learning_rate": 7.4183210646738875e-06, + "loss": 0.7546, + "step": 5486 + }, + { + "epoch": 2.9788273615635177, + "grad_norm": 16.28337626285995, + "learning_rate": 7.414923605599329e-06, + "loss": 1.2774, + "step": 5487 + }, + { + "epoch": 2.979370249728556, + "grad_norm": 11.936294560502736, + "learning_rate": 7.411526466251135e-06, + "loss": 0.9397, + "step": 5488 + }, + { + "epoch": 2.979913137893594, + "grad_norm": 15.443918342772207, + "learning_rate": 7.408129647049474e-06, + "loss": 0.7591, + "step": 5489 + }, + { + "epoch": 2.980456026058632, + "grad_norm": 9.303363471188012, + "learning_rate": 7.404733148414471e-06, + "loss": 0.4917, + "step": 5490 + }, + { + "epoch": 2.9809989142236697, + "grad_norm": 16.111589693524394, + "learning_rate": 7.401336970766207e-06, + "loss": 0.8724, + "step": 5491 + }, + { + "epoch": 2.981541802388708, + "grad_norm": 15.913107093535366, + "learning_rate": 7.397941114524727e-06, + "loss": 1.0915, + "step": 5492 + }, + { + "epoch": 2.982084690553746, + "grad_norm": 12.432387434616404, + "learning_rate": 7.3945455801100404e-06, + "loss": 1.1143, + "step": 5493 + }, + { + "epoch": 2.982627578718784, + "grad_norm": 11.65910861056599, + "learning_rate": 7.39115036794211e-06, + "loss": 0.7012, + "step": 5494 + }, + { + "epoch": 2.9831704668838217, + "grad_norm": 10.645886415744744, + "learning_rate": 7.387755478440855e-06, + "loss": 0.5327, + "step": 5495 + }, + { + "epoch": 2.98371335504886, + "grad_norm": 11.340492191845808, + "learning_rate": 7.384360912026167e-06, + "loss": 0.8393, + "step": 5496 + }, + { + "epoch": 2.984256243213898, + "grad_norm": 11.222712079595647, + "learning_rate": 7.38096666911789e-06, + "loss": 0.4731, + "step": 5497 + }, + { + "epoch": 2.984799131378936, + "grad_norm": 14.06199638971423, + "learning_rate": 7.377572750135833e-06, + "loss": 1.0083, + "step": 5498 + }, + { + "epoch": 2.9853420195439737, + "grad_norm": 11.609516661749161, + "learning_rate": 7.374179155499752e-06, + "loss": 0.5397, + "step": 5499 + }, + { + "epoch": 2.985884907709012, + "grad_norm": 11.809076211376697, + "learning_rate": 7.370785885629384e-06, + "loss": 0.5726, + "step": 5500 + }, + { + "epoch": 2.98642779587405, + "grad_norm": 12.767099228571903, + "learning_rate": 7.367392940944403e-06, + "loss": 0.8919, + "step": 5501 + }, + { + "epoch": 2.986970684039088, + "grad_norm": 8.511059646511987, + "learning_rate": 7.364000321864464e-06, + "loss": 0.4431, + "step": 5502 + }, + { + "epoch": 2.9875135722041257, + "grad_norm": 14.166198872023001, + "learning_rate": 7.360608028809161e-06, + "loss": 0.8709, + "step": 5503 + }, + { + "epoch": 2.988056460369164, + "grad_norm": 11.774867588892974, + "learning_rate": 7.357216062198066e-06, + "loss": 0.4857, + "step": 5504 + }, + { + "epoch": 2.988599348534202, + "grad_norm": 9.622067683785492, + "learning_rate": 7.353824422450702e-06, + "loss": 0.4385, + "step": 5505 + }, + { + "epoch": 2.98914223669924, + "grad_norm": 9.959533869172905, + "learning_rate": 7.3504331099865474e-06, + "loss": 0.6041, + "step": 5506 + }, + { + "epoch": 2.9896851248642777, + "grad_norm": 12.611515654376284, + "learning_rate": 7.347042125225052e-06, + "loss": 0.5667, + "step": 5507 + }, + { + "epoch": 2.990228013029316, + "grad_norm": 16.346151732068346, + "learning_rate": 7.343651468585611e-06, + "loss": 0.8017, + "step": 5508 + }, + { + "epoch": 2.990770901194354, + "grad_norm": 14.039933672423732, + "learning_rate": 7.340261140487593e-06, + "loss": 0.6905, + "step": 5509 + }, + { + "epoch": 2.991313789359392, + "grad_norm": 9.68037751098424, + "learning_rate": 7.336871141350313e-06, + "loss": 0.7553, + "step": 5510 + }, + { + "epoch": 2.9918566775244297, + "grad_norm": 11.6044953546391, + "learning_rate": 7.333481471593058e-06, + "loss": 0.7525, + "step": 5511 + }, + { + "epoch": 2.992399565689468, + "grad_norm": 10.257246759089158, + "learning_rate": 7.330092131635061e-06, + "loss": 0.5479, + "step": 5512 + }, + { + "epoch": 2.992942453854506, + "grad_norm": 9.37947259585779, + "learning_rate": 7.326703121895528e-06, + "loss": 0.6117, + "step": 5513 + }, + { + "epoch": 2.993485342019544, + "grad_norm": 10.344183596900065, + "learning_rate": 7.323314442793614e-06, + "loss": 0.5074, + "step": 5514 + }, + { + "epoch": 2.9940282301845818, + "grad_norm": 8.940259405187144, + "learning_rate": 7.319926094748433e-06, + "loss": 0.2863, + "step": 5515 + }, + { + "epoch": 2.99457111834962, + "grad_norm": 10.920924886109798, + "learning_rate": 7.316538078179065e-06, + "loss": 0.7706, + "step": 5516 + }, + { + "epoch": 2.995114006514658, + "grad_norm": 4.8493484123114445, + "learning_rate": 7.3131503935045424e-06, + "loss": 0.2066, + "step": 5517 + }, + { + "epoch": 2.995656894679696, + "grad_norm": 11.385458894978902, + "learning_rate": 7.309763041143865e-06, + "loss": 0.5706, + "step": 5518 + }, + { + "epoch": 2.9961997828447338, + "grad_norm": 8.263697834815542, + "learning_rate": 7.306376021515977e-06, + "loss": 0.667, + "step": 5519 + }, + { + "epoch": 2.996742671009772, + "grad_norm": 11.30709409496411, + "learning_rate": 7.302989335039801e-06, + "loss": 0.5522, + "step": 5520 + }, + { + "epoch": 2.99728555917481, + "grad_norm": 9.914253293704078, + "learning_rate": 7.299602982134199e-06, + "loss": 0.6109, + "step": 5521 + }, + { + "epoch": 2.997828447339848, + "grad_norm": 15.471798238376472, + "learning_rate": 7.2962169632180055e-06, + "loss": 0.8321, + "step": 5522 + }, + { + "epoch": 2.9983713355048858, + "grad_norm": 10.956384342902831, + "learning_rate": 7.292831278710007e-06, + "loss": 0.5791, + "step": 5523 + }, + { + "epoch": 2.998914223669924, + "grad_norm": 8.8527583386149, + "learning_rate": 7.289445929028951e-06, + "loss": 0.5807, + "step": 5524 + }, + { + "epoch": 2.999457111834962, + "grad_norm": 10.520949488289421, + "learning_rate": 7.286060914593544e-06, + "loss": 0.6861, + "step": 5525 + }, + { + "epoch": 3.0, + "grad_norm": 11.327416720286067, + "learning_rate": 7.282676235822444e-06, + "loss": 0.4657, + "step": 5526 + }, + { + "epoch": 3.000542888165038, + "grad_norm": 8.97355367199935, + "learning_rate": 7.279291893134283e-06, + "loss": 0.2931, + "step": 5527 + }, + { + "epoch": 3.001085776330076, + "grad_norm": 8.169786294777076, + "learning_rate": 7.275907886947632e-06, + "loss": 0.3065, + "step": 5528 + }, + { + "epoch": 3.001628664495114, + "grad_norm": 9.498228913363004, + "learning_rate": 7.272524217681036e-06, + "loss": 0.3704, + "step": 5529 + }, + { + "epoch": 3.002171552660152, + "grad_norm": 11.241888839923089, + "learning_rate": 7.269140885752992e-06, + "loss": 0.5917, + "step": 5530 + }, + { + "epoch": 3.00271444082519, + "grad_norm": 8.725213535150235, + "learning_rate": 7.26575789158196e-06, + "loss": 0.4699, + "step": 5531 + }, + { + "epoch": 3.003257328990228, + "grad_norm": 7.545852057856303, + "learning_rate": 7.262375235586343e-06, + "loss": 0.5126, + "step": 5532 + }, + { + "epoch": 3.003800217155266, + "grad_norm": 11.581979137257708, + "learning_rate": 7.2589929181845255e-06, + "loss": 0.4424, + "step": 5533 + }, + { + "epoch": 3.004343105320304, + "grad_norm": 9.70311467595886, + "learning_rate": 7.255610939794831e-06, + "loss": 0.4074, + "step": 5534 + }, + { + "epoch": 3.004885993485342, + "grad_norm": 7.016594112536396, + "learning_rate": 7.252229300835545e-06, + "loss": 0.2748, + "step": 5535 + }, + { + "epoch": 3.00542888165038, + "grad_norm": 11.131101403236986, + "learning_rate": 7.248848001724924e-06, + "loss": 0.4696, + "step": 5536 + }, + { + "epoch": 3.005971769815418, + "grad_norm": 9.402705826901306, + "learning_rate": 7.2454670428811625e-06, + "loss": 0.4554, + "step": 5537 + }, + { + "epoch": 3.006514657980456, + "grad_norm": 10.04584435406275, + "learning_rate": 7.242086424722432e-06, + "loss": 0.5017, + "step": 5538 + }, + { + "epoch": 3.007057546145494, + "grad_norm": 9.814079675077302, + "learning_rate": 7.238706147666843e-06, + "loss": 0.3366, + "step": 5539 + }, + { + "epoch": 3.007600434310532, + "grad_norm": 8.970687114442066, + "learning_rate": 7.235326212132483e-06, + "loss": 0.5007, + "step": 5540 + }, + { + "epoch": 3.00814332247557, + "grad_norm": 10.57886281786491, + "learning_rate": 7.23194661853738e-06, + "loss": 0.3796, + "step": 5541 + }, + { + "epoch": 3.008686210640608, + "grad_norm": 10.781975432954557, + "learning_rate": 7.228567367299532e-06, + "loss": 0.5567, + "step": 5542 + }, + { + "epoch": 3.009229098805646, + "grad_norm": 11.394439293433617, + "learning_rate": 7.2251884588368916e-06, + "loss": 0.521, + "step": 5543 + }, + { + "epoch": 3.009771986970684, + "grad_norm": 10.408091638796, + "learning_rate": 7.22180989356736e-06, + "loss": 0.4805, + "step": 5544 + }, + { + "epoch": 3.010314875135722, + "grad_norm": 11.389146042667898, + "learning_rate": 7.218431671908812e-06, + "loss": 0.49, + "step": 5545 + }, + { + "epoch": 3.01085776330076, + "grad_norm": 12.57922467984807, + "learning_rate": 7.215053794279066e-06, + "loss": 0.5101, + "step": 5546 + }, + { + "epoch": 3.011400651465798, + "grad_norm": 14.591941806963382, + "learning_rate": 7.211676261095905e-06, + "loss": 0.6942, + "step": 5547 + }, + { + "epoch": 3.011943539630836, + "grad_norm": 12.696865069952242, + "learning_rate": 7.20829907277707e-06, + "loss": 0.6091, + "step": 5548 + }, + { + "epoch": 3.012486427795874, + "grad_norm": 8.118652375371086, + "learning_rate": 7.204922229740255e-06, + "loss": 0.346, + "step": 5549 + }, + { + "epoch": 3.013029315960912, + "grad_norm": 10.942362893265422, + "learning_rate": 7.2015457324031105e-06, + "loss": 0.6571, + "step": 5550 + }, + { + "epoch": 3.01357220412595, + "grad_norm": 10.204677993952515, + "learning_rate": 7.198169581183254e-06, + "loss": 0.5563, + "step": 5551 + }, + { + "epoch": 3.014115092290988, + "grad_norm": 9.634479609863043, + "learning_rate": 7.194793776498245e-06, + "loss": 0.5768, + "step": 5552 + }, + { + "epoch": 3.014657980456026, + "grad_norm": 9.591830793887732, + "learning_rate": 7.191418318765616e-06, + "loss": 0.3283, + "step": 5553 + }, + { + "epoch": 3.015200868621064, + "grad_norm": 13.201493691800454, + "learning_rate": 7.188043208402844e-06, + "loss": 0.6985, + "step": 5554 + }, + { + "epoch": 3.015743756786102, + "grad_norm": 8.863552231556248, + "learning_rate": 7.184668445827367e-06, + "loss": 0.569, + "step": 5555 + }, + { + "epoch": 3.01628664495114, + "grad_norm": 16.6519621896966, + "learning_rate": 7.181294031456589e-06, + "loss": 0.5339, + "step": 5556 + }, + { + "epoch": 3.016829533116178, + "grad_norm": 10.16813004001019, + "learning_rate": 7.177919965707853e-06, + "loss": 0.4042, + "step": 5557 + }, + { + "epoch": 3.017372421281216, + "grad_norm": 9.578956705487652, + "learning_rate": 7.174546248998477e-06, + "loss": 0.454, + "step": 5558 + }, + { + "epoch": 3.017915309446254, + "grad_norm": 12.884511869394336, + "learning_rate": 7.1711728817457204e-06, + "loss": 0.769, + "step": 5559 + }, + { + "epoch": 3.018458197611292, + "grad_norm": 9.438716836740053, + "learning_rate": 7.1677998643668124e-06, + "loss": 0.3148, + "step": 5560 + }, + { + "epoch": 3.01900108577633, + "grad_norm": 10.176350005774413, + "learning_rate": 7.164427197278931e-06, + "loss": 0.4634, + "step": 5561 + }, + { + "epoch": 3.019543973941368, + "grad_norm": 13.390457057860667, + "learning_rate": 7.161054880899212e-06, + "loss": 0.8095, + "step": 5562 + }, + { + "epoch": 3.020086862106406, + "grad_norm": 15.075886108978443, + "learning_rate": 7.157682915644754e-06, + "loss": 0.4937, + "step": 5563 + }, + { + "epoch": 3.020629750271444, + "grad_norm": 10.999178702680004, + "learning_rate": 7.154311301932597e-06, + "loss": 0.6033, + "step": 5564 + }, + { + "epoch": 3.021172638436482, + "grad_norm": 11.230531078526122, + "learning_rate": 7.150940040179759e-06, + "loss": 0.4776, + "step": 5565 + }, + { + "epoch": 3.02171552660152, + "grad_norm": 10.476946027170913, + "learning_rate": 7.147569130803193e-06, + "loss": 0.3336, + "step": 5566 + }, + { + "epoch": 3.022258414766558, + "grad_norm": 11.753790547662213, + "learning_rate": 7.144198574219824e-06, + "loss": 0.5004, + "step": 5567 + }, + { + "epoch": 3.022801302931596, + "grad_norm": 8.64883980987418, + "learning_rate": 7.140828370846525e-06, + "loss": 0.3042, + "step": 5568 + }, + { + "epoch": 3.023344191096634, + "grad_norm": 13.493035832393069, + "learning_rate": 7.137458521100136e-06, + "loss": 0.5849, + "step": 5569 + }, + { + "epoch": 3.023887079261672, + "grad_norm": 13.029908881843049, + "learning_rate": 7.1340890253974324e-06, + "loss": 0.583, + "step": 5570 + }, + { + "epoch": 3.02442996742671, + "grad_norm": 11.852047436985602, + "learning_rate": 7.130719884155173e-06, + "loss": 0.4575, + "step": 5571 + }, + { + "epoch": 3.024972855591748, + "grad_norm": 13.476829762238244, + "learning_rate": 7.127351097790045e-06, + "loss": 0.5722, + "step": 5572 + }, + { + "epoch": 3.025515743756786, + "grad_norm": 11.710899279093313, + "learning_rate": 7.123982666718716e-06, + "loss": 0.5369, + "step": 5573 + }, + { + "epoch": 3.026058631921824, + "grad_norm": 12.982691203975197, + "learning_rate": 7.120614591357793e-06, + "loss": 0.4587, + "step": 5574 + }, + { + "epoch": 3.026601520086862, + "grad_norm": 13.494356693487626, + "learning_rate": 7.117246872123843e-06, + "loss": 0.5962, + "step": 5575 + }, + { + "epoch": 3.0271444082519, + "grad_norm": 9.838797633608028, + "learning_rate": 7.113879509433399e-06, + "loss": 0.3549, + "step": 5576 + }, + { + "epoch": 3.027687296416938, + "grad_norm": 15.816616895711462, + "learning_rate": 7.110512503702933e-06, + "loss": 0.6669, + "step": 5577 + }, + { + "epoch": 3.028230184581976, + "grad_norm": 11.513602729284349, + "learning_rate": 7.1071458553488904e-06, + "loss": 0.3573, + "step": 5578 + }, + { + "epoch": 3.028773072747014, + "grad_norm": 14.040376133978413, + "learning_rate": 7.103779564787654e-06, + "loss": 0.5898, + "step": 5579 + }, + { + "epoch": 3.029315960912052, + "grad_norm": 12.839234013609035, + "learning_rate": 7.10041363243558e-06, + "loss": 0.525, + "step": 5580 + }, + { + "epoch": 3.02985884907709, + "grad_norm": 12.773102345996788, + "learning_rate": 7.097048058708966e-06, + "loss": 0.7338, + "step": 5581 + }, + { + "epoch": 3.030401737242128, + "grad_norm": 9.757116728754855, + "learning_rate": 7.093682844024079e-06, + "loss": 0.4107, + "step": 5582 + }, + { + "epoch": 3.030944625407166, + "grad_norm": 13.284417042689952, + "learning_rate": 7.090317988797131e-06, + "loss": 0.4909, + "step": 5583 + }, + { + "epoch": 3.031487513572204, + "grad_norm": 12.858923373950708, + "learning_rate": 7.086953493444286e-06, + "loss": 0.4164, + "step": 5584 + }, + { + "epoch": 3.032030401737242, + "grad_norm": 11.460500988764892, + "learning_rate": 7.083589358381681e-06, + "loss": 0.6339, + "step": 5585 + }, + { + "epoch": 3.03257328990228, + "grad_norm": 11.76507179913234, + "learning_rate": 7.0802255840253865e-06, + "loss": 0.33, + "step": 5586 + }, + { + "epoch": 3.033116178067318, + "grad_norm": 14.990076792803963, + "learning_rate": 7.076862170791449e-06, + "loss": 0.5647, + "step": 5587 + }, + { + "epoch": 3.033659066232356, + "grad_norm": 12.953159557836509, + "learning_rate": 7.0734991190958545e-06, + "loss": 0.5641, + "step": 5588 + }, + { + "epoch": 3.034201954397394, + "grad_norm": 9.746562255040114, + "learning_rate": 7.070136429354557e-06, + "loss": 0.5755, + "step": 5589 + }, + { + "epoch": 3.034744842562432, + "grad_norm": 14.867296485205143, + "learning_rate": 7.066774101983452e-06, + "loss": 0.4915, + "step": 5590 + }, + { + "epoch": 3.03528773072747, + "grad_norm": 12.725313171891754, + "learning_rate": 7.063412137398402e-06, + "loss": 0.541, + "step": 5591 + }, + { + "epoch": 3.035830618892508, + "grad_norm": 11.023036821532774, + "learning_rate": 7.060050536015219e-06, + "loss": 0.3524, + "step": 5592 + }, + { + "epoch": 3.036373507057546, + "grad_norm": 9.747447295285967, + "learning_rate": 7.05668929824967e-06, + "loss": 0.4588, + "step": 5593 + }, + { + "epoch": 3.036916395222584, + "grad_norm": 12.477427859321583, + "learning_rate": 7.0533284245174826e-06, + "loss": 0.382, + "step": 5594 + }, + { + "epoch": 3.037459283387622, + "grad_norm": 9.62366763242101, + "learning_rate": 7.049967915234329e-06, + "loss": 0.2984, + "step": 5595 + }, + { + "epoch": 3.03800217155266, + "grad_norm": 13.368913470296512, + "learning_rate": 7.046607770815849e-06, + "loss": 0.6245, + "step": 5596 + }, + { + "epoch": 3.038545059717698, + "grad_norm": 10.380221393303437, + "learning_rate": 7.043247991677622e-06, + "loss": 0.4542, + "step": 5597 + }, + { + "epoch": 3.039087947882736, + "grad_norm": 9.778914279269939, + "learning_rate": 7.039888578235201e-06, + "loss": 0.3281, + "step": 5598 + }, + { + "epoch": 3.039630836047774, + "grad_norm": 7.360801727275878, + "learning_rate": 7.036529530904073e-06, + "loss": 0.2668, + "step": 5599 + }, + { + "epoch": 3.040173724212812, + "grad_norm": 7.962043569210736, + "learning_rate": 7.033170850099699e-06, + "loss": 0.2437, + "step": 5600 + }, + { + "epoch": 3.04071661237785, + "grad_norm": 13.434858118835157, + "learning_rate": 7.029812536237479e-06, + "loss": 0.5657, + "step": 5601 + }, + { + "epoch": 3.041259500542888, + "grad_norm": 12.720111358562143, + "learning_rate": 7.026454589732784e-06, + "loss": 0.5799, + "step": 5602 + }, + { + "epoch": 3.041802388707926, + "grad_norm": 13.271222462410066, + "learning_rate": 7.023097011000923e-06, + "loss": 0.4124, + "step": 5603 + }, + { + "epoch": 3.042345276872964, + "grad_norm": 12.713200398376557, + "learning_rate": 7.019739800457166e-06, + "loss": 0.5288, + "step": 5604 + }, + { + "epoch": 3.042888165038002, + "grad_norm": 8.938306532181711, + "learning_rate": 7.016382958516741e-06, + "loss": 0.3436, + "step": 5605 + }, + { + "epoch": 3.04343105320304, + "grad_norm": 11.284866978643858, + "learning_rate": 7.0130264855948245e-06, + "loss": 0.3379, + "step": 5606 + }, + { + "epoch": 3.043973941368078, + "grad_norm": 13.192559496466789, + "learning_rate": 7.009670382106558e-06, + "loss": 0.4862, + "step": 5607 + }, + { + "epoch": 3.044516829533116, + "grad_norm": 11.479680729723038, + "learning_rate": 7.006314648467021e-06, + "loss": 0.3483, + "step": 5608 + }, + { + "epoch": 3.045059717698154, + "grad_norm": 12.417526118207578, + "learning_rate": 7.002959285091262e-06, + "loss": 0.4803, + "step": 5609 + }, + { + "epoch": 3.045602605863192, + "grad_norm": 10.73078215472077, + "learning_rate": 6.999604292394271e-06, + "loss": 0.391, + "step": 5610 + }, + { + "epoch": 3.04614549402823, + "grad_norm": 11.713115986580751, + "learning_rate": 6.9962496707910085e-06, + "loss": 0.4674, + "step": 5611 + }, + { + "epoch": 3.046688382193268, + "grad_norm": 11.477819508258428, + "learning_rate": 6.992895420696371e-06, + "loss": 0.4974, + "step": 5612 + }, + { + "epoch": 3.047231270358306, + "grad_norm": 13.625151931379403, + "learning_rate": 6.989541542525218e-06, + "loss": 0.7687, + "step": 5613 + }, + { + "epoch": 3.047774158523344, + "grad_norm": 11.52719950465663, + "learning_rate": 6.986188036692369e-06, + "loss": 0.5544, + "step": 5614 + }, + { + "epoch": 3.048317046688382, + "grad_norm": 13.283507119767767, + "learning_rate": 6.982834903612584e-06, + "loss": 0.6122, + "step": 5615 + }, + { + "epoch": 3.04885993485342, + "grad_norm": 13.11709341600981, + "learning_rate": 6.979482143700591e-06, + "loss": 0.5184, + "step": 5616 + }, + { + "epoch": 3.049402823018458, + "grad_norm": 14.076144429789945, + "learning_rate": 6.976129757371055e-06, + "loss": 0.5594, + "step": 5617 + }, + { + "epoch": 3.049945711183496, + "grad_norm": 13.73222380193425, + "learning_rate": 6.9727777450386124e-06, + "loss": 0.4793, + "step": 5618 + }, + { + "epoch": 3.050488599348534, + "grad_norm": 10.519849404041729, + "learning_rate": 6.969426107117845e-06, + "loss": 0.3503, + "step": 5619 + }, + { + "epoch": 3.0510314875135722, + "grad_norm": 11.598423103761737, + "learning_rate": 6.966074844023284e-06, + "loss": 0.8233, + "step": 5620 + }, + { + "epoch": 3.05157437567861, + "grad_norm": 10.260836656524097, + "learning_rate": 6.9627239561694205e-06, + "loss": 0.4298, + "step": 5621 + }, + { + "epoch": 3.0521172638436482, + "grad_norm": 12.202759177421122, + "learning_rate": 6.9593734439707024e-06, + "loss": 0.5644, + "step": 5622 + }, + { + "epoch": 3.052660152008686, + "grad_norm": 21.73953611396027, + "learning_rate": 6.956023307841524e-06, + "loss": 0.9458, + "step": 5623 + }, + { + "epoch": 3.0532030401737242, + "grad_norm": 12.609994641445128, + "learning_rate": 6.952673548196229e-06, + "loss": 0.5591, + "step": 5624 + }, + { + "epoch": 3.053745928338762, + "grad_norm": 9.019788767418095, + "learning_rate": 6.949324165449129e-06, + "loss": 0.2823, + "step": 5625 + }, + { + "epoch": 3.0542888165038002, + "grad_norm": 9.597041690728307, + "learning_rate": 6.945975160014476e-06, + "loss": 0.339, + "step": 5626 + }, + { + "epoch": 3.054831704668838, + "grad_norm": 9.413655638291356, + "learning_rate": 6.9426265323064865e-06, + "loss": 0.3789, + "step": 5627 + }, + { + "epoch": 3.0553745928338762, + "grad_norm": 11.15027104482265, + "learning_rate": 6.939278282739317e-06, + "loss": 0.5436, + "step": 5628 + }, + { + "epoch": 3.055917480998914, + "grad_norm": 8.752949901728641, + "learning_rate": 6.935930411727091e-06, + "loss": 0.3561, + "step": 5629 + }, + { + "epoch": 3.0564603691639523, + "grad_norm": 13.044879197154156, + "learning_rate": 6.93258291968387e-06, + "loss": 0.5744, + "step": 5630 + }, + { + "epoch": 3.05700325732899, + "grad_norm": 8.940911398612299, + "learning_rate": 6.929235807023689e-06, + "loss": 0.3882, + "step": 5631 + }, + { + "epoch": 3.0575461454940283, + "grad_norm": 11.140820597123383, + "learning_rate": 6.925889074160512e-06, + "loss": 0.4042, + "step": 5632 + }, + { + "epoch": 3.058089033659066, + "grad_norm": 11.477795419029349, + "learning_rate": 6.9225427215082745e-06, + "loss": 0.531, + "step": 5633 + }, + { + "epoch": 3.0586319218241043, + "grad_norm": 11.606939647718864, + "learning_rate": 6.919196749480861e-06, + "loss": 0.3965, + "step": 5634 + }, + { + "epoch": 3.059174809989142, + "grad_norm": 12.112684503158684, + "learning_rate": 6.915851158492099e-06, + "loss": 0.3518, + "step": 5635 + }, + { + "epoch": 3.0597176981541803, + "grad_norm": 11.996768432451738, + "learning_rate": 6.9125059489557855e-06, + "loss": 0.4507, + "step": 5636 + }, + { + "epoch": 3.060260586319218, + "grad_norm": 11.43143134688868, + "learning_rate": 6.909161121285653e-06, + "loss": 0.3375, + "step": 5637 + }, + { + "epoch": 3.0608034744842563, + "grad_norm": 11.944059447932277, + "learning_rate": 6.9058166758954015e-06, + "loss": 0.5155, + "step": 5638 + }, + { + "epoch": 3.061346362649294, + "grad_norm": 13.048822571970177, + "learning_rate": 6.902472613198672e-06, + "loss": 0.4834, + "step": 5639 + }, + { + "epoch": 3.0618892508143323, + "grad_norm": 12.031888695640616, + "learning_rate": 6.899128933609071e-06, + "loss": 0.5558, + "step": 5640 + }, + { + "epoch": 3.06243213897937, + "grad_norm": 12.352798864366527, + "learning_rate": 6.8957856375401466e-06, + "loss": 0.6299, + "step": 5641 + }, + { + "epoch": 3.0629750271444083, + "grad_norm": 8.205311159067556, + "learning_rate": 6.892442725405396e-06, + "loss": 0.2726, + "step": 5642 + }, + { + "epoch": 3.063517915309446, + "grad_norm": 10.36891977803926, + "learning_rate": 6.8891001976182845e-06, + "loss": 0.3787, + "step": 5643 + }, + { + "epoch": 3.0640608034744843, + "grad_norm": 13.053698798672075, + "learning_rate": 6.88575805459222e-06, + "loss": 0.8244, + "step": 5644 + }, + { + "epoch": 3.064603691639522, + "grad_norm": 9.553838049974857, + "learning_rate": 6.88241629674056e-06, + "loss": 0.3611, + "step": 5645 + }, + { + "epoch": 3.0651465798045603, + "grad_norm": 13.221077748060404, + "learning_rate": 6.879074924476621e-06, + "loss": 0.3718, + "step": 5646 + }, + { + "epoch": 3.065689467969598, + "grad_norm": 13.072535493727939, + "learning_rate": 6.8757339382136735e-06, + "loss": 0.4338, + "step": 5647 + }, + { + "epoch": 3.0662323561346363, + "grad_norm": 13.8353344886429, + "learning_rate": 6.872393338364927e-06, + "loss": 0.4158, + "step": 5648 + }, + { + "epoch": 3.066775244299674, + "grad_norm": 12.713981935614036, + "learning_rate": 6.869053125343561e-06, + "loss": 0.4537, + "step": 5649 + }, + { + "epoch": 3.0673181324647123, + "grad_norm": 9.412074844790709, + "learning_rate": 6.865713299562691e-06, + "loss": 0.3842, + "step": 5650 + }, + { + "epoch": 3.06786102062975, + "grad_norm": 12.671904220941121, + "learning_rate": 6.862373861435397e-06, + "loss": 0.5343, + "step": 5651 + }, + { + "epoch": 3.0684039087947883, + "grad_norm": 8.312202499711958, + "learning_rate": 6.859034811374707e-06, + "loss": 0.4265, + "step": 5652 + }, + { + "epoch": 3.068946796959826, + "grad_norm": 8.89496055231579, + "learning_rate": 6.8556961497935925e-06, + "loss": 0.2747, + "step": 5653 + }, + { + "epoch": 3.0694896851248643, + "grad_norm": 12.131106832543901, + "learning_rate": 6.852357877104995e-06, + "loss": 0.4964, + "step": 5654 + }, + { + "epoch": 3.070032573289902, + "grad_norm": 11.860190856186797, + "learning_rate": 6.849019993721786e-06, + "loss": 0.6045, + "step": 5655 + }, + { + "epoch": 3.0705754614549403, + "grad_norm": 9.678033292318382, + "learning_rate": 6.845682500056811e-06, + "loss": 0.331, + "step": 5656 + }, + { + "epoch": 3.071118349619978, + "grad_norm": 15.602537064720222, + "learning_rate": 6.8423453965228485e-06, + "loss": 0.586, + "step": 5657 + }, + { + "epoch": 3.0716612377850163, + "grad_norm": 14.701485694410227, + "learning_rate": 6.839008683532641e-06, + "loss": 0.4876, + "step": 5658 + }, + { + "epoch": 3.072204125950054, + "grad_norm": 10.535230213477728, + "learning_rate": 6.835672361498875e-06, + "loss": 0.6123, + "step": 5659 + }, + { + "epoch": 3.0727470141150923, + "grad_norm": 11.131327768224557, + "learning_rate": 6.832336430834199e-06, + "loss": 0.6257, + "step": 5660 + }, + { + "epoch": 3.07328990228013, + "grad_norm": 8.624075967779987, + "learning_rate": 6.829000891951202e-06, + "loss": 0.3361, + "step": 5661 + }, + { + "epoch": 3.0738327904451683, + "grad_norm": 13.128895205370725, + "learning_rate": 6.825665745262424e-06, + "loss": 0.5965, + "step": 5662 + }, + { + "epoch": 3.074375678610206, + "grad_norm": 12.64008926373084, + "learning_rate": 6.822330991180368e-06, + "loss": 0.62, + "step": 5663 + }, + { + "epoch": 3.0749185667752443, + "grad_norm": 13.416120809361393, + "learning_rate": 6.8189966301174785e-06, + "loss": 0.6674, + "step": 5664 + }, + { + "epoch": 3.075461454940282, + "grad_norm": 9.203438718678973, + "learning_rate": 6.815662662486158e-06, + "loss": 0.3816, + "step": 5665 + }, + { + "epoch": 3.0760043431053203, + "grad_norm": 13.276570487788803, + "learning_rate": 6.81232908869875e-06, + "loss": 0.8179, + "step": 5666 + }, + { + "epoch": 3.076547231270358, + "grad_norm": 12.280157213200626, + "learning_rate": 6.808995909167566e-06, + "loss": 0.5559, + "step": 5667 + }, + { + "epoch": 3.0770901194353963, + "grad_norm": 10.175753421889684, + "learning_rate": 6.805663124304848e-06, + "loss": 0.6191, + "step": 5668 + }, + { + "epoch": 3.077633007600434, + "grad_norm": 12.523871003768551, + "learning_rate": 6.802330734522813e-06, + "loss": 0.4992, + "step": 5669 + }, + { + "epoch": 3.0781758957654723, + "grad_norm": 9.976975930711898, + "learning_rate": 6.798998740233602e-06, + "loss": 0.4104, + "step": 5670 + }, + { + "epoch": 3.07871878393051, + "grad_norm": 11.57486842983445, + "learning_rate": 6.795667141849333e-06, + "loss": 0.5803, + "step": 5671 + }, + { + "epoch": 3.0792616720955484, + "grad_norm": 13.330648489276534, + "learning_rate": 6.79233593978206e-06, + "loss": 0.5116, + "step": 5672 + }, + { + "epoch": 3.079804560260586, + "grad_norm": 10.454358932554449, + "learning_rate": 6.789005134443785e-06, + "loss": 0.4015, + "step": 5673 + }, + { + "epoch": 3.0803474484256244, + "grad_norm": 10.39956810217567, + "learning_rate": 6.785674726246477e-06, + "loss": 0.5832, + "step": 5674 + }, + { + "epoch": 3.080890336590662, + "grad_norm": 10.388603711886958, + "learning_rate": 6.782344715602038e-06, + "loss": 0.2995, + "step": 5675 + }, + { + "epoch": 3.0814332247557004, + "grad_norm": 14.02200163557337, + "learning_rate": 6.779015102922335e-06, + "loss": 0.5132, + "step": 5676 + }, + { + "epoch": 3.081976112920738, + "grad_norm": 9.590771798908678, + "learning_rate": 6.775685888619174e-06, + "loss": 0.2876, + "step": 5677 + }, + { + "epoch": 3.0825190010857764, + "grad_norm": 15.505233514983502, + "learning_rate": 6.7723570731043275e-06, + "loss": 0.8031, + "step": 5678 + }, + { + "epoch": 3.083061889250814, + "grad_norm": 11.180083592225891, + "learning_rate": 6.769028656789497e-06, + "loss": 0.5965, + "step": 5679 + }, + { + "epoch": 3.0836047774158524, + "grad_norm": 12.733678586969875, + "learning_rate": 6.765700640086356e-06, + "loss": 0.4956, + "step": 5680 + }, + { + "epoch": 3.08414766558089, + "grad_norm": 6.888856085950901, + "learning_rate": 6.762373023406515e-06, + "loss": 0.2367, + "step": 5681 + }, + { + "epoch": 3.0846905537459284, + "grad_norm": 12.334377562502302, + "learning_rate": 6.759045807161532e-06, + "loss": 0.5297, + "step": 5682 + }, + { + "epoch": 3.085233441910966, + "grad_norm": 10.637487438782616, + "learning_rate": 6.755718991762934e-06, + "loss": 0.4714, + "step": 5683 + }, + { + "epoch": 3.0857763300760044, + "grad_norm": 10.636835957652801, + "learning_rate": 6.7523925776221775e-06, + "loss": 0.5023, + "step": 5684 + }, + { + "epoch": 3.086319218241042, + "grad_norm": 9.416617539617917, + "learning_rate": 6.749066565150688e-06, + "loss": 0.4012, + "step": 5685 + }, + { + "epoch": 3.0868621064060804, + "grad_norm": 12.384873763197831, + "learning_rate": 6.745740954759823e-06, + "loss": 0.4588, + "step": 5686 + }, + { + "epoch": 3.087404994571118, + "grad_norm": 16.770871731703775, + "learning_rate": 6.7424157468609075e-06, + "loss": 0.8317, + "step": 5687 + }, + { + "epoch": 3.0879478827361564, + "grad_norm": 8.972498742892359, + "learning_rate": 6.7390909418652e-06, + "loss": 0.376, + "step": 5688 + }, + { + "epoch": 3.088490770901194, + "grad_norm": 12.96316859851228, + "learning_rate": 6.735766540183923e-06, + "loss": 0.5285, + "step": 5689 + }, + { + "epoch": 3.0890336590662324, + "grad_norm": 15.891256870281978, + "learning_rate": 6.732442542228245e-06, + "loss": 0.6191, + "step": 5690 + }, + { + "epoch": 3.08957654723127, + "grad_norm": 17.35289550101855, + "learning_rate": 6.729118948409278e-06, + "loss": 0.6878, + "step": 5691 + }, + { + "epoch": 3.0901194353963084, + "grad_norm": 9.022024416599917, + "learning_rate": 6.7257957591380965e-06, + "loss": 0.417, + "step": 5692 + }, + { + "epoch": 3.090662323561346, + "grad_norm": 14.27557655416772, + "learning_rate": 6.722472974825709e-06, + "loss": 0.5535, + "step": 5693 + }, + { + "epoch": 3.0912052117263844, + "grad_norm": 13.610102273496864, + "learning_rate": 6.7191505958830916e-06, + "loss": 0.542, + "step": 5694 + }, + { + "epoch": 3.091748099891422, + "grad_norm": 9.424378370747705, + "learning_rate": 6.715828622721154e-06, + "loss": 0.3447, + "step": 5695 + }, + { + "epoch": 3.0922909880564604, + "grad_norm": 10.815334224243674, + "learning_rate": 6.712507055750768e-06, + "loss": 0.5058, + "step": 5696 + }, + { + "epoch": 3.092833876221498, + "grad_norm": 9.390605474272599, + "learning_rate": 6.709185895382746e-06, + "loss": 0.3778, + "step": 5697 + }, + { + "epoch": 3.0933767643865364, + "grad_norm": 9.621100751873954, + "learning_rate": 6.705865142027863e-06, + "loss": 0.5128, + "step": 5698 + }, + { + "epoch": 3.093919652551574, + "grad_norm": 10.148509981020817, + "learning_rate": 6.7025447960968236e-06, + "loss": 0.3857, + "step": 5699 + }, + { + "epoch": 3.0944625407166124, + "grad_norm": 12.950847437294314, + "learning_rate": 6.699224858000305e-06, + "loss": 0.5364, + "step": 5700 + }, + { + "epoch": 3.09500542888165, + "grad_norm": 11.169096513096612, + "learning_rate": 6.695905328148914e-06, + "loss": 0.7072, + "step": 5701 + }, + { + "epoch": 3.0955483170466884, + "grad_norm": 9.231874354349005, + "learning_rate": 6.692586206953218e-06, + "loss": 0.5096, + "step": 5702 + }, + { + "epoch": 3.096091205211726, + "grad_norm": 13.629340417763906, + "learning_rate": 6.68926749482373e-06, + "loss": 0.5454, + "step": 5703 + }, + { + "epoch": 3.0966340933767644, + "grad_norm": 12.18004127970413, + "learning_rate": 6.685949192170913e-06, + "loss": 0.4463, + "step": 5704 + }, + { + "epoch": 3.097176981541802, + "grad_norm": 10.488473413353695, + "learning_rate": 6.682631299405187e-06, + "loss": 0.4041, + "step": 5705 + }, + { + "epoch": 3.0977198697068404, + "grad_norm": 9.624411116526867, + "learning_rate": 6.679313816936904e-06, + "loss": 0.5316, + "step": 5706 + }, + { + "epoch": 3.098262757871878, + "grad_norm": 13.907233176818153, + "learning_rate": 6.675996745176385e-06, + "loss": 0.6742, + "step": 5707 + }, + { + "epoch": 3.0988056460369164, + "grad_norm": 9.086606489202428, + "learning_rate": 6.672680084533882e-06, + "loss": 0.4101, + "step": 5708 + }, + { + "epoch": 3.099348534201954, + "grad_norm": 9.283349632521654, + "learning_rate": 6.669363835419611e-06, + "loss": 0.318, + "step": 5709 + }, + { + "epoch": 3.0998914223669924, + "grad_norm": 9.668872732852824, + "learning_rate": 6.666047998243732e-06, + "loss": 0.3352, + "step": 5710 + }, + { + "epoch": 3.1004343105320302, + "grad_norm": 9.343099296154486, + "learning_rate": 6.662732573416345e-06, + "loss": 0.4697, + "step": 5711 + }, + { + "epoch": 3.1009771986970684, + "grad_norm": 11.924198282652524, + "learning_rate": 6.659417561347516e-06, + "loss": 0.5743, + "step": 5712 + }, + { + "epoch": 3.1015200868621062, + "grad_norm": 8.560855249443614, + "learning_rate": 6.656102962447244e-06, + "loss": 0.3406, + "step": 5713 + }, + { + "epoch": 3.1020629750271445, + "grad_norm": 14.378777363618713, + "learning_rate": 6.652788777125488e-06, + "loss": 0.6911, + "step": 5714 + }, + { + "epoch": 3.1026058631921822, + "grad_norm": 10.29068899101719, + "learning_rate": 6.649475005792152e-06, + "loss": 0.4449, + "step": 5715 + }, + { + "epoch": 3.1031487513572205, + "grad_norm": 14.02995967568585, + "learning_rate": 6.646161648857085e-06, + "loss": 0.6167, + "step": 5716 + }, + { + "epoch": 3.1036916395222582, + "grad_norm": 9.518238552951749, + "learning_rate": 6.642848706730092e-06, + "loss": 0.3222, + "step": 5717 + }, + { + "epoch": 3.1042345276872965, + "grad_norm": 11.82162879354163, + "learning_rate": 6.639536179820922e-06, + "loss": 0.541, + "step": 5718 + }, + { + "epoch": 3.1047774158523342, + "grad_norm": 10.063084650592682, + "learning_rate": 6.636224068539271e-06, + "loss": 0.3842, + "step": 5719 + }, + { + "epoch": 3.1053203040173725, + "grad_norm": 10.94706260488087, + "learning_rate": 6.632912373294792e-06, + "loss": 0.4513, + "step": 5720 + }, + { + "epoch": 3.1058631921824102, + "grad_norm": 18.39843990173123, + "learning_rate": 6.6296010944970745e-06, + "loss": 0.6059, + "step": 5721 + }, + { + "epoch": 3.1064060803474485, + "grad_norm": 12.609614695754882, + "learning_rate": 6.6262902325556635e-06, + "loss": 0.5504, + "step": 5722 + }, + { + "epoch": 3.1069489685124863, + "grad_norm": 10.109861615464935, + "learning_rate": 6.6229797878800575e-06, + "loss": 0.4609, + "step": 5723 + }, + { + "epoch": 3.1074918566775245, + "grad_norm": 10.309200757998537, + "learning_rate": 6.61966976087969e-06, + "loss": 0.3823, + "step": 5724 + }, + { + "epoch": 3.1080347448425623, + "grad_norm": 13.21038451142413, + "learning_rate": 6.616360151963956e-06, + "loss": 1.0012, + "step": 5725 + }, + { + "epoch": 3.1085776330076005, + "grad_norm": 10.685970514335464, + "learning_rate": 6.613050961542189e-06, + "loss": 0.4587, + "step": 5726 + }, + { + "epoch": 3.1091205211726383, + "grad_norm": 10.001481339897822, + "learning_rate": 6.60974219002368e-06, + "loss": 0.4954, + "step": 5727 + }, + { + "epoch": 3.1096634093376765, + "grad_norm": 10.414474916866299, + "learning_rate": 6.6064338378176585e-06, + "loss": 0.5087, + "step": 5728 + }, + { + "epoch": 3.1102062975027143, + "grad_norm": 12.456658351321629, + "learning_rate": 6.603125905333307e-06, + "loss": 0.5004, + "step": 5729 + }, + { + "epoch": 3.1107491856677525, + "grad_norm": 11.228566931054628, + "learning_rate": 6.599818392979762e-06, + "loss": 0.6843, + "step": 5730 + }, + { + "epoch": 3.1112920738327903, + "grad_norm": 11.262471951903663, + "learning_rate": 6.596511301166093e-06, + "loss": 0.4458, + "step": 5731 + }, + { + "epoch": 3.1118349619978285, + "grad_norm": 14.452306212538934, + "learning_rate": 6.593204630301333e-06, + "loss": 0.4319, + "step": 5732 + }, + { + "epoch": 3.1123778501628663, + "grad_norm": 11.527781393494147, + "learning_rate": 6.589898380794451e-06, + "loss": 0.6469, + "step": 5733 + }, + { + "epoch": 3.1129207383279045, + "grad_norm": 13.390783514445054, + "learning_rate": 6.586592553054374e-06, + "loss": 0.5311, + "step": 5734 + }, + { + "epoch": 3.1134636264929423, + "grad_norm": 9.033138513383328, + "learning_rate": 6.583287147489968e-06, + "loss": 0.4243, + "step": 5735 + }, + { + "epoch": 3.1140065146579805, + "grad_norm": 8.69133142787827, + "learning_rate": 6.579982164510057e-06, + "loss": 0.3874, + "step": 5736 + }, + { + "epoch": 3.1145494028230183, + "grad_norm": 8.998980230350567, + "learning_rate": 6.576677604523399e-06, + "loss": 0.3375, + "step": 5737 + }, + { + "epoch": 3.1150922909880565, + "grad_norm": 13.718371009849239, + "learning_rate": 6.573373467938715e-06, + "loss": 0.8846, + "step": 5738 + }, + { + "epoch": 3.1156351791530943, + "grad_norm": 9.65489128269965, + "learning_rate": 6.5700697551646595e-06, + "loss": 0.5469, + "step": 5739 + }, + { + "epoch": 3.1161780673181325, + "grad_norm": 12.925637209034505, + "learning_rate": 6.566766466609844e-06, + "loss": 0.4983, + "step": 5740 + }, + { + "epoch": 3.1167209554831703, + "grad_norm": 8.77821937172237, + "learning_rate": 6.563463602682823e-06, + "loss": 0.3698, + "step": 5741 + }, + { + "epoch": 3.1172638436482085, + "grad_norm": 9.642465181954359, + "learning_rate": 6.560161163792098e-06, + "loss": 0.3764, + "step": 5742 + }, + { + "epoch": 3.1178067318132463, + "grad_norm": 13.798683831974408, + "learning_rate": 6.5568591503461266e-06, + "loss": 0.6993, + "step": 5743 + }, + { + "epoch": 3.1183496199782845, + "grad_norm": 9.872788283771353, + "learning_rate": 6.553557562753299e-06, + "loss": 0.3386, + "step": 5744 + }, + { + "epoch": 3.1188925081433223, + "grad_norm": 8.862249503192198, + "learning_rate": 6.550256401421969e-06, + "loss": 0.2179, + "step": 5745 + }, + { + "epoch": 3.1194353963083605, + "grad_norm": 12.08451508737181, + "learning_rate": 6.546955666760421e-06, + "loss": 0.6611, + "step": 5746 + }, + { + "epoch": 3.1199782844733983, + "grad_norm": 12.735515016235034, + "learning_rate": 6.543655359176904e-06, + "loss": 0.497, + "step": 5747 + }, + { + "epoch": 3.1205211726384365, + "grad_norm": 11.84865199124047, + "learning_rate": 6.5403554790795974e-06, + "loss": 0.7862, + "step": 5748 + }, + { + "epoch": 3.1210640608034743, + "grad_norm": 10.714296855189131, + "learning_rate": 6.53705602687664e-06, + "loss": 0.4053, + "step": 5749 + }, + { + "epoch": 3.1216069489685125, + "grad_norm": 13.603289352492634, + "learning_rate": 6.533757002976115e-06, + "loss": 0.8514, + "step": 5750 + }, + { + "epoch": 3.1221498371335503, + "grad_norm": 13.12224567132668, + "learning_rate": 6.530458407786045e-06, + "loss": 0.5521, + "step": 5751 + }, + { + "epoch": 3.1226927252985885, + "grad_norm": 12.175088945770037, + "learning_rate": 6.527160241714412e-06, + "loss": 0.5578, + "step": 5752 + }, + { + "epoch": 3.1232356134636263, + "grad_norm": 8.84737172176379, + "learning_rate": 6.52386250516913e-06, + "loss": 0.628, + "step": 5753 + }, + { + "epoch": 3.1237785016286646, + "grad_norm": 9.436585430499584, + "learning_rate": 6.520565198558076e-06, + "loss": 0.3497, + "step": 5754 + }, + { + "epoch": 3.1243213897937023, + "grad_norm": 19.037012868331093, + "learning_rate": 6.517268322289063e-06, + "loss": 0.7797, + "step": 5755 + }, + { + "epoch": 3.1248642779587406, + "grad_norm": 10.13012645392841, + "learning_rate": 6.513971876769857e-06, + "loss": 0.4892, + "step": 5756 + }, + { + "epoch": 3.1254071661237783, + "grad_norm": 12.402167588058873, + "learning_rate": 6.510675862408162e-06, + "loss": 0.4358, + "step": 5757 + }, + { + "epoch": 3.1259500542888166, + "grad_norm": 13.090215605994585, + "learning_rate": 6.5073802796116415e-06, + "loss": 0.4757, + "step": 5758 + }, + { + "epoch": 3.1264929424538543, + "grad_norm": 12.046620866150537, + "learning_rate": 6.504085128787892e-06, + "loss": 0.6104, + "step": 5759 + }, + { + "epoch": 3.1270358306188926, + "grad_norm": 13.860233354070779, + "learning_rate": 6.500790410344463e-06, + "loss": 0.5933, + "step": 5760 + }, + { + "epoch": 3.1275787187839303, + "grad_norm": 10.049188091595793, + "learning_rate": 6.497496124688858e-06, + "loss": 0.294, + "step": 5761 + }, + { + "epoch": 3.1281216069489686, + "grad_norm": 14.96534379979502, + "learning_rate": 6.494202272228509e-06, + "loss": 0.5983, + "step": 5762 + }, + { + "epoch": 3.1286644951140063, + "grad_norm": 12.09845857427764, + "learning_rate": 6.490908853370816e-06, + "loss": 0.6995, + "step": 5763 + }, + { + "epoch": 3.1292073832790446, + "grad_norm": 11.420498115973434, + "learning_rate": 6.487615868523105e-06, + "loss": 0.5189, + "step": 5764 + }, + { + "epoch": 3.1297502714440824, + "grad_norm": 9.433130521020503, + "learning_rate": 6.4843233180926646e-06, + "loss": 0.5145, + "step": 5765 + }, + { + "epoch": 3.1302931596091206, + "grad_norm": 13.10906497158192, + "learning_rate": 6.481031202486714e-06, + "loss": 0.4662, + "step": 5766 + }, + { + "epoch": 3.1308360477741584, + "grad_norm": 11.61717934785526, + "learning_rate": 6.477739522112436e-06, + "loss": 0.4496, + "step": 5767 + }, + { + "epoch": 3.1313789359391966, + "grad_norm": 15.125919718339924, + "learning_rate": 6.474448277376946e-06, + "loss": 0.6205, + "step": 5768 + }, + { + "epoch": 3.1319218241042344, + "grad_norm": 11.779008588612678, + "learning_rate": 6.471157468687315e-06, + "loss": 0.337, + "step": 5769 + }, + { + "epoch": 3.1324647122692726, + "grad_norm": 11.960250814404343, + "learning_rate": 6.467867096450553e-06, + "loss": 0.6284, + "step": 5770 + }, + { + "epoch": 3.1330076004343104, + "grad_norm": 11.356052401682772, + "learning_rate": 6.464577161073613e-06, + "loss": 0.5146, + "step": 5771 + }, + { + "epoch": 3.1335504885993486, + "grad_norm": 12.23671034812476, + "learning_rate": 6.461287662963407e-06, + "loss": 0.4118, + "step": 5772 + }, + { + "epoch": 3.1340933767643864, + "grad_norm": 9.3897356425462, + "learning_rate": 6.457998602526783e-06, + "loss": 0.3312, + "step": 5773 + }, + { + "epoch": 3.1346362649294246, + "grad_norm": 8.285183279593793, + "learning_rate": 6.454709980170537e-06, + "loss": 0.3857, + "step": 5774 + }, + { + "epoch": 3.1351791530944624, + "grad_norm": 10.70178891805582, + "learning_rate": 6.451421796301407e-06, + "loss": 0.4596, + "step": 5775 + }, + { + "epoch": 3.1357220412595006, + "grad_norm": 9.966009510827126, + "learning_rate": 6.448134051326092e-06, + "loss": 0.5223, + "step": 5776 + }, + { + "epoch": 3.1362649294245384, + "grad_norm": 11.165316341245944, + "learning_rate": 6.444846745651212e-06, + "loss": 0.482, + "step": 5777 + }, + { + "epoch": 3.1368078175895766, + "grad_norm": 10.091328730281877, + "learning_rate": 6.441559879683357e-06, + "loss": 0.4232, + "step": 5778 + }, + { + "epoch": 3.1373507057546144, + "grad_norm": 9.885329689068836, + "learning_rate": 6.438273453829047e-06, + "loss": 0.3008, + "step": 5779 + }, + { + "epoch": 3.1378935939196526, + "grad_norm": 11.500516853477965, + "learning_rate": 6.434987468494749e-06, + "loss": 0.5076, + "step": 5780 + }, + { + "epoch": 3.1384364820846904, + "grad_norm": 11.00413836223355, + "learning_rate": 6.431701924086887e-06, + "loss": 0.3688, + "step": 5781 + }, + { + "epoch": 3.1389793702497286, + "grad_norm": 11.998875560827074, + "learning_rate": 6.428416821011814e-06, + "loss": 0.6377, + "step": 5782 + }, + { + "epoch": 3.1395222584147664, + "grad_norm": 13.007472970217544, + "learning_rate": 6.425132159675845e-06, + "loss": 0.5721, + "step": 5783 + }, + { + "epoch": 3.1400651465798046, + "grad_norm": 11.356611803430905, + "learning_rate": 6.4218479404852244e-06, + "loss": 0.3753, + "step": 5784 + }, + { + "epoch": 3.1406080347448424, + "grad_norm": 12.12920282380304, + "learning_rate": 6.418564163846155e-06, + "loss": 0.4842, + "step": 5785 + }, + { + "epoch": 3.1411509229098806, + "grad_norm": 10.041071912027489, + "learning_rate": 6.415280830164779e-06, + "loss": 0.4939, + "step": 5786 + }, + { + "epoch": 3.1416938110749184, + "grad_norm": 12.998407630945819, + "learning_rate": 6.411997939847181e-06, + "loss": 0.636, + "step": 5787 + }, + { + "epoch": 3.1422366992399566, + "grad_norm": 9.655708839558747, + "learning_rate": 6.408715493299398e-06, + "loss": 0.4262, + "step": 5788 + }, + { + "epoch": 3.1427795874049944, + "grad_norm": 18.92187273336628, + "learning_rate": 6.405433490927404e-06, + "loss": 0.6645, + "step": 5789 + }, + { + "epoch": 3.1433224755700326, + "grad_norm": 7.288861789816135, + "learning_rate": 6.402151933137128e-06, + "loss": 0.3247, + "step": 5790 + }, + { + "epoch": 3.1438653637350704, + "grad_norm": 15.171324384191708, + "learning_rate": 6.398870820334431e-06, + "loss": 0.5835, + "step": 5791 + }, + { + "epoch": 3.1444082519001086, + "grad_norm": 19.537893777194515, + "learning_rate": 6.395590152925133e-06, + "loss": 0.4794, + "step": 5792 + }, + { + "epoch": 3.1449511400651464, + "grad_norm": 15.604738367299253, + "learning_rate": 6.3923099313149865e-06, + "loss": 0.8757, + "step": 5793 + }, + { + "epoch": 3.1454940282301846, + "grad_norm": 11.401596673156138, + "learning_rate": 6.389030155909702e-06, + "loss": 0.5278, + "step": 5794 + }, + { + "epoch": 3.1460369163952224, + "grad_norm": 15.026441679185012, + "learning_rate": 6.385750827114919e-06, + "loss": 0.7473, + "step": 5795 + }, + { + "epoch": 3.1465798045602607, + "grad_norm": 10.75641281038809, + "learning_rate": 6.382471945336238e-06, + "loss": 0.7022, + "step": 5796 + }, + { + "epoch": 3.1471226927252984, + "grad_norm": 13.14436770595673, + "learning_rate": 6.379193510979188e-06, + "loss": 0.5036, + "step": 5797 + }, + { + "epoch": 3.1476655808903367, + "grad_norm": 12.687239439350824, + "learning_rate": 6.375915524449259e-06, + "loss": 0.7799, + "step": 5798 + }, + { + "epoch": 3.1482084690553744, + "grad_norm": 12.06538329633808, + "learning_rate": 6.372637986151873e-06, + "loss": 0.6135, + "step": 5799 + }, + { + "epoch": 3.1487513572204127, + "grad_norm": 14.280970019829384, + "learning_rate": 6.369360896492398e-06, + "loss": 0.5631, + "step": 5800 + }, + { + "epoch": 3.1492942453854504, + "grad_norm": 13.49612526729136, + "learning_rate": 6.366084255876159e-06, + "loss": 0.8147, + "step": 5801 + }, + { + "epoch": 3.1498371335504887, + "grad_norm": 16.44633506143303, + "learning_rate": 6.3628080647084074e-06, + "loss": 0.705, + "step": 5802 + }, + { + "epoch": 3.1503800217155264, + "grad_norm": 11.659496492366037, + "learning_rate": 6.359532323394355e-06, + "loss": 0.4872, + "step": 5803 + }, + { + "epoch": 3.1509229098805647, + "grad_norm": 9.20526792646835, + "learning_rate": 6.356257032339141e-06, + "loss": 0.3279, + "step": 5804 + }, + { + "epoch": 3.1514657980456025, + "grad_norm": 15.869195168975136, + "learning_rate": 6.35298219194787e-06, + "loss": 0.6893, + "step": 5805 + }, + { + "epoch": 3.1520086862106407, + "grad_norm": 12.918024481861224, + "learning_rate": 6.349707802625569e-06, + "loss": 0.3561, + "step": 5806 + }, + { + "epoch": 3.1525515743756785, + "grad_norm": 11.948249403281372, + "learning_rate": 6.34643386477723e-06, + "loss": 0.6678, + "step": 5807 + }, + { + "epoch": 3.1530944625407167, + "grad_norm": 10.53882181570608, + "learning_rate": 6.3431603788077736e-06, + "loss": 0.2843, + "step": 5808 + }, + { + "epoch": 3.1536373507057545, + "grad_norm": 14.890522782673512, + "learning_rate": 6.3398873451220644e-06, + "loss": 0.7516, + "step": 5809 + }, + { + "epoch": 3.1541802388707927, + "grad_norm": 13.860961190955267, + "learning_rate": 6.3366147641249266e-06, + "loss": 0.697, + "step": 5810 + }, + { + "epoch": 3.1547231270358305, + "grad_norm": 13.048567845980694, + "learning_rate": 6.3333426362211096e-06, + "loss": 0.5362, + "step": 5811 + }, + { + "epoch": 3.1552660152008687, + "grad_norm": 11.430913085361082, + "learning_rate": 6.33007096181532e-06, + "loss": 0.3698, + "step": 5812 + }, + { + "epoch": 3.1558089033659065, + "grad_norm": 13.799412170578423, + "learning_rate": 6.326799741312202e-06, + "loss": 0.579, + "step": 5813 + }, + { + "epoch": 3.1563517915309447, + "grad_norm": 12.205698077175509, + "learning_rate": 6.323528975116349e-06, + "loss": 0.6956, + "step": 5814 + }, + { + "epoch": 3.1568946796959825, + "grad_norm": 14.923158817152583, + "learning_rate": 6.320258663632288e-06, + "loss": 0.7094, + "step": 5815 + }, + { + "epoch": 3.1574375678610207, + "grad_norm": 9.221494784749213, + "learning_rate": 6.316988807264506e-06, + "loss": 0.3346, + "step": 5816 + }, + { + "epoch": 3.1579804560260585, + "grad_norm": 12.263733681763519, + "learning_rate": 6.3137194064174124e-06, + "loss": 0.6395, + "step": 5817 + }, + { + "epoch": 3.1585233441910967, + "grad_norm": 12.198844326189183, + "learning_rate": 6.31045046149538e-06, + "loss": 0.4894, + "step": 5818 + }, + { + "epoch": 3.1590662323561345, + "grad_norm": 11.16927244530179, + "learning_rate": 6.3071819729027175e-06, + "loss": 0.4653, + "step": 5819 + }, + { + "epoch": 3.1596091205211727, + "grad_norm": 11.182448096771813, + "learning_rate": 6.303913941043669e-06, + "loss": 0.3997, + "step": 5820 + }, + { + "epoch": 3.1601520086862105, + "grad_norm": 9.56488246617197, + "learning_rate": 6.300646366322439e-06, + "loss": 0.6558, + "step": 5821 + }, + { + "epoch": 3.1606948968512487, + "grad_norm": 12.485736110882538, + "learning_rate": 6.297379249143158e-06, + "loss": 0.2747, + "step": 5822 + }, + { + "epoch": 3.1612377850162865, + "grad_norm": 10.29008763858223, + "learning_rate": 6.294112589909919e-06, + "loss": 0.3593, + "step": 5823 + }, + { + "epoch": 3.1617806731813247, + "grad_norm": 10.281965116706075, + "learning_rate": 6.2908463890267345e-06, + "loss": 0.4736, + "step": 5824 + }, + { + "epoch": 3.1623235613463625, + "grad_norm": 9.464092471296539, + "learning_rate": 6.287580646897584e-06, + "loss": 0.4445, + "step": 5825 + }, + { + "epoch": 3.1628664495114007, + "grad_norm": 17.834508294283445, + "learning_rate": 6.284315363926372e-06, + "loss": 0.7063, + "step": 5826 + }, + { + "epoch": 3.1634093376764385, + "grad_norm": 13.155791246426181, + "learning_rate": 6.281050540516963e-06, + "loss": 0.5295, + "step": 5827 + }, + { + "epoch": 3.1639522258414767, + "grad_norm": 11.87732424107669, + "learning_rate": 6.277786177073151e-06, + "loss": 0.6095, + "step": 5828 + }, + { + "epoch": 3.1644951140065145, + "grad_norm": 9.937871872571362, + "learning_rate": 6.274522273998671e-06, + "loss": 0.3788, + "step": 5829 + }, + { + "epoch": 3.1650380021715527, + "grad_norm": 10.26395895690379, + "learning_rate": 6.2712588316972155e-06, + "loss": 0.4352, + "step": 5830 + }, + { + "epoch": 3.1655808903365905, + "grad_norm": 8.627642026627477, + "learning_rate": 6.26799585057241e-06, + "loss": 0.3298, + "step": 5831 + }, + { + "epoch": 3.1661237785016287, + "grad_norm": 12.940542818393178, + "learning_rate": 6.2647333310278295e-06, + "loss": 0.5464, + "step": 5832 + }, + { + "epoch": 3.1666666666666665, + "grad_norm": 11.72079573230472, + "learning_rate": 6.261471273466979e-06, + "loss": 0.3785, + "step": 5833 + }, + { + "epoch": 3.1672095548317047, + "grad_norm": 13.338454285585954, + "learning_rate": 6.258209678293324e-06, + "loss": 0.4455, + "step": 5834 + }, + { + "epoch": 3.1677524429967425, + "grad_norm": 10.755191139797633, + "learning_rate": 6.254948545910256e-06, + "loss": 0.492, + "step": 5835 + }, + { + "epoch": 3.1682953311617807, + "grad_norm": 19.472452158150972, + "learning_rate": 6.251687876721124e-06, + "loss": 0.5031, + "step": 5836 + }, + { + "epoch": 3.1688382193268185, + "grad_norm": 17.265109190709392, + "learning_rate": 6.248427671129207e-06, + "loss": 0.7912, + "step": 5837 + }, + { + "epoch": 3.1693811074918568, + "grad_norm": 15.65566631740727, + "learning_rate": 6.245167929537733e-06, + "loss": 0.5431, + "step": 5838 + }, + { + "epoch": 3.1699239956568945, + "grad_norm": 11.65166282725162, + "learning_rate": 6.241908652349879e-06, + "loss": 0.3528, + "step": 5839 + }, + { + "epoch": 3.1704668838219328, + "grad_norm": 10.29071410809139, + "learning_rate": 6.238649839968746e-06, + "loss": 0.314, + "step": 5840 + }, + { + "epoch": 3.1710097719869705, + "grad_norm": 8.17516522370587, + "learning_rate": 6.2353914927974024e-06, + "loss": 0.3906, + "step": 5841 + }, + { + "epoch": 3.1715526601520088, + "grad_norm": 10.555776499314874, + "learning_rate": 6.232133611238833e-06, + "loss": 0.4222, + "step": 5842 + }, + { + "epoch": 3.1720955483170465, + "grad_norm": 15.841225391604809, + "learning_rate": 6.228876195695986e-06, + "loss": 0.8112, + "step": 5843 + }, + { + "epoch": 3.1726384364820848, + "grad_norm": 12.430872813466289, + "learning_rate": 6.225619246571741e-06, + "loss": 0.5456, + "step": 5844 + }, + { + "epoch": 3.1731813246471225, + "grad_norm": 13.424033486912041, + "learning_rate": 6.222362764268924e-06, + "loss": 0.5942, + "step": 5845 + }, + { + "epoch": 3.1737242128121608, + "grad_norm": 13.261228349891576, + "learning_rate": 6.219106749190298e-06, + "loss": 0.6697, + "step": 5846 + }, + { + "epoch": 3.1742671009771986, + "grad_norm": 12.710362718713364, + "learning_rate": 6.215851201738578e-06, + "loss": 0.5753, + "step": 5847 + }, + { + "epoch": 3.1748099891422368, + "grad_norm": 14.841367294849315, + "learning_rate": 6.212596122316415e-06, + "loss": 0.965, + "step": 5848 + }, + { + "epoch": 3.1753528773072746, + "grad_norm": 13.885344637636523, + "learning_rate": 6.209341511326394e-06, + "loss": 0.5931, + "step": 5849 + }, + { + "epoch": 3.175895765472313, + "grad_norm": 11.460331492201762, + "learning_rate": 6.20608736917106e-06, + "loss": 0.5989, + "step": 5850 + }, + { + "epoch": 3.1764386536373506, + "grad_norm": 11.70057563506978, + "learning_rate": 6.202833696252884e-06, + "loss": 0.4908, + "step": 5851 + }, + { + "epoch": 3.176981541802389, + "grad_norm": 10.645588299206919, + "learning_rate": 6.199580492974291e-06, + "loss": 0.4098, + "step": 5852 + }, + { + "epoch": 3.1775244299674266, + "grad_norm": 12.755952884154276, + "learning_rate": 6.196327759737637e-06, + "loss": 0.6992, + "step": 5853 + }, + { + "epoch": 3.178067318132465, + "grad_norm": 14.709812641950414, + "learning_rate": 6.193075496945231e-06, + "loss": 0.6841, + "step": 5854 + }, + { + "epoch": 3.1786102062975026, + "grad_norm": 13.875280510305055, + "learning_rate": 6.189823704999312e-06, + "loss": 0.7106, + "step": 5855 + }, + { + "epoch": 3.179153094462541, + "grad_norm": 9.281312738229682, + "learning_rate": 6.186572384302072e-06, + "loss": 0.3546, + "step": 5856 + }, + { + "epoch": 3.1796959826275786, + "grad_norm": 12.958826598950415, + "learning_rate": 6.1833215352556375e-06, + "loss": 0.6019, + "step": 5857 + }, + { + "epoch": 3.180238870792617, + "grad_norm": 11.464928690074025, + "learning_rate": 6.180071158262075e-06, + "loss": 0.408, + "step": 5858 + }, + { + "epoch": 3.1807817589576546, + "grad_norm": 9.768529641559107, + "learning_rate": 6.176821253723404e-06, + "loss": 0.4072, + "step": 5859 + }, + { + "epoch": 3.181324647122693, + "grad_norm": 9.678665348787685, + "learning_rate": 6.173571822041568e-06, + "loss": 0.4732, + "step": 5860 + }, + { + "epoch": 3.1818675352877306, + "grad_norm": 14.30027690416035, + "learning_rate": 6.170322863618474e-06, + "loss": 0.7642, + "step": 5861 + }, + { + "epoch": 3.182410423452769, + "grad_norm": 14.961370236846983, + "learning_rate": 6.167074378855946e-06, + "loss": 0.6362, + "step": 5862 + }, + { + "epoch": 3.1829533116178066, + "grad_norm": 9.357640920276147, + "learning_rate": 6.1638263681557695e-06, + "loss": 0.4049, + "step": 5863 + }, + { + "epoch": 3.183496199782845, + "grad_norm": 10.782208269160673, + "learning_rate": 6.160578831919662e-06, + "loss": 0.336, + "step": 5864 + }, + { + "epoch": 3.1840390879478826, + "grad_norm": 14.384452607422604, + "learning_rate": 6.157331770549285e-06, + "loss": 0.5396, + "step": 5865 + }, + { + "epoch": 3.184581976112921, + "grad_norm": 9.250100344453772, + "learning_rate": 6.154085184446237e-06, + "loss": 0.3785, + "step": 5866 + }, + { + "epoch": 3.1851248642779586, + "grad_norm": 15.376158526777193, + "learning_rate": 6.150839074012065e-06, + "loss": 0.5189, + "step": 5867 + }, + { + "epoch": 3.185667752442997, + "grad_norm": 10.177494395248477, + "learning_rate": 6.14759343964825e-06, + "loss": 0.3756, + "step": 5868 + }, + { + "epoch": 3.1862106406080346, + "grad_norm": 17.67659076253852, + "learning_rate": 6.144348281756218e-06, + "loss": 0.7381, + "step": 5869 + }, + { + "epoch": 3.186753528773073, + "grad_norm": 11.56583944116506, + "learning_rate": 6.141103600737337e-06, + "loss": 0.7712, + "step": 5870 + }, + { + "epoch": 3.1872964169381106, + "grad_norm": 15.096190908138054, + "learning_rate": 6.137859396992909e-06, + "loss": 0.6608, + "step": 5871 + }, + { + "epoch": 3.187839305103149, + "grad_norm": 14.055042562247136, + "learning_rate": 6.134615670924191e-06, + "loss": 0.6762, + "step": 5872 + }, + { + "epoch": 3.1883821932681866, + "grad_norm": 12.50506943616142, + "learning_rate": 6.131372422932363e-06, + "loss": 0.4721, + "step": 5873 + }, + { + "epoch": 3.188925081433225, + "grad_norm": 12.679042521649755, + "learning_rate": 6.128129653418562e-06, + "loss": 0.4277, + "step": 5874 + }, + { + "epoch": 3.1894679695982626, + "grad_norm": 13.177050387778413, + "learning_rate": 6.124887362783854e-06, + "loss": 0.7357, + "step": 5875 + }, + { + "epoch": 3.190010857763301, + "grad_norm": 9.589104761627354, + "learning_rate": 6.121645551429255e-06, + "loss": 0.3541, + "step": 5876 + }, + { + "epoch": 3.1905537459283386, + "grad_norm": 15.986950102257723, + "learning_rate": 6.118404219755717e-06, + "loss": 0.5148, + "step": 5877 + }, + { + "epoch": 3.191096634093377, + "grad_norm": 14.855627037048395, + "learning_rate": 6.1151633681641275e-06, + "loss": 0.6874, + "step": 5878 + }, + { + "epoch": 3.1916395222584146, + "grad_norm": 9.68244350868187, + "learning_rate": 6.111922997055328e-06, + "loss": 0.2941, + "step": 5879 + }, + { + "epoch": 3.192182410423453, + "grad_norm": 10.654011696718703, + "learning_rate": 6.108683106830085e-06, + "loss": 0.5009, + "step": 5880 + }, + { + "epoch": 3.1927252985884906, + "grad_norm": 12.855166875044116, + "learning_rate": 6.105443697889121e-06, + "loss": 0.4806, + "step": 5881 + }, + { + "epoch": 3.193268186753529, + "grad_norm": 10.436692542811286, + "learning_rate": 6.102204770633083e-06, + "loss": 0.5723, + "step": 5882 + }, + { + "epoch": 3.1938110749185666, + "grad_norm": 9.990036928653613, + "learning_rate": 6.098966325462574e-06, + "loss": 0.5778, + "step": 5883 + }, + { + "epoch": 3.194353963083605, + "grad_norm": 12.046567955894364, + "learning_rate": 6.095728362778125e-06, + "loss": 0.5015, + "step": 5884 + }, + { + "epoch": 3.1948968512486426, + "grad_norm": 8.725155173492308, + "learning_rate": 6.092490882980219e-06, + "loss": 0.3857, + "step": 5885 + }, + { + "epoch": 3.195439739413681, + "grad_norm": 11.013208621693263, + "learning_rate": 6.089253886469267e-06, + "loss": 0.6389, + "step": 5886 + }, + { + "epoch": 3.1959826275787186, + "grad_norm": 12.794704970363892, + "learning_rate": 6.086017373645624e-06, + "loss": 0.6362, + "step": 5887 + }, + { + "epoch": 3.196525515743757, + "grad_norm": 10.413523492938069, + "learning_rate": 6.082781344909592e-06, + "loss": 0.477, + "step": 5888 + }, + { + "epoch": 3.1970684039087947, + "grad_norm": 19.28502803064406, + "learning_rate": 6.079545800661404e-06, + "loss": 0.6435, + "step": 5889 + }, + { + "epoch": 3.197611292073833, + "grad_norm": 12.8419010375632, + "learning_rate": 6.076310741301244e-06, + "loss": 0.816, + "step": 5890 + }, + { + "epoch": 3.1981541802388707, + "grad_norm": 13.169440134578803, + "learning_rate": 6.0730761672292215e-06, + "loss": 0.4307, + "step": 5891 + }, + { + "epoch": 3.198697068403909, + "grad_norm": 12.487488215101045, + "learning_rate": 6.0698420788454e-06, + "loss": 0.6785, + "step": 5892 + }, + { + "epoch": 3.1992399565689467, + "grad_norm": 13.786560936789252, + "learning_rate": 6.066608476549771e-06, + "loss": 0.866, + "step": 5893 + }, + { + "epoch": 3.199782844733985, + "grad_norm": 14.699571921737691, + "learning_rate": 6.0633753607422785e-06, + "loss": 0.682, + "step": 5894 + }, + { + "epoch": 3.2003257328990227, + "grad_norm": 10.298137592669294, + "learning_rate": 6.060142731822792e-06, + "loss": 0.4765, + "step": 5895 + }, + { + "epoch": 3.200868621064061, + "grad_norm": 11.57501814659113, + "learning_rate": 6.056910590191132e-06, + "loss": 0.4907, + "step": 5896 + }, + { + "epoch": 3.2014115092290987, + "grad_norm": 10.861108561617078, + "learning_rate": 6.0536789362470575e-06, + "loss": 0.4201, + "step": 5897 + }, + { + "epoch": 3.201954397394137, + "grad_norm": 14.726234720260377, + "learning_rate": 6.050447770390258e-06, + "loss": 0.5714, + "step": 5898 + }, + { + "epoch": 3.2024972855591747, + "grad_norm": 10.353294495922004, + "learning_rate": 6.047217093020374e-06, + "loss": 0.3656, + "step": 5899 + }, + { + "epoch": 3.203040173724213, + "grad_norm": 7.811262042752878, + "learning_rate": 6.043986904536979e-06, + "loss": 0.3013, + "step": 5900 + }, + { + "epoch": 3.2035830618892507, + "grad_norm": 9.961324210163603, + "learning_rate": 6.040757205339589e-06, + "loss": 0.3498, + "step": 5901 + }, + { + "epoch": 3.204125950054289, + "grad_norm": 11.897050086158284, + "learning_rate": 6.037527995827659e-06, + "loss": 0.4876, + "step": 5902 + }, + { + "epoch": 3.2046688382193267, + "grad_norm": 10.075300527690798, + "learning_rate": 6.034299276400582e-06, + "loss": 0.3781, + "step": 5903 + }, + { + "epoch": 3.205211726384365, + "grad_norm": 20.471624027064, + "learning_rate": 6.031071047457689e-06, + "loss": 0.7781, + "step": 5904 + }, + { + "epoch": 3.2057546145494027, + "grad_norm": 13.759518261461583, + "learning_rate": 6.027843309398256e-06, + "loss": 0.6616, + "step": 5905 + }, + { + "epoch": 3.206297502714441, + "grad_norm": 11.734095364989608, + "learning_rate": 6.024616062621496e-06, + "loss": 0.4283, + "step": 5906 + }, + { + "epoch": 3.2068403908794787, + "grad_norm": 11.862203608845647, + "learning_rate": 6.02138930752655e-06, + "loss": 0.5662, + "step": 5907 + }, + { + "epoch": 3.207383279044517, + "grad_norm": 15.72669110441235, + "learning_rate": 6.018163044512521e-06, + "loss": 0.5579, + "step": 5908 + }, + { + "epoch": 3.2079261672095547, + "grad_norm": 19.31549988156055, + "learning_rate": 6.014937273978429e-06, + "loss": 0.7692, + "step": 5909 + }, + { + "epoch": 3.208469055374593, + "grad_norm": 10.302000372364747, + "learning_rate": 6.011711996323251e-06, + "loss": 0.5133, + "step": 5910 + }, + { + "epoch": 3.2090119435396307, + "grad_norm": 12.224649649591383, + "learning_rate": 6.008487211945884e-06, + "loss": 0.7677, + "step": 5911 + }, + { + "epoch": 3.209554831704669, + "grad_norm": 17.460170685666615, + "learning_rate": 6.005262921245185e-06, + "loss": 0.6836, + "step": 5912 + }, + { + "epoch": 3.2100977198697067, + "grad_norm": 19.301567461364495, + "learning_rate": 6.00203912461993e-06, + "loss": 0.9095, + "step": 5913 + }, + { + "epoch": 3.210640608034745, + "grad_norm": 10.852675563314923, + "learning_rate": 5.998815822468851e-06, + "loss": 0.5342, + "step": 5914 + }, + { + "epoch": 3.2111834961997827, + "grad_norm": 10.649138913539469, + "learning_rate": 5.995593015190608e-06, + "loss": 0.5665, + "step": 5915 + }, + { + "epoch": 3.211726384364821, + "grad_norm": 10.875890162606636, + "learning_rate": 5.992370703183803e-06, + "loss": 0.3634, + "step": 5916 + }, + { + "epoch": 3.2122692725298587, + "grad_norm": 16.640296114311138, + "learning_rate": 5.9891488868469775e-06, + "loss": 0.7804, + "step": 5917 + }, + { + "epoch": 3.212812160694897, + "grad_norm": 9.73069162630754, + "learning_rate": 5.9859275665786065e-06, + "loss": 0.35, + "step": 5918 + }, + { + "epoch": 3.2133550488599347, + "grad_norm": 10.657768777546748, + "learning_rate": 5.982706742777116e-06, + "loss": 0.529, + "step": 5919 + }, + { + "epoch": 3.213897937024973, + "grad_norm": 12.700699759951497, + "learning_rate": 5.9794864158408535e-06, + "loss": 0.5159, + "step": 5920 + }, + { + "epoch": 3.2144408251900107, + "grad_norm": 8.375283482104031, + "learning_rate": 5.976266586168121e-06, + "loss": 0.2818, + "step": 5921 + }, + { + "epoch": 3.214983713355049, + "grad_norm": 18.061811128959963, + "learning_rate": 5.973047254157148e-06, + "loss": 1.2271, + "step": 5922 + }, + { + "epoch": 3.2155266015200867, + "grad_norm": 10.630424335111323, + "learning_rate": 5.969828420206114e-06, + "loss": 0.3874, + "step": 5923 + }, + { + "epoch": 3.216069489685125, + "grad_norm": 11.324230234064501, + "learning_rate": 5.966610084713118e-06, + "loss": 0.6559, + "step": 5924 + }, + { + "epoch": 3.2166123778501627, + "grad_norm": 12.436990138030758, + "learning_rate": 5.9633922480762205e-06, + "loss": 0.6468, + "step": 5925 + }, + { + "epoch": 3.217155266015201, + "grad_norm": 16.354922341711003, + "learning_rate": 5.960174910693401e-06, + "loss": 0.8733, + "step": 5926 + }, + { + "epoch": 3.2176981541802387, + "grad_norm": 12.756426829873451, + "learning_rate": 5.956958072962586e-06, + "loss": 0.5399, + "step": 5927 + }, + { + "epoch": 3.218241042345277, + "grad_norm": 11.493661345744254, + "learning_rate": 5.953741735281641e-06, + "loss": 0.5487, + "step": 5928 + }, + { + "epoch": 3.2187839305103148, + "grad_norm": 11.431750755108203, + "learning_rate": 5.9505258980483645e-06, + "loss": 0.3997, + "step": 5929 + }, + { + "epoch": 3.219326818675353, + "grad_norm": 12.940367227326554, + "learning_rate": 5.947310561660503e-06, + "loss": 0.8178, + "step": 5930 + }, + { + "epoch": 3.2198697068403908, + "grad_norm": 11.421594805016625, + "learning_rate": 5.9440957265157265e-06, + "loss": 0.5015, + "step": 5931 + }, + { + "epoch": 3.220412595005429, + "grad_norm": 12.215300050799616, + "learning_rate": 5.940881393011659e-06, + "loss": 0.4044, + "step": 5932 + }, + { + "epoch": 3.2209554831704668, + "grad_norm": 11.265138511012573, + "learning_rate": 5.937667561545845e-06, + "loss": 0.6294, + "step": 5933 + }, + { + "epoch": 3.221498371335505, + "grad_norm": 11.883510126446058, + "learning_rate": 5.934454232515785e-06, + "loss": 0.5257, + "step": 5934 + }, + { + "epoch": 3.2220412595005428, + "grad_norm": 14.862260710742142, + "learning_rate": 5.931241406318906e-06, + "loss": 0.7435, + "step": 5935 + }, + { + "epoch": 3.222584147665581, + "grad_norm": 11.010381613248398, + "learning_rate": 5.92802908335257e-06, + "loss": 0.4852, + "step": 5936 + }, + { + "epoch": 3.2231270358306188, + "grad_norm": 10.894562800140209, + "learning_rate": 5.924817264014091e-06, + "loss": 0.4694, + "step": 5937 + }, + { + "epoch": 3.223669923995657, + "grad_norm": 9.91297174934936, + "learning_rate": 5.921605948700704e-06, + "loss": 0.4031, + "step": 5938 + }, + { + "epoch": 3.2242128121606948, + "grad_norm": 10.829816377096837, + "learning_rate": 5.918395137809596e-06, + "loss": 0.5139, + "step": 5939 + }, + { + "epoch": 3.224755700325733, + "grad_norm": 9.940894286293783, + "learning_rate": 5.915184831737881e-06, + "loss": 0.3475, + "step": 5940 + }, + { + "epoch": 3.225298588490771, + "grad_norm": 12.818677443045534, + "learning_rate": 5.911975030882617e-06, + "loss": 0.4293, + "step": 5941 + }, + { + "epoch": 3.225841476655809, + "grad_norm": 8.859395484195392, + "learning_rate": 5.908765735640795e-06, + "loss": 0.4065, + "step": 5942 + }, + { + "epoch": 3.226384364820847, + "grad_norm": 10.145624567540983, + "learning_rate": 5.90555694640935e-06, + "loss": 0.3713, + "step": 5943 + }, + { + "epoch": 3.226927252985885, + "grad_norm": 12.596453073542072, + "learning_rate": 5.902348663585144e-06, + "loss": 0.5128, + "step": 5944 + }, + { + "epoch": 3.227470141150923, + "grad_norm": 13.220304215230376, + "learning_rate": 5.899140887564991e-06, + "loss": 0.7293, + "step": 5945 + }, + { + "epoch": 3.228013029315961, + "grad_norm": 15.436135404867363, + "learning_rate": 5.8959336187456265e-06, + "loss": 0.8522, + "step": 5946 + }, + { + "epoch": 3.228555917480999, + "grad_norm": 11.078819738735637, + "learning_rate": 5.89272685752373e-06, + "loss": 0.5559, + "step": 5947 + }, + { + "epoch": 3.229098805646037, + "grad_norm": 11.732786443102684, + "learning_rate": 5.889520604295926e-06, + "loss": 0.472, + "step": 5948 + }, + { + "epoch": 3.229641693811075, + "grad_norm": 12.744163420827304, + "learning_rate": 5.88631485945876e-06, + "loss": 0.5843, + "step": 5949 + }, + { + "epoch": 3.230184581976113, + "grad_norm": 12.717140075489619, + "learning_rate": 5.8831096234087335e-06, + "loss": 0.4826, + "step": 5950 + }, + { + "epoch": 3.230727470141151, + "grad_norm": 14.460440748409685, + "learning_rate": 5.8799048965422665e-06, + "loss": 0.5057, + "step": 5951 + }, + { + "epoch": 3.231270358306189, + "grad_norm": 13.034475639862867, + "learning_rate": 5.8767006792557316e-06, + "loss": 0.6811, + "step": 5952 + }, + { + "epoch": 3.231813246471227, + "grad_norm": 12.638392990019069, + "learning_rate": 5.873496971945425e-06, + "loss": 0.4532, + "step": 5953 + }, + { + "epoch": 3.232356134636265, + "grad_norm": 11.04321550599593, + "learning_rate": 5.8702937750075914e-06, + "loss": 0.5711, + "step": 5954 + }, + { + "epoch": 3.232899022801303, + "grad_norm": 14.964397050426088, + "learning_rate": 5.867091088838406e-06, + "loss": 0.6479, + "step": 5955 + }, + { + "epoch": 3.233441910966341, + "grad_norm": 10.647120251880176, + "learning_rate": 5.863888913833979e-06, + "loss": 0.5262, + "step": 5956 + }, + { + "epoch": 3.233984799131379, + "grad_norm": 11.288671625393498, + "learning_rate": 5.860687250390367e-06, + "loss": 0.479, + "step": 5957 + }, + { + "epoch": 3.234527687296417, + "grad_norm": 11.49866073465499, + "learning_rate": 5.857486098903549e-06, + "loss": 0.2939, + "step": 5958 + }, + { + "epoch": 3.235070575461455, + "grad_norm": 11.340345898343635, + "learning_rate": 5.854285459769454e-06, + "loss": 0.3908, + "step": 5959 + }, + { + "epoch": 3.235613463626493, + "grad_norm": 10.20477531793921, + "learning_rate": 5.851085333383939e-06, + "loss": 0.4133, + "step": 5960 + }, + { + "epoch": 3.236156351791531, + "grad_norm": 9.818700824916304, + "learning_rate": 5.8478857201428075e-06, + "loss": 0.3324, + "step": 5961 + }, + { + "epoch": 3.236699239956569, + "grad_norm": 9.73255562956133, + "learning_rate": 5.844686620441782e-06, + "loss": 0.7815, + "step": 5962 + }, + { + "epoch": 3.237242128121607, + "grad_norm": 17.504088445923752, + "learning_rate": 5.841488034676542e-06, + "loss": 0.693, + "step": 5963 + }, + { + "epoch": 3.237785016286645, + "grad_norm": 15.010551187805474, + "learning_rate": 5.838289963242684e-06, + "loss": 0.453, + "step": 5964 + }, + { + "epoch": 3.238327904451683, + "grad_norm": 11.512605593104578, + "learning_rate": 5.835092406535763e-06, + "loss": 0.792, + "step": 5965 + }, + { + "epoch": 3.238870792616721, + "grad_norm": 15.034973727458302, + "learning_rate": 5.831895364951247e-06, + "loss": 0.6216, + "step": 5966 + }, + { + "epoch": 3.239413680781759, + "grad_norm": 15.522277540385527, + "learning_rate": 5.828698838884558e-06, + "loss": 0.4771, + "step": 5967 + }, + { + "epoch": 3.239956568946797, + "grad_norm": 11.391715611260866, + "learning_rate": 5.825502828731042e-06, + "loss": 0.4213, + "step": 5968 + }, + { + "epoch": 3.240499457111835, + "grad_norm": 15.534988403627082, + "learning_rate": 5.82230733488599e-06, + "loss": 0.6494, + "step": 5969 + }, + { + "epoch": 3.241042345276873, + "grad_norm": 18.181760899119496, + "learning_rate": 5.819112357744626e-06, + "loss": 0.5825, + "step": 5970 + }, + { + "epoch": 3.241585233441911, + "grad_norm": 10.971869300023029, + "learning_rate": 5.815917897702104e-06, + "loss": 0.7647, + "step": 5971 + }, + { + "epoch": 3.242128121606949, + "grad_norm": 11.760933490969961, + "learning_rate": 5.812723955153533e-06, + "loss": 0.6203, + "step": 5972 + }, + { + "epoch": 3.242671009771987, + "grad_norm": 17.51328901218973, + "learning_rate": 5.809530530493929e-06, + "loss": 0.6598, + "step": 5973 + }, + { + "epoch": 3.243213897937025, + "grad_norm": 12.500120263524371, + "learning_rate": 5.8063376241182745e-06, + "loss": 0.4958, + "step": 5974 + }, + { + "epoch": 3.243756786102063, + "grad_norm": 11.050563890247632, + "learning_rate": 5.803145236421464e-06, + "loss": 0.4204, + "step": 5975 + }, + { + "epoch": 3.244299674267101, + "grad_norm": 6.058184111347752, + "learning_rate": 5.799953367798335e-06, + "loss": 0.2181, + "step": 5976 + }, + { + "epoch": 3.244842562432139, + "grad_norm": 9.648373148728487, + "learning_rate": 5.796762018643675e-06, + "loss": 0.4304, + "step": 5977 + }, + { + "epoch": 3.245385450597177, + "grad_norm": 11.906735912997219, + "learning_rate": 5.793571189352179e-06, + "loss": 0.5365, + "step": 5978 + }, + { + "epoch": 3.245928338762215, + "grad_norm": 12.12325593772168, + "learning_rate": 5.790380880318511e-06, + "loss": 0.3811, + "step": 5979 + }, + { + "epoch": 3.246471226927253, + "grad_norm": 18.625337422182657, + "learning_rate": 5.787191091937236e-06, + "loss": 0.6784, + "step": 5980 + }, + { + "epoch": 3.247014115092291, + "grad_norm": 9.093728620631289, + "learning_rate": 5.784001824602891e-06, + "loss": 0.4545, + "step": 5981 + }, + { + "epoch": 3.247557003257329, + "grad_norm": 12.360404516544222, + "learning_rate": 5.78081307870991e-06, + "loss": 0.5982, + "step": 5982 + }, + { + "epoch": 3.248099891422367, + "grad_norm": 12.023338602328888, + "learning_rate": 5.777624854652696e-06, + "loss": 0.4595, + "step": 5983 + }, + { + "epoch": 3.248642779587405, + "grad_norm": 11.60238178463911, + "learning_rate": 5.774437152825574e-06, + "loss": 0.4674, + "step": 5984 + }, + { + "epoch": 3.249185667752443, + "grad_norm": 13.826357134916073, + "learning_rate": 5.77124997362279e-06, + "loss": 0.4552, + "step": 5985 + }, + { + "epoch": 3.249728555917481, + "grad_norm": 13.531388043207002, + "learning_rate": 5.7680633174385595e-06, + "loss": 0.4926, + "step": 5986 + }, + { + "epoch": 3.250271444082519, + "grad_norm": 11.384053416563145, + "learning_rate": 5.764877184666993e-06, + "loss": 0.5127, + "step": 5987 + }, + { + "epoch": 3.250814332247557, + "grad_norm": 12.390918037366044, + "learning_rate": 5.761691575702168e-06, + "loss": 0.6428, + "step": 5988 + }, + { + "epoch": 3.251357220412595, + "grad_norm": 12.158948584345922, + "learning_rate": 5.758506490938087e-06, + "loss": 0.5374, + "step": 5989 + }, + { + "epoch": 3.251900108577633, + "grad_norm": 12.352114570156113, + "learning_rate": 5.755321930768683e-06, + "loss": 0.4149, + "step": 5990 + }, + { + "epoch": 3.252442996742671, + "grad_norm": 11.451850663530495, + "learning_rate": 5.752137895587826e-06, + "loss": 0.4864, + "step": 5991 + }, + { + "epoch": 3.252985884907709, + "grad_norm": 14.59595398014034, + "learning_rate": 5.748954385789325e-06, + "loss": 0.6791, + "step": 5992 + }, + { + "epoch": 3.253528773072747, + "grad_norm": 10.572257438915425, + "learning_rate": 5.74577140176692e-06, + "loss": 0.3851, + "step": 5993 + }, + { + "epoch": 3.254071661237785, + "grad_norm": 13.496500778871066, + "learning_rate": 5.742588943914289e-06, + "loss": 0.8371, + "step": 5994 + }, + { + "epoch": 3.254614549402823, + "grad_norm": 10.366843508091446, + "learning_rate": 5.739407012625043e-06, + "loss": 0.5289, + "step": 5995 + }, + { + "epoch": 3.255157437567861, + "grad_norm": 13.59778716822848, + "learning_rate": 5.736225608292727e-06, + "loss": 0.6353, + "step": 5996 + }, + { + "epoch": 3.255700325732899, + "grad_norm": 8.13083770907386, + "learning_rate": 5.7330447313108236e-06, + "loss": 0.2591, + "step": 5997 + }, + { + "epoch": 3.256243213897937, + "grad_norm": 11.705959913597166, + "learning_rate": 5.72986438207275e-06, + "loss": 0.4058, + "step": 5998 + }, + { + "epoch": 3.256786102062975, + "grad_norm": 9.25230495085822, + "learning_rate": 5.726684560971852e-06, + "loss": 0.3647, + "step": 5999 + }, + { + "epoch": 3.257328990228013, + "grad_norm": 10.705513260623167, + "learning_rate": 5.72350526840142e-06, + "loss": 0.6986, + "step": 6000 + }, + { + "epoch": 3.257871878393051, + "grad_norm": 11.426736023101553, + "learning_rate": 5.720326504754672e-06, + "loss": 0.548, + "step": 6001 + }, + { + "epoch": 3.258414766558089, + "grad_norm": 11.105330737807535, + "learning_rate": 5.71714827042476e-06, + "loss": 0.5345, + "step": 6002 + }, + { + "epoch": 3.258957654723127, + "grad_norm": 11.734545793783724, + "learning_rate": 5.713970565804782e-06, + "loss": 0.4435, + "step": 6003 + }, + { + "epoch": 3.259500542888165, + "grad_norm": 13.234029513207872, + "learning_rate": 5.710793391287751e-06, + "loss": 0.538, + "step": 6004 + }, + { + "epoch": 3.260043431053203, + "grad_norm": 9.429540505637686, + "learning_rate": 5.707616747266631e-06, + "loss": 0.4042, + "step": 6005 + }, + { + "epoch": 3.260586319218241, + "grad_norm": 10.73830038342973, + "learning_rate": 5.704440634134312e-06, + "loss": 0.4394, + "step": 6006 + }, + { + "epoch": 3.261129207383279, + "grad_norm": 15.154563307319142, + "learning_rate": 5.70126505228362e-06, + "loss": 0.5733, + "step": 6007 + }, + { + "epoch": 3.261672095548317, + "grad_norm": 11.231865527791925, + "learning_rate": 5.6980900021073196e-06, + "loss": 0.3472, + "step": 6008 + }, + { + "epoch": 3.262214983713355, + "grad_norm": 10.315608836480246, + "learning_rate": 5.694915483998099e-06, + "loss": 0.3231, + "step": 6009 + }, + { + "epoch": 3.262757871878393, + "grad_norm": 13.017299228674494, + "learning_rate": 5.691741498348601e-06, + "loss": 0.4652, + "step": 6010 + }, + { + "epoch": 3.263300760043431, + "grad_norm": 13.428357165583693, + "learning_rate": 5.688568045551373e-06, + "loss": 0.51, + "step": 6011 + }, + { + "epoch": 3.263843648208469, + "grad_norm": 12.887391562992057, + "learning_rate": 5.685395125998927e-06, + "loss": 0.7156, + "step": 6012 + }, + { + "epoch": 3.264386536373507, + "grad_norm": 11.112402459616494, + "learning_rate": 5.682222740083683e-06, + "loss": 0.3077, + "step": 6013 + }, + { + "epoch": 3.264929424538545, + "grad_norm": 17.334186329467677, + "learning_rate": 5.6790508881980136e-06, + "loss": 0.487, + "step": 6014 + }, + { + "epoch": 3.265472312703583, + "grad_norm": 11.272673037009941, + "learning_rate": 5.675879570734222e-06, + "loss": 0.5596, + "step": 6015 + }, + { + "epoch": 3.266015200868621, + "grad_norm": 10.57062964165952, + "learning_rate": 5.672708788084528e-06, + "loss": 0.4693, + "step": 6016 + }, + { + "epoch": 3.266558089033659, + "grad_norm": 12.532994743872049, + "learning_rate": 5.669538540641116e-06, + "loss": 0.5709, + "step": 6017 + }, + { + "epoch": 3.267100977198697, + "grad_norm": 12.467434828440602, + "learning_rate": 5.666368828796072e-06, + "loss": 0.5328, + "step": 6018 + }, + { + "epoch": 3.267643865363735, + "grad_norm": 10.349926895360463, + "learning_rate": 5.663199652941444e-06, + "loss": 0.3587, + "step": 6019 + }, + { + "epoch": 3.268186753528773, + "grad_norm": 10.889914966114198, + "learning_rate": 5.660031013469189e-06, + "loss": 0.469, + "step": 6020 + }, + { + "epoch": 3.268729641693811, + "grad_norm": 16.63961667582981, + "learning_rate": 5.656862910771218e-06, + "loss": 0.7146, + "step": 6021 + }, + { + "epoch": 3.269272529858849, + "grad_norm": 10.447294419775371, + "learning_rate": 5.653695345239365e-06, + "loss": 0.3975, + "step": 6022 + }, + { + "epoch": 3.269815418023887, + "grad_norm": 12.85070217834024, + "learning_rate": 5.650528317265398e-06, + "loss": 0.7395, + "step": 6023 + }, + { + "epoch": 3.270358306188925, + "grad_norm": 8.341742485534219, + "learning_rate": 5.647361827241025e-06, + "loss": 0.2861, + "step": 6024 + }, + { + "epoch": 3.270901194353963, + "grad_norm": 9.898309581037385, + "learning_rate": 5.64419587555787e-06, + "loss": 0.6192, + "step": 6025 + }, + { + "epoch": 3.271444082519001, + "grad_norm": 12.335345691658691, + "learning_rate": 5.641030462607514e-06, + "loss": 0.5161, + "step": 6026 + }, + { + "epoch": 3.271986970684039, + "grad_norm": 10.304455881328025, + "learning_rate": 5.637865588781459e-06, + "loss": 0.5773, + "step": 6027 + }, + { + "epoch": 3.272529858849077, + "grad_norm": 11.527834619584397, + "learning_rate": 5.63470125447114e-06, + "loss": 0.4415, + "step": 6028 + }, + { + "epoch": 3.273072747014115, + "grad_norm": 12.835285732552766, + "learning_rate": 5.631537460067926e-06, + "loss": 0.4513, + "step": 6029 + }, + { + "epoch": 3.273615635179153, + "grad_norm": 15.34754753916867, + "learning_rate": 5.628374205963123e-06, + "loss": 0.9626, + "step": 6030 + }, + { + "epoch": 3.274158523344191, + "grad_norm": 13.537553719954104, + "learning_rate": 5.625211492547965e-06, + "loss": 0.6022, + "step": 6031 + }, + { + "epoch": 3.274701411509229, + "grad_norm": 13.03189959063881, + "learning_rate": 5.622049320213622e-06, + "loss": 0.3918, + "step": 6032 + }, + { + "epoch": 3.275244299674267, + "grad_norm": 11.883924111949161, + "learning_rate": 5.618887689351195e-06, + "loss": 0.6166, + "step": 6033 + }, + { + "epoch": 3.2757871878393052, + "grad_norm": 12.574278462374288, + "learning_rate": 5.615726600351723e-06, + "loss": 0.364, + "step": 6034 + }, + { + "epoch": 3.276330076004343, + "grad_norm": 11.646167790458746, + "learning_rate": 5.612566053606172e-06, + "loss": 0.4893, + "step": 6035 + }, + { + "epoch": 3.2768729641693812, + "grad_norm": 15.464095407664189, + "learning_rate": 5.6094060495054435e-06, + "loss": 0.5033, + "step": 6036 + }, + { + "epoch": 3.277415852334419, + "grad_norm": 9.470859172197285, + "learning_rate": 5.606246588440374e-06, + "loss": 0.3201, + "step": 6037 + }, + { + "epoch": 3.2779587404994572, + "grad_norm": 10.186080712958546, + "learning_rate": 5.603087670801728e-06, + "loss": 0.375, + "step": 6038 + }, + { + "epoch": 3.278501628664495, + "grad_norm": 12.069192340552016, + "learning_rate": 5.599929296980205e-06, + "loss": 0.521, + "step": 6039 + }, + { + "epoch": 3.2790445168295332, + "grad_norm": 13.470203218211036, + "learning_rate": 5.596771467366442e-06, + "loss": 0.6999, + "step": 6040 + }, + { + "epoch": 3.279587404994571, + "grad_norm": 15.234328984284964, + "learning_rate": 5.5936141823509995e-06, + "loss": 0.658, + "step": 6041 + }, + { + "epoch": 3.2801302931596092, + "grad_norm": 10.28714486065803, + "learning_rate": 5.590457442324374e-06, + "loss": 0.49, + "step": 6042 + }, + { + "epoch": 3.280673181324647, + "grad_norm": 13.229082345644139, + "learning_rate": 5.587301247677009e-06, + "loss": 0.8617, + "step": 6043 + }, + { + "epoch": 3.2812160694896852, + "grad_norm": 19.161743435780973, + "learning_rate": 5.5841455987992536e-06, + "loss": 0.673, + "step": 6044 + }, + { + "epoch": 3.281758957654723, + "grad_norm": 14.576260981710798, + "learning_rate": 5.580990496081407e-06, + "loss": 0.4843, + "step": 6045 + }, + { + "epoch": 3.2823018458197613, + "grad_norm": 10.509886153036398, + "learning_rate": 5.577835939913701e-06, + "loss": 0.3372, + "step": 6046 + }, + { + "epoch": 3.282844733984799, + "grad_norm": 14.23830213747384, + "learning_rate": 5.574681930686289e-06, + "loss": 0.5041, + "step": 6047 + }, + { + "epoch": 3.2833876221498373, + "grad_norm": 16.736802729512746, + "learning_rate": 5.571528468789276e-06, + "loss": 0.6261, + "step": 6048 + }, + { + "epoch": 3.283930510314875, + "grad_norm": 16.53917203611799, + "learning_rate": 5.5683755546126724e-06, + "loss": 0.6671, + "step": 6049 + }, + { + "epoch": 3.2844733984799133, + "grad_norm": 9.26057962235423, + "learning_rate": 5.565223188546452e-06, + "loss": 0.4135, + "step": 6050 + }, + { + "epoch": 3.285016286644951, + "grad_norm": 13.238225399242737, + "learning_rate": 5.562071370980486e-06, + "loss": 0.558, + "step": 6051 + }, + { + "epoch": 3.2855591748099893, + "grad_norm": 9.005338285525495, + "learning_rate": 5.558920102304615e-06, + "loss": 0.3677, + "step": 6052 + }, + { + "epoch": 3.286102062975027, + "grad_norm": 14.118931865854162, + "learning_rate": 5.55576938290858e-06, + "loss": 0.8303, + "step": 6053 + }, + { + "epoch": 3.2866449511400653, + "grad_norm": 9.596979737331942, + "learning_rate": 5.5526192131820665e-06, + "loss": 0.3322, + "step": 6054 + }, + { + "epoch": 3.287187839305103, + "grad_norm": 11.143732344567033, + "learning_rate": 5.549469593514705e-06, + "loss": 0.4143, + "step": 6055 + }, + { + "epoch": 3.2877307274701413, + "grad_norm": 12.135282131920155, + "learning_rate": 5.546320524296028e-06, + "loss": 0.5302, + "step": 6056 + }, + { + "epoch": 3.288273615635179, + "grad_norm": 12.209025529059957, + "learning_rate": 5.543172005915536e-06, + "loss": 0.5643, + "step": 6057 + }, + { + "epoch": 3.2888165038002173, + "grad_norm": 16.880330158776534, + "learning_rate": 5.540024038762623e-06, + "loss": 0.6743, + "step": 6058 + }, + { + "epoch": 3.289359391965255, + "grad_norm": 13.623361718010658, + "learning_rate": 5.536876623226652e-06, + "loss": 0.6925, + "step": 6059 + }, + { + "epoch": 3.2899022801302933, + "grad_norm": 11.161796182879957, + "learning_rate": 5.533729759696893e-06, + "loss": 0.5875, + "step": 6060 + }, + { + "epoch": 3.290445168295331, + "grad_norm": 14.324802623539963, + "learning_rate": 5.5305834485625545e-06, + "loss": 0.9071, + "step": 6061 + }, + { + "epoch": 3.2909880564603693, + "grad_norm": 12.43112855099258, + "learning_rate": 5.527437690212778e-06, + "loss": 0.3761, + "step": 6062 + }, + { + "epoch": 3.291530944625407, + "grad_norm": 14.231447955496243, + "learning_rate": 5.524292485036638e-06, + "loss": 0.7001, + "step": 6063 + }, + { + "epoch": 3.2920738327904453, + "grad_norm": 11.474598912194304, + "learning_rate": 5.521147833423135e-06, + "loss": 0.419, + "step": 6064 + }, + { + "epoch": 3.292616720955483, + "grad_norm": 13.615753477381617, + "learning_rate": 5.518003735761206e-06, + "loss": 0.7142, + "step": 6065 + }, + { + "epoch": 3.2931596091205213, + "grad_norm": 10.341494655988399, + "learning_rate": 5.514860192439719e-06, + "loss": 0.4286, + "step": 6066 + }, + { + "epoch": 3.293702497285559, + "grad_norm": 11.37742694154389, + "learning_rate": 5.511717203847472e-06, + "loss": 0.4302, + "step": 6067 + }, + { + "epoch": 3.2942453854505973, + "grad_norm": 14.04977404747662, + "learning_rate": 5.508574770373194e-06, + "loss": 0.4538, + "step": 6068 + }, + { + "epoch": 3.294788273615635, + "grad_norm": 13.008904129849688, + "learning_rate": 5.505432892405547e-06, + "loss": 0.419, + "step": 6069 + }, + { + "epoch": 3.2953311617806733, + "grad_norm": 8.086590828855625, + "learning_rate": 5.502291570333122e-06, + "loss": 0.276, + "step": 6070 + }, + { + "epoch": 3.295874049945711, + "grad_norm": 8.20494775069595, + "learning_rate": 5.499150804544445e-06, + "loss": 0.349, + "step": 6071 + }, + { + "epoch": 3.2964169381107493, + "grad_norm": 13.336711424526726, + "learning_rate": 5.496010595427969e-06, + "loss": 0.3958, + "step": 6072 + }, + { + "epoch": 3.296959826275787, + "grad_norm": 21.927197817240263, + "learning_rate": 5.492870943372082e-06, + "loss": 0.5561, + "step": 6073 + }, + { + "epoch": 3.2975027144408253, + "grad_norm": 14.168102563986727, + "learning_rate": 5.489731848765098e-06, + "loss": 0.6231, + "step": 6074 + }, + { + "epoch": 3.298045602605863, + "grad_norm": 9.428786171987674, + "learning_rate": 5.48659331199527e-06, + "loss": 0.3004, + "step": 6075 + }, + { + "epoch": 3.2985884907709013, + "grad_norm": 13.58619098409203, + "learning_rate": 5.483455333450774e-06, + "loss": 0.7655, + "step": 6076 + }, + { + "epoch": 3.299131378935939, + "grad_norm": 10.799157987601124, + "learning_rate": 5.480317913519718e-06, + "loss": 0.4688, + "step": 6077 + }, + { + "epoch": 3.2996742671009773, + "grad_norm": 11.88277850667338, + "learning_rate": 5.477181052590148e-06, + "loss": 0.576, + "step": 6078 + }, + { + "epoch": 3.300217155266015, + "grad_norm": 9.399361473432771, + "learning_rate": 5.474044751050032e-06, + "loss": 0.3612, + "step": 6079 + }, + { + "epoch": 3.3007600434310533, + "grad_norm": 12.83439735807638, + "learning_rate": 5.470909009287273e-06, + "loss": 0.4007, + "step": 6080 + }, + { + "epoch": 3.301302931596091, + "grad_norm": 10.935369972874188, + "learning_rate": 5.467773827689712e-06, + "loss": 0.5349, + "step": 6081 + }, + { + "epoch": 3.3018458197611293, + "grad_norm": 13.981231839826476, + "learning_rate": 5.464639206645104e-06, + "loss": 0.4123, + "step": 6082 + }, + { + "epoch": 3.302388707926167, + "grad_norm": 10.388999075429536, + "learning_rate": 5.461505146541147e-06, + "loss": 0.666, + "step": 6083 + }, + { + "epoch": 3.3029315960912053, + "grad_norm": 10.483727188037403, + "learning_rate": 5.4583716477654665e-06, + "loss": 0.4385, + "step": 6084 + }, + { + "epoch": 3.303474484256243, + "grad_norm": 15.004888765224539, + "learning_rate": 5.455238710705616e-06, + "loss": 0.6089, + "step": 6085 + }, + { + "epoch": 3.3040173724212814, + "grad_norm": 12.510560050118539, + "learning_rate": 5.452106335749092e-06, + "loss": 0.5481, + "step": 6086 + }, + { + "epoch": 3.304560260586319, + "grad_norm": 12.402037551065197, + "learning_rate": 5.448974523283297e-06, + "loss": 0.5909, + "step": 6087 + }, + { + "epoch": 3.3051031487513574, + "grad_norm": 9.456679882341124, + "learning_rate": 5.445843273695595e-06, + "loss": 0.4422, + "step": 6088 + }, + { + "epoch": 3.305646036916395, + "grad_norm": 12.165182627083702, + "learning_rate": 5.4427125873732455e-06, + "loss": 0.8632, + "step": 6089 + }, + { + "epoch": 3.3061889250814334, + "grad_norm": 10.815051939615287, + "learning_rate": 5.439582464703476e-06, + "loss": 0.5143, + "step": 6090 + }, + { + "epoch": 3.306731813246471, + "grad_norm": 8.659965055909717, + "learning_rate": 5.436452906073406e-06, + "loss": 0.3732, + "step": 6091 + }, + { + "epoch": 3.3072747014115094, + "grad_norm": 12.566885836457084, + "learning_rate": 5.433323911870119e-06, + "loss": 0.7708, + "step": 6092 + }, + { + "epoch": 3.307817589576547, + "grad_norm": 12.162201810127142, + "learning_rate": 5.430195482480611e-06, + "loss": 0.5414, + "step": 6093 + }, + { + "epoch": 3.3083604777415854, + "grad_norm": 10.50777462312628, + "learning_rate": 5.427067618291804e-06, + "loss": 0.4301, + "step": 6094 + }, + { + "epoch": 3.308903365906623, + "grad_norm": 10.640857438281406, + "learning_rate": 5.423940319690568e-06, + "loss": 0.4682, + "step": 6095 + }, + { + "epoch": 3.3094462540716614, + "grad_norm": 10.370610442117792, + "learning_rate": 5.420813587063678e-06, + "loss": 0.3056, + "step": 6096 + }, + { + "epoch": 3.309989142236699, + "grad_norm": 10.60821300635486, + "learning_rate": 5.417687420797867e-06, + "loss": 0.3344, + "step": 6097 + }, + { + "epoch": 3.3105320304017374, + "grad_norm": 10.353507117674113, + "learning_rate": 5.414561821279778e-06, + "loss": 0.656, + "step": 6098 + }, + { + "epoch": 3.311074918566775, + "grad_norm": 8.344962502930883, + "learning_rate": 5.411436788895992e-06, + "loss": 0.3413, + "step": 6099 + }, + { + "epoch": 3.3116178067318134, + "grad_norm": 9.483199213032712, + "learning_rate": 5.408312324033016e-06, + "loss": 0.4156, + "step": 6100 + }, + { + "epoch": 3.312160694896851, + "grad_norm": 10.849967748759967, + "learning_rate": 5.40518842707729e-06, + "loss": 0.3849, + "step": 6101 + }, + { + "epoch": 3.3127035830618894, + "grad_norm": 11.385869965644408, + "learning_rate": 5.402065098415188e-06, + "loss": 0.5767, + "step": 6102 + }, + { + "epoch": 3.313246471226927, + "grad_norm": 9.846646631682042, + "learning_rate": 5.398942338432993e-06, + "loss": 0.3517, + "step": 6103 + }, + { + "epoch": 3.3137893593919654, + "grad_norm": 12.208760557718977, + "learning_rate": 5.3958201475169455e-06, + "loss": 0.3989, + "step": 6104 + }, + { + "epoch": 3.314332247557003, + "grad_norm": 14.044852806162847, + "learning_rate": 5.392698526053203e-06, + "loss": 0.6161, + "step": 6105 + }, + { + "epoch": 3.3148751357220414, + "grad_norm": 16.302573867426716, + "learning_rate": 5.389577474427848e-06, + "loss": 0.5668, + "step": 6106 + }, + { + "epoch": 3.315418023887079, + "grad_norm": 15.241837519747031, + "learning_rate": 5.3864569930268986e-06, + "loss": 0.531, + "step": 6107 + }, + { + "epoch": 3.3159609120521174, + "grad_norm": 16.054606664926766, + "learning_rate": 5.383337082236302e-06, + "loss": 0.5549, + "step": 6108 + }, + { + "epoch": 3.316503800217155, + "grad_norm": 7.873315866697446, + "learning_rate": 5.3802177424419335e-06, + "loss": 0.3505, + "step": 6109 + }, + { + "epoch": 3.3170466883821934, + "grad_norm": 7.233865234327245, + "learning_rate": 5.377098974029595e-06, + "loss": 0.3402, + "step": 6110 + }, + { + "epoch": 3.317589576547231, + "grad_norm": 13.489019875241555, + "learning_rate": 5.373980777385026e-06, + "loss": 0.5829, + "step": 6111 + }, + { + "epoch": 3.3181324647122694, + "grad_norm": 8.944352234530756, + "learning_rate": 5.370863152893886e-06, + "loss": 0.2591, + "step": 6112 + }, + { + "epoch": 3.318675352877307, + "grad_norm": 11.421952434092379, + "learning_rate": 5.367746100941769e-06, + "loss": 0.4549, + "step": 6113 + }, + { + "epoch": 3.3192182410423454, + "grad_norm": 9.61822328391865, + "learning_rate": 5.364629621914197e-06, + "loss": 0.3805, + "step": 6114 + }, + { + "epoch": 3.319761129207383, + "grad_norm": 9.466765582290753, + "learning_rate": 5.3615137161966205e-06, + "loss": 0.2804, + "step": 6115 + }, + { + "epoch": 3.3203040173724214, + "grad_norm": 10.228565001712072, + "learning_rate": 5.358398384174421e-06, + "loss": 0.3802, + "step": 6116 + }, + { + "epoch": 3.320846905537459, + "grad_norm": 12.471667235628182, + "learning_rate": 5.355283626232907e-06, + "loss": 0.7176, + "step": 6117 + }, + { + "epoch": 3.3213897937024974, + "grad_norm": 8.698625687388546, + "learning_rate": 5.352169442757312e-06, + "loss": 0.2962, + "step": 6118 + }, + { + "epoch": 3.321932681867535, + "grad_norm": 14.02974515499815, + "learning_rate": 5.349055834132817e-06, + "loss": 0.4804, + "step": 6119 + }, + { + "epoch": 3.3224755700325734, + "grad_norm": 14.91877137041168, + "learning_rate": 5.345942800744499e-06, + "loss": 0.4992, + "step": 6120 + }, + { + "epoch": 3.323018458197611, + "grad_norm": 13.54006240651423, + "learning_rate": 5.342830342977403e-06, + "loss": 0.822, + "step": 6121 + }, + { + "epoch": 3.3235613463626494, + "grad_norm": 11.467004290914351, + "learning_rate": 5.3397184612164676e-06, + "loss": 0.4878, + "step": 6122 + }, + { + "epoch": 3.324104234527687, + "grad_norm": 11.864551909227606, + "learning_rate": 5.33660715584658e-06, + "loss": 0.5312, + "step": 6123 + }, + { + "epoch": 3.3246471226927254, + "grad_norm": 10.288196378126806, + "learning_rate": 5.333496427252551e-06, + "loss": 0.4778, + "step": 6124 + }, + { + "epoch": 3.3251900108577632, + "grad_norm": 11.010968624281599, + "learning_rate": 5.330386275819119e-06, + "loss": 0.4471, + "step": 6125 + }, + { + "epoch": 3.3257328990228014, + "grad_norm": 9.902175725807343, + "learning_rate": 5.327276701930961e-06, + "loss": 0.3581, + "step": 6126 + }, + { + "epoch": 3.3262757871878392, + "grad_norm": 14.462369688525524, + "learning_rate": 5.324167705972661e-06, + "loss": 0.4262, + "step": 6127 + }, + { + "epoch": 3.3268186753528775, + "grad_norm": 12.522646032715151, + "learning_rate": 5.32105928832876e-06, + "loss": 0.3605, + "step": 6128 + }, + { + "epoch": 3.3273615635179152, + "grad_norm": 8.804303263881113, + "learning_rate": 5.317951449383693e-06, + "loss": 0.2589, + "step": 6129 + }, + { + "epoch": 3.3279044516829535, + "grad_norm": 16.231015853377418, + "learning_rate": 5.314844189521859e-06, + "loss": 0.6517, + "step": 6130 + }, + { + "epoch": 3.3284473398479912, + "grad_norm": 15.588313322776349, + "learning_rate": 5.311737509127561e-06, + "loss": 0.6392, + "step": 6131 + }, + { + "epoch": 3.3289902280130295, + "grad_norm": 8.118329570429164, + "learning_rate": 5.308631408585041e-06, + "loss": 0.2433, + "step": 6132 + }, + { + "epoch": 3.3295331161780672, + "grad_norm": 17.927534810088048, + "learning_rate": 5.305525888278469e-06, + "loss": 0.8314, + "step": 6133 + }, + { + "epoch": 3.3300760043431055, + "grad_norm": 12.7719407908283, + "learning_rate": 5.302420948591929e-06, + "loss": 0.5422, + "step": 6134 + }, + { + "epoch": 3.3306188925081432, + "grad_norm": 13.715317083116945, + "learning_rate": 5.29931658990946e-06, + "loss": 0.4756, + "step": 6135 + }, + { + "epoch": 3.3311617806731815, + "grad_norm": 14.232906623385299, + "learning_rate": 5.296212812615001e-06, + "loss": 0.5507, + "step": 6136 + }, + { + "epoch": 3.3317046688382193, + "grad_norm": 11.07024486678308, + "learning_rate": 5.29310961709244e-06, + "loss": 0.4916, + "step": 6137 + }, + { + "epoch": 3.3322475570032575, + "grad_norm": 12.322287003987269, + "learning_rate": 5.290007003725585e-06, + "loss": 0.426, + "step": 6138 + }, + { + "epoch": 3.3327904451682953, + "grad_norm": 11.084747737931762, + "learning_rate": 5.286904972898168e-06, + "loss": 0.2768, + "step": 6139 + }, + { + "epoch": 3.3333333333333335, + "grad_norm": 12.758536777666697, + "learning_rate": 5.283803524993858e-06, + "loss": 0.9266, + "step": 6140 + }, + { + "epoch": 3.3338762214983713, + "grad_norm": 13.098970900751402, + "learning_rate": 5.280702660396243e-06, + "loss": 0.512, + "step": 6141 + }, + { + "epoch": 3.3344191096634095, + "grad_norm": 12.317106256281363, + "learning_rate": 5.277602379488844e-06, + "loss": 0.4155, + "step": 6142 + }, + { + "epoch": 3.3349619978284473, + "grad_norm": 11.38584068826573, + "learning_rate": 5.274502682655112e-06, + "loss": 0.4051, + "step": 6143 + }, + { + "epoch": 3.3355048859934855, + "grad_norm": 14.501004041816849, + "learning_rate": 5.271403570278417e-06, + "loss": 0.5548, + "step": 6144 + }, + { + "epoch": 3.3360477741585233, + "grad_norm": 12.998451360538299, + "learning_rate": 5.268305042742065e-06, + "loss": 0.4021, + "step": 6145 + }, + { + "epoch": 3.3365906623235615, + "grad_norm": 9.038839543671822, + "learning_rate": 5.2652071004292855e-06, + "loss": 0.2625, + "step": 6146 + }, + { + "epoch": 3.3371335504885993, + "grad_norm": 14.728843022444776, + "learning_rate": 5.26210974372324e-06, + "loss": 0.7124, + "step": 6147 + }, + { + "epoch": 3.3376764386536375, + "grad_norm": 12.24691457854214, + "learning_rate": 5.259012973007011e-06, + "loss": 0.3889, + "step": 6148 + }, + { + "epoch": 3.3382193268186753, + "grad_norm": 12.889746719808702, + "learning_rate": 5.255916788663614e-06, + "loss": 0.6392, + "step": 6149 + }, + { + "epoch": 3.3387622149837135, + "grad_norm": 15.16855584494826, + "learning_rate": 5.252821191075989e-06, + "loss": 0.6308, + "step": 6150 + }, + { + "epoch": 3.3393051031487513, + "grad_norm": 15.137292971607204, + "learning_rate": 5.249726180627006e-06, + "loss": 0.4732, + "step": 6151 + }, + { + "epoch": 3.3398479913137895, + "grad_norm": 12.827169669915257, + "learning_rate": 5.246631757699461e-06, + "loss": 0.4235, + "step": 6152 + }, + { + "epoch": 3.3403908794788273, + "grad_norm": 8.000573554523031, + "learning_rate": 5.243537922676074e-06, + "loss": 0.3355, + "step": 6153 + }, + { + "epoch": 3.3409337676438655, + "grad_norm": 11.071696970347325, + "learning_rate": 5.240444675939498e-06, + "loss": 0.3129, + "step": 6154 + }, + { + "epoch": 3.3414766558089033, + "grad_norm": 13.751963016664462, + "learning_rate": 5.237352017872308e-06, + "loss": 0.7001, + "step": 6155 + }, + { + "epoch": 3.3420195439739415, + "grad_norm": 11.736421365008567, + "learning_rate": 5.23425994885701e-06, + "loss": 0.5775, + "step": 6156 + }, + { + "epoch": 3.3425624321389793, + "grad_norm": 13.034138205582762, + "learning_rate": 5.231168469276044e-06, + "loss": 0.4829, + "step": 6157 + }, + { + "epoch": 3.3431053203040175, + "grad_norm": 13.897122730322542, + "learning_rate": 5.228077579511754e-06, + "loss": 0.6586, + "step": 6158 + }, + { + "epoch": 3.3436482084690553, + "grad_norm": 13.765426696215664, + "learning_rate": 5.224987279946441e-06, + "loss": 0.4817, + "step": 6159 + }, + { + "epoch": 3.3441910966340935, + "grad_norm": 13.446494398430909, + "learning_rate": 5.221897570962304e-06, + "loss": 0.5121, + "step": 6160 + }, + { + "epoch": 3.3447339847991313, + "grad_norm": 9.817713520136255, + "learning_rate": 5.2188084529415e-06, + "loss": 0.2662, + "step": 6161 + }, + { + "epoch": 3.3452768729641695, + "grad_norm": 10.025373895120913, + "learning_rate": 5.215719926266082e-06, + "loss": 0.4928, + "step": 6162 + }, + { + "epoch": 3.3458197611292073, + "grad_norm": 13.267267741014804, + "learning_rate": 5.212631991318044e-06, + "loss": 0.6133, + "step": 6163 + }, + { + "epoch": 3.3463626492942455, + "grad_norm": 13.654831265704226, + "learning_rate": 5.209544648479319e-06, + "loss": 0.4202, + "step": 6164 + }, + { + "epoch": 3.3469055374592833, + "grad_norm": 14.003058740046605, + "learning_rate": 5.20645789813174e-06, + "loss": 0.6629, + "step": 6165 + }, + { + "epoch": 3.3474484256243215, + "grad_norm": 8.761766797305249, + "learning_rate": 5.203371740657095e-06, + "loss": 0.3107, + "step": 6166 + }, + { + "epoch": 3.3479913137893593, + "grad_norm": 13.961944426146035, + "learning_rate": 5.2002861764370705e-06, + "loss": 0.4029, + "step": 6167 + }, + { + "epoch": 3.3485342019543975, + "grad_norm": 11.942844125447042, + "learning_rate": 5.1972012058533035e-06, + "loss": 0.2327, + "step": 6168 + }, + { + "epoch": 3.3490770901194353, + "grad_norm": 15.94835762092326, + "learning_rate": 5.194116829287348e-06, + "loss": 0.4844, + "step": 6169 + }, + { + "epoch": 3.3496199782844736, + "grad_norm": 14.492034902597197, + "learning_rate": 5.191033047120682e-06, + "loss": 0.7219, + "step": 6170 + }, + { + "epoch": 3.3501628664495113, + "grad_norm": 16.04200227643707, + "learning_rate": 5.187949859734715e-06, + "loss": 0.4706, + "step": 6171 + }, + { + "epoch": 3.3507057546145496, + "grad_norm": 16.558616346746106, + "learning_rate": 5.184867267510774e-06, + "loss": 0.6355, + "step": 6172 + }, + { + "epoch": 3.3512486427795873, + "grad_norm": 14.202794625614665, + "learning_rate": 5.18178527083013e-06, + "loss": 0.6307, + "step": 6173 + }, + { + "epoch": 3.3517915309446256, + "grad_norm": 13.034510344424431, + "learning_rate": 5.178703870073954e-06, + "loss": 0.614, + "step": 6174 + }, + { + "epoch": 3.3523344191096633, + "grad_norm": 8.768533900570704, + "learning_rate": 5.1756230656233715e-06, + "loss": 0.4036, + "step": 6175 + }, + { + "epoch": 3.3528773072747016, + "grad_norm": 16.580405700504947, + "learning_rate": 5.172542857859418e-06, + "loss": 0.5804, + "step": 6176 + }, + { + "epoch": 3.3534201954397393, + "grad_norm": 11.843457868265668, + "learning_rate": 5.169463247163058e-06, + "loss": 0.5303, + "step": 6177 + }, + { + "epoch": 3.3539630836047776, + "grad_norm": 12.879389378162015, + "learning_rate": 5.166384233915182e-06, + "loss": 0.5253, + "step": 6178 + }, + { + "epoch": 3.3545059717698154, + "grad_norm": 9.5479358372862, + "learning_rate": 5.163305818496607e-06, + "loss": 0.2689, + "step": 6179 + }, + { + "epoch": 3.3550488599348536, + "grad_norm": 11.148812089885375, + "learning_rate": 5.160228001288077e-06, + "loss": 0.5212, + "step": 6180 + }, + { + "epoch": 3.3555917480998914, + "grad_norm": 10.77047326599204, + "learning_rate": 5.157150782670261e-06, + "loss": 0.3674, + "step": 6181 + }, + { + "epoch": 3.3561346362649296, + "grad_norm": 11.804189327484117, + "learning_rate": 5.154074163023756e-06, + "loss": 0.5972, + "step": 6182 + }, + { + "epoch": 3.3566775244299674, + "grad_norm": 11.540083386779209, + "learning_rate": 5.15099814272908e-06, + "loss": 0.703, + "step": 6183 + }, + { + "epoch": 3.3572204125950056, + "grad_norm": 12.780871236150638, + "learning_rate": 5.147922722166683e-06, + "loss": 0.4786, + "step": 6184 + }, + { + "epoch": 3.3577633007600434, + "grad_norm": 11.170640555611453, + "learning_rate": 5.144847901716936e-06, + "loss": 0.6836, + "step": 6185 + }, + { + "epoch": 3.3583061889250816, + "grad_norm": 11.202376321881136, + "learning_rate": 5.1417736817601386e-06, + "loss": 0.327, + "step": 6186 + }, + { + "epoch": 3.3588490770901194, + "grad_norm": 10.62227316496751, + "learning_rate": 5.138700062676516e-06, + "loss": 0.4885, + "step": 6187 + }, + { + "epoch": 3.3593919652551576, + "grad_norm": 11.784247976226597, + "learning_rate": 5.135627044846216e-06, + "loss": 0.4551, + "step": 6188 + }, + { + "epoch": 3.3599348534201954, + "grad_norm": 9.425213682768682, + "learning_rate": 5.132554628649313e-06, + "loss": 0.3285, + "step": 6189 + }, + { + "epoch": 3.3604777415852336, + "grad_norm": 10.622928852300923, + "learning_rate": 5.1294828144658185e-06, + "loss": 0.4265, + "step": 6190 + }, + { + "epoch": 3.3610206297502714, + "grad_norm": 11.666591539214851, + "learning_rate": 5.126411602675649e-06, + "loss": 0.2681, + "step": 6191 + }, + { + "epoch": 3.3615635179153096, + "grad_norm": 19.598513215176627, + "learning_rate": 5.123340993658658e-06, + "loss": 0.7577, + "step": 6192 + }, + { + "epoch": 3.3621064060803474, + "grad_norm": 13.077915822970434, + "learning_rate": 5.120270987794627e-06, + "loss": 0.653, + "step": 6193 + }, + { + "epoch": 3.3626492942453856, + "grad_norm": 15.592283919989452, + "learning_rate": 5.117201585463256e-06, + "loss": 0.4488, + "step": 6194 + }, + { + "epoch": 3.3631921824104234, + "grad_norm": 9.288688824641968, + "learning_rate": 5.114132787044175e-06, + "loss": 0.4185, + "step": 6195 + }, + { + "epoch": 3.3637350705754616, + "grad_norm": 15.306973087542367, + "learning_rate": 5.111064592916935e-06, + "loss": 0.8354, + "step": 6196 + }, + { + "epoch": 3.3642779587404994, + "grad_norm": 11.499674684661942, + "learning_rate": 5.107997003461023e-06, + "loss": 0.5953, + "step": 6197 + }, + { + "epoch": 3.3648208469055376, + "grad_norm": 12.276252733631965, + "learning_rate": 5.104930019055834e-06, + "loss": 0.5574, + "step": 6198 + }, + { + "epoch": 3.3653637350705754, + "grad_norm": 15.0044996135686, + "learning_rate": 5.1018636400807075e-06, + "loss": 0.403, + "step": 6199 + }, + { + "epoch": 3.3659066232356136, + "grad_norm": 8.91478305391761, + "learning_rate": 5.098797866914889e-06, + "loss": 0.3447, + "step": 6200 + }, + { + "epoch": 3.3664495114006514, + "grad_norm": 12.707988441543339, + "learning_rate": 5.095732699937559e-06, + "loss": 0.6121, + "step": 6201 + }, + { + "epoch": 3.3669923995656896, + "grad_norm": 11.893526329599132, + "learning_rate": 5.092668139527831e-06, + "loss": 0.4349, + "step": 6202 + }, + { + "epoch": 3.3675352877307274, + "grad_norm": 9.573100623698757, + "learning_rate": 5.08960418606472e-06, + "loss": 0.3979, + "step": 6203 + }, + { + "epoch": 3.3680781758957656, + "grad_norm": 12.055211170267912, + "learning_rate": 5.0865408399271995e-06, + "loss": 0.7061, + "step": 6204 + }, + { + "epoch": 3.3686210640608034, + "grad_norm": 10.746677950946136, + "learning_rate": 5.08347810149413e-06, + "loss": 0.4687, + "step": 6205 + }, + { + "epoch": 3.3691639522258416, + "grad_norm": 11.592472853876425, + "learning_rate": 5.080415971144332e-06, + "loss": 0.4447, + "step": 6206 + }, + { + "epoch": 3.3697068403908794, + "grad_norm": 14.931359329572437, + "learning_rate": 5.077354449256521e-06, + "loss": 0.7571, + "step": 6207 + }, + { + "epoch": 3.3702497285559176, + "grad_norm": 11.016079444425253, + "learning_rate": 5.07429353620936e-06, + "loss": 0.3964, + "step": 6208 + }, + { + "epoch": 3.3707926167209554, + "grad_norm": 9.567910765955927, + "learning_rate": 5.071233232381425e-06, + "loss": 0.395, + "step": 6209 + }, + { + "epoch": 3.3713355048859937, + "grad_norm": 14.85470239623418, + "learning_rate": 5.0681735381512195e-06, + "loss": 0.557, + "step": 6210 + }, + { + "epoch": 3.3718783930510314, + "grad_norm": 11.383297704214762, + "learning_rate": 5.0651144538971746e-06, + "loss": 0.4522, + "step": 6211 + }, + { + "epoch": 3.3724212812160697, + "grad_norm": 13.445463967563244, + "learning_rate": 5.062055979997631e-06, + "loss": 0.5483, + "step": 6212 + }, + { + "epoch": 3.3729641693811074, + "grad_norm": 12.602441542764687, + "learning_rate": 5.058998116830878e-06, + "loss": 0.6498, + "step": 6213 + }, + { + "epoch": 3.3735070575461457, + "grad_norm": 11.110392326995717, + "learning_rate": 5.055940864775113e-06, + "loss": 0.4893, + "step": 6214 + }, + { + "epoch": 3.3740499457111834, + "grad_norm": 12.47623770525584, + "learning_rate": 5.052884224208461e-06, + "loss": 0.4237, + "step": 6215 + }, + { + "epoch": 3.3745928338762217, + "grad_norm": 9.744420395590724, + "learning_rate": 5.049828195508972e-06, + "loss": 0.2669, + "step": 6216 + }, + { + "epoch": 3.3751357220412594, + "grad_norm": 15.871527506273058, + "learning_rate": 5.046772779054622e-06, + "loss": 0.4374, + "step": 6217 + }, + { + "epoch": 3.3756786102062977, + "grad_norm": 9.442731773451024, + "learning_rate": 5.043717975223308e-06, + "loss": 0.3363, + "step": 6218 + }, + { + "epoch": 3.3762214983713354, + "grad_norm": 9.122598498293966, + "learning_rate": 5.040663784392855e-06, + "loss": 0.366, + "step": 6219 + }, + { + "epoch": 3.3767643865363737, + "grad_norm": 11.713343466114539, + "learning_rate": 5.037610206941009e-06, + "loss": 0.4296, + "step": 6220 + }, + { + "epoch": 3.3773072747014115, + "grad_norm": 9.906918527878904, + "learning_rate": 5.034557243245441e-06, + "loss": 0.3189, + "step": 6221 + }, + { + "epoch": 3.3778501628664497, + "grad_norm": 12.91483922171168, + "learning_rate": 5.031504893683748e-06, + "loss": 0.4095, + "step": 6222 + }, + { + "epoch": 3.3783930510314875, + "grad_norm": 12.288487294242882, + "learning_rate": 5.028453158633448e-06, + "loss": 0.5457, + "step": 6223 + }, + { + "epoch": 3.3789359391965257, + "grad_norm": 8.504130528670414, + "learning_rate": 5.025402038471984e-06, + "loss": 0.3373, + "step": 6224 + }, + { + "epoch": 3.3794788273615635, + "grad_norm": 13.190667855419926, + "learning_rate": 5.022351533576725e-06, + "loss": 0.4901, + "step": 6225 + }, + { + "epoch": 3.3800217155266017, + "grad_norm": 15.552140760061034, + "learning_rate": 5.019301644324961e-06, + "loss": 1.0988, + "step": 6226 + }, + { + "epoch": 3.3805646036916395, + "grad_norm": 9.04502851979091, + "learning_rate": 5.016252371093904e-06, + "loss": 0.3088, + "step": 6227 + }, + { + "epoch": 3.3811074918566777, + "grad_norm": 10.595816125042607, + "learning_rate": 5.0132037142607035e-06, + "loss": 0.4645, + "step": 6228 + }, + { + "epoch": 3.3816503800217155, + "grad_norm": 12.60889622464924, + "learning_rate": 5.010155674202409e-06, + "loss": 0.4484, + "step": 6229 + }, + { + "epoch": 3.3821932681867537, + "grad_norm": 11.424705925472448, + "learning_rate": 5.00710825129602e-06, + "loss": 0.3788, + "step": 6230 + }, + { + "epoch": 3.3827361563517915, + "grad_norm": 12.859629128090067, + "learning_rate": 5.004061445918438e-06, + "loss": 0.6222, + "step": 6231 + }, + { + "epoch": 3.3832790445168297, + "grad_norm": 9.436094386037109, + "learning_rate": 5.001015258446497e-06, + "loss": 0.4005, + "step": 6232 + }, + { + "epoch": 3.3838219326818675, + "grad_norm": 13.270237716590554, + "learning_rate": 4.997969689256957e-06, + "loss": 0.5421, + "step": 6233 + }, + { + "epoch": 3.3843648208469057, + "grad_norm": 8.823605358833275, + "learning_rate": 4.994924738726493e-06, + "loss": 0.3196, + "step": 6234 + }, + { + "epoch": 3.3849077090119435, + "grad_norm": 15.35267771565606, + "learning_rate": 4.991880407231722e-06, + "loss": 0.7645, + "step": 6235 + }, + { + "epoch": 3.3854505971769817, + "grad_norm": 12.948816197321962, + "learning_rate": 4.988836695149156e-06, + "loss": 0.4939, + "step": 6236 + }, + { + "epoch": 3.3859934853420195, + "grad_norm": 11.08333997503354, + "learning_rate": 4.985793602855264e-06, + "loss": 0.386, + "step": 6237 + }, + { + "epoch": 3.3865363735070577, + "grad_norm": 11.954902260666392, + "learning_rate": 4.9827511307264006e-06, + "loss": 0.3809, + "step": 6238 + }, + { + "epoch": 3.3870792616720955, + "grad_norm": 13.826055371275897, + "learning_rate": 4.979709279138879e-06, + "loss": 0.488, + "step": 6239 + }, + { + "epoch": 3.3876221498371337, + "grad_norm": 12.084219110446965, + "learning_rate": 4.976668048468918e-06, + "loss": 0.5239, + "step": 6240 + }, + { + "epoch": 3.3881650380021715, + "grad_norm": 10.947811118815906, + "learning_rate": 4.973627439092651e-06, + "loss": 0.3714, + "step": 6241 + }, + { + "epoch": 3.3887079261672097, + "grad_norm": 12.522059184173699, + "learning_rate": 4.97058745138616e-06, + "loss": 0.689, + "step": 6242 + }, + { + "epoch": 3.3892508143322475, + "grad_norm": 10.527706094398768, + "learning_rate": 4.967548085725423e-06, + "loss": 0.6032, + "step": 6243 + }, + { + "epoch": 3.3897937024972857, + "grad_norm": 13.88984081976927, + "learning_rate": 4.964509342486365e-06, + "loss": 0.7107, + "step": 6244 + }, + { + "epoch": 3.3903365906623235, + "grad_norm": 12.254968564513517, + "learning_rate": 4.961471222044811e-06, + "loss": 0.5961, + "step": 6245 + }, + { + "epoch": 3.3908794788273617, + "grad_norm": 14.460789356302293, + "learning_rate": 4.95843372477653e-06, + "loss": 0.4468, + "step": 6246 + }, + { + "epoch": 3.3914223669923995, + "grad_norm": 11.641301023056258, + "learning_rate": 4.955396851057201e-06, + "loss": 0.4814, + "step": 6247 + }, + { + "epoch": 3.3919652551574377, + "grad_norm": 16.08871509775364, + "learning_rate": 4.9523606012624285e-06, + "loss": 0.6589, + "step": 6248 + }, + { + "epoch": 3.3925081433224755, + "grad_norm": 10.999987937636945, + "learning_rate": 4.9493249757677454e-06, + "loss": 0.4282, + "step": 6249 + }, + { + "epoch": 3.3930510314875137, + "grad_norm": 13.140914544731421, + "learning_rate": 4.946289974948591e-06, + "loss": 0.5447, + "step": 6250 + }, + { + "epoch": 3.3935939196525515, + "grad_norm": 16.948288206914583, + "learning_rate": 4.943255599180352e-06, + "loss": 0.3733, + "step": 6251 + }, + { + "epoch": 3.3941368078175898, + "grad_norm": 13.872557767694476, + "learning_rate": 4.940221848838319e-06, + "loss": 0.407, + "step": 6252 + }, + { + "epoch": 3.3946796959826275, + "grad_norm": 12.01068343710924, + "learning_rate": 4.937188724297713e-06, + "loss": 0.3866, + "step": 6253 + }, + { + "epoch": 3.3952225841476658, + "grad_norm": 12.980406743985375, + "learning_rate": 4.934156225933673e-06, + "loss": 0.6519, + "step": 6254 + }, + { + "epoch": 3.3957654723127035, + "grad_norm": 11.103766528107464, + "learning_rate": 4.931124354121265e-06, + "loss": 0.2982, + "step": 6255 + }, + { + "epoch": 3.3963083604777418, + "grad_norm": 7.2969268723114675, + "learning_rate": 4.928093109235476e-06, + "loss": 0.3553, + "step": 6256 + }, + { + "epoch": 3.3968512486427795, + "grad_norm": 8.742701237748966, + "learning_rate": 4.925062491651213e-06, + "loss": 0.3673, + "step": 6257 + }, + { + "epoch": 3.3973941368078178, + "grad_norm": 11.343078244137898, + "learning_rate": 4.922032501743311e-06, + "loss": 0.5849, + "step": 6258 + }, + { + "epoch": 3.3979370249728555, + "grad_norm": 10.498271159334314, + "learning_rate": 4.919003139886522e-06, + "loss": 0.4092, + "step": 6259 + }, + { + "epoch": 3.3984799131378938, + "grad_norm": 10.97821864845177, + "learning_rate": 4.915974406455522e-06, + "loss": 0.3967, + "step": 6260 + }, + { + "epoch": 3.3990228013029316, + "grad_norm": 13.345358070056497, + "learning_rate": 4.912946301824911e-06, + "loss": 0.61, + "step": 6261 + }, + { + "epoch": 3.3995656894679698, + "grad_norm": 11.235526091701708, + "learning_rate": 4.909918826369209e-06, + "loss": 0.577, + "step": 6262 + }, + { + "epoch": 3.4001085776330076, + "grad_norm": 11.63018111198774, + "learning_rate": 4.9068919804628575e-06, + "loss": 0.5611, + "step": 6263 + }, + { + "epoch": 3.400651465798046, + "grad_norm": 14.484517727976659, + "learning_rate": 4.903865764480224e-06, + "loss": 0.8804, + "step": 6264 + }, + { + "epoch": 3.4011943539630836, + "grad_norm": 13.063076605432173, + "learning_rate": 4.9008401787955964e-06, + "loss": 0.6872, + "step": 6265 + }, + { + "epoch": 3.401737242128122, + "grad_norm": 10.26193448056392, + "learning_rate": 4.89781522378318e-06, + "loss": 0.4086, + "step": 6266 + }, + { + "epoch": 3.4022801302931596, + "grad_norm": 14.7618293483203, + "learning_rate": 4.894790899817106e-06, + "loss": 0.5349, + "step": 6267 + }, + { + "epoch": 3.402823018458198, + "grad_norm": 15.246141002277422, + "learning_rate": 4.8917672072714364e-06, + "loss": 0.4647, + "step": 6268 + }, + { + "epoch": 3.4033659066232356, + "grad_norm": 12.999965847014597, + "learning_rate": 4.888744146520137e-06, + "loss": 0.6288, + "step": 6269 + }, + { + "epoch": 3.403908794788274, + "grad_norm": 12.403108581616499, + "learning_rate": 4.885721717937106e-06, + "loss": 0.4168, + "step": 6270 + }, + { + "epoch": 3.4044516829533116, + "grad_norm": 9.010405275529319, + "learning_rate": 4.882699921896166e-06, + "loss": 0.2848, + "step": 6271 + }, + { + "epoch": 3.40499457111835, + "grad_norm": 9.731197165536528, + "learning_rate": 4.87967875877105e-06, + "loss": 0.4276, + "step": 6272 + }, + { + "epoch": 3.4055374592833876, + "grad_norm": 16.141362422365116, + "learning_rate": 4.876658228935434e-06, + "loss": 0.425, + "step": 6273 + }, + { + "epoch": 3.406080347448426, + "grad_norm": 10.09711287157442, + "learning_rate": 4.873638332762887e-06, + "loss": 0.6085, + "step": 6274 + }, + { + "epoch": 3.4066232356134636, + "grad_norm": 15.877829530437113, + "learning_rate": 4.8706190706269276e-06, + "loss": 0.7737, + "step": 6275 + }, + { + "epoch": 3.407166123778502, + "grad_norm": 16.848740635164727, + "learning_rate": 4.867600442900969e-06, + "loss": 0.9117, + "step": 6276 + }, + { + "epoch": 3.4077090119435396, + "grad_norm": 12.439996859861845, + "learning_rate": 4.8645824499583764e-06, + "loss": 0.6655, + "step": 6277 + }, + { + "epoch": 3.408251900108578, + "grad_norm": 15.742400137527083, + "learning_rate": 4.861565092172402e-06, + "loss": 0.4276, + "step": 6278 + }, + { + "epoch": 3.4087947882736156, + "grad_norm": 10.574321270699965, + "learning_rate": 4.8585483699162505e-06, + "loss": 0.3849, + "step": 6279 + }, + { + "epoch": 3.409337676438654, + "grad_norm": 14.691129894472994, + "learning_rate": 4.8555322835630345e-06, + "loss": 0.8061, + "step": 6280 + }, + { + "epoch": 3.4098805646036916, + "grad_norm": 15.252415938946724, + "learning_rate": 4.852516833485778e-06, + "loss": 0.8019, + "step": 6281 + }, + { + "epoch": 3.41042345276873, + "grad_norm": 15.141237929554537, + "learning_rate": 4.849502020057449e-06, + "loss": 0.8201, + "step": 6282 + }, + { + "epoch": 3.4109663409337676, + "grad_norm": 12.751756702153507, + "learning_rate": 4.846487843650914e-06, + "loss": 0.5871, + "step": 6283 + }, + { + "epoch": 3.411509229098806, + "grad_norm": 10.510847411766605, + "learning_rate": 4.843474304638977e-06, + "loss": 0.5543, + "step": 6284 + }, + { + "epoch": 3.4120521172638436, + "grad_norm": 11.747946589267375, + "learning_rate": 4.8404614033943586e-06, + "loss": 0.5981, + "step": 6285 + }, + { + "epoch": 3.412595005428882, + "grad_norm": 10.03814207660896, + "learning_rate": 4.837449140289696e-06, + "loss": 0.3838, + "step": 6286 + }, + { + "epoch": 3.4131378935939196, + "grad_norm": 9.955014298679115, + "learning_rate": 4.8344375156975525e-06, + "loss": 0.4807, + "step": 6287 + }, + { + "epoch": 3.413680781758958, + "grad_norm": 16.70257123694898, + "learning_rate": 4.8314265299904085e-06, + "loss": 0.5431, + "step": 6288 + }, + { + "epoch": 3.4142236699239956, + "grad_norm": 12.494957483834643, + "learning_rate": 4.828416183540668e-06, + "loss": 0.428, + "step": 6289 + }, + { + "epoch": 3.414766558089034, + "grad_norm": 13.556838718061059, + "learning_rate": 4.825406476720658e-06, + "loss": 0.5487, + "step": 6290 + }, + { + "epoch": 3.4153094462540716, + "grad_norm": 11.716720789445835, + "learning_rate": 4.822397409902622e-06, + "loss": 0.8306, + "step": 6291 + }, + { + "epoch": 3.41585233441911, + "grad_norm": 18.495547243648275, + "learning_rate": 4.819388983458725e-06, + "loss": 1.0386, + "step": 6292 + }, + { + "epoch": 3.4163952225841476, + "grad_norm": 11.600851672889053, + "learning_rate": 4.816381197761055e-06, + "loss": 0.3111, + "step": 6293 + }, + { + "epoch": 3.416938110749186, + "grad_norm": 10.513433670937403, + "learning_rate": 4.813374053181621e-06, + "loss": 0.3881, + "step": 6294 + }, + { + "epoch": 3.4174809989142236, + "grad_norm": 12.120445230198527, + "learning_rate": 4.810367550092349e-06, + "loss": 0.6089, + "step": 6295 + }, + { + "epoch": 3.418023887079262, + "grad_norm": 14.617438317762286, + "learning_rate": 4.807361688865091e-06, + "loss": 0.725, + "step": 6296 + }, + { + "epoch": 3.4185667752442996, + "grad_norm": 13.459119769065767, + "learning_rate": 4.804356469871615e-06, + "loss": 0.678, + "step": 6297 + }, + { + "epoch": 3.419109663409338, + "grad_norm": 11.259674657240234, + "learning_rate": 4.801351893483611e-06, + "loss": 0.4629, + "step": 6298 + }, + { + "epoch": 3.4196525515743756, + "grad_norm": 12.169301874189793, + "learning_rate": 4.7983479600726904e-06, + "loss": 0.6371, + "step": 6299 + }, + { + "epoch": 3.420195439739414, + "grad_norm": 10.211316992248761, + "learning_rate": 4.795344670010385e-06, + "loss": 0.591, + "step": 6300 + }, + { + "epoch": 3.4207383279044516, + "grad_norm": 10.221898080949522, + "learning_rate": 4.792342023668144e-06, + "loss": 0.4058, + "step": 6301 + }, + { + "epoch": 3.42128121606949, + "grad_norm": 11.161796287973816, + "learning_rate": 4.789340021417343e-06, + "loss": 0.6044, + "step": 6302 + }, + { + "epoch": 3.4218241042345277, + "grad_norm": 13.57180649881648, + "learning_rate": 4.7863386636292705e-06, + "loss": 0.5895, + "step": 6303 + }, + { + "epoch": 3.422366992399566, + "grad_norm": 14.249903450132871, + "learning_rate": 4.783337950675143e-06, + "loss": 0.9115, + "step": 6304 + }, + { + "epoch": 3.4229098805646037, + "grad_norm": 8.778670420410771, + "learning_rate": 4.780337882926088e-06, + "loss": 0.368, + "step": 6305 + }, + { + "epoch": 3.423452768729642, + "grad_norm": 14.418437807091719, + "learning_rate": 4.77733846075317e-06, + "loss": 0.7248, + "step": 6306 + }, + { + "epoch": 3.4239956568946797, + "grad_norm": 12.207383551156253, + "learning_rate": 4.774339684527348e-06, + "loss": 0.715, + "step": 6307 + }, + { + "epoch": 3.424538545059718, + "grad_norm": 12.625928846965705, + "learning_rate": 4.7713415546195285e-06, + "loss": 0.4691, + "step": 6308 + }, + { + "epoch": 3.4250814332247557, + "grad_norm": 12.54338617690595, + "learning_rate": 4.768344071400516e-06, + "loss": 0.7746, + "step": 6309 + }, + { + "epoch": 3.425624321389794, + "grad_norm": 12.757091279673624, + "learning_rate": 4.765347235241042e-06, + "loss": 0.5129, + "step": 6310 + }, + { + "epoch": 3.4261672095548317, + "grad_norm": 14.719338319419373, + "learning_rate": 4.762351046511774e-06, + "loss": 0.6281, + "step": 6311 + }, + { + "epoch": 3.42671009771987, + "grad_norm": 14.600430485134494, + "learning_rate": 4.759355505583267e-06, + "loss": 0.6332, + "step": 6312 + }, + { + "epoch": 3.4272529858849077, + "grad_norm": 12.499065575741712, + "learning_rate": 4.756360612826032e-06, + "loss": 0.5188, + "step": 6313 + }, + { + "epoch": 3.427795874049946, + "grad_norm": 13.595282233685417, + "learning_rate": 4.753366368610466e-06, + "loss": 0.6934, + "step": 6314 + }, + { + "epoch": 3.4283387622149837, + "grad_norm": 18.513402859066886, + "learning_rate": 4.750372773306916e-06, + "loss": 0.8336, + "step": 6315 + }, + { + "epoch": 3.428881650380022, + "grad_norm": 12.541674490834746, + "learning_rate": 4.747379827285621e-06, + "loss": 0.4741, + "step": 6316 + }, + { + "epoch": 3.4294245385450597, + "grad_norm": 13.06139830486867, + "learning_rate": 4.744387530916764e-06, + "loss": 0.6489, + "step": 6317 + }, + { + "epoch": 3.429967426710098, + "grad_norm": 9.57665402322532, + "learning_rate": 4.741395884570437e-06, + "loss": 0.4976, + "step": 6318 + }, + { + "epoch": 3.4305103148751357, + "grad_norm": 9.338701533011545, + "learning_rate": 4.738404888616641e-06, + "loss": 0.4947, + "step": 6319 + }, + { + "epoch": 3.431053203040174, + "grad_norm": 13.17347762890872, + "learning_rate": 4.735414543425321e-06, + "loss": 0.5455, + "step": 6320 + }, + { + "epoch": 3.4315960912052117, + "grad_norm": 14.408997044186131, + "learning_rate": 4.732424849366314e-06, + "loss": 0.5761, + "step": 6321 + }, + { + "epoch": 3.43213897937025, + "grad_norm": 10.201121637580389, + "learning_rate": 4.729435806809401e-06, + "loss": 0.4271, + "step": 6322 + }, + { + "epoch": 3.4326818675352877, + "grad_norm": 11.714865692839483, + "learning_rate": 4.726447416124266e-06, + "loss": 0.4182, + "step": 6323 + }, + { + "epoch": 3.433224755700326, + "grad_norm": 21.671744658819392, + "learning_rate": 4.72345967768052e-06, + "loss": 1.1518, + "step": 6324 + }, + { + "epoch": 3.4337676438653637, + "grad_norm": 10.904130560504068, + "learning_rate": 4.72047259184769e-06, + "loss": 0.533, + "step": 6325 + }, + { + "epoch": 3.434310532030402, + "grad_norm": 17.26082351569084, + "learning_rate": 4.717486158995225e-06, + "loss": 0.5579, + "step": 6326 + }, + { + "epoch": 3.4348534201954397, + "grad_norm": 13.562029451071414, + "learning_rate": 4.7145003794924905e-06, + "loss": 0.6275, + "step": 6327 + }, + { + "epoch": 3.435396308360478, + "grad_norm": 10.679128408418185, + "learning_rate": 4.711515253708774e-06, + "loss": 0.3116, + "step": 6328 + }, + { + "epoch": 3.4359391965255157, + "grad_norm": 6.813829973820327, + "learning_rate": 4.708530782013277e-06, + "loss": 0.2313, + "step": 6329 + }, + { + "epoch": 3.436482084690554, + "grad_norm": 10.696300631140947, + "learning_rate": 4.705546964775128e-06, + "loss": 0.4345, + "step": 6330 + }, + { + "epoch": 3.4370249728555917, + "grad_norm": 15.665510743577896, + "learning_rate": 4.702563802363369e-06, + "loss": 0.6155, + "step": 6331 + }, + { + "epoch": 3.4375678610206295, + "grad_norm": 9.437457272265185, + "learning_rate": 4.699581295146961e-06, + "loss": 0.3972, + "step": 6332 + }, + { + "epoch": 3.4381107491856677, + "grad_norm": 11.253233608357364, + "learning_rate": 4.696599443494787e-06, + "loss": 0.4871, + "step": 6333 + }, + { + "epoch": 3.438653637350706, + "grad_norm": 13.323571403470238, + "learning_rate": 4.693618247775645e-06, + "loss": 0.4443, + "step": 6334 + }, + { + "epoch": 3.4391965255157437, + "grad_norm": 10.853976251462788, + "learning_rate": 4.6906377083582556e-06, + "loss": 0.4012, + "step": 6335 + }, + { + "epoch": 3.4397394136807815, + "grad_norm": 11.483324693854579, + "learning_rate": 4.687657825611256e-06, + "loss": 0.4913, + "step": 6336 + }, + { + "epoch": 3.4402823018458197, + "grad_norm": 16.90515830826938, + "learning_rate": 4.684678599903204e-06, + "loss": 0.5872, + "step": 6337 + }, + { + "epoch": 3.440825190010858, + "grad_norm": 11.339161987387357, + "learning_rate": 4.681700031602573e-06, + "loss": 0.484, + "step": 6338 + }, + { + "epoch": 3.4413680781758957, + "grad_norm": 13.116517374892055, + "learning_rate": 4.678722121077759e-06, + "loss": 0.8588, + "step": 6339 + }, + { + "epoch": 3.4419109663409335, + "grad_norm": 10.313990325741717, + "learning_rate": 4.675744868697073e-06, + "loss": 0.4864, + "step": 6340 + }, + { + "epoch": 3.4424538545059717, + "grad_norm": 12.464768763015064, + "learning_rate": 4.672768274828748e-06, + "loss": 0.7793, + "step": 6341 + }, + { + "epoch": 3.44299674267101, + "grad_norm": 7.491565779729114, + "learning_rate": 4.669792339840933e-06, + "loss": 0.2551, + "step": 6342 + }, + { + "epoch": 3.4435396308360477, + "grad_norm": 11.582532266159841, + "learning_rate": 4.666817064101693e-06, + "loss": 0.4687, + "step": 6343 + }, + { + "epoch": 3.4440825190010855, + "grad_norm": 13.223648734250315, + "learning_rate": 4.663842447979026e-06, + "loss": 0.5676, + "step": 6344 + }, + { + "epoch": 3.4446254071661238, + "grad_norm": 9.420907388669441, + "learning_rate": 4.660868491840821e-06, + "loss": 0.4091, + "step": 6345 + }, + { + "epoch": 3.445168295331162, + "grad_norm": 10.937341000869324, + "learning_rate": 4.657895196054919e-06, + "loss": 0.4299, + "step": 6346 + }, + { + "epoch": 3.4457111834961998, + "grad_norm": 10.997320970308403, + "learning_rate": 4.654922560989049e-06, + "loss": 0.4176, + "step": 6347 + }, + { + "epoch": 3.4462540716612375, + "grad_norm": 11.008951815202312, + "learning_rate": 4.651950587010875e-06, + "loss": 0.3343, + "step": 6348 + }, + { + "epoch": 3.4467969598262758, + "grad_norm": 13.714805717613569, + "learning_rate": 4.6489792744879755e-06, + "loss": 0.8672, + "step": 6349 + }, + { + "epoch": 3.447339847991314, + "grad_norm": 12.722366469633661, + "learning_rate": 4.646008623787845e-06, + "loss": 0.601, + "step": 6350 + }, + { + "epoch": 3.4478827361563518, + "grad_norm": 10.3662276435613, + "learning_rate": 4.643038635277908e-06, + "loss": 0.4597, + "step": 6351 + }, + { + "epoch": 3.4484256243213895, + "grad_norm": 11.28259690900768, + "learning_rate": 4.640069309325484e-06, + "loss": 0.4909, + "step": 6352 + }, + { + "epoch": 3.4489685124864278, + "grad_norm": 15.163346859296523, + "learning_rate": 4.6371006462978355e-06, + "loss": 0.8422, + "step": 6353 + }, + { + "epoch": 3.449511400651466, + "grad_norm": 9.625566221898923, + "learning_rate": 4.634132646562119e-06, + "loss": 0.2858, + "step": 6354 + }, + { + "epoch": 3.450054288816504, + "grad_norm": 14.927715657666415, + "learning_rate": 4.631165310485434e-06, + "loss": 0.5556, + "step": 6355 + }, + { + "epoch": 3.4505971769815416, + "grad_norm": 14.575103603282974, + "learning_rate": 4.62819863843478e-06, + "loss": 0.5422, + "step": 6356 + }, + { + "epoch": 3.45114006514658, + "grad_norm": 11.84763157429604, + "learning_rate": 4.625232630777079e-06, + "loss": 0.5382, + "step": 6357 + }, + { + "epoch": 3.451682953311618, + "grad_norm": 14.334352386311854, + "learning_rate": 4.622267287879176e-06, + "loss": 0.6771, + "step": 6358 + }, + { + "epoch": 3.452225841476656, + "grad_norm": 13.03121716154961, + "learning_rate": 4.619302610107819e-06, + "loss": 0.702, + "step": 6359 + }, + { + "epoch": 3.4527687296416936, + "grad_norm": 9.331217131822793, + "learning_rate": 4.616338597829697e-06, + "loss": 0.4664, + "step": 6360 + }, + { + "epoch": 3.453311617806732, + "grad_norm": 16.384420267273573, + "learning_rate": 4.61337525141139e-06, + "loss": 1.1202, + "step": 6361 + }, + { + "epoch": 3.45385450597177, + "grad_norm": 12.656834373054709, + "learning_rate": 4.610412571219421e-06, + "loss": 0.8405, + "step": 6362 + }, + { + "epoch": 3.454397394136808, + "grad_norm": 10.6285388663689, + "learning_rate": 4.607450557620216e-06, + "loss": 0.4027, + "step": 6363 + }, + { + "epoch": 3.4549402823018456, + "grad_norm": 14.149860401799353, + "learning_rate": 4.60448921098012e-06, + "loss": 0.5762, + "step": 6364 + }, + { + "epoch": 3.455483170466884, + "grad_norm": 15.249693808816065, + "learning_rate": 4.601528531665397e-06, + "loss": 1.0724, + "step": 6365 + }, + { + "epoch": 3.456026058631922, + "grad_norm": 11.967042611547605, + "learning_rate": 4.598568520042229e-06, + "loss": 0.572, + "step": 6366 + }, + { + "epoch": 3.45656894679696, + "grad_norm": 13.806580531191415, + "learning_rate": 4.595609176476715e-06, + "loss": 0.7437, + "step": 6367 + }, + { + "epoch": 3.4571118349619976, + "grad_norm": 15.062073427987208, + "learning_rate": 4.592650501334872e-06, + "loss": 0.6357, + "step": 6368 + }, + { + "epoch": 3.457654723127036, + "grad_norm": 8.804207344489894, + "learning_rate": 4.589692494982632e-06, + "loss": 0.3561, + "step": 6369 + }, + { + "epoch": 3.458197611292074, + "grad_norm": 13.940651501182272, + "learning_rate": 4.5867351577858475e-06, + "loss": 0.6055, + "step": 6370 + }, + { + "epoch": 3.458740499457112, + "grad_norm": 12.874778392486414, + "learning_rate": 4.583778490110287e-06, + "loss": 0.466, + "step": 6371 + }, + { + "epoch": 3.4592833876221496, + "grad_norm": 13.362591165530915, + "learning_rate": 4.580822492321634e-06, + "loss": 0.837, + "step": 6372 + }, + { + "epoch": 3.459826275787188, + "grad_norm": 13.501338998690153, + "learning_rate": 4.577867164785492e-06, + "loss": 0.4919, + "step": 6373 + }, + { + "epoch": 3.460369163952226, + "grad_norm": 8.394152440486542, + "learning_rate": 4.574912507867382e-06, + "loss": 0.3555, + "step": 6374 + }, + { + "epoch": 3.460912052117264, + "grad_norm": 11.611965148146934, + "learning_rate": 4.571958521932738e-06, + "loss": 0.4808, + "step": 6375 + }, + { + "epoch": 3.4614549402823016, + "grad_norm": 11.297446435669148, + "learning_rate": 4.569005207346911e-06, + "loss": 0.4093, + "step": 6376 + }, + { + "epoch": 3.46199782844734, + "grad_norm": 13.040685712021983, + "learning_rate": 4.566052564475184e-06, + "loss": 0.5487, + "step": 6377 + }, + { + "epoch": 3.462540716612378, + "grad_norm": 10.160191141450042, + "learning_rate": 4.563100593682732e-06, + "loss": 0.5555, + "step": 6378 + }, + { + "epoch": 3.463083604777416, + "grad_norm": 10.641601715805423, + "learning_rate": 4.560149295334664e-06, + "loss": 0.4319, + "step": 6379 + }, + { + "epoch": 3.4636264929424536, + "grad_norm": 11.742982035764232, + "learning_rate": 4.557198669796001e-06, + "loss": 0.77, + "step": 6380 + }, + { + "epoch": 3.464169381107492, + "grad_norm": 15.297545906946976, + "learning_rate": 4.554248717431678e-06, + "loss": 0.753, + "step": 6381 + }, + { + "epoch": 3.46471226927253, + "grad_norm": 11.061047467202666, + "learning_rate": 4.551299438606559e-06, + "loss": 0.428, + "step": 6382 + }, + { + "epoch": 3.465255157437568, + "grad_norm": 12.238086621939612, + "learning_rate": 4.548350833685402e-06, + "loss": 0.4874, + "step": 6383 + }, + { + "epoch": 3.4657980456026056, + "grad_norm": 13.316349313453275, + "learning_rate": 4.54540290303291e-06, + "loss": 0.7391, + "step": 6384 + }, + { + "epoch": 3.466340933767644, + "grad_norm": 14.524029697923662, + "learning_rate": 4.5424556470136735e-06, + "loss": 0.6782, + "step": 6385 + }, + { + "epoch": 3.466883821932682, + "grad_norm": 11.699705038001243, + "learning_rate": 4.539509065992229e-06, + "loss": 0.6561, + "step": 6386 + }, + { + "epoch": 3.46742671009772, + "grad_norm": 16.5251401693806, + "learning_rate": 4.536563160333001e-06, + "loss": 0.6788, + "step": 6387 + }, + { + "epoch": 3.4679695982627576, + "grad_norm": 8.228129211788293, + "learning_rate": 4.533617930400345e-06, + "loss": 0.4383, + "step": 6388 + }, + { + "epoch": 3.468512486427796, + "grad_norm": 10.016870671639179, + "learning_rate": 4.530673376558543e-06, + "loss": 0.5374, + "step": 6389 + }, + { + "epoch": 3.469055374592834, + "grad_norm": 12.529286864818076, + "learning_rate": 4.527729499171767e-06, + "loss": 0.3919, + "step": 6390 + }, + { + "epoch": 3.469598262757872, + "grad_norm": 13.31372906596201, + "learning_rate": 4.524786298604136e-06, + "loss": 0.7436, + "step": 6391 + }, + { + "epoch": 3.4701411509229096, + "grad_norm": 11.420174160084546, + "learning_rate": 4.521843775219654e-06, + "loss": 0.4285, + "step": 6392 + }, + { + "epoch": 3.470684039087948, + "grad_norm": 11.461560335128281, + "learning_rate": 4.518901929382267e-06, + "loss": 0.5651, + "step": 6393 + }, + { + "epoch": 3.471226927252986, + "grad_norm": 14.298269585226816, + "learning_rate": 4.515960761455827e-06, + "loss": 0.3989, + "step": 6394 + }, + { + "epoch": 3.471769815418024, + "grad_norm": 12.421864505486035, + "learning_rate": 4.5130202718041004e-06, + "loss": 0.8773, + "step": 6395 + }, + { + "epoch": 3.4723127035830617, + "grad_norm": 12.947288721903634, + "learning_rate": 4.510080460790775e-06, + "loss": 0.5396, + "step": 6396 + }, + { + "epoch": 3.4728555917481, + "grad_norm": 10.339717202308544, + "learning_rate": 4.507141328779439e-06, + "loss": 0.5154, + "step": 6397 + }, + { + "epoch": 3.473398479913138, + "grad_norm": 8.810723375318315, + "learning_rate": 4.504202876133627e-06, + "loss": 0.4513, + "step": 6398 + }, + { + "epoch": 3.473941368078176, + "grad_norm": 13.655631619886284, + "learning_rate": 4.501265103216755e-06, + "loss": 0.5043, + "step": 6399 + }, + { + "epoch": 3.4744842562432137, + "grad_norm": 9.947920274498516, + "learning_rate": 4.49832801039218e-06, + "loss": 0.6907, + "step": 6400 + }, + { + "epoch": 3.475027144408252, + "grad_norm": 9.800586651795994, + "learning_rate": 4.495391598023167e-06, + "loss": 0.3988, + "step": 6401 + }, + { + "epoch": 3.47557003257329, + "grad_norm": 14.696382692377027, + "learning_rate": 4.492455866472894e-06, + "loss": 0.5805, + "step": 6402 + }, + { + "epoch": 3.476112920738328, + "grad_norm": 13.06051257277803, + "learning_rate": 4.489520816104455e-06, + "loss": 0.7204, + "step": 6403 + }, + { + "epoch": 3.4766558089033657, + "grad_norm": 12.99399781439784, + "learning_rate": 4.486586447280866e-06, + "loss": 0.7112, + "step": 6404 + }, + { + "epoch": 3.477198697068404, + "grad_norm": 12.422813336616981, + "learning_rate": 4.483652760365052e-06, + "loss": 0.3927, + "step": 6405 + }, + { + "epoch": 3.477741585233442, + "grad_norm": 10.948217532391471, + "learning_rate": 4.480719755719857e-06, + "loss": 0.5587, + "step": 6406 + }, + { + "epoch": 3.47828447339848, + "grad_norm": 11.657205565439646, + "learning_rate": 4.477787433708038e-06, + "loss": 0.5584, + "step": 6407 + }, + { + "epoch": 3.4788273615635177, + "grad_norm": 7.795661051863107, + "learning_rate": 4.474855794692271e-06, + "loss": 0.3232, + "step": 6408 + }, + { + "epoch": 3.479370249728556, + "grad_norm": 7.775932230943303, + "learning_rate": 4.4719248390351446e-06, + "loss": 0.3613, + "step": 6409 + }, + { + "epoch": 3.479913137893594, + "grad_norm": 9.097718161588768, + "learning_rate": 4.468994567099165e-06, + "loss": 0.3896, + "step": 6410 + }, + { + "epoch": 3.480456026058632, + "grad_norm": 7.087231016744389, + "learning_rate": 4.466064979246751e-06, + "loss": 0.2486, + "step": 6411 + }, + { + "epoch": 3.4809989142236697, + "grad_norm": 17.009119363822553, + "learning_rate": 4.463136075840242e-06, + "loss": 0.7777, + "step": 6412 + }, + { + "epoch": 3.481541802388708, + "grad_norm": 11.54915190749092, + "learning_rate": 4.460207857241887e-06, + "loss": 0.4528, + "step": 6413 + }, + { + "epoch": 3.482084690553746, + "grad_norm": 14.237806827589361, + "learning_rate": 4.45728032381385e-06, + "loss": 0.8941, + "step": 6414 + }, + { + "epoch": 3.482627578718784, + "grad_norm": 11.85837340138276, + "learning_rate": 4.454353475918223e-06, + "loss": 0.5289, + "step": 6415 + }, + { + "epoch": 3.4831704668838217, + "grad_norm": 13.987937632959078, + "learning_rate": 4.4514273139169925e-06, + "loss": 0.5152, + "step": 6416 + }, + { + "epoch": 3.48371335504886, + "grad_norm": 10.643372071284285, + "learning_rate": 4.4485018381720755e-06, + "loss": 0.4758, + "step": 6417 + }, + { + "epoch": 3.484256243213898, + "grad_norm": 8.692254926051978, + "learning_rate": 4.445577049045299e-06, + "loss": 0.382, + "step": 6418 + }, + { + "epoch": 3.484799131378936, + "grad_norm": 9.767782914786244, + "learning_rate": 4.4426529468984055e-06, + "loss": 0.4674, + "step": 6419 + }, + { + "epoch": 3.4853420195439737, + "grad_norm": 14.118719943852668, + "learning_rate": 4.4397295320930525e-06, + "loss": 0.5859, + "step": 6420 + }, + { + "epoch": 3.485884907709012, + "grad_norm": 9.120918716734044, + "learning_rate": 4.4368068049908085e-06, + "loss": 0.37, + "step": 6421 + }, + { + "epoch": 3.48642779587405, + "grad_norm": 12.706215416375834, + "learning_rate": 4.4338847659531735e-06, + "loss": 0.8322, + "step": 6422 + }, + { + "epoch": 3.486970684039088, + "grad_norm": 10.926968726715666, + "learning_rate": 4.430963415341533e-06, + "loss": 0.4146, + "step": 6423 + }, + { + "epoch": 3.4875135722041257, + "grad_norm": 14.53230992435977, + "learning_rate": 4.428042753517222e-06, + "loss": 0.5612, + "step": 6424 + }, + { + "epoch": 3.488056460369164, + "grad_norm": 12.614428085329433, + "learning_rate": 4.425122780841456e-06, + "loss": 0.4886, + "step": 6425 + }, + { + "epoch": 3.488599348534202, + "grad_norm": 9.07248478126251, + "learning_rate": 4.422203497675394e-06, + "loss": 0.3342, + "step": 6426 + }, + { + "epoch": 3.48914223669924, + "grad_norm": 7.63850442052429, + "learning_rate": 4.419284904380095e-06, + "loss": 0.2709, + "step": 6427 + }, + { + "epoch": 3.4896851248642777, + "grad_norm": 8.139657158702438, + "learning_rate": 4.416367001316526e-06, + "loss": 0.3276, + "step": 6428 + }, + { + "epoch": 3.490228013029316, + "grad_norm": 12.800427760538255, + "learning_rate": 4.413449788845594e-06, + "loss": 0.8097, + "step": 6429 + }, + { + "epoch": 3.490770901194354, + "grad_norm": 9.88357123496715, + "learning_rate": 4.410533267328087e-06, + "loss": 0.3224, + "step": 6430 + }, + { + "epoch": 3.491313789359392, + "grad_norm": 10.15573231324521, + "learning_rate": 4.407617437124741e-06, + "loss": 0.5535, + "step": 6431 + }, + { + "epoch": 3.4918566775244297, + "grad_norm": 17.972024507407102, + "learning_rate": 4.404702298596177e-06, + "loss": 0.9413, + "step": 6432 + }, + { + "epoch": 3.492399565689468, + "grad_norm": 11.366140193603957, + "learning_rate": 4.401787852102955e-06, + "loss": 0.7224, + "step": 6433 + }, + { + "epoch": 3.492942453854506, + "grad_norm": 11.183293759896978, + "learning_rate": 4.398874098005532e-06, + "loss": 0.6597, + "step": 6434 + }, + { + "epoch": 3.493485342019544, + "grad_norm": 11.409722406543372, + "learning_rate": 4.395961036664288e-06, + "loss": 0.6165, + "step": 6435 + }, + { + "epoch": 3.4940282301845818, + "grad_norm": 8.532518886519954, + "learning_rate": 4.393048668439518e-06, + "loss": 0.3045, + "step": 6436 + }, + { + "epoch": 3.49457111834962, + "grad_norm": 15.499278790978442, + "learning_rate": 4.390136993691417e-06, + "loss": 0.6505, + "step": 6437 + }, + { + "epoch": 3.495114006514658, + "grad_norm": 9.855565799480534, + "learning_rate": 4.387226012780117e-06, + "loss": 0.4787, + "step": 6438 + }, + { + "epoch": 3.495656894679696, + "grad_norm": 13.510584259969823, + "learning_rate": 4.38431572606565e-06, + "loss": 0.5388, + "step": 6439 + }, + { + "epoch": 3.4961997828447338, + "grad_norm": 8.040242399067179, + "learning_rate": 4.381406133907964e-06, + "loss": 0.3596, + "step": 6440 + }, + { + "epoch": 3.496742671009772, + "grad_norm": 9.402081395938351, + "learning_rate": 4.378497236666922e-06, + "loss": 0.3326, + "step": 6441 + }, + { + "epoch": 3.49728555917481, + "grad_norm": 15.188738381211488, + "learning_rate": 4.3755890347023e-06, + "loss": 0.7356, + "step": 6442 + }, + { + "epoch": 3.497828447339848, + "grad_norm": 13.644691532416777, + "learning_rate": 4.372681528373791e-06, + "loss": 0.5727, + "step": 6443 + }, + { + "epoch": 3.4983713355048858, + "grad_norm": 12.036428423892223, + "learning_rate": 4.369774718041e-06, + "loss": 0.4333, + "step": 6444 + }, + { + "epoch": 3.498914223669924, + "grad_norm": 9.814264562422439, + "learning_rate": 4.366868604063444e-06, + "loss": 0.4396, + "step": 6445 + }, + { + "epoch": 3.499457111834962, + "grad_norm": 10.623481464500886, + "learning_rate": 4.363963186800557e-06, + "loss": 0.509, + "step": 6446 + }, + { + "epoch": 3.5, + "grad_norm": 9.931711278540687, + "learning_rate": 4.361058466611686e-06, + "loss": 0.352, + "step": 6447 + }, + { + "epoch": 3.500542888165038, + "grad_norm": 10.709530760490054, + "learning_rate": 4.358154443856091e-06, + "loss": 0.3233, + "step": 6448 + }, + { + "epoch": 3.501085776330076, + "grad_norm": 12.489815680915486, + "learning_rate": 4.355251118892946e-06, + "loss": 0.8317, + "step": 6449 + }, + { + "epoch": 3.5016286644951142, + "grad_norm": 13.400967427781875, + "learning_rate": 4.35234849208134e-06, + "loss": 0.601, + "step": 6450 + }, + { + "epoch": 3.502171552660152, + "grad_norm": 11.710683499628859, + "learning_rate": 4.349446563780272e-06, + "loss": 0.7362, + "step": 6451 + }, + { + "epoch": 3.50271444082519, + "grad_norm": 14.51448957958314, + "learning_rate": 4.346545334348658e-06, + "loss": 0.8007, + "step": 6452 + }, + { + "epoch": 3.503257328990228, + "grad_norm": 10.464776296844045, + "learning_rate": 4.343644804145329e-06, + "loss": 0.3993, + "step": 6453 + }, + { + "epoch": 3.5038002171552662, + "grad_norm": 9.718005412495163, + "learning_rate": 4.340744973529022e-06, + "loss": 0.269, + "step": 6454 + }, + { + "epoch": 3.504343105320304, + "grad_norm": 10.85577962939331, + "learning_rate": 4.337845842858402e-06, + "loss": 0.5342, + "step": 6455 + }, + { + "epoch": 3.504885993485342, + "grad_norm": 12.181308768808934, + "learning_rate": 4.33494741249203e-06, + "loss": 0.4005, + "step": 6456 + }, + { + "epoch": 3.50542888165038, + "grad_norm": 11.779746087085687, + "learning_rate": 4.332049682788391e-06, + "loss": 0.4414, + "step": 6457 + }, + { + "epoch": 3.5059717698154182, + "grad_norm": 11.664912784701475, + "learning_rate": 4.3291526541058795e-06, + "loss": 0.548, + "step": 6458 + }, + { + "epoch": 3.506514657980456, + "grad_norm": 13.373343929291233, + "learning_rate": 4.326256326802802e-06, + "loss": 0.564, + "step": 6459 + }, + { + "epoch": 3.507057546145494, + "grad_norm": 11.223466720082536, + "learning_rate": 4.3233607012373925e-06, + "loss": 0.3611, + "step": 6460 + }, + { + "epoch": 3.507600434310532, + "grad_norm": 13.530570397967848, + "learning_rate": 4.32046577776777e-06, + "loss": 0.6113, + "step": 6461 + }, + { + "epoch": 3.5081433224755703, + "grad_norm": 12.62657182669542, + "learning_rate": 4.3175715567520015e-06, + "loss": 0.4457, + "step": 6462 + }, + { + "epoch": 3.508686210640608, + "grad_norm": 14.69985392835765, + "learning_rate": 4.31467803854803e-06, + "loss": 0.8015, + "step": 6463 + }, + { + "epoch": 3.509229098805646, + "grad_norm": 12.910242277352095, + "learning_rate": 4.311785223513744e-06, + "loss": 0.5957, + "step": 6464 + }, + { + "epoch": 3.509771986970684, + "grad_norm": 14.38667526612287, + "learning_rate": 4.308893112006932e-06, + "loss": 0.6333, + "step": 6465 + }, + { + "epoch": 3.5103148751357223, + "grad_norm": 10.811689404016661, + "learning_rate": 4.306001704385282e-06, + "loss": 0.379, + "step": 6466 + }, + { + "epoch": 3.51085776330076, + "grad_norm": 12.950655194202454, + "learning_rate": 4.303111001006424e-06, + "loss": 0.3791, + "step": 6467 + }, + { + "epoch": 3.511400651465798, + "grad_norm": 10.403413919758718, + "learning_rate": 4.300221002227869e-06, + "loss": 0.5023, + "step": 6468 + }, + { + "epoch": 3.511943539630836, + "grad_norm": 14.86724875684143, + "learning_rate": 4.297331708407072e-06, + "loss": 0.5985, + "step": 6469 + }, + { + "epoch": 3.5124864277958743, + "grad_norm": 9.211711120681478, + "learning_rate": 4.29444311990137e-06, + "loss": 0.2738, + "step": 6470 + }, + { + "epoch": 3.513029315960912, + "grad_norm": 11.398896094597806, + "learning_rate": 4.29155523706804e-06, + "loss": 0.6349, + "step": 6471 + }, + { + "epoch": 3.51357220412595, + "grad_norm": 9.97395797119566, + "learning_rate": 4.288668060264257e-06, + "loss": 0.3945, + "step": 6472 + }, + { + "epoch": 3.514115092290988, + "grad_norm": 10.965919373159068, + "learning_rate": 4.2857815898471114e-06, + "loss": 0.6313, + "step": 6473 + }, + { + "epoch": 3.5146579804560263, + "grad_norm": 14.272003668320222, + "learning_rate": 4.2828958261736045e-06, + "loss": 0.8566, + "step": 6474 + }, + { + "epoch": 3.515200868621064, + "grad_norm": 9.155542594309077, + "learning_rate": 4.280010769600653e-06, + "loss": 0.4155, + "step": 6475 + }, + { + "epoch": 3.515743756786102, + "grad_norm": 12.664316762774577, + "learning_rate": 4.277126420485087e-06, + "loss": 0.4668, + "step": 6476 + }, + { + "epoch": 3.51628664495114, + "grad_norm": 13.218980598800185, + "learning_rate": 4.274242779183646e-06, + "loss": 0.5638, + "step": 6477 + }, + { + "epoch": 3.5168295331161783, + "grad_norm": 10.791302144957823, + "learning_rate": 4.271359846052983e-06, + "loss": 0.5458, + "step": 6478 + }, + { + "epoch": 3.517372421281216, + "grad_norm": 11.294781461737724, + "learning_rate": 4.268477621449666e-06, + "loss": 0.5746, + "step": 6479 + }, + { + "epoch": 3.517915309446254, + "grad_norm": 8.318333652865173, + "learning_rate": 4.2655961057301695e-06, + "loss": 0.3447, + "step": 6480 + }, + { + "epoch": 3.518458197611292, + "grad_norm": 14.570486409127577, + "learning_rate": 4.2627152992508865e-06, + "loss": 0.6757, + "step": 6481 + }, + { + "epoch": 3.5190010857763303, + "grad_norm": 12.036103572116089, + "learning_rate": 4.25983520236812e-06, + "loss": 0.4259, + "step": 6482 + }, + { + "epoch": 3.519543973941368, + "grad_norm": 9.254932202972295, + "learning_rate": 4.256955815438084e-06, + "loss": 0.4689, + "step": 6483 + }, + { + "epoch": 3.520086862106406, + "grad_norm": 10.562719431036207, + "learning_rate": 4.254077138816905e-06, + "loss": 0.3951, + "step": 6484 + }, + { + "epoch": 3.520629750271444, + "grad_norm": 10.475306089406416, + "learning_rate": 4.251199172860624e-06, + "loss": 0.6583, + "step": 6485 + }, + { + "epoch": 3.5211726384364823, + "grad_norm": 11.146639535946015, + "learning_rate": 4.2483219179251915e-06, + "loss": 0.5873, + "step": 6486 + }, + { + "epoch": 3.52171552660152, + "grad_norm": 13.607190728220559, + "learning_rate": 4.245445374366469e-06, + "loss": 0.5111, + "step": 6487 + }, + { + "epoch": 3.522258414766558, + "grad_norm": 13.73052324797901, + "learning_rate": 4.242569542540236e-06, + "loss": 0.6546, + "step": 6488 + }, + { + "epoch": 3.522801302931596, + "grad_norm": 11.846325790998266, + "learning_rate": 4.239694422802177e-06, + "loss": 0.3475, + "step": 6489 + }, + { + "epoch": 3.5233441910966343, + "grad_norm": 9.944462539493102, + "learning_rate": 4.236820015507893e-06, + "loss": 0.3798, + "step": 6490 + }, + { + "epoch": 3.523887079261672, + "grad_norm": 11.901268081862414, + "learning_rate": 4.233946321012895e-06, + "loss": 0.503, + "step": 6491 + }, + { + "epoch": 3.52442996742671, + "grad_norm": 11.373264829742677, + "learning_rate": 4.231073339672601e-06, + "loss": 0.3985, + "step": 6492 + }, + { + "epoch": 3.524972855591748, + "grad_norm": 8.462656388647403, + "learning_rate": 4.228201071842358e-06, + "loss": 0.3526, + "step": 6493 + }, + { + "epoch": 3.5255157437567863, + "grad_norm": 13.319319520255643, + "learning_rate": 4.2253295178774036e-06, + "loss": 0.3119, + "step": 6494 + }, + { + "epoch": 3.526058631921824, + "grad_norm": 11.873988390722962, + "learning_rate": 4.222458678132896e-06, + "loss": 0.3681, + "step": 6495 + }, + { + "epoch": 3.526601520086862, + "grad_norm": 14.615618896842006, + "learning_rate": 4.219588552963907e-06, + "loss": 0.5678, + "step": 6496 + }, + { + "epoch": 3.5271444082519, + "grad_norm": 12.902856396087687, + "learning_rate": 4.216719142725415e-06, + "loss": 0.4386, + "step": 6497 + }, + { + "epoch": 3.5276872964169383, + "grad_norm": 14.124175164997677, + "learning_rate": 4.213850447772324e-06, + "loss": 0.752, + "step": 6498 + }, + { + "epoch": 3.528230184581976, + "grad_norm": 17.881101453026922, + "learning_rate": 4.210982468459425e-06, + "loss": 1.0417, + "step": 6499 + }, + { + "epoch": 3.528773072747014, + "grad_norm": 12.44176561816019, + "learning_rate": 4.2081152051414464e-06, + "loss": 0.5422, + "step": 6500 + }, + { + "epoch": 3.529315960912052, + "grad_norm": 11.595014008440414, + "learning_rate": 4.205248658173005e-06, + "loss": 0.5242, + "step": 6501 + }, + { + "epoch": 3.5298588490770904, + "grad_norm": 13.42551753781462, + "learning_rate": 4.20238282790865e-06, + "loss": 0.4613, + "step": 6502 + }, + { + "epoch": 3.530401737242128, + "grad_norm": 14.699795423214242, + "learning_rate": 4.1995177147028195e-06, + "loss": 0.7931, + "step": 6503 + }, + { + "epoch": 3.530944625407166, + "grad_norm": 13.875599451841557, + "learning_rate": 4.196653318909888e-06, + "loss": 0.7584, + "step": 6504 + }, + { + "epoch": 3.531487513572204, + "grad_norm": 10.802073588313325, + "learning_rate": 4.193789640884126e-06, + "loss": 0.4051, + "step": 6505 + }, + { + "epoch": 3.5320304017372424, + "grad_norm": 12.895721698436635, + "learning_rate": 4.190926680979708e-06, + "loss": 0.5394, + "step": 6506 + }, + { + "epoch": 3.53257328990228, + "grad_norm": 9.06469669507275, + "learning_rate": 4.188064439550743e-06, + "loss": 0.3421, + "step": 6507 + }, + { + "epoch": 3.533116178067318, + "grad_norm": 9.872074130422732, + "learning_rate": 4.185202916951224e-06, + "loss": 0.4934, + "step": 6508 + }, + { + "epoch": 3.533659066232356, + "grad_norm": 14.460060519698754, + "learning_rate": 4.1823421135350796e-06, + "loss": 0.9002, + "step": 6509 + }, + { + "epoch": 3.5342019543973944, + "grad_norm": 10.80679699799082, + "learning_rate": 4.179482029656134e-06, + "loss": 0.5466, + "step": 6510 + }, + { + "epoch": 3.534744842562432, + "grad_norm": 17.39434448772562, + "learning_rate": 4.176622665668127e-06, + "loss": 0.7686, + "step": 6511 + }, + { + "epoch": 3.53528773072747, + "grad_norm": 18.02312862709536, + "learning_rate": 4.173764021924711e-06, + "loss": 0.6285, + "step": 6512 + }, + { + "epoch": 3.535830618892508, + "grad_norm": 16.367851061651017, + "learning_rate": 4.170906098779446e-06, + "loss": 0.7154, + "step": 6513 + }, + { + "epoch": 3.5363735070575464, + "grad_norm": 14.114924740024234, + "learning_rate": 4.168048896585809e-06, + "loss": 0.6001, + "step": 6514 + }, + { + "epoch": 3.536916395222584, + "grad_norm": 11.148204893242914, + "learning_rate": 4.165192415697171e-06, + "loss": 0.4029, + "step": 6515 + }, + { + "epoch": 3.537459283387622, + "grad_norm": 12.586445991725615, + "learning_rate": 4.162336656466839e-06, + "loss": 0.53, + "step": 6516 + }, + { + "epoch": 3.53800217155266, + "grad_norm": 10.601436138316005, + "learning_rate": 4.159481619248012e-06, + "loss": 0.5671, + "step": 6517 + }, + { + "epoch": 3.5385450597176984, + "grad_norm": 21.51381975898863, + "learning_rate": 4.156627304393808e-06, + "loss": 0.8123, + "step": 6518 + }, + { + "epoch": 3.539087947882736, + "grad_norm": 11.786864237322536, + "learning_rate": 4.153773712257251e-06, + "loss": 0.5383, + "step": 6519 + }, + { + "epoch": 3.539630836047774, + "grad_norm": 13.251189336378648, + "learning_rate": 4.1509208431912785e-06, + "loss": 0.4292, + "step": 6520 + }, + { + "epoch": 3.540173724212812, + "grad_norm": 11.978453022863066, + "learning_rate": 4.14806869754874e-06, + "loss": 0.5242, + "step": 6521 + }, + { + "epoch": 3.5407166123778504, + "grad_norm": 14.020243792854204, + "learning_rate": 4.145217275682389e-06, + "loss": 0.4147, + "step": 6522 + }, + { + "epoch": 3.541259500542888, + "grad_norm": 12.871747424704319, + "learning_rate": 4.142366577944897e-06, + "loss": 0.479, + "step": 6523 + }, + { + "epoch": 3.541802388707926, + "grad_norm": 16.745498758328235, + "learning_rate": 4.139516604688843e-06, + "loss": 0.732, + "step": 6524 + }, + { + "epoch": 3.542345276872964, + "grad_norm": 18.803948548990295, + "learning_rate": 4.136667356266716e-06, + "loss": 0.6007, + "step": 6525 + }, + { + "epoch": 3.5428881650380024, + "grad_norm": 12.83744283380288, + "learning_rate": 4.133818833030915e-06, + "loss": 0.556, + "step": 6526 + }, + { + "epoch": 3.54343105320304, + "grad_norm": 10.490416435869522, + "learning_rate": 4.13097103533375e-06, + "loss": 0.3933, + "step": 6527 + }, + { + "epoch": 3.543973941368078, + "grad_norm": 13.443902999757336, + "learning_rate": 4.128123963527441e-06, + "loss": 0.7184, + "step": 6528 + }, + { + "epoch": 3.544516829533116, + "grad_norm": 9.68823576515441, + "learning_rate": 4.125277617964119e-06, + "loss": 0.4799, + "step": 6529 + }, + { + "epoch": 3.5450597176981544, + "grad_norm": 8.977409878928121, + "learning_rate": 4.122431998995819e-06, + "loss": 0.2998, + "step": 6530 + }, + { + "epoch": 3.545602605863192, + "grad_norm": 10.738202771196406, + "learning_rate": 4.119587106974505e-06, + "loss": 0.4038, + "step": 6531 + }, + { + "epoch": 3.54614549402823, + "grad_norm": 13.534073718206239, + "learning_rate": 4.116742942252024e-06, + "loss": 0.6041, + "step": 6532 + }, + { + "epoch": 3.546688382193268, + "grad_norm": 9.554994157525224, + "learning_rate": 4.113899505180157e-06, + "loss": 0.3955, + "step": 6533 + }, + { + "epoch": 3.5472312703583064, + "grad_norm": 14.083745919964446, + "learning_rate": 4.1110567961105775e-06, + "loss": 0.5232, + "step": 6534 + }, + { + "epoch": 3.547774158523344, + "grad_norm": 9.270850995609294, + "learning_rate": 4.108214815394876e-06, + "loss": 0.3932, + "step": 6535 + }, + { + "epoch": 3.548317046688382, + "grad_norm": 8.658156641649512, + "learning_rate": 4.105373563384563e-06, + "loss": 0.3587, + "step": 6536 + }, + { + "epoch": 3.54885993485342, + "grad_norm": 22.179187215860352, + "learning_rate": 4.102533040431034e-06, + "loss": 0.6502, + "step": 6537 + }, + { + "epoch": 3.5494028230184584, + "grad_norm": 12.115851203141874, + "learning_rate": 4.0996932468856265e-06, + "loss": 0.6969, + "step": 6538 + }, + { + "epoch": 3.549945711183496, + "grad_norm": 9.686560788090254, + "learning_rate": 4.096854183099554e-06, + "loss": 0.4376, + "step": 6539 + }, + { + "epoch": 3.550488599348534, + "grad_norm": 14.280429673703154, + "learning_rate": 4.0940158494239725e-06, + "loss": 0.7158, + "step": 6540 + }, + { + "epoch": 3.5510314875135722, + "grad_norm": 12.581936500402087, + "learning_rate": 4.091178246209916e-06, + "loss": 0.5323, + "step": 6541 + }, + { + "epoch": 3.5515743756786105, + "grad_norm": 14.994050222936808, + "learning_rate": 4.088341373808354e-06, + "loss": 0.6231, + "step": 6542 + }, + { + "epoch": 3.5521172638436482, + "grad_norm": 11.419896198480089, + "learning_rate": 4.085505232570157e-06, + "loss": 0.4173, + "step": 6543 + }, + { + "epoch": 3.552660152008686, + "grad_norm": 9.002355774397408, + "learning_rate": 4.082669822846092e-06, + "loss": 0.4046, + "step": 6544 + }, + { + "epoch": 3.5532030401737242, + "grad_norm": 8.59747484019196, + "learning_rate": 4.079835144986861e-06, + "loss": 0.3399, + "step": 6545 + }, + { + "epoch": 3.5537459283387625, + "grad_norm": 15.204451286324845, + "learning_rate": 4.0770011993430465e-06, + "loss": 0.6488, + "step": 6546 + }, + { + "epoch": 3.5542888165038002, + "grad_norm": 12.454918944013878, + "learning_rate": 4.074167986265169e-06, + "loss": 0.7895, + "step": 6547 + }, + { + "epoch": 3.554831704668838, + "grad_norm": 12.640544797370467, + "learning_rate": 4.071335506103639e-06, + "loss": 0.3486, + "step": 6548 + }, + { + "epoch": 3.5553745928338762, + "grad_norm": 14.54014915273436, + "learning_rate": 4.068503759208782e-06, + "loss": 0.8116, + "step": 6549 + }, + { + "epoch": 3.5559174809989145, + "grad_norm": 13.706653138925564, + "learning_rate": 4.065672745930833e-06, + "loss": 0.6351, + "step": 6550 + }, + { + "epoch": 3.5564603691639523, + "grad_norm": 7.990687725437862, + "learning_rate": 4.062842466619937e-06, + "loss": 0.356, + "step": 6551 + }, + { + "epoch": 3.55700325732899, + "grad_norm": 12.892934153832375, + "learning_rate": 4.060012921626146e-06, + "loss": 0.5433, + "step": 6552 + }, + { + "epoch": 3.5575461454940283, + "grad_norm": 11.135918453001386, + "learning_rate": 4.057184111299426e-06, + "loss": 0.5487, + "step": 6553 + }, + { + "epoch": 3.5580890336590665, + "grad_norm": 8.023158669990128, + "learning_rate": 4.054356035989645e-06, + "loss": 0.2137, + "step": 6554 + }, + { + "epoch": 3.5586319218241043, + "grad_norm": 15.187504399840787, + "learning_rate": 4.051528696046586e-06, + "loss": 0.513, + "step": 6555 + }, + { + "epoch": 3.559174809989142, + "grad_norm": 11.432478939914837, + "learning_rate": 4.048702091819938e-06, + "loss": 0.4349, + "step": 6556 + }, + { + "epoch": 3.5597176981541803, + "grad_norm": 8.365367809450516, + "learning_rate": 4.045876223659301e-06, + "loss": 0.3435, + "step": 6557 + }, + { + "epoch": 3.5602605863192185, + "grad_norm": 12.186833110184857, + "learning_rate": 4.0430510919141815e-06, + "loss": 0.6071, + "step": 6558 + }, + { + "epoch": 3.5608034744842563, + "grad_norm": 12.640045326528098, + "learning_rate": 4.040226696933998e-06, + "loss": 0.4771, + "step": 6559 + }, + { + "epoch": 3.561346362649294, + "grad_norm": 12.859046455514006, + "learning_rate": 4.037403039068073e-06, + "loss": 0.6651, + "step": 6560 + }, + { + "epoch": 3.5618892508143323, + "grad_norm": 10.219552402377104, + "learning_rate": 4.034580118665644e-06, + "loss": 0.3973, + "step": 6561 + }, + { + "epoch": 3.5624321389793705, + "grad_norm": 11.468743456972462, + "learning_rate": 4.031757936075854e-06, + "loss": 0.5699, + "step": 6562 + }, + { + "epoch": 3.5629750271444083, + "grad_norm": 14.399787814630601, + "learning_rate": 4.028936491647753e-06, + "loss": 0.7208, + "step": 6563 + }, + { + "epoch": 3.563517915309446, + "grad_norm": 12.303594318406011, + "learning_rate": 4.026115785730305e-06, + "loss": 0.4314, + "step": 6564 + }, + { + "epoch": 3.5640608034744843, + "grad_norm": 9.21496866787658, + "learning_rate": 4.023295818672377e-06, + "loss": 0.4055, + "step": 6565 + }, + { + "epoch": 3.5646036916395225, + "grad_norm": 13.534731784395452, + "learning_rate": 4.0204765908227475e-06, + "loss": 0.4986, + "step": 6566 + }, + { + "epoch": 3.5651465798045603, + "grad_norm": 10.67198457926589, + "learning_rate": 4.017658102530103e-06, + "loss": 0.424, + "step": 6567 + }, + { + "epoch": 3.565689467969598, + "grad_norm": 10.45173885950783, + "learning_rate": 4.014840354143035e-06, + "loss": 0.3562, + "step": 6568 + }, + { + "epoch": 3.5662323561346363, + "grad_norm": 17.950537818667204, + "learning_rate": 4.012023346010059e-06, + "loss": 0.6163, + "step": 6569 + }, + { + "epoch": 3.5667752442996745, + "grad_norm": 12.324798504977831, + "learning_rate": 4.009207078479571e-06, + "loss": 0.4225, + "step": 6570 + }, + { + "epoch": 3.5673181324647123, + "grad_norm": 14.602931671691664, + "learning_rate": 4.006391551899906e-06, + "loss": 0.5519, + "step": 6571 + }, + { + "epoch": 3.56786102062975, + "grad_norm": 11.70892852397016, + "learning_rate": 4.00357676661928e-06, + "loss": 0.483, + "step": 6572 + }, + { + "epoch": 3.5684039087947883, + "grad_norm": 9.372765530009781, + "learning_rate": 4.000762722985844e-06, + "loss": 0.2418, + "step": 6573 + }, + { + "epoch": 3.5689467969598265, + "grad_norm": 9.365246601466223, + "learning_rate": 3.997949421347631e-06, + "loss": 0.3748, + "step": 6574 + }, + { + "epoch": 3.5694896851248643, + "grad_norm": 21.027585535755822, + "learning_rate": 3.995136862052597e-06, + "loss": 0.6116, + "step": 6575 + }, + { + "epoch": 3.570032573289902, + "grad_norm": 9.558242987641426, + "learning_rate": 3.992325045448613e-06, + "loss": 0.3608, + "step": 6576 + }, + { + "epoch": 3.5705754614549403, + "grad_norm": 15.196445736424556, + "learning_rate": 3.989513971883434e-06, + "loss": 0.5847, + "step": 6577 + }, + { + "epoch": 3.5711183496199785, + "grad_norm": 13.049400371079434, + "learning_rate": 3.9867036417047546e-06, + "loss": 0.9054, + "step": 6578 + }, + { + "epoch": 3.5716612377850163, + "grad_norm": 10.359620340512786, + "learning_rate": 3.983894055260146e-06, + "loss": 0.3928, + "step": 6579 + }, + { + "epoch": 3.572204125950054, + "grad_norm": 11.96747351920452, + "learning_rate": 3.981085212897111e-06, + "loss": 0.4581, + "step": 6580 + }, + { + "epoch": 3.5727470141150923, + "grad_norm": 13.611366572987158, + "learning_rate": 3.97827711496305e-06, + "loss": 0.8939, + "step": 6581 + }, + { + "epoch": 3.5732899022801305, + "grad_norm": 12.527292279303031, + "learning_rate": 3.975469761805273e-06, + "loss": 0.3713, + "step": 6582 + }, + { + "epoch": 3.5738327904451683, + "grad_norm": 11.7570760452819, + "learning_rate": 3.9726631537710005e-06, + "loss": 0.442, + "step": 6583 + }, + { + "epoch": 3.574375678610206, + "grad_norm": 10.8606079711249, + "learning_rate": 3.969857291207349e-06, + "loss": 0.3717, + "step": 6584 + }, + { + "epoch": 3.5749185667752443, + "grad_norm": 10.89094058468825, + "learning_rate": 3.9670521744613645e-06, + "loss": 0.5867, + "step": 6585 + }, + { + "epoch": 3.5754614549402826, + "grad_norm": 11.08527189234627, + "learning_rate": 3.964247803879976e-06, + "loss": 0.4183, + "step": 6586 + }, + { + "epoch": 3.5760043431053203, + "grad_norm": 14.584182784882016, + "learning_rate": 3.9614441798100415e-06, + "loss": 0.6388, + "step": 6587 + }, + { + "epoch": 3.576547231270358, + "grad_norm": 12.70805104327354, + "learning_rate": 3.958641302598315e-06, + "loss": 0.4211, + "step": 6588 + }, + { + "epoch": 3.5770901194353963, + "grad_norm": 13.935385901164445, + "learning_rate": 3.95583917259146e-06, + "loss": 0.4955, + "step": 6589 + }, + { + "epoch": 3.5776330076004346, + "grad_norm": 8.25831878298027, + "learning_rate": 3.953037790136051e-06, + "loss": 0.3929, + "step": 6590 + }, + { + "epoch": 3.5781758957654723, + "grad_norm": 13.224695924777794, + "learning_rate": 3.950237155578563e-06, + "loss": 0.9157, + "step": 6591 + }, + { + "epoch": 3.57871878393051, + "grad_norm": 14.476470240562103, + "learning_rate": 3.947437269265387e-06, + "loss": 0.6185, + "step": 6592 + }, + { + "epoch": 3.5792616720955484, + "grad_norm": 14.623277057541618, + "learning_rate": 3.944638131542816e-06, + "loss": 0.415, + "step": 6593 + }, + { + "epoch": 3.5798045602605866, + "grad_norm": 10.928897970897149, + "learning_rate": 3.941839742757052e-06, + "loss": 0.4612, + "step": 6594 + }, + { + "epoch": 3.5803474484256244, + "grad_norm": 11.795057907246525, + "learning_rate": 3.939042103254204e-06, + "loss": 0.449, + "step": 6595 + }, + { + "epoch": 3.580890336590662, + "grad_norm": 13.011176416375783, + "learning_rate": 3.9362452133802866e-06, + "loss": 0.4674, + "step": 6596 + }, + { + "epoch": 3.5814332247557004, + "grad_norm": 12.733097120202023, + "learning_rate": 3.933449073481227e-06, + "loss": 0.5925, + "step": 6597 + }, + { + "epoch": 3.5819761129207386, + "grad_norm": 11.53786371590528, + "learning_rate": 3.930653683902854e-06, + "loss": 0.6089, + "step": 6598 + }, + { + "epoch": 3.5825190010857764, + "grad_norm": 10.840934046340113, + "learning_rate": 3.927859044990907e-06, + "loss": 0.3937, + "step": 6599 + }, + { + "epoch": 3.583061889250814, + "grad_norm": 13.435349756607788, + "learning_rate": 3.92506515709103e-06, + "loss": 0.5642, + "step": 6600 + }, + { + "epoch": 3.5836047774158524, + "grad_norm": 15.725827415410208, + "learning_rate": 3.922272020548773e-06, + "loss": 0.7689, + "step": 6601 + }, + { + "epoch": 3.5841476655808906, + "grad_norm": 10.5862295430844, + "learning_rate": 3.919479635709606e-06, + "loss": 0.4151, + "step": 6602 + }, + { + "epoch": 3.5846905537459284, + "grad_norm": 13.70191819426941, + "learning_rate": 3.9166880029188855e-06, + "loss": 0.5176, + "step": 6603 + }, + { + "epoch": 3.585233441910966, + "grad_norm": 20.653564386817997, + "learning_rate": 3.9138971225218895e-06, + "loss": 0.5782, + "step": 6604 + }, + { + "epoch": 3.5857763300760044, + "grad_norm": 13.964645747661887, + "learning_rate": 3.911106994863794e-06, + "loss": 0.629, + "step": 6605 + }, + { + "epoch": 3.5863192182410426, + "grad_norm": 15.152047460754902, + "learning_rate": 3.908317620289689e-06, + "loss": 0.7885, + "step": 6606 + }, + { + "epoch": 3.5868621064060804, + "grad_norm": 10.596374095925908, + "learning_rate": 3.905528999144576e-06, + "loss": 0.3883, + "step": 6607 + }, + { + "epoch": 3.587404994571118, + "grad_norm": 11.600211796857451, + "learning_rate": 3.902741131773341e-06, + "loss": 0.4589, + "step": 6608 + }, + { + "epoch": 3.5879478827361564, + "grad_norm": 9.294471096012794, + "learning_rate": 3.899954018520808e-06, + "loss": 0.334, + "step": 6609 + }, + { + "epoch": 3.5884907709011946, + "grad_norm": 12.555701694026313, + "learning_rate": 3.8971676597316775e-06, + "loss": 0.3707, + "step": 6610 + }, + { + "epoch": 3.5890336590662324, + "grad_norm": 10.572122929609495, + "learning_rate": 3.894382055750584e-06, + "loss": 0.5635, + "step": 6611 + }, + { + "epoch": 3.58957654723127, + "grad_norm": 9.702545363892131, + "learning_rate": 3.891597206922046e-06, + "loss": 0.3634, + "step": 6612 + }, + { + "epoch": 3.5901194353963084, + "grad_norm": 10.430633527197228, + "learning_rate": 3.888813113590496e-06, + "loss": 0.3777, + "step": 6613 + }, + { + "epoch": 3.5906623235613466, + "grad_norm": 16.05235504376427, + "learning_rate": 3.886029776100287e-06, + "loss": 0.5759, + "step": 6614 + }, + { + "epoch": 3.5912052117263844, + "grad_norm": 9.348911210053458, + "learning_rate": 3.883247194795653e-06, + "loss": 0.2619, + "step": 6615 + }, + { + "epoch": 3.591748099891422, + "grad_norm": 8.462834598999683, + "learning_rate": 3.8804653700207615e-06, + "loss": 0.3012, + "step": 6616 + }, + { + "epoch": 3.5922909880564604, + "grad_norm": 16.64696159175473, + "learning_rate": 3.87768430211966e-06, + "loss": 0.9137, + "step": 6617 + }, + { + "epoch": 3.5928338762214986, + "grad_norm": 12.223148301717917, + "learning_rate": 3.874903991436324e-06, + "loss": 0.4287, + "step": 6618 + }, + { + "epoch": 3.5933767643865364, + "grad_norm": 8.953587319006854, + "learning_rate": 3.872124438314624e-06, + "loss": 0.3932, + "step": 6619 + }, + { + "epoch": 3.593919652551574, + "grad_norm": 13.721613828536574, + "learning_rate": 3.869345643098341e-06, + "loss": 0.5449, + "step": 6620 + }, + { + "epoch": 3.5944625407166124, + "grad_norm": 10.931345128421508, + "learning_rate": 3.866567606131159e-06, + "loss": 0.3144, + "step": 6621 + }, + { + "epoch": 3.5950054288816506, + "grad_norm": 16.911421303952586, + "learning_rate": 3.86379032775667e-06, + "loss": 0.8115, + "step": 6622 + }, + { + "epoch": 3.5955483170466884, + "grad_norm": 13.179299764136056, + "learning_rate": 3.861013808318378e-06, + "loss": 0.557, + "step": 6623 + }, + { + "epoch": 3.596091205211726, + "grad_norm": 14.513913457624946, + "learning_rate": 3.858238048159674e-06, + "loss": 0.7132, + "step": 6624 + }, + { + "epoch": 3.5966340933767644, + "grad_norm": 18.20977238222908, + "learning_rate": 3.8554630476238805e-06, + "loss": 0.9001, + "step": 6625 + }, + { + "epoch": 3.5971769815418027, + "grad_norm": 13.411920870051427, + "learning_rate": 3.852688807054211e-06, + "loss": 0.6691, + "step": 6626 + }, + { + "epoch": 3.5977198697068404, + "grad_norm": 12.94431912351694, + "learning_rate": 3.849915326793786e-06, + "loss": 0.8245, + "step": 6627 + }, + { + "epoch": 3.598262757871878, + "grad_norm": 11.378136214690231, + "learning_rate": 3.847142607185636e-06, + "loss": 0.5369, + "step": 6628 + }, + { + "epoch": 3.5988056460369164, + "grad_norm": 18.107740640174118, + "learning_rate": 3.844370648572694e-06, + "loss": 0.5532, + "step": 6629 + }, + { + "epoch": 3.5993485342019547, + "grad_norm": 13.376518628320737, + "learning_rate": 3.841599451297802e-06, + "loss": 0.6283, + "step": 6630 + }, + { + "epoch": 3.5998914223669924, + "grad_norm": 17.68256418630599, + "learning_rate": 3.8388290157037034e-06, + "loss": 0.5916, + "step": 6631 + }, + { + "epoch": 3.6004343105320302, + "grad_norm": 9.436411945807468, + "learning_rate": 3.836059342133052e-06, + "loss": 0.488, + "step": 6632 + }, + { + "epoch": 3.6009771986970684, + "grad_norm": 12.744056075258795, + "learning_rate": 3.8332904309284045e-06, + "loss": 0.568, + "step": 6633 + }, + { + "epoch": 3.6015200868621067, + "grad_norm": 15.484356986063371, + "learning_rate": 3.830522282432227e-06, + "loss": 0.5619, + "step": 6634 + }, + { + "epoch": 3.6020629750271445, + "grad_norm": 22.466452052149346, + "learning_rate": 3.827754896986884e-06, + "loss": 0.754, + "step": 6635 + }, + { + "epoch": 3.6026058631921822, + "grad_norm": 10.907712543607193, + "learning_rate": 3.824988274934654e-06, + "loss": 0.3684, + "step": 6636 + }, + { + "epoch": 3.6031487513572205, + "grad_norm": 11.436863024139019, + "learning_rate": 3.822222416617714e-06, + "loss": 0.3772, + "step": 6637 + }, + { + "epoch": 3.6036916395222587, + "grad_norm": 8.249398777832742, + "learning_rate": 3.8194573223781515e-06, + "loss": 0.3908, + "step": 6638 + }, + { + "epoch": 3.6042345276872965, + "grad_norm": 10.129204140499496, + "learning_rate": 3.816692992557955e-06, + "loss": 0.5793, + "step": 6639 + }, + { + "epoch": 3.6047774158523342, + "grad_norm": 16.22320006976241, + "learning_rate": 3.8139294274990313e-06, + "loss": 0.7805, + "step": 6640 + }, + { + "epoch": 3.6053203040173725, + "grad_norm": 16.13099309974979, + "learning_rate": 3.8111666275431715e-06, + "loss": 0.5848, + "step": 6641 + }, + { + "epoch": 3.6058631921824107, + "grad_norm": 16.513602515759278, + "learning_rate": 3.8084045930320855e-06, + "loss": 0.5467, + "step": 6642 + }, + { + "epoch": 3.6064060803474485, + "grad_norm": 17.235827177127607, + "learning_rate": 3.8056433243073888e-06, + "loss": 0.7889, + "step": 6643 + }, + { + "epoch": 3.6069489685124863, + "grad_norm": 8.437058780004433, + "learning_rate": 3.8028828217105974e-06, + "loss": 0.4132, + "step": 6644 + }, + { + "epoch": 3.6074918566775245, + "grad_norm": 10.802576839366907, + "learning_rate": 3.800123085583135e-06, + "loss": 0.4123, + "step": 6645 + }, + { + "epoch": 3.6080347448425627, + "grad_norm": 11.08981584907767, + "learning_rate": 3.7973641162663276e-06, + "loss": 0.3947, + "step": 6646 + }, + { + "epoch": 3.6085776330076005, + "grad_norm": 9.95657545789331, + "learning_rate": 3.794605914101419e-06, + "loss": 0.6197, + "step": 6647 + }, + { + "epoch": 3.6091205211726383, + "grad_norm": 8.688663473460801, + "learning_rate": 3.791848479429534e-06, + "loss": 0.2914, + "step": 6648 + }, + { + "epoch": 3.6096634093376765, + "grad_norm": 13.569135821875664, + "learning_rate": 3.78909181259173e-06, + "loss": 0.4108, + "step": 6649 + }, + { + "epoch": 3.6102062975027147, + "grad_norm": 12.845103922950058, + "learning_rate": 3.7863359139289425e-06, + "loss": 0.5261, + "step": 6650 + }, + { + "epoch": 3.6107491856677525, + "grad_norm": 19.751616468555607, + "learning_rate": 3.783580783782035e-06, + "loss": 0.7023, + "step": 6651 + }, + { + "epoch": 3.6112920738327903, + "grad_norm": 13.094345121958364, + "learning_rate": 3.780826422491768e-06, + "loss": 0.5129, + "step": 6652 + }, + { + "epoch": 3.6118349619978285, + "grad_norm": 9.608815012306174, + "learning_rate": 3.7780728303987935e-06, + "loss": 0.3692, + "step": 6653 + }, + { + "epoch": 3.6123778501628667, + "grad_norm": 12.846531429400725, + "learning_rate": 3.7753200078436947e-06, + "loss": 0.5278, + "step": 6654 + }, + { + "epoch": 3.6129207383279045, + "grad_norm": 10.62078519937276, + "learning_rate": 3.7725679551669316e-06, + "loss": 0.5868, + "step": 6655 + }, + { + "epoch": 3.6134636264929423, + "grad_norm": 12.42176232319324, + "learning_rate": 3.769816672708895e-06, + "loss": 0.5746, + "step": 6656 + }, + { + "epoch": 3.6140065146579805, + "grad_norm": 15.37568549128414, + "learning_rate": 3.7670661608098545e-06, + "loss": 0.5671, + "step": 6657 + }, + { + "epoch": 3.6145494028230187, + "grad_norm": 14.651318253713534, + "learning_rate": 3.764316419810009e-06, + "loss": 0.4817, + "step": 6658 + }, + { + "epoch": 3.6150922909880565, + "grad_norm": 17.772625754725265, + "learning_rate": 3.761567450049448e-06, + "loss": 0.6106, + "step": 6659 + }, + { + "epoch": 3.6156351791530943, + "grad_norm": 12.785953562325862, + "learning_rate": 3.7588192518681664e-06, + "loss": 0.458, + "step": 6660 + }, + { + "epoch": 3.6161780673181325, + "grad_norm": 14.872223886784575, + "learning_rate": 3.7560718256060704e-06, + "loss": 0.5833, + "step": 6661 + }, + { + "epoch": 3.6167209554831707, + "grad_norm": 10.482699567034725, + "learning_rate": 3.753325171602955e-06, + "loss": 0.4262, + "step": 6662 + }, + { + "epoch": 3.6172638436482085, + "grad_norm": 12.097306912879557, + "learning_rate": 3.7505792901985417e-06, + "loss": 0.4847, + "step": 6663 + }, + { + "epoch": 3.6178067318132463, + "grad_norm": 11.56080631055532, + "learning_rate": 3.7478341817324414e-06, + "loss": 0.4421, + "step": 6664 + }, + { + "epoch": 3.6183496199782845, + "grad_norm": 17.08246749281531, + "learning_rate": 3.7450898465441744e-06, + "loss": 0.6724, + "step": 6665 + }, + { + "epoch": 3.6188925081433228, + "grad_norm": 14.083035375418559, + "learning_rate": 3.742346284973164e-06, + "loss": 0.7099, + "step": 6666 + }, + { + "epoch": 3.6194353963083605, + "grad_norm": 12.257812595787682, + "learning_rate": 3.7396034973587392e-06, + "loss": 0.5108, + "step": 6667 + }, + { + "epoch": 3.6199782844733983, + "grad_norm": 16.583982679880926, + "learning_rate": 3.7368614840401297e-06, + "loss": 0.5608, + "step": 6668 + }, + { + "epoch": 3.6205211726384365, + "grad_norm": 9.31804613123421, + "learning_rate": 3.734120245356475e-06, + "loss": 0.2673, + "step": 6669 + }, + { + "epoch": 3.6210640608034748, + "grad_norm": 8.73729193728082, + "learning_rate": 3.7313797816468145e-06, + "loss": 0.2867, + "step": 6670 + }, + { + "epoch": 3.6216069489685125, + "grad_norm": 10.574625431794137, + "learning_rate": 3.728640093250093e-06, + "loss": 0.6508, + "step": 6671 + }, + { + "epoch": 3.6221498371335503, + "grad_norm": 13.116215370859619, + "learning_rate": 3.7259011805051605e-06, + "loss": 0.8055, + "step": 6672 + }, + { + "epoch": 3.6226927252985885, + "grad_norm": 12.674214863430533, + "learning_rate": 3.723163043750768e-06, + "loss": 0.4149, + "step": 6673 + }, + { + "epoch": 3.6232356134636268, + "grad_norm": 15.457787982715566, + "learning_rate": 3.720425683325576e-06, + "loss": 0.5407, + "step": 6674 + }, + { + "epoch": 3.6237785016286646, + "grad_norm": 13.509659223593033, + "learning_rate": 3.7176890995681426e-06, + "loss": 0.3499, + "step": 6675 + }, + { + "epoch": 3.6243213897937023, + "grad_norm": 16.313287760559906, + "learning_rate": 3.714953292816934e-06, + "loss": 0.697, + "step": 6676 + }, + { + "epoch": 3.6248642779587406, + "grad_norm": 14.440828134958558, + "learning_rate": 3.712218263410319e-06, + "loss": 0.5698, + "step": 6677 + }, + { + "epoch": 3.6254071661237783, + "grad_norm": 9.354441770636624, + "learning_rate": 3.7094840116865693e-06, + "loss": 0.3452, + "step": 6678 + }, + { + "epoch": 3.6259500542888166, + "grad_norm": 9.584690153683102, + "learning_rate": 3.706750537983861e-06, + "loss": 0.4305, + "step": 6679 + }, + { + "epoch": 3.6264929424538543, + "grad_norm": 18.051709305408423, + "learning_rate": 3.7040178426402817e-06, + "loss": 0.7387, + "step": 6680 + }, + { + "epoch": 3.6270358306188926, + "grad_norm": 10.05950366415881, + "learning_rate": 3.7012859259938074e-06, + "loss": 0.3443, + "step": 6681 + }, + { + "epoch": 3.6275787187839303, + "grad_norm": 12.034931076516903, + "learning_rate": 3.698554788382328e-06, + "loss": 0.542, + "step": 6682 + }, + { + "epoch": 3.6281216069489686, + "grad_norm": 12.814483154092642, + "learning_rate": 3.6958244301436353e-06, + "loss": 0.6022, + "step": 6683 + }, + { + "epoch": 3.6286644951140063, + "grad_norm": 12.602310642914164, + "learning_rate": 3.693094851615421e-06, + "loss": 0.5764, + "step": 6684 + }, + { + "epoch": 3.6292073832790446, + "grad_norm": 13.840451235628782, + "learning_rate": 3.6903660531352947e-06, + "loss": 0.673, + "step": 6685 + }, + { + "epoch": 3.6297502714440824, + "grad_norm": 11.319388737736466, + "learning_rate": 3.6876380350407428e-06, + "loss": 0.5201, + "step": 6686 + }, + { + "epoch": 3.6302931596091206, + "grad_norm": 8.195460173712483, + "learning_rate": 3.6849107976691867e-06, + "loss": 0.2834, + "step": 6687 + }, + { + "epoch": 3.6308360477741584, + "grad_norm": 11.24454405051031, + "learning_rate": 3.6821843413579217e-06, + "loss": 0.4174, + "step": 6688 + }, + { + "epoch": 3.6313789359391966, + "grad_norm": 13.272520776849182, + "learning_rate": 3.679458666444169e-06, + "loss": 0.5333, + "step": 6689 + }, + { + "epoch": 3.6319218241042344, + "grad_norm": 11.044629755858516, + "learning_rate": 3.6767337732650454e-06, + "loss": 0.4734, + "step": 6690 + }, + { + "epoch": 3.6324647122692726, + "grad_norm": 9.202516399757954, + "learning_rate": 3.6740096621575604e-06, + "loss": 0.28, + "step": 6691 + }, + { + "epoch": 3.6330076004343104, + "grad_norm": 11.820877288548346, + "learning_rate": 3.6712863334586478e-06, + "loss": 0.2672, + "step": 6692 + }, + { + "epoch": 3.6335504885993486, + "grad_norm": 11.093852797013197, + "learning_rate": 3.668563787505123e-06, + "loss": 0.3923, + "step": 6693 + }, + { + "epoch": 3.6340933767643864, + "grad_norm": 10.10902067133091, + "learning_rate": 3.6658420246337256e-06, + "loss": 0.3633, + "step": 6694 + }, + { + "epoch": 3.6346362649294246, + "grad_norm": 11.211935387408433, + "learning_rate": 3.663121045181075e-06, + "loss": 0.4325, + "step": 6695 + }, + { + "epoch": 3.6351791530944624, + "grad_norm": 7.899984820961417, + "learning_rate": 3.660400849483716e-06, + "loss": 0.2753, + "step": 6696 + }, + { + "epoch": 3.6357220412595006, + "grad_norm": 15.593004160513892, + "learning_rate": 3.6576814378780844e-06, + "loss": 0.5612, + "step": 6697 + }, + { + "epoch": 3.6362649294245384, + "grad_norm": 14.607948256302272, + "learning_rate": 3.654962810700521e-06, + "loss": 0.5311, + "step": 6698 + }, + { + "epoch": 3.6368078175895766, + "grad_norm": 14.159583660700658, + "learning_rate": 3.6522449682872687e-06, + "loss": 0.4704, + "step": 6699 + }, + { + "epoch": 3.6373507057546144, + "grad_norm": 12.5139018167431, + "learning_rate": 3.6495279109744764e-06, + "loss": 0.5437, + "step": 6700 + }, + { + "epoch": 3.6378935939196526, + "grad_norm": 13.188903276553058, + "learning_rate": 3.6468116390981922e-06, + "loss": 0.5713, + "step": 6701 + }, + { + "epoch": 3.6384364820846904, + "grad_norm": 9.269396910628474, + "learning_rate": 3.644096152994371e-06, + "loss": 0.3362, + "step": 6702 + }, + { + "epoch": 3.6389793702497286, + "grad_norm": 15.76289925225522, + "learning_rate": 3.6413814529988666e-06, + "loss": 0.6549, + "step": 6703 + }, + { + "epoch": 3.6395222584147664, + "grad_norm": 14.218821330155238, + "learning_rate": 3.6386675394474393e-06, + "loss": 0.5017, + "step": 6704 + }, + { + "epoch": 3.6400651465798046, + "grad_norm": 13.035114169534788, + "learning_rate": 3.6359544126757483e-06, + "loss": 0.4331, + "step": 6705 + }, + { + "epoch": 3.6406080347448424, + "grad_norm": 20.82770787825795, + "learning_rate": 3.6332420730193584e-06, + "loss": 0.7242, + "step": 6706 + }, + { + "epoch": 3.6411509229098806, + "grad_norm": 14.084479195896975, + "learning_rate": 3.630530520813735e-06, + "loss": 0.5915, + "step": 6707 + }, + { + "epoch": 3.6416938110749184, + "grad_norm": 12.24391802241245, + "learning_rate": 3.6278197563942497e-06, + "loss": 0.3637, + "step": 6708 + }, + { + "epoch": 3.6422366992399566, + "grad_norm": 19.068246989978615, + "learning_rate": 3.625109780096173e-06, + "loss": 0.5401, + "step": 6709 + }, + { + "epoch": 3.6427795874049944, + "grad_norm": 12.923571390979863, + "learning_rate": 3.6224005922546777e-06, + "loss": 0.4372, + "step": 6710 + }, + { + "epoch": 3.6433224755700326, + "grad_norm": 18.72956431478281, + "learning_rate": 3.6196921932048424e-06, + "loss": 0.5212, + "step": 6711 + }, + { + "epoch": 3.6438653637350704, + "grad_norm": 9.246217525160361, + "learning_rate": 3.6169845832816443e-06, + "loss": 0.2875, + "step": 6712 + }, + { + "epoch": 3.6444082519001086, + "grad_norm": 12.538873429253352, + "learning_rate": 3.6142777628199667e-06, + "loss": 0.7951, + "step": 6713 + }, + { + "epoch": 3.6449511400651464, + "grad_norm": 9.970157389027023, + "learning_rate": 3.611571732154592e-06, + "loss": 0.3658, + "step": 6714 + }, + { + "epoch": 3.6454940282301846, + "grad_norm": 12.203072582682385, + "learning_rate": 3.6088664916202077e-06, + "loss": 0.5721, + "step": 6715 + }, + { + "epoch": 3.6460369163952224, + "grad_norm": 12.279248739912763, + "learning_rate": 3.606162041551402e-06, + "loss": 0.4408, + "step": 6716 + }, + { + "epoch": 3.6465798045602607, + "grad_norm": 12.802459902824975, + "learning_rate": 3.6034583822826607e-06, + "loss": 0.6177, + "step": 6717 + }, + { + "epoch": 3.6471226927252984, + "grad_norm": 14.71635978144862, + "learning_rate": 3.6007555141483897e-06, + "loss": 0.6333, + "step": 6718 + }, + { + "epoch": 3.6476655808903367, + "grad_norm": 15.254287066882977, + "learning_rate": 3.5980534374828667e-06, + "loss": 0.6541, + "step": 6719 + }, + { + "epoch": 3.6482084690553744, + "grad_norm": 17.306678251097242, + "learning_rate": 3.595352152620306e-06, + "loss": 0.5605, + "step": 6720 + }, + { + "epoch": 3.6487513572204127, + "grad_norm": 13.28374885202886, + "learning_rate": 3.5926516598947946e-06, + "loss": 0.553, + "step": 6721 + }, + { + "epoch": 3.6492942453854504, + "grad_norm": 15.674888174232077, + "learning_rate": 3.5899519596403356e-06, + "loss": 0.5473, + "step": 6722 + }, + { + "epoch": 3.6498371335504887, + "grad_norm": 9.387854198306167, + "learning_rate": 3.58725305219084e-06, + "loss": 0.4405, + "step": 6723 + }, + { + "epoch": 3.6503800217155264, + "grad_norm": 9.686200542254724, + "learning_rate": 3.5845549378801005e-06, + "loss": 0.4019, + "step": 6724 + }, + { + "epoch": 3.6509229098805647, + "grad_norm": 15.112849538872657, + "learning_rate": 3.5818576170418373e-06, + "loss": 0.7064, + "step": 6725 + }, + { + "epoch": 3.6514657980456025, + "grad_norm": 11.208536381794502, + "learning_rate": 3.5791610900096465e-06, + "loss": 0.3352, + "step": 6726 + }, + { + "epoch": 3.6520086862106407, + "grad_norm": 7.334988680617685, + "learning_rate": 3.5764653571170527e-06, + "loss": 0.288, + "step": 6727 + }, + { + "epoch": 3.6525515743756785, + "grad_norm": 11.838685637818697, + "learning_rate": 3.573770418697454e-06, + "loss": 0.6447, + "step": 6728 + }, + { + "epoch": 3.6530944625407167, + "grad_norm": 15.976115688824532, + "learning_rate": 3.5710762750841756e-06, + "loss": 0.4407, + "step": 6729 + }, + { + "epoch": 3.6536373507057545, + "grad_norm": 12.25069694802214, + "learning_rate": 3.5683829266104318e-06, + "loss": 0.4395, + "step": 6730 + }, + { + "epoch": 3.6541802388707927, + "grad_norm": 9.192123102498591, + "learning_rate": 3.5656903736093328e-06, + "loss": 0.3363, + "step": 6731 + }, + { + "epoch": 3.6547231270358305, + "grad_norm": 14.524314713398583, + "learning_rate": 3.5629986164139095e-06, + "loss": 0.3778, + "step": 6732 + }, + { + "epoch": 3.6552660152008687, + "grad_norm": 16.927339643078007, + "learning_rate": 3.5603076553570692e-06, + "loss": 0.6248, + "step": 6733 + }, + { + "epoch": 3.6558089033659065, + "grad_norm": 12.599717116787803, + "learning_rate": 3.5576174907716455e-06, + "loss": 0.6944, + "step": 6734 + }, + { + "epoch": 3.6563517915309447, + "grad_norm": 10.646444115088963, + "learning_rate": 3.5549281229903575e-06, + "loss": 0.3772, + "step": 6735 + }, + { + "epoch": 3.6568946796959825, + "grad_norm": 15.498764867333419, + "learning_rate": 3.552239552345831e-06, + "loss": 0.5415, + "step": 6736 + }, + { + "epoch": 3.6574375678610207, + "grad_norm": 15.747483640049493, + "learning_rate": 3.5495517791705937e-06, + "loss": 0.4526, + "step": 6737 + }, + { + "epoch": 3.6579804560260585, + "grad_norm": 8.811218308037992, + "learning_rate": 3.546864803797072e-06, + "loss": 0.2853, + "step": 6738 + }, + { + "epoch": 3.6585233441910967, + "grad_norm": 13.307618617651949, + "learning_rate": 3.5441786265576006e-06, + "loss": 0.5929, + "step": 6739 + }, + { + "epoch": 3.6590662323561345, + "grad_norm": 16.35655181341609, + "learning_rate": 3.541493247784398e-06, + "loss": 0.5969, + "step": 6740 + }, + { + "epoch": 3.6596091205211727, + "grad_norm": 13.468394744914283, + "learning_rate": 3.5388086678096077e-06, + "loss": 0.7131, + "step": 6741 + }, + { + "epoch": 3.6601520086862105, + "grad_norm": 10.610479158018709, + "learning_rate": 3.5361248869652586e-06, + "loss": 0.3491, + "step": 6742 + }, + { + "epoch": 3.6606948968512487, + "grad_norm": 10.63033092512971, + "learning_rate": 3.5334419055832857e-06, + "loss": 0.4148, + "step": 6743 + }, + { + "epoch": 3.6612377850162865, + "grad_norm": 12.066673558862, + "learning_rate": 3.530759723995524e-06, + "loss": 0.5922, + "step": 6744 + }, + { + "epoch": 3.6617806731813247, + "grad_norm": 13.669464818674667, + "learning_rate": 3.5280783425337097e-06, + "loss": 0.4984, + "step": 6745 + }, + { + "epoch": 3.6623235613463625, + "grad_norm": 11.385901052590954, + "learning_rate": 3.525397761529481e-06, + "loss": 0.4451, + "step": 6746 + }, + { + "epoch": 3.6628664495114007, + "grad_norm": 13.127975944644733, + "learning_rate": 3.5227179813143763e-06, + "loss": 0.4786, + "step": 6747 + }, + { + "epoch": 3.6634093376764385, + "grad_norm": 8.286555642541957, + "learning_rate": 3.520039002219835e-06, + "loss": 0.3308, + "step": 6748 + }, + { + "epoch": 3.6639522258414767, + "grad_norm": 13.989369716944896, + "learning_rate": 3.5173608245771964e-06, + "loss": 0.6301, + "step": 6749 + }, + { + "epoch": 3.6644951140065145, + "grad_norm": 11.280241380801893, + "learning_rate": 3.514683448717704e-06, + "loss": 0.4333, + "step": 6750 + }, + { + "epoch": 3.6650380021715527, + "grad_norm": 14.109069012865124, + "learning_rate": 3.5120068749724977e-06, + "loss": 0.4928, + "step": 6751 + }, + { + "epoch": 3.6655808903365905, + "grad_norm": 9.420479780236201, + "learning_rate": 3.509331103672623e-06, + "loss": 0.3409, + "step": 6752 + }, + { + "epoch": 3.6661237785016287, + "grad_norm": 13.634334295467264, + "learning_rate": 3.5066561351490202e-06, + "loss": 0.4104, + "step": 6753 + }, + { + "epoch": 3.6666666666666665, + "grad_norm": 12.151334809320293, + "learning_rate": 3.503981969732536e-06, + "loss": 0.6371, + "step": 6754 + }, + { + "epoch": 3.6672095548317047, + "grad_norm": 15.410529391564157, + "learning_rate": 3.5013086077539125e-06, + "loss": 0.6359, + "step": 6755 + }, + { + "epoch": 3.6677524429967425, + "grad_norm": 14.234740703322796, + "learning_rate": 3.4986360495438033e-06, + "loss": 0.5397, + "step": 6756 + }, + { + "epoch": 3.6682953311617807, + "grad_norm": 12.465036402535198, + "learning_rate": 3.4959642954327435e-06, + "loss": 0.5274, + "step": 6757 + }, + { + "epoch": 3.6688382193268185, + "grad_norm": 12.485874013130712, + "learning_rate": 3.4932933457511907e-06, + "loss": 0.4669, + "step": 6758 + }, + { + "epoch": 3.6693811074918568, + "grad_norm": 13.300754823528772, + "learning_rate": 3.490623200829485e-06, + "loss": 0.5991, + "step": 6759 + }, + { + "epoch": 3.6699239956568945, + "grad_norm": 13.6162114238038, + "learning_rate": 3.4879538609978724e-06, + "loss": 0.5753, + "step": 6760 + }, + { + "epoch": 3.6704668838219328, + "grad_norm": 15.055930735004507, + "learning_rate": 3.4852853265865117e-06, + "loss": 0.5937, + "step": 6761 + }, + { + "epoch": 3.6710097719869705, + "grad_norm": 19.021493329558147, + "learning_rate": 3.4826175979254382e-06, + "loss": 0.7309, + "step": 6762 + }, + { + "epoch": 3.6715526601520088, + "grad_norm": 7.879252217720616, + "learning_rate": 3.4799506753446133e-06, + "loss": 0.2324, + "step": 6763 + }, + { + "epoch": 3.6720955483170465, + "grad_norm": 15.566429874979375, + "learning_rate": 3.477284559173875e-06, + "loss": 0.8685, + "step": 6764 + }, + { + "epoch": 3.6726384364820848, + "grad_norm": 14.932288315207366, + "learning_rate": 3.474619249742983e-06, + "loss": 0.4121, + "step": 6765 + }, + { + "epoch": 3.6731813246471225, + "grad_norm": 12.826517078026129, + "learning_rate": 3.471954747381577e-06, + "loss": 0.4172, + "step": 6766 + }, + { + "epoch": 3.6737242128121608, + "grad_norm": 11.27094341605798, + "learning_rate": 3.4692910524192145e-06, + "loss": 0.6047, + "step": 6767 + }, + { + "epoch": 3.6742671009771986, + "grad_norm": 10.227384928715546, + "learning_rate": 3.4666281651853426e-06, + "loss": 0.4978, + "step": 6768 + }, + { + "epoch": 3.6748099891422368, + "grad_norm": 9.140068130601126, + "learning_rate": 3.463966086009313e-06, + "loss": 0.3314, + "step": 6769 + }, + { + "epoch": 3.6753528773072746, + "grad_norm": 8.57213348310477, + "learning_rate": 3.4613048152203776e-06, + "loss": 0.3135, + "step": 6770 + }, + { + "epoch": 3.675895765472313, + "grad_norm": 18.06864303821088, + "learning_rate": 3.4586443531476777e-06, + "loss": 0.7083, + "step": 6771 + }, + { + "epoch": 3.6764386536373506, + "grad_norm": 12.06612974538307, + "learning_rate": 3.455984700120272e-06, + "loss": 0.4263, + "step": 6772 + }, + { + "epoch": 3.676981541802389, + "grad_norm": 15.51068609579363, + "learning_rate": 3.453325856467108e-06, + "loss": 0.5577, + "step": 6773 + }, + { + "epoch": 3.6775244299674266, + "grad_norm": 10.276565099526616, + "learning_rate": 3.4506678225170363e-06, + "loss": 0.3197, + "step": 6774 + }, + { + "epoch": 3.678067318132465, + "grad_norm": 9.001248775476501, + "learning_rate": 3.4480105985988065e-06, + "loss": 0.376, + "step": 6775 + }, + { + "epoch": 3.6786102062975026, + "grad_norm": 10.679849220418427, + "learning_rate": 3.4453541850410686e-06, + "loss": 0.3383, + "step": 6776 + }, + { + "epoch": 3.679153094462541, + "grad_norm": 10.331306790662499, + "learning_rate": 3.4426985821723714e-06, + "loss": 0.5359, + "step": 6777 + }, + { + "epoch": 3.6796959826275786, + "grad_norm": 12.802407333669652, + "learning_rate": 3.4400437903211635e-06, + "loss": 0.4397, + "step": 6778 + }, + { + "epoch": 3.680238870792617, + "grad_norm": 14.73981252395514, + "learning_rate": 3.4373898098157953e-06, + "loss": 0.5717, + "step": 6779 + }, + { + "epoch": 3.6807817589576546, + "grad_norm": 16.362575863101522, + "learning_rate": 3.434736640984515e-06, + "loss": 0.5125, + "step": 6780 + }, + { + "epoch": 3.681324647122693, + "grad_norm": 11.576699027697012, + "learning_rate": 3.4320842841554702e-06, + "loss": 0.3924, + "step": 6781 + }, + { + "epoch": 3.6818675352877306, + "grad_norm": 12.430829059754513, + "learning_rate": 3.4294327396567085e-06, + "loss": 0.5816, + "step": 6782 + }, + { + "epoch": 3.682410423452769, + "grad_norm": 12.320695413208654, + "learning_rate": 3.4267820078161785e-06, + "loss": 0.5416, + "step": 6783 + }, + { + "epoch": 3.6829533116178066, + "grad_norm": 13.000253896056515, + "learning_rate": 3.4241320889617245e-06, + "loss": 0.536, + "step": 6784 + }, + { + "epoch": 3.683496199782845, + "grad_norm": 10.91584072091396, + "learning_rate": 3.4214829834210948e-06, + "loss": 0.4491, + "step": 6785 + }, + { + "epoch": 3.6840390879478826, + "grad_norm": 13.085054837243629, + "learning_rate": 3.418834691521934e-06, + "loss": 0.378, + "step": 6786 + }, + { + "epoch": 3.684581976112921, + "grad_norm": 11.116151442052336, + "learning_rate": 3.4161872135917873e-06, + "loss": 0.4053, + "step": 6787 + }, + { + "epoch": 3.6851248642779586, + "grad_norm": 14.044458874643864, + "learning_rate": 3.4135405499580998e-06, + "loss": 0.3851, + "step": 6788 + }, + { + "epoch": 3.685667752442997, + "grad_norm": 15.351406534798858, + "learning_rate": 3.410894700948214e-06, + "loss": 0.7129, + "step": 6789 + }, + { + "epoch": 3.6862106406080346, + "grad_norm": 18.553137629768052, + "learning_rate": 3.408249666889373e-06, + "loss": 0.9605, + "step": 6790 + }, + { + "epoch": 3.686753528773073, + "grad_norm": 12.059229166245107, + "learning_rate": 3.4056054481087187e-06, + "loss": 0.6246, + "step": 6791 + }, + { + "epoch": 3.6872964169381106, + "grad_norm": 8.164629976709701, + "learning_rate": 3.4029620449332925e-06, + "loss": 0.2484, + "step": 6792 + }, + { + "epoch": 3.687839305103149, + "grad_norm": 12.956574155251525, + "learning_rate": 3.400319457690031e-06, + "loss": 0.4846, + "step": 6793 + }, + { + "epoch": 3.6883821932681866, + "grad_norm": 13.218562644434163, + "learning_rate": 3.397677686705785e-06, + "loss": 0.4956, + "step": 6794 + }, + { + "epoch": 3.688925081433225, + "grad_norm": 12.52521210536022, + "learning_rate": 3.3950367323072777e-06, + "loss": 0.5609, + "step": 6795 + }, + { + "epoch": 3.6894679695982626, + "grad_norm": 14.858442377207329, + "learning_rate": 3.3923965948211612e-06, + "loss": 0.5361, + "step": 6796 + }, + { + "epoch": 3.690010857763301, + "grad_norm": 10.357147809279667, + "learning_rate": 3.3897572745739584e-06, + "loss": 0.4834, + "step": 6797 + }, + { + "epoch": 3.6905537459283386, + "grad_norm": 10.579748653808505, + "learning_rate": 3.3871187718921173e-06, + "loss": 0.4685, + "step": 6798 + }, + { + "epoch": 3.691096634093377, + "grad_norm": 16.620322523965804, + "learning_rate": 3.3844810871019617e-06, + "loss": 0.6373, + "step": 6799 + }, + { + "epoch": 3.6916395222584146, + "grad_norm": 15.999425270589017, + "learning_rate": 3.3818442205297262e-06, + "loss": 0.895, + "step": 6800 + }, + { + "epoch": 3.692182410423453, + "grad_norm": 10.78206305263471, + "learning_rate": 3.3792081725015525e-06, + "loss": 0.5564, + "step": 6801 + }, + { + "epoch": 3.6927252985884906, + "grad_norm": 11.641189355388065, + "learning_rate": 3.376572943343457e-06, + "loss": 0.5319, + "step": 6802 + }, + { + "epoch": 3.693268186753529, + "grad_norm": 14.695430550731649, + "learning_rate": 3.3739385333813824e-06, + "loss": 0.5868, + "step": 6803 + }, + { + "epoch": 3.6938110749185666, + "grad_norm": 10.078064673313023, + "learning_rate": 3.3713049429411435e-06, + "loss": 0.5002, + "step": 6804 + }, + { + "epoch": 3.694353963083605, + "grad_norm": 13.958804490496405, + "learning_rate": 3.368672172348478e-06, + "loss": 0.5521, + "step": 6805 + }, + { + "epoch": 3.6948968512486426, + "grad_norm": 14.349798638841735, + "learning_rate": 3.366040221929007e-06, + "loss": 0.8658, + "step": 6806 + }, + { + "epoch": 3.695439739413681, + "grad_norm": 19.918583954387678, + "learning_rate": 3.363409092008253e-06, + "loss": 0.7743, + "step": 6807 + }, + { + "epoch": 3.6959826275787186, + "grad_norm": 17.742647882095138, + "learning_rate": 3.3607787829116446e-06, + "loss": 0.4515, + "step": 6808 + }, + { + "epoch": 3.696525515743757, + "grad_norm": 10.287874260684818, + "learning_rate": 3.35814929496449e-06, + "loss": 0.4691, + "step": 6809 + }, + { + "epoch": 3.6970684039087947, + "grad_norm": 13.988040829949796, + "learning_rate": 3.3555206284920226e-06, + "loss": 0.5869, + "step": 6810 + }, + { + "epoch": 3.697611292073833, + "grad_norm": 12.599509307267757, + "learning_rate": 3.352892783819347e-06, + "loss": 0.5899, + "step": 6811 + }, + { + "epoch": 3.6981541802388707, + "grad_norm": 9.433309182729886, + "learning_rate": 3.3502657612714894e-06, + "loss": 0.3442, + "step": 6812 + }, + { + "epoch": 3.698697068403909, + "grad_norm": 16.010015776538683, + "learning_rate": 3.3476395611733616e-06, + "loss": 0.5644, + "step": 6813 + }, + { + "epoch": 3.6992399565689467, + "grad_norm": 12.170086596108076, + "learning_rate": 3.345014183849774e-06, + "loss": 0.5416, + "step": 6814 + }, + { + "epoch": 3.699782844733985, + "grad_norm": 9.466907002380811, + "learning_rate": 3.342389629625439e-06, + "loss": 0.5223, + "step": 6815 + }, + { + "epoch": 3.7003257328990227, + "grad_norm": 10.374676588070637, + "learning_rate": 3.3397658988249647e-06, + "loss": 0.327, + "step": 6816 + }, + { + "epoch": 3.700868621064061, + "grad_norm": 16.837753149163387, + "learning_rate": 3.337142991772859e-06, + "loss": 0.3853, + "step": 6817 + }, + { + "epoch": 3.7014115092290987, + "grad_norm": 19.187866841083977, + "learning_rate": 3.334520908793527e-06, + "loss": 0.7495, + "step": 6818 + }, + { + "epoch": 3.701954397394137, + "grad_norm": 12.434378548251997, + "learning_rate": 3.3318996502112733e-06, + "loss": 0.4615, + "step": 6819 + }, + { + "epoch": 3.7024972855591747, + "grad_norm": 14.472283943179146, + "learning_rate": 3.3292792163502973e-06, + "loss": 0.6001, + "step": 6820 + }, + { + "epoch": 3.703040173724213, + "grad_norm": 10.597662524177974, + "learning_rate": 3.3266596075346993e-06, + "loss": 0.4247, + "step": 6821 + }, + { + "epoch": 3.7035830618892507, + "grad_norm": 10.287091606357066, + "learning_rate": 3.324040824088477e-06, + "loss": 0.3372, + "step": 6822 + }, + { + "epoch": 3.704125950054289, + "grad_norm": 16.36152453578531, + "learning_rate": 3.3214228663355252e-06, + "loss": 0.4507, + "step": 6823 + }, + { + "epoch": 3.7046688382193267, + "grad_norm": 11.166011131196704, + "learning_rate": 3.3188057345996372e-06, + "loss": 0.4297, + "step": 6824 + }, + { + "epoch": 3.705211726384365, + "grad_norm": 13.97058303490548, + "learning_rate": 3.316189429204505e-06, + "loss": 0.5737, + "step": 6825 + }, + { + "epoch": 3.7057546145494027, + "grad_norm": 13.078068910284376, + "learning_rate": 3.313573950473713e-06, + "loss": 0.3998, + "step": 6826 + }, + { + "epoch": 3.706297502714441, + "grad_norm": 17.144621177815008, + "learning_rate": 3.3109592987307583e-06, + "loss": 0.6534, + "step": 6827 + }, + { + "epoch": 3.7068403908794787, + "grad_norm": 12.887234362542056, + "learning_rate": 3.308345474299014e-06, + "loss": 0.6388, + "step": 6828 + }, + { + "epoch": 3.707383279044517, + "grad_norm": 14.117206302860215, + "learning_rate": 3.305732477501766e-06, + "loss": 0.8498, + "step": 6829 + }, + { + "epoch": 3.7079261672095547, + "grad_norm": 16.668445763886474, + "learning_rate": 3.3031203086621953e-06, + "loss": 0.5491, + "step": 6830 + }, + { + "epoch": 3.708469055374593, + "grad_norm": 9.687167039444047, + "learning_rate": 3.300508968103375e-06, + "loss": 0.3493, + "step": 6831 + }, + { + "epoch": 3.7090119435396307, + "grad_norm": 8.884631092991913, + "learning_rate": 3.2978984561482898e-06, + "loss": 0.3227, + "step": 6832 + }, + { + "epoch": 3.709554831704669, + "grad_norm": 10.708989687832828, + "learning_rate": 3.295288773119797e-06, + "loss": 0.5266, + "step": 6833 + }, + { + "epoch": 3.7100977198697067, + "grad_norm": 11.602581058072978, + "learning_rate": 3.292679919340682e-06, + "loss": 0.4903, + "step": 6834 + }, + { + "epoch": 3.710640608034745, + "grad_norm": 11.22107577843576, + "learning_rate": 3.2900718951335973e-06, + "loss": 0.2758, + "step": 6835 + }, + { + "epoch": 3.7111834961997827, + "grad_norm": 11.654780601665442, + "learning_rate": 3.2874647008211223e-06, + "loss": 0.4478, + "step": 6836 + }, + { + "epoch": 3.711726384364821, + "grad_norm": 13.952341454584447, + "learning_rate": 3.2848583367257037e-06, + "loss": 0.6784, + "step": 6837 + }, + { + "epoch": 3.7122692725298587, + "grad_norm": 9.937530151430312, + "learning_rate": 3.2822528031697123e-06, + "loss": 0.3417, + "step": 6838 + }, + { + "epoch": 3.712812160694897, + "grad_norm": 12.66955971380617, + "learning_rate": 3.279648100475403e-06, + "loss": 0.5978, + "step": 6839 + }, + { + "epoch": 3.7133550488599347, + "grad_norm": 11.746942398326137, + "learning_rate": 3.2770442289649218e-06, + "loss": 0.422, + "step": 6840 + }, + { + "epoch": 3.713897937024973, + "grad_norm": 9.742871784049646, + "learning_rate": 3.274441188960331e-06, + "loss": 0.2878, + "step": 6841 + }, + { + "epoch": 3.7144408251900107, + "grad_norm": 16.102195766631716, + "learning_rate": 3.271838980783567e-06, + "loss": 0.6011, + "step": 6842 + }, + { + "epoch": 3.714983713355049, + "grad_norm": 12.78519494193334, + "learning_rate": 3.2692376047564843e-06, + "loss": 0.5416, + "step": 6843 + }, + { + "epoch": 3.7155266015200867, + "grad_norm": 7.51189016165619, + "learning_rate": 3.266637061200821e-06, + "loss": 0.2993, + "step": 6844 + }, + { + "epoch": 3.716069489685125, + "grad_norm": 11.448398733111881, + "learning_rate": 3.264037350438218e-06, + "loss": 0.469, + "step": 6845 + }, + { + "epoch": 3.7166123778501627, + "grad_norm": 11.057137423687962, + "learning_rate": 3.2614384727902106e-06, + "loss": 0.5315, + "step": 6846 + }, + { + "epoch": 3.717155266015201, + "grad_norm": 10.948114702379822, + "learning_rate": 3.2588404285782327e-06, + "loss": 0.5224, + "step": 6847 + }, + { + "epoch": 3.7176981541802387, + "grad_norm": 14.169026887876377, + "learning_rate": 3.2562432181236182e-06, + "loss": 0.4101, + "step": 6848 + }, + { + "epoch": 3.718241042345277, + "grad_norm": 10.211418365179567, + "learning_rate": 3.253646841747583e-06, + "loss": 0.356, + "step": 6849 + }, + { + "epoch": 3.7187839305103148, + "grad_norm": 10.750722744873066, + "learning_rate": 3.251051299771263e-06, + "loss": 0.5381, + "step": 6850 + }, + { + "epoch": 3.719326818675353, + "grad_norm": 10.07832966516926, + "learning_rate": 3.2484565925156732e-06, + "loss": 0.3729, + "step": 6851 + }, + { + "epoch": 3.7198697068403908, + "grad_norm": 12.95382559219359, + "learning_rate": 3.2458627203017335e-06, + "loss": 0.5247, + "step": 6852 + }, + { + "epoch": 3.720412595005429, + "grad_norm": 11.982999422321642, + "learning_rate": 3.243269683450256e-06, + "loss": 0.4333, + "step": 6853 + }, + { + "epoch": 3.7209554831704668, + "grad_norm": 14.445578614379583, + "learning_rate": 3.240677482281953e-06, + "loss": 0.6296, + "step": 6854 + }, + { + "epoch": 3.721498371335505, + "grad_norm": 14.769553336755148, + "learning_rate": 3.2380861171174317e-06, + "loss": 0.48, + "step": 6855 + }, + { + "epoch": 3.7220412595005428, + "grad_norm": 13.97227360884286, + "learning_rate": 3.2354955882771965e-06, + "loss": 0.8019, + "step": 6856 + }, + { + "epoch": 3.722584147665581, + "grad_norm": 10.17994965485879, + "learning_rate": 3.2329058960816483e-06, + "loss": 0.4003, + "step": 6857 + }, + { + "epoch": 3.7231270358306188, + "grad_norm": 11.647057134479718, + "learning_rate": 3.2303170408510832e-06, + "loss": 0.367, + "step": 6858 + }, + { + "epoch": 3.723669923995657, + "grad_norm": 13.75957119148339, + "learning_rate": 3.227729022905697e-06, + "loss": 0.5791, + "step": 6859 + }, + { + "epoch": 3.7242128121606948, + "grad_norm": 9.205512068687561, + "learning_rate": 3.225141842565577e-06, + "loss": 0.3107, + "step": 6860 + }, + { + "epoch": 3.724755700325733, + "grad_norm": 11.81216687228549, + "learning_rate": 3.2225555001507124e-06, + "loss": 0.555, + "step": 6861 + }, + { + "epoch": 3.725298588490771, + "grad_norm": 9.593384384681405, + "learning_rate": 3.219969995980985e-06, + "loss": 0.4445, + "step": 6862 + }, + { + "epoch": 3.725841476655809, + "grad_norm": 6.990210033223102, + "learning_rate": 3.217385330376174e-06, + "loss": 0.1827, + "step": 6863 + }, + { + "epoch": 3.726384364820847, + "grad_norm": 16.17725365334651, + "learning_rate": 3.2148015036559522e-06, + "loss": 0.6442, + "step": 6864 + }, + { + "epoch": 3.726927252985885, + "grad_norm": 9.936484066410749, + "learning_rate": 3.2122185161399012e-06, + "loss": 0.3459, + "step": 6865 + }, + { + "epoch": 3.727470141150923, + "grad_norm": 16.59297040971604, + "learning_rate": 3.209636368147475e-06, + "loss": 0.5169, + "step": 6866 + }, + { + "epoch": 3.728013029315961, + "grad_norm": 14.156312439459766, + "learning_rate": 3.207055059998052e-06, + "loss": 0.6236, + "step": 6867 + }, + { + "epoch": 3.728555917480999, + "grad_norm": 10.18075916418277, + "learning_rate": 3.2044745920108832e-06, + "loss": 0.3108, + "step": 6868 + }, + { + "epoch": 3.729098805646037, + "grad_norm": 11.677366741727328, + "learning_rate": 3.2018949645051257e-06, + "loss": 0.5047, + "step": 6869 + }, + { + "epoch": 3.729641693811075, + "grad_norm": 12.56566346772868, + "learning_rate": 3.199316177799835e-06, + "loss": 0.4698, + "step": 6870 + }, + { + "epoch": 3.730184581976113, + "grad_norm": 14.518568356077754, + "learning_rate": 3.1967382322139538e-06, + "loss": 0.6106, + "step": 6871 + }, + { + "epoch": 3.730727470141151, + "grad_norm": 12.595862204241472, + "learning_rate": 3.194161128066338e-06, + "loss": 0.5617, + "step": 6872 + }, + { + "epoch": 3.731270358306189, + "grad_norm": 14.836027445862813, + "learning_rate": 3.191584865675713e-06, + "loss": 0.5989, + "step": 6873 + }, + { + "epoch": 3.731813246471227, + "grad_norm": 15.308949249609173, + "learning_rate": 3.189009445360731e-06, + "loss": 0.5385, + "step": 6874 + }, + { + "epoch": 3.732356134636265, + "grad_norm": 12.000029094911469, + "learning_rate": 3.1864348674399083e-06, + "loss": 0.4519, + "step": 6875 + }, + { + "epoch": 3.732899022801303, + "grad_norm": 15.618812442704373, + "learning_rate": 3.1838611322316836e-06, + "loss": 0.7854, + "step": 6876 + }, + { + "epoch": 3.733441910966341, + "grad_norm": 11.71029558715352, + "learning_rate": 3.181288240054381e-06, + "loss": 0.4303, + "step": 6877 + }, + { + "epoch": 3.733984799131379, + "grad_norm": 10.712473236039532, + "learning_rate": 3.17871619122621e-06, + "loss": 0.4896, + "step": 6878 + }, + { + "epoch": 3.734527687296417, + "grad_norm": 10.854545722563955, + "learning_rate": 3.176144986065298e-06, + "loss": 0.4795, + "step": 6879 + }, + { + "epoch": 3.735070575461455, + "grad_norm": 10.845848101056692, + "learning_rate": 3.1735746248896436e-06, + "loss": 0.3512, + "step": 6880 + }, + { + "epoch": 3.735613463626493, + "grad_norm": 11.675294486688255, + "learning_rate": 3.1710051080171657e-06, + "loss": 0.3521, + "step": 6881 + }, + { + "epoch": 3.736156351791531, + "grad_norm": 13.325204505254437, + "learning_rate": 3.168436435765654e-06, + "loss": 0.9413, + "step": 6882 + }, + { + "epoch": 3.736699239956569, + "grad_norm": 11.623023017043877, + "learning_rate": 3.1658686084528146e-06, + "loss": 0.491, + "step": 6883 + }, + { + "epoch": 3.737242128121607, + "grad_norm": 7.846579009143442, + "learning_rate": 3.163301626396237e-06, + "loss": 0.2837, + "step": 6884 + }, + { + "epoch": 3.737785016286645, + "grad_norm": 13.308504927624488, + "learning_rate": 3.1607354899134102e-06, + "loss": 0.5548, + "step": 6885 + }, + { + "epoch": 3.738327904451683, + "grad_norm": 8.7685040860308, + "learning_rate": 3.1581701993217197e-06, + "loss": 0.2844, + "step": 6886 + }, + { + "epoch": 3.738870792616721, + "grad_norm": 12.979195325374219, + "learning_rate": 3.1556057549384424e-06, + "loss": 0.6242, + "step": 6887 + }, + { + "epoch": 3.739413680781759, + "grad_norm": 15.955341391994917, + "learning_rate": 3.1530421570807536e-06, + "loss": 0.5053, + "step": 6888 + }, + { + "epoch": 3.739956568946797, + "grad_norm": 15.39892028189493, + "learning_rate": 3.1504794060657228e-06, + "loss": 0.8675, + "step": 6889 + }, + { + "epoch": 3.740499457111835, + "grad_norm": 15.189388217562758, + "learning_rate": 3.147917502210316e-06, + "loss": 0.6521, + "step": 6890 + }, + { + "epoch": 3.741042345276873, + "grad_norm": 15.85941037907273, + "learning_rate": 3.1453564458313923e-06, + "loss": 0.6711, + "step": 6891 + }, + { + "epoch": 3.741585233441911, + "grad_norm": 10.599235827332738, + "learning_rate": 3.142796237245709e-06, + "loss": 0.3489, + "step": 6892 + }, + { + "epoch": 3.742128121606949, + "grad_norm": 11.991913744676399, + "learning_rate": 3.1402368767699152e-06, + "loss": 0.4865, + "step": 6893 + }, + { + "epoch": 3.742671009771987, + "grad_norm": 15.429458030904042, + "learning_rate": 3.137678364720559e-06, + "loss": 0.8425, + "step": 6894 + }, + { + "epoch": 3.743213897937025, + "grad_norm": 13.89952284029879, + "learning_rate": 3.135120701414077e-06, + "loss": 0.5919, + "step": 6895 + }, + { + "epoch": 3.743756786102063, + "grad_norm": 14.355639876320694, + "learning_rate": 3.1325638871668094e-06, + "loss": 0.517, + "step": 6896 + }, + { + "epoch": 3.744299674267101, + "grad_norm": 9.998711427681842, + "learning_rate": 3.1300079222949854e-06, + "loss": 0.4837, + "step": 6897 + }, + { + "epoch": 3.744842562432139, + "grad_norm": 12.948353433621449, + "learning_rate": 3.1274528071147303e-06, + "loss": 0.4702, + "step": 6898 + }, + { + "epoch": 3.745385450597177, + "grad_norm": 11.300075344262957, + "learning_rate": 3.124898541942065e-06, + "loss": 0.4163, + "step": 6899 + }, + { + "epoch": 3.745928338762215, + "grad_norm": 10.014842311292028, + "learning_rate": 3.122345127092905e-06, + "loss": 0.4597, + "step": 6900 + }, + { + "epoch": 3.746471226927253, + "grad_norm": 14.937187078292135, + "learning_rate": 3.1197925628830616e-06, + "loss": 1.0359, + "step": 6901 + }, + { + "epoch": 3.747014115092291, + "grad_norm": 11.821812911772865, + "learning_rate": 3.117240849628239e-06, + "loss": 0.369, + "step": 6902 + }, + { + "epoch": 3.747557003257329, + "grad_norm": 18.601455125565817, + "learning_rate": 3.1146899876440383e-06, + "loss": 0.3967, + "step": 6903 + }, + { + "epoch": 3.748099891422367, + "grad_norm": 26.38167902221625, + "learning_rate": 3.1121399772459495e-06, + "loss": 0.7955, + "step": 6904 + }, + { + "epoch": 3.748642779587405, + "grad_norm": 9.036657138578375, + "learning_rate": 3.109590818749373e-06, + "loss": 0.3234, + "step": 6905 + }, + { + "epoch": 3.749185667752443, + "grad_norm": 12.029611743069063, + "learning_rate": 3.1070425124695825e-06, + "loss": 0.4689, + "step": 6906 + }, + { + "epoch": 3.749728555917481, + "grad_norm": 11.16075799831913, + "learning_rate": 3.1044950587217595e-06, + "loss": 0.455, + "step": 6907 + }, + { + "epoch": 3.750271444082519, + "grad_norm": 12.64123781208839, + "learning_rate": 3.1019484578209768e-06, + "loss": 0.4525, + "step": 6908 + }, + { + "epoch": 3.750814332247557, + "grad_norm": 10.873865130895668, + "learning_rate": 3.0994027100822e-06, + "loss": 0.7957, + "step": 6909 + }, + { + "epoch": 3.751357220412595, + "grad_norm": 15.826997708730714, + "learning_rate": 3.0968578158203e-06, + "loss": 0.4531, + "step": 6910 + }, + { + "epoch": 3.751900108577633, + "grad_norm": 16.864174578861103, + "learning_rate": 3.094313775350022e-06, + "loss": 0.6061, + "step": 6911 + }, + { + "epoch": 3.752442996742671, + "grad_norm": 16.360017183257842, + "learning_rate": 3.0917705889860283e-06, + "loss": 0.849, + "step": 6912 + }, + { + "epoch": 3.752985884907709, + "grad_norm": 11.473341112084933, + "learning_rate": 3.0892282570428513e-06, + "loss": 0.4494, + "step": 6913 + }, + { + "epoch": 3.753528773072747, + "grad_norm": 12.456658860480365, + "learning_rate": 3.0866867798349407e-06, + "loss": 0.4774, + "step": 6914 + }, + { + "epoch": 3.754071661237785, + "grad_norm": 16.476960663849457, + "learning_rate": 3.0841461576766284e-06, + "loss": 0.8323, + "step": 6915 + }, + { + "epoch": 3.754614549402823, + "grad_norm": 13.888913904073622, + "learning_rate": 3.0816063908821403e-06, + "loss": 0.5861, + "step": 6916 + }, + { + "epoch": 3.755157437567861, + "grad_norm": 10.764080500005052, + "learning_rate": 3.079067479765604e-06, + "loss": 0.3865, + "step": 6917 + }, + { + "epoch": 3.755700325732899, + "grad_norm": 13.47504295740381, + "learning_rate": 3.076529424641026e-06, + "loss": 0.628, + "step": 6918 + }, + { + "epoch": 3.756243213897937, + "grad_norm": 14.742498671541298, + "learning_rate": 3.0739922258223286e-06, + "loss": 0.6586, + "step": 6919 + }, + { + "epoch": 3.756786102062975, + "grad_norm": 11.002483024658138, + "learning_rate": 3.0714558836233056e-06, + "loss": 0.539, + "step": 6920 + }, + { + "epoch": 3.757328990228013, + "grad_norm": 13.952905049231472, + "learning_rate": 3.068920398357663e-06, + "loss": 0.8338, + "step": 6921 + }, + { + "epoch": 3.757871878393051, + "grad_norm": 10.43091762446858, + "learning_rate": 3.0663857703389943e-06, + "loss": 0.4779, + "step": 6922 + }, + { + "epoch": 3.758414766558089, + "grad_norm": 16.415629535468835, + "learning_rate": 3.0638519998807826e-06, + "loss": 0.4277, + "step": 6923 + }, + { + "epoch": 3.758957654723127, + "grad_norm": 8.867208839527924, + "learning_rate": 3.0613190872964104e-06, + "loss": 0.3777, + "step": 6924 + }, + { + "epoch": 3.759500542888165, + "grad_norm": 11.642294707216093, + "learning_rate": 3.058787032899152e-06, + "loss": 0.5056, + "step": 6925 + }, + { + "epoch": 3.760043431053203, + "grad_norm": 11.206786718860933, + "learning_rate": 3.0562558370021765e-06, + "loss": 0.446, + "step": 6926 + }, + { + "epoch": 3.760586319218241, + "grad_norm": 10.068491984488928, + "learning_rate": 3.053725499918545e-06, + "loss": 0.5076, + "step": 6927 + }, + { + "epoch": 3.761129207383279, + "grad_norm": 16.07206135268971, + "learning_rate": 3.0511960219612147e-06, + "loss": 0.5143, + "step": 6928 + }, + { + "epoch": 3.761672095548317, + "grad_norm": 11.602660115945492, + "learning_rate": 3.0486674034430352e-06, + "loss": 0.508, + "step": 6929 + }, + { + "epoch": 3.762214983713355, + "grad_norm": 13.3862945751605, + "learning_rate": 3.046139644676751e-06, + "loss": 0.3828, + "step": 6930 + }, + { + "epoch": 3.762757871878393, + "grad_norm": 10.733311225895106, + "learning_rate": 3.0436127459749975e-06, + "loss": 0.37, + "step": 6931 + }, + { + "epoch": 3.763300760043431, + "grad_norm": 11.261346397862807, + "learning_rate": 3.041086707650306e-06, + "loss": 0.4808, + "step": 6932 + }, + { + "epoch": 3.763843648208469, + "grad_norm": 16.501582349438152, + "learning_rate": 3.0385615300151018e-06, + "loss": 0.6239, + "step": 6933 + }, + { + "epoch": 3.764386536373507, + "grad_norm": 11.553010310131304, + "learning_rate": 3.036037213381702e-06, + "loss": 0.4541, + "step": 6934 + }, + { + "epoch": 3.764929424538545, + "grad_norm": 11.723556454514561, + "learning_rate": 3.0335137580623154e-06, + "loss": 0.4217, + "step": 6935 + }, + { + "epoch": 3.765472312703583, + "grad_norm": 15.843208183133134, + "learning_rate": 3.030991164369057e-06, + "loss": 0.628, + "step": 6936 + }, + { + "epoch": 3.766015200868621, + "grad_norm": 8.046500189967162, + "learning_rate": 3.028469432613915e-06, + "loss": 0.288, + "step": 6937 + }, + { + "epoch": 3.766558089033659, + "grad_norm": 10.115796532018264, + "learning_rate": 3.0259485631087846e-06, + "loss": 0.3899, + "step": 6938 + }, + { + "epoch": 3.767100977198697, + "grad_norm": 13.289116360712832, + "learning_rate": 3.023428556165451e-06, + "loss": 0.6068, + "step": 6939 + }, + { + "epoch": 3.767643865363735, + "grad_norm": 11.871048514073903, + "learning_rate": 3.020909412095592e-06, + "loss": 0.6089, + "step": 6940 + }, + { + "epoch": 3.768186753528773, + "grad_norm": 8.084122023965138, + "learning_rate": 3.0183911312107806e-06, + "loss": 0.3247, + "step": 6941 + }, + { + "epoch": 3.768729641693811, + "grad_norm": 10.871234822080087, + "learning_rate": 3.015873713822478e-06, + "loss": 0.4928, + "step": 6942 + }, + { + "epoch": 3.769272529858849, + "grad_norm": 13.956354691016482, + "learning_rate": 3.013357160242051e-06, + "loss": 0.7454, + "step": 6943 + }, + { + "epoch": 3.769815418023887, + "grad_norm": 17.342203574294, + "learning_rate": 3.01084147078074e-06, + "loss": 0.6629, + "step": 6944 + }, + { + "epoch": 3.770358306188925, + "grad_norm": 9.435776085708705, + "learning_rate": 3.008326645749701e-06, + "loss": 0.2969, + "step": 6945 + }, + { + "epoch": 3.770901194353963, + "grad_norm": 14.35523924414514, + "learning_rate": 3.0058126854599624e-06, + "loss": 0.6859, + "step": 6946 + }, + { + "epoch": 3.771444082519001, + "grad_norm": 10.162640497221313, + "learning_rate": 3.003299590222454e-06, + "loss": 0.4144, + "step": 6947 + }, + { + "epoch": 3.771986970684039, + "grad_norm": 10.611516903885335, + "learning_rate": 3.0007873603480107e-06, + "loss": 0.5975, + "step": 6948 + }, + { + "epoch": 3.772529858849077, + "grad_norm": 13.809727664609994, + "learning_rate": 2.998275996147335e-06, + "loss": 0.4714, + "step": 6949 + }, + { + "epoch": 3.773072747014115, + "grad_norm": 11.317211520834723, + "learning_rate": 2.995765497931049e-06, + "loss": 0.4646, + "step": 6950 + }, + { + "epoch": 3.773615635179153, + "grad_norm": 10.86923796914853, + "learning_rate": 2.9932558660096443e-06, + "loss": 0.3228, + "step": 6951 + }, + { + "epoch": 3.774158523344191, + "grad_norm": 7.470997903259584, + "learning_rate": 2.990747100693526e-06, + "loss": 0.2531, + "step": 6952 + }, + { + "epoch": 3.774701411509229, + "grad_norm": 8.627824874780737, + "learning_rate": 2.988239202292972e-06, + "loss": 0.361, + "step": 6953 + }, + { + "epoch": 3.775244299674267, + "grad_norm": 9.14607939221467, + "learning_rate": 2.985732171118172e-06, + "loss": 0.3116, + "step": 6954 + }, + { + "epoch": 3.7757871878393052, + "grad_norm": 10.155112379070209, + "learning_rate": 2.9832260074791984e-06, + "loss": 0.4723, + "step": 6955 + }, + { + "epoch": 3.776330076004343, + "grad_norm": 10.506852840497306, + "learning_rate": 2.9807207116860094e-06, + "loss": 0.4232, + "step": 6956 + }, + { + "epoch": 3.7768729641693812, + "grad_norm": 11.193678853889507, + "learning_rate": 2.9782162840484765e-06, + "loss": 0.4915, + "step": 6957 + }, + { + "epoch": 3.777415852334419, + "grad_norm": 14.323380391376173, + "learning_rate": 2.9757127248763375e-06, + "loss": 0.6207, + "step": 6958 + }, + { + "epoch": 3.7779587404994572, + "grad_norm": 11.773695664037353, + "learning_rate": 2.973210034479247e-06, + "loss": 0.3957, + "step": 6959 + }, + { + "epoch": 3.778501628664495, + "grad_norm": 8.978033406900165, + "learning_rate": 2.97070821316674e-06, + "loss": 0.2935, + "step": 6960 + }, + { + "epoch": 3.7790445168295332, + "grad_norm": 12.4187581608995, + "learning_rate": 2.9682072612482427e-06, + "loss": 0.515, + "step": 6961 + }, + { + "epoch": 3.779587404994571, + "grad_norm": 12.551466191074608, + "learning_rate": 2.965707179033078e-06, + "loss": 0.4014, + "step": 6962 + }, + { + "epoch": 3.7801302931596092, + "grad_norm": 8.193678184944888, + "learning_rate": 2.9632079668304613e-06, + "loss": 0.2947, + "step": 6963 + }, + { + "epoch": 3.780673181324647, + "grad_norm": 12.277305597806611, + "learning_rate": 2.960709624949498e-06, + "loss": 0.4283, + "step": 6964 + }, + { + "epoch": 3.7812160694896852, + "grad_norm": 12.873825056251757, + "learning_rate": 2.958212153699187e-06, + "loss": 0.3719, + "step": 6965 + }, + { + "epoch": 3.781758957654723, + "grad_norm": 11.733322711394022, + "learning_rate": 2.955715553388421e-06, + "loss": 0.4547, + "step": 6966 + }, + { + "epoch": 3.7823018458197613, + "grad_norm": 13.421562379885513, + "learning_rate": 2.95321982432598e-06, + "loss": 0.8095, + "step": 6967 + }, + { + "epoch": 3.782844733984799, + "grad_norm": 10.751529697853956, + "learning_rate": 2.950724966820544e-06, + "loss": 0.3795, + "step": 6968 + }, + { + "epoch": 3.7833876221498373, + "grad_norm": 11.23769589292945, + "learning_rate": 2.9482309811806785e-06, + "loss": 0.4324, + "step": 6969 + }, + { + "epoch": 3.783930510314875, + "grad_norm": 10.760821500115565, + "learning_rate": 2.945737867714844e-06, + "loss": 0.4112, + "step": 6970 + }, + { + "epoch": 3.7844733984799133, + "grad_norm": 10.577088653058517, + "learning_rate": 2.9432456267313924e-06, + "loss": 0.3819, + "step": 6971 + }, + { + "epoch": 3.785016286644951, + "grad_norm": 9.877099189181129, + "learning_rate": 2.9407542585385683e-06, + "loss": 0.2344, + "step": 6972 + }, + { + "epoch": 3.7855591748099893, + "grad_norm": 18.577198876664223, + "learning_rate": 2.9382637634445087e-06, + "loss": 0.3254, + "step": 6973 + }, + { + "epoch": 3.786102062975027, + "grad_norm": 14.663768331220036, + "learning_rate": 2.9357741417572415e-06, + "loss": 0.5764, + "step": 6974 + }, + { + "epoch": 3.7866449511400653, + "grad_norm": 12.695557652262636, + "learning_rate": 2.9332853937846873e-06, + "loss": 0.4016, + "step": 6975 + }, + { + "epoch": 3.787187839305103, + "grad_norm": 13.456679312848303, + "learning_rate": 2.930797519834658e-06, + "loss": 0.438, + "step": 6976 + }, + { + "epoch": 3.7877307274701413, + "grad_norm": 10.441195425871536, + "learning_rate": 2.928310520214859e-06, + "loss": 0.4062, + "step": 6977 + }, + { + "epoch": 3.788273615635179, + "grad_norm": 10.300692057422426, + "learning_rate": 2.9258243952328846e-06, + "loss": 0.368, + "step": 6978 + }, + { + "epoch": 3.7888165038002173, + "grad_norm": 11.532450200794159, + "learning_rate": 2.9233391451962233e-06, + "loss": 0.5835, + "step": 6979 + }, + { + "epoch": 3.789359391965255, + "grad_norm": 13.465329246550354, + "learning_rate": 2.9208547704122525e-06, + "loss": 0.5226, + "step": 6980 + }, + { + "epoch": 3.7899022801302933, + "grad_norm": 14.107590055474905, + "learning_rate": 2.9183712711882526e-06, + "loss": 0.6195, + "step": 6981 + }, + { + "epoch": 3.790445168295331, + "grad_norm": 10.577784567014287, + "learning_rate": 2.915888647831374e-06, + "loss": 0.3655, + "step": 6982 + }, + { + "epoch": 3.7909880564603693, + "grad_norm": 13.136687160957166, + "learning_rate": 2.913406900648683e-06, + "loss": 0.6708, + "step": 6983 + }, + { + "epoch": 3.791530944625407, + "grad_norm": 15.57844160050923, + "learning_rate": 2.9109260299471165e-06, + "loss": 0.7229, + "step": 6984 + }, + { + "epoch": 3.7920738327904453, + "grad_norm": 14.528051815895886, + "learning_rate": 2.908446036033519e-06, + "loss": 0.4592, + "step": 6985 + }, + { + "epoch": 3.792616720955483, + "grad_norm": 20.212297672580796, + "learning_rate": 2.90596691921462e-06, + "loss": 0.759, + "step": 6986 + }, + { + "epoch": 3.7931596091205213, + "grad_norm": 9.84665764740946, + "learning_rate": 2.9034886797970342e-06, + "loss": 0.3681, + "step": 6987 + }, + { + "epoch": 3.793702497285559, + "grad_norm": 8.475948428830446, + "learning_rate": 2.9010113180872847e-06, + "loss": 0.2304, + "step": 6988 + }, + { + "epoch": 3.7942453854505973, + "grad_norm": 11.955148009858453, + "learning_rate": 2.898534834391762e-06, + "loss": 0.5278, + "step": 6989 + }, + { + "epoch": 3.794788273615635, + "grad_norm": 18.8535666021868, + "learning_rate": 2.8960592290167754e-06, + "loss": 1.1773, + "step": 6990 + }, + { + "epoch": 3.7953311617806733, + "grad_norm": 14.603134582974139, + "learning_rate": 2.8935845022685006e-06, + "loss": 0.7039, + "step": 6991 + }, + { + "epoch": 3.795874049945711, + "grad_norm": 17.09419124573872, + "learning_rate": 2.891110654453022e-06, + "loss": 0.6918, + "step": 6992 + }, + { + "epoch": 3.7964169381107493, + "grad_norm": 14.480524368793315, + "learning_rate": 2.888637685876309e-06, + "loss": 0.5987, + "step": 6993 + }, + { + "epoch": 3.796959826275787, + "grad_norm": 13.958888615564357, + "learning_rate": 2.8861655968442192e-06, + "loss": 0.7819, + "step": 6994 + }, + { + "epoch": 3.7975027144408253, + "grad_norm": 13.785514235662655, + "learning_rate": 2.883694387662511e-06, + "loss": 0.6469, + "step": 6995 + }, + { + "epoch": 3.798045602605863, + "grad_norm": 15.113325361621122, + "learning_rate": 2.8812240586368157e-06, + "loss": 0.6732, + "step": 6996 + }, + { + "epoch": 3.7985884907709013, + "grad_norm": 13.447315748094697, + "learning_rate": 2.8787546100726773e-06, + "loss": 0.8885, + "step": 6997 + }, + { + "epoch": 3.799131378935939, + "grad_norm": 15.942374082087648, + "learning_rate": 2.876286042275518e-06, + "loss": 0.5496, + "step": 6998 + }, + { + "epoch": 3.7996742671009773, + "grad_norm": 13.841633437355465, + "learning_rate": 2.873818355550655e-06, + "loss": 0.2924, + "step": 6999 + }, + { + "epoch": 3.800217155266015, + "grad_norm": 12.814203732560337, + "learning_rate": 2.871351550203295e-06, + "loss": 0.5538, + "step": 7000 + }, + { + "epoch": 3.8007600434310533, + "grad_norm": 12.191028501042764, + "learning_rate": 2.8688856265385367e-06, + "loss": 0.4576, + "step": 7001 + }, + { + "epoch": 3.801302931596091, + "grad_norm": 10.517009605166134, + "learning_rate": 2.86642058486137e-06, + "loss": 0.4275, + "step": 7002 + }, + { + "epoch": 3.8018458197611293, + "grad_norm": 13.976454434703868, + "learning_rate": 2.863956425476674e-06, + "loss": 0.5466, + "step": 7003 + }, + { + "epoch": 3.802388707926167, + "grad_norm": 11.5016640245614, + "learning_rate": 2.8614931486892207e-06, + "loss": 0.366, + "step": 7004 + }, + { + "epoch": 3.8029315960912053, + "grad_norm": 11.939582442434286, + "learning_rate": 2.8590307548036712e-06, + "loss": 0.4141, + "step": 7005 + }, + { + "epoch": 3.803474484256243, + "grad_norm": 12.987510367990025, + "learning_rate": 2.8565692441245796e-06, + "loss": 0.5152, + "step": 7006 + }, + { + "epoch": 3.8040173724212814, + "grad_norm": 12.832896642663844, + "learning_rate": 2.8541086169563894e-06, + "loss": 0.4585, + "step": 7007 + }, + { + "epoch": 3.804560260586319, + "grad_norm": 11.863347803098113, + "learning_rate": 2.851648873603433e-06, + "loss": 0.4196, + "step": 7008 + }, + { + "epoch": 3.8051031487513574, + "grad_norm": 13.239707667599935, + "learning_rate": 2.8491900143699367e-06, + "loss": 0.5748, + "step": 7009 + }, + { + "epoch": 3.805646036916395, + "grad_norm": 12.665696921336465, + "learning_rate": 2.846732039560017e-06, + "loss": 0.4688, + "step": 7010 + }, + { + "epoch": 3.8061889250814334, + "grad_norm": 10.268698833296764, + "learning_rate": 2.844274949477679e-06, + "loss": 0.3858, + "step": 7011 + }, + { + "epoch": 3.806731813246471, + "grad_norm": 12.660908637644715, + "learning_rate": 2.8418187444268198e-06, + "loss": 0.445, + "step": 7012 + }, + { + "epoch": 3.8072747014115094, + "grad_norm": 10.571667634245497, + "learning_rate": 2.8393634247112232e-06, + "loss": 0.4066, + "step": 7013 + }, + { + "epoch": 3.807817589576547, + "grad_norm": 19.94546394586301, + "learning_rate": 2.836908990634578e-06, + "loss": 0.5101, + "step": 7014 + }, + { + "epoch": 3.8083604777415854, + "grad_norm": 13.85083509805195, + "learning_rate": 2.834455442500441e-06, + "loss": 0.4073, + "step": 7015 + }, + { + "epoch": 3.808903365906623, + "grad_norm": 10.847008608128625, + "learning_rate": 2.8320027806122753e-06, + "loss": 0.3582, + "step": 7016 + }, + { + "epoch": 3.8094462540716614, + "grad_norm": 13.798257428735926, + "learning_rate": 2.82955100527343e-06, + "loss": 0.7057, + "step": 7017 + }, + { + "epoch": 3.809989142236699, + "grad_norm": 15.062376260700791, + "learning_rate": 2.8271001167871405e-06, + "loss": 0.6019, + "step": 7018 + }, + { + "epoch": 3.8105320304017374, + "grad_norm": 11.109879992701345, + "learning_rate": 2.8246501154565476e-06, + "loss": 0.3265, + "step": 7019 + }, + { + "epoch": 3.811074918566775, + "grad_norm": 13.70633009915418, + "learning_rate": 2.822201001584657e-06, + "loss": 0.4019, + "step": 7020 + }, + { + "epoch": 3.8116178067318134, + "grad_norm": 12.339132223947464, + "learning_rate": 2.8197527754743926e-06, + "loss": 0.3272, + "step": 7021 + }, + { + "epoch": 3.812160694896851, + "grad_norm": 9.817254835111205, + "learning_rate": 2.8173054374285434e-06, + "loss": 0.4118, + "step": 7022 + }, + { + "epoch": 3.8127035830618894, + "grad_norm": 15.18363193003276, + "learning_rate": 2.81485898774981e-06, + "loss": 0.7261, + "step": 7023 + }, + { + "epoch": 3.813246471226927, + "grad_norm": 11.984870936412664, + "learning_rate": 2.8124134267407665e-06, + "loss": 0.3734, + "step": 7024 + }, + { + "epoch": 3.8137893593919654, + "grad_norm": 10.678523359563961, + "learning_rate": 2.8099687547038824e-06, + "loss": 0.5318, + "step": 7025 + }, + { + "epoch": 3.814332247557003, + "grad_norm": 18.489591388355425, + "learning_rate": 2.807524971941528e-06, + "loss": 0.8301, + "step": 7026 + }, + { + "epoch": 3.8148751357220414, + "grad_norm": 10.459115256144122, + "learning_rate": 2.805082078755942e-06, + "loss": 0.3638, + "step": 7027 + }, + { + "epoch": 3.815418023887079, + "grad_norm": 12.778148708574054, + "learning_rate": 2.8026400754492767e-06, + "loss": 0.6202, + "step": 7028 + }, + { + "epoch": 3.8159609120521174, + "grad_norm": 14.76075800024929, + "learning_rate": 2.8001989623235528e-06, + "loss": 0.8516, + "step": 7029 + }, + { + "epoch": 3.816503800217155, + "grad_norm": 15.243047141569159, + "learning_rate": 2.797758739680698e-06, + "loss": 0.7363, + "step": 7030 + }, + { + "epoch": 3.8170466883821934, + "grad_norm": 12.332962959117395, + "learning_rate": 2.7953194078225223e-06, + "loss": 0.5528, + "step": 7031 + }, + { + "epoch": 3.817589576547231, + "grad_norm": 14.15721396911131, + "learning_rate": 2.792880967050724e-06, + "loss": 0.6093, + "step": 7032 + }, + { + "epoch": 3.8181324647122694, + "grad_norm": 16.149812128817175, + "learning_rate": 2.790443417666894e-06, + "loss": 0.7232, + "step": 7033 + }, + { + "epoch": 3.818675352877307, + "grad_norm": 6.884614594609857, + "learning_rate": 2.7880067599725123e-06, + "loss": 0.2341, + "step": 7034 + }, + { + "epoch": 3.8192182410423454, + "grad_norm": 13.604715542482158, + "learning_rate": 2.7855709942689525e-06, + "loss": 0.5522, + "step": 7035 + }, + { + "epoch": 3.819761129207383, + "grad_norm": 12.822857567671976, + "learning_rate": 2.7831361208574636e-06, + "loss": 0.459, + "step": 7036 + }, + { + "epoch": 3.8203040173724214, + "grad_norm": 16.80060104026762, + "learning_rate": 2.780702140039204e-06, + "loss": 0.6408, + "step": 7037 + }, + { + "epoch": 3.820846905537459, + "grad_norm": 9.25304440984111, + "learning_rate": 2.77826905211521e-06, + "loss": 0.3628, + "step": 7038 + }, + { + "epoch": 3.8213897937024974, + "grad_norm": 9.5263200358754, + "learning_rate": 2.775836857386409e-06, + "loss": 0.4398, + "step": 7039 + }, + { + "epoch": 3.821932681867535, + "grad_norm": 9.394535823999632, + "learning_rate": 2.77340555615362e-06, + "loss": 0.4318, + "step": 7040 + }, + { + "epoch": 3.8224755700325734, + "grad_norm": 16.75570479373163, + "learning_rate": 2.7709751487175486e-06, + "loss": 0.621, + "step": 7041 + }, + { + "epoch": 3.823018458197611, + "grad_norm": 11.68838601944592, + "learning_rate": 2.768545635378792e-06, + "loss": 0.4332, + "step": 7042 + }, + { + "epoch": 3.8235613463626494, + "grad_norm": 13.068740742878463, + "learning_rate": 2.7661170164378372e-06, + "loss": 0.6055, + "step": 7043 + }, + { + "epoch": 3.824104234527687, + "grad_norm": 12.808750119841285, + "learning_rate": 2.7636892921950586e-06, + "loss": 0.3966, + "step": 7044 + }, + { + "epoch": 3.8246471226927254, + "grad_norm": 13.368527360017245, + "learning_rate": 2.7612624629507213e-06, + "loss": 0.4222, + "step": 7045 + }, + { + "epoch": 3.8251900108577632, + "grad_norm": 14.624704158928514, + "learning_rate": 2.758836529004979e-06, + "loss": 0.5196, + "step": 7046 + }, + { + "epoch": 3.8257328990228014, + "grad_norm": 11.425494087567786, + "learning_rate": 2.756411490657875e-06, + "loss": 0.5476, + "step": 7047 + }, + { + "epoch": 3.8262757871878392, + "grad_norm": 11.438998129235056, + "learning_rate": 2.753987348209344e-06, + "loss": 0.3902, + "step": 7048 + }, + { + "epoch": 3.8268186753528775, + "grad_norm": 13.662522916551874, + "learning_rate": 2.7515641019592053e-06, + "loss": 0.494, + "step": 7049 + }, + { + "epoch": 3.8273615635179152, + "grad_norm": 10.296671849651366, + "learning_rate": 2.7491417522071706e-06, + "loss": 0.3226, + "step": 7050 + }, + { + "epoch": 3.8279044516829535, + "grad_norm": 10.454941310693076, + "learning_rate": 2.7467202992528376e-06, + "loss": 0.2803, + "step": 7051 + }, + { + "epoch": 3.8284473398479912, + "grad_norm": 11.648384847247753, + "learning_rate": 2.744299743395703e-06, + "loss": 0.5601, + "step": 7052 + }, + { + "epoch": 3.8289902280130295, + "grad_norm": 8.944598429537628, + "learning_rate": 2.741880084935138e-06, + "loss": 0.3296, + "step": 7053 + }, + { + "epoch": 3.8295331161780672, + "grad_norm": 13.218169787059113, + "learning_rate": 2.7394613241704117e-06, + "loss": 0.5443, + "step": 7054 + }, + { + "epoch": 3.8300760043431055, + "grad_norm": 9.804324070437223, + "learning_rate": 2.7370434614006803e-06, + "loss": 0.4269, + "step": 7055 + }, + { + "epoch": 3.8306188925081432, + "grad_norm": 16.684309046147042, + "learning_rate": 2.734626496924986e-06, + "loss": 0.7608, + "step": 7056 + }, + { + "epoch": 3.8311617806731815, + "grad_norm": 11.594370437365152, + "learning_rate": 2.7322104310422713e-06, + "loss": 0.4286, + "step": 7057 + }, + { + "epoch": 3.8317046688382193, + "grad_norm": 9.387708800824917, + "learning_rate": 2.7297952640513483e-06, + "loss": 0.4659, + "step": 7058 + }, + { + "epoch": 3.8322475570032575, + "grad_norm": 15.36709432735338, + "learning_rate": 2.727380996250939e-06, + "loss": 0.4572, + "step": 7059 + }, + { + "epoch": 3.8327904451682953, + "grad_norm": 10.866529562034719, + "learning_rate": 2.724967627939634e-06, + "loss": 0.5208, + "step": 7060 + }, + { + "epoch": 3.8333333333333335, + "grad_norm": 11.215185901261178, + "learning_rate": 2.722555159415934e-06, + "loss": 0.4077, + "step": 7061 + }, + { + "epoch": 3.8338762214983713, + "grad_norm": 13.560818928747905, + "learning_rate": 2.7201435909782027e-06, + "loss": 0.4849, + "step": 7062 + }, + { + "epoch": 3.8344191096634095, + "grad_norm": 12.76281749363346, + "learning_rate": 2.7177329229247186e-06, + "loss": 0.6375, + "step": 7063 + }, + { + "epoch": 3.8349619978284473, + "grad_norm": 11.109626055006531, + "learning_rate": 2.715323155553635e-06, + "loss": 0.6538, + "step": 7064 + }, + { + "epoch": 3.8355048859934855, + "grad_norm": 16.73906444542459, + "learning_rate": 2.712914289162989e-06, + "loss": 0.4548, + "step": 7065 + }, + { + "epoch": 3.8360477741585233, + "grad_norm": 15.832313422557572, + "learning_rate": 2.7105063240507222e-06, + "loss": 0.5417, + "step": 7066 + }, + { + "epoch": 3.8365906623235615, + "grad_norm": 10.615862418603832, + "learning_rate": 2.708099260514645e-06, + "loss": 0.6034, + "step": 7067 + }, + { + "epoch": 3.8371335504885993, + "grad_norm": 16.513578882781225, + "learning_rate": 2.7056930988524763e-06, + "loss": 0.9556, + "step": 7068 + }, + { + "epoch": 3.8376764386536375, + "grad_norm": 17.31989234293219, + "learning_rate": 2.703287839361811e-06, + "loss": 0.744, + "step": 7069 + }, + { + "epoch": 3.8382193268186753, + "grad_norm": 10.336380611592002, + "learning_rate": 2.7008834823401344e-06, + "loss": 0.5354, + "step": 7070 + }, + { + "epoch": 3.8387622149837135, + "grad_norm": 13.577288380592488, + "learning_rate": 2.698480028084821e-06, + "loss": 0.5647, + "step": 7071 + }, + { + "epoch": 3.8393051031487513, + "grad_norm": 15.062012119646445, + "learning_rate": 2.6960774768931365e-06, + "loss": 0.6897, + "step": 7072 + }, + { + "epoch": 3.8398479913137895, + "grad_norm": 9.679061664826015, + "learning_rate": 2.6936758290622324e-06, + "loss": 0.3364, + "step": 7073 + }, + { + "epoch": 3.8403908794788273, + "grad_norm": 12.06717080820273, + "learning_rate": 2.69127508488914e-06, + "loss": 0.473, + "step": 7074 + }, + { + "epoch": 3.8409337676438655, + "grad_norm": 12.515990246581914, + "learning_rate": 2.6888752446707965e-06, + "loss": 0.7188, + "step": 7075 + }, + { + "epoch": 3.8414766558089033, + "grad_norm": 12.151182717520161, + "learning_rate": 2.6864763087040145e-06, + "loss": 0.4494, + "step": 7076 + }, + { + "epoch": 3.8420195439739415, + "grad_norm": 13.03785855389188, + "learning_rate": 2.6840782772855002e-06, + "loss": 0.6012, + "step": 7077 + }, + { + "epoch": 3.8425624321389793, + "grad_norm": 13.049506054659528, + "learning_rate": 2.6816811507118436e-06, + "loss": 0.6763, + "step": 7078 + }, + { + "epoch": 3.8431053203040175, + "grad_norm": 10.797345758374234, + "learning_rate": 2.6792849292795253e-06, + "loss": 0.3712, + "step": 7079 + }, + { + "epoch": 3.8436482084690553, + "grad_norm": 7.303579416108625, + "learning_rate": 2.6768896132849144e-06, + "loss": 0.2622, + "step": 7080 + }, + { + "epoch": 3.8441910966340935, + "grad_norm": 12.992302429769211, + "learning_rate": 2.674495203024269e-06, + "loss": 0.5022, + "step": 7081 + }, + { + "epoch": 3.8447339847991313, + "grad_norm": 11.355323594435934, + "learning_rate": 2.67210169879373e-06, + "loss": 0.3898, + "step": 7082 + }, + { + "epoch": 3.8452768729641695, + "grad_norm": 10.043141357807462, + "learning_rate": 2.669709100889333e-06, + "loss": 0.4849, + "step": 7083 + }, + { + "epoch": 3.8458197611292073, + "grad_norm": 10.154968040172625, + "learning_rate": 2.6673174096069976e-06, + "loss": 0.6587, + "step": 7084 + }, + { + "epoch": 3.8463626492942455, + "grad_norm": 15.332496592868017, + "learning_rate": 2.6649266252425297e-06, + "loss": 0.5297, + "step": 7085 + }, + { + "epoch": 3.8469055374592833, + "grad_norm": 12.176208814956482, + "learning_rate": 2.6625367480916285e-06, + "loss": 0.603, + "step": 7086 + }, + { + "epoch": 3.8474484256243215, + "grad_norm": 12.08600431730549, + "learning_rate": 2.660147778449876e-06, + "loss": 0.6609, + "step": 7087 + }, + { + "epoch": 3.8479913137893593, + "grad_norm": 12.157193141505328, + "learning_rate": 2.6577597166127455e-06, + "loss": 0.4869, + "step": 7088 + }, + { + "epoch": 3.8485342019543975, + "grad_norm": 13.235050132598122, + "learning_rate": 2.655372562875591e-06, + "loss": 0.4279, + "step": 7089 + }, + { + "epoch": 3.8490770901194353, + "grad_norm": 12.171527155216868, + "learning_rate": 2.652986317533669e-06, + "loss": 0.4388, + "step": 7090 + }, + { + "epoch": 3.8496199782844736, + "grad_norm": 10.123066437797494, + "learning_rate": 2.650600980882104e-06, + "loss": 0.3086, + "step": 7091 + }, + { + "epoch": 3.8501628664495113, + "grad_norm": 12.006465921202063, + "learning_rate": 2.6482165532159275e-06, + "loss": 0.4117, + "step": 7092 + }, + { + "epoch": 3.8507057546145496, + "grad_norm": 15.608180907150922, + "learning_rate": 2.645833034830043e-06, + "loss": 0.5819, + "step": 7093 + }, + { + "epoch": 3.8512486427795873, + "grad_norm": 18.214304962348578, + "learning_rate": 2.6434504260192485e-06, + "loss": 0.6414, + "step": 7094 + }, + { + "epoch": 3.8517915309446256, + "grad_norm": 9.263854181848101, + "learning_rate": 2.64106872707823e-06, + "loss": 0.3855, + "step": 7095 + }, + { + "epoch": 3.8523344191096633, + "grad_norm": 11.56441830671296, + "learning_rate": 2.638687938301557e-06, + "loss": 0.3565, + "step": 7096 + }, + { + "epoch": 3.8528773072747016, + "grad_norm": 13.250463401216958, + "learning_rate": 2.636308059983699e-06, + "loss": 0.6425, + "step": 7097 + }, + { + "epoch": 3.8534201954397393, + "grad_norm": 14.438578870413643, + "learning_rate": 2.6339290924189886e-06, + "loss": 0.5401, + "step": 7098 + }, + { + "epoch": 3.8539630836047776, + "grad_norm": 19.454185693944922, + "learning_rate": 2.6315510359016736e-06, + "loss": 0.5113, + "step": 7099 + }, + { + "epoch": 3.8545059717698154, + "grad_norm": 14.281154826114456, + "learning_rate": 2.629173890725866e-06, + "loss": 0.6589, + "step": 7100 + }, + { + "epoch": 3.8550488599348536, + "grad_norm": 10.71424151041571, + "learning_rate": 2.62679765718558e-06, + "loss": 0.5477, + "step": 7101 + }, + { + "epoch": 3.8555917480998914, + "grad_norm": 15.907412596919608, + "learning_rate": 2.6244223355747144e-06, + "loss": 0.5586, + "step": 7102 + }, + { + "epoch": 3.8561346362649296, + "grad_norm": 14.500284217616562, + "learning_rate": 2.6220479261870436e-06, + "loss": 0.4131, + "step": 7103 + }, + { + "epoch": 3.8566775244299674, + "grad_norm": 16.114243210903357, + "learning_rate": 2.6196744293162503e-06, + "loss": 0.8063, + "step": 7104 + }, + { + "epoch": 3.8572204125950056, + "grad_norm": 10.762654220587946, + "learning_rate": 2.6173018452558786e-06, + "loss": 0.3736, + "step": 7105 + }, + { + "epoch": 3.8577633007600434, + "grad_norm": 8.619518240537495, + "learning_rate": 2.614930174299388e-06, + "loss": 0.4294, + "step": 7106 + }, + { + "epoch": 3.8583061889250816, + "grad_norm": 13.731123733938277, + "learning_rate": 2.612559416740098e-06, + "loss": 0.5039, + "step": 7107 + }, + { + "epoch": 3.8588490770901194, + "grad_norm": 9.925845082716284, + "learning_rate": 2.6101895728712355e-06, + "loss": 0.4914, + "step": 7108 + }, + { + "epoch": 3.8593919652551576, + "grad_norm": 9.305166559901256, + "learning_rate": 2.6078206429859044e-06, + "loss": 0.2875, + "step": 7109 + }, + { + "epoch": 3.8599348534201954, + "grad_norm": 11.732054868905763, + "learning_rate": 2.6054526273770975e-06, + "loss": 0.3843, + "step": 7110 + }, + { + "epoch": 3.8604777415852336, + "grad_norm": 11.857351610067585, + "learning_rate": 2.603085526337694e-06, + "loss": 0.431, + "step": 7111 + }, + { + "epoch": 3.8610206297502714, + "grad_norm": 8.758502750249699, + "learning_rate": 2.6007193401604626e-06, + "loss": 0.2799, + "step": 7112 + }, + { + "epoch": 3.8615635179153096, + "grad_norm": 11.732493468231032, + "learning_rate": 2.598354069138056e-06, + "loss": 0.4618, + "step": 7113 + }, + { + "epoch": 3.8621064060803474, + "grad_norm": 12.849925851696895, + "learning_rate": 2.595989713563014e-06, + "loss": 0.6998, + "step": 7114 + }, + { + "epoch": 3.8626492942453856, + "grad_norm": 8.79473787534307, + "learning_rate": 2.593626273727765e-06, + "loss": 0.2626, + "step": 7115 + }, + { + "epoch": 3.8631921824104234, + "grad_norm": 10.941535841138563, + "learning_rate": 2.5912637499246218e-06, + "loss": 0.4595, + "step": 7116 + }, + { + "epoch": 3.8637350705754616, + "grad_norm": 12.201214800033737, + "learning_rate": 2.588902142445786e-06, + "loss": 0.3282, + "step": 7117 + }, + { + "epoch": 3.8642779587404994, + "grad_norm": 12.348184440426978, + "learning_rate": 2.5865414515833455e-06, + "loss": 0.5828, + "step": 7118 + }, + { + "epoch": 3.8648208469055376, + "grad_norm": 13.160180910276585, + "learning_rate": 2.5841816776292727e-06, + "loss": 0.3516, + "step": 7119 + }, + { + "epoch": 3.8653637350705754, + "grad_norm": 19.57808856867751, + "learning_rate": 2.581822820875429e-06, + "loss": 0.8939, + "step": 7120 + }, + { + "epoch": 3.8659066232356136, + "grad_norm": 10.92966233467966, + "learning_rate": 2.5794648816135627e-06, + "loss": 0.3509, + "step": 7121 + }, + { + "epoch": 3.8664495114006514, + "grad_norm": 14.189811177607414, + "learning_rate": 2.577107860135305e-06, + "loss": 0.5024, + "step": 7122 + }, + { + "epoch": 3.8669923995656896, + "grad_norm": 10.963014195472391, + "learning_rate": 2.5747517567321776e-06, + "loss": 0.3253, + "step": 7123 + }, + { + "epoch": 3.8675352877307274, + "grad_norm": 16.06753478220038, + "learning_rate": 2.5723965716955878e-06, + "loss": 0.5281, + "step": 7124 + }, + { + "epoch": 3.8680781758957656, + "grad_norm": 10.96537764415174, + "learning_rate": 2.5700423053168276e-06, + "loss": 0.5059, + "step": 7125 + }, + { + "epoch": 3.8686210640608034, + "grad_norm": 9.185810250309261, + "learning_rate": 2.567688957887077e-06, + "loss": 0.3768, + "step": 7126 + }, + { + "epoch": 3.8691639522258416, + "grad_norm": 13.984895212503027, + "learning_rate": 2.565336529697401e-06, + "loss": 0.5294, + "step": 7127 + }, + { + "epoch": 3.8697068403908794, + "grad_norm": 14.97489290708819, + "learning_rate": 2.562985021038752e-06, + "loss": 0.6177, + "step": 7128 + }, + { + "epoch": 3.8702497285559176, + "grad_norm": 15.68383969722191, + "learning_rate": 2.560634432201966e-06, + "loss": 0.4874, + "step": 7129 + }, + { + "epoch": 3.8707926167209554, + "grad_norm": 13.06746691168529, + "learning_rate": 2.5582847634777753e-06, + "loss": 0.771, + "step": 7130 + }, + { + "epoch": 3.8713355048859937, + "grad_norm": 11.661980178584857, + "learning_rate": 2.55593601515678e-06, + "loss": 0.4957, + "step": 7131 + }, + { + "epoch": 3.8718783930510314, + "grad_norm": 9.687387371930855, + "learning_rate": 2.5535881875294875e-06, + "loss": 0.238, + "step": 7132 + }, + { + "epoch": 3.8724212812160697, + "grad_norm": 12.53249739465081, + "learning_rate": 2.551241280886274e-06, + "loss": 0.3521, + "step": 7133 + }, + { + "epoch": 3.8729641693811074, + "grad_norm": 18.057246108983335, + "learning_rate": 2.548895295517406e-06, + "loss": 0.9307, + "step": 7134 + }, + { + "epoch": 3.8735070575461457, + "grad_norm": 15.513117876289204, + "learning_rate": 2.546550231713051e-06, + "loss": 0.6529, + "step": 7135 + }, + { + "epoch": 3.8740499457111834, + "grad_norm": 11.883743610451969, + "learning_rate": 2.544206089763235e-06, + "loss": 0.7861, + "step": 7136 + }, + { + "epoch": 3.8745928338762217, + "grad_norm": 14.684059381744154, + "learning_rate": 2.5418628699578986e-06, + "loss": 0.412, + "step": 7137 + }, + { + "epoch": 3.8751357220412594, + "grad_norm": 7.135402796193207, + "learning_rate": 2.539520572586843e-06, + "loss": 0.2079, + "step": 7138 + }, + { + "epoch": 3.8756786102062977, + "grad_norm": 10.289771258088079, + "learning_rate": 2.5371791979397766e-06, + "loss": 0.3274, + "step": 7139 + }, + { + "epoch": 3.8762214983713354, + "grad_norm": 9.949109620860526, + "learning_rate": 2.5348387463062808e-06, + "loss": 0.4355, + "step": 7140 + }, + { + "epoch": 3.8767643865363732, + "grad_norm": 10.991037946192776, + "learning_rate": 2.5324992179758268e-06, + "loss": 0.4432, + "step": 7141 + }, + { + "epoch": 3.8773072747014115, + "grad_norm": 15.18957067683597, + "learning_rate": 2.5301606132377733e-06, + "loss": 0.8666, + "step": 7142 + }, + { + "epoch": 3.8778501628664497, + "grad_norm": 14.818504355332566, + "learning_rate": 2.5278229323813553e-06, + "loss": 0.6253, + "step": 7143 + }, + { + "epoch": 3.8783930510314875, + "grad_norm": 13.754502247150116, + "learning_rate": 2.5254861756957115e-06, + "loss": 0.5842, + "step": 7144 + }, + { + "epoch": 3.8789359391965252, + "grad_norm": 19.3020908869945, + "learning_rate": 2.5231503434698435e-06, + "loss": 0.6372, + "step": 7145 + }, + { + "epoch": 3.8794788273615635, + "grad_norm": 8.868163440639846, + "learning_rate": 2.5208154359926606e-06, + "loss": 0.4229, + "step": 7146 + }, + { + "epoch": 3.8800217155266017, + "grad_norm": 10.130535143618188, + "learning_rate": 2.5184814535529457e-06, + "loss": 0.3024, + "step": 7147 + }, + { + "epoch": 3.8805646036916395, + "grad_norm": 10.20257149827293, + "learning_rate": 2.5161483964393676e-06, + "loss": 0.3697, + "step": 7148 + }, + { + "epoch": 3.8811074918566772, + "grad_norm": 17.03033594136655, + "learning_rate": 2.513816264940483e-06, + "loss": 0.6084, + "step": 7149 + }, + { + "epoch": 3.8816503800217155, + "grad_norm": 9.767302667294265, + "learning_rate": 2.5114850593447336e-06, + "loss": 0.2872, + "step": 7150 + }, + { + "epoch": 3.8821932681867537, + "grad_norm": 11.550149537445538, + "learning_rate": 2.509154779940447e-06, + "loss": 0.4796, + "step": 7151 + }, + { + "epoch": 3.8827361563517915, + "grad_norm": 13.607872338694929, + "learning_rate": 2.5068254270158364e-06, + "loss": 0.6471, + "step": 7152 + }, + { + "epoch": 3.8832790445168293, + "grad_norm": 13.786046110369192, + "learning_rate": 2.504497000859e-06, + "loss": 0.6666, + "step": 7153 + }, + { + "epoch": 3.8838219326818675, + "grad_norm": 11.281665232889665, + "learning_rate": 2.5021695017579193e-06, + "loss": 0.5587, + "step": 7154 + }, + { + "epoch": 3.8843648208469057, + "grad_norm": 11.481761454219619, + "learning_rate": 2.4998429300004657e-06, + "loss": 0.3071, + "step": 7155 + }, + { + "epoch": 3.8849077090119435, + "grad_norm": 13.28075154587505, + "learning_rate": 2.4975172858743914e-06, + "loss": 0.3665, + "step": 7156 + }, + { + "epoch": 3.8854505971769813, + "grad_norm": 9.802453085562453, + "learning_rate": 2.4951925696673352e-06, + "loss": 0.3148, + "step": 7157 + }, + { + "epoch": 3.8859934853420195, + "grad_norm": 14.520688569604944, + "learning_rate": 2.492868781666824e-06, + "loss": 0.4942, + "step": 7158 + }, + { + "epoch": 3.8865363735070577, + "grad_norm": 10.711275878066797, + "learning_rate": 2.4905459221602667e-06, + "loss": 0.3353, + "step": 7159 + }, + { + "epoch": 3.8870792616720955, + "grad_norm": 12.549002018795433, + "learning_rate": 2.488223991434955e-06, + "loss": 0.3241, + "step": 7160 + }, + { + "epoch": 3.8876221498371333, + "grad_norm": 13.961472675915795, + "learning_rate": 2.485902989778077e-06, + "loss": 0.5143, + "step": 7161 + }, + { + "epoch": 3.8881650380021715, + "grad_norm": 10.194253119170678, + "learning_rate": 2.483582917476691e-06, + "loss": 0.3732, + "step": 7162 + }, + { + "epoch": 3.8887079261672097, + "grad_norm": 11.408277950635817, + "learning_rate": 2.481263774817748e-06, + "loss": 0.5475, + "step": 7163 + }, + { + "epoch": 3.8892508143322475, + "grad_norm": 11.720624145965107, + "learning_rate": 2.4789455620880855e-06, + "loss": 0.4152, + "step": 7164 + }, + { + "epoch": 3.8897937024972853, + "grad_norm": 17.498990068476704, + "learning_rate": 2.4766282795744225e-06, + "loss": 0.4959, + "step": 7165 + }, + { + "epoch": 3.8903365906623235, + "grad_norm": 10.025081073758155, + "learning_rate": 2.474311927563364e-06, + "loss": 0.2563, + "step": 7166 + }, + { + "epoch": 3.8908794788273617, + "grad_norm": 11.064975553515668, + "learning_rate": 2.4719965063413975e-06, + "loss": 0.4892, + "step": 7167 + }, + { + "epoch": 3.8914223669923995, + "grad_norm": 11.4169177186445, + "learning_rate": 2.4696820161949076e-06, + "loss": 0.4661, + "step": 7168 + }, + { + "epoch": 3.8919652551574373, + "grad_norm": 11.127726171001362, + "learning_rate": 2.4673684574101407e-06, + "loss": 0.5133, + "step": 7169 + }, + { + "epoch": 3.8925081433224755, + "grad_norm": 14.4532269361977, + "learning_rate": 2.4650558302732554e-06, + "loss": 0.6211, + "step": 7170 + }, + { + "epoch": 3.8930510314875137, + "grad_norm": 11.501842216375938, + "learning_rate": 2.4627441350702697e-06, + "loss": 0.43, + "step": 7171 + }, + { + "epoch": 3.8935939196525515, + "grad_norm": 11.827951767183402, + "learning_rate": 2.460433372087099e-06, + "loss": 0.4745, + "step": 7172 + }, + { + "epoch": 3.8941368078175893, + "grad_norm": 16.63470564409548, + "learning_rate": 2.4581235416095516e-06, + "loss": 0.7871, + "step": 7173 + }, + { + "epoch": 3.8946796959826275, + "grad_norm": 13.957180122419292, + "learning_rate": 2.455814643923298e-06, + "loss": 0.435, + "step": 7174 + }, + { + "epoch": 3.8952225841476658, + "grad_norm": 14.844456130947902, + "learning_rate": 2.4535066793139194e-06, + "loss": 0.3498, + "step": 7175 + }, + { + "epoch": 3.8957654723127035, + "grad_norm": 13.063011297960943, + "learning_rate": 2.4511996480668554e-06, + "loss": 0.4807, + "step": 7176 + }, + { + "epoch": 3.8963083604777413, + "grad_norm": 11.343167538907203, + "learning_rate": 2.448893550467456e-06, + "loss": 0.6566, + "step": 7177 + }, + { + "epoch": 3.8968512486427795, + "grad_norm": 21.134566936836904, + "learning_rate": 2.4465883868009323e-06, + "loss": 0.7527, + "step": 7178 + }, + { + "epoch": 3.8973941368078178, + "grad_norm": 12.658548113922626, + "learning_rate": 2.4442841573523967e-06, + "loss": 0.388, + "step": 7179 + }, + { + "epoch": 3.8979370249728555, + "grad_norm": 10.941499945328948, + "learning_rate": 2.4419808624068396e-06, + "loss": 0.585, + "step": 7180 + }, + { + "epoch": 3.8984799131378933, + "grad_norm": 14.25189382311016, + "learning_rate": 2.4396785022491343e-06, + "loss": 0.6702, + "step": 7181 + }, + { + "epoch": 3.8990228013029316, + "grad_norm": 16.90343774367846, + "learning_rate": 2.4373770771640448e-06, + "loss": 0.5498, + "step": 7182 + }, + { + "epoch": 3.8995656894679698, + "grad_norm": 14.478488557861258, + "learning_rate": 2.4350765874362047e-06, + "loss": 0.6703, + "step": 7183 + }, + { + "epoch": 3.9001085776330076, + "grad_norm": 11.280629928495003, + "learning_rate": 2.4327770333501522e-06, + "loss": 0.387, + "step": 7184 + }, + { + "epoch": 3.9006514657980453, + "grad_norm": 10.389018994464337, + "learning_rate": 2.430478415190297e-06, + "loss": 0.4396, + "step": 7185 + }, + { + "epoch": 3.9011943539630836, + "grad_norm": 12.229139584290909, + "learning_rate": 2.4281807332409358e-06, + "loss": 0.639, + "step": 7186 + }, + { + "epoch": 3.901737242128122, + "grad_norm": 18.238591257772253, + "learning_rate": 2.425883987786248e-06, + "loss": 0.6162, + "step": 7187 + }, + { + "epoch": 3.9022801302931596, + "grad_norm": 14.326334376777995, + "learning_rate": 2.423588179110301e-06, + "loss": 0.8227, + "step": 7188 + }, + { + "epoch": 3.9028230184581973, + "grad_norm": 14.43500891639456, + "learning_rate": 2.4212933074970423e-06, + "loss": 0.5057, + "step": 7189 + }, + { + "epoch": 3.9033659066232356, + "grad_norm": 14.30031793020835, + "learning_rate": 2.4189993732303063e-06, + "loss": 0.4514, + "step": 7190 + }, + { + "epoch": 3.903908794788274, + "grad_norm": 9.814097341569619, + "learning_rate": 2.4167063765938103e-06, + "loss": 0.3212, + "step": 7191 + }, + { + "epoch": 3.9044516829533116, + "grad_norm": 13.100322061273497, + "learning_rate": 2.4144143178711555e-06, + "loss": 0.5785, + "step": 7192 + }, + { + "epoch": 3.9049945711183494, + "grad_norm": 15.547319070400764, + "learning_rate": 2.412123197345827e-06, + "loss": 0.6, + "step": 7193 + }, + { + "epoch": 3.9055374592833876, + "grad_norm": 10.568502118830102, + "learning_rate": 2.409833015301195e-06, + "loss": 0.4499, + "step": 7194 + }, + { + "epoch": 3.906080347448426, + "grad_norm": 12.216551577948717, + "learning_rate": 2.4075437720205132e-06, + "loss": 0.3987, + "step": 7195 + }, + { + "epoch": 3.9066232356134636, + "grad_norm": 14.203010493004756, + "learning_rate": 2.4052554677869165e-06, + "loss": 0.7274, + "step": 7196 + }, + { + "epoch": 3.9071661237785014, + "grad_norm": 16.842358223683046, + "learning_rate": 2.4029681028834293e-06, + "loss": 0.6862, + "step": 7197 + }, + { + "epoch": 3.9077090119435396, + "grad_norm": 17.087528448096148, + "learning_rate": 2.4006816775929553e-06, + "loss": 0.6367, + "step": 7198 + }, + { + "epoch": 3.908251900108578, + "grad_norm": 13.559074748741233, + "learning_rate": 2.3983961921982823e-06, + "loss": 0.497, + "step": 7199 + }, + { + "epoch": 3.9087947882736156, + "grad_norm": 11.82349002805914, + "learning_rate": 2.3961116469820834e-06, + "loss": 0.3808, + "step": 7200 + }, + { + "epoch": 3.9093376764386534, + "grad_norm": 9.536111713709516, + "learning_rate": 2.3938280422269143e-06, + "loss": 0.2577, + "step": 7201 + }, + { + "epoch": 3.9098805646036916, + "grad_norm": 13.40690087309448, + "learning_rate": 2.3915453782152166e-06, + "loss": 0.5696, + "step": 7202 + }, + { + "epoch": 3.91042345276873, + "grad_norm": 11.767180310470792, + "learning_rate": 2.3892636552293114e-06, + "loss": 0.447, + "step": 7203 + }, + { + "epoch": 3.9109663409337676, + "grad_norm": 15.579650605609013, + "learning_rate": 2.3869828735514076e-06, + "loss": 0.5241, + "step": 7204 + }, + { + "epoch": 3.9115092290988054, + "grad_norm": 10.113811350963385, + "learning_rate": 2.3847030334635923e-06, + "loss": 0.3542, + "step": 7205 + }, + { + "epoch": 3.9120521172638436, + "grad_norm": 10.589793992165902, + "learning_rate": 2.3824241352478484e-06, + "loss": 0.2795, + "step": 7206 + }, + { + "epoch": 3.912595005428882, + "grad_norm": 7.9421075555522505, + "learning_rate": 2.3801461791860226e-06, + "loss": 0.2936, + "step": 7207 + }, + { + "epoch": 3.9131378935939196, + "grad_norm": 9.535162659306835, + "learning_rate": 2.377869165559867e-06, + "loss": 0.5889, + "step": 7208 + }, + { + "epoch": 3.9136807817589574, + "grad_norm": 11.900400034485365, + "learning_rate": 2.375593094650995e-06, + "loss": 0.5422, + "step": 7209 + }, + { + "epoch": 3.9142236699239956, + "grad_norm": 11.395565814320083, + "learning_rate": 2.3733179667409247e-06, + "loss": 0.4194, + "step": 7210 + }, + { + "epoch": 3.914766558089034, + "grad_norm": 13.14302661684852, + "learning_rate": 2.3710437821110456e-06, + "loss": 0.5611, + "step": 7211 + }, + { + "epoch": 3.9153094462540716, + "grad_norm": 14.854169528681664, + "learning_rate": 2.3687705410426242e-06, + "loss": 0.676, + "step": 7212 + }, + { + "epoch": 3.9158523344191094, + "grad_norm": 14.377106944895534, + "learning_rate": 2.3664982438168305e-06, + "loss": 0.5029, + "step": 7213 + }, + { + "epoch": 3.9163952225841476, + "grad_norm": 16.2949751889776, + "learning_rate": 2.364226890714694e-06, + "loss": 0.4424, + "step": 7214 + }, + { + "epoch": 3.916938110749186, + "grad_norm": 12.453892587793705, + "learning_rate": 2.3619564820171515e-06, + "loss": 0.3498, + "step": 7215 + }, + { + "epoch": 3.9174809989142236, + "grad_norm": 10.510207014160567, + "learning_rate": 2.359687018004998e-06, + "loss": 0.4104, + "step": 7216 + }, + { + "epoch": 3.9180238870792614, + "grad_norm": 10.817797408933902, + "learning_rate": 2.3574184989589345e-06, + "loss": 0.466, + "step": 7217 + }, + { + "epoch": 3.9185667752442996, + "grad_norm": 9.615419411562353, + "learning_rate": 2.3551509251595315e-06, + "loss": 0.3975, + "step": 7218 + }, + { + "epoch": 3.919109663409338, + "grad_norm": 10.629184513933357, + "learning_rate": 2.3528842968872456e-06, + "loss": 0.3305, + "step": 7219 + }, + { + "epoch": 3.9196525515743756, + "grad_norm": 14.741934283339353, + "learning_rate": 2.350618614422421e-06, + "loss": 0.5822, + "step": 7220 + }, + { + "epoch": 3.9201954397394134, + "grad_norm": 10.145231872720867, + "learning_rate": 2.348353878045272e-06, + "loss": 0.3223, + "step": 7221 + }, + { + "epoch": 3.9207383279044516, + "grad_norm": 9.719525157659627, + "learning_rate": 2.346090088035913e-06, + "loss": 0.323, + "step": 7222 + }, + { + "epoch": 3.92128121606949, + "grad_norm": 11.153861466428655, + "learning_rate": 2.3438272446743293e-06, + "loss": 0.4453, + "step": 7223 + }, + { + "epoch": 3.9218241042345277, + "grad_norm": 16.881204557198902, + "learning_rate": 2.3415653482403954e-06, + "loss": 0.8422, + "step": 7224 + }, + { + "epoch": 3.9223669923995654, + "grad_norm": 13.409647968671297, + "learning_rate": 2.339304399013864e-06, + "loss": 0.5316, + "step": 7225 + }, + { + "epoch": 3.9229098805646037, + "grad_norm": 17.78528943034469, + "learning_rate": 2.337044397274375e-06, + "loss": 1.0815, + "step": 7226 + }, + { + "epoch": 3.923452768729642, + "grad_norm": 13.521330366302392, + "learning_rate": 2.3347853433014467e-06, + "loss": 0.6285, + "step": 7227 + }, + { + "epoch": 3.9239956568946797, + "grad_norm": 15.033911656714881, + "learning_rate": 2.3325272373744844e-06, + "loss": 0.4578, + "step": 7228 + }, + { + "epoch": 3.9245385450597174, + "grad_norm": 13.063924995112636, + "learning_rate": 2.330270079772774e-06, + "loss": 0.4246, + "step": 7229 + }, + { + "epoch": 3.9250814332247557, + "grad_norm": 11.872225399295557, + "learning_rate": 2.328013870775483e-06, + "loss": 0.354, + "step": 7230 + }, + { + "epoch": 3.925624321389794, + "grad_norm": 10.802228004866322, + "learning_rate": 2.325758610661664e-06, + "loss": 0.2926, + "step": 7231 + }, + { + "epoch": 3.9261672095548317, + "grad_norm": 10.24423290229698, + "learning_rate": 2.323504299710251e-06, + "loss": 0.3887, + "step": 7232 + }, + { + "epoch": 3.9267100977198695, + "grad_norm": 10.40790875242795, + "learning_rate": 2.32125093820006e-06, + "loss": 0.3774, + "step": 7233 + }, + { + "epoch": 3.9272529858849077, + "grad_norm": 10.819416071470908, + "learning_rate": 2.3189985264097925e-06, + "loss": 0.4433, + "step": 7234 + }, + { + "epoch": 3.927795874049946, + "grad_norm": 14.400478824494995, + "learning_rate": 2.316747064618028e-06, + "loss": 0.6139, + "step": 7235 + }, + { + "epoch": 3.9283387622149837, + "grad_norm": 13.400576026026663, + "learning_rate": 2.3144965531032314e-06, + "loss": 0.6807, + "step": 7236 + }, + { + "epoch": 3.9288816503800215, + "grad_norm": 12.144971157259214, + "learning_rate": 2.3122469921437507e-06, + "loss": 0.406, + "step": 7237 + }, + { + "epoch": 3.9294245385450597, + "grad_norm": 13.566955931406746, + "learning_rate": 2.3099983820178116e-06, + "loss": 0.6583, + "step": 7238 + }, + { + "epoch": 3.929967426710098, + "grad_norm": 13.453764479761338, + "learning_rate": 2.3077507230035345e-06, + "loss": 0.8706, + "step": 7239 + }, + { + "epoch": 3.9305103148751357, + "grad_norm": 15.149717368244767, + "learning_rate": 2.305504015378904e-06, + "loss": 0.6808, + "step": 7240 + }, + { + "epoch": 3.9310532030401735, + "grad_norm": 15.68496135669598, + "learning_rate": 2.303258259421801e-06, + "loss": 0.4163, + "step": 7241 + }, + { + "epoch": 3.9315960912052117, + "grad_norm": 18.73387442818072, + "learning_rate": 2.301013455409983e-06, + "loss": 0.972, + "step": 7242 + }, + { + "epoch": 3.93213897937025, + "grad_norm": 11.659704005587447, + "learning_rate": 2.298769603621088e-06, + "loss": 0.4766, + "step": 7243 + }, + { + "epoch": 3.9326818675352877, + "grad_norm": 12.907778650132892, + "learning_rate": 2.296526704332648e-06, + "loss": 0.5669, + "step": 7244 + }, + { + "epoch": 3.9332247557003255, + "grad_norm": 14.095811490299557, + "learning_rate": 2.294284757822057e-06, + "loss": 0.6699, + "step": 7245 + }, + { + "epoch": 3.9337676438653637, + "grad_norm": 14.347126906125892, + "learning_rate": 2.292043764366615e-06, + "loss": 0.5478, + "step": 7246 + }, + { + "epoch": 3.934310532030402, + "grad_norm": 14.15574104757881, + "learning_rate": 2.289803724243478e-06, + "loss": 0.3451, + "step": 7247 + }, + { + "epoch": 3.9348534201954397, + "grad_norm": 16.273077860218656, + "learning_rate": 2.28756463772971e-06, + "loss": 0.433, + "step": 7248 + }, + { + "epoch": 3.9353963083604775, + "grad_norm": 12.905791138340465, + "learning_rate": 2.2853265051022376e-06, + "loss": 0.3549, + "step": 7249 + }, + { + "epoch": 3.9359391965255157, + "grad_norm": 10.740402790195867, + "learning_rate": 2.283089326637875e-06, + "loss": 0.294, + "step": 7250 + }, + { + "epoch": 3.936482084690554, + "grad_norm": 16.206654981483148, + "learning_rate": 2.2808531026133297e-06, + "loss": 0.4773, + "step": 7251 + }, + { + "epoch": 3.9370249728555917, + "grad_norm": 13.723535114050927, + "learning_rate": 2.278617833305169e-06, + "loss": 0.5248, + "step": 7252 + }, + { + "epoch": 3.9375678610206295, + "grad_norm": 12.638788461264767, + "learning_rate": 2.2763835189898665e-06, + "loss": 0.6087, + "step": 7253 + }, + { + "epoch": 3.9381107491856677, + "grad_norm": 12.333038982803954, + "learning_rate": 2.2741501599437543e-06, + "loss": 0.3829, + "step": 7254 + }, + { + "epoch": 3.938653637350706, + "grad_norm": 11.07699141293442, + "learning_rate": 2.2719177564430662e-06, + "loss": 0.6625, + "step": 7255 + }, + { + "epoch": 3.9391965255157437, + "grad_norm": 14.354975596263804, + "learning_rate": 2.2696863087639063e-06, + "loss": 0.6906, + "step": 7256 + }, + { + "epoch": 3.9397394136807815, + "grad_norm": 10.159046635813173, + "learning_rate": 2.2674558171822646e-06, + "loss": 0.4371, + "step": 7257 + }, + { + "epoch": 3.9402823018458197, + "grad_norm": 18.017466534966523, + "learning_rate": 2.265226281974011e-06, + "loss": 0.5347, + "step": 7258 + }, + { + "epoch": 3.940825190010858, + "grad_norm": 13.962772198264076, + "learning_rate": 2.2629977034148988e-06, + "loss": 0.4712, + "step": 7259 + }, + { + "epoch": 3.9413680781758957, + "grad_norm": 10.058544580261096, + "learning_rate": 2.2607700817805635e-06, + "loss": 0.3168, + "step": 7260 + }, + { + "epoch": 3.9419109663409335, + "grad_norm": 18.61661644453424, + "learning_rate": 2.258543417346514e-06, + "loss": 0.5639, + "step": 7261 + }, + { + "epoch": 3.9424538545059717, + "grad_norm": 12.38740485358688, + "learning_rate": 2.256317710388155e-06, + "loss": 0.4182, + "step": 7262 + }, + { + "epoch": 3.94299674267101, + "grad_norm": 12.633562464203537, + "learning_rate": 2.254092961180764e-06, + "loss": 0.5299, + "step": 7263 + }, + { + "epoch": 3.9435396308360477, + "grad_norm": 15.51128932246205, + "learning_rate": 2.2518691699995e-06, + "loss": 0.6563, + "step": 7264 + }, + { + "epoch": 3.9440825190010855, + "grad_norm": 14.124949663680717, + "learning_rate": 2.2496463371194065e-06, + "loss": 0.7836, + "step": 7265 + }, + { + "epoch": 3.9446254071661238, + "grad_norm": 14.035611568552724, + "learning_rate": 2.247424462815405e-06, + "loss": 0.4113, + "step": 7266 + }, + { + "epoch": 3.945168295331162, + "grad_norm": 13.30563635940546, + "learning_rate": 2.2452035473623022e-06, + "loss": 0.449, + "step": 7267 + }, + { + "epoch": 3.9457111834961998, + "grad_norm": 12.093768184800798, + "learning_rate": 2.242983591034784e-06, + "loss": 0.5745, + "step": 7268 + }, + { + "epoch": 3.9462540716612375, + "grad_norm": 13.19161002685443, + "learning_rate": 2.2407645941074185e-06, + "loss": 0.6073, + "step": 7269 + }, + { + "epoch": 3.9467969598262758, + "grad_norm": 10.454523102469183, + "learning_rate": 2.238546556854655e-06, + "loss": 0.265, + "step": 7270 + }, + { + "epoch": 3.947339847991314, + "grad_norm": 12.97022293790845, + "learning_rate": 2.2363294795508217e-06, + "loss": 0.5445, + "step": 7271 + }, + { + "epoch": 3.9478827361563518, + "grad_norm": 12.83418150037974, + "learning_rate": 2.234113362470133e-06, + "loss": 0.4941, + "step": 7272 + }, + { + "epoch": 3.9484256243213895, + "grad_norm": 12.615689050586889, + "learning_rate": 2.23189820588668e-06, + "loss": 0.8588, + "step": 7273 + }, + { + "epoch": 3.9489685124864278, + "grad_norm": 14.878456274053418, + "learning_rate": 2.2296840100744375e-06, + "loss": 0.4019, + "step": 7274 + }, + { + "epoch": 3.949511400651466, + "grad_norm": 13.017249525734119, + "learning_rate": 2.227470775307261e-06, + "loss": 0.5343, + "step": 7275 + }, + { + "epoch": 3.950054288816504, + "grad_norm": 11.575062166378828, + "learning_rate": 2.2252585018588836e-06, + "loss": 0.3775, + "step": 7276 + }, + { + "epoch": 3.9505971769815416, + "grad_norm": 15.290489125989911, + "learning_rate": 2.2230471900029303e-06, + "loss": 0.4912, + "step": 7277 + }, + { + "epoch": 3.95114006514658, + "grad_norm": 17.764334443185927, + "learning_rate": 2.220836840012891e-06, + "loss": 0.7215, + "step": 7278 + }, + { + "epoch": 3.951682953311618, + "grad_norm": 14.154407180706324, + "learning_rate": 2.218627452162154e-06, + "loss": 0.6472, + "step": 7279 + }, + { + "epoch": 3.952225841476656, + "grad_norm": 11.225658955685093, + "learning_rate": 2.2164190267239737e-06, + "loss": 0.2478, + "step": 7280 + }, + { + "epoch": 3.9527687296416936, + "grad_norm": 12.814332016509574, + "learning_rate": 2.2142115639714935e-06, + "loss": 0.5622, + "step": 7281 + }, + { + "epoch": 3.953311617806732, + "grad_norm": 19.419294319476467, + "learning_rate": 2.2120050641777345e-06, + "loss": 0.7996, + "step": 7282 + }, + { + "epoch": 3.95385450597177, + "grad_norm": 19.022041488664335, + "learning_rate": 2.209799527615599e-06, + "loss": 0.7117, + "step": 7283 + }, + { + "epoch": 3.954397394136808, + "grad_norm": 12.40634141862146, + "learning_rate": 2.20759495455788e-06, + "loss": 0.3793, + "step": 7284 + }, + { + "epoch": 3.9549402823018456, + "grad_norm": 8.873847512885336, + "learning_rate": 2.20539134527723e-06, + "loss": 0.3723, + "step": 7285 + }, + { + "epoch": 3.955483170466884, + "grad_norm": 15.833917455665446, + "learning_rate": 2.203188700046207e-06, + "loss": 0.6743, + "step": 7286 + }, + { + "epoch": 3.956026058631922, + "grad_norm": 12.353488572936646, + "learning_rate": 2.2009870191372263e-06, + "loss": 0.4195, + "step": 7287 + }, + { + "epoch": 3.95656894679696, + "grad_norm": 19.03028178846917, + "learning_rate": 2.198786302822603e-06, + "loss": 0.9055, + "step": 7288 + }, + { + "epoch": 3.9571118349619976, + "grad_norm": 15.81794212526427, + "learning_rate": 2.1965865513745265e-06, + "loss": 0.4088, + "step": 7289 + }, + { + "epoch": 3.957654723127036, + "grad_norm": 14.909566500608973, + "learning_rate": 2.1943877650650556e-06, + "loss": 0.9969, + "step": 7290 + }, + { + "epoch": 3.958197611292074, + "grad_norm": 12.332907917932245, + "learning_rate": 2.192189944166153e-06, + "loss": 0.3762, + "step": 7291 + }, + { + "epoch": 3.958740499457112, + "grad_norm": 13.036980814291068, + "learning_rate": 2.189993088949636e-06, + "loss": 0.4931, + "step": 7292 + }, + { + "epoch": 3.9592833876221496, + "grad_norm": 15.24047006419779, + "learning_rate": 2.187797199687224e-06, + "loss": 0.5143, + "step": 7293 + }, + { + "epoch": 3.959826275787188, + "grad_norm": 12.214872928235861, + "learning_rate": 2.185602276650505e-06, + "loss": 0.4249, + "step": 7294 + }, + { + "epoch": 3.960369163952226, + "grad_norm": 11.734417487796039, + "learning_rate": 2.18340832011095e-06, + "loss": 0.5159, + "step": 7295 + }, + { + "epoch": 3.960912052117264, + "grad_norm": 16.316727205367513, + "learning_rate": 2.181215330339912e-06, + "loss": 0.5141, + "step": 7296 + }, + { + "epoch": 3.9614549402823016, + "grad_norm": 13.896917482997276, + "learning_rate": 2.1790233076086243e-06, + "loss": 1.0396, + "step": 7297 + }, + { + "epoch": 3.96199782844734, + "grad_norm": 17.3480021128049, + "learning_rate": 2.1768322521882003e-06, + "loss": 0.43, + "step": 7298 + }, + { + "epoch": 3.962540716612378, + "grad_norm": 11.942778635499716, + "learning_rate": 2.1746421643496264e-06, + "loss": 0.383, + "step": 7299 + }, + { + "epoch": 3.963083604777416, + "grad_norm": 10.46804392176075, + "learning_rate": 2.1724530443637836e-06, + "loss": 0.6297, + "step": 7300 + }, + { + "epoch": 3.9636264929424536, + "grad_norm": 12.678863612803186, + "learning_rate": 2.1702648925014248e-06, + "loss": 0.4674, + "step": 7301 + }, + { + "epoch": 3.964169381107492, + "grad_norm": 7.353045500973433, + "learning_rate": 2.1680777090331816e-06, + "loss": 0.3103, + "step": 7302 + }, + { + "epoch": 3.96471226927253, + "grad_norm": 12.29878912174572, + "learning_rate": 2.1658914942295706e-06, + "loss": 0.5663, + "step": 7303 + }, + { + "epoch": 3.965255157437568, + "grad_norm": 8.497343343626396, + "learning_rate": 2.163706248360985e-06, + "loss": 0.3524, + "step": 7304 + }, + { + "epoch": 3.9657980456026056, + "grad_norm": 11.764041467674721, + "learning_rate": 2.1615219716977e-06, + "loss": 0.4548, + "step": 7305 + }, + { + "epoch": 3.966340933767644, + "grad_norm": 17.313346132201072, + "learning_rate": 2.1593386645098692e-06, + "loss": 1.0529, + "step": 7306 + }, + { + "epoch": 3.966883821932682, + "grad_norm": 13.90349930971142, + "learning_rate": 2.15715632706753e-06, + "loss": 0.5808, + "step": 7307 + }, + { + "epoch": 3.96742671009772, + "grad_norm": 13.776936265783004, + "learning_rate": 2.1549749596405945e-06, + "loss": 0.6697, + "step": 7308 + }, + { + "epoch": 3.9679695982627576, + "grad_norm": 16.504532446022996, + "learning_rate": 2.15279456249886e-06, + "loss": 0.5508, + "step": 7309 + }, + { + "epoch": 3.968512486427796, + "grad_norm": 9.948445138257954, + "learning_rate": 2.1506151359119997e-06, + "loss": 0.4461, + "step": 7310 + }, + { + "epoch": 3.969055374592834, + "grad_norm": 14.495691133016251, + "learning_rate": 2.1484366801495705e-06, + "loss": 0.7312, + "step": 7311 + }, + { + "epoch": 3.969598262757872, + "grad_norm": 16.82715961877643, + "learning_rate": 2.1462591954810054e-06, + "loss": 0.7524, + "step": 7312 + }, + { + "epoch": 3.9701411509229096, + "grad_norm": 11.225827985724983, + "learning_rate": 2.1440826821756213e-06, + "loss": 0.4354, + "step": 7313 + }, + { + "epoch": 3.970684039087948, + "grad_norm": 9.865289892982581, + "learning_rate": 2.141907140502607e-06, + "loss": 0.2906, + "step": 7314 + }, + { + "epoch": 3.971226927252986, + "grad_norm": 8.988273380587568, + "learning_rate": 2.139732570731048e-06, + "loss": 0.3062, + "step": 7315 + }, + { + "epoch": 3.971769815418024, + "grad_norm": 13.659816709804225, + "learning_rate": 2.1375589731298864e-06, + "loss": 0.382, + "step": 7316 + }, + { + "epoch": 3.9723127035830617, + "grad_norm": 15.628687568254344, + "learning_rate": 2.135386347967967e-06, + "loss": 0.6687, + "step": 7317 + }, + { + "epoch": 3.9728555917481, + "grad_norm": 10.910710203208446, + "learning_rate": 2.1332146955139963e-06, + "loss": 0.3909, + "step": 7318 + }, + { + "epoch": 3.973398479913138, + "grad_norm": 13.270954739017608, + "learning_rate": 2.13104401603657e-06, + "loss": 0.4808, + "step": 7319 + }, + { + "epoch": 3.973941368078176, + "grad_norm": 13.799228933349236, + "learning_rate": 2.12887430980416e-06, + "loss": 0.4717, + "step": 7320 + }, + { + "epoch": 3.9744842562432137, + "grad_norm": 12.928285689167119, + "learning_rate": 2.1267055770851185e-06, + "loss": 0.5013, + "step": 7321 + }, + { + "epoch": 3.975027144408252, + "grad_norm": 11.400185046713785, + "learning_rate": 2.124537818147684e-06, + "loss": 0.304, + "step": 7322 + }, + { + "epoch": 3.97557003257329, + "grad_norm": 10.537157565034327, + "learning_rate": 2.122371033259959e-06, + "loss": 0.3975, + "step": 7323 + }, + { + "epoch": 3.976112920738328, + "grad_norm": 14.40105822826117, + "learning_rate": 2.120205222689944e-06, + "loss": 0.6778, + "step": 7324 + }, + { + "epoch": 3.9766558089033657, + "grad_norm": 13.229824974968945, + "learning_rate": 2.118040386705501e-06, + "loss": 0.6712, + "step": 7325 + }, + { + "epoch": 3.977198697068404, + "grad_norm": 13.793638194835786, + "learning_rate": 2.1158765255743872e-06, + "loss": 0.4601, + "step": 7326 + }, + { + "epoch": 3.977741585233442, + "grad_norm": 15.505430296521286, + "learning_rate": 2.113713639564231e-06, + "loss": 0.8212, + "step": 7327 + }, + { + "epoch": 3.97828447339848, + "grad_norm": 17.09119757360545, + "learning_rate": 2.11155172894254e-06, + "loss": 0.7094, + "step": 7328 + }, + { + "epoch": 3.9788273615635177, + "grad_norm": 13.556531251232467, + "learning_rate": 2.1093907939767065e-06, + "loss": 0.4832, + "step": 7329 + }, + { + "epoch": 3.979370249728556, + "grad_norm": 14.237329918169115, + "learning_rate": 2.10723083493399e-06, + "loss": 0.3663, + "step": 7330 + }, + { + "epoch": 3.979913137893594, + "grad_norm": 9.012447122900646, + "learning_rate": 2.1050718520815485e-06, + "loss": 0.5217, + "step": 7331 + }, + { + "epoch": 3.980456026058632, + "grad_norm": 12.91832304111153, + "learning_rate": 2.1029138456863973e-06, + "loss": 0.7424, + "step": 7332 + }, + { + "epoch": 3.9809989142236697, + "grad_norm": 17.140779656879335, + "learning_rate": 2.1007568160154502e-06, + "loss": 0.5607, + "step": 7333 + }, + { + "epoch": 3.981541802388708, + "grad_norm": 15.535441443258492, + "learning_rate": 2.09860076333549e-06, + "loss": 0.6682, + "step": 7334 + }, + { + "epoch": 3.982084690553746, + "grad_norm": 9.108166871624157, + "learning_rate": 2.09644568791318e-06, + "loss": 0.3471, + "step": 7335 + }, + { + "epoch": 3.982627578718784, + "grad_norm": 13.8585496563886, + "learning_rate": 2.094291590015064e-06, + "loss": 0.4907, + "step": 7336 + }, + { + "epoch": 3.9831704668838217, + "grad_norm": 17.309058810158508, + "learning_rate": 2.0921384699075644e-06, + "loss": 0.7292, + "step": 7337 + }, + { + "epoch": 3.98371335504886, + "grad_norm": 11.888938470455745, + "learning_rate": 2.089986327856981e-06, + "loss": 0.5343, + "step": 7338 + }, + { + "epoch": 3.984256243213898, + "grad_norm": 14.96573909686316, + "learning_rate": 2.087835164129496e-06, + "loss": 0.436, + "step": 7339 + }, + { + "epoch": 3.984799131378936, + "grad_norm": 11.965090702213786, + "learning_rate": 2.085684978991168e-06, + "loss": 0.4307, + "step": 7340 + }, + { + "epoch": 3.9853420195439737, + "grad_norm": 9.089611131661139, + "learning_rate": 2.083535772707935e-06, + "loss": 0.3474, + "step": 7341 + }, + { + "epoch": 3.985884907709012, + "grad_norm": 15.457161419441805, + "learning_rate": 2.0813875455456156e-06, + "loss": 0.6491, + "step": 7342 + }, + { + "epoch": 3.98642779587405, + "grad_norm": 10.250596479107204, + "learning_rate": 2.0792402977699033e-06, + "loss": 0.3589, + "step": 7343 + }, + { + "epoch": 3.986970684039088, + "grad_norm": 14.949877353645148, + "learning_rate": 2.077094029646376e-06, + "loss": 0.8685, + "step": 7344 + }, + { + "epoch": 3.9875135722041257, + "grad_norm": 15.23923375936191, + "learning_rate": 2.074948741440486e-06, + "loss": 0.744, + "step": 7345 + }, + { + "epoch": 3.988056460369164, + "grad_norm": 11.133935275967207, + "learning_rate": 2.0728044334175667e-06, + "loss": 0.7823, + "step": 7346 + }, + { + "epoch": 3.988599348534202, + "grad_norm": 10.744562653679704, + "learning_rate": 2.0706611058428285e-06, + "loss": 0.3669, + "step": 7347 + }, + { + "epoch": 3.98914223669924, + "grad_norm": 12.814655704488615, + "learning_rate": 2.0685187589813625e-06, + "loss": 0.3812, + "step": 7348 + }, + { + "epoch": 3.9896851248642777, + "grad_norm": 10.282758906730834, + "learning_rate": 2.0663773930981367e-06, + "loss": 0.3229, + "step": 7349 + }, + { + "epoch": 3.990228013029316, + "grad_norm": 14.644174681230187, + "learning_rate": 2.064237008458e-06, + "loss": 0.7615, + "step": 7350 + }, + { + "epoch": 3.990770901194354, + "grad_norm": 16.98502298622554, + "learning_rate": 2.062097605325678e-06, + "loss": 0.5847, + "step": 7351 + }, + { + "epoch": 3.991313789359392, + "grad_norm": 12.815769655827422, + "learning_rate": 2.059959183965775e-06, + "loss": 0.3939, + "step": 7352 + }, + { + "epoch": 3.9918566775244297, + "grad_norm": 13.188644656309128, + "learning_rate": 2.057821744642774e-06, + "loss": 0.6187, + "step": 7353 + }, + { + "epoch": 3.992399565689468, + "grad_norm": 12.875041049936685, + "learning_rate": 2.0556852876210354e-06, + "loss": 0.3272, + "step": 7354 + }, + { + "epoch": 3.992942453854506, + "grad_norm": 13.754004827000545, + "learning_rate": 2.053549813164808e-06, + "loss": 0.4637, + "step": 7355 + }, + { + "epoch": 3.993485342019544, + "grad_norm": 10.58203405942224, + "learning_rate": 2.0514153215381983e-06, + "loss": 0.5784, + "step": 7356 + }, + { + "epoch": 3.9940282301845818, + "grad_norm": 10.96146456022653, + "learning_rate": 2.049281813005215e-06, + "loss": 0.4389, + "step": 7357 + }, + { + "epoch": 3.99457111834962, + "grad_norm": 12.017836415062291, + "learning_rate": 2.047149287829726e-06, + "loss": 0.5839, + "step": 7358 + }, + { + "epoch": 3.995114006514658, + "grad_norm": 10.822043166102832, + "learning_rate": 2.045017746275485e-06, + "loss": 0.4467, + "step": 7359 + }, + { + "epoch": 3.995656894679696, + "grad_norm": 13.151255799546034, + "learning_rate": 2.0428871886061343e-06, + "loss": 0.6173, + "step": 7360 + }, + { + "epoch": 3.9961997828447338, + "grad_norm": 11.118960416841484, + "learning_rate": 2.0407576150851705e-06, + "loss": 0.3935, + "step": 7361 + }, + { + "epoch": 3.996742671009772, + "grad_norm": 12.928271003217045, + "learning_rate": 2.0386290259759967e-06, + "loss": 0.5222, + "step": 7362 + }, + { + "epoch": 3.99728555917481, + "grad_norm": 14.237952289938509, + "learning_rate": 2.0365014215418676e-06, + "loss": 0.495, + "step": 7363 + }, + { + "epoch": 3.997828447339848, + "grad_norm": 15.550337043796231, + "learning_rate": 2.0343748020459374e-06, + "loss": 0.4867, + "step": 7364 + }, + { + "epoch": 3.9983713355048858, + "grad_norm": 12.364600226217249, + "learning_rate": 2.032249167751228e-06, + "loss": 0.6296, + "step": 7365 + }, + { + "epoch": 3.998914223669924, + "grad_norm": 13.569416929145026, + "learning_rate": 2.0301245189206385e-06, + "loss": 0.5857, + "step": 7366 + }, + { + "epoch": 3.999457111834962, + "grad_norm": 14.915873617390455, + "learning_rate": 2.028000855816954e-06, + "loss": 0.7152, + "step": 7367 + }, + { + "epoch": 4.0, + "grad_norm": 13.686794191538802, + "learning_rate": 2.025878178702825e-06, + "loss": 0.3451, + "step": 7368 + } + ], + "logging_steps": 1.0, + "max_steps": 9210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8791217965891584.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/training_args.bin b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b16de63f342ea53d59c4d97aa8a3bd463a79531f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9801fdd09c6ba798e12b77f77221db3b5747b6e1296785789d44cf2bc492ff9e +size 7160 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/zero_to_fp32.py b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-7368/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/README.md b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/README.md new file mode 100644 index 0000000000000000000000000000000000000000..972a036203d877262d3c6673f4d81814e7409dc5 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/README.md @@ -0,0 +1,202 @@ +--- +base_model: llava-hf/LLaVA-NeXT-Video-7B-hf +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/adapter_config.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..719f6e47b29561ff93d519680ccd115b480d825e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/adapter_config.json @@ -0,0 +1,132 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "llava-hf/LLaVA-NeXT-Video-7B-hf", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 8, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [], + "peft_type": "LORA", + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "model.layers.17.self_attn.q_proj", + "model.layers.17.self_attn.k_proj", + "model.layers.13.self_attn.k_proj", + "model.layers.20.self_attn.q_proj", + "model.layers.18.self_attn.k_proj", + "29.self_attn.k_proj", + "model.layers.16.self_attn.v_proj", + "model.layers.14.self_attn.q_proj", + "29.self_attn.v_proj", + "24.self_attn.k_proj", + "model.layers.18.self_attn.v_proj", + "model.layers.5.self_attn.q_proj", + "model.layers.11.self_attn.k_proj", + "27.self_attn.v_proj", + "model.layers.16.self_attn.q_proj", + "model.layers.10.self_attn.q_proj", + "model.layers.12.self_attn.v_proj", + "model.layers.20.self_attn.v_proj", + "model.layers.0.self_attn.v_proj", + "model.layers.7.self_attn.q_proj", + "model.layers.23.self_attn.q_proj", + "model.layers.1.self_attn.q_proj", + "25.self_attn.v_proj", + "model.layers.4.self_attn.v_proj", + "model.layers.19.self_attn.q_proj", + "model.layers.10.self_attn.k_proj", + "model.layers.22.self_attn.k_proj", + "model.layers.14.self_attn.v_proj", + "model.layers.13.self_attn.v_proj", + "25.self_attn.k_proj", + "24.self_attn.v_proj", + "model.layers.21.self_attn.q_proj", + "model.layers.8.self_attn.q_proj", + "model.layers.6.self_attn.q_proj", + "model.layers.6.self_attn.k_proj", + "model.layers.1.self_attn.v_proj", + "model.layers.9.self_attn.k_proj", + "model.layers.13.self_attn.q_proj", + "o_proj", + "model.layers.16.self_attn.k_proj", + "model.layers.6.self_attn.v_proj", + "model.layers.21.self_attn.k_proj", + "model.layers.5.self_attn.k_proj", + "27.self_attn.q_proj", + "model.layers.20.self_attn.k_proj", + "model.layers.11.self_attn.v_proj", + "model.layers.22.self_attn.q_proj", + "model.layers.14.self_attn.k_proj", + "28.self_attn.v_proj", + "model.layers.22.self_attn.v_proj", + "gate_proj", + "model.layers.0.self_attn.k_proj", + "model.layers.5.self_attn.v_proj", + "30.self_attn.q_proj", + "model.layers.3.self_attn.k_proj", + "29.self_attn.q_proj", + "model.layers.19.self_attn.k_proj", + "model.layers.10.self_attn.v_proj", + "model.layers.15.self_attn.k_proj", + "model.layers.7.self_attn.v_proj", + "model.layers.9.self_attn.v_proj", + "up_proj", + "model.layers.4.self_attn.q_proj", + "28.self_attn.q_proj", + "model.layers.3.self_attn.q_proj", + "24.self_attn.q_proj", + "30.self_attn.v_proj", + "model.layers.12.self_attn.k_proj", + "model.layers.23.self_attn.k_proj", + "26.self_attn.q_proj", + "28.self_attn.k_proj", + "31.self_attn.k_proj", + "down_proj", + "model.layers.2.self_attn.q_proj", + "31.self_attn.q_proj", + "model.layers.15.self_attn.q_proj", + "model.layers.12.self_attn.q_proj", + "model.layers.0.self_attn.q_proj", + "model.layers.2.self_attn.k_proj", + "model.layers.3.self_attn.v_proj", + "model.layers.11.self_attn.q_proj", + "25.self_attn.q_proj", + "model.layers.23.self_attn.v_proj", + "30.self_attn.k_proj", + "31.self_attn.v_proj", + "model.layers.15.self_attn.v_proj", + "model.layers.8.self_attn.v_proj", + "26.self_attn.k_proj", + "model.layers.19.self_attn.v_proj", + "model.layers.2.self_attn.v_proj", + "26.self_attn.v_proj", + "model.layers.21.self_attn.v_proj", + "model.layers.17.self_attn.v_proj", + "model.layers.8.self_attn.k_proj", + "model.layers.18.self_attn.q_proj", + "model.layers.7.self_attn.k_proj", + "model.layers.4.self_attn.k_proj", + "model.layers.9.self_attn.q_proj", + "27.self_attn.k_proj", + "model.layers.1.self_attn.k_proj" + ], + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/adapter_model.safetensors b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fca48c618e79d6c9d59165c4271e1468a35bf25e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcffdf981d90fd323630fa03bc9b94cc6c0d2e47a9bf7387c891fbcc8c4b54bc +size 40043208 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..2860b450ea264a7a3583d735c88d3cafcca6268c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b2a9e99f2032f27b09b8cd89e2f8d2d03e33d3bbaf2b9d435121c402fbbc85f +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..1369c2cd43d0e3cfb76c3eb122dbed117331872c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64ba30a2ab27aac6682f8473c27e3507a6eb1c16c933ad49ee0700216c172d0b +size 119934704 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/zero_pp_rank_0_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/zero_pp_rank_0_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..9e9ebae86b9cb560ba8f54dc8f2b4fad1acd149e --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/zero_pp_rank_0_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c10599120608faed98e06ee4ec5c477511e94b87060926f90b049355a5d6bfb +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/zero_pp_rank_1_mp_rank_00_model_states.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/zero_pp_rank_1_mp_rank_00_model_states.pt new file mode 100644 index 0000000000000000000000000000000000000000..461e4cf8143def469afe38b30a6dd3a1bf3d88c2 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/global_step9210/zero_pp_rank_1_mp_rank_00_model_states.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e27b36e51ca2afa7ed52cd59167b231b8be65b6017cc152c64d2b30eb00fcb6e +size 528366 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/latest b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/latest new file mode 100644 index 0000000000000000000000000000000000000000..be490a80a1364f8c3d783c86581ee547159789ca --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/latest @@ -0,0 +1 @@ +global_step9210 \ No newline at end of file diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/rng_state_0.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..333f9cc825e638c8f5e021edaa46214b60c2053a --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a8a277cdef7730ffbb6334c57d822c714c2d6734581db9ebbcb1a40685c0456 +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/rng_state_1.pth b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..514c033d3217c1b5ade80431a47733762044437c --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36957c3886252d05e1c4d6bc36fb39538dc1d914eb9a8d5d93f7fcb581ae0d4c +size 14512 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/scheduler.pt b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..052b34aa8a6d96f058f2737a3f189bf81a4838f6 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a88f6d85aca92323415307a7db4b215c53496d9d217284c5951e1388e3da658 +size 1064 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/trainer_state.json b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..e89ec1f05c23daa565313521a14dbc694bc067ca --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/trainer_state.json @@ -0,0 +1,64503 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 9210, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005428881650380022, + "grad_norm": 2.1320506281889067, + "learning_rate": 7.220216606498195e-08, + "loss": 5.1936, + "step": 1 + }, + { + "epoch": 0.0010857763300760044, + "grad_norm": 1.8633488766761366, + "learning_rate": 1.444043321299639e-07, + "loss": 4.6995, + "step": 2 + }, + { + "epoch": 0.0016286644951140066, + "grad_norm": 2.063286098879776, + "learning_rate": 2.1660649819494586e-07, + "loss": 4.4902, + "step": 3 + }, + { + "epoch": 0.002171552660152009, + "grad_norm": 1.7123050230907495, + "learning_rate": 2.888086642599278e-07, + "loss": 4.9125, + "step": 4 + }, + { + "epoch": 0.0027144408251900108, + "grad_norm": 1.7499657276799891, + "learning_rate": 3.610108303249098e-07, + "loss": 3.7648, + "step": 5 + }, + { + "epoch": 0.003257328990228013, + "grad_norm": 2.0419045636066886, + "learning_rate": 4.332129963898917e-07, + "loss": 4.7842, + "step": 6 + }, + { + "epoch": 0.003800217155266015, + "grad_norm": 1.4960282529105982, + "learning_rate": 5.054151624548736e-07, + "loss": 4.9333, + "step": 7 + }, + { + "epoch": 0.004343105320304018, + "grad_norm": 1.640954967476452, + "learning_rate": 5.776173285198556e-07, + "loss": 3.9576, + "step": 8 + }, + { + "epoch": 0.004885993485342019, + "grad_norm": 2.0590500459610244, + "learning_rate": 6.498194945848375e-07, + "loss": 4.3833, + "step": 9 + }, + { + "epoch": 0.0054288816503800215, + "grad_norm": 1.07709353257816, + "learning_rate": 7.220216606498196e-07, + "loss": 3.7669, + "step": 10 + }, + { + "epoch": 0.005971769815418024, + "grad_norm": 1.9506235655782926, + "learning_rate": 7.942238267148016e-07, + "loss": 4.5975, + "step": 11 + }, + { + "epoch": 0.006514657980456026, + "grad_norm": 1.411983096954528, + "learning_rate": 8.664259927797834e-07, + "loss": 3.9706, + "step": 12 + }, + { + "epoch": 0.007057546145494028, + "grad_norm": 1.7110053131090028, + "learning_rate": 9.386281588447654e-07, + "loss": 4.9081, + "step": 13 + }, + { + "epoch": 0.00760043431053203, + "grad_norm": 1.3689315757199692, + "learning_rate": 1.0108303249097473e-06, + "loss": 4.2426, + "step": 14 + }, + { + "epoch": 0.008143322475570033, + "grad_norm": 1.5612091671030481, + "learning_rate": 1.0830324909747293e-06, + "loss": 4.5144, + "step": 15 + }, + { + "epoch": 0.008686210640608035, + "grad_norm": 1.3725168953865754, + "learning_rate": 1.1552346570397112e-06, + "loss": 4.9378, + "step": 16 + }, + { + "epoch": 0.009229098805646038, + "grad_norm": 2.027359913808926, + "learning_rate": 1.2274368231046932e-06, + "loss": 3.7477, + "step": 17 + }, + { + "epoch": 0.009771986970684038, + "grad_norm": 2.1695456962160224, + "learning_rate": 1.299638989169675e-06, + "loss": 5.7636, + "step": 18 + }, + { + "epoch": 0.01031487513572204, + "grad_norm": 3.2637002287680428, + "learning_rate": 1.3718411552346572e-06, + "loss": 6.1883, + "step": 19 + }, + { + "epoch": 0.010857763300760043, + "grad_norm": 2.5723616120643857, + "learning_rate": 1.4440433212996392e-06, + "loss": 4.0267, + "step": 20 + }, + { + "epoch": 0.011400651465798045, + "grad_norm": 1.4473754577877487, + "learning_rate": 1.516245487364621e-06, + "loss": 4.526, + "step": 21 + }, + { + "epoch": 0.011943539630836048, + "grad_norm": 1.6098208731761, + "learning_rate": 1.5884476534296031e-06, + "loss": 4.5163, + "step": 22 + }, + { + "epoch": 0.01248642779587405, + "grad_norm": 2.899624682573202, + "learning_rate": 1.6606498194945849e-06, + "loss": 4.8188, + "step": 23 + }, + { + "epoch": 0.013029315960912053, + "grad_norm": 2.273571202553131, + "learning_rate": 1.7328519855595669e-06, + "loss": 5.0074, + "step": 24 + }, + { + "epoch": 0.013572204125950055, + "grad_norm": 1.1877323869171315, + "learning_rate": 1.805054151624549e-06, + "loss": 4.4331, + "step": 25 + }, + { + "epoch": 0.014115092290988056, + "grad_norm": 1.9365501019385494, + "learning_rate": 1.8772563176895308e-06, + "loss": 5.0525, + "step": 26 + }, + { + "epoch": 0.014657980456026058, + "grad_norm": 1.7710485953105986, + "learning_rate": 1.949458483754513e-06, + "loss": 4.1346, + "step": 27 + }, + { + "epoch": 0.01520086862106406, + "grad_norm": 1.231848881292397, + "learning_rate": 2.0216606498194946e-06, + "loss": 4.2431, + "step": 28 + }, + { + "epoch": 0.015743756786102063, + "grad_norm": 1.6372838462475754, + "learning_rate": 2.0938628158844768e-06, + "loss": 5.4317, + "step": 29 + }, + { + "epoch": 0.016286644951140065, + "grad_norm": 1.7326568199523718, + "learning_rate": 2.1660649819494585e-06, + "loss": 4.4774, + "step": 30 + }, + { + "epoch": 0.016829533116178068, + "grad_norm": 2.1315641946923387, + "learning_rate": 2.2382671480144407e-06, + "loss": 4.1418, + "step": 31 + }, + { + "epoch": 0.01737242128121607, + "grad_norm": 1.916538366799816, + "learning_rate": 2.3104693140794225e-06, + "loss": 5.7926, + "step": 32 + }, + { + "epoch": 0.017915309446254073, + "grad_norm": 2.057531274363787, + "learning_rate": 2.3826714801444047e-06, + "loss": 4.2862, + "step": 33 + }, + { + "epoch": 0.018458197611292075, + "grad_norm": 2.6727877745045023, + "learning_rate": 2.4548736462093864e-06, + "loss": 5.1323, + "step": 34 + }, + { + "epoch": 0.019001085776330078, + "grad_norm": 1.5043525736330898, + "learning_rate": 2.527075812274368e-06, + "loss": 4.0634, + "step": 35 + }, + { + "epoch": 0.019543973941368076, + "grad_norm": 2.6324248758076862, + "learning_rate": 2.59927797833935e-06, + "loss": 5.4211, + "step": 36 + }, + { + "epoch": 0.02008686210640608, + "grad_norm": 1.6581905646911415, + "learning_rate": 2.6714801444043326e-06, + "loss": 4.9825, + "step": 37 + }, + { + "epoch": 0.02062975027144408, + "grad_norm": 1.6667970880898455, + "learning_rate": 2.7436823104693144e-06, + "loss": 4.3219, + "step": 38 + }, + { + "epoch": 0.021172638436482084, + "grad_norm": 1.7448150724822071, + "learning_rate": 2.815884476534296e-06, + "loss": 4.3912, + "step": 39 + }, + { + "epoch": 0.021715526601520086, + "grad_norm": 3.720503777726443, + "learning_rate": 2.8880866425992783e-06, + "loss": 6.0592, + "step": 40 + }, + { + "epoch": 0.02225841476655809, + "grad_norm": 1.6853001074010774, + "learning_rate": 2.96028880866426e-06, + "loss": 3.8395, + "step": 41 + }, + { + "epoch": 0.02280130293159609, + "grad_norm": 1.797029067943667, + "learning_rate": 3.032490974729242e-06, + "loss": 4.4651, + "step": 42 + }, + { + "epoch": 0.023344191096634093, + "grad_norm": 1.6298824221707349, + "learning_rate": 3.1046931407942245e-06, + "loss": 4.4507, + "step": 43 + }, + { + "epoch": 0.023887079261672096, + "grad_norm": 1.4793189209059368, + "learning_rate": 3.1768953068592062e-06, + "loss": 4.7849, + "step": 44 + }, + { + "epoch": 0.024429967426710098, + "grad_norm": 1.1741346570472608, + "learning_rate": 3.249097472924188e-06, + "loss": 3.5897, + "step": 45 + }, + { + "epoch": 0.0249728555917481, + "grad_norm": 2.1468867306169064, + "learning_rate": 3.3212996389891698e-06, + "loss": 3.8969, + "step": 46 + }, + { + "epoch": 0.025515743756786103, + "grad_norm": 1.657494000697021, + "learning_rate": 3.393501805054152e-06, + "loss": 4.3412, + "step": 47 + }, + { + "epoch": 0.026058631921824105, + "grad_norm": 1.7125864055710684, + "learning_rate": 3.4657039711191337e-06, + "loss": 3.8059, + "step": 48 + }, + { + "epoch": 0.026601520086862108, + "grad_norm": 2.9515347000925622, + "learning_rate": 3.5379061371841155e-06, + "loss": 4.1069, + "step": 49 + }, + { + "epoch": 0.02714440825190011, + "grad_norm": 3.0185188918098262, + "learning_rate": 3.610108303249098e-06, + "loss": 5.2646, + "step": 50 + }, + { + "epoch": 0.02768729641693811, + "grad_norm": 1.5546972252207274, + "learning_rate": 3.68231046931408e-06, + "loss": 3.7212, + "step": 51 + }, + { + "epoch": 0.02823018458197611, + "grad_norm": 2.6376587426282474, + "learning_rate": 3.7545126353790616e-06, + "loss": 4.2862, + "step": 52 + }, + { + "epoch": 0.028773072747014114, + "grad_norm": 2.053905428064986, + "learning_rate": 3.826714801444043e-06, + "loss": 3.8304, + "step": 53 + }, + { + "epoch": 0.029315960912052116, + "grad_norm": 2.9083183656849725, + "learning_rate": 3.898916967509026e-06, + "loss": 4.8223, + "step": 54 + }, + { + "epoch": 0.02985884907709012, + "grad_norm": 1.9919239162425852, + "learning_rate": 3.971119133574007e-06, + "loss": 3.9054, + "step": 55 + }, + { + "epoch": 0.03040173724212812, + "grad_norm": 2.4076122411223135, + "learning_rate": 4.043321299638989e-06, + "loss": 4.8568, + "step": 56 + }, + { + "epoch": 0.030944625407166124, + "grad_norm": 3.854434602830178, + "learning_rate": 4.115523465703971e-06, + "loss": 5.7771, + "step": 57 + }, + { + "epoch": 0.031487513572204126, + "grad_norm": 2.510303564095041, + "learning_rate": 4.1877256317689535e-06, + "loss": 5.5851, + "step": 58 + }, + { + "epoch": 0.03203040173724213, + "grad_norm": 2.8884448366576305, + "learning_rate": 4.259927797833936e-06, + "loss": 4.1774, + "step": 59 + }, + { + "epoch": 0.03257328990228013, + "grad_norm": 2.0893778739721, + "learning_rate": 4.332129963898917e-06, + "loss": 4.4679, + "step": 60 + }, + { + "epoch": 0.03311617806731813, + "grad_norm": 1.5816439039579229, + "learning_rate": 4.404332129963899e-06, + "loss": 4.1999, + "step": 61 + }, + { + "epoch": 0.033659066232356136, + "grad_norm": 1.9591574137365761, + "learning_rate": 4.4765342960288814e-06, + "loss": 4.1539, + "step": 62 + }, + { + "epoch": 0.03420195439739414, + "grad_norm": 3.1533909413995924, + "learning_rate": 4.548736462093864e-06, + "loss": 4.5842, + "step": 63 + }, + { + "epoch": 0.03474484256243214, + "grad_norm": 2.543715437360663, + "learning_rate": 4.620938628158845e-06, + "loss": 4.1473, + "step": 64 + }, + { + "epoch": 0.03528773072747014, + "grad_norm": 2.7608093188531955, + "learning_rate": 4.693140794223827e-06, + "loss": 4.9354, + "step": 65 + }, + { + "epoch": 0.035830618892508145, + "grad_norm": 4.192163162864289, + "learning_rate": 4.765342960288809e-06, + "loss": 5.4263, + "step": 66 + }, + { + "epoch": 0.03637350705754615, + "grad_norm": 2.809206875027472, + "learning_rate": 4.837545126353791e-06, + "loss": 4.4259, + "step": 67 + }, + { + "epoch": 0.03691639522258415, + "grad_norm": 3.9318073942863534, + "learning_rate": 4.909747292418773e-06, + "loss": 4.1483, + "step": 68 + }, + { + "epoch": 0.03745928338762215, + "grad_norm": 2.935065385226121, + "learning_rate": 4.981949458483755e-06, + "loss": 4.6854, + "step": 69 + }, + { + "epoch": 0.038002171552660155, + "grad_norm": 2.3687140876301305, + "learning_rate": 5.054151624548736e-06, + "loss": 4.5875, + "step": 70 + }, + { + "epoch": 0.03854505971769816, + "grad_norm": 2.987709377109457, + "learning_rate": 5.126353790613719e-06, + "loss": 5.5784, + "step": 71 + }, + { + "epoch": 0.03908794788273615, + "grad_norm": 2.618514747422557, + "learning_rate": 5.1985559566787e-06, + "loss": 3.9901, + "step": 72 + }, + { + "epoch": 0.039630836047774155, + "grad_norm": 3.729239515680902, + "learning_rate": 5.270758122743683e-06, + "loss": 4.6192, + "step": 73 + }, + { + "epoch": 0.04017372421281216, + "grad_norm": 2.488658936956188, + "learning_rate": 5.342960288808665e-06, + "loss": 3.9889, + "step": 74 + }, + { + "epoch": 0.04071661237785016, + "grad_norm": 2.6055640979236223, + "learning_rate": 5.415162454873647e-06, + "loss": 4.3166, + "step": 75 + }, + { + "epoch": 0.04125950054288816, + "grad_norm": 2.091905540809132, + "learning_rate": 5.487364620938629e-06, + "loss": 3.5848, + "step": 76 + }, + { + "epoch": 0.041802388707926165, + "grad_norm": 4.153140315460812, + "learning_rate": 5.559566787003611e-06, + "loss": 4.4889, + "step": 77 + }, + { + "epoch": 0.04234527687296417, + "grad_norm": 4.828033123535034, + "learning_rate": 5.631768953068592e-06, + "loss": 4.886, + "step": 78 + }, + { + "epoch": 0.04288816503800217, + "grad_norm": 3.106146715385438, + "learning_rate": 5.7039711191335744e-06, + "loss": 4.3698, + "step": 79 + }, + { + "epoch": 0.04343105320304017, + "grad_norm": 4.458698539182903, + "learning_rate": 5.776173285198557e-06, + "loss": 4.7018, + "step": 80 + }, + { + "epoch": 0.043973941368078175, + "grad_norm": 4.619902873416999, + "learning_rate": 5.848375451263538e-06, + "loss": 4.3449, + "step": 81 + }, + { + "epoch": 0.04451682953311618, + "grad_norm": 2.411841615517987, + "learning_rate": 5.92057761732852e-06, + "loss": 3.9952, + "step": 82 + }, + { + "epoch": 0.04505971769815418, + "grad_norm": 3.6594837742651705, + "learning_rate": 5.992779783393502e-06, + "loss": 4.1308, + "step": 83 + }, + { + "epoch": 0.04560260586319218, + "grad_norm": 2.3760888273966074, + "learning_rate": 6.064981949458484e-06, + "loss": 4.0315, + "step": 84 + }, + { + "epoch": 0.046145494028230184, + "grad_norm": 4.579655616004927, + "learning_rate": 6.137184115523466e-06, + "loss": 4.1051, + "step": 85 + }, + { + "epoch": 0.04668838219326819, + "grad_norm": 3.5909339610827846, + "learning_rate": 6.209386281588449e-06, + "loss": 4.3126, + "step": 86 + }, + { + "epoch": 0.04723127035830619, + "grad_norm": 3.1030422711991625, + "learning_rate": 6.28158844765343e-06, + "loss": 4.3038, + "step": 87 + }, + { + "epoch": 0.04777415852334419, + "grad_norm": 4.134211324790428, + "learning_rate": 6.3537906137184125e-06, + "loss": 4.906, + "step": 88 + }, + { + "epoch": 0.048317046688382194, + "grad_norm": 3.538946913288816, + "learning_rate": 6.425992779783395e-06, + "loss": 4.2182, + "step": 89 + }, + { + "epoch": 0.048859934853420196, + "grad_norm": 1.9828152543749473, + "learning_rate": 6.498194945848376e-06, + "loss": 3.8487, + "step": 90 + }, + { + "epoch": 0.0494028230184582, + "grad_norm": 2.127969936938475, + "learning_rate": 6.570397111913358e-06, + "loss": 3.9221, + "step": 91 + }, + { + "epoch": 0.0499457111834962, + "grad_norm": 2.740653026382279, + "learning_rate": 6.6425992779783395e-06, + "loss": 4.4661, + "step": 92 + }, + { + "epoch": 0.050488599348534204, + "grad_norm": 3.278496788216742, + "learning_rate": 6.714801444043322e-06, + "loss": 4.1995, + "step": 93 + }, + { + "epoch": 0.051031487513572206, + "grad_norm": 1.5455749978381657, + "learning_rate": 6.787003610108304e-06, + "loss": 3.3941, + "step": 94 + }, + { + "epoch": 0.05157437567861021, + "grad_norm": 2.6938438745410975, + "learning_rate": 6.859205776173285e-06, + "loss": 3.9863, + "step": 95 + }, + { + "epoch": 0.05211726384364821, + "grad_norm": 2.754193515919841, + "learning_rate": 6.9314079422382674e-06, + "loss": 4.1621, + "step": 96 + }, + { + "epoch": 0.05266015200868621, + "grad_norm": 1.787757508585929, + "learning_rate": 7.00361010830325e-06, + "loss": 3.728, + "step": 97 + }, + { + "epoch": 0.053203040173724216, + "grad_norm": 1.6756343581979376, + "learning_rate": 7.075812274368231e-06, + "loss": 3.4483, + "step": 98 + }, + { + "epoch": 0.05374592833876222, + "grad_norm": 3.0418610323064317, + "learning_rate": 7.148014440433214e-06, + "loss": 4.1131, + "step": 99 + }, + { + "epoch": 0.05428881650380022, + "grad_norm": 2.710846119482301, + "learning_rate": 7.220216606498196e-06, + "loss": 5.0449, + "step": 100 + }, + { + "epoch": 0.054831704668838216, + "grad_norm": 1.906382949957388, + "learning_rate": 7.2924187725631776e-06, + "loss": 3.3181, + "step": 101 + }, + { + "epoch": 0.05537459283387622, + "grad_norm": 2.4669878697837, + "learning_rate": 7.36462093862816e-06, + "loss": 3.7996, + "step": 102 + }, + { + "epoch": 0.05591748099891422, + "grad_norm": 2.3755842712374586, + "learning_rate": 7.436823104693142e-06, + "loss": 4.0321, + "step": 103 + }, + { + "epoch": 0.05646036916395222, + "grad_norm": 2.032793864149989, + "learning_rate": 7.509025270758123e-06, + "loss": 3.4121, + "step": 104 + }, + { + "epoch": 0.057003257328990226, + "grad_norm": 2.1561554352810686, + "learning_rate": 7.5812274368231055e-06, + "loss": 4.2585, + "step": 105 + }, + { + "epoch": 0.05754614549402823, + "grad_norm": 1.9627108654626784, + "learning_rate": 7.653429602888087e-06, + "loss": 4.3085, + "step": 106 + }, + { + "epoch": 0.05808903365906623, + "grad_norm": 1.8622753095278954, + "learning_rate": 7.72563176895307e-06, + "loss": 3.8459, + "step": 107 + }, + { + "epoch": 0.05863192182410423, + "grad_norm": 1.4438765470100654, + "learning_rate": 7.797833935018051e-06, + "loss": 3.4914, + "step": 108 + }, + { + "epoch": 0.059174809989142235, + "grad_norm": 1.6164085905580508, + "learning_rate": 7.870036101083033e-06, + "loss": 3.4958, + "step": 109 + }, + { + "epoch": 0.05971769815418024, + "grad_norm": 1.6075014519578008, + "learning_rate": 7.942238267148014e-06, + "loss": 3.202, + "step": 110 + }, + { + "epoch": 0.06026058631921824, + "grad_norm": 1.7630994896021268, + "learning_rate": 8.014440433212997e-06, + "loss": 3.0916, + "step": 111 + }, + { + "epoch": 0.06080347448425624, + "grad_norm": 1.7600666006428431, + "learning_rate": 8.086642599277978e-06, + "loss": 3.5791, + "step": 112 + }, + { + "epoch": 0.061346362649294245, + "grad_norm": 2.324243905399748, + "learning_rate": 8.158844765342961e-06, + "loss": 3.6247, + "step": 113 + }, + { + "epoch": 0.06188925081433225, + "grad_norm": 1.2715449173786617, + "learning_rate": 8.231046931407943e-06, + "loss": 3.2381, + "step": 114 + }, + { + "epoch": 0.06243213897937025, + "grad_norm": 1.9029014476619814, + "learning_rate": 8.303249097472926e-06, + "loss": 3.5269, + "step": 115 + }, + { + "epoch": 0.06297502714440825, + "grad_norm": 1.8782574049467182, + "learning_rate": 8.375451263537907e-06, + "loss": 2.8698, + "step": 116 + }, + { + "epoch": 0.06351791530944625, + "grad_norm": 1.8522383705853351, + "learning_rate": 8.447653429602888e-06, + "loss": 3.5003, + "step": 117 + }, + { + "epoch": 0.06406080347448426, + "grad_norm": 1.6480852413427443, + "learning_rate": 8.519855595667871e-06, + "loss": 3.5665, + "step": 118 + }, + { + "epoch": 0.06460369163952226, + "grad_norm": 2.1628448626122685, + "learning_rate": 8.592057761732853e-06, + "loss": 3.343, + "step": 119 + }, + { + "epoch": 0.06514657980456026, + "grad_norm": 1.568456113423694, + "learning_rate": 8.664259927797834e-06, + "loss": 2.474, + "step": 120 + }, + { + "epoch": 0.06568946796959826, + "grad_norm": 1.8054357065473743, + "learning_rate": 8.736462093862817e-06, + "loss": 2.9228, + "step": 121 + }, + { + "epoch": 0.06623235613463627, + "grad_norm": 1.9418450135728782, + "learning_rate": 8.808664259927798e-06, + "loss": 3.1995, + "step": 122 + }, + { + "epoch": 0.06677524429967427, + "grad_norm": 1.321851521267606, + "learning_rate": 8.88086642599278e-06, + "loss": 3.5515, + "step": 123 + }, + { + "epoch": 0.06731813246471227, + "grad_norm": 1.390198419578882, + "learning_rate": 8.953068592057763e-06, + "loss": 2.7741, + "step": 124 + }, + { + "epoch": 0.06786102062975027, + "grad_norm": 2.119370139523294, + "learning_rate": 9.025270758122744e-06, + "loss": 3.3427, + "step": 125 + }, + { + "epoch": 0.06840390879478828, + "grad_norm": 2.042189734142807, + "learning_rate": 9.097472924187727e-06, + "loss": 3.1414, + "step": 126 + }, + { + "epoch": 0.06894679695982628, + "grad_norm": 2.0399787373405123, + "learning_rate": 9.169675090252709e-06, + "loss": 3.0862, + "step": 127 + }, + { + "epoch": 0.06948968512486428, + "grad_norm": 2.0523727295223915, + "learning_rate": 9.24187725631769e-06, + "loss": 2.6461, + "step": 128 + }, + { + "epoch": 0.07003257328990228, + "grad_norm": 1.459190908283032, + "learning_rate": 9.314079422382673e-06, + "loss": 2.9907, + "step": 129 + }, + { + "epoch": 0.07057546145494029, + "grad_norm": 1.7403521896723462, + "learning_rate": 9.386281588447654e-06, + "loss": 3.0681, + "step": 130 + }, + { + "epoch": 0.07111834961997829, + "grad_norm": 2.2283604183714383, + "learning_rate": 9.458483754512636e-06, + "loss": 3.4619, + "step": 131 + }, + { + "epoch": 0.07166123778501629, + "grad_norm": 1.3291461832292721, + "learning_rate": 9.530685920577619e-06, + "loss": 3.096, + "step": 132 + }, + { + "epoch": 0.0722041259500543, + "grad_norm": 1.7839357457835638, + "learning_rate": 9.6028880866426e-06, + "loss": 2.7512, + "step": 133 + }, + { + "epoch": 0.0727470141150923, + "grad_norm": 1.7836465130823984, + "learning_rate": 9.675090252707581e-06, + "loss": 3.3118, + "step": 134 + }, + { + "epoch": 0.0732899022801303, + "grad_norm": 2.0407172747925135, + "learning_rate": 9.747292418772564e-06, + "loss": 2.1032, + "step": 135 + }, + { + "epoch": 0.0738327904451683, + "grad_norm": 2.335226908691354, + "learning_rate": 9.819494584837546e-06, + "loss": 2.9039, + "step": 136 + }, + { + "epoch": 0.0743756786102063, + "grad_norm": 2.567081845339778, + "learning_rate": 9.891696750902527e-06, + "loss": 3.1071, + "step": 137 + }, + { + "epoch": 0.0749185667752443, + "grad_norm": 1.9549184920601244, + "learning_rate": 9.96389891696751e-06, + "loss": 2.8253, + "step": 138 + }, + { + "epoch": 0.07546145494028231, + "grad_norm": 1.6389815727698025, + "learning_rate": 1.0036101083032491e-05, + "loss": 2.5701, + "step": 139 + }, + { + "epoch": 0.07600434310532031, + "grad_norm": 1.6558924195158018, + "learning_rate": 1.0108303249097473e-05, + "loss": 2.6428, + "step": 140 + }, + { + "epoch": 0.07654723127035831, + "grad_norm": 2.0257343427634087, + "learning_rate": 1.0180505415162456e-05, + "loss": 2.2828, + "step": 141 + }, + { + "epoch": 0.07709011943539631, + "grad_norm": 1.879991343961764, + "learning_rate": 1.0252707581227437e-05, + "loss": 3.1436, + "step": 142 + }, + { + "epoch": 0.07763300760043432, + "grad_norm": 1.8718236498893788, + "learning_rate": 1.032490974729242e-05, + "loss": 3.2434, + "step": 143 + }, + { + "epoch": 0.0781758957654723, + "grad_norm": 1.8252190287923398, + "learning_rate": 1.03971119133574e-05, + "loss": 2.4851, + "step": 144 + }, + { + "epoch": 0.07871878393051031, + "grad_norm": 2.6564413093838213, + "learning_rate": 1.0469314079422383e-05, + "loss": 2.7315, + "step": 145 + }, + { + "epoch": 0.07926167209554831, + "grad_norm": 2.054556112475997, + "learning_rate": 1.0541516245487366e-05, + "loss": 2.647, + "step": 146 + }, + { + "epoch": 0.07980456026058631, + "grad_norm": 1.072343898976271, + "learning_rate": 1.0613718411552347e-05, + "loss": 2.5868, + "step": 147 + }, + { + "epoch": 0.08034744842562432, + "grad_norm": 1.1443585178571194, + "learning_rate": 1.068592057761733e-05, + "loss": 2.3247, + "step": 148 + }, + { + "epoch": 0.08089033659066232, + "grad_norm": 2.0424670761496646, + "learning_rate": 1.0758122743682312e-05, + "loss": 1.989, + "step": 149 + }, + { + "epoch": 0.08143322475570032, + "grad_norm": 1.3715615162918033, + "learning_rate": 1.0830324909747295e-05, + "loss": 2.3989, + "step": 150 + }, + { + "epoch": 0.08197611292073832, + "grad_norm": 1.3740116291532023, + "learning_rate": 1.0902527075812274e-05, + "loss": 1.8726, + "step": 151 + }, + { + "epoch": 0.08251900108577633, + "grad_norm": 2.5824759155435655, + "learning_rate": 1.0974729241877257e-05, + "loss": 2.3019, + "step": 152 + }, + { + "epoch": 0.08306188925081433, + "grad_norm": 1.5584514490971877, + "learning_rate": 1.1046931407942239e-05, + "loss": 2.4209, + "step": 153 + }, + { + "epoch": 0.08360477741585233, + "grad_norm": 1.6184456176727229, + "learning_rate": 1.1119133574007222e-05, + "loss": 2.2377, + "step": 154 + }, + { + "epoch": 0.08414766558089033, + "grad_norm": 2.6665151568462586, + "learning_rate": 1.1191335740072201e-05, + "loss": 2.6381, + "step": 155 + }, + { + "epoch": 0.08469055374592833, + "grad_norm": 1.442929928712006, + "learning_rate": 1.1263537906137184e-05, + "loss": 2.3405, + "step": 156 + }, + { + "epoch": 0.08523344191096634, + "grad_norm": 1.906547769868643, + "learning_rate": 1.1335740072202166e-05, + "loss": 2.4391, + "step": 157 + }, + { + "epoch": 0.08577633007600434, + "grad_norm": 1.2153727696829317, + "learning_rate": 1.1407942238267149e-05, + "loss": 1.8401, + "step": 158 + }, + { + "epoch": 0.08631921824104234, + "grad_norm": 1.4810729195972283, + "learning_rate": 1.1480144404332132e-05, + "loss": 2.7248, + "step": 159 + }, + { + "epoch": 0.08686210640608034, + "grad_norm": 2.9401118999625058, + "learning_rate": 1.1552346570397113e-05, + "loss": 2.6582, + "step": 160 + }, + { + "epoch": 0.08740499457111835, + "grad_norm": 1.145106458573816, + "learning_rate": 1.1624548736462096e-05, + "loss": 2.6648, + "step": 161 + }, + { + "epoch": 0.08794788273615635, + "grad_norm": 1.9228128025198137, + "learning_rate": 1.1696750902527076e-05, + "loss": 2.3702, + "step": 162 + }, + { + "epoch": 0.08849077090119435, + "grad_norm": 1.318978856783607, + "learning_rate": 1.1768953068592059e-05, + "loss": 1.9388, + "step": 163 + }, + { + "epoch": 0.08903365906623235, + "grad_norm": 1.58104775083856, + "learning_rate": 1.184115523465704e-05, + "loss": 2.0774, + "step": 164 + }, + { + "epoch": 0.08957654723127036, + "grad_norm": 1.8206646323260607, + "learning_rate": 1.1913357400722023e-05, + "loss": 2.5221, + "step": 165 + }, + { + "epoch": 0.09011943539630836, + "grad_norm": 1.0815960351491627, + "learning_rate": 1.1985559566787005e-05, + "loss": 2.128, + "step": 166 + }, + { + "epoch": 0.09066232356134636, + "grad_norm": 1.3341486432977188, + "learning_rate": 1.2057761732851988e-05, + "loss": 1.9418, + "step": 167 + }, + { + "epoch": 0.09120521172638436, + "grad_norm": 1.3772486036170464, + "learning_rate": 1.2129963898916967e-05, + "loss": 1.84, + "step": 168 + }, + { + "epoch": 0.09174809989142237, + "grad_norm": 1.42874152479176, + "learning_rate": 1.220216606498195e-05, + "loss": 2.2172, + "step": 169 + }, + { + "epoch": 0.09229098805646037, + "grad_norm": 1.9691394708584433, + "learning_rate": 1.2274368231046932e-05, + "loss": 2.9247, + "step": 170 + }, + { + "epoch": 0.09283387622149837, + "grad_norm": 1.4773974635910447, + "learning_rate": 1.2346570397111915e-05, + "loss": 2.1028, + "step": 171 + }, + { + "epoch": 0.09337676438653637, + "grad_norm": 1.5937344901476087, + "learning_rate": 1.2418772563176898e-05, + "loss": 2.4073, + "step": 172 + }, + { + "epoch": 0.09391965255157438, + "grad_norm": 1.630783959399718, + "learning_rate": 1.2490974729241878e-05, + "loss": 2.6978, + "step": 173 + }, + { + "epoch": 0.09446254071661238, + "grad_norm": 1.44736447594164, + "learning_rate": 1.256317689530686e-05, + "loss": 1.8851, + "step": 174 + }, + { + "epoch": 0.09500542888165038, + "grad_norm": 1.3909459856510267, + "learning_rate": 1.2635379061371842e-05, + "loss": 1.576, + "step": 175 + }, + { + "epoch": 0.09554831704668838, + "grad_norm": 1.7838393677575515, + "learning_rate": 1.2707581227436825e-05, + "loss": 2.3895, + "step": 176 + }, + { + "epoch": 0.09609120521172639, + "grad_norm": 1.1294127385799255, + "learning_rate": 1.2779783393501806e-05, + "loss": 2.4855, + "step": 177 + }, + { + "epoch": 0.09663409337676439, + "grad_norm": 1.259662421149648, + "learning_rate": 1.285198555956679e-05, + "loss": 2.5469, + "step": 178 + }, + { + "epoch": 0.09717698154180239, + "grad_norm": 1.4621367635995284, + "learning_rate": 1.2924187725631769e-05, + "loss": 1.9015, + "step": 179 + }, + { + "epoch": 0.09771986970684039, + "grad_norm": 1.6960126954903645, + "learning_rate": 1.2996389891696752e-05, + "loss": 2.9545, + "step": 180 + }, + { + "epoch": 0.0982627578718784, + "grad_norm": 1.5419022503429414, + "learning_rate": 1.3068592057761733e-05, + "loss": 1.9912, + "step": 181 + }, + { + "epoch": 0.0988056460369164, + "grad_norm": 1.4073422496120207, + "learning_rate": 1.3140794223826716e-05, + "loss": 2.3353, + "step": 182 + }, + { + "epoch": 0.0993485342019544, + "grad_norm": 1.2881659254471862, + "learning_rate": 1.3212996389891696e-05, + "loss": 2.7784, + "step": 183 + }, + { + "epoch": 0.0998914223669924, + "grad_norm": 1.361969663712106, + "learning_rate": 1.3285198555956679e-05, + "loss": 2.523, + "step": 184 + }, + { + "epoch": 0.1004343105320304, + "grad_norm": 2.5841731119141316, + "learning_rate": 1.3357400722021662e-05, + "loss": 2.43, + "step": 185 + }, + { + "epoch": 0.10097719869706841, + "grad_norm": 1.3203537847943392, + "learning_rate": 1.3429602888086643e-05, + "loss": 1.9404, + "step": 186 + }, + { + "epoch": 0.10152008686210641, + "grad_norm": 1.393097901493087, + "learning_rate": 1.3501805054151626e-05, + "loss": 2.6202, + "step": 187 + }, + { + "epoch": 0.10206297502714441, + "grad_norm": 0.9414377380652101, + "learning_rate": 1.3574007220216608e-05, + "loss": 2.0879, + "step": 188 + }, + { + "epoch": 0.10260586319218241, + "grad_norm": 1.6547581806701488, + "learning_rate": 1.3646209386281591e-05, + "loss": 2.8912, + "step": 189 + }, + { + "epoch": 0.10314875135722042, + "grad_norm": 1.1236261864015666, + "learning_rate": 1.371841155234657e-05, + "loss": 2.1802, + "step": 190 + }, + { + "epoch": 0.10369163952225842, + "grad_norm": 1.1766384878462062, + "learning_rate": 1.3790613718411554e-05, + "loss": 1.9127, + "step": 191 + }, + { + "epoch": 0.10423452768729642, + "grad_norm": 2.0453724866488656, + "learning_rate": 1.3862815884476535e-05, + "loss": 1.8522, + "step": 192 + }, + { + "epoch": 0.10477741585233442, + "grad_norm": 1.550964370734908, + "learning_rate": 1.3935018050541518e-05, + "loss": 2.5852, + "step": 193 + }, + { + "epoch": 0.10532030401737243, + "grad_norm": 1.303227180334529, + "learning_rate": 1.40072202166065e-05, + "loss": 2.407, + "step": 194 + }, + { + "epoch": 0.10586319218241043, + "grad_norm": 1.2009193115283474, + "learning_rate": 1.4079422382671482e-05, + "loss": 1.8978, + "step": 195 + }, + { + "epoch": 0.10640608034744843, + "grad_norm": 1.353648963417667, + "learning_rate": 1.4151624548736462e-05, + "loss": 2.283, + "step": 196 + }, + { + "epoch": 0.10694896851248643, + "grad_norm": 1.266154745219424, + "learning_rate": 1.4223826714801445e-05, + "loss": 2.0679, + "step": 197 + }, + { + "epoch": 0.10749185667752444, + "grad_norm": 4.186447043886807, + "learning_rate": 1.4296028880866428e-05, + "loss": 1.9175, + "step": 198 + }, + { + "epoch": 0.10803474484256244, + "grad_norm": 1.147065238667786, + "learning_rate": 1.436823104693141e-05, + "loss": 2.2748, + "step": 199 + }, + { + "epoch": 0.10857763300760044, + "grad_norm": 1.2760699256137895, + "learning_rate": 1.4440433212996392e-05, + "loss": 2.3031, + "step": 200 + }, + { + "epoch": 0.10912052117263844, + "grad_norm": 0.9452152631580967, + "learning_rate": 1.4512635379061372e-05, + "loss": 1.4833, + "step": 201 + }, + { + "epoch": 0.10966340933767643, + "grad_norm": 1.3980490034032074, + "learning_rate": 1.4584837545126355e-05, + "loss": 1.5865, + "step": 202 + }, + { + "epoch": 0.11020629750271443, + "grad_norm": 1.6440838011007093, + "learning_rate": 1.4657039711191336e-05, + "loss": 1.632, + "step": 203 + }, + { + "epoch": 0.11074918566775244, + "grad_norm": 1.3969764794458068, + "learning_rate": 1.472924187725632e-05, + "loss": 1.4301, + "step": 204 + }, + { + "epoch": 0.11129207383279044, + "grad_norm": 1.223663090997016, + "learning_rate": 1.48014440433213e-05, + "loss": 2.3993, + "step": 205 + }, + { + "epoch": 0.11183496199782844, + "grad_norm": 1.183881027275754, + "learning_rate": 1.4873646209386284e-05, + "loss": 1.4987, + "step": 206 + }, + { + "epoch": 0.11237785016286644, + "grad_norm": 1.3788483357578956, + "learning_rate": 1.4945848375451264e-05, + "loss": 2.7742, + "step": 207 + }, + { + "epoch": 0.11292073832790445, + "grad_norm": 1.1260622559595288, + "learning_rate": 1.5018050541516247e-05, + "loss": 1.755, + "step": 208 + }, + { + "epoch": 0.11346362649294245, + "grad_norm": 1.8439396062381286, + "learning_rate": 1.5090252707581228e-05, + "loss": 2.9875, + "step": 209 + }, + { + "epoch": 0.11400651465798045, + "grad_norm": 1.2789617727573501, + "learning_rate": 1.5162454873646211e-05, + "loss": 2.0119, + "step": 210 + }, + { + "epoch": 0.11454940282301845, + "grad_norm": 1.3385913561767113, + "learning_rate": 1.5234657039711192e-05, + "loss": 1.7693, + "step": 211 + }, + { + "epoch": 0.11509229098805646, + "grad_norm": 1.049162702021749, + "learning_rate": 1.5306859205776174e-05, + "loss": 1.7541, + "step": 212 + }, + { + "epoch": 0.11563517915309446, + "grad_norm": 1.0797407129441892, + "learning_rate": 1.537906137184116e-05, + "loss": 2.0427, + "step": 213 + }, + { + "epoch": 0.11617806731813246, + "grad_norm": 1.3661735952693994, + "learning_rate": 1.545126353790614e-05, + "loss": 2.722, + "step": 214 + }, + { + "epoch": 0.11672095548317046, + "grad_norm": 1.7084933140661187, + "learning_rate": 1.552346570397112e-05, + "loss": 2.2125, + "step": 215 + }, + { + "epoch": 0.11726384364820847, + "grad_norm": 1.3814428141700028, + "learning_rate": 1.5595667870036102e-05, + "loss": 2.448, + "step": 216 + }, + { + "epoch": 0.11780673181324647, + "grad_norm": 1.298309196615407, + "learning_rate": 1.5667870036101084e-05, + "loss": 2.2706, + "step": 217 + }, + { + "epoch": 0.11834961997828447, + "grad_norm": 1.4553148411944858, + "learning_rate": 1.5740072202166065e-05, + "loss": 2.0087, + "step": 218 + }, + { + "epoch": 0.11889250814332247, + "grad_norm": 1.5083115470870359, + "learning_rate": 1.581227436823105e-05, + "loss": 1.7179, + "step": 219 + }, + { + "epoch": 0.11943539630836048, + "grad_norm": 1.2335179878839475, + "learning_rate": 1.5884476534296028e-05, + "loss": 2.5337, + "step": 220 + }, + { + "epoch": 0.11997828447339848, + "grad_norm": 1.1729282902419653, + "learning_rate": 1.5956678700361013e-05, + "loss": 1.9604, + "step": 221 + }, + { + "epoch": 0.12052117263843648, + "grad_norm": 1.5700347334422509, + "learning_rate": 1.6028880866425994e-05, + "loss": 2.3655, + "step": 222 + }, + { + "epoch": 0.12106406080347448, + "grad_norm": 1.435764746128967, + "learning_rate": 1.6101083032490975e-05, + "loss": 2.109, + "step": 223 + }, + { + "epoch": 0.12160694896851248, + "grad_norm": 1.0866216936301165, + "learning_rate": 1.6173285198555957e-05, + "loss": 2.3334, + "step": 224 + }, + { + "epoch": 0.12214983713355049, + "grad_norm": 1.8060789361277199, + "learning_rate": 1.624548736462094e-05, + "loss": 1.6605, + "step": 225 + }, + { + "epoch": 0.12269272529858849, + "grad_norm": 1.539101200174871, + "learning_rate": 1.6317689530685923e-05, + "loss": 2.2597, + "step": 226 + }, + { + "epoch": 0.12323561346362649, + "grad_norm": 1.7392984350243024, + "learning_rate": 1.6389891696750904e-05, + "loss": 2.5979, + "step": 227 + }, + { + "epoch": 0.1237785016286645, + "grad_norm": 1.6792982939949113, + "learning_rate": 1.6462093862815885e-05, + "loss": 2.0626, + "step": 228 + }, + { + "epoch": 0.1243213897937025, + "grad_norm": 2.067856070276011, + "learning_rate": 1.6534296028880867e-05, + "loss": 2.056, + "step": 229 + }, + { + "epoch": 0.1248642779587405, + "grad_norm": 1.2773224600362139, + "learning_rate": 1.660649819494585e-05, + "loss": 1.5852, + "step": 230 + }, + { + "epoch": 0.1254071661237785, + "grad_norm": 1.381245750855955, + "learning_rate": 1.6678700361010833e-05, + "loss": 2.0675, + "step": 231 + }, + { + "epoch": 0.1259500542888165, + "grad_norm": 1.4215755121774605, + "learning_rate": 1.6750902527075814e-05, + "loss": 2.1002, + "step": 232 + }, + { + "epoch": 0.1264929424538545, + "grad_norm": 1.4581673093175722, + "learning_rate": 1.6823104693140795e-05, + "loss": 1.9279, + "step": 233 + }, + { + "epoch": 0.1270358306188925, + "grad_norm": 1.2656102558648035, + "learning_rate": 1.6895306859205777e-05, + "loss": 1.4913, + "step": 234 + }, + { + "epoch": 0.1275787187839305, + "grad_norm": 1.3985246778048182, + "learning_rate": 1.6967509025270758e-05, + "loss": 2.1629, + "step": 235 + }, + { + "epoch": 0.1281216069489685, + "grad_norm": 1.757194392177075, + "learning_rate": 1.7039711191335743e-05, + "loss": 2.5688, + "step": 236 + }, + { + "epoch": 0.12866449511400652, + "grad_norm": 1.3766112824147358, + "learning_rate": 1.711191335740072e-05, + "loss": 2.3344, + "step": 237 + }, + { + "epoch": 0.12920738327904452, + "grad_norm": 1.2985043105700416, + "learning_rate": 1.7184115523465706e-05, + "loss": 2.0626, + "step": 238 + }, + { + "epoch": 0.12975027144408252, + "grad_norm": 1.715796674825951, + "learning_rate": 1.7256317689530687e-05, + "loss": 2.1559, + "step": 239 + }, + { + "epoch": 0.13029315960912052, + "grad_norm": 1.6374209306607368, + "learning_rate": 1.7328519855595668e-05, + "loss": 2.0801, + "step": 240 + }, + { + "epoch": 0.13083604777415853, + "grad_norm": 1.244682506367769, + "learning_rate": 1.7400722021660653e-05, + "loss": 1.6125, + "step": 241 + }, + { + "epoch": 0.13137893593919653, + "grad_norm": 1.3001838781366861, + "learning_rate": 1.7472924187725634e-05, + "loss": 2.0934, + "step": 242 + }, + { + "epoch": 0.13192182410423453, + "grad_norm": 1.1600879910417528, + "learning_rate": 1.7545126353790616e-05, + "loss": 1.7862, + "step": 243 + }, + { + "epoch": 0.13246471226927253, + "grad_norm": 1.3503312968411385, + "learning_rate": 1.7617328519855597e-05, + "loss": 2.1636, + "step": 244 + }, + { + "epoch": 0.13300760043431054, + "grad_norm": 2.1299761628528167, + "learning_rate": 1.768953068592058e-05, + "loss": 2.4826, + "step": 245 + }, + { + "epoch": 0.13355048859934854, + "grad_norm": 1.3232052544377066, + "learning_rate": 1.776173285198556e-05, + "loss": 2.0497, + "step": 246 + }, + { + "epoch": 0.13409337676438654, + "grad_norm": 1.5079892467642568, + "learning_rate": 1.7833935018050544e-05, + "loss": 2.4552, + "step": 247 + }, + { + "epoch": 0.13463626492942454, + "grad_norm": 1.716355318914646, + "learning_rate": 1.7906137184115526e-05, + "loss": 1.5309, + "step": 248 + }, + { + "epoch": 0.13517915309446255, + "grad_norm": 2.1523878955969193, + "learning_rate": 1.7978339350180507e-05, + "loss": 2.7403, + "step": 249 + }, + { + "epoch": 0.13572204125950055, + "grad_norm": 1.8169301927997485, + "learning_rate": 1.805054151624549e-05, + "loss": 2.3324, + "step": 250 + }, + { + "epoch": 0.13626492942453855, + "grad_norm": 1.4464924096148426, + "learning_rate": 1.812274368231047e-05, + "loss": 1.8399, + "step": 251 + }, + { + "epoch": 0.13680781758957655, + "grad_norm": 1.9818298103877654, + "learning_rate": 1.8194945848375454e-05, + "loss": 2.4237, + "step": 252 + }, + { + "epoch": 0.13735070575461455, + "grad_norm": 1.6371003739671923, + "learning_rate": 1.8267148014440436e-05, + "loss": 2.0477, + "step": 253 + }, + { + "epoch": 0.13789359391965256, + "grad_norm": 1.2944832148952543, + "learning_rate": 1.8339350180505417e-05, + "loss": 2.2269, + "step": 254 + }, + { + "epoch": 0.13843648208469056, + "grad_norm": 1.4284289887933213, + "learning_rate": 1.84115523465704e-05, + "loss": 2.2465, + "step": 255 + }, + { + "epoch": 0.13897937024972856, + "grad_norm": 1.4633226693151502, + "learning_rate": 1.848375451263538e-05, + "loss": 2.1582, + "step": 256 + }, + { + "epoch": 0.13952225841476656, + "grad_norm": 1.6169546473154692, + "learning_rate": 1.855595667870036e-05, + "loss": 1.9526, + "step": 257 + }, + { + "epoch": 0.14006514657980457, + "grad_norm": 1.47082651679944, + "learning_rate": 1.8628158844765346e-05, + "loss": 2.0314, + "step": 258 + }, + { + "epoch": 0.14060803474484257, + "grad_norm": 1.2279964838302115, + "learning_rate": 1.8700361010830327e-05, + "loss": 1.4345, + "step": 259 + }, + { + "epoch": 0.14115092290988057, + "grad_norm": 1.5727648286246085, + "learning_rate": 1.877256317689531e-05, + "loss": 2.7371, + "step": 260 + }, + { + "epoch": 0.14169381107491857, + "grad_norm": 1.5536143039045338, + "learning_rate": 1.884476534296029e-05, + "loss": 1.6691, + "step": 261 + }, + { + "epoch": 0.14223669923995658, + "grad_norm": 1.504953118413033, + "learning_rate": 1.891696750902527e-05, + "loss": 1.8422, + "step": 262 + }, + { + "epoch": 0.14277958740499458, + "grad_norm": 1.6092868309857633, + "learning_rate": 1.8989169675090253e-05, + "loss": 1.6446, + "step": 263 + }, + { + "epoch": 0.14332247557003258, + "grad_norm": 1.434484546738067, + "learning_rate": 1.9061371841155237e-05, + "loss": 1.5945, + "step": 264 + }, + { + "epoch": 0.14386536373507058, + "grad_norm": 1.3472204852366787, + "learning_rate": 1.913357400722022e-05, + "loss": 1.4511, + "step": 265 + }, + { + "epoch": 0.1444082519001086, + "grad_norm": 2.493842366649246, + "learning_rate": 1.92057761732852e-05, + "loss": 2.1144, + "step": 266 + }, + { + "epoch": 0.1449511400651466, + "grad_norm": 1.3138428233046642, + "learning_rate": 1.927797833935018e-05, + "loss": 1.6592, + "step": 267 + }, + { + "epoch": 0.1454940282301846, + "grad_norm": 1.7201458829291465, + "learning_rate": 1.9350180505415163e-05, + "loss": 1.6895, + "step": 268 + }, + { + "epoch": 0.1460369163952226, + "grad_norm": 1.9777013455620547, + "learning_rate": 1.9422382671480147e-05, + "loss": 2.2628, + "step": 269 + }, + { + "epoch": 0.1465798045602606, + "grad_norm": 1.8559862870685497, + "learning_rate": 1.949458483754513e-05, + "loss": 1.3374, + "step": 270 + }, + { + "epoch": 0.1471226927252986, + "grad_norm": 1.919404449298866, + "learning_rate": 1.956678700361011e-05, + "loss": 2.3805, + "step": 271 + }, + { + "epoch": 0.1476655808903366, + "grad_norm": 1.4760010543412088, + "learning_rate": 1.963898916967509e-05, + "loss": 1.9791, + "step": 272 + }, + { + "epoch": 0.1482084690553746, + "grad_norm": 1.4491842831658432, + "learning_rate": 1.9711191335740073e-05, + "loss": 1.8119, + "step": 273 + }, + { + "epoch": 0.1487513572204126, + "grad_norm": 1.4098796806726177, + "learning_rate": 1.9783393501805054e-05, + "loss": 1.7398, + "step": 274 + }, + { + "epoch": 0.1492942453854506, + "grad_norm": 1.5009369942229867, + "learning_rate": 1.985559566787004e-05, + "loss": 1.5194, + "step": 275 + }, + { + "epoch": 0.1498371335504886, + "grad_norm": 1.5073462671342261, + "learning_rate": 1.992779783393502e-05, + "loss": 1.8125, + "step": 276 + }, + { + "epoch": 0.1503800217155266, + "grad_norm": 1.9315604890805302, + "learning_rate": 2e-05, + "loss": 1.5429, + "step": 277 + }, + { + "epoch": 0.15092290988056462, + "grad_norm": 1.5059106777628657, + "learning_rate": 1.999999938159203e-05, + "loss": 1.6402, + "step": 278 + }, + { + "epoch": 0.15146579804560262, + "grad_norm": 1.7255630320192266, + "learning_rate": 1.9999997526368205e-05, + "loss": 2.2806, + "step": 279 + }, + { + "epoch": 0.15200868621064062, + "grad_norm": 1.7985407961594548, + "learning_rate": 1.999999443432874e-05, + "loss": 2.5213, + "step": 280 + }, + { + "epoch": 0.15255157437567862, + "grad_norm": 1.801471230172198, + "learning_rate": 1.999999010547403e-05, + "loss": 1.8362, + "step": 281 + }, + { + "epoch": 0.15309446254071662, + "grad_norm": 3.236881896883975, + "learning_rate": 1.999998453980461e-05, + "loss": 2.7148, + "step": 282 + }, + { + "epoch": 0.15363735070575463, + "grad_norm": 1.7343786362856708, + "learning_rate": 1.9999977737321156e-05, + "loss": 1.8195, + "step": 283 + }, + { + "epoch": 0.15418023887079263, + "grad_norm": 1.5006942215232695, + "learning_rate": 1.999996969802452e-05, + "loss": 1.8744, + "step": 284 + }, + { + "epoch": 0.15472312703583063, + "grad_norm": 4.171967924853298, + "learning_rate": 1.99999604219157e-05, + "loss": 2.4329, + "step": 285 + }, + { + "epoch": 0.15526601520086863, + "grad_norm": 1.7546486556973808, + "learning_rate": 1.9999949908995832e-05, + "loss": 1.3931, + "step": 286 + }, + { + "epoch": 0.15580890336590664, + "grad_norm": 2.0223702721102375, + "learning_rate": 1.999993815926622e-05, + "loss": 1.6704, + "step": 287 + }, + { + "epoch": 0.1563517915309446, + "grad_norm": 1.8736762651082781, + "learning_rate": 1.9999925172728324e-05, + "loss": 1.9895, + "step": 288 + }, + { + "epoch": 0.15689467969598261, + "grad_norm": 1.8878827291463698, + "learning_rate": 1.9999910949383742e-05, + "loss": 2.4939, + "step": 289 + }, + { + "epoch": 0.15743756786102062, + "grad_norm": 1.7277080078786893, + "learning_rate": 1.9999895489234234e-05, + "loss": 1.9265, + "step": 290 + }, + { + "epoch": 0.15798045602605862, + "grad_norm": 1.7402172969699101, + "learning_rate": 1.999987879228172e-05, + "loss": 1.8101, + "step": 291 + }, + { + "epoch": 0.15852334419109662, + "grad_norm": 2.077278713661093, + "learning_rate": 1.9999860858528257e-05, + "loss": 2.1763, + "step": 292 + }, + { + "epoch": 0.15906623235613462, + "grad_norm": 2.247624069618552, + "learning_rate": 1.9999841687976067e-05, + "loss": 1.9727, + "step": 293 + }, + { + "epoch": 0.15960912052117263, + "grad_norm": 1.6712153667666279, + "learning_rate": 1.999982128062752e-05, + "loss": 1.9921, + "step": 294 + }, + { + "epoch": 0.16015200868621063, + "grad_norm": 1.7126188149003763, + "learning_rate": 1.999979963648514e-05, + "loss": 1.9375, + "step": 295 + }, + { + "epoch": 0.16069489685124863, + "grad_norm": 1.949228534239786, + "learning_rate": 1.99997767555516e-05, + "loss": 2.1628, + "step": 296 + }, + { + "epoch": 0.16123778501628663, + "grad_norm": 1.8253305155157105, + "learning_rate": 1.9999752637829734e-05, + "loss": 1.8293, + "step": 297 + }, + { + "epoch": 0.16178067318132464, + "grad_norm": 1.5881557560095252, + "learning_rate": 1.9999727283322524e-05, + "loss": 1.3688, + "step": 298 + }, + { + "epoch": 0.16232356134636264, + "grad_norm": 2.4008202914116503, + "learning_rate": 1.9999700692033112e-05, + "loss": 1.5919, + "step": 299 + }, + { + "epoch": 0.16286644951140064, + "grad_norm": 1.610839420604411, + "learning_rate": 1.9999672863964778e-05, + "loss": 1.8724, + "step": 300 + }, + { + "epoch": 0.16340933767643864, + "grad_norm": 1.4663439746230646, + "learning_rate": 1.9999643799120964e-05, + "loss": 1.4628, + "step": 301 + }, + { + "epoch": 0.16395222584147665, + "grad_norm": 1.7051101810518976, + "learning_rate": 1.9999613497505272e-05, + "loss": 1.2916, + "step": 302 + }, + { + "epoch": 0.16449511400651465, + "grad_norm": 1.8967399031374104, + "learning_rate": 1.9999581959121443e-05, + "loss": 1.7559, + "step": 303 + }, + { + "epoch": 0.16503800217155265, + "grad_norm": 1.6206520761555416, + "learning_rate": 1.9999549183973382e-05, + "loss": 1.9196, + "step": 304 + }, + { + "epoch": 0.16558089033659065, + "grad_norm": 2.3545652344963544, + "learning_rate": 1.999951517206514e-05, + "loss": 1.2731, + "step": 305 + }, + { + "epoch": 0.16612377850162866, + "grad_norm": 1.7952736257095967, + "learning_rate": 1.9999479923400926e-05, + "loss": 2.0502, + "step": 306 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 1.7199420678903319, + "learning_rate": 1.99994434379851e-05, + "loss": 1.7059, + "step": 307 + }, + { + "epoch": 0.16720955483170466, + "grad_norm": 2.3318553047346, + "learning_rate": 1.9999405715822167e-05, + "loss": 1.4667, + "step": 308 + }, + { + "epoch": 0.16775244299674266, + "grad_norm": 1.618961362672766, + "learning_rate": 1.9999366756916804e-05, + "loss": 1.1866, + "step": 309 + }, + { + "epoch": 0.16829533116178066, + "grad_norm": 2.0805239071290114, + "learning_rate": 1.999932656127382e-05, + "loss": 1.7601, + "step": 310 + }, + { + "epoch": 0.16883821932681867, + "grad_norm": 1.782022796323131, + "learning_rate": 1.9999285128898193e-05, + "loss": 1.6207, + "step": 311 + }, + { + "epoch": 0.16938110749185667, + "grad_norm": 2.2438870289045028, + "learning_rate": 1.9999242459795045e-05, + "loss": 2.1329, + "step": 312 + }, + { + "epoch": 0.16992399565689467, + "grad_norm": 2.2550840823235587, + "learning_rate": 1.9999198553969652e-05, + "loss": 1.6055, + "step": 313 + }, + { + "epoch": 0.17046688382193267, + "grad_norm": 1.8012459587153988, + "learning_rate": 1.9999153411427445e-05, + "loss": 2.018, + "step": 314 + }, + { + "epoch": 0.17100977198697068, + "grad_norm": 1.8498840013582993, + "learning_rate": 1.9999107032174007e-05, + "loss": 1.8113, + "step": 315 + }, + { + "epoch": 0.17155266015200868, + "grad_norm": 2.2799557163953965, + "learning_rate": 1.9999059416215078e-05, + "loss": 1.9494, + "step": 316 + }, + { + "epoch": 0.17209554831704668, + "grad_norm": 2.383891228288174, + "learning_rate": 1.999901056355654e-05, + "loss": 1.8323, + "step": 317 + }, + { + "epoch": 0.17263843648208468, + "grad_norm": 1.9081077971876759, + "learning_rate": 1.9998960474204443e-05, + "loss": 1.5476, + "step": 318 + }, + { + "epoch": 0.1731813246471227, + "grad_norm": 1.9010783437994137, + "learning_rate": 1.999890914816498e-05, + "loss": 1.4602, + "step": 319 + }, + { + "epoch": 0.1737242128121607, + "grad_norm": 2.295573334252978, + "learning_rate": 1.9998856585444493e-05, + "loss": 1.0965, + "step": 320 + }, + { + "epoch": 0.1742671009771987, + "grad_norm": 2.1014332124066954, + "learning_rate": 1.999880278604949e-05, + "loss": 1.3726, + "step": 321 + }, + { + "epoch": 0.1748099891422367, + "grad_norm": 2.0418722317428206, + "learning_rate": 1.9998747749986625e-05, + "loss": 1.8875, + "step": 322 + }, + { + "epoch": 0.1753528773072747, + "grad_norm": 1.768897707524193, + "learning_rate": 1.99986914772627e-05, + "loss": 1.8841, + "step": 323 + }, + { + "epoch": 0.1758957654723127, + "grad_norm": 2.603335158781435, + "learning_rate": 1.9998633967884676e-05, + "loss": 1.4262, + "step": 324 + }, + { + "epoch": 0.1764386536373507, + "grad_norm": 2.1485447383197607, + "learning_rate": 1.999857522185967e-05, + "loss": 1.9283, + "step": 325 + }, + { + "epoch": 0.1769815418023887, + "grad_norm": 1.8398472624384716, + "learning_rate": 1.9998515239194945e-05, + "loss": 1.9089, + "step": 326 + }, + { + "epoch": 0.1775244299674267, + "grad_norm": 2.6988014220525374, + "learning_rate": 1.9998454019897918e-05, + "loss": 1.3636, + "step": 327 + }, + { + "epoch": 0.1780673181324647, + "grad_norm": 2.182794188412918, + "learning_rate": 1.9998391563976166e-05, + "loss": 1.7041, + "step": 328 + }, + { + "epoch": 0.1786102062975027, + "grad_norm": 1.970858895313084, + "learning_rate": 1.9998327871437405e-05, + "loss": 1.4917, + "step": 329 + }, + { + "epoch": 0.1791530944625407, + "grad_norm": 1.829655959511745, + "learning_rate": 1.9998262942289524e-05, + "loss": 1.9694, + "step": 330 + }, + { + "epoch": 0.17969598262757872, + "grad_norm": 1.9990017706312653, + "learning_rate": 1.9998196776540545e-05, + "loss": 2.067, + "step": 331 + }, + { + "epoch": 0.18023887079261672, + "grad_norm": 2.2829395740777234, + "learning_rate": 1.9998129374198655e-05, + "loss": 1.6707, + "step": 332 + }, + { + "epoch": 0.18078175895765472, + "grad_norm": 2.174263401724357, + "learning_rate": 1.9998060735272186e-05, + "loss": 1.6897, + "step": 333 + }, + { + "epoch": 0.18132464712269272, + "grad_norm": 1.8438064757899841, + "learning_rate": 1.9997990859769633e-05, + "loss": 1.4446, + "step": 334 + }, + { + "epoch": 0.18186753528773072, + "grad_norm": 2.0988396037795507, + "learning_rate": 1.9997919747699638e-05, + "loss": 1.8837, + "step": 335 + }, + { + "epoch": 0.18241042345276873, + "grad_norm": 2.582616969632433, + "learning_rate": 1.999784739907099e-05, + "loss": 2.459, + "step": 336 + }, + { + "epoch": 0.18295331161780673, + "grad_norm": 2.5198677585207707, + "learning_rate": 1.9997773813892644e-05, + "loss": 1.5117, + "step": 337 + }, + { + "epoch": 0.18349619978284473, + "grad_norm": 2.1361317951692635, + "learning_rate": 1.9997698992173697e-05, + "loss": 1.5892, + "step": 338 + }, + { + "epoch": 0.18403908794788273, + "grad_norm": 1.7284990063627486, + "learning_rate": 1.9997622933923406e-05, + "loss": 1.5151, + "step": 339 + }, + { + "epoch": 0.18458197611292074, + "grad_norm": 2.076752898704788, + "learning_rate": 1.9997545639151176e-05, + "loss": 2.3424, + "step": 340 + }, + { + "epoch": 0.18512486427795874, + "grad_norm": 2.479369276757794, + "learning_rate": 1.999746710786657e-05, + "loss": 1.5923, + "step": 341 + }, + { + "epoch": 0.18566775244299674, + "grad_norm": 1.9193986648366674, + "learning_rate": 1.9997387340079294e-05, + "loss": 1.6222, + "step": 342 + }, + { + "epoch": 0.18621064060803474, + "grad_norm": 1.8622986153150223, + "learning_rate": 1.999730633579922e-05, + "loss": 1.8402, + "step": 343 + }, + { + "epoch": 0.18675352877307275, + "grad_norm": 2.4012194427686238, + "learning_rate": 1.999722409503637e-05, + "loss": 1.9406, + "step": 344 + }, + { + "epoch": 0.18729641693811075, + "grad_norm": 2.3641518463127538, + "learning_rate": 1.9997140617800907e-05, + "loss": 2.0108, + "step": 345 + }, + { + "epoch": 0.18783930510314875, + "grad_norm": 2.406749148558192, + "learning_rate": 1.9997055904103156e-05, + "loss": 1.1323, + "step": 346 + }, + { + "epoch": 0.18838219326818675, + "grad_norm": 2.002785642517938, + "learning_rate": 1.99969699539536e-05, + "loss": 1.8745, + "step": 347 + }, + { + "epoch": 0.18892508143322476, + "grad_norm": 2.0002882431712816, + "learning_rate": 1.9996882767362874e-05, + "loss": 1.6311, + "step": 348 + }, + { + "epoch": 0.18946796959826276, + "grad_norm": 2.14418043520414, + "learning_rate": 1.9996794344341744e-05, + "loss": 1.7789, + "step": 349 + }, + { + "epoch": 0.19001085776330076, + "grad_norm": 2.0976983604919144, + "learning_rate": 1.9996704684901163e-05, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.19055374592833876, + "grad_norm": 2.017704706993719, + "learning_rate": 1.9996613789052214e-05, + "loss": 1.7974, + "step": 351 + }, + { + "epoch": 0.19109663409337677, + "grad_norm": 1.9907769296520095, + "learning_rate": 1.999652165680614e-05, + "loss": 1.9079, + "step": 352 + }, + { + "epoch": 0.19163952225841477, + "grad_norm": 2.3031451756435746, + "learning_rate": 1.999642828817433e-05, + "loss": 1.7022, + "step": 353 + }, + { + "epoch": 0.19218241042345277, + "grad_norm": 1.6722822353766713, + "learning_rate": 1.9996333683168342e-05, + "loss": 1.6655, + "step": 354 + }, + { + "epoch": 0.19272529858849077, + "grad_norm": 2.2974087830498027, + "learning_rate": 1.9996237841799874e-05, + "loss": 1.761, + "step": 355 + }, + { + "epoch": 0.19326818675352878, + "grad_norm": 1.9181254747298617, + "learning_rate": 1.9996140764080777e-05, + "loss": 1.8259, + "step": 356 + }, + { + "epoch": 0.19381107491856678, + "grad_norm": 2.07933834321742, + "learning_rate": 1.9996042450023053e-05, + "loss": 1.3783, + "step": 357 + }, + { + "epoch": 0.19435396308360478, + "grad_norm": 2.8332787295887436, + "learning_rate": 1.9995942899638875e-05, + "loss": 1.9796, + "step": 358 + }, + { + "epoch": 0.19489685124864278, + "grad_norm": 1.7924418183793758, + "learning_rate": 1.9995842112940545e-05, + "loss": 1.5063, + "step": 359 + }, + { + "epoch": 0.19543973941368079, + "grad_norm": 2.78627522830663, + "learning_rate": 1.9995740089940532e-05, + "loss": 1.891, + "step": 360 + }, + { + "epoch": 0.1959826275787188, + "grad_norm": 2.067311381839208, + "learning_rate": 1.9995636830651453e-05, + "loss": 1.9235, + "step": 361 + }, + { + "epoch": 0.1965255157437568, + "grad_norm": 2.539112412358468, + "learning_rate": 1.9995532335086078e-05, + "loss": 2.3062, + "step": 362 + }, + { + "epoch": 0.1970684039087948, + "grad_norm": 1.9040915545393904, + "learning_rate": 1.999542660325734e-05, + "loss": 1.9263, + "step": 363 + }, + { + "epoch": 0.1976112920738328, + "grad_norm": 1.7664296640025938, + "learning_rate": 1.9995319635178305e-05, + "loss": 1.5562, + "step": 364 + }, + { + "epoch": 0.1981541802388708, + "grad_norm": 1.687746217581272, + "learning_rate": 1.9995211430862206e-05, + "loss": 1.738, + "step": 365 + }, + { + "epoch": 0.1986970684039088, + "grad_norm": 1.9292930460598687, + "learning_rate": 1.9995101990322428e-05, + "loss": 1.5446, + "step": 366 + }, + { + "epoch": 0.1992399565689468, + "grad_norm": 1.9890390069401638, + "learning_rate": 1.9994991313572508e-05, + "loss": 1.9086, + "step": 367 + }, + { + "epoch": 0.1997828447339848, + "grad_norm": 1.8632430045841502, + "learning_rate": 1.999487940062613e-05, + "loss": 1.9747, + "step": 368 + }, + { + "epoch": 0.2003257328990228, + "grad_norm": 2.0677818303742237, + "learning_rate": 1.999476625149714e-05, + "loss": 1.4377, + "step": 369 + }, + { + "epoch": 0.2008686210640608, + "grad_norm": 1.7412205305358945, + "learning_rate": 1.9994651866199527e-05, + "loss": 1.9352, + "step": 370 + }, + { + "epoch": 0.2014115092290988, + "grad_norm": 1.8562175126383418, + "learning_rate": 1.9994536244747448e-05, + "loss": 1.5469, + "step": 371 + }, + { + "epoch": 0.20195439739413681, + "grad_norm": 1.8688345857586197, + "learning_rate": 1.9994419387155194e-05, + "loss": 1.8327, + "step": 372 + }, + { + "epoch": 0.20249728555917482, + "grad_norm": 2.1538407513899647, + "learning_rate": 1.9994301293437223e-05, + "loss": 1.9621, + "step": 373 + }, + { + "epoch": 0.20304017372421282, + "grad_norm": 2.4849664096044752, + "learning_rate": 1.999418196360814e-05, + "loss": 1.8179, + "step": 374 + }, + { + "epoch": 0.20358306188925082, + "grad_norm": 2.098949652063271, + "learning_rate": 1.99940613976827e-05, + "loss": 1.8719, + "step": 375 + }, + { + "epoch": 0.20412595005428882, + "grad_norm": 1.8802547433747465, + "learning_rate": 1.999393959567582e-05, + "loss": 1.5697, + "step": 376 + }, + { + "epoch": 0.20466883821932683, + "grad_norm": 2.090861017905227, + "learning_rate": 1.9993816557602567e-05, + "loss": 1.8249, + "step": 377 + }, + { + "epoch": 0.20521172638436483, + "grad_norm": 1.775481861127165, + "learning_rate": 1.999369228347815e-05, + "loss": 1.3095, + "step": 378 + }, + { + "epoch": 0.20575461454940283, + "grad_norm": 2.260330600420479, + "learning_rate": 1.999356677331794e-05, + "loss": 1.9153, + "step": 379 + }, + { + "epoch": 0.20629750271444083, + "grad_norm": 2.0769016143551484, + "learning_rate": 1.999344002713747e-05, + "loss": 1.958, + "step": 380 + }, + { + "epoch": 0.20684039087947884, + "grad_norm": 2.063584775140761, + "learning_rate": 1.9993312044952408e-05, + "loss": 1.7887, + "step": 381 + }, + { + "epoch": 0.20738327904451684, + "grad_norm": 1.9638736053910157, + "learning_rate": 1.9993182826778588e-05, + "loss": 1.3178, + "step": 382 + }, + { + "epoch": 0.20792616720955484, + "grad_norm": 1.7752329715548703, + "learning_rate": 1.9993052372631988e-05, + "loss": 1.4473, + "step": 383 + }, + { + "epoch": 0.20846905537459284, + "grad_norm": 2.1002641405528952, + "learning_rate": 1.999292068252874e-05, + "loss": 1.9698, + "step": 384 + }, + { + "epoch": 0.20901194353963085, + "grad_norm": 2.757803188261632, + "learning_rate": 1.999278775648514e-05, + "loss": 1.1517, + "step": 385 + }, + { + "epoch": 0.20955483170466885, + "grad_norm": 3.0341284449704378, + "learning_rate": 1.9992653594517624e-05, + "loss": 1.987, + "step": 386 + }, + { + "epoch": 0.21009771986970685, + "grad_norm": 2.445830333821562, + "learning_rate": 1.9992518196642786e-05, + "loss": 1.7656, + "step": 387 + }, + { + "epoch": 0.21064060803474485, + "grad_norm": 2.737783962568244, + "learning_rate": 1.9992381562877368e-05, + "loss": 2.3012, + "step": 388 + }, + { + "epoch": 0.21118349619978286, + "grad_norm": 2.0108432149096793, + "learning_rate": 1.9992243693238275e-05, + "loss": 2.1096, + "step": 389 + }, + { + "epoch": 0.21172638436482086, + "grad_norm": 2.1448465445226397, + "learning_rate": 1.9992104587742558e-05, + "loss": 1.6912, + "step": 390 + }, + { + "epoch": 0.21226927252985886, + "grad_norm": 2.842739011518123, + "learning_rate": 1.999196424640742e-05, + "loss": 1.9987, + "step": 391 + }, + { + "epoch": 0.21281216069489686, + "grad_norm": 2.026900017823922, + "learning_rate": 1.9991822669250216e-05, + "loss": 1.6251, + "step": 392 + }, + { + "epoch": 0.21335504885993486, + "grad_norm": 2.139818571931489, + "learning_rate": 1.9991679856288462e-05, + "loss": 1.4181, + "step": 393 + }, + { + "epoch": 0.21389793702497287, + "grad_norm": 2.061327407258535, + "learning_rate": 1.999153580753982e-05, + "loss": 2.0976, + "step": 394 + }, + { + "epoch": 0.21444082519001087, + "grad_norm": 1.9510646261617408, + "learning_rate": 1.9991390523022105e-05, + "loss": 1.3091, + "step": 395 + }, + { + "epoch": 0.21498371335504887, + "grad_norm": 2.0892427860245664, + "learning_rate": 1.9991244002753287e-05, + "loss": 1.3693, + "step": 396 + }, + { + "epoch": 0.21552660152008687, + "grad_norm": 1.9528708238712815, + "learning_rate": 1.9991096246751483e-05, + "loss": 1.2807, + "step": 397 + }, + { + "epoch": 0.21606948968512488, + "grad_norm": 2.0084632909809983, + "learning_rate": 1.9990947255034977e-05, + "loss": 1.7429, + "step": 398 + }, + { + "epoch": 0.21661237785016288, + "grad_norm": 2.19414527352725, + "learning_rate": 1.999079702762219e-05, + "loss": 1.5962, + "step": 399 + }, + { + "epoch": 0.21715526601520088, + "grad_norm": 2.184484203373308, + "learning_rate": 1.9990645564531702e-05, + "loss": 1.6887, + "step": 400 + }, + { + "epoch": 0.21769815418023888, + "grad_norm": 2.214473402714986, + "learning_rate": 1.9990492865782248e-05, + "loss": 1.5226, + "step": 401 + }, + { + "epoch": 0.2182410423452769, + "grad_norm": 2.1467748369867414, + "learning_rate": 1.9990338931392714e-05, + "loss": 1.5634, + "step": 402 + }, + { + "epoch": 0.21878393051031486, + "grad_norm": 2.1969353910959684, + "learning_rate": 1.999018376138214e-05, + "loss": 1.8652, + "step": 403 + }, + { + "epoch": 0.21932681867535286, + "grad_norm": 2.245031544698939, + "learning_rate": 1.9990027355769715e-05, + "loss": 2.1811, + "step": 404 + }, + { + "epoch": 0.21986970684039087, + "grad_norm": 2.381275606702572, + "learning_rate": 1.9989869714574784e-05, + "loss": 1.6356, + "step": 405 + }, + { + "epoch": 0.22041259500542887, + "grad_norm": 2.8847531540936053, + "learning_rate": 1.9989710837816846e-05, + "loss": 1.9953, + "step": 406 + }, + { + "epoch": 0.22095548317046687, + "grad_norm": 2.383074182943421, + "learning_rate": 1.9989550725515553e-05, + "loss": 1.5589, + "step": 407 + }, + { + "epoch": 0.22149837133550487, + "grad_norm": 2.3867744628533987, + "learning_rate": 1.99893893776907e-05, + "loss": 1.5351, + "step": 408 + }, + { + "epoch": 0.22204125950054288, + "grad_norm": 2.4139767232435463, + "learning_rate": 1.998922679436225e-05, + "loss": 1.5949, + "step": 409 + }, + { + "epoch": 0.22258414766558088, + "grad_norm": 3.034840719531937, + "learning_rate": 1.9989062975550313e-05, + "loss": 1.7994, + "step": 410 + }, + { + "epoch": 0.22312703583061888, + "grad_norm": 2.6245673280102544, + "learning_rate": 1.9988897921275144e-05, + "loss": 2.2593, + "step": 411 + }, + { + "epoch": 0.22366992399565688, + "grad_norm": 2.403934048820897, + "learning_rate": 1.998873163155716e-05, + "loss": 1.3179, + "step": 412 + }, + { + "epoch": 0.22421281216069489, + "grad_norm": 2.0174357446052524, + "learning_rate": 1.998856410641693e-05, + "loss": 1.8698, + "step": 413 + }, + { + "epoch": 0.2247557003257329, + "grad_norm": 2.25869172419699, + "learning_rate": 1.998839534587517e-05, + "loss": 1.596, + "step": 414 + }, + { + "epoch": 0.2252985884907709, + "grad_norm": 3.384596068816443, + "learning_rate": 1.9988225349952758e-05, + "loss": 1.5232, + "step": 415 + }, + { + "epoch": 0.2258414766558089, + "grad_norm": 2.2109577343653246, + "learning_rate": 1.9988054118670712e-05, + "loss": 1.2403, + "step": 416 + }, + { + "epoch": 0.2263843648208469, + "grad_norm": 2.1825505717325053, + "learning_rate": 1.9987881652050215e-05, + "loss": 1.7205, + "step": 417 + }, + { + "epoch": 0.2269272529858849, + "grad_norm": 2.7992033856118583, + "learning_rate": 1.99877079501126e-05, + "loss": 1.7486, + "step": 418 + }, + { + "epoch": 0.2274701411509229, + "grad_norm": 2.6351781282432793, + "learning_rate": 1.9987533012879344e-05, + "loss": 1.6693, + "step": 419 + }, + { + "epoch": 0.2280130293159609, + "grad_norm": 2.2385207734696384, + "learning_rate": 1.9987356840372088e-05, + "loss": 1.1296, + "step": 420 + }, + { + "epoch": 0.2285559174809989, + "grad_norm": 2.4921881643709107, + "learning_rate": 1.998717943261262e-05, + "loss": 1.1379, + "step": 421 + }, + { + "epoch": 0.2290988056460369, + "grad_norm": 2.842685080646979, + "learning_rate": 1.9987000789622884e-05, + "loss": 1.7963, + "step": 422 + }, + { + "epoch": 0.2296416938110749, + "grad_norm": 2.4776027353414145, + "learning_rate": 1.9986820911424972e-05, + "loss": 1.6939, + "step": 423 + }, + { + "epoch": 0.2301845819761129, + "grad_norm": 2.688785422525701, + "learning_rate": 1.9986639798041134e-05, + "loss": 1.2282, + "step": 424 + }, + { + "epoch": 0.23072747014115091, + "grad_norm": 3.4758865457335495, + "learning_rate": 1.998645744949377e-05, + "loss": 1.7664, + "step": 425 + }, + { + "epoch": 0.23127035830618892, + "grad_norm": 2.5586740557561605, + "learning_rate": 1.9986273865805432e-05, + "loss": 1.5637, + "step": 426 + }, + { + "epoch": 0.23181324647122692, + "grad_norm": 2.645306116903878, + "learning_rate": 1.9986089046998827e-05, + "loss": 1.5331, + "step": 427 + }, + { + "epoch": 0.23235613463626492, + "grad_norm": 2.748145883999443, + "learning_rate": 1.998590299309681e-05, + "loss": 1.7316, + "step": 428 + }, + { + "epoch": 0.23289902280130292, + "grad_norm": 2.4357923833746438, + "learning_rate": 1.99857157041224e-05, + "loss": 1.1479, + "step": 429 + }, + { + "epoch": 0.23344191096634093, + "grad_norm": 2.7322633978331377, + "learning_rate": 1.9985527180098755e-05, + "loss": 1.6089, + "step": 430 + }, + { + "epoch": 0.23398479913137893, + "grad_norm": 2.625531215380387, + "learning_rate": 1.9985337421049193e-05, + "loss": 1.8973, + "step": 431 + }, + { + "epoch": 0.23452768729641693, + "grad_norm": 2.1586487666504754, + "learning_rate": 1.9985146426997185e-05, + "loss": 1.664, + "step": 432 + }, + { + "epoch": 0.23507057546145493, + "grad_norm": 2.748704307086674, + "learning_rate": 1.9984954197966355e-05, + "loss": 1.677, + "step": 433 + }, + { + "epoch": 0.23561346362649294, + "grad_norm": 2.345532526959197, + "learning_rate": 1.9984760733980476e-05, + "loss": 1.5133, + "step": 434 + }, + { + "epoch": 0.23615635179153094, + "grad_norm": 2.1655585346308848, + "learning_rate": 1.9984566035063473e-05, + "loss": 1.4206, + "step": 435 + }, + { + "epoch": 0.23669923995656894, + "grad_norm": 2.697584333049271, + "learning_rate": 1.9984370101239434e-05, + "loss": 1.8131, + "step": 436 + }, + { + "epoch": 0.23724212812160694, + "grad_norm": 2.235069158786981, + "learning_rate": 1.9984172932532583e-05, + "loss": 1.7839, + "step": 437 + }, + { + "epoch": 0.23778501628664495, + "grad_norm": 2.548562657856099, + "learning_rate": 1.998397452896731e-05, + "loss": 1.4876, + "step": 438 + }, + { + "epoch": 0.23832790445168295, + "grad_norm": 2.1698874138883673, + "learning_rate": 1.9983774890568163e-05, + "loss": 1.2916, + "step": 439 + }, + { + "epoch": 0.23887079261672095, + "grad_norm": 2.3237954527681084, + "learning_rate": 1.998357401735982e-05, + "loss": 1.398, + "step": 440 + }, + { + "epoch": 0.23941368078175895, + "grad_norm": 2.448364938314344, + "learning_rate": 1.9983371909367135e-05, + "loss": 1.2663, + "step": 441 + }, + { + "epoch": 0.23995656894679696, + "grad_norm": 1.935899726785714, + "learning_rate": 1.99831685666151e-05, + "loss": 1.2448, + "step": 442 + }, + { + "epoch": 0.24049945711183496, + "grad_norm": 2.14357016947003, + "learning_rate": 1.9982963989128864e-05, + "loss": 1.4786, + "step": 443 + }, + { + "epoch": 0.24104234527687296, + "grad_norm": 2.5830135196717148, + "learning_rate": 1.998275817693373e-05, + "loss": 1.8324, + "step": 444 + }, + { + "epoch": 0.24158523344191096, + "grad_norm": 3.159289299734819, + "learning_rate": 1.9982551130055157e-05, + "loss": 1.9091, + "step": 445 + }, + { + "epoch": 0.24212812160694897, + "grad_norm": 3.2712239674501755, + "learning_rate": 1.9982342848518753e-05, + "loss": 1.7869, + "step": 446 + }, + { + "epoch": 0.24267100977198697, + "grad_norm": 3.35834841128443, + "learning_rate": 1.998213333235027e-05, + "loss": 1.2772, + "step": 447 + }, + { + "epoch": 0.24321389793702497, + "grad_norm": 2.9379666136057354, + "learning_rate": 1.998192258157563e-05, + "loss": 1.3497, + "step": 448 + }, + { + "epoch": 0.24375678610206297, + "grad_norm": 2.334119075845609, + "learning_rate": 1.9981710596220897e-05, + "loss": 1.5541, + "step": 449 + }, + { + "epoch": 0.24429967426710097, + "grad_norm": 2.781469856763384, + "learning_rate": 1.998149737631229e-05, + "loss": 1.9413, + "step": 450 + }, + { + "epoch": 0.24484256243213898, + "grad_norm": 3.0196781534130452, + "learning_rate": 1.9981282921876177e-05, + "loss": 1.3238, + "step": 451 + }, + { + "epoch": 0.24538545059717698, + "grad_norm": 3.547169321727429, + "learning_rate": 1.9981067232939086e-05, + "loss": 1.9952, + "step": 452 + }, + { + "epoch": 0.24592833876221498, + "grad_norm": 3.2411499530913535, + "learning_rate": 1.9980850309527693e-05, + "loss": 1.8244, + "step": 453 + }, + { + "epoch": 0.24647122692725298, + "grad_norm": 2.670313260104859, + "learning_rate": 1.9980632151668822e-05, + "loss": 1.607, + "step": 454 + }, + { + "epoch": 0.247014115092291, + "grad_norm": 2.8822100638306143, + "learning_rate": 1.9980412759389468e-05, + "loss": 1.8868, + "step": 455 + }, + { + "epoch": 0.247557003257329, + "grad_norm": 2.4776152417583317, + "learning_rate": 1.9980192132716748e-05, + "loss": 1.778, + "step": 456 + }, + { + "epoch": 0.248099891422367, + "grad_norm": 2.2001923672712076, + "learning_rate": 1.9979970271677967e-05, + "loss": 1.3544, + "step": 457 + }, + { + "epoch": 0.248642779587405, + "grad_norm": 2.7694932683911837, + "learning_rate": 1.9979747176300553e-05, + "loss": 1.6521, + "step": 458 + }, + { + "epoch": 0.249185667752443, + "grad_norm": 2.8464573075472845, + "learning_rate": 1.99795228466121e-05, + "loss": 1.8803, + "step": 459 + }, + { + "epoch": 0.249728555917481, + "grad_norm": 2.577989994947286, + "learning_rate": 1.9979297282640365e-05, + "loss": 1.8838, + "step": 460 + }, + { + "epoch": 0.250271444082519, + "grad_norm": 2.9859834231033164, + "learning_rate": 1.997907048441323e-05, + "loss": 1.5122, + "step": 461 + }, + { + "epoch": 0.250814332247557, + "grad_norm": 3.0857580735568098, + "learning_rate": 1.9978842451958757e-05, + "loss": 1.7789, + "step": 462 + }, + { + "epoch": 0.251357220412595, + "grad_norm": 2.523783848348806, + "learning_rate": 1.9978613185305145e-05, + "loss": 1.5815, + "step": 463 + }, + { + "epoch": 0.251900108577633, + "grad_norm": 2.699094362013039, + "learning_rate": 1.9978382684480747e-05, + "loss": 1.7448, + "step": 464 + }, + { + "epoch": 0.252442996742671, + "grad_norm": 2.736268550628698, + "learning_rate": 1.997815094951408e-05, + "loss": 1.7052, + "step": 465 + }, + { + "epoch": 0.252985884907709, + "grad_norm": 3.1150828428090014, + "learning_rate": 1.99779179804338e-05, + "loss": 1.7743, + "step": 466 + }, + { + "epoch": 0.253528773072747, + "grad_norm": 3.3513164619888482, + "learning_rate": 1.997768377726872e-05, + "loss": 2.3905, + "step": 467 + }, + { + "epoch": 0.254071661237785, + "grad_norm": 2.400886982379507, + "learning_rate": 1.9977448340047808e-05, + "loss": 1.6096, + "step": 468 + }, + { + "epoch": 0.254614549402823, + "grad_norm": 2.633533044966171, + "learning_rate": 1.9977211668800186e-05, + "loss": 1.7796, + "step": 469 + }, + { + "epoch": 0.255157437567861, + "grad_norm": 3.9297139007235042, + "learning_rate": 1.997697376355512e-05, + "loss": 1.4602, + "step": 470 + }, + { + "epoch": 0.255700325732899, + "grad_norm": 5.588260619709643, + "learning_rate": 1.9976734624342044e-05, + "loss": 2.0389, + "step": 471 + }, + { + "epoch": 0.256243213897937, + "grad_norm": 2.600723171476426, + "learning_rate": 1.9976494251190522e-05, + "loss": 1.3676, + "step": 472 + }, + { + "epoch": 0.25678610206297503, + "grad_norm": 2.61945002649116, + "learning_rate": 1.9976252644130297e-05, + "loss": 1.7902, + "step": 473 + }, + { + "epoch": 0.25732899022801303, + "grad_norm": 3.7513561790803838, + "learning_rate": 1.997600980319124e-05, + "loss": 1.5997, + "step": 474 + }, + { + "epoch": 0.25787187839305103, + "grad_norm": 2.8832359552778737, + "learning_rate": 1.9975765728403395e-05, + "loss": 1.9636, + "step": 475 + }, + { + "epoch": 0.25841476655808904, + "grad_norm": 3.0975018752600243, + "learning_rate": 1.9975520419796942e-05, + "loss": 1.1165, + "step": 476 + }, + { + "epoch": 0.25895765472312704, + "grad_norm": 2.121708502818221, + "learning_rate": 1.9975273877402227e-05, + "loss": 1.5108, + "step": 477 + }, + { + "epoch": 0.25950054288816504, + "grad_norm": 2.4073592870530116, + "learning_rate": 1.997502610124974e-05, + "loss": 1.7828, + "step": 478 + }, + { + "epoch": 0.26004343105320304, + "grad_norm": 2.6858679986632974, + "learning_rate": 1.997477709137013e-05, + "loss": 1.8483, + "step": 479 + }, + { + "epoch": 0.26058631921824105, + "grad_norm": 3.283752190131325, + "learning_rate": 1.997452684779419e-05, + "loss": 1.6105, + "step": 480 + }, + { + "epoch": 0.26112920738327905, + "grad_norm": 2.500181185675909, + "learning_rate": 1.997427537055287e-05, + "loss": 1.5475, + "step": 481 + }, + { + "epoch": 0.26167209554831705, + "grad_norm": 2.3992324550953885, + "learning_rate": 1.9974022659677278e-05, + "loss": 1.6062, + "step": 482 + }, + { + "epoch": 0.26221498371335505, + "grad_norm": 2.7692293822867837, + "learning_rate": 1.9973768715198667e-05, + "loss": 1.4995, + "step": 483 + }, + { + "epoch": 0.26275787187839306, + "grad_norm": 2.971423364277874, + "learning_rate": 1.9973513537148447e-05, + "loss": 1.7904, + "step": 484 + }, + { + "epoch": 0.26330076004343106, + "grad_norm": 2.2769736321644105, + "learning_rate": 1.9973257125558177e-05, + "loss": 1.3121, + "step": 485 + }, + { + "epoch": 0.26384364820846906, + "grad_norm": 2.5949046051899254, + "learning_rate": 1.997299948045957e-05, + "loss": 1.4555, + "step": 486 + }, + { + "epoch": 0.26438653637350706, + "grad_norm": 3.0210593683445204, + "learning_rate": 1.997274060188449e-05, + "loss": 1.8211, + "step": 487 + }, + { + "epoch": 0.26492942453854507, + "grad_norm": 3.076626224616319, + "learning_rate": 1.9972480489864962e-05, + "loss": 1.6366, + "step": 488 + }, + { + "epoch": 0.26547231270358307, + "grad_norm": 3.1369728365663536, + "learning_rate": 1.9972219144433148e-05, + "loss": 1.5027, + "step": 489 + }, + { + "epoch": 0.26601520086862107, + "grad_norm": 2.2870450349164635, + "learning_rate": 1.9971956565621383e-05, + "loss": 1.2784, + "step": 490 + }, + { + "epoch": 0.2665580890336591, + "grad_norm": 3.0130036065633776, + "learning_rate": 1.9971692753462134e-05, + "loss": 1.2083, + "step": 491 + }, + { + "epoch": 0.2671009771986971, + "grad_norm": 2.6063157323029733, + "learning_rate": 1.9971427707988034e-05, + "loss": 1.0083, + "step": 492 + }, + { + "epoch": 0.2676438653637351, + "grad_norm": 2.579371053895234, + "learning_rate": 1.997116142923186e-05, + "loss": 1.1937, + "step": 493 + }, + { + "epoch": 0.2681867535287731, + "grad_norm": 2.532537971800688, + "learning_rate": 1.9970893917226554e-05, + "loss": 1.4735, + "step": 494 + }, + { + "epoch": 0.2687296416938111, + "grad_norm": 2.3483150144294105, + "learning_rate": 1.997062517200519e-05, + "loss": 1.7269, + "step": 495 + }, + { + "epoch": 0.2692725298588491, + "grad_norm": 2.594809867192747, + "learning_rate": 1.997035519360102e-05, + "loss": 1.8283, + "step": 496 + }, + { + "epoch": 0.2698154180238871, + "grad_norm": 2.613813750609998, + "learning_rate": 1.9970083982047428e-05, + "loss": 1.2302, + "step": 497 + }, + { + "epoch": 0.2703583061889251, + "grad_norm": 2.648279162964909, + "learning_rate": 1.9969811537377956e-05, + "loss": 1.6225, + "step": 498 + }, + { + "epoch": 0.2709011943539631, + "grad_norm": 2.3790190706794325, + "learning_rate": 1.9969537859626308e-05, + "loss": 1.5172, + "step": 499 + }, + { + "epoch": 0.2714440825190011, + "grad_norm": 2.7054998578606364, + "learning_rate": 1.9969262948826326e-05, + "loss": 1.4525, + "step": 500 + }, + { + "epoch": 0.2719869706840391, + "grad_norm": 2.412151508264948, + "learning_rate": 1.9968986805012012e-05, + "loss": 1.3299, + "step": 501 + }, + { + "epoch": 0.2725298588490771, + "grad_norm": 2.4836460319285414, + "learning_rate": 1.9968709428217525e-05, + "loss": 1.5217, + "step": 502 + }, + { + "epoch": 0.2730727470141151, + "grad_norm": 3.3112944949184606, + "learning_rate": 1.9968430818477168e-05, + "loss": 2.0643, + "step": 503 + }, + { + "epoch": 0.2736156351791531, + "grad_norm": 3.266395629954733, + "learning_rate": 1.9968150975825397e-05, + "loss": 1.1719, + "step": 504 + }, + { + "epoch": 0.2741585233441911, + "grad_norm": 2.3487351103507073, + "learning_rate": 1.996786990029683e-05, + "loss": 1.6876, + "step": 505 + }, + { + "epoch": 0.2747014115092291, + "grad_norm": 3.0672442719402673, + "learning_rate": 1.9967587591926227e-05, + "loss": 1.3946, + "step": 506 + }, + { + "epoch": 0.2752442996742671, + "grad_norm": 2.84320621483769, + "learning_rate": 1.99673040507485e-05, + "loss": 1.3259, + "step": 507 + }, + { + "epoch": 0.2757871878393051, + "grad_norm": 2.5078449617010707, + "learning_rate": 1.9967019276798728e-05, + "loss": 0.8478, + "step": 508 + }, + { + "epoch": 0.2763300760043431, + "grad_norm": 3.137784699454796, + "learning_rate": 1.9966733270112126e-05, + "loss": 1.1688, + "step": 509 + }, + { + "epoch": 0.2768729641693811, + "grad_norm": 3.7277826491955017, + "learning_rate": 1.996644603072407e-05, + "loss": 1.1091, + "step": 510 + }, + { + "epoch": 0.2774158523344191, + "grad_norm": 2.5276781604415635, + "learning_rate": 1.996615755867008e-05, + "loss": 1.1299, + "step": 511 + }, + { + "epoch": 0.2779587404994571, + "grad_norm": 3.1012700661738744, + "learning_rate": 1.996586785398584e-05, + "loss": 1.7218, + "step": 512 + }, + { + "epoch": 0.2785016286644951, + "grad_norm": 2.285166712515903, + "learning_rate": 1.9965576916707182e-05, + "loss": 1.2868, + "step": 513 + }, + { + "epoch": 0.27904451682953313, + "grad_norm": 2.33097906349044, + "learning_rate": 1.9965284746870088e-05, + "loss": 0.9887, + "step": 514 + }, + { + "epoch": 0.27958740499457113, + "grad_norm": 2.6473787082237927, + "learning_rate": 1.9964991344510697e-05, + "loss": 1.8543, + "step": 515 + }, + { + "epoch": 0.28013029315960913, + "grad_norm": 2.4628160599533366, + "learning_rate": 1.996469670966529e-05, + "loss": 1.2263, + "step": 516 + }, + { + "epoch": 0.28067318132464714, + "grad_norm": 3.2897847068350905, + "learning_rate": 1.9964400842370314e-05, + "loss": 1.6338, + "step": 517 + }, + { + "epoch": 0.28121606948968514, + "grad_norm": 2.4439319341540324, + "learning_rate": 1.9964103742662363e-05, + "loss": 1.0836, + "step": 518 + }, + { + "epoch": 0.28175895765472314, + "grad_norm": 2.3221991020412003, + "learning_rate": 1.996380541057818e-05, + "loss": 1.2331, + "step": 519 + }, + { + "epoch": 0.28230184581976114, + "grad_norm": 2.9571040634251564, + "learning_rate": 1.9963505846154662e-05, + "loss": 1.3066, + "step": 520 + }, + { + "epoch": 0.28284473398479915, + "grad_norm": 3.7512706020225624, + "learning_rate": 1.996320504942886e-05, + "loss": 1.7482, + "step": 521 + }, + { + "epoch": 0.28338762214983715, + "grad_norm": 2.4620109793388267, + "learning_rate": 1.9962903020437983e-05, + "loss": 1.5334, + "step": 522 + }, + { + "epoch": 0.28393051031487515, + "grad_norm": 3.030374272795485, + "learning_rate": 1.9962599759219383e-05, + "loss": 1.8957, + "step": 523 + }, + { + "epoch": 0.28447339847991315, + "grad_norm": 2.452389821491403, + "learning_rate": 1.9962295265810563e-05, + "loss": 1.5438, + "step": 524 + }, + { + "epoch": 0.28501628664495116, + "grad_norm": 2.713028369466205, + "learning_rate": 1.996198954024919e-05, + "loss": 1.4272, + "step": 525 + }, + { + "epoch": 0.28555917480998916, + "grad_norm": 3.810321275175567, + "learning_rate": 1.996168258257307e-05, + "loss": 1.9028, + "step": 526 + }, + { + "epoch": 0.28610206297502716, + "grad_norm": 2.7774100977441236, + "learning_rate": 1.9961374392820173e-05, + "loss": 1.5644, + "step": 527 + }, + { + "epoch": 0.28664495114006516, + "grad_norm": 2.6798712089104186, + "learning_rate": 1.9961064971028616e-05, + "loss": 1.188, + "step": 528 + }, + { + "epoch": 0.28718783930510317, + "grad_norm": 2.889465990486677, + "learning_rate": 1.9960754317236666e-05, + "loss": 1.5393, + "step": 529 + }, + { + "epoch": 0.28773072747014117, + "grad_norm": 3.465251366831076, + "learning_rate": 1.996044243148275e-05, + "loss": 2.1899, + "step": 530 + }, + { + "epoch": 0.28827361563517917, + "grad_norm": 2.645941940974219, + "learning_rate": 1.9960129313805437e-05, + "loss": 1.3691, + "step": 531 + }, + { + "epoch": 0.2888165038002172, + "grad_norm": 2.3914199977194293, + "learning_rate": 1.9959814964243455e-05, + "loss": 1.5219, + "step": 532 + }, + { + "epoch": 0.2893593919652552, + "grad_norm": 2.3023810529281343, + "learning_rate": 1.995949938283569e-05, + "loss": 1.5147, + "step": 533 + }, + { + "epoch": 0.2899022801302932, + "grad_norm": 2.7362205671791155, + "learning_rate": 1.9959182569621164e-05, + "loss": 1.7571, + "step": 534 + }, + { + "epoch": 0.2904451682953312, + "grad_norm": 3.971162331076012, + "learning_rate": 1.9958864524639066e-05, + "loss": 1.3425, + "step": 535 + }, + { + "epoch": 0.2909880564603692, + "grad_norm": 3.4144928239616514, + "learning_rate": 1.9958545247928727e-05, + "loss": 1.6962, + "step": 536 + }, + { + "epoch": 0.2915309446254072, + "grad_norm": 3.5063126675319043, + "learning_rate": 1.9958224739529647e-05, + "loss": 1.6406, + "step": 537 + }, + { + "epoch": 0.2920738327904452, + "grad_norm": 2.9013783116047547, + "learning_rate": 1.995790299948146e-05, + "loss": 1.6376, + "step": 538 + }, + { + "epoch": 0.2926167209554832, + "grad_norm": 2.4827691033904693, + "learning_rate": 1.9957580027823957e-05, + "loss": 1.8672, + "step": 539 + }, + { + "epoch": 0.2931596091205212, + "grad_norm": 3.1245563354940242, + "learning_rate": 1.9957255824597087e-05, + "loss": 1.885, + "step": 540 + }, + { + "epoch": 0.2937024972855592, + "grad_norm": 2.7941860466759896, + "learning_rate": 1.9956930389840945e-05, + "loss": 1.0903, + "step": 541 + }, + { + "epoch": 0.2942453854505972, + "grad_norm": 3.1348104146717772, + "learning_rate": 1.9956603723595784e-05, + "loss": 2.1446, + "step": 542 + }, + { + "epoch": 0.2947882736156352, + "grad_norm": 2.4248222663066747, + "learning_rate": 1.995627582590201e-05, + "loss": 1.3503, + "step": 543 + }, + { + "epoch": 0.2953311617806732, + "grad_norm": 3.1538245864476337, + "learning_rate": 1.995594669680017e-05, + "loss": 1.3275, + "step": 544 + }, + { + "epoch": 0.2958740499457112, + "grad_norm": 3.436905240480997, + "learning_rate": 1.9955616336330976e-05, + "loss": 1.7249, + "step": 545 + }, + { + "epoch": 0.2964169381107492, + "grad_norm": 3.9145352507266393, + "learning_rate": 1.9955284744535287e-05, + "loss": 2.0089, + "step": 546 + }, + { + "epoch": 0.2969598262757872, + "grad_norm": 2.8600509325831185, + "learning_rate": 1.9954951921454113e-05, + "loss": 1.4527, + "step": 547 + }, + { + "epoch": 0.2975027144408252, + "grad_norm": 2.865760632888347, + "learning_rate": 1.995461786712862e-05, + "loss": 1.4553, + "step": 548 + }, + { + "epoch": 0.2980456026058632, + "grad_norm": 3.3559882259900706, + "learning_rate": 1.9954282581600127e-05, + "loss": 2.0456, + "step": 549 + }, + { + "epoch": 0.2985884907709012, + "grad_norm": 3.3424656149019008, + "learning_rate": 1.9953946064910098e-05, + "loss": 1.8253, + "step": 550 + }, + { + "epoch": 0.2991313789359392, + "grad_norm": 2.569119318410074, + "learning_rate": 1.9953608317100153e-05, + "loss": 1.3623, + "step": 551 + }, + { + "epoch": 0.2996742671009772, + "grad_norm": 2.8918953822102424, + "learning_rate": 1.995326933821207e-05, + "loss": 1.7521, + "step": 552 + }, + { + "epoch": 0.3002171552660152, + "grad_norm": 2.7627900710192246, + "learning_rate": 1.995292912828777e-05, + "loss": 1.775, + "step": 553 + }, + { + "epoch": 0.3007600434310532, + "grad_norm": 2.8053609300694804, + "learning_rate": 1.9952587687369334e-05, + "loss": 1.7536, + "step": 554 + }, + { + "epoch": 0.30130293159609123, + "grad_norm": 2.9775715301146803, + "learning_rate": 1.995224501549899e-05, + "loss": 1.6715, + "step": 555 + }, + { + "epoch": 0.30184581976112923, + "grad_norm": 3.146995410263436, + "learning_rate": 1.9951901112719123e-05, + "loss": 1.1032, + "step": 556 + }, + { + "epoch": 0.30238870792616723, + "grad_norm": 3.219197817112143, + "learning_rate": 1.9951555979072266e-05, + "loss": 1.6326, + "step": 557 + }, + { + "epoch": 0.30293159609120524, + "grad_norm": 2.676508071644292, + "learning_rate": 1.99512096146011e-05, + "loss": 1.4836, + "step": 558 + }, + { + "epoch": 0.30347448425624324, + "grad_norm": 3.8806900857620374, + "learning_rate": 1.9950862019348474e-05, + "loss": 1.7794, + "step": 559 + }, + { + "epoch": 0.30401737242128124, + "grad_norm": 2.547409032322543, + "learning_rate": 1.995051319335737e-05, + "loss": 1.3263, + "step": 560 + }, + { + "epoch": 0.30456026058631924, + "grad_norm": 3.025189850713409, + "learning_rate": 1.995016313667094e-05, + "loss": 1.2409, + "step": 561 + }, + { + "epoch": 0.30510314875135724, + "grad_norm": 3.3644665856402614, + "learning_rate": 1.9949811849332476e-05, + "loss": 1.2988, + "step": 562 + }, + { + "epoch": 0.30564603691639525, + "grad_norm": 3.4126932100522755, + "learning_rate": 1.9949459331385422e-05, + "loss": 1.6126, + "step": 563 + }, + { + "epoch": 0.30618892508143325, + "grad_norm": 3.3751971677759416, + "learning_rate": 1.994910558287338e-05, + "loss": 1.6243, + "step": 564 + }, + { + "epoch": 0.30673181324647125, + "grad_norm": 3.0948952637866105, + "learning_rate": 1.9948750603840102e-05, + "loss": 1.6553, + "step": 565 + }, + { + "epoch": 0.30727470141150925, + "grad_norm": 3.1116929004314224, + "learning_rate": 1.9948394394329494e-05, + "loss": 1.0466, + "step": 566 + }, + { + "epoch": 0.30781758957654726, + "grad_norm": 2.649244034262683, + "learning_rate": 1.9948036954385613e-05, + "loss": 1.2914, + "step": 567 + }, + { + "epoch": 0.30836047774158526, + "grad_norm": 3.9338023936000965, + "learning_rate": 1.9947678284052667e-05, + "loss": 1.7532, + "step": 568 + }, + { + "epoch": 0.30890336590662326, + "grad_norm": 2.887133065759567, + "learning_rate": 1.9947318383375017e-05, + "loss": 1.8001, + "step": 569 + }, + { + "epoch": 0.30944625407166126, + "grad_norm": 3.6996119456889915, + "learning_rate": 1.9946957252397173e-05, + "loss": 2.4852, + "step": 570 + }, + { + "epoch": 0.30998914223669927, + "grad_norm": 2.829855772438557, + "learning_rate": 1.9946594891163808e-05, + "loss": 1.6048, + "step": 571 + }, + { + "epoch": 0.31053203040173727, + "grad_norm": 4.288772356209683, + "learning_rate": 1.9946231299719732e-05, + "loss": 1.4841, + "step": 572 + }, + { + "epoch": 0.31107491856677527, + "grad_norm": 2.37958986026152, + "learning_rate": 1.9945866478109914e-05, + "loss": 1.1797, + "step": 573 + }, + { + "epoch": 0.3116178067318133, + "grad_norm": 2.9977970906442932, + "learning_rate": 1.9945500426379483e-05, + "loss": 1.224, + "step": 574 + }, + { + "epoch": 0.3121606948968513, + "grad_norm": 3.406596416686285, + "learning_rate": 1.9945133144573705e-05, + "loss": 1.4793, + "step": 575 + }, + { + "epoch": 0.3127035830618892, + "grad_norm": 3.3772239188722244, + "learning_rate": 1.994476463273801e-05, + "loss": 1.3696, + "step": 576 + }, + { + "epoch": 0.3132464712269272, + "grad_norm": 2.729299768057245, + "learning_rate": 1.9944394890917977e-05, + "loss": 1.448, + "step": 577 + }, + { + "epoch": 0.31378935939196523, + "grad_norm": 2.6526406591248297, + "learning_rate": 1.9944023919159335e-05, + "loss": 1.6905, + "step": 578 + }, + { + "epoch": 0.31433224755700323, + "grad_norm": 2.9512501195242944, + "learning_rate": 1.9943651717507965e-05, + "loss": 1.6277, + "step": 579 + }, + { + "epoch": 0.31487513572204123, + "grad_norm": 2.8292191327941723, + "learning_rate": 1.9943278286009903e-05, + "loss": 1.0532, + "step": 580 + }, + { + "epoch": 0.31541802388707924, + "grad_norm": 3.2105870294745436, + "learning_rate": 1.9942903624711335e-05, + "loss": 1.1823, + "step": 581 + }, + { + "epoch": 0.31596091205211724, + "grad_norm": 3.224137115744835, + "learning_rate": 1.9942527733658602e-05, + "loss": 1.5409, + "step": 582 + }, + { + "epoch": 0.31650380021715524, + "grad_norm": 3.126134123334164, + "learning_rate": 1.9942150612898194e-05, + "loss": 1.2423, + "step": 583 + }, + { + "epoch": 0.31704668838219324, + "grad_norm": 2.709954974374804, + "learning_rate": 1.994177226247675e-05, + "loss": 1.3419, + "step": 584 + }, + { + "epoch": 0.31758957654723124, + "grad_norm": 2.9030367366777927, + "learning_rate": 1.9941392682441066e-05, + "loss": 1.427, + "step": 585 + }, + { + "epoch": 0.31813246471226925, + "grad_norm": 3.272875981776567, + "learning_rate": 1.9941011872838092e-05, + "loss": 2.0196, + "step": 586 + }, + { + "epoch": 0.31867535287730725, + "grad_norm": 3.3480185179867, + "learning_rate": 1.994062983371493e-05, + "loss": 1.6038, + "step": 587 + }, + { + "epoch": 0.31921824104234525, + "grad_norm": 2.801738772846361, + "learning_rate": 1.9940246565118822e-05, + "loss": 1.7505, + "step": 588 + }, + { + "epoch": 0.31976112920738325, + "grad_norm": 2.743550967049156, + "learning_rate": 1.993986206709718e-05, + "loss": 1.3914, + "step": 589 + }, + { + "epoch": 0.32030401737242126, + "grad_norm": 2.481845489278486, + "learning_rate": 1.9939476339697555e-05, + "loss": 1.6927, + "step": 590 + }, + { + "epoch": 0.32084690553745926, + "grad_norm": 3.145019330878407, + "learning_rate": 1.993908938296765e-05, + "loss": 1.42, + "step": 591 + }, + { + "epoch": 0.32138979370249726, + "grad_norm": 3.326427662456394, + "learning_rate": 1.9938701196955335e-05, + "loss": 1.2516, + "step": 592 + }, + { + "epoch": 0.32193268186753526, + "grad_norm": 3.195815443331326, + "learning_rate": 1.9938311781708616e-05, + "loss": 2.1428, + "step": 593 + }, + { + "epoch": 0.32247557003257327, + "grad_norm": 2.461395361556941, + "learning_rate": 1.9937921137275657e-05, + "loss": 1.4448, + "step": 594 + }, + { + "epoch": 0.32301845819761127, + "grad_norm": 3.1793351214549794, + "learning_rate": 1.993752926370477e-05, + "loss": 1.4609, + "step": 595 + }, + { + "epoch": 0.32356134636264927, + "grad_norm": 2.783909288864463, + "learning_rate": 1.9937136161044427e-05, + "loss": 1.3355, + "step": 596 + }, + { + "epoch": 0.3241042345276873, + "grad_norm": 2.8156929696256734, + "learning_rate": 1.9936741829343247e-05, + "loss": 2.101, + "step": 597 + }, + { + "epoch": 0.3246471226927253, + "grad_norm": 3.793352093788154, + "learning_rate": 1.993634626865e-05, + "loss": 2.19, + "step": 598 + }, + { + "epoch": 0.3251900108577633, + "grad_norm": 2.6680863350639545, + "learning_rate": 1.993594947901361e-05, + "loss": 1.5199, + "step": 599 + }, + { + "epoch": 0.3257328990228013, + "grad_norm": 2.6498586813134297, + "learning_rate": 1.9935551460483155e-05, + "loss": 0.9282, + "step": 600 + }, + { + "epoch": 0.3262757871878393, + "grad_norm": 3.443063689484479, + "learning_rate": 1.993515221310786e-05, + "loss": 1.9107, + "step": 601 + }, + { + "epoch": 0.3268186753528773, + "grad_norm": 2.45686855619251, + "learning_rate": 1.9934751736937103e-05, + "loss": 1.2929, + "step": 602 + }, + { + "epoch": 0.3273615635179153, + "grad_norm": 3.3974155688994077, + "learning_rate": 1.9934350032020417e-05, + "loss": 1.546, + "step": 603 + }, + { + "epoch": 0.3279044516829533, + "grad_norm": 2.3692191456624783, + "learning_rate": 1.993394709840749e-05, + "loss": 1.2239, + "step": 604 + }, + { + "epoch": 0.3284473398479913, + "grad_norm": 4.044352925058249, + "learning_rate": 1.993354293614815e-05, + "loss": 1.801, + "step": 605 + }, + { + "epoch": 0.3289902280130293, + "grad_norm": 4.051192057617293, + "learning_rate": 1.993313754529239e-05, + "loss": 1.8043, + "step": 606 + }, + { + "epoch": 0.3295331161780673, + "grad_norm": 3.3680281079594634, + "learning_rate": 1.9932730925890344e-05, + "loss": 1.4915, + "step": 607 + }, + { + "epoch": 0.3300760043431053, + "grad_norm": 2.993677842102555, + "learning_rate": 1.9932323077992312e-05, + "loss": 1.3457, + "step": 608 + }, + { + "epoch": 0.3306188925081433, + "grad_norm": 3.325666927669253, + "learning_rate": 1.9931914001648726e-05, + "loss": 1.4221, + "step": 609 + }, + { + "epoch": 0.3311617806731813, + "grad_norm": 2.459382097232459, + "learning_rate": 1.993150369691019e-05, + "loss": 1.2551, + "step": 610 + }, + { + "epoch": 0.3317046688382193, + "grad_norm": 5.281638549560053, + "learning_rate": 1.993109216382745e-05, + "loss": 1.142, + "step": 611 + }, + { + "epoch": 0.3322475570032573, + "grad_norm": 2.6931033883715374, + "learning_rate": 1.99306794024514e-05, + "loss": 1.2573, + "step": 612 + }, + { + "epoch": 0.3327904451682953, + "grad_norm": 2.8066393604642714, + "learning_rate": 1.9930265412833097e-05, + "loss": 0.9847, + "step": 613 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 3.682172765988149, + "learning_rate": 1.992985019502374e-05, + "loss": 1.4482, + "step": 614 + }, + { + "epoch": 0.3338762214983713, + "grad_norm": 3.7190248655227873, + "learning_rate": 1.9929433749074684e-05, + "loss": 1.3682, + "step": 615 + }, + { + "epoch": 0.3344191096634093, + "grad_norm": 2.8594798198555527, + "learning_rate": 1.9929016075037438e-05, + "loss": 1.2396, + "step": 616 + }, + { + "epoch": 0.3349619978284473, + "grad_norm": 3.5439348644341035, + "learning_rate": 1.992859717296366e-05, + "loss": 1.5843, + "step": 617 + }, + { + "epoch": 0.3355048859934853, + "grad_norm": 2.3447575030783283, + "learning_rate": 1.992817704290516e-05, + "loss": 1.2025, + "step": 618 + }, + { + "epoch": 0.3360477741585233, + "grad_norm": 3.027039186876229, + "learning_rate": 1.99277556849139e-05, + "loss": 1.7133, + "step": 619 + }, + { + "epoch": 0.33659066232356133, + "grad_norm": 2.9868315770124343, + "learning_rate": 1.9927333099041992e-05, + "loss": 1.2309, + "step": 620 + }, + { + "epoch": 0.33713355048859933, + "grad_norm": 3.215889472723888, + "learning_rate": 1.9926909285341705e-05, + "loss": 1.2217, + "step": 621 + }, + { + "epoch": 0.33767643865363733, + "grad_norm": 3.293119071439456, + "learning_rate": 1.992648424386546e-05, + "loss": 1.7828, + "step": 622 + }, + { + "epoch": 0.33821932681867534, + "grad_norm": 2.7023629491691623, + "learning_rate": 1.992605797466582e-05, + "loss": 1.3745, + "step": 623 + }, + { + "epoch": 0.33876221498371334, + "grad_norm": 4.234534944070885, + "learning_rate": 1.9925630477795514e-05, + "loss": 1.3415, + "step": 624 + }, + { + "epoch": 0.33930510314875134, + "grad_norm": 3.6994706259321815, + "learning_rate": 1.9925201753307406e-05, + "loss": 1.4234, + "step": 625 + }, + { + "epoch": 0.33984799131378934, + "grad_norm": 3.156992028368943, + "learning_rate": 1.992477180125453e-05, + "loss": 1.4845, + "step": 626 + }, + { + "epoch": 0.34039087947882735, + "grad_norm": 3.1286393814611397, + "learning_rate": 1.9924340621690058e-05, + "loss": 0.9685, + "step": 627 + }, + { + "epoch": 0.34093376764386535, + "grad_norm": 3.336970766050665, + "learning_rate": 1.9923908214667323e-05, + "loss": 1.3268, + "step": 628 + }, + { + "epoch": 0.34147665580890335, + "grad_norm": 4.290496076646803, + "learning_rate": 1.99234745802398e-05, + "loss": 1.2785, + "step": 629 + }, + { + "epoch": 0.34201954397394135, + "grad_norm": 2.7491777959300023, + "learning_rate": 1.9923039718461127e-05, + "loss": 1.3798, + "step": 630 + }, + { + "epoch": 0.34256243213897936, + "grad_norm": 5.312529540414799, + "learning_rate": 1.992260362938509e-05, + "loss": 2.2512, + "step": 631 + }, + { + "epoch": 0.34310532030401736, + "grad_norm": 3.9557409230221676, + "learning_rate": 1.9922166313065618e-05, + "loss": 1.849, + "step": 632 + }, + { + "epoch": 0.34364820846905536, + "grad_norm": 3.7327574664165786, + "learning_rate": 1.9921727769556805e-05, + "loss": 1.4078, + "step": 633 + }, + { + "epoch": 0.34419109663409336, + "grad_norm": 3.7644055867642594, + "learning_rate": 1.992128799891289e-05, + "loss": 1.714, + "step": 634 + }, + { + "epoch": 0.34473398479913137, + "grad_norm": 3.923952524278675, + "learning_rate": 1.9920847001188258e-05, + "loss": 1.7613, + "step": 635 + }, + { + "epoch": 0.34527687296416937, + "grad_norm": 3.1523646374054155, + "learning_rate": 1.992040477643746e-05, + "loss": 1.2563, + "step": 636 + }, + { + "epoch": 0.34581976112920737, + "grad_norm": 3.6157151039083444, + "learning_rate": 1.991996132471519e-05, + "loss": 1.1795, + "step": 637 + }, + { + "epoch": 0.3463626492942454, + "grad_norm": 4.222007576307266, + "learning_rate": 1.991951664607629e-05, + "loss": 1.6605, + "step": 638 + }, + { + "epoch": 0.3469055374592834, + "grad_norm": 3.1555785318795544, + "learning_rate": 1.9919070740575764e-05, + "loss": 1.526, + "step": 639 + }, + { + "epoch": 0.3474484256243214, + "grad_norm": 3.2645269107818677, + "learning_rate": 1.991862360826876e-05, + "loss": 1.3182, + "step": 640 + }, + { + "epoch": 0.3479913137893594, + "grad_norm": 3.53044955128402, + "learning_rate": 1.991817524921058e-05, + "loss": 1.5583, + "step": 641 + }, + { + "epoch": 0.3485342019543974, + "grad_norm": 3.24710260787145, + "learning_rate": 1.9917725663456678e-05, + "loss": 0.9878, + "step": 642 + }, + { + "epoch": 0.3490770901194354, + "grad_norm": 3.9154742975791654, + "learning_rate": 1.991727485106266e-05, + "loss": 1.6034, + "step": 643 + }, + { + "epoch": 0.3496199782844734, + "grad_norm": 3.8199965805178318, + "learning_rate": 1.9916822812084282e-05, + "loss": 1.3768, + "step": 644 + }, + { + "epoch": 0.3501628664495114, + "grad_norm": 3.198793670499035, + "learning_rate": 1.9916369546577455e-05, + "loss": 1.4308, + "step": 645 + }, + { + "epoch": 0.3507057546145494, + "grad_norm": 2.908177540585374, + "learning_rate": 1.9915915054598237e-05, + "loss": 1.3964, + "step": 646 + }, + { + "epoch": 0.3512486427795874, + "grad_norm": 2.920889003436479, + "learning_rate": 1.9915459336202844e-05, + "loss": 1.056, + "step": 647 + }, + { + "epoch": 0.3517915309446254, + "grad_norm": 3.7046400158340864, + "learning_rate": 1.991500239144763e-05, + "loss": 1.9052, + "step": 648 + }, + { + "epoch": 0.3523344191096634, + "grad_norm": 4.412988121206581, + "learning_rate": 1.9914544220389124e-05, + "loss": 1.944, + "step": 649 + }, + { + "epoch": 0.3528773072747014, + "grad_norm": 3.87124697251994, + "learning_rate": 1.9914084823083988e-05, + "loss": 1.4951, + "step": 650 + }, + { + "epoch": 0.3534201954397394, + "grad_norm": 3.8734663848108584, + "learning_rate": 1.9913624199589037e-05, + "loss": 1.5462, + "step": 651 + }, + { + "epoch": 0.3539630836047774, + "grad_norm": 2.7283359997443126, + "learning_rate": 1.9913162349961248e-05, + "loss": 0.9188, + "step": 652 + }, + { + "epoch": 0.3545059717698154, + "grad_norm": 2.885333878853398, + "learning_rate": 1.991269927425774e-05, + "loss": 1.2292, + "step": 653 + }, + { + "epoch": 0.3550488599348534, + "grad_norm": 3.2425772460053257, + "learning_rate": 1.9912234972535788e-05, + "loss": 1.2863, + "step": 654 + }, + { + "epoch": 0.3555917480998914, + "grad_norm": 3.060966260943164, + "learning_rate": 1.991176944485281e-05, + "loss": 1.5056, + "step": 655 + }, + { + "epoch": 0.3561346362649294, + "grad_norm": 4.533553635387312, + "learning_rate": 1.99113026912664e-05, + "loss": 1.8229, + "step": 656 + }, + { + "epoch": 0.3566775244299674, + "grad_norm": 2.8582150809054045, + "learning_rate": 1.9910834711834267e-05, + "loss": 1.3233, + "step": 657 + }, + { + "epoch": 0.3572204125950054, + "grad_norm": 4.930680768753159, + "learning_rate": 1.9910365506614308e-05, + "loss": 1.5997, + "step": 658 + }, + { + "epoch": 0.3577633007600434, + "grad_norm": 3.0733335483268083, + "learning_rate": 1.9909895075664545e-05, + "loss": 0.9206, + "step": 659 + }, + { + "epoch": 0.3583061889250814, + "grad_norm": 3.4659704922712686, + "learning_rate": 1.990942341904317e-05, + "loss": 1.2752, + "step": 660 + }, + { + "epoch": 0.35884907709011943, + "grad_norm": 3.493850558129149, + "learning_rate": 1.9908950536808508e-05, + "loss": 1.588, + "step": 661 + }, + { + "epoch": 0.35939196525515743, + "grad_norm": 3.61256229272583, + "learning_rate": 1.9908476429019056e-05, + "loss": 1.469, + "step": 662 + }, + { + "epoch": 0.35993485342019543, + "grad_norm": 3.3430037514089803, + "learning_rate": 1.9908001095733445e-05, + "loss": 1.3038, + "step": 663 + }, + { + "epoch": 0.36047774158523344, + "grad_norm": 3.8918915484374557, + "learning_rate": 1.9907524537010467e-05, + "loss": 1.3683, + "step": 664 + }, + { + "epoch": 0.36102062975027144, + "grad_norm": 3.6032272045245053, + "learning_rate": 1.9907046752909064e-05, + "loss": 1.8694, + "step": 665 + }, + { + "epoch": 0.36156351791530944, + "grad_norm": 4.546268377661869, + "learning_rate": 1.9906567743488326e-05, + "loss": 1.2871, + "step": 666 + }, + { + "epoch": 0.36210640608034744, + "grad_norm": 3.5283169402860777, + "learning_rate": 1.9906087508807504e-05, + "loss": 1.5334, + "step": 667 + }, + { + "epoch": 0.36264929424538545, + "grad_norm": 2.7984191262784504, + "learning_rate": 1.9905606048925993e-05, + "loss": 1.1924, + "step": 668 + }, + { + "epoch": 0.36319218241042345, + "grad_norm": 3.54703835950394, + "learning_rate": 1.9905123363903335e-05, + "loss": 1.5972, + "step": 669 + }, + { + "epoch": 0.36373507057546145, + "grad_norm": 3.0126938064772855, + "learning_rate": 1.9904639453799236e-05, + "loss": 1.2294, + "step": 670 + }, + { + "epoch": 0.36427795874049945, + "grad_norm": 3.16358553989926, + "learning_rate": 1.990415431867354e-05, + "loss": 1.3564, + "step": 671 + }, + { + "epoch": 0.36482084690553745, + "grad_norm": 4.118910539945259, + "learning_rate": 1.990366795858626e-05, + "loss": 1.6838, + "step": 672 + }, + { + "epoch": 0.36536373507057546, + "grad_norm": 3.8466533902335502, + "learning_rate": 1.9903180373597534e-05, + "loss": 1.7986, + "step": 673 + }, + { + "epoch": 0.36590662323561346, + "grad_norm": 3.6195384409794684, + "learning_rate": 1.990269156376768e-05, + "loss": 1.6113, + "step": 674 + }, + { + "epoch": 0.36644951140065146, + "grad_norm": 3.509724379422402, + "learning_rate": 1.9902201529157152e-05, + "loss": 1.5496, + "step": 675 + }, + { + "epoch": 0.36699239956568946, + "grad_norm": 3.579737463592409, + "learning_rate": 1.9901710269826554e-05, + "loss": 1.4856, + "step": 676 + }, + { + "epoch": 0.36753528773072747, + "grad_norm": 3.780431786449365, + "learning_rate": 1.9901217785836655e-05, + "loss": 1.8519, + "step": 677 + }, + { + "epoch": 0.36807817589576547, + "grad_norm": 4.0659352627131735, + "learning_rate": 1.9900724077248354e-05, + "loss": 1.4382, + "step": 678 + }, + { + "epoch": 0.36862106406080347, + "grad_norm": 2.9524656376446274, + "learning_rate": 1.9900229144122723e-05, + "loss": 1.4541, + "step": 679 + }, + { + "epoch": 0.3691639522258415, + "grad_norm": 2.388269642575342, + "learning_rate": 1.989973298652097e-05, + "loss": 0.9567, + "step": 680 + }, + { + "epoch": 0.3697068403908795, + "grad_norm": 2.8322419431529453, + "learning_rate": 1.9899235604504467e-05, + "loss": 1.1187, + "step": 681 + }, + { + "epoch": 0.3702497285559175, + "grad_norm": 3.0469435898757613, + "learning_rate": 1.9898736998134726e-05, + "loss": 1.5571, + "step": 682 + }, + { + "epoch": 0.3707926167209555, + "grad_norm": 3.623982705749655, + "learning_rate": 1.9898237167473416e-05, + "loss": 1.2047, + "step": 683 + }, + { + "epoch": 0.3713355048859935, + "grad_norm": 3.669362267695381, + "learning_rate": 1.9897736112582357e-05, + "loss": 1.7747, + "step": 684 + }, + { + "epoch": 0.3718783930510315, + "grad_norm": 2.9109325710576353, + "learning_rate": 1.989723383352352e-05, + "loss": 1.0044, + "step": 685 + }, + { + "epoch": 0.3724212812160695, + "grad_norm": 3.4478434843760146, + "learning_rate": 1.9896730330359032e-05, + "loss": 1.1245, + "step": 686 + }, + { + "epoch": 0.3729641693811075, + "grad_norm": 3.2957514959337275, + "learning_rate": 1.989622560315116e-05, + "loss": 1.3243, + "step": 687 + }, + { + "epoch": 0.3735070575461455, + "grad_norm": 3.5500423086355988, + "learning_rate": 1.989571965196234e-05, + "loss": 2.0478, + "step": 688 + }, + { + "epoch": 0.3740499457111835, + "grad_norm": 2.621987663839103, + "learning_rate": 1.9895212476855136e-05, + "loss": 1.3135, + "step": 689 + }, + { + "epoch": 0.3745928338762215, + "grad_norm": 4.055390555970487, + "learning_rate": 1.989470407789228e-05, + "loss": 1.5165, + "step": 690 + }, + { + "epoch": 0.3751357220412595, + "grad_norm": 2.94204110816024, + "learning_rate": 1.989419445513666e-05, + "loss": 1.3458, + "step": 691 + }, + { + "epoch": 0.3756786102062975, + "grad_norm": 3.5714132672245977, + "learning_rate": 1.98936836086513e-05, + "loss": 1.3034, + "step": 692 + }, + { + "epoch": 0.3762214983713355, + "grad_norm": 3.659137782783938, + "learning_rate": 1.9893171538499382e-05, + "loss": 1.6203, + "step": 693 + }, + { + "epoch": 0.3767643865363735, + "grad_norm": 5.116097300755018, + "learning_rate": 1.9892658244744236e-05, + "loss": 2.1071, + "step": 694 + }, + { + "epoch": 0.3773072747014115, + "grad_norm": 2.8449662018307005, + "learning_rate": 1.9892143727449357e-05, + "loss": 1.0477, + "step": 695 + }, + { + "epoch": 0.3778501628664495, + "grad_norm": 4.021030963638319, + "learning_rate": 1.989162798667838e-05, + "loss": 1.9528, + "step": 696 + }, + { + "epoch": 0.3783930510314875, + "grad_norm": 3.377368666687089, + "learning_rate": 1.989111102249508e-05, + "loss": 1.2481, + "step": 697 + }, + { + "epoch": 0.3789359391965255, + "grad_norm": 3.403268816169458, + "learning_rate": 1.9890592834963406e-05, + "loss": 1.1864, + "step": 698 + }, + { + "epoch": 0.3794788273615635, + "grad_norm": 3.082879601892987, + "learning_rate": 1.9890073424147453e-05, + "loss": 1.7365, + "step": 699 + }, + { + "epoch": 0.3800217155266015, + "grad_norm": 4.122610250215441, + "learning_rate": 1.988955279011145e-05, + "loss": 1.4139, + "step": 700 + }, + { + "epoch": 0.3805646036916395, + "grad_norm": 3.298087120576001, + "learning_rate": 1.98890309329198e-05, + "loss": 1.5904, + "step": 701 + }, + { + "epoch": 0.3811074918566775, + "grad_norm": 3.1183850173888703, + "learning_rate": 1.9888507852637043e-05, + "loss": 1.2146, + "step": 702 + }, + { + "epoch": 0.38165038002171553, + "grad_norm": 3.128042971411966, + "learning_rate": 1.9887983549327873e-05, + "loss": 1.2751, + "step": 703 + }, + { + "epoch": 0.38219326818675353, + "grad_norm": 3.362555156234201, + "learning_rate": 1.988745802305714e-05, + "loss": 1.1363, + "step": 704 + }, + { + "epoch": 0.38273615635179153, + "grad_norm": 3.29574344438116, + "learning_rate": 1.988693127388984e-05, + "loss": 1.227, + "step": 705 + }, + { + "epoch": 0.38327904451682954, + "grad_norm": 4.2418853227437205, + "learning_rate": 1.9886403301891123e-05, + "loss": 1.7091, + "step": 706 + }, + { + "epoch": 0.38382193268186754, + "grad_norm": 4.165071133964158, + "learning_rate": 1.9885874107126287e-05, + "loss": 1.9403, + "step": 707 + }, + { + "epoch": 0.38436482084690554, + "grad_norm": 4.639609386348944, + "learning_rate": 1.9885343689660787e-05, + "loss": 1.221, + "step": 708 + }, + { + "epoch": 0.38490770901194354, + "grad_norm": 3.0746775894794034, + "learning_rate": 1.9884812049560226e-05, + "loss": 1.4958, + "step": 709 + }, + { + "epoch": 0.38545059717698155, + "grad_norm": 4.822804859325608, + "learning_rate": 1.9884279186890357e-05, + "loss": 1.6928, + "step": 710 + }, + { + "epoch": 0.38599348534201955, + "grad_norm": 4.299811491105524, + "learning_rate": 1.9883745101717084e-05, + "loss": 1.3075, + "step": 711 + }, + { + "epoch": 0.38653637350705755, + "grad_norm": 3.882170994492157, + "learning_rate": 1.9883209794106464e-05, + "loss": 1.9712, + "step": 712 + }, + { + "epoch": 0.38707926167209555, + "grad_norm": 4.164019029255246, + "learning_rate": 1.9882673264124705e-05, + "loss": 1.6109, + "step": 713 + }, + { + "epoch": 0.38762214983713356, + "grad_norm": 3.691226317297577, + "learning_rate": 1.9882135511838167e-05, + "loss": 1.3225, + "step": 714 + }, + { + "epoch": 0.38816503800217156, + "grad_norm": 5.777282505586183, + "learning_rate": 1.988159653731336e-05, + "loss": 1.7945, + "step": 715 + }, + { + "epoch": 0.38870792616720956, + "grad_norm": 3.3926512133724165, + "learning_rate": 1.9881056340616944e-05, + "loss": 1.5797, + "step": 716 + }, + { + "epoch": 0.38925081433224756, + "grad_norm": 3.7756965718668467, + "learning_rate": 1.988051492181573e-05, + "loss": 1.5515, + "step": 717 + }, + { + "epoch": 0.38979370249728557, + "grad_norm": 3.5509658648264613, + "learning_rate": 1.987997228097668e-05, + "loss": 1.2004, + "step": 718 + }, + { + "epoch": 0.39033659066232357, + "grad_norm": 4.682493032945398, + "learning_rate": 1.987942841816692e-05, + "loss": 1.2957, + "step": 719 + }, + { + "epoch": 0.39087947882736157, + "grad_norm": 2.923319971884201, + "learning_rate": 1.9878883333453704e-05, + "loss": 0.753, + "step": 720 + }, + { + "epoch": 0.3914223669923996, + "grad_norm": 2.8706465146507845, + "learning_rate": 1.987833702690445e-05, + "loss": 1.1057, + "step": 721 + }, + { + "epoch": 0.3919652551574376, + "grad_norm": 3.4516078574834057, + "learning_rate": 1.987778949858673e-05, + "loss": 1.1275, + "step": 722 + }, + { + "epoch": 0.3925081433224756, + "grad_norm": 3.3400553957950567, + "learning_rate": 1.9877240748568263e-05, + "loss": 1.1538, + "step": 723 + }, + { + "epoch": 0.3930510314875136, + "grad_norm": 3.155540544638446, + "learning_rate": 1.987669077691692e-05, + "loss": 1.2486, + "step": 724 + }, + { + "epoch": 0.3935939196525516, + "grad_norm": 3.1450276486107054, + "learning_rate": 1.987613958370072e-05, + "loss": 1.265, + "step": 725 + }, + { + "epoch": 0.3941368078175896, + "grad_norm": 3.0493327058864406, + "learning_rate": 1.9875587168987834e-05, + "loss": 1.2097, + "step": 726 + }, + { + "epoch": 0.3946796959826276, + "grad_norm": 3.716259137065701, + "learning_rate": 1.987503353284659e-05, + "loss": 1.5386, + "step": 727 + }, + { + "epoch": 0.3952225841476656, + "grad_norm": 3.441725300410296, + "learning_rate": 1.9874478675345458e-05, + "loss": 1.3936, + "step": 728 + }, + { + "epoch": 0.3957654723127036, + "grad_norm": 3.773353353497139, + "learning_rate": 1.9873922596553067e-05, + "loss": 1.731, + "step": 729 + }, + { + "epoch": 0.3963083604777416, + "grad_norm": 3.527645597828566, + "learning_rate": 1.987336529653819e-05, + "loss": 1.708, + "step": 730 + }, + { + "epoch": 0.3968512486427796, + "grad_norm": 3.8440953147603643, + "learning_rate": 1.9872806775369762e-05, + "loss": 1.6102, + "step": 731 + }, + { + "epoch": 0.3973941368078176, + "grad_norm": 3.4036622029265966, + "learning_rate": 1.9872247033116855e-05, + "loss": 0.9298, + "step": 732 + }, + { + "epoch": 0.3979370249728556, + "grad_norm": 4.034027632823603, + "learning_rate": 1.98716860698487e-05, + "loss": 1.78, + "step": 733 + }, + { + "epoch": 0.3984799131378936, + "grad_norm": 4.10524690369694, + "learning_rate": 1.987112388563468e-05, + "loss": 1.2354, + "step": 734 + }, + { + "epoch": 0.3990228013029316, + "grad_norm": 4.059289261823474, + "learning_rate": 1.9870560480544325e-05, + "loss": 1.804, + "step": 735 + }, + { + "epoch": 0.3995656894679696, + "grad_norm": 3.2544839104292667, + "learning_rate": 1.986999585464732e-05, + "loss": 1.1228, + "step": 736 + }, + { + "epoch": 0.4001085776330076, + "grad_norm": 3.741117309135591, + "learning_rate": 1.9869430008013496e-05, + "loss": 1.329, + "step": 737 + }, + { + "epoch": 0.4006514657980456, + "grad_norm": 4.446330579981585, + "learning_rate": 1.9868862940712838e-05, + "loss": 1.6506, + "step": 738 + }, + { + "epoch": 0.4011943539630836, + "grad_norm": 3.139296317304318, + "learning_rate": 1.9868294652815483e-05, + "loss": 1.144, + "step": 739 + }, + { + "epoch": 0.4017372421281216, + "grad_norm": 4.025270331554587, + "learning_rate": 1.986772514439172e-05, + "loss": 1.336, + "step": 740 + }, + { + "epoch": 0.4022801302931596, + "grad_norm": 3.4545518889756557, + "learning_rate": 1.986715441551198e-05, + "loss": 1.6006, + "step": 741 + }, + { + "epoch": 0.4028230184581976, + "grad_norm": 3.748598610621174, + "learning_rate": 1.986658246624686e-05, + "loss": 1.4812, + "step": 742 + }, + { + "epoch": 0.4033659066232356, + "grad_norm": 5.126058311631635, + "learning_rate": 1.9866009296667093e-05, + "loss": 2.0129, + "step": 743 + }, + { + "epoch": 0.40390879478827363, + "grad_norm": 8.145692299678405, + "learning_rate": 1.9865434906843574e-05, + "loss": 1.8321, + "step": 744 + }, + { + "epoch": 0.40445168295331163, + "grad_norm": 4.23918362592579, + "learning_rate": 1.9864859296847343e-05, + "loss": 1.5765, + "step": 745 + }, + { + "epoch": 0.40499457111834963, + "grad_norm": 4.85145039296263, + "learning_rate": 1.986428246674959e-05, + "loss": 1.6486, + "step": 746 + }, + { + "epoch": 0.40553745928338764, + "grad_norm": 3.7681072947070415, + "learning_rate": 1.986370441662166e-05, + "loss": 1.4466, + "step": 747 + }, + { + "epoch": 0.40608034744842564, + "grad_norm": 4.206136090938125, + "learning_rate": 1.986312514653505e-05, + "loss": 1.6717, + "step": 748 + }, + { + "epoch": 0.40662323561346364, + "grad_norm": 3.7336223866799965, + "learning_rate": 1.9862544656561403e-05, + "loss": 0.9599, + "step": 749 + }, + { + "epoch": 0.40716612377850164, + "grad_norm": 4.037759008136602, + "learning_rate": 1.986196294677251e-05, + "loss": 1.9387, + "step": 750 + }, + { + "epoch": 0.40770901194353965, + "grad_norm": 3.8832525632606156, + "learning_rate": 1.9861380017240324e-05, + "loss": 2.0184, + "step": 751 + }, + { + "epoch": 0.40825190010857765, + "grad_norm": 3.9521376069923546, + "learning_rate": 1.986079586803694e-05, + "loss": 1.0926, + "step": 752 + }, + { + "epoch": 0.40879478827361565, + "grad_norm": 5.128973903987411, + "learning_rate": 1.986021049923461e-05, + "loss": 1.5943, + "step": 753 + }, + { + "epoch": 0.40933767643865365, + "grad_norm": 4.5372892013581865, + "learning_rate": 1.9859623910905728e-05, + "loss": 1.643, + "step": 754 + }, + { + "epoch": 0.40988056460369166, + "grad_norm": 4.2137978897196415, + "learning_rate": 1.985903610312285e-05, + "loss": 1.7237, + "step": 755 + }, + { + "epoch": 0.41042345276872966, + "grad_norm": 2.713265124864733, + "learning_rate": 1.985844707595867e-05, + "loss": 1.0521, + "step": 756 + }, + { + "epoch": 0.41096634093376766, + "grad_norm": 3.362872283003281, + "learning_rate": 1.9857856829486045e-05, + "loss": 1.2879, + "step": 757 + }, + { + "epoch": 0.41150922909880566, + "grad_norm": 4.022235417044952, + "learning_rate": 1.9857265363777975e-05, + "loss": 1.398, + "step": 758 + }, + { + "epoch": 0.41205211726384366, + "grad_norm": 4.083280210123792, + "learning_rate": 1.9856672678907616e-05, + "loss": 1.6652, + "step": 759 + }, + { + "epoch": 0.41259500542888167, + "grad_norm": 4.469771898345589, + "learning_rate": 1.985607877494827e-05, + "loss": 1.6242, + "step": 760 + }, + { + "epoch": 0.41313789359391967, + "grad_norm": 3.0490808076449016, + "learning_rate": 1.9855483651973396e-05, + "loss": 1.0783, + "step": 761 + }, + { + "epoch": 0.41368078175895767, + "grad_norm": 3.488230005104396, + "learning_rate": 1.9854887310056593e-05, + "loss": 1.3953, + "step": 762 + }, + { + "epoch": 0.4142236699239957, + "grad_norm": 3.5610333355465653, + "learning_rate": 1.9854289749271624e-05, + "loss": 1.3692, + "step": 763 + }, + { + "epoch": 0.4147665580890337, + "grad_norm": 3.5572639096784533, + "learning_rate": 1.9853690969692393e-05, + "loss": 1.646, + "step": 764 + }, + { + "epoch": 0.4153094462540717, + "grad_norm": 3.811741925703646, + "learning_rate": 1.9853090971392953e-05, + "loss": 1.6637, + "step": 765 + }, + { + "epoch": 0.4158523344191097, + "grad_norm": 2.912609699248709, + "learning_rate": 1.9852489754447526e-05, + "loss": 0.9086, + "step": 766 + }, + { + "epoch": 0.4163952225841477, + "grad_norm": 3.7608762982374646, + "learning_rate": 1.985188731893046e-05, + "loss": 1.5131, + "step": 767 + }, + { + "epoch": 0.4169381107491857, + "grad_norm": 3.776000457813264, + "learning_rate": 1.985128366491627e-05, + "loss": 1.5929, + "step": 768 + }, + { + "epoch": 0.4174809989142237, + "grad_norm": 4.129873624081247, + "learning_rate": 1.9850678792479613e-05, + "loss": 1.5461, + "step": 769 + }, + { + "epoch": 0.4180238870792617, + "grad_norm": 3.2233425858963645, + "learning_rate": 1.9850072701695306e-05, + "loss": 0.9614, + "step": 770 + }, + { + "epoch": 0.4185667752442997, + "grad_norm": 3.8215876727885054, + "learning_rate": 1.984946539263831e-05, + "loss": 1.2549, + "step": 771 + }, + { + "epoch": 0.4191096634093377, + "grad_norm": 3.0062072101032125, + "learning_rate": 1.9848856865383732e-05, + "loss": 0.9796, + "step": 772 + }, + { + "epoch": 0.4196525515743757, + "grad_norm": 2.937513473620578, + "learning_rate": 1.984824712000684e-05, + "loss": 1.0204, + "step": 773 + }, + { + "epoch": 0.4201954397394137, + "grad_norm": 2.949279428268976, + "learning_rate": 1.984763615658305e-05, + "loss": 1.332, + "step": 774 + }, + { + "epoch": 0.4207383279044517, + "grad_norm": 3.455520053658832, + "learning_rate": 1.9847023975187925e-05, + "loss": 1.3786, + "step": 775 + }, + { + "epoch": 0.4212812160694897, + "grad_norm": 3.8352233275650733, + "learning_rate": 1.9846410575897183e-05, + "loss": 1.4897, + "step": 776 + }, + { + "epoch": 0.4218241042345277, + "grad_norm": 3.262791090623995, + "learning_rate": 1.984579595878669e-05, + "loss": 1.5097, + "step": 777 + }, + { + "epoch": 0.4223669923995657, + "grad_norm": 3.603809745585289, + "learning_rate": 1.9845180123932456e-05, + "loss": 1.8419, + "step": 778 + }, + { + "epoch": 0.4229098805646037, + "grad_norm": 3.9648489260279196, + "learning_rate": 1.9844563071410656e-05, + "loss": 1.5159, + "step": 779 + }, + { + "epoch": 0.4234527687296417, + "grad_norm": 4.576987136894564, + "learning_rate": 1.9843944801297605e-05, + "loss": 1.6866, + "step": 780 + }, + { + "epoch": 0.4239956568946797, + "grad_norm": 3.4164331578089406, + "learning_rate": 1.9843325313669774e-05, + "loss": 1.5533, + "step": 781 + }, + { + "epoch": 0.4245385450597177, + "grad_norm": 3.2121441901059367, + "learning_rate": 1.9842704608603774e-05, + "loss": 0.8879, + "step": 782 + }, + { + "epoch": 0.4250814332247557, + "grad_norm": 3.6213877563860537, + "learning_rate": 1.9842082686176388e-05, + "loss": 1.3098, + "step": 783 + }, + { + "epoch": 0.4256243213897937, + "grad_norm": 3.719993086277978, + "learning_rate": 1.9841459546464527e-05, + "loss": 1.331, + "step": 784 + }, + { + "epoch": 0.4261672095548317, + "grad_norm": 2.7952357248876516, + "learning_rate": 1.9840835189545266e-05, + "loss": 1.1298, + "step": 785 + }, + { + "epoch": 0.42671009771986973, + "grad_norm": 3.450597586563723, + "learning_rate": 1.9840209615495822e-05, + "loss": 1.0023, + "step": 786 + }, + { + "epoch": 0.42725298588490773, + "grad_norm": 4.338480682869939, + "learning_rate": 1.983958282439357e-05, + "loss": 1.7044, + "step": 787 + }, + { + "epoch": 0.42779587404994573, + "grad_norm": 3.4831324579282765, + "learning_rate": 1.983895481631603e-05, + "loss": 1.4948, + "step": 788 + }, + { + "epoch": 0.42833876221498374, + "grad_norm": 6.382810905837672, + "learning_rate": 1.9838325591340885e-05, + "loss": 2.3008, + "step": 789 + }, + { + "epoch": 0.42888165038002174, + "grad_norm": 3.063848408601439, + "learning_rate": 1.9837695149545945e-05, + "loss": 1.2045, + "step": 790 + }, + { + "epoch": 0.42942453854505974, + "grad_norm": 4.80361465821882, + "learning_rate": 1.9837063491009193e-05, + "loss": 1.2552, + "step": 791 + }, + { + "epoch": 0.42996742671009774, + "grad_norm": 4.16884813133452, + "learning_rate": 1.9836430615808745e-05, + "loss": 1.6618, + "step": 792 + }, + { + "epoch": 0.43051031487513575, + "grad_norm": 4.215747683493222, + "learning_rate": 1.9835796524022886e-05, + "loss": 1.2571, + "step": 793 + }, + { + "epoch": 0.43105320304017375, + "grad_norm": 4.894853560761769, + "learning_rate": 1.9835161215730038e-05, + "loss": 0.793, + "step": 794 + }, + { + "epoch": 0.43159609120521175, + "grad_norm": 3.264854272979086, + "learning_rate": 1.983452469100877e-05, + "loss": 1.0721, + "step": 795 + }, + { + "epoch": 0.43213897937024975, + "grad_norm": 3.2346237132359335, + "learning_rate": 1.9833886949937823e-05, + "loss": 1.2396, + "step": 796 + }, + { + "epoch": 0.43268186753528776, + "grad_norm": 3.7265266370894023, + "learning_rate": 1.9833247992596058e-05, + "loss": 1.8845, + "step": 797 + }, + { + "epoch": 0.43322475570032576, + "grad_norm": 3.0655010730082126, + "learning_rate": 1.9832607819062513e-05, + "loss": 1.556, + "step": 798 + }, + { + "epoch": 0.43376764386536376, + "grad_norm": 3.209417077142042, + "learning_rate": 1.983196642941636e-05, + "loss": 1.2183, + "step": 799 + }, + { + "epoch": 0.43431053203040176, + "grad_norm": 4.384697511143271, + "learning_rate": 1.9831323823736933e-05, + "loss": 2.0889, + "step": 800 + }, + { + "epoch": 0.43485342019543977, + "grad_norm": 3.2773759507081084, + "learning_rate": 1.9830680002103703e-05, + "loss": 0.919, + "step": 801 + }, + { + "epoch": 0.43539630836047777, + "grad_norm": 4.1436406822638805, + "learning_rate": 1.9830034964596304e-05, + "loss": 1.7759, + "step": 802 + }, + { + "epoch": 0.43593919652551577, + "grad_norm": 5.715436881791243, + "learning_rate": 1.9829388711294512e-05, + "loss": 1.4277, + "step": 803 + }, + { + "epoch": 0.4364820846905538, + "grad_norm": 3.166985124043999, + "learning_rate": 1.982874124227826e-05, + "loss": 1.56, + "step": 804 + }, + { + "epoch": 0.4370249728555918, + "grad_norm": 4.237249076660815, + "learning_rate": 1.9828092557627626e-05, + "loss": 1.2366, + "step": 805 + }, + { + "epoch": 0.4375678610206297, + "grad_norm": 3.073398749866342, + "learning_rate": 1.982744265742284e-05, + "loss": 1.1475, + "step": 806 + }, + { + "epoch": 0.4381107491856677, + "grad_norm": 3.356777818996544, + "learning_rate": 1.9826791541744285e-05, + "loss": 1.2666, + "step": 807 + }, + { + "epoch": 0.4386536373507057, + "grad_norm": 3.171383453112648, + "learning_rate": 1.982613921067249e-05, + "loss": 1.1407, + "step": 808 + }, + { + "epoch": 0.43919652551574373, + "grad_norm": 3.334493176451811, + "learning_rate": 1.9825485664288138e-05, + "loss": 1.3309, + "step": 809 + }, + { + "epoch": 0.43973941368078173, + "grad_norm": 3.199045603289457, + "learning_rate": 1.982483090267206e-05, + "loss": 1.1634, + "step": 810 + }, + { + "epoch": 0.44028230184581973, + "grad_norm": 2.9970299057571856, + "learning_rate": 1.9824174925905235e-05, + "loss": 0.9642, + "step": 811 + }, + { + "epoch": 0.44082519001085774, + "grad_norm": 3.9851247078132994, + "learning_rate": 1.98235177340688e-05, + "loss": 1.4926, + "step": 812 + }, + { + "epoch": 0.44136807817589574, + "grad_norm": 3.586604601363014, + "learning_rate": 1.9822859327244034e-05, + "loss": 1.3105, + "step": 813 + }, + { + "epoch": 0.44191096634093374, + "grad_norm": 4.198634753759357, + "learning_rate": 1.9822199705512372e-05, + "loss": 1.4437, + "step": 814 + }, + { + "epoch": 0.44245385450597174, + "grad_norm": 3.6499375531563096, + "learning_rate": 1.9821538868955394e-05, + "loss": 1.2063, + "step": 815 + }, + { + "epoch": 0.44299674267100975, + "grad_norm": 4.3973905860052716, + "learning_rate": 1.9820876817654836e-05, + "loss": 1.5002, + "step": 816 + }, + { + "epoch": 0.44353963083604775, + "grad_norm": 2.7871960857121287, + "learning_rate": 1.9820213551692585e-05, + "loss": 0.6589, + "step": 817 + }, + { + "epoch": 0.44408251900108575, + "grad_norm": 3.4126665543063184, + "learning_rate": 1.981954907115067e-05, + "loss": 1.1651, + "step": 818 + }, + { + "epoch": 0.44462540716612375, + "grad_norm": 4.229651941034759, + "learning_rate": 1.981888337611127e-05, + "loss": 1.6444, + "step": 819 + }, + { + "epoch": 0.44516829533116176, + "grad_norm": 3.4135993440364056, + "learning_rate": 1.981821646665673e-05, + "loss": 1.2614, + "step": 820 + }, + { + "epoch": 0.44571118349619976, + "grad_norm": 4.274078711060124, + "learning_rate": 1.9817548342869527e-05, + "loss": 1.66, + "step": 821 + }, + { + "epoch": 0.44625407166123776, + "grad_norm": 2.9406863812447903, + "learning_rate": 1.98168790048323e-05, + "loss": 1.0412, + "step": 822 + }, + { + "epoch": 0.44679695982627576, + "grad_norm": 4.300279758003301, + "learning_rate": 1.981620845262783e-05, + "loss": 1.3689, + "step": 823 + }, + { + "epoch": 0.44733984799131377, + "grad_norm": 3.130525530503507, + "learning_rate": 1.9815536686339056e-05, + "loss": 1.1771, + "step": 824 + }, + { + "epoch": 0.44788273615635177, + "grad_norm": 3.6606030046920597, + "learning_rate": 1.981486370604906e-05, + "loss": 1.0323, + "step": 825 + }, + { + "epoch": 0.44842562432138977, + "grad_norm": 3.495865966877518, + "learning_rate": 1.981418951184108e-05, + "loss": 1.4493, + "step": 826 + }, + { + "epoch": 0.4489685124864278, + "grad_norm": 3.621229661878418, + "learning_rate": 1.9813514103798498e-05, + "loss": 1.5392, + "step": 827 + }, + { + "epoch": 0.4495114006514658, + "grad_norm": 3.6798995423063485, + "learning_rate": 1.9812837482004853e-05, + "loss": 1.0593, + "step": 828 + }, + { + "epoch": 0.4500542888165038, + "grad_norm": 3.4514040085362954, + "learning_rate": 1.9812159646543824e-05, + "loss": 0.9262, + "step": 829 + }, + { + "epoch": 0.4505971769815418, + "grad_norm": 3.195454430712678, + "learning_rate": 1.9811480597499257e-05, + "loss": 1.1469, + "step": 830 + }, + { + "epoch": 0.4511400651465798, + "grad_norm": 4.590874916307442, + "learning_rate": 1.981080033495513e-05, + "loss": 1.6534, + "step": 831 + }, + { + "epoch": 0.4516829533116178, + "grad_norm": 3.668702667937113, + "learning_rate": 1.981011885899558e-05, + "loss": 1.0597, + "step": 832 + }, + { + "epoch": 0.4522258414766558, + "grad_norm": 2.7669639542025046, + "learning_rate": 1.98094361697049e-05, + "loss": 1.0387, + "step": 833 + }, + { + "epoch": 0.4527687296416938, + "grad_norm": 4.734510249253564, + "learning_rate": 1.9808752267167515e-05, + "loss": 1.6757, + "step": 834 + }, + { + "epoch": 0.4533116178067318, + "grad_norm": 9.006078348758516, + "learning_rate": 1.9808067151468018e-05, + "loss": 2.1559, + "step": 835 + }, + { + "epoch": 0.4538545059717698, + "grad_norm": 4.169764802563779, + "learning_rate": 1.9807380822691146e-05, + "loss": 1.2557, + "step": 836 + }, + { + "epoch": 0.4543973941368078, + "grad_norm": 3.970541817152741, + "learning_rate": 1.9806693280921783e-05, + "loss": 1.5586, + "step": 837 + }, + { + "epoch": 0.4549402823018458, + "grad_norm": 5.042448582457387, + "learning_rate": 1.9806004526244966e-05, + "loss": 1.3191, + "step": 838 + }, + { + "epoch": 0.4554831704668838, + "grad_norm": 3.7467475303782236, + "learning_rate": 1.9805314558745876e-05, + "loss": 1.3318, + "step": 839 + }, + { + "epoch": 0.4560260586319218, + "grad_norm": 4.214820131429537, + "learning_rate": 1.980462337850986e-05, + "loss": 1.3958, + "step": 840 + }, + { + "epoch": 0.4565689467969598, + "grad_norm": 3.9655409992603965, + "learning_rate": 1.980393098562239e-05, + "loss": 0.9978, + "step": 841 + }, + { + "epoch": 0.4571118349619978, + "grad_norm": 5.522638140748038, + "learning_rate": 1.980323738016912e-05, + "loss": 1.4151, + "step": 842 + }, + { + "epoch": 0.4576547231270358, + "grad_norm": 3.84359618797871, + "learning_rate": 1.9802542562235822e-05, + "loss": 1.3567, + "step": 843 + }, + { + "epoch": 0.4581976112920738, + "grad_norm": 5.115138943098397, + "learning_rate": 1.9801846531908437e-05, + "loss": 1.6437, + "step": 844 + }, + { + "epoch": 0.4587404994571118, + "grad_norm": 3.811847477591368, + "learning_rate": 1.9801149289273054e-05, + "loss": 1.2335, + "step": 845 + }, + { + "epoch": 0.4592833876221498, + "grad_norm": 4.4009848681180115, + "learning_rate": 1.98004508344159e-05, + "loss": 1.2147, + "step": 846 + }, + { + "epoch": 0.4598262757871878, + "grad_norm": 3.850991850744113, + "learning_rate": 1.9799751167423376e-05, + "loss": 1.4865, + "step": 847 + }, + { + "epoch": 0.4603691639522258, + "grad_norm": 3.1303042065802917, + "learning_rate": 1.9799050288382004e-05, + "loss": 1.4926, + "step": 848 + }, + { + "epoch": 0.4609120521172638, + "grad_norm": 4.225513214699549, + "learning_rate": 1.9798348197378475e-05, + "loss": 1.2044, + "step": 849 + }, + { + "epoch": 0.46145494028230183, + "grad_norm": 4.139979880818486, + "learning_rate": 1.9797644894499625e-05, + "loss": 1.5682, + "step": 850 + }, + { + "epoch": 0.46199782844733983, + "grad_norm": 4.414209735978709, + "learning_rate": 1.979694037983244e-05, + "loss": 1.9033, + "step": 851 + }, + { + "epoch": 0.46254071661237783, + "grad_norm": 5.079031167256292, + "learning_rate": 1.9796234653464057e-05, + "loss": 1.8745, + "step": 852 + }, + { + "epoch": 0.46308360477741584, + "grad_norm": 6.59971413476838, + "learning_rate": 1.9795527715481755e-05, + "loss": 1.2449, + "step": 853 + }, + { + "epoch": 0.46362649294245384, + "grad_norm": 3.863740127637264, + "learning_rate": 1.9794819565972973e-05, + "loss": 1.3912, + "step": 854 + }, + { + "epoch": 0.46416938110749184, + "grad_norm": 3.8578186204961327, + "learning_rate": 1.9794110205025302e-05, + "loss": 1.3487, + "step": 855 + }, + { + "epoch": 0.46471226927252984, + "grad_norm": 4.19751015099396, + "learning_rate": 1.9793399632726466e-05, + "loss": 1.129, + "step": 856 + }, + { + "epoch": 0.46525515743756785, + "grad_norm": 4.261483190159906, + "learning_rate": 1.979268784916436e-05, + "loss": 1.7298, + "step": 857 + }, + { + "epoch": 0.46579804560260585, + "grad_norm": 4.1774094914047435, + "learning_rate": 1.9791974854427008e-05, + "loss": 0.8716, + "step": 858 + }, + { + "epoch": 0.46634093376764385, + "grad_norm": 3.845723101011648, + "learning_rate": 1.9791260648602603e-05, + "loss": 0.9086, + "step": 859 + }, + { + "epoch": 0.46688382193268185, + "grad_norm": 3.4922208632687393, + "learning_rate": 1.9790545231779476e-05, + "loss": 1.3688, + "step": 860 + }, + { + "epoch": 0.46742671009771986, + "grad_norm": 2.976761133296452, + "learning_rate": 1.978982860404611e-05, + "loss": 0.7865, + "step": 861 + }, + { + "epoch": 0.46796959826275786, + "grad_norm": 3.934765402497862, + "learning_rate": 1.978911076549114e-05, + "loss": 1.2993, + "step": 862 + }, + { + "epoch": 0.46851248642779586, + "grad_norm": 3.029198772099951, + "learning_rate": 1.978839171620335e-05, + "loss": 1.1737, + "step": 863 + }, + { + "epoch": 0.46905537459283386, + "grad_norm": 3.5637545602445946, + "learning_rate": 1.978767145627167e-05, + "loss": 1.1932, + "step": 864 + }, + { + "epoch": 0.46959826275787186, + "grad_norm": 3.131289093388952, + "learning_rate": 1.9786949985785187e-05, + "loss": 0.69, + "step": 865 + }, + { + "epoch": 0.47014115092290987, + "grad_norm": 3.522468470152659, + "learning_rate": 1.978622730483313e-05, + "loss": 1.0277, + "step": 866 + }, + { + "epoch": 0.47068403908794787, + "grad_norm": 4.046439598363532, + "learning_rate": 1.9785503413504883e-05, + "loss": 1.2917, + "step": 867 + }, + { + "epoch": 0.47122692725298587, + "grad_norm": 5.55259287782346, + "learning_rate": 1.978477831188998e-05, + "loss": 1.866, + "step": 868 + }, + { + "epoch": 0.4717698154180239, + "grad_norm": 5.874766516228883, + "learning_rate": 1.97840520000781e-05, + "loss": 1.6221, + "step": 869 + }, + { + "epoch": 0.4723127035830619, + "grad_norm": 5.530080877073305, + "learning_rate": 1.9783324478159074e-05, + "loss": 1.7617, + "step": 870 + }, + { + "epoch": 0.4728555917480999, + "grad_norm": 4.996268624326209, + "learning_rate": 1.9782595746222886e-05, + "loss": 1.5794, + "step": 871 + }, + { + "epoch": 0.4733984799131379, + "grad_norm": 3.934049315300115, + "learning_rate": 1.9781865804359663e-05, + "loss": 1.2603, + "step": 872 + }, + { + "epoch": 0.4739413680781759, + "grad_norm": 4.534039775579426, + "learning_rate": 1.978113465265969e-05, + "loss": 1.2766, + "step": 873 + }, + { + "epoch": 0.4744842562432139, + "grad_norm": 4.281791787646871, + "learning_rate": 1.9780402291213393e-05, + "loss": 1.6424, + "step": 874 + }, + { + "epoch": 0.4750271444082519, + "grad_norm": 4.443571747878234, + "learning_rate": 1.977966872011135e-05, + "loss": 1.514, + "step": 875 + }, + { + "epoch": 0.4755700325732899, + "grad_norm": 4.432653652742011, + "learning_rate": 1.9778933939444298e-05, + "loss": 1.2405, + "step": 876 + }, + { + "epoch": 0.4761129207383279, + "grad_norm": 3.408851551186198, + "learning_rate": 1.9778197949303107e-05, + "loss": 0.9597, + "step": 877 + }, + { + "epoch": 0.4766558089033659, + "grad_norm": 3.13258933410851, + "learning_rate": 1.9777460749778812e-05, + "loss": 0.8015, + "step": 878 + }, + { + "epoch": 0.4771986970684039, + "grad_norm": 3.535707436018844, + "learning_rate": 1.977672234096259e-05, + "loss": 1.3375, + "step": 879 + }, + { + "epoch": 0.4777415852334419, + "grad_norm": 5.210333639552661, + "learning_rate": 1.9775982722945764e-05, + "loss": 1.5039, + "step": 880 + }, + { + "epoch": 0.4782844733984799, + "grad_norm": 4.14115837125392, + "learning_rate": 1.9775241895819818e-05, + "loss": 1.0267, + "step": 881 + }, + { + "epoch": 0.4788273615635179, + "grad_norm": 4.390843049257208, + "learning_rate": 1.977449985967637e-05, + "loss": 1.6639, + "step": 882 + }, + { + "epoch": 0.4793702497285559, + "grad_norm": 3.5906251649150405, + "learning_rate": 1.9773756614607205e-05, + "loss": 1.2382, + "step": 883 + }, + { + "epoch": 0.4799131378935939, + "grad_norm": 4.78880370244169, + "learning_rate": 1.9773012160704242e-05, + "loss": 1.9436, + "step": 884 + }, + { + "epoch": 0.4804560260586319, + "grad_norm": 4.024679917575935, + "learning_rate": 1.977226649805956e-05, + "loss": 1.7126, + "step": 885 + }, + { + "epoch": 0.4809989142236699, + "grad_norm": 4.789782875875197, + "learning_rate": 1.9771519626765384e-05, + "loss": 1.4639, + "step": 886 + }, + { + "epoch": 0.4815418023887079, + "grad_norm": 4.188325588019175, + "learning_rate": 1.9770771546914088e-05, + "loss": 0.9698, + "step": 887 + }, + { + "epoch": 0.4820846905537459, + "grad_norm": 4.491436069698424, + "learning_rate": 1.9770022258598192e-05, + "loss": 1.5504, + "step": 888 + }, + { + "epoch": 0.4826275787187839, + "grad_norm": 4.889270563723842, + "learning_rate": 1.9769271761910373e-05, + "loss": 1.6052, + "step": 889 + }, + { + "epoch": 0.4831704668838219, + "grad_norm": 3.7615089990618493, + "learning_rate": 1.9768520056943454e-05, + "loss": 1.2445, + "step": 890 + }, + { + "epoch": 0.4837133550488599, + "grad_norm": 4.153654659347469, + "learning_rate": 1.9767767143790403e-05, + "loss": 1.4304, + "step": 891 + }, + { + "epoch": 0.48425624321389793, + "grad_norm": 4.040032964648755, + "learning_rate": 1.9767013022544346e-05, + "loss": 1.2906, + "step": 892 + }, + { + "epoch": 0.48479913137893593, + "grad_norm": 4.6559235863058275, + "learning_rate": 1.976625769329855e-05, + "loss": 1.5214, + "step": 893 + }, + { + "epoch": 0.48534201954397393, + "grad_norm": 4.224494142526306, + "learning_rate": 1.976550115614644e-05, + "loss": 1.2299, + "step": 894 + }, + { + "epoch": 0.48588490770901194, + "grad_norm": 5.225621473809677, + "learning_rate": 1.9764743411181585e-05, + "loss": 1.3489, + "step": 895 + }, + { + "epoch": 0.48642779587404994, + "grad_norm": 3.298440710545404, + "learning_rate": 1.97639844584977e-05, + "loss": 1.2107, + "step": 896 + }, + { + "epoch": 0.48697068403908794, + "grad_norm": 4.8373671778516485, + "learning_rate": 1.9763224298188652e-05, + "loss": 1.7748, + "step": 897 + }, + { + "epoch": 0.48751357220412594, + "grad_norm": 3.9332967347674224, + "learning_rate": 1.9762462930348465e-05, + "loss": 1.1036, + "step": 898 + }, + { + "epoch": 0.48805646036916395, + "grad_norm": 3.555513906222738, + "learning_rate": 1.9761700355071306e-05, + "loss": 1.5988, + "step": 899 + }, + { + "epoch": 0.48859934853420195, + "grad_norm": 4.115001448738943, + "learning_rate": 1.9760936572451487e-05, + "loss": 0.9789, + "step": 900 + }, + { + "epoch": 0.48914223669923995, + "grad_norm": 3.955269044937872, + "learning_rate": 1.9760171582583476e-05, + "loss": 1.3833, + "step": 901 + }, + { + "epoch": 0.48968512486427795, + "grad_norm": 2.718909991804504, + "learning_rate": 1.975940538556189e-05, + "loss": 1.1708, + "step": 902 + }, + { + "epoch": 0.49022801302931596, + "grad_norm": 4.234679090051759, + "learning_rate": 1.975863798148149e-05, + "loss": 1.3002, + "step": 903 + }, + { + "epoch": 0.49077090119435396, + "grad_norm": 4.162638017444828, + "learning_rate": 1.975786937043719e-05, + "loss": 1.246, + "step": 904 + }, + { + "epoch": 0.49131378935939196, + "grad_norm": 4.466027820294199, + "learning_rate": 1.975709955252406e-05, + "loss": 1.7345, + "step": 905 + }, + { + "epoch": 0.49185667752442996, + "grad_norm": 4.120456111509664, + "learning_rate": 1.97563285278373e-05, + "loss": 1.1339, + "step": 906 + }, + { + "epoch": 0.49239956568946797, + "grad_norm": 3.7932999801551346, + "learning_rate": 1.9755556296472286e-05, + "loss": 1.241, + "step": 907 + }, + { + "epoch": 0.49294245385450597, + "grad_norm": 3.3311865780622107, + "learning_rate": 1.9754782858524515e-05, + "loss": 1.2329, + "step": 908 + }, + { + "epoch": 0.49348534201954397, + "grad_norm": 4.21189840914184, + "learning_rate": 1.9754008214089654e-05, + "loss": 1.089, + "step": 909 + }, + { + "epoch": 0.494028230184582, + "grad_norm": 4.882048591403942, + "learning_rate": 1.9753232363263513e-05, + "loss": 1.3579, + "step": 910 + }, + { + "epoch": 0.49457111834962, + "grad_norm": 3.8495412115497833, + "learning_rate": 1.9752455306142052e-05, + "loss": 0.876, + "step": 911 + }, + { + "epoch": 0.495114006514658, + "grad_norm": 5.35378911928701, + "learning_rate": 1.975167704282137e-05, + "loss": 1.5025, + "step": 912 + }, + { + "epoch": 0.495656894679696, + "grad_norm": 3.4776107222093176, + "learning_rate": 1.9750897573397733e-05, + "loss": 1.1576, + "step": 913 + }, + { + "epoch": 0.496199782844734, + "grad_norm": 4.3680755226003525, + "learning_rate": 1.9750116897967545e-05, + "loss": 1.2519, + "step": 914 + }, + { + "epoch": 0.496742671009772, + "grad_norm": 4.154395835162648, + "learning_rate": 1.9749335016627355e-05, + "loss": 1.3397, + "step": 915 + }, + { + "epoch": 0.49728555917481, + "grad_norm": 5.465210053229008, + "learning_rate": 1.9748551929473873e-05, + "loss": 1.7555, + "step": 916 + }, + { + "epoch": 0.497828447339848, + "grad_norm": 4.391089608848125, + "learning_rate": 1.9747767636603953e-05, + "loss": 0.9685, + "step": 917 + }, + { + "epoch": 0.498371335504886, + "grad_norm": 4.215127655197256, + "learning_rate": 1.9746982138114597e-05, + "loss": 1.3889, + "step": 918 + }, + { + "epoch": 0.498914223669924, + "grad_norm": 4.559291789845839, + "learning_rate": 1.9746195434102956e-05, + "loss": 1.2411, + "step": 919 + }, + { + "epoch": 0.499457111834962, + "grad_norm": 4.226267506705079, + "learning_rate": 1.9745407524666327e-05, + "loss": 0.9517, + "step": 920 + }, + { + "epoch": 0.5, + "grad_norm": 4.289151336000736, + "learning_rate": 1.9744618409902164e-05, + "loss": 0.9123, + "step": 921 + }, + { + "epoch": 0.500542888165038, + "grad_norm": 5.356043714827742, + "learning_rate": 1.9743828089908067e-05, + "loss": 1.4039, + "step": 922 + }, + { + "epoch": 0.501085776330076, + "grad_norm": 4.867388642606108, + "learning_rate": 1.9743036564781785e-05, + "loss": 1.877, + "step": 923 + }, + { + "epoch": 0.501628664495114, + "grad_norm": 4.034273796102648, + "learning_rate": 1.9742243834621207e-05, + "loss": 1.6151, + "step": 924 + }, + { + "epoch": 0.502171552660152, + "grad_norm": 4.67131234305618, + "learning_rate": 1.9741449899524393e-05, + "loss": 1.6415, + "step": 925 + }, + { + "epoch": 0.50271444082519, + "grad_norm": 4.609211834325983, + "learning_rate": 1.9740654759589524e-05, + "loss": 1.2702, + "step": 926 + }, + { + "epoch": 0.503257328990228, + "grad_norm": 3.8693712873725064, + "learning_rate": 1.973985841491495e-05, + "loss": 1.0518, + "step": 927 + }, + { + "epoch": 0.503800217155266, + "grad_norm": 4.224951197105749, + "learning_rate": 1.9739060865599163e-05, + "loss": 1.1367, + "step": 928 + }, + { + "epoch": 0.504343105320304, + "grad_norm": 4.851656117293343, + "learning_rate": 1.973826211174081e-05, + "loss": 1.2753, + "step": 929 + }, + { + "epoch": 0.504885993485342, + "grad_norm": 4.255261482914816, + "learning_rate": 1.9737462153438675e-05, + "loss": 1.1723, + "step": 930 + }, + { + "epoch": 0.50542888165038, + "grad_norm": 3.8038614228711616, + "learning_rate": 1.9736660990791705e-05, + "loss": 1.1919, + "step": 931 + }, + { + "epoch": 0.505971769815418, + "grad_norm": 3.934227455395002, + "learning_rate": 1.9735858623898984e-05, + "loss": 1.1311, + "step": 932 + }, + { + "epoch": 0.506514657980456, + "grad_norm": 6.087081162565673, + "learning_rate": 1.9735055052859752e-05, + "loss": 1.4226, + "step": 933 + }, + { + "epoch": 0.507057546145494, + "grad_norm": 3.901680463400142, + "learning_rate": 1.9734250277773398e-05, + "loss": 0.8336, + "step": 934 + }, + { + "epoch": 0.507600434310532, + "grad_norm": 4.29804205692996, + "learning_rate": 1.9733444298739454e-05, + "loss": 1.5767, + "step": 935 + }, + { + "epoch": 0.50814332247557, + "grad_norm": 4.114842645062852, + "learning_rate": 1.9732637115857606e-05, + "loss": 1.3353, + "step": 936 + }, + { + "epoch": 0.508686210640608, + "grad_norm": 3.8525228530533706, + "learning_rate": 1.9731828729227686e-05, + "loss": 0.8208, + "step": 937 + }, + { + "epoch": 0.509229098805646, + "grad_norm": 4.592070280997817, + "learning_rate": 1.973101913894968e-05, + "loss": 1.3908, + "step": 938 + }, + { + "epoch": 0.509771986970684, + "grad_norm": 3.5842386277465628, + "learning_rate": 1.9730208345123718e-05, + "loss": 1.1996, + "step": 939 + }, + { + "epoch": 0.510314875135722, + "grad_norm": 6.053647281932304, + "learning_rate": 1.9729396347850082e-05, + "loss": 1.565, + "step": 940 + }, + { + "epoch": 0.51085776330076, + "grad_norm": 4.515664489508703, + "learning_rate": 1.9728583147229196e-05, + "loss": 1.4823, + "step": 941 + }, + { + "epoch": 0.511400651465798, + "grad_norm": 4.7752750595730475, + "learning_rate": 1.9727768743361644e-05, + "loss": 1.3867, + "step": 942 + }, + { + "epoch": 0.511943539630836, + "grad_norm": 3.983262894008015, + "learning_rate": 1.972695313634815e-05, + "loss": 1.1327, + "step": 943 + }, + { + "epoch": 0.512486427795874, + "grad_norm": 3.8510372454489326, + "learning_rate": 1.9726136326289586e-05, + "loss": 1.0233, + "step": 944 + }, + { + "epoch": 0.5130293159609121, + "grad_norm": 3.7881258264140634, + "learning_rate": 1.9725318313286985e-05, + "loss": 1.0239, + "step": 945 + }, + { + "epoch": 0.5135722041259501, + "grad_norm": 5.532602386318144, + "learning_rate": 1.9724499097441513e-05, + "loss": 1.8325, + "step": 946 + }, + { + "epoch": 0.5141150922909881, + "grad_norm": 5.978249809716352, + "learning_rate": 1.9723678678854493e-05, + "loss": 1.4899, + "step": 947 + }, + { + "epoch": 0.5146579804560261, + "grad_norm": 4.164217442125778, + "learning_rate": 1.9722857057627398e-05, + "loss": 1.5428, + "step": 948 + }, + { + "epoch": 0.5152008686210641, + "grad_norm": 3.315839086509451, + "learning_rate": 1.9722034233861848e-05, + "loss": 1.161, + "step": 949 + }, + { + "epoch": 0.5157437567861021, + "grad_norm": 4.362448119896709, + "learning_rate": 1.9721210207659608e-05, + "loss": 1.3402, + "step": 950 + }, + { + "epoch": 0.5162866449511401, + "grad_norm": 4.196652705893848, + "learning_rate": 1.9720384979122594e-05, + "loss": 1.4782, + "step": 951 + }, + { + "epoch": 0.5168295331161781, + "grad_norm": 4.093486911730782, + "learning_rate": 1.9719558548352876e-05, + "loss": 0.8508, + "step": 952 + }, + { + "epoch": 0.5173724212812161, + "grad_norm": 3.623412868803193, + "learning_rate": 1.9718730915452664e-05, + "loss": 1.4979, + "step": 953 + }, + { + "epoch": 0.5179153094462541, + "grad_norm": 3.4964911024452845, + "learning_rate": 1.9717902080524324e-05, + "loss": 1.0495, + "step": 954 + }, + { + "epoch": 0.5184581976112921, + "grad_norm": 4.021867363378223, + "learning_rate": 1.9717072043670367e-05, + "loss": 1.5284, + "step": 955 + }, + { + "epoch": 0.5190010857763301, + "grad_norm": 4.383918034859983, + "learning_rate": 1.9716240804993454e-05, + "loss": 1.3092, + "step": 956 + }, + { + "epoch": 0.5195439739413681, + "grad_norm": 4.61200677033352, + "learning_rate": 1.971540836459639e-05, + "loss": 1.8809, + "step": 957 + }, + { + "epoch": 0.5200868621064061, + "grad_norm": 4.490685725956694, + "learning_rate": 1.9714574722582142e-05, + "loss": 1.4004, + "step": 958 + }, + { + "epoch": 0.5206297502714441, + "grad_norm": 3.4814367857792736, + "learning_rate": 1.9713739879053802e-05, + "loss": 0.7918, + "step": 959 + }, + { + "epoch": 0.5211726384364821, + "grad_norm": 4.751597566480749, + "learning_rate": 1.9712903834114635e-05, + "loss": 1.6595, + "step": 960 + }, + { + "epoch": 0.5217155266015201, + "grad_norm": 4.535333205317335, + "learning_rate": 1.9712066587868042e-05, + "loss": 1.5676, + "step": 961 + }, + { + "epoch": 0.5222584147665581, + "grad_norm": 4.285001993552699, + "learning_rate": 1.9711228140417577e-05, + "loss": 1.0156, + "step": 962 + }, + { + "epoch": 0.5228013029315961, + "grad_norm": 4.61361421766516, + "learning_rate": 1.9710388491866934e-05, + "loss": 0.9552, + "step": 963 + }, + { + "epoch": 0.5233441910966341, + "grad_norm": 5.760482951437034, + "learning_rate": 1.9709547642319968e-05, + "loss": 2.2573, + "step": 964 + }, + { + "epoch": 0.5238870792616721, + "grad_norm": 4.38974180509626, + "learning_rate": 1.9708705591880674e-05, + "loss": 0.9939, + "step": 965 + }, + { + "epoch": 0.5244299674267101, + "grad_norm": 4.856793718321527, + "learning_rate": 1.97078623406532e-05, + "loss": 1.6161, + "step": 966 + }, + { + "epoch": 0.5249728555917481, + "grad_norm": 4.164767885626303, + "learning_rate": 1.9707017888741838e-05, + "loss": 1.4327, + "step": 967 + }, + { + "epoch": 0.5255157437567861, + "grad_norm": 5.163022804287371, + "learning_rate": 1.970617223625104e-05, + "loss": 1.2063, + "step": 968 + }, + { + "epoch": 0.5260586319218241, + "grad_norm": 4.826390195584734, + "learning_rate": 1.9705325383285384e-05, + "loss": 1.609, + "step": 969 + }, + { + "epoch": 0.5266015200868621, + "grad_norm": 6.3570394668292405, + "learning_rate": 1.9704477329949617e-05, + "loss": 1.5845, + "step": 970 + }, + { + "epoch": 0.5271444082519001, + "grad_norm": 5.176208926963624, + "learning_rate": 1.9703628076348628e-05, + "loss": 1.5792, + "step": 971 + }, + { + "epoch": 0.5276872964169381, + "grad_norm": 4.968725075311641, + "learning_rate": 1.9702777622587452e-05, + "loss": 1.774, + "step": 972 + }, + { + "epoch": 0.5282301845819761, + "grad_norm": 3.733580423042607, + "learning_rate": 1.9701925968771277e-05, + "loss": 0.8424, + "step": 973 + }, + { + "epoch": 0.5287730727470141, + "grad_norm": 4.330993358163065, + "learning_rate": 1.9701073115005437e-05, + "loss": 1.6198, + "step": 974 + }, + { + "epoch": 0.5293159609120521, + "grad_norm": 4.64032700965104, + "learning_rate": 1.9700219061395408e-05, + "loss": 1.2958, + "step": 975 + }, + { + "epoch": 0.5298588490770901, + "grad_norm": 4.08538215893539, + "learning_rate": 1.969936380804683e-05, + "loss": 1.5371, + "step": 976 + }, + { + "epoch": 0.5304017372421281, + "grad_norm": 3.7672102191405012, + "learning_rate": 1.9698507355065478e-05, + "loss": 1.4831, + "step": 977 + }, + { + "epoch": 0.5309446254071661, + "grad_norm": 4.237599031471779, + "learning_rate": 1.969764970255728e-05, + "loss": 1.4561, + "step": 978 + }, + { + "epoch": 0.5314875135722041, + "grad_norm": 4.296489590459447, + "learning_rate": 1.9696790850628308e-05, + "loss": 1.3813, + "step": 979 + }, + { + "epoch": 0.5320304017372421, + "grad_norm": 4.037479773042692, + "learning_rate": 1.9695930799384787e-05, + "loss": 1.103, + "step": 980 + }, + { + "epoch": 0.5325732899022801, + "grad_norm": 3.388491089813824, + "learning_rate": 1.9695069548933097e-05, + "loss": 1.0416, + "step": 981 + }, + { + "epoch": 0.5331161780673181, + "grad_norm": 3.6808593992848873, + "learning_rate": 1.969420709937975e-05, + "loss": 1.3158, + "step": 982 + }, + { + "epoch": 0.5336590662323561, + "grad_norm": 3.574434479685402, + "learning_rate": 1.969334345083142e-05, + "loss": 0.9542, + "step": 983 + }, + { + "epoch": 0.5342019543973942, + "grad_norm": 4.205886917864195, + "learning_rate": 1.9692478603394926e-05, + "loss": 1.1542, + "step": 984 + }, + { + "epoch": 0.5347448425624322, + "grad_norm": 3.89660120144171, + "learning_rate": 1.9691612557177225e-05, + "loss": 1.2374, + "step": 985 + }, + { + "epoch": 0.5352877307274702, + "grad_norm": 5.508260629607227, + "learning_rate": 1.9690745312285443e-05, + "loss": 2.0836, + "step": 986 + }, + { + "epoch": 0.5358306188925082, + "grad_norm": 5.482641917586033, + "learning_rate": 1.968987686882683e-05, + "loss": 1.5608, + "step": 987 + }, + { + "epoch": 0.5363735070575462, + "grad_norm": 5.4640893579957845, + "learning_rate": 1.9689007226908807e-05, + "loss": 1.3632, + "step": 988 + }, + { + "epoch": 0.5369163952225842, + "grad_norm": 4.163793092253212, + "learning_rate": 1.9688136386638926e-05, + "loss": 1.0509, + "step": 989 + }, + { + "epoch": 0.5374592833876222, + "grad_norm": 7.431185891169344, + "learning_rate": 1.96872643481249e-05, + "loss": 2.2147, + "step": 990 + }, + { + "epoch": 0.5380021715526602, + "grad_norm": 4.243102861037138, + "learning_rate": 1.9686391111474574e-05, + "loss": 0.9978, + "step": 991 + }, + { + "epoch": 0.5385450597176982, + "grad_norm": 4.250454886057412, + "learning_rate": 1.968551667679596e-05, + "loss": 1.5265, + "step": 992 + }, + { + "epoch": 0.5390879478827362, + "grad_norm": 5.989466882905416, + "learning_rate": 1.9684641044197207e-05, + "loss": 1.3344, + "step": 993 + }, + { + "epoch": 0.5396308360477742, + "grad_norm": 6.003524822087355, + "learning_rate": 1.9683764213786617e-05, + "loss": 2.0785, + "step": 994 + }, + { + "epoch": 0.5401737242128122, + "grad_norm": 4.535858561115482, + "learning_rate": 1.9682886185672633e-05, + "loss": 1.7654, + "step": 995 + }, + { + "epoch": 0.5407166123778502, + "grad_norm": 3.304287633610362, + "learning_rate": 1.9682006959963854e-05, + "loss": 0.9151, + "step": 996 + }, + { + "epoch": 0.5412595005428882, + "grad_norm": 3.6154194334954797, + "learning_rate": 1.9681126536769022e-05, + "loss": 1.1871, + "step": 997 + }, + { + "epoch": 0.5418023887079262, + "grad_norm": 5.596808777173772, + "learning_rate": 1.968024491619703e-05, + "loss": 1.1773, + "step": 998 + }, + { + "epoch": 0.5423452768729642, + "grad_norm": 4.486728043355676, + "learning_rate": 1.9679362098356923e-05, + "loss": 1.6495, + "step": 999 + }, + { + "epoch": 0.5428881650380022, + "grad_norm": 5.254525660355539, + "learning_rate": 1.9678478083357882e-05, + "loss": 1.7632, + "step": 1000 + }, + { + "epoch": 0.5434310532030402, + "grad_norm": 4.084222439189464, + "learning_rate": 1.9677592871309248e-05, + "loss": 1.4108, + "step": 1001 + }, + { + "epoch": 0.5439739413680782, + "grad_norm": 3.7647794378542248, + "learning_rate": 1.9676706462320504e-05, + "loss": 1.0893, + "step": 1002 + }, + { + "epoch": 0.5445168295331162, + "grad_norm": 3.658128786211824, + "learning_rate": 1.967581885650128e-05, + "loss": 1.0626, + "step": 1003 + }, + { + "epoch": 0.5450597176981542, + "grad_norm": 4.111525087755858, + "learning_rate": 1.967493005396136e-05, + "loss": 1.1915, + "step": 1004 + }, + { + "epoch": 0.5456026058631922, + "grad_norm": 4.130761699023899, + "learning_rate": 1.967404005481067e-05, + "loss": 1.1007, + "step": 1005 + }, + { + "epoch": 0.5461454940282302, + "grad_norm": 4.610164776763781, + "learning_rate": 1.9673148859159292e-05, + "loss": 1.3626, + "step": 1006 + }, + { + "epoch": 0.5466883821932682, + "grad_norm": 3.3516729554956686, + "learning_rate": 1.9672256467117445e-05, + "loss": 1.1138, + "step": 1007 + }, + { + "epoch": 0.5472312703583062, + "grad_norm": 5.185808250719794, + "learning_rate": 1.9671362878795502e-05, + "loss": 1.5639, + "step": 1008 + }, + { + "epoch": 0.5477741585233442, + "grad_norm": 3.9549787550673323, + "learning_rate": 1.9670468094303983e-05, + "loss": 1.0094, + "step": 1009 + }, + { + "epoch": 0.5483170466883822, + "grad_norm": 5.025417088271204, + "learning_rate": 1.966957211375356e-05, + "loss": 1.9741, + "step": 1010 + }, + { + "epoch": 0.5488599348534202, + "grad_norm": 5.1560673391175, + "learning_rate": 1.9668674937255044e-05, + "loss": 0.9653, + "step": 1011 + }, + { + "epoch": 0.5494028230184582, + "grad_norm": 4.843540796107876, + "learning_rate": 1.9667776564919404e-05, + "loss": 1.4745, + "step": 1012 + }, + { + "epoch": 0.5499457111834962, + "grad_norm": 3.8379573247832193, + "learning_rate": 1.966687699685775e-05, + "loss": 1.537, + "step": 1013 + }, + { + "epoch": 0.5504885993485342, + "grad_norm": 4.291804816024517, + "learning_rate": 1.9665976233181342e-05, + "loss": 1.6859, + "step": 1014 + }, + { + "epoch": 0.5510314875135722, + "grad_norm": 4.827630960232388, + "learning_rate": 1.966507427400159e-05, + "loss": 1.3205, + "step": 1015 + }, + { + "epoch": 0.5515743756786102, + "grad_norm": 4.251540690047392, + "learning_rate": 1.9664171119430044e-05, + "loss": 0.8213, + "step": 1016 + }, + { + "epoch": 0.5521172638436482, + "grad_norm": 3.7693342275615476, + "learning_rate": 1.9663266769578414e-05, + "loss": 1.0349, + "step": 1017 + }, + { + "epoch": 0.5526601520086862, + "grad_norm": 4.075472936638236, + "learning_rate": 1.966236122455855e-05, + "loss": 1.0853, + "step": 1018 + }, + { + "epoch": 0.5532030401737242, + "grad_norm": 4.771167333111926, + "learning_rate": 1.9661454484482448e-05, + "loss": 1.2135, + "step": 1019 + }, + { + "epoch": 0.5537459283387622, + "grad_norm": 3.834854303591261, + "learning_rate": 1.9660546549462262e-05, + "loss": 1.0972, + "step": 1020 + }, + { + "epoch": 0.5542888165038002, + "grad_norm": 4.712560042125338, + "learning_rate": 1.9659637419610278e-05, + "loss": 0.8779, + "step": 1021 + }, + { + "epoch": 0.5548317046688382, + "grad_norm": 3.8174143787104566, + "learning_rate": 1.9658727095038942e-05, + "loss": 0.7395, + "step": 1022 + }, + { + "epoch": 0.5553745928338762, + "grad_norm": 5.970646625648977, + "learning_rate": 1.965781557586085e-05, + "loss": 1.4233, + "step": 1023 + }, + { + "epoch": 0.5559174809989142, + "grad_norm": 3.872798187838392, + "learning_rate": 1.9656902862188732e-05, + "loss": 0.8147, + "step": 1024 + }, + { + "epoch": 0.5564603691639523, + "grad_norm": 5.632609234869673, + "learning_rate": 1.9655988954135473e-05, + "loss": 1.1581, + "step": 1025 + }, + { + "epoch": 0.5570032573289903, + "grad_norm": 4.141188191911578, + "learning_rate": 1.9655073851814117e-05, + "loss": 1.1707, + "step": 1026 + }, + { + "epoch": 0.5575461454940283, + "grad_norm": 5.7835323867879564, + "learning_rate": 1.9654157555337837e-05, + "loss": 1.2383, + "step": 1027 + }, + { + "epoch": 0.5580890336590663, + "grad_norm": 6.410603262158535, + "learning_rate": 1.9653240064819965e-05, + "loss": 2.3461, + "step": 1028 + }, + { + "epoch": 0.5586319218241043, + "grad_norm": 4.661944777700264, + "learning_rate": 1.9652321380373974e-05, + "loss": 1.1718, + "step": 1029 + }, + { + "epoch": 0.5591748099891423, + "grad_norm": 4.224334614804915, + "learning_rate": 1.9651401502113497e-05, + "loss": 1.0937, + "step": 1030 + }, + { + "epoch": 0.5597176981541803, + "grad_norm": 4.292561821167698, + "learning_rate": 1.9650480430152295e-05, + "loss": 0.8306, + "step": 1031 + }, + { + "epoch": 0.5602605863192183, + "grad_norm": 4.515070808084738, + "learning_rate": 1.9649558164604293e-05, + "loss": 1.4042, + "step": 1032 + }, + { + "epoch": 0.5608034744842563, + "grad_norm": 4.906478484057087, + "learning_rate": 1.964863470558356e-05, + "loss": 1.3396, + "step": 1033 + }, + { + "epoch": 0.5613463626492943, + "grad_norm": 3.9530548146664306, + "learning_rate": 1.9647710053204307e-05, + "loss": 0.917, + "step": 1034 + }, + { + "epoch": 0.5618892508143323, + "grad_norm": 4.640160859146004, + "learning_rate": 1.96467842075809e-05, + "loss": 1.6625, + "step": 1035 + }, + { + "epoch": 0.5624321389793703, + "grad_norm": 4.921887921875984, + "learning_rate": 1.964585716882785e-05, + "loss": 1.5133, + "step": 1036 + }, + { + "epoch": 0.5629750271444083, + "grad_norm": 4.667679101301949, + "learning_rate": 1.964492893705981e-05, + "loss": 0.9713, + "step": 1037 + }, + { + "epoch": 0.5635179153094463, + "grad_norm": 6.0457527064758825, + "learning_rate": 1.9643999512391586e-05, + "loss": 1.1101, + "step": 1038 + }, + { + "epoch": 0.5640608034744843, + "grad_norm": 5.188403972218186, + "learning_rate": 1.964306889493813e-05, + "loss": 1.8276, + "step": 1039 + }, + { + "epoch": 0.5646036916395223, + "grad_norm": 4.449370010668479, + "learning_rate": 1.9642137084814548e-05, + "loss": 1.0797, + "step": 1040 + }, + { + "epoch": 0.5651465798045603, + "grad_norm": 5.97660491495462, + "learning_rate": 1.9641204082136085e-05, + "loss": 1.913, + "step": 1041 + }, + { + "epoch": 0.5656894679695983, + "grad_norm": 5.611282387277076, + "learning_rate": 1.9640269887018135e-05, + "loss": 2.4593, + "step": 1042 + }, + { + "epoch": 0.5662323561346363, + "grad_norm": 3.6226559029227996, + "learning_rate": 1.9639334499576237e-05, + "loss": 0.7939, + "step": 1043 + }, + { + "epoch": 0.5667752442996743, + "grad_norm": 4.701569628633213, + "learning_rate": 1.963839791992609e-05, + "loss": 1.4092, + "step": 1044 + }, + { + "epoch": 0.5673181324647123, + "grad_norm": 5.185222096624812, + "learning_rate": 1.9637460148183525e-05, + "loss": 1.3851, + "step": 1045 + }, + { + "epoch": 0.5678610206297503, + "grad_norm": 4.094772946058271, + "learning_rate": 1.963652118446453e-05, + "loss": 1.1872, + "step": 1046 + }, + { + "epoch": 0.5684039087947883, + "grad_norm": 4.653748340189311, + "learning_rate": 1.9635581028885233e-05, + "loss": 1.1546, + "step": 1047 + }, + { + "epoch": 0.5689467969598263, + "grad_norm": 3.9426831069951844, + "learning_rate": 1.9634639681561924e-05, + "loss": 0.9324, + "step": 1048 + }, + { + "epoch": 0.5694896851248643, + "grad_norm": 4.949304407204856, + "learning_rate": 1.9633697142611017e-05, + "loss": 1.2372, + "step": 1049 + }, + { + "epoch": 0.5700325732899023, + "grad_norm": 5.364689744345368, + "learning_rate": 1.9632753412149096e-05, + "loss": 1.5104, + "step": 1050 + }, + { + "epoch": 0.5705754614549403, + "grad_norm": 5.307969223857254, + "learning_rate": 1.9631808490292884e-05, + "loss": 1.212, + "step": 1051 + }, + { + "epoch": 0.5711183496199783, + "grad_norm": 5.197310781163775, + "learning_rate": 1.963086237715924e-05, + "loss": 1.0175, + "step": 1052 + }, + { + "epoch": 0.5716612377850163, + "grad_norm": 4.095170540464123, + "learning_rate": 1.9629915072865194e-05, + "loss": 1.2711, + "step": 1053 + }, + { + "epoch": 0.5722041259500543, + "grad_norm": 4.816474093418903, + "learning_rate": 1.9628966577527902e-05, + "loss": 0.7274, + "step": 1054 + }, + { + "epoch": 0.5727470141150923, + "grad_norm": 4.000604469041167, + "learning_rate": 1.962801689126468e-05, + "loss": 1.2913, + "step": 1055 + }, + { + "epoch": 0.5732899022801303, + "grad_norm": 4.31668064977018, + "learning_rate": 1.962706601419298e-05, + "loss": 1.2598, + "step": 1056 + }, + { + "epoch": 0.5738327904451683, + "grad_norm": 5.033983318277113, + "learning_rate": 1.9626113946430414e-05, + "loss": 1.5149, + "step": 1057 + }, + { + "epoch": 0.5743756786102063, + "grad_norm": 4.51738649586986, + "learning_rate": 1.9625160688094733e-05, + "loss": 1.2389, + "step": 1058 + }, + { + "epoch": 0.5749185667752443, + "grad_norm": 3.6899120907028853, + "learning_rate": 1.9624206239303837e-05, + "loss": 1.1632, + "step": 1059 + }, + { + "epoch": 0.5754614549402823, + "grad_norm": 4.423539955787714, + "learning_rate": 1.9623250600175775e-05, + "loss": 0.9526, + "step": 1060 + }, + { + "epoch": 0.5760043431053203, + "grad_norm": 5.062986942342047, + "learning_rate": 1.962229377082874e-05, + "loss": 1.6768, + "step": 1061 + }, + { + "epoch": 0.5765472312703583, + "grad_norm": 5.044415439938055, + "learning_rate": 1.962133575138108e-05, + "loss": 1.3808, + "step": 1062 + }, + { + "epoch": 0.5770901194353963, + "grad_norm": 3.1405283495786587, + "learning_rate": 1.962037654195128e-05, + "loss": 0.7588, + "step": 1063 + }, + { + "epoch": 0.5776330076004343, + "grad_norm": 5.449873922495929, + "learning_rate": 1.9619416142657974e-05, + "loss": 1.1544, + "step": 1064 + }, + { + "epoch": 0.5781758957654723, + "grad_norm": 4.268527750000251, + "learning_rate": 1.961845455361995e-05, + "loss": 1.1941, + "step": 1065 + }, + { + "epoch": 0.5787187839305103, + "grad_norm": 4.6641878464575095, + "learning_rate": 1.9617491774956137e-05, + "loss": 1.7162, + "step": 1066 + }, + { + "epoch": 0.5792616720955484, + "grad_norm": 4.583539478633983, + "learning_rate": 1.9616527806785612e-05, + "loss": 1.1415, + "step": 1067 + }, + { + "epoch": 0.5798045602605864, + "grad_norm": 4.325093393868848, + "learning_rate": 1.96155626492276e-05, + "loss": 1.3284, + "step": 1068 + }, + { + "epoch": 0.5803474484256244, + "grad_norm": 4.180109104666535, + "learning_rate": 1.9614596302401478e-05, + "loss": 1.136, + "step": 1069 + }, + { + "epoch": 0.5808903365906624, + "grad_norm": 4.108700840520527, + "learning_rate": 1.9613628766426762e-05, + "loss": 1.6142, + "step": 1070 + }, + { + "epoch": 0.5814332247557004, + "grad_norm": 5.1931558422083555, + "learning_rate": 1.9612660041423116e-05, + "loss": 0.8479, + "step": 1071 + }, + { + "epoch": 0.5819761129207384, + "grad_norm": 6.008601497593343, + "learning_rate": 1.9611690127510358e-05, + "loss": 1.4743, + "step": 1072 + }, + { + "epoch": 0.5825190010857764, + "grad_norm": 4.531622764164067, + "learning_rate": 1.9610719024808444e-05, + "loss": 1.5755, + "step": 1073 + }, + { + "epoch": 0.5830618892508144, + "grad_norm": 6.18647874774261, + "learning_rate": 1.9609746733437486e-05, + "loss": 1.7369, + "step": 1074 + }, + { + "epoch": 0.5836047774158524, + "grad_norm": 5.5880249590216025, + "learning_rate": 1.9608773253517738e-05, + "loss": 1.4402, + "step": 1075 + }, + { + "epoch": 0.5841476655808904, + "grad_norm": 4.798692544278573, + "learning_rate": 1.9607798585169595e-05, + "loss": 1.2257, + "step": 1076 + }, + { + "epoch": 0.5846905537459284, + "grad_norm": 5.635786862796798, + "learning_rate": 1.960682272851361e-05, + "loss": 1.183, + "step": 1077 + }, + { + "epoch": 0.5852334419109664, + "grad_norm": 3.3979344506068037, + "learning_rate": 1.9605845683670484e-05, + "loss": 0.6463, + "step": 1078 + }, + { + "epoch": 0.5857763300760044, + "grad_norm": 4.5152680576801405, + "learning_rate": 1.960486745076105e-05, + "loss": 0.8476, + "step": 1079 + }, + { + "epoch": 0.5863192182410424, + "grad_norm": 4.4606140059612835, + "learning_rate": 1.9603888029906305e-05, + "loss": 1.185, + "step": 1080 + }, + { + "epoch": 0.5868621064060804, + "grad_norm": 4.229219175605004, + "learning_rate": 1.960290742122738e-05, + "loss": 1.3672, + "step": 1081 + }, + { + "epoch": 0.5874049945711184, + "grad_norm": 4.964172517040214, + "learning_rate": 1.960192562484556e-05, + "loss": 1.0561, + "step": 1082 + }, + { + "epoch": 0.5879478827361564, + "grad_norm": 4.8348336861653785, + "learning_rate": 1.960094264088228e-05, + "loss": 1.1005, + "step": 1083 + }, + { + "epoch": 0.5884907709011944, + "grad_norm": 4.3153540955607514, + "learning_rate": 1.959995846945911e-05, + "loss": 1.6513, + "step": 1084 + }, + { + "epoch": 0.5890336590662324, + "grad_norm": 4.733352525701168, + "learning_rate": 1.9598973110697773e-05, + "loss": 1.6162, + "step": 1085 + }, + { + "epoch": 0.5895765472312704, + "grad_norm": 5.402815147148315, + "learning_rate": 1.959798656472015e-05, + "loss": 1.4178, + "step": 1086 + }, + { + "epoch": 0.5901194353963084, + "grad_norm": 3.8276911935691147, + "learning_rate": 1.9596998831648247e-05, + "loss": 1.1554, + "step": 1087 + }, + { + "epoch": 0.5906623235613464, + "grad_norm": 4.499779496087119, + "learning_rate": 1.9596009911604232e-05, + "loss": 0.9786, + "step": 1088 + }, + { + "epoch": 0.5912052117263844, + "grad_norm": 4.3705352636049595, + "learning_rate": 1.959501980471042e-05, + "loss": 1.0827, + "step": 1089 + }, + { + "epoch": 0.5917480998914224, + "grad_norm": 4.9770508472387025, + "learning_rate": 1.9594028511089264e-05, + "loss": 1.2512, + "step": 1090 + }, + { + "epoch": 0.5922909880564604, + "grad_norm": 4.23068807802663, + "learning_rate": 1.9593036030863376e-05, + "loss": 0.7528, + "step": 1091 + }, + { + "epoch": 0.5928338762214984, + "grad_norm": 4.321323716192747, + "learning_rate": 1.9592042364155496e-05, + "loss": 1.0773, + "step": 1092 + }, + { + "epoch": 0.5933767643865364, + "grad_norm": 3.380036112638087, + "learning_rate": 1.9591047511088535e-05, + "loss": 0.7238, + "step": 1093 + }, + { + "epoch": 0.5939196525515744, + "grad_norm": 3.9722637658815447, + "learning_rate": 1.959005147178553e-05, + "loss": 0.8132, + "step": 1094 + }, + { + "epoch": 0.5944625407166124, + "grad_norm": 4.354488865716506, + "learning_rate": 1.9589054246369673e-05, + "loss": 1.1214, + "step": 1095 + }, + { + "epoch": 0.5950054288816504, + "grad_norm": 4.337355429292092, + "learning_rate": 1.9588055834964307e-05, + "loss": 0.9039, + "step": 1096 + }, + { + "epoch": 0.5955483170466884, + "grad_norm": 5.864047590425705, + "learning_rate": 1.9587056237692912e-05, + "loss": 1.6648, + "step": 1097 + }, + { + "epoch": 0.5960912052117264, + "grad_norm": 4.904173131706714, + "learning_rate": 1.9586055454679123e-05, + "loss": 1.1465, + "step": 1098 + }, + { + "epoch": 0.5966340933767644, + "grad_norm": 4.411877930208604, + "learning_rate": 1.9585053486046717e-05, + "loss": 1.4579, + "step": 1099 + }, + { + "epoch": 0.5971769815418024, + "grad_norm": 4.985171448881133, + "learning_rate": 1.9584050331919616e-05, + "loss": 1.5497, + "step": 1100 + }, + { + "epoch": 0.5977198697068404, + "grad_norm": 4.950532301400581, + "learning_rate": 1.9583045992421902e-05, + "loss": 1.3811, + "step": 1101 + }, + { + "epoch": 0.5982627578718784, + "grad_norm": 4.154119544907876, + "learning_rate": 1.9582040467677782e-05, + "loss": 1.186, + "step": 1102 + }, + { + "epoch": 0.5988056460369164, + "grad_norm": 4.069018604381756, + "learning_rate": 1.9581033757811628e-05, + "loss": 0.9312, + "step": 1103 + }, + { + "epoch": 0.5993485342019544, + "grad_norm": 5.33070083129983, + "learning_rate": 1.9580025862947948e-05, + "loss": 1.4274, + "step": 1104 + }, + { + "epoch": 0.5998914223669924, + "grad_norm": 4.855606632567906, + "learning_rate": 1.95790167832114e-05, + "loss": 1.0875, + "step": 1105 + }, + { + "epoch": 0.6004343105320304, + "grad_norm": 5.30182267546013, + "learning_rate": 1.957800651872679e-05, + "loss": 1.6501, + "step": 1106 + }, + { + "epoch": 0.6009771986970684, + "grad_norm": 4.433903391255917, + "learning_rate": 1.957699506961907e-05, + "loss": 1.3897, + "step": 1107 + }, + { + "epoch": 0.6015200868621065, + "grad_norm": 4.687653782723915, + "learning_rate": 1.9575982436013335e-05, + "loss": 0.8484, + "step": 1108 + }, + { + "epoch": 0.6020629750271445, + "grad_norm": 5.650426140105992, + "learning_rate": 1.957496861803483e-05, + "loss": 1.923, + "step": 1109 + }, + { + "epoch": 0.6026058631921825, + "grad_norm": 4.024650772512973, + "learning_rate": 1.957395361580895e-05, + "loss": 1.2713, + "step": 1110 + }, + { + "epoch": 0.6031487513572205, + "grad_norm": 4.343921133786845, + "learning_rate": 1.9572937429461223e-05, + "loss": 1.3152, + "step": 1111 + }, + { + "epoch": 0.6036916395222585, + "grad_norm": 4.1467176317930265, + "learning_rate": 1.957192005911734e-05, + "loss": 1.3325, + "step": 1112 + }, + { + "epoch": 0.6042345276872965, + "grad_norm": 4.305465221711973, + "learning_rate": 1.9570901504903128e-05, + "loss": 0.9237, + "step": 1113 + }, + { + "epoch": 0.6047774158523345, + "grad_norm": 4.101045495900469, + "learning_rate": 1.9569881766944564e-05, + "loss": 1.3932, + "step": 1114 + }, + { + "epoch": 0.6053203040173725, + "grad_norm": 4.0920748344339835, + "learning_rate": 1.956886084536777e-05, + "loss": 1.0067, + "step": 1115 + }, + { + "epoch": 0.6058631921824105, + "grad_norm": 4.844431281019285, + "learning_rate": 1.956783874029902e-05, + "loss": 1.2726, + "step": 1116 + }, + { + "epoch": 0.6064060803474485, + "grad_norm": 4.329792546128024, + "learning_rate": 1.9566815451864723e-05, + "loss": 1.1144, + "step": 1117 + }, + { + "epoch": 0.6069489685124865, + "grad_norm": 5.430322578642178, + "learning_rate": 1.9565790980191447e-05, + "loss": 1.3307, + "step": 1118 + }, + { + "epoch": 0.6074918566775245, + "grad_norm": 3.672369600344093, + "learning_rate": 1.9564765325405895e-05, + "loss": 0.8514, + "step": 1119 + }, + { + "epoch": 0.6080347448425625, + "grad_norm": 3.6433282223345334, + "learning_rate": 1.9563738487634924e-05, + "loss": 1.1972, + "step": 1120 + }, + { + "epoch": 0.6085776330076005, + "grad_norm": 4.23249445053715, + "learning_rate": 1.956271046700553e-05, + "loss": 1.0638, + "step": 1121 + }, + { + "epoch": 0.6091205211726385, + "grad_norm": 3.9162807435580134, + "learning_rate": 1.956168126364487e-05, + "loss": 1.121, + "step": 1122 + }, + { + "epoch": 0.6096634093376765, + "grad_norm": 3.3651337591577812, + "learning_rate": 1.956065087768023e-05, + "loss": 0.9604, + "step": 1123 + }, + { + "epoch": 0.6102062975027145, + "grad_norm": 5.5905307606778685, + "learning_rate": 1.955961930923905e-05, + "loss": 1.4509, + "step": 1124 + }, + { + "epoch": 0.6107491856677525, + "grad_norm": 3.6120233107886404, + "learning_rate": 1.955858655844892e-05, + "loss": 1.0106, + "step": 1125 + }, + { + "epoch": 0.6112920738327905, + "grad_norm": 3.6242967727445468, + "learning_rate": 1.9557552625437574e-05, + "loss": 1.0768, + "step": 1126 + }, + { + "epoch": 0.6118349619978285, + "grad_norm": 4.856763858492221, + "learning_rate": 1.9556517510332883e-05, + "loss": 1.6372, + "step": 1127 + }, + { + "epoch": 0.6123778501628665, + "grad_norm": 4.648427809898363, + "learning_rate": 1.9555481213262873e-05, + "loss": 1.1622, + "step": 1128 + }, + { + "epoch": 0.6129207383279045, + "grad_norm": 4.228089396781667, + "learning_rate": 1.9554443734355723e-05, + "loss": 1.067, + "step": 1129 + }, + { + "epoch": 0.6134636264929425, + "grad_norm": 4.069786144676624, + "learning_rate": 1.9553405073739743e-05, + "loss": 1.1427, + "step": 1130 + }, + { + "epoch": 0.6140065146579805, + "grad_norm": 3.112182391585594, + "learning_rate": 1.9552365231543395e-05, + "loss": 0.7862, + "step": 1131 + }, + { + "epoch": 0.6145494028230185, + "grad_norm": 4.783518734299315, + "learning_rate": 1.955132420789529e-05, + "loss": 1.0366, + "step": 1132 + }, + { + "epoch": 0.6150922909880565, + "grad_norm": 4.905128048667836, + "learning_rate": 1.9550282002924187e-05, + "loss": 1.2566, + "step": 1133 + }, + { + "epoch": 0.6156351791530945, + "grad_norm": 4.643678450407587, + "learning_rate": 1.9549238616758987e-05, + "loss": 1.2749, + "step": 1134 + }, + { + "epoch": 0.6161780673181325, + "grad_norm": 4.18363664236292, + "learning_rate": 1.954819404952873e-05, + "loss": 1.1523, + "step": 1135 + }, + { + "epoch": 0.6167209554831705, + "grad_norm": 3.9776065520060335, + "learning_rate": 1.9547148301362623e-05, + "loss": 0.8672, + "step": 1136 + }, + { + "epoch": 0.6172638436482085, + "grad_norm": 5.417170242762362, + "learning_rate": 1.9546101372389994e-05, + "loss": 1.6434, + "step": 1137 + }, + { + "epoch": 0.6178067318132465, + "grad_norm": 3.3325133642660068, + "learning_rate": 1.9545053262740335e-05, + "loss": 1.1447, + "step": 1138 + }, + { + "epoch": 0.6183496199782845, + "grad_norm": 5.4727757064949545, + "learning_rate": 1.9544003972543273e-05, + "loss": 1.5013, + "step": 1139 + }, + { + "epoch": 0.6188925081433225, + "grad_norm": 4.619714236951128, + "learning_rate": 1.954295350192859e-05, + "loss": 0.9758, + "step": 1140 + }, + { + "epoch": 0.6194353963083605, + "grad_norm": 5.189787937143953, + "learning_rate": 1.954190185102621e-05, + "loss": 1.2771, + "step": 1141 + }, + { + "epoch": 0.6199782844733985, + "grad_norm": 6.571114312185869, + "learning_rate": 1.9540849019966198e-05, + "loss": 1.2805, + "step": 1142 + }, + { + "epoch": 0.6205211726384365, + "grad_norm": 5.6343515388196455, + "learning_rate": 1.9539795008878774e-05, + "loss": 1.4061, + "step": 1143 + }, + { + "epoch": 0.6210640608034745, + "grad_norm": 5.582702517092888, + "learning_rate": 1.9538739817894302e-05, + "loss": 1.613, + "step": 1144 + }, + { + "epoch": 0.6216069489685125, + "grad_norm": 6.194616996988272, + "learning_rate": 1.9537683447143287e-05, + "loss": 1.114, + "step": 1145 + }, + { + "epoch": 0.6221498371335505, + "grad_norm": 6.0395907363668515, + "learning_rate": 1.9536625896756377e-05, + "loss": 1.5233, + "step": 1146 + }, + { + "epoch": 0.6226927252985885, + "grad_norm": 4.672511820840566, + "learning_rate": 1.9535567166864382e-05, + "loss": 1.1491, + "step": 1147 + }, + { + "epoch": 0.6232356134636265, + "grad_norm": 4.954133576744497, + "learning_rate": 1.9534507257598244e-05, + "loss": 1.1974, + "step": 1148 + }, + { + "epoch": 0.6237785016286646, + "grad_norm": 4.447205012745297, + "learning_rate": 1.953344616908905e-05, + "loss": 1.019, + "step": 1149 + }, + { + "epoch": 0.6243213897937026, + "grad_norm": 4.284087771325503, + "learning_rate": 1.9532383901468038e-05, + "loss": 1.0814, + "step": 1150 + }, + { + "epoch": 0.6248642779587406, + "grad_norm": 3.576067836847051, + "learning_rate": 1.9531320454866595e-05, + "loss": 1.1025, + "step": 1151 + }, + { + "epoch": 0.6254071661237784, + "grad_norm": 3.6831292979487777, + "learning_rate": 1.9530255829416246e-05, + "loss": 0.967, + "step": 1152 + }, + { + "epoch": 0.6259500542888164, + "grad_norm": 3.842496728015195, + "learning_rate": 1.952919002524867e-05, + "loss": 1.0903, + "step": 1153 + }, + { + "epoch": 0.6264929424538545, + "grad_norm": 4.077633480381933, + "learning_rate": 1.952812304249568e-05, + "loss": 0.9227, + "step": 1154 + }, + { + "epoch": 0.6270358306188925, + "grad_norm": 4.157004955121598, + "learning_rate": 1.952705488128925e-05, + "loss": 0.9427, + "step": 1155 + }, + { + "epoch": 0.6275787187839305, + "grad_norm": 5.267614104282471, + "learning_rate": 1.952598554176149e-05, + "loss": 1.9875, + "step": 1156 + }, + { + "epoch": 0.6281216069489685, + "grad_norm": 3.8938350314376735, + "learning_rate": 1.952491502404465e-05, + "loss": 1.1662, + "step": 1157 + }, + { + "epoch": 0.6286644951140065, + "grad_norm": 4.851409026807682, + "learning_rate": 1.9523843328271144e-05, + "loss": 1.9559, + "step": 1158 + }, + { + "epoch": 0.6292073832790445, + "grad_norm": 4.4104613077358845, + "learning_rate": 1.9522770454573513e-05, + "loss": 0.8011, + "step": 1159 + }, + { + "epoch": 0.6297502714440825, + "grad_norm": 3.6883335641947856, + "learning_rate": 1.952169640308446e-05, + "loss": 1.1838, + "step": 1160 + }, + { + "epoch": 0.6302931596091205, + "grad_norm": 3.728203039459734, + "learning_rate": 1.9520621173936818e-05, + "loss": 1.1423, + "step": 1161 + }, + { + "epoch": 0.6308360477741585, + "grad_norm": 3.6272612276564296, + "learning_rate": 1.9519544767263574e-05, + "loss": 0.8376, + "step": 1162 + }, + { + "epoch": 0.6313789359391965, + "grad_norm": 3.928276333701112, + "learning_rate": 1.951846718319786e-05, + "loss": 1.0749, + "step": 1163 + }, + { + "epoch": 0.6319218241042345, + "grad_norm": 5.941949883397345, + "learning_rate": 1.951738842187296e-05, + "loss": 1.8535, + "step": 1164 + }, + { + "epoch": 0.6324647122692725, + "grad_norm": 3.4741200080360644, + "learning_rate": 1.951630848342229e-05, + "loss": 0.7537, + "step": 1165 + }, + { + "epoch": 0.6330076004343105, + "grad_norm": 5.392480994365815, + "learning_rate": 1.9515227367979416e-05, + "loss": 0.8776, + "step": 1166 + }, + { + "epoch": 0.6335504885993485, + "grad_norm": 5.203169892005258, + "learning_rate": 1.951414507567806e-05, + "loss": 1.2678, + "step": 1167 + }, + { + "epoch": 0.6340933767643865, + "grad_norm": 7.499226723318992, + "learning_rate": 1.9513061606652076e-05, + "loss": 1.3295, + "step": 1168 + }, + { + "epoch": 0.6346362649294245, + "grad_norm": 4.355172848230307, + "learning_rate": 1.9511976961035474e-05, + "loss": 0.9188, + "step": 1169 + }, + { + "epoch": 0.6351791530944625, + "grad_norm": 4.983220943925928, + "learning_rate": 1.9510891138962398e-05, + "loss": 1.1951, + "step": 1170 + }, + { + "epoch": 0.6357220412595005, + "grad_norm": 5.3098934312394395, + "learning_rate": 1.950980414056715e-05, + "loss": 0.855, + "step": 1171 + }, + { + "epoch": 0.6362649294245385, + "grad_norm": 5.125683116898323, + "learning_rate": 1.950871596598417e-05, + "loss": 1.1371, + "step": 1172 + }, + { + "epoch": 0.6368078175895765, + "grad_norm": 4.3010814264850215, + "learning_rate": 1.950762661534804e-05, + "loss": 1.7401, + "step": 1173 + }, + { + "epoch": 0.6373507057546145, + "grad_norm": 5.860625964437982, + "learning_rate": 1.95065360887935e-05, + "loss": 1.3557, + "step": 1174 + }, + { + "epoch": 0.6378935939196525, + "grad_norm": 6.3356570754666155, + "learning_rate": 1.9505444386455426e-05, + "loss": 1.2281, + "step": 1175 + }, + { + "epoch": 0.6384364820846905, + "grad_norm": 4.788366086551355, + "learning_rate": 1.9504351508468842e-05, + "loss": 1.3458, + "step": 1176 + }, + { + "epoch": 0.6389793702497285, + "grad_norm": 4.5553889359557065, + "learning_rate": 1.9503257454968914e-05, + "loss": 0.803, + "step": 1177 + }, + { + "epoch": 0.6395222584147665, + "grad_norm": 4.9063187966995345, + "learning_rate": 1.950216222609096e-05, + "loss": 0.8306, + "step": 1178 + }, + { + "epoch": 0.6400651465798045, + "grad_norm": 4.775138114897886, + "learning_rate": 1.9501065821970435e-05, + "loss": 0.9783, + "step": 1179 + }, + { + "epoch": 0.6406080347448425, + "grad_norm": 4.53672284240203, + "learning_rate": 1.9499968242742948e-05, + "loss": 1.3076, + "step": 1180 + }, + { + "epoch": 0.6411509229098805, + "grad_norm": 4.539860050199487, + "learning_rate": 1.949886948854425e-05, + "loss": 1.4212, + "step": 1181 + }, + { + "epoch": 0.6416938110749185, + "grad_norm": 4.251462432005476, + "learning_rate": 1.9497769559510232e-05, + "loss": 1.0274, + "step": 1182 + }, + { + "epoch": 0.6422366992399565, + "grad_norm": 4.49791289564495, + "learning_rate": 1.9496668455776938e-05, + "loss": 1.7531, + "step": 1183 + }, + { + "epoch": 0.6427795874049945, + "grad_norm": 4.996740483321485, + "learning_rate": 1.9495566177480555e-05, + "loss": 1.201, + "step": 1184 + }, + { + "epoch": 0.6433224755700325, + "grad_norm": 5.347352236818569, + "learning_rate": 1.9494462724757413e-05, + "loss": 1.5723, + "step": 1185 + }, + { + "epoch": 0.6438653637350705, + "grad_norm": 4.2242558589374015, + "learning_rate": 1.9493358097743988e-05, + "loss": 1.8014, + "step": 1186 + }, + { + "epoch": 0.6444082519001085, + "grad_norm": 5.042556201755935, + "learning_rate": 1.9492252296576906e-05, + "loss": 1.5466, + "step": 1187 + }, + { + "epoch": 0.6449511400651465, + "grad_norm": 4.284314238035697, + "learning_rate": 1.949114532139293e-05, + "loss": 0.9515, + "step": 1188 + }, + { + "epoch": 0.6454940282301845, + "grad_norm": 4.699826119048575, + "learning_rate": 1.9490037172328974e-05, + "loss": 1.1677, + "step": 1189 + }, + { + "epoch": 0.6460369163952225, + "grad_norm": 6.035247135554916, + "learning_rate": 1.9488927849522095e-05, + "loss": 1.957, + "step": 1190 + }, + { + "epoch": 0.6465798045602605, + "grad_norm": 4.042004235065528, + "learning_rate": 1.94878173531095e-05, + "loss": 1.2182, + "step": 1191 + }, + { + "epoch": 0.6471226927252985, + "grad_norm": 4.114859597350168, + "learning_rate": 1.948670568322853e-05, + "loss": 1.1787, + "step": 1192 + }, + { + "epoch": 0.6476655808903365, + "grad_norm": 4.380177461001504, + "learning_rate": 1.9485592840016682e-05, + "loss": 1.3912, + "step": 1193 + }, + { + "epoch": 0.6482084690553745, + "grad_norm": 5.160710982722015, + "learning_rate": 1.948447882361159e-05, + "loss": 1.2642, + "step": 1194 + }, + { + "epoch": 0.6487513572204126, + "grad_norm": 3.950160510485583, + "learning_rate": 1.9483363634151046e-05, + "loss": 1.1246, + "step": 1195 + }, + { + "epoch": 0.6492942453854506, + "grad_norm": 4.587449618376974, + "learning_rate": 1.9482247271772974e-05, + "loss": 1.3275, + "step": 1196 + }, + { + "epoch": 0.6498371335504886, + "grad_norm": 4.919831513668322, + "learning_rate": 1.9481129736615445e-05, + "loss": 1.4007, + "step": 1197 + }, + { + "epoch": 0.6503800217155266, + "grad_norm": 4.298954423192263, + "learning_rate": 1.948001102881668e-05, + "loss": 0.8076, + "step": 1198 + }, + { + "epoch": 0.6509229098805646, + "grad_norm": 5.525772014179547, + "learning_rate": 1.9478891148515043e-05, + "loss": 1.4558, + "step": 1199 + }, + { + "epoch": 0.6514657980456026, + "grad_norm": 5.473549669240299, + "learning_rate": 1.947777009584904e-05, + "loss": 1.0976, + "step": 1200 + }, + { + "epoch": 0.6520086862106406, + "grad_norm": 5.360010284326786, + "learning_rate": 1.947664787095733e-05, + "loss": 1.0462, + "step": 1201 + }, + { + "epoch": 0.6525515743756786, + "grad_norm": 4.696691081444868, + "learning_rate": 1.9475524473978705e-05, + "loss": 0.933, + "step": 1202 + }, + { + "epoch": 0.6530944625407166, + "grad_norm": 6.0503072087093965, + "learning_rate": 1.947439990505211e-05, + "loss": 0.9628, + "step": 1203 + }, + { + "epoch": 0.6536373507057546, + "grad_norm": 5.476272854511399, + "learning_rate": 1.9473274164316637e-05, + "loss": 0.9716, + "step": 1204 + }, + { + "epoch": 0.6541802388707926, + "grad_norm": 3.910448715873912, + "learning_rate": 1.9472147251911517e-05, + "loss": 1.0732, + "step": 1205 + }, + { + "epoch": 0.6547231270358306, + "grad_norm": 4.596114022138067, + "learning_rate": 1.9471019167976126e-05, + "loss": 0.9399, + "step": 1206 + }, + { + "epoch": 0.6552660152008686, + "grad_norm": 6.764840749864346, + "learning_rate": 1.946988991264999e-05, + "loss": 2.0529, + "step": 1207 + }, + { + "epoch": 0.6558089033659066, + "grad_norm": 5.217328230116913, + "learning_rate": 1.9468759486072778e-05, + "loss": 1.0477, + "step": 1208 + }, + { + "epoch": 0.6563517915309446, + "grad_norm": 4.03429219354135, + "learning_rate": 1.9467627888384303e-05, + "loss": 1.1103, + "step": 1209 + }, + { + "epoch": 0.6568946796959826, + "grad_norm": 4.301388894222403, + "learning_rate": 1.946649511972452e-05, + "loss": 0.8515, + "step": 1210 + }, + { + "epoch": 0.6574375678610206, + "grad_norm": 5.285446092583105, + "learning_rate": 1.9465361180233536e-05, + "loss": 1.2881, + "step": 1211 + }, + { + "epoch": 0.6579804560260586, + "grad_norm": 5.879897026479874, + "learning_rate": 1.9464226070051593e-05, + "loss": 0.9362, + "step": 1212 + }, + { + "epoch": 0.6585233441910966, + "grad_norm": 5.335257600571856, + "learning_rate": 1.9463089789319083e-05, + "loss": 1.4372, + "step": 1213 + }, + { + "epoch": 0.6590662323561346, + "grad_norm": 5.272070125431448, + "learning_rate": 1.9461952338176552e-05, + "loss": 1.2063, + "step": 1214 + }, + { + "epoch": 0.6596091205211726, + "grad_norm": 5.188765715525815, + "learning_rate": 1.946081371676467e-05, + "loss": 0.8912, + "step": 1215 + }, + { + "epoch": 0.6601520086862106, + "grad_norm": 4.747539421286818, + "learning_rate": 1.9459673925224275e-05, + "loss": 1.0381, + "step": 1216 + }, + { + "epoch": 0.6606948968512486, + "grad_norm": 4.486254331013965, + "learning_rate": 1.945853296369633e-05, + "loss": 1.2584, + "step": 1217 + }, + { + "epoch": 0.6612377850162866, + "grad_norm": 4.730012415000151, + "learning_rate": 1.945739083232195e-05, + "loss": 1.1061, + "step": 1218 + }, + { + "epoch": 0.6617806731813246, + "grad_norm": 5.713566521874704, + "learning_rate": 1.9456247531242405e-05, + "loss": 1.4628, + "step": 1219 + }, + { + "epoch": 0.6623235613463626, + "grad_norm": 6.08134803588137, + "learning_rate": 1.9455103060599093e-05, + "loss": 1.6743, + "step": 1220 + }, + { + "epoch": 0.6628664495114006, + "grad_norm": 4.778151760566848, + "learning_rate": 1.9453957420533562e-05, + "loss": 0.8454, + "step": 1221 + }, + { + "epoch": 0.6634093376764386, + "grad_norm": 5.007171333863363, + "learning_rate": 1.945281061118751e-05, + "loss": 1.079, + "step": 1222 + }, + { + "epoch": 0.6639522258414766, + "grad_norm": 4.034543591270571, + "learning_rate": 1.945166263270278e-05, + "loss": 0.8378, + "step": 1223 + }, + { + "epoch": 0.6644951140065146, + "grad_norm": 4.191899722379318, + "learning_rate": 1.9450513485221352e-05, + "loss": 1.3602, + "step": 1224 + }, + { + "epoch": 0.6650380021715526, + "grad_norm": 6.349672663190246, + "learning_rate": 1.944936316888535e-05, + "loss": 1.861, + "step": 1225 + }, + { + "epoch": 0.6655808903365906, + "grad_norm": 5.991894404952759, + "learning_rate": 1.9448211683837055e-05, + "loss": 1.0931, + "step": 1226 + }, + { + "epoch": 0.6661237785016286, + "grad_norm": 4.405247018509644, + "learning_rate": 1.9447059030218876e-05, + "loss": 1.2239, + "step": 1227 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 4.810361623786047, + "learning_rate": 1.9445905208173387e-05, + "loss": 1.1468, + "step": 1228 + }, + { + "epoch": 0.6672095548317046, + "grad_norm": 4.6967843893557015, + "learning_rate": 1.944475021784328e-05, + "loss": 1.5304, + "step": 1229 + }, + { + "epoch": 0.6677524429967426, + "grad_norm": 5.325911382516086, + "learning_rate": 1.9443594059371417e-05, + "loss": 1.2218, + "step": 1230 + }, + { + "epoch": 0.6682953311617806, + "grad_norm": 4.416437363913103, + "learning_rate": 1.9442436732900788e-05, + "loss": 1.223, + "step": 1231 + }, + { + "epoch": 0.6688382193268186, + "grad_norm": 3.720560087272861, + "learning_rate": 1.9441278238574537e-05, + "loss": 0.8155, + "step": 1232 + }, + { + "epoch": 0.6693811074918566, + "grad_norm": 5.273159297494141, + "learning_rate": 1.9440118576535947e-05, + "loss": 1.1958, + "step": 1233 + }, + { + "epoch": 0.6699239956568946, + "grad_norm": 5.684867015184436, + "learning_rate": 1.9438957746928443e-05, + "loss": 1.2384, + "step": 1234 + }, + { + "epoch": 0.6704668838219326, + "grad_norm": 4.952541482610055, + "learning_rate": 1.9437795749895604e-05, + "loss": 1.337, + "step": 1235 + }, + { + "epoch": 0.6710097719869706, + "grad_norm": 4.788863413585636, + "learning_rate": 1.9436632585581145e-05, + "loss": 0.5182, + "step": 1236 + }, + { + "epoch": 0.6715526601520087, + "grad_norm": 5.723878805619219, + "learning_rate": 1.9435468254128925e-05, + "loss": 1.1444, + "step": 1237 + }, + { + "epoch": 0.6720955483170467, + "grad_norm": 4.3068839361839, + "learning_rate": 1.9434302755682958e-05, + "loss": 1.1286, + "step": 1238 + }, + { + "epoch": 0.6726384364820847, + "grad_norm": 3.7588872381668477, + "learning_rate": 1.9433136090387384e-05, + "loss": 0.9718, + "step": 1239 + }, + { + "epoch": 0.6731813246471227, + "grad_norm": 6.802207677110681, + "learning_rate": 1.9431968258386508e-05, + "loss": 1.2389, + "step": 1240 + }, + { + "epoch": 0.6737242128121607, + "grad_norm": 3.6180158875124735, + "learning_rate": 1.9430799259824766e-05, + "loss": 0.644, + "step": 1241 + }, + { + "epoch": 0.6742671009771987, + "grad_norm": 5.559822163300458, + "learning_rate": 1.9429629094846742e-05, + "loss": 1.1157, + "step": 1242 + }, + { + "epoch": 0.6748099891422367, + "grad_norm": 4.084355750224745, + "learning_rate": 1.942845776359716e-05, + "loss": 0.91, + "step": 1243 + }, + { + "epoch": 0.6753528773072747, + "grad_norm": 3.9402358983627583, + "learning_rate": 1.9427285266220895e-05, + "loss": 0.9409, + "step": 1244 + }, + { + "epoch": 0.6758957654723127, + "grad_norm": 5.093172430302549, + "learning_rate": 1.9426111602862968e-05, + "loss": 1.5905, + "step": 1245 + }, + { + "epoch": 0.6764386536373507, + "grad_norm": 4.02536040473524, + "learning_rate": 1.942493677366853e-05, + "loss": 0.9277, + "step": 1246 + }, + { + "epoch": 0.6769815418023887, + "grad_norm": 4.741533426757492, + "learning_rate": 1.942376077878289e-05, + "loss": 1.1027, + "step": 1247 + }, + { + "epoch": 0.6775244299674267, + "grad_norm": 5.365058237804207, + "learning_rate": 1.9422583618351503e-05, + "loss": 1.072, + "step": 1248 + }, + { + "epoch": 0.6780673181324647, + "grad_norm": 4.479339123691231, + "learning_rate": 1.9421405292519956e-05, + "loss": 1.474, + "step": 1249 + }, + { + "epoch": 0.6786102062975027, + "grad_norm": 7.79086989645888, + "learning_rate": 1.942022580143398e-05, + "loss": 1.6973, + "step": 1250 + }, + { + "epoch": 0.6791530944625407, + "grad_norm": 5.621467578044424, + "learning_rate": 1.9419045145239474e-05, + "loss": 1.5432, + "step": 1251 + }, + { + "epoch": 0.6796959826275787, + "grad_norm": 5.041038447258229, + "learning_rate": 1.9417863324082444e-05, + "loss": 1.2977, + "step": 1252 + }, + { + "epoch": 0.6802388707926167, + "grad_norm": 4.717189339301973, + "learning_rate": 1.941668033810907e-05, + "loss": 0.924, + "step": 1253 + }, + { + "epoch": 0.6807817589576547, + "grad_norm": 5.086833595497465, + "learning_rate": 1.9415496187465667e-05, + "loss": 1.5545, + "step": 1254 + }, + { + "epoch": 0.6813246471226927, + "grad_norm": 4.790403842984273, + "learning_rate": 1.941431087229869e-05, + "loss": 1.2545, + "step": 1255 + }, + { + "epoch": 0.6818675352877307, + "grad_norm": 4.470785518257313, + "learning_rate": 1.9413124392754735e-05, + "loss": 1.2036, + "step": 1256 + }, + { + "epoch": 0.6824104234527687, + "grad_norm": 3.820710537112263, + "learning_rate": 1.9411936748980557e-05, + "loss": 1.0856, + "step": 1257 + }, + { + "epoch": 0.6829533116178067, + "grad_norm": 4.453720544384727, + "learning_rate": 1.9410747941123036e-05, + "loss": 1.1645, + "step": 1258 + }, + { + "epoch": 0.6834961997828447, + "grad_norm": 4.692503519890958, + "learning_rate": 1.9409557969329218e-05, + "loss": 1.1174, + "step": 1259 + }, + { + "epoch": 0.6840390879478827, + "grad_norm": 5.666971506629844, + "learning_rate": 1.940836683374627e-05, + "loss": 1.4904, + "step": 1260 + }, + { + "epoch": 0.6845819761129207, + "grad_norm": 5.468580615919992, + "learning_rate": 1.940717453452152e-05, + "loss": 1.2834, + "step": 1261 + }, + { + "epoch": 0.6851248642779587, + "grad_norm": 6.42319990350808, + "learning_rate": 1.9405981071802425e-05, + "loss": 1.1144, + "step": 1262 + }, + { + "epoch": 0.6856677524429967, + "grad_norm": 6.078674980286373, + "learning_rate": 1.9404786445736607e-05, + "loss": 1.4214, + "step": 1263 + }, + { + "epoch": 0.6862106406080347, + "grad_norm": 4.465631588283012, + "learning_rate": 1.9403590656471806e-05, + "loss": 0.7813, + "step": 1264 + }, + { + "epoch": 0.6867535287730727, + "grad_norm": 5.220582351132811, + "learning_rate": 1.940239370415593e-05, + "loss": 1.2112, + "step": 1265 + }, + { + "epoch": 0.6872964169381107, + "grad_norm": 5.1374760487841895, + "learning_rate": 1.9401195588937014e-05, + "loss": 1.2402, + "step": 1266 + }, + { + "epoch": 0.6878393051031487, + "grad_norm": 5.106543779494583, + "learning_rate": 1.9399996310963243e-05, + "loss": 1.1846, + "step": 1267 + }, + { + "epoch": 0.6883821932681867, + "grad_norm": 3.9521299273320363, + "learning_rate": 1.939879587038295e-05, + "loss": 0.8975, + "step": 1268 + }, + { + "epoch": 0.6889250814332247, + "grad_norm": 5.398700470714175, + "learning_rate": 1.9397594267344604e-05, + "loss": 1.4686, + "step": 1269 + }, + { + "epoch": 0.6894679695982627, + "grad_norm": 5.8287973073589265, + "learning_rate": 1.939639150199682e-05, + "loss": 1.011, + "step": 1270 + }, + { + "epoch": 0.6900108577633007, + "grad_norm": 4.812101240482426, + "learning_rate": 1.9395187574488358e-05, + "loss": 0.948, + "step": 1271 + }, + { + "epoch": 0.6905537459283387, + "grad_norm": 5.828731649028842, + "learning_rate": 1.939398248496813e-05, + "loss": 1.2233, + "step": 1272 + }, + { + "epoch": 0.6910966340933767, + "grad_norm": 4.890080173319348, + "learning_rate": 1.9392776233585167e-05, + "loss": 0.8859, + "step": 1273 + }, + { + "epoch": 0.6916395222584147, + "grad_norm": 4.920508250004042, + "learning_rate": 1.9391568820488674e-05, + "loss": 0.73, + "step": 1274 + }, + { + "epoch": 0.6921824104234527, + "grad_norm": 5.635401553567158, + "learning_rate": 1.9390360245827983e-05, + "loss": 1.1432, + "step": 1275 + }, + { + "epoch": 0.6927252985884907, + "grad_norm": 4.588336191443314, + "learning_rate": 1.9389150509752566e-05, + "loss": 1.3326, + "step": 1276 + }, + { + "epoch": 0.6932681867535287, + "grad_norm": 4.631694372660094, + "learning_rate": 1.9387939612412056e-05, + "loss": 1.3762, + "step": 1277 + }, + { + "epoch": 0.6938110749185668, + "grad_norm": 5.585985863722577, + "learning_rate": 1.938672755395621e-05, + "loss": 1.3606, + "step": 1278 + }, + { + "epoch": 0.6943539630836048, + "grad_norm": 4.764783952643876, + "learning_rate": 1.938551433453494e-05, + "loss": 0.9858, + "step": 1279 + }, + { + "epoch": 0.6948968512486428, + "grad_norm": 4.234252918062867, + "learning_rate": 1.9384299954298297e-05, + "loss": 0.7702, + "step": 1280 + }, + { + "epoch": 0.6954397394136808, + "grad_norm": 4.36117548276201, + "learning_rate": 1.938308441339648e-05, + "loss": 1.1846, + "step": 1281 + }, + { + "epoch": 0.6959826275787188, + "grad_norm": 4.311957696864392, + "learning_rate": 1.938186771197983e-05, + "loss": 0.9934, + "step": 1282 + }, + { + "epoch": 0.6965255157437568, + "grad_norm": 5.117807004985933, + "learning_rate": 1.9380649850198824e-05, + "loss": 1.4474, + "step": 1283 + }, + { + "epoch": 0.6970684039087948, + "grad_norm": 5.695329527120276, + "learning_rate": 1.93794308282041e-05, + "loss": 1.3001, + "step": 1284 + }, + { + "epoch": 0.6976112920738328, + "grad_norm": 4.305244126285458, + "learning_rate": 1.937821064614642e-05, + "loss": 1.1421, + "step": 1285 + }, + { + "epoch": 0.6981541802388708, + "grad_norm": 4.574797635236639, + "learning_rate": 1.93769893041767e-05, + "loss": 1.1104, + "step": 1286 + }, + { + "epoch": 0.6986970684039088, + "grad_norm": 5.051193103711606, + "learning_rate": 1.9375766802446002e-05, + "loss": 1.2858, + "step": 1287 + }, + { + "epoch": 0.6992399565689468, + "grad_norm": 5.002235181102662, + "learning_rate": 1.9374543141105518e-05, + "loss": 1.2931, + "step": 1288 + }, + { + "epoch": 0.6997828447339848, + "grad_norm": 5.259386594746194, + "learning_rate": 1.93733183203066e-05, + "loss": 1.0432, + "step": 1289 + }, + { + "epoch": 0.7003257328990228, + "grad_norm": 3.7218449672815117, + "learning_rate": 1.9372092340200736e-05, + "loss": 0.9913, + "step": 1290 + }, + { + "epoch": 0.7008686210640608, + "grad_norm": 4.45109753373239, + "learning_rate": 1.937086520093955e-05, + "loss": 1.2012, + "step": 1291 + }, + { + "epoch": 0.7014115092290988, + "grad_norm": 4.73529586440727, + "learning_rate": 1.9369636902674823e-05, + "loss": 1.3995, + "step": 1292 + }, + { + "epoch": 0.7019543973941368, + "grad_norm": 3.845443933095641, + "learning_rate": 1.936840744555847e-05, + "loss": 1.0971, + "step": 1293 + }, + { + "epoch": 0.7024972855591748, + "grad_norm": 6.67046211091574, + "learning_rate": 1.9367176829742553e-05, + "loss": 1.5273, + "step": 1294 + }, + { + "epoch": 0.7030401737242128, + "grad_norm": 4.454761961182613, + "learning_rate": 1.9365945055379275e-05, + "loss": 1.1282, + "step": 1295 + }, + { + "epoch": 0.7035830618892508, + "grad_norm": 6.552953195473038, + "learning_rate": 1.936471212262099e-05, + "loss": 2.171, + "step": 1296 + }, + { + "epoch": 0.7041259500542888, + "grad_norm": 4.894073730847946, + "learning_rate": 1.9363478031620182e-05, + "loss": 1.2303, + "step": 1297 + }, + { + "epoch": 0.7046688382193268, + "grad_norm": 3.863202427039748, + "learning_rate": 1.936224278252949e-05, + "loss": 0.8515, + "step": 1298 + }, + { + "epoch": 0.7052117263843648, + "grad_norm": 5.503562663348476, + "learning_rate": 1.9361006375501685e-05, + "loss": 1.1834, + "step": 1299 + }, + { + "epoch": 0.7057546145494028, + "grad_norm": 5.318422186143503, + "learning_rate": 1.9359768810689697e-05, + "loss": 1.3378, + "step": 1300 + }, + { + "epoch": 0.7062975027144408, + "grad_norm": 4.46194766163203, + "learning_rate": 1.9358530088246582e-05, + "loss": 1.3318, + "step": 1301 + }, + { + "epoch": 0.7068403908794788, + "grad_norm": 5.6623901137247765, + "learning_rate": 1.9357290208325552e-05, + "loss": 1.3512, + "step": 1302 + }, + { + "epoch": 0.7073832790445168, + "grad_norm": 5.805010217787694, + "learning_rate": 1.9356049171079957e-05, + "loss": 1.2383, + "step": 1303 + }, + { + "epoch": 0.7079261672095548, + "grad_norm": 4.847713425075124, + "learning_rate": 1.9354806976663286e-05, + "loss": 0.7329, + "step": 1304 + }, + { + "epoch": 0.7084690553745928, + "grad_norm": 4.207200307039602, + "learning_rate": 1.935356362522918e-05, + "loss": 1.0119, + "step": 1305 + }, + { + "epoch": 0.7090119435396308, + "grad_norm": 5.76731204400203, + "learning_rate": 1.9352319116931417e-05, + "loss": 1.7383, + "step": 1306 + }, + { + "epoch": 0.7095548317046688, + "grad_norm": 4.989804434976344, + "learning_rate": 1.935107345192392e-05, + "loss": 1.1958, + "step": 1307 + }, + { + "epoch": 0.7100977198697068, + "grad_norm": 4.6099552434159135, + "learning_rate": 1.9349826630360757e-05, + "loss": 0.9933, + "step": 1308 + }, + { + "epoch": 0.7106406080347448, + "grad_norm": 4.693361205108708, + "learning_rate": 1.9348578652396136e-05, + "loss": 1.3553, + "step": 1309 + }, + { + "epoch": 0.7111834961997828, + "grad_norm": 5.473441077550542, + "learning_rate": 1.9347329518184406e-05, + "loss": 1.2202, + "step": 1310 + }, + { + "epoch": 0.7117263843648208, + "grad_norm": 5.885376904454794, + "learning_rate": 1.9346079227880062e-05, + "loss": 1.3698, + "step": 1311 + }, + { + "epoch": 0.7122692725298588, + "grad_norm": 4.115405735194196, + "learning_rate": 1.9344827781637744e-05, + "loss": 1.3361, + "step": 1312 + }, + { + "epoch": 0.7128121606948968, + "grad_norm": 6.745730370060005, + "learning_rate": 1.9343575179612236e-05, + "loss": 1.6411, + "step": 1313 + }, + { + "epoch": 0.7133550488599348, + "grad_norm": 4.363445628534919, + "learning_rate": 1.9342321421958455e-05, + "loss": 0.792, + "step": 1314 + }, + { + "epoch": 0.7138979370249728, + "grad_norm": 6.167581448099132, + "learning_rate": 1.9341066508831472e-05, + "loss": 1.4264, + "step": 1315 + }, + { + "epoch": 0.7144408251900108, + "grad_norm": 4.973571230650737, + "learning_rate": 1.9339810440386495e-05, + "loss": 0.9769, + "step": 1316 + }, + { + "epoch": 0.7149837133550488, + "grad_norm": 5.2637238951774945, + "learning_rate": 1.933855321677888e-05, + "loss": 1.2478, + "step": 1317 + }, + { + "epoch": 0.7155266015200868, + "grad_norm": 5.2296234862723505, + "learning_rate": 1.9337294838164118e-05, + "loss": 1.3999, + "step": 1318 + }, + { + "epoch": 0.7160694896851248, + "grad_norm": 4.720322767524389, + "learning_rate": 1.9336035304697848e-05, + "loss": 1.3561, + "step": 1319 + }, + { + "epoch": 0.7166123778501629, + "grad_norm": 4.2513159313917015, + "learning_rate": 1.9334774616535854e-05, + "loss": 1.176, + "step": 1320 + }, + { + "epoch": 0.7171552660152009, + "grad_norm": 4.5267349508423225, + "learning_rate": 1.9333512773834057e-05, + "loss": 1.1211, + "step": 1321 + }, + { + "epoch": 0.7176981541802389, + "grad_norm": 5.384909301403834, + "learning_rate": 1.9332249776748523e-05, + "loss": 1.3904, + "step": 1322 + }, + { + "epoch": 0.7182410423452769, + "grad_norm": 4.67839864320759, + "learning_rate": 1.9330985625435468e-05, + "loss": 0.8608, + "step": 1323 + }, + { + "epoch": 0.7187839305103149, + "grad_norm": 5.193826637693982, + "learning_rate": 1.9329720320051233e-05, + "loss": 1.3172, + "step": 1324 + }, + { + "epoch": 0.7193268186753529, + "grad_norm": 4.8605293911870096, + "learning_rate": 1.9328453860752324e-05, + "loss": 1.3134, + "step": 1325 + }, + { + "epoch": 0.7198697068403909, + "grad_norm": 5.685265101122876, + "learning_rate": 1.9327186247695377e-05, + "loss": 1.2036, + "step": 1326 + }, + { + "epoch": 0.7204125950054289, + "grad_norm": 5.159715270999642, + "learning_rate": 1.9325917481037164e-05, + "loss": 1.4898, + "step": 1327 + }, + { + "epoch": 0.7209554831704669, + "grad_norm": 5.793178164657722, + "learning_rate": 1.9324647560934613e-05, + "loss": 1.3909, + "step": 1328 + }, + { + "epoch": 0.7214983713355049, + "grad_norm": 3.7442056178189267, + "learning_rate": 1.9323376487544795e-05, + "loss": 0.774, + "step": 1329 + }, + { + "epoch": 0.7220412595005429, + "grad_norm": 4.962252385233631, + "learning_rate": 1.9322104261024912e-05, + "loss": 1.3528, + "step": 1330 + }, + { + "epoch": 0.7225841476655809, + "grad_norm": 5.4534463408495, + "learning_rate": 1.9320830881532316e-05, + "loss": 1.6754, + "step": 1331 + }, + { + "epoch": 0.7231270358306189, + "grad_norm": 6.301262380555263, + "learning_rate": 1.93195563492245e-05, + "loss": 1.5583, + "step": 1332 + }, + { + "epoch": 0.7236699239956569, + "grad_norm": 3.970618104957748, + "learning_rate": 1.9318280664259103e-05, + "loss": 0.8964, + "step": 1333 + }, + { + "epoch": 0.7242128121606949, + "grad_norm": 3.490105469488415, + "learning_rate": 1.9317003826793904e-05, + "loss": 0.8098, + "step": 1334 + }, + { + "epoch": 0.7247557003257329, + "grad_norm": 5.240119278714349, + "learning_rate": 1.9315725836986822e-05, + "loss": 1.1688, + "step": 1335 + }, + { + "epoch": 0.7252985884907709, + "grad_norm": 4.550130688769963, + "learning_rate": 1.931444669499592e-05, + "loss": 1.2569, + "step": 1336 + }, + { + "epoch": 0.7258414766558089, + "grad_norm": 5.058353172472431, + "learning_rate": 1.9313166400979404e-05, + "loss": 1.3471, + "step": 1337 + }, + { + "epoch": 0.7263843648208469, + "grad_norm": 4.528967594836655, + "learning_rate": 1.931188495509563e-05, + "loss": 1.1831, + "step": 1338 + }, + { + "epoch": 0.7269272529858849, + "grad_norm": 4.493564615513058, + "learning_rate": 1.931060235750308e-05, + "loss": 1.4889, + "step": 1339 + }, + { + "epoch": 0.7274701411509229, + "grad_norm": 5.176344629525538, + "learning_rate": 1.9309318608360392e-05, + "loss": 1.6704, + "step": 1340 + }, + { + "epoch": 0.7280130293159609, + "grad_norm": 6.738195156936204, + "learning_rate": 1.930803370782634e-05, + "loss": 1.2147, + "step": 1341 + }, + { + "epoch": 0.7285559174809989, + "grad_norm": 4.248439922934747, + "learning_rate": 1.9306747656059847e-05, + "loss": 1.1943, + "step": 1342 + }, + { + "epoch": 0.7290988056460369, + "grad_norm": 5.175389864390461, + "learning_rate": 1.930546045321997e-05, + "loss": 1.4587, + "step": 1343 + }, + { + "epoch": 0.7296416938110749, + "grad_norm": 5.19934365393201, + "learning_rate": 1.9304172099465914e-05, + "loss": 0.9433, + "step": 1344 + }, + { + "epoch": 0.7301845819761129, + "grad_norm": 6.18681550522946, + "learning_rate": 1.9302882594957025e-05, + "loss": 1.5576, + "step": 1345 + }, + { + "epoch": 0.7307274701411509, + "grad_norm": 4.311510847746277, + "learning_rate": 1.930159193985279e-05, + "loss": 1.327, + "step": 1346 + }, + { + "epoch": 0.7312703583061889, + "grad_norm": 4.044345899710192, + "learning_rate": 1.9300300134312838e-05, + "loss": 1.2106, + "step": 1347 + }, + { + "epoch": 0.7318132464712269, + "grad_norm": 4.181894752624144, + "learning_rate": 1.929900717849694e-05, + "loss": 0.8703, + "step": 1348 + }, + { + "epoch": 0.7323561346362649, + "grad_norm": 5.37566626429507, + "learning_rate": 1.929771307256502e-05, + "loss": 1.446, + "step": 1349 + }, + { + "epoch": 0.7328990228013029, + "grad_norm": 5.37247658500195, + "learning_rate": 1.9296417816677123e-05, + "loss": 0.9239, + "step": 1350 + }, + { + "epoch": 0.7334419109663409, + "grad_norm": 5.749510968066806, + "learning_rate": 1.929512141099346e-05, + "loss": 1.4799, + "step": 1351 + }, + { + "epoch": 0.7339847991313789, + "grad_norm": 5.079635654383841, + "learning_rate": 1.929382385567436e-05, + "loss": 1.2571, + "step": 1352 + }, + { + "epoch": 0.7345276872964169, + "grad_norm": 4.923400923431618, + "learning_rate": 1.929252515088032e-05, + "loss": 1.1967, + "step": 1353 + }, + { + "epoch": 0.7350705754614549, + "grad_norm": 4.397561354835364, + "learning_rate": 1.9291225296771957e-05, + "loss": 0.9088, + "step": 1354 + }, + { + "epoch": 0.7356134636264929, + "grad_norm": 4.13744065040371, + "learning_rate": 1.9289924293510037e-05, + "loss": 1.023, + "step": 1355 + }, + { + "epoch": 0.7361563517915309, + "grad_norm": 4.266553618249588, + "learning_rate": 1.9288622141255477e-05, + "loss": 1.019, + "step": 1356 + }, + { + "epoch": 0.7366992399565689, + "grad_norm": 4.52588883698767, + "learning_rate": 1.928731884016933e-05, + "loss": 1.0935, + "step": 1357 + }, + { + "epoch": 0.7372421281216069, + "grad_norm": 5.46382933116543, + "learning_rate": 1.9286014390412786e-05, + "loss": 1.2833, + "step": 1358 + }, + { + "epoch": 0.737785016286645, + "grad_norm": 4.88651312061201, + "learning_rate": 1.928470879214718e-05, + "loss": 1.6064, + "step": 1359 + }, + { + "epoch": 0.738327904451683, + "grad_norm": 5.403882256899124, + "learning_rate": 1.9283402045533995e-05, + "loss": 1.1537, + "step": 1360 + }, + { + "epoch": 0.738870792616721, + "grad_norm": 6.381532932652649, + "learning_rate": 1.928209415073485e-05, + "loss": 1.527, + "step": 1361 + }, + { + "epoch": 0.739413680781759, + "grad_norm": 4.795244725313847, + "learning_rate": 1.9280785107911505e-05, + "loss": 1.4696, + "step": 1362 + }, + { + "epoch": 0.739956568946797, + "grad_norm": 5.8579061053451955, + "learning_rate": 1.9279474917225866e-05, + "loss": 1.5566, + "step": 1363 + }, + { + "epoch": 0.740499457111835, + "grad_norm": 4.5018596923036185, + "learning_rate": 1.927816357883998e-05, + "loss": 0.9842, + "step": 1364 + }, + { + "epoch": 0.741042345276873, + "grad_norm": 3.596935443077127, + "learning_rate": 1.927685109291604e-05, + "loss": 0.632, + "step": 1365 + }, + { + "epoch": 0.741585233441911, + "grad_norm": 4.535193817325199, + "learning_rate": 1.9275537459616364e-05, + "loss": 1.332, + "step": 1366 + }, + { + "epoch": 0.742128121606949, + "grad_norm": 6.678844775309246, + "learning_rate": 1.9274222679103437e-05, + "loss": 1.6826, + "step": 1367 + }, + { + "epoch": 0.742671009771987, + "grad_norm": 4.444554760823683, + "learning_rate": 1.927290675153987e-05, + "loss": 0.9597, + "step": 1368 + }, + { + "epoch": 0.743213897937025, + "grad_norm": 5.19917061501543, + "learning_rate": 1.927158967708841e-05, + "loss": 1.4333, + "step": 1369 + }, + { + "epoch": 0.743756786102063, + "grad_norm": 4.742331177279521, + "learning_rate": 1.927027145591197e-05, + "loss": 1.1659, + "step": 1370 + }, + { + "epoch": 0.744299674267101, + "grad_norm": 3.713496752478956, + "learning_rate": 1.926895208817358e-05, + "loss": 0.8688, + "step": 1371 + }, + { + "epoch": 0.744842562432139, + "grad_norm": 4.424173231823696, + "learning_rate": 1.9267631574036417e-05, + "loss": 1.0751, + "step": 1372 + }, + { + "epoch": 0.745385450597177, + "grad_norm": 4.7028219143924055, + "learning_rate": 1.9266309913663815e-05, + "loss": 1.3478, + "step": 1373 + }, + { + "epoch": 0.745928338762215, + "grad_norm": 5.498103807374834, + "learning_rate": 1.9264987107219237e-05, + "loss": 1.168, + "step": 1374 + }, + { + "epoch": 0.746471226927253, + "grad_norm": 4.326699254091958, + "learning_rate": 1.9263663154866285e-05, + "loss": 1.0713, + "step": 1375 + }, + { + "epoch": 0.747014115092291, + "grad_norm": 5.58697502498226, + "learning_rate": 1.926233805676871e-05, + "loss": 1.2474, + "step": 1376 + }, + { + "epoch": 0.747557003257329, + "grad_norm": 6.475832093356038, + "learning_rate": 1.92610118130904e-05, + "loss": 1.2805, + "step": 1377 + }, + { + "epoch": 0.748099891422367, + "grad_norm": 5.194166453296571, + "learning_rate": 1.925968442399539e-05, + "loss": 0.9736, + "step": 1378 + }, + { + "epoch": 0.748642779587405, + "grad_norm": 3.867607548943643, + "learning_rate": 1.9258355889647855e-05, + "loss": 1.2325, + "step": 1379 + }, + { + "epoch": 0.749185667752443, + "grad_norm": 5.602424635230674, + "learning_rate": 1.925702621021211e-05, + "loss": 1.0153, + "step": 1380 + }, + { + "epoch": 0.749728555917481, + "grad_norm": 5.011820196133583, + "learning_rate": 1.9255695385852604e-05, + "loss": 1.1823, + "step": 1381 + }, + { + "epoch": 0.750271444082519, + "grad_norm": 5.238515717778492, + "learning_rate": 1.9254363416733944e-05, + "loss": 1.4279, + "step": 1382 + }, + { + "epoch": 0.750814332247557, + "grad_norm": 5.600089025591521, + "learning_rate": 1.925303030302087e-05, + "loss": 0.9212, + "step": 1383 + }, + { + "epoch": 0.751357220412595, + "grad_norm": 4.318444011047328, + "learning_rate": 1.9251696044878255e-05, + "loss": 0.9623, + "step": 1384 + }, + { + "epoch": 0.751900108577633, + "grad_norm": 5.471464995723797, + "learning_rate": 1.925036064247113e-05, + "loss": 1.5446, + "step": 1385 + }, + { + "epoch": 0.752442996742671, + "grad_norm": 5.153911794581426, + "learning_rate": 1.9249024095964663e-05, + "loss": 1.7223, + "step": 1386 + }, + { + "epoch": 0.752985884907709, + "grad_norm": 3.7202924563691537, + "learning_rate": 1.924768640552415e-05, + "loss": 1.1469, + "step": 1387 + }, + { + "epoch": 0.753528773072747, + "grad_norm": 4.612135079041308, + "learning_rate": 1.9246347571315043e-05, + "loss": 0.9572, + "step": 1388 + }, + { + "epoch": 0.754071661237785, + "grad_norm": 4.618898432554816, + "learning_rate": 1.9245007593502937e-05, + "loss": 1.0443, + "step": 1389 + }, + { + "epoch": 0.754614549402823, + "grad_norm": 5.2787203720857345, + "learning_rate": 1.9243666472253554e-05, + "loss": 1.2034, + "step": 1390 + }, + { + "epoch": 0.755157437567861, + "grad_norm": 4.510719578546688, + "learning_rate": 1.9242324207732766e-05, + "loss": 0.8055, + "step": 1391 + }, + { + "epoch": 0.755700325732899, + "grad_norm": 6.256234902874701, + "learning_rate": 1.9240980800106596e-05, + "loss": 1.5985, + "step": 1392 + }, + { + "epoch": 0.756243213897937, + "grad_norm": 9.01213198061367, + "learning_rate": 1.923963624954119e-05, + "loss": 1.7921, + "step": 1393 + }, + { + "epoch": 0.756786102062975, + "grad_norm": 5.149498439254725, + "learning_rate": 1.923829055620285e-05, + "loss": 1.2406, + "step": 1394 + }, + { + "epoch": 0.757328990228013, + "grad_norm": 5.083183084254609, + "learning_rate": 1.9236943720258007e-05, + "loss": 0.9887, + "step": 1395 + }, + { + "epoch": 0.757871878393051, + "grad_norm": 4.5118989088500685, + "learning_rate": 1.9235595741873247e-05, + "loss": 1.3528, + "step": 1396 + }, + { + "epoch": 0.758414766558089, + "grad_norm": 4.583236692880461, + "learning_rate": 1.923424662121528e-05, + "loss": 1.0185, + "step": 1397 + }, + { + "epoch": 0.758957654723127, + "grad_norm": 5.3602373686113625, + "learning_rate": 1.9232896358450976e-05, + "loss": 1.1827, + "step": 1398 + }, + { + "epoch": 0.759500542888165, + "grad_norm": 4.510149132944334, + "learning_rate": 1.9231544953747336e-05, + "loss": 0.9981, + "step": 1399 + }, + { + "epoch": 0.760043431053203, + "grad_norm": 5.008678356958532, + "learning_rate": 1.9230192407271506e-05, + "loss": 1.4957, + "step": 1400 + }, + { + "epoch": 0.760586319218241, + "grad_norm": 4.6938647576746995, + "learning_rate": 1.9228838719190765e-05, + "loss": 1.361, + "step": 1401 + }, + { + "epoch": 0.761129207383279, + "grad_norm": 5.263462060803471, + "learning_rate": 1.9227483889672544e-05, + "loss": 1.1716, + "step": 1402 + }, + { + "epoch": 0.761672095548317, + "grad_norm": 4.737410394333335, + "learning_rate": 1.9226127918884407e-05, + "loss": 1.3924, + "step": 1403 + }, + { + "epoch": 0.762214983713355, + "grad_norm": 5.117476933198257, + "learning_rate": 1.9224770806994066e-05, + "loss": 1.1215, + "step": 1404 + }, + { + "epoch": 0.7627578718783931, + "grad_norm": 4.525492833460315, + "learning_rate": 1.922341255416937e-05, + "loss": 1.0626, + "step": 1405 + }, + { + "epoch": 0.7633007600434311, + "grad_norm": 4.928417801176309, + "learning_rate": 1.9222053160578312e-05, + "loss": 0.9576, + "step": 1406 + }, + { + "epoch": 0.7638436482084691, + "grad_norm": 4.44076437857558, + "learning_rate": 1.9220692626389018e-05, + "loss": 0.9186, + "step": 1407 + }, + { + "epoch": 0.7643865363735071, + "grad_norm": 4.779705057681976, + "learning_rate": 1.9219330951769763e-05, + "loss": 1.3392, + "step": 1408 + }, + { + "epoch": 0.7649294245385451, + "grad_norm": 5.37856562129718, + "learning_rate": 1.9217968136888965e-05, + "loss": 1.043, + "step": 1409 + }, + { + "epoch": 0.7654723127035831, + "grad_norm": 5.596447438931628, + "learning_rate": 1.9216604181915178e-05, + "loss": 1.3223, + "step": 1410 + }, + { + "epoch": 0.7660152008686211, + "grad_norm": 5.267245970848837, + "learning_rate": 1.9215239087017093e-05, + "loss": 1.0484, + "step": 1411 + }, + { + "epoch": 0.7665580890336591, + "grad_norm": 6.124227707475327, + "learning_rate": 1.9213872852363552e-05, + "loss": 1.5361, + "step": 1412 + }, + { + "epoch": 0.7671009771986971, + "grad_norm": 5.632804630433347, + "learning_rate": 1.9212505478123532e-05, + "loss": 1.2227, + "step": 1413 + }, + { + "epoch": 0.7676438653637351, + "grad_norm": 4.903635376409972, + "learning_rate": 1.9211136964466152e-05, + "loss": 0.9045, + "step": 1414 + }, + { + "epoch": 0.7681867535287731, + "grad_norm": 4.995962297168909, + "learning_rate": 1.9209767311560673e-05, + "loss": 1.2364, + "step": 1415 + }, + { + "epoch": 0.7687296416938111, + "grad_norm": 4.76685077713632, + "learning_rate": 1.9208396519576494e-05, + "loss": 1.4849, + "step": 1416 + }, + { + "epoch": 0.7692725298588491, + "grad_norm": 5.66289239913894, + "learning_rate": 1.9207024588683158e-05, + "loss": 1.389, + "step": 1417 + }, + { + "epoch": 0.7698154180238871, + "grad_norm": 4.044195406366437, + "learning_rate": 1.920565151905035e-05, + "loss": 0.5736, + "step": 1418 + }, + { + "epoch": 0.7703583061889251, + "grad_norm": 5.610527750585898, + "learning_rate": 1.9204277310847887e-05, + "loss": 1.5147, + "step": 1419 + }, + { + "epoch": 0.7709011943539631, + "grad_norm": 3.8629108074125424, + "learning_rate": 1.9202901964245734e-05, + "loss": 0.9184, + "step": 1420 + }, + { + "epoch": 0.7714440825190011, + "grad_norm": 4.582445031278247, + "learning_rate": 1.9201525479414e-05, + "loss": 1.071, + "step": 1421 + }, + { + "epoch": 0.7719869706840391, + "grad_norm": 5.016846104390101, + "learning_rate": 1.9200147856522933e-05, + "loss": 1.3673, + "step": 1422 + }, + { + "epoch": 0.7725298588490771, + "grad_norm": 4.798189213061551, + "learning_rate": 1.9198769095742914e-05, + "loss": 1.3483, + "step": 1423 + }, + { + "epoch": 0.7730727470141151, + "grad_norm": 5.871902023790772, + "learning_rate": 1.9197389197244473e-05, + "loss": 1.7625, + "step": 1424 + }, + { + "epoch": 0.7736156351791531, + "grad_norm": 5.895934775040147, + "learning_rate": 1.9196008161198277e-05, + "loss": 0.999, + "step": 1425 + }, + { + "epoch": 0.7741585233441911, + "grad_norm": 5.0199402476408155, + "learning_rate": 1.9194625987775138e-05, + "loss": 1.3251, + "step": 1426 + }, + { + "epoch": 0.7747014115092291, + "grad_norm": 6.650397672217608, + "learning_rate": 1.9193242677146e-05, + "loss": 1.2162, + "step": 1427 + }, + { + "epoch": 0.7752442996742671, + "grad_norm": 5.515267491505962, + "learning_rate": 1.9191858229481958e-05, + "loss": 1.3849, + "step": 1428 + }, + { + "epoch": 0.7757871878393051, + "grad_norm": 5.694611687374825, + "learning_rate": 1.9190472644954236e-05, + "loss": 1.0831, + "step": 1429 + }, + { + "epoch": 0.7763300760043431, + "grad_norm": 4.548114219835821, + "learning_rate": 1.9189085923734215e-05, + "loss": 1.2549, + "step": 1430 + }, + { + "epoch": 0.7768729641693811, + "grad_norm": 6.119997613777156, + "learning_rate": 1.9187698065993398e-05, + "loss": 1.6137, + "step": 1431 + }, + { + "epoch": 0.7774158523344191, + "grad_norm": 4.574150272616086, + "learning_rate": 1.9186309071903445e-05, + "loss": 1.3015, + "step": 1432 + }, + { + "epoch": 0.7779587404994571, + "grad_norm": 4.77854731853541, + "learning_rate": 1.9184918941636142e-05, + "loss": 0.6973, + "step": 1433 + }, + { + "epoch": 0.7785016286644951, + "grad_norm": 6.926122760031406, + "learning_rate": 1.9183527675363425e-05, + "loss": 1.4034, + "step": 1434 + }, + { + "epoch": 0.7790445168295331, + "grad_norm": 4.6748890605309645, + "learning_rate": 1.9182135273257372e-05, + "loss": 0.9854, + "step": 1435 + }, + { + "epoch": 0.7795874049945711, + "grad_norm": 5.072338861625223, + "learning_rate": 1.9180741735490194e-05, + "loss": 1.1604, + "step": 1436 + }, + { + "epoch": 0.7801302931596091, + "grad_norm": 5.995183838581222, + "learning_rate": 1.9179347062234245e-05, + "loss": 1.3913, + "step": 1437 + }, + { + "epoch": 0.7806731813246471, + "grad_norm": 5.494006248546126, + "learning_rate": 1.917795125366202e-05, + "loss": 1.2541, + "step": 1438 + }, + { + "epoch": 0.7812160694896851, + "grad_norm": 3.8223879163574694, + "learning_rate": 1.917655430994616e-05, + "loss": 0.7292, + "step": 1439 + }, + { + "epoch": 0.7817589576547231, + "grad_norm": 4.1476284953657405, + "learning_rate": 1.9175156231259434e-05, + "loss": 0.608, + "step": 1440 + }, + { + "epoch": 0.7823018458197611, + "grad_norm": 5.5144637244676495, + "learning_rate": 1.9173757017774764e-05, + "loss": 1.2674, + "step": 1441 + }, + { + "epoch": 0.7828447339847991, + "grad_norm": 5.4660654663594945, + "learning_rate": 1.9172356669665206e-05, + "loss": 1.3043, + "step": 1442 + }, + { + "epoch": 0.7833876221498371, + "grad_norm": 3.526312966998694, + "learning_rate": 1.9170955187103957e-05, + "loss": 0.8721, + "step": 1443 + }, + { + "epoch": 0.7839305103148752, + "grad_norm": 4.395429767339641, + "learning_rate": 1.9169552570264355e-05, + "loss": 0.985, + "step": 1444 + }, + { + "epoch": 0.7844733984799132, + "grad_norm": 4.608635000415997, + "learning_rate": 1.9168148819319874e-05, + "loss": 1.3492, + "step": 1445 + }, + { + "epoch": 0.7850162866449512, + "grad_norm": 5.729965497961937, + "learning_rate": 1.9166743934444137e-05, + "loss": 1.3405, + "step": 1446 + }, + { + "epoch": 0.7855591748099892, + "grad_norm": 6.60908414105679, + "learning_rate": 1.91653379158109e-05, + "loss": 1.5271, + "step": 1447 + }, + { + "epoch": 0.7861020629750272, + "grad_norm": 4.91386728282805, + "learning_rate": 1.916393076359406e-05, + "loss": 1.475, + "step": 1448 + }, + { + "epoch": 0.7866449511400652, + "grad_norm": 5.286158051966196, + "learning_rate": 1.916252247796766e-05, + "loss": 1.5235, + "step": 1449 + }, + { + "epoch": 0.7871878393051032, + "grad_norm": 4.699459928019414, + "learning_rate": 1.916111305910588e-05, + "loss": 1.151, + "step": 1450 + }, + { + "epoch": 0.7877307274701412, + "grad_norm": 5.125659016648778, + "learning_rate": 1.915970250718303e-05, + "loss": 1.2952, + "step": 1451 + }, + { + "epoch": 0.7882736156351792, + "grad_norm": 4.976170683421487, + "learning_rate": 1.915829082237358e-05, + "loss": 1.3291, + "step": 1452 + }, + { + "epoch": 0.7888165038002172, + "grad_norm": 5.237853683905863, + "learning_rate": 1.9156878004852123e-05, + "loss": 1.4775, + "step": 1453 + }, + { + "epoch": 0.7893593919652552, + "grad_norm": 4.914941394388547, + "learning_rate": 1.9155464054793404e-05, + "loss": 1.2151, + "step": 1454 + }, + { + "epoch": 0.7899022801302932, + "grad_norm": 5.050785426148085, + "learning_rate": 1.9154048972372293e-05, + "loss": 1.163, + "step": 1455 + }, + { + "epoch": 0.7904451682953312, + "grad_norm": 4.2016259300832255, + "learning_rate": 1.915263275776382e-05, + "loss": 0.9601, + "step": 1456 + }, + { + "epoch": 0.7909880564603692, + "grad_norm": 5.48804064939896, + "learning_rate": 1.915121541114314e-05, + "loss": 1.3026, + "step": 1457 + }, + { + "epoch": 0.7915309446254072, + "grad_norm": 5.230495684608947, + "learning_rate": 1.9149796932685552e-05, + "loss": 1.1923, + "step": 1458 + }, + { + "epoch": 0.7920738327904452, + "grad_norm": 4.788655104859546, + "learning_rate": 1.91483773225665e-05, + "loss": 1.2437, + "step": 1459 + }, + { + "epoch": 0.7926167209554832, + "grad_norm": 4.506216689801701, + "learning_rate": 1.9146956580961556e-05, + "loss": 0.9364, + "step": 1460 + }, + { + "epoch": 0.7931596091205212, + "grad_norm": 4.2335852812311865, + "learning_rate": 1.9145534708046446e-05, + "loss": 0.7104, + "step": 1461 + }, + { + "epoch": 0.7937024972855592, + "grad_norm": 6.204688311211956, + "learning_rate": 1.914411170399703e-05, + "loss": 1.0825, + "step": 1462 + }, + { + "epoch": 0.7942453854505972, + "grad_norm": 3.445489329210515, + "learning_rate": 1.91426875689893e-05, + "loss": 0.9921, + "step": 1463 + }, + { + "epoch": 0.7947882736156352, + "grad_norm": 5.204416925095863, + "learning_rate": 1.9141262303199403e-05, + "loss": 1.3043, + "step": 1464 + }, + { + "epoch": 0.7953311617806732, + "grad_norm": 6.3486214559668985, + "learning_rate": 1.9139835906803612e-05, + "loss": 1.3193, + "step": 1465 + }, + { + "epoch": 0.7958740499457112, + "grad_norm": 5.610159156463615, + "learning_rate": 1.913840837997835e-05, + "loss": 1.0455, + "step": 1466 + }, + { + "epoch": 0.7964169381107492, + "grad_norm": 4.625080334899242, + "learning_rate": 1.913697972290018e-05, + "loss": 0.7981, + "step": 1467 + }, + { + "epoch": 0.7969598262757872, + "grad_norm": 5.993005746484773, + "learning_rate": 1.9135549935745792e-05, + "loss": 1.1674, + "step": 1468 + }, + { + "epoch": 0.7975027144408252, + "grad_norm": 5.084136125969368, + "learning_rate": 1.913411901869203e-05, + "loss": 1.0362, + "step": 1469 + }, + { + "epoch": 0.7980456026058632, + "grad_norm": 5.701736254232889, + "learning_rate": 1.913268697191587e-05, + "loss": 1.2159, + "step": 1470 + }, + { + "epoch": 0.7985884907709012, + "grad_norm": 4.715190322082246, + "learning_rate": 1.9131253795594428e-05, + "loss": 0.9848, + "step": 1471 + }, + { + "epoch": 0.7991313789359392, + "grad_norm": 5.375233653105075, + "learning_rate": 1.9129819489904964e-05, + "loss": 1.0476, + "step": 1472 + }, + { + "epoch": 0.7996742671009772, + "grad_norm": 6.352674291337691, + "learning_rate": 1.9128384055024874e-05, + "loss": 1.2362, + "step": 1473 + }, + { + "epoch": 0.8002171552660152, + "grad_norm": 5.46716259791096, + "learning_rate": 1.91269474911317e-05, + "loss": 1.2949, + "step": 1474 + }, + { + "epoch": 0.8007600434310532, + "grad_norm": 5.3150808603597826, + "learning_rate": 1.912550979840311e-05, + "loss": 1.1587, + "step": 1475 + }, + { + "epoch": 0.8013029315960912, + "grad_norm": 4.46814887402293, + "learning_rate": 1.9124070977016926e-05, + "loss": 0.9649, + "step": 1476 + }, + { + "epoch": 0.8018458197611292, + "grad_norm": 4.916267658604107, + "learning_rate": 1.9122631027151103e-05, + "loss": 1.2117, + "step": 1477 + }, + { + "epoch": 0.8023887079261672, + "grad_norm": 5.506988713852874, + "learning_rate": 1.9121189948983733e-05, + "loss": 1.3387, + "step": 1478 + }, + { + "epoch": 0.8029315960912052, + "grad_norm": 4.7704152930487895, + "learning_rate": 1.911974774269305e-05, + "loss": 1.3379, + "step": 1479 + }, + { + "epoch": 0.8034744842562432, + "grad_norm": 5.667769720352476, + "learning_rate": 1.9118304408457435e-05, + "loss": 1.0552, + "step": 1480 + }, + { + "epoch": 0.8040173724212812, + "grad_norm": 4.9471232370904925, + "learning_rate": 1.91168599464554e-05, + "loss": 1.0013, + "step": 1481 + }, + { + "epoch": 0.8045602605863192, + "grad_norm": 4.856623754785127, + "learning_rate": 1.9115414356865594e-05, + "loss": 0.9001, + "step": 1482 + }, + { + "epoch": 0.8051031487513572, + "grad_norm": 4.004976425614515, + "learning_rate": 1.9113967639866815e-05, + "loss": 0.8114, + "step": 1483 + }, + { + "epoch": 0.8056460369163952, + "grad_norm": 5.983237168687411, + "learning_rate": 1.911251979563799e-05, + "loss": 1.1678, + "step": 1484 + }, + { + "epoch": 0.8061889250814332, + "grad_norm": 4.808701495369871, + "learning_rate": 1.9111070824358196e-05, + "loss": 0.9181, + "step": 1485 + }, + { + "epoch": 0.8067318132464713, + "grad_norm": 5.080557830592386, + "learning_rate": 1.910962072620664e-05, + "loss": 1.0751, + "step": 1486 + }, + { + "epoch": 0.8072747014115093, + "grad_norm": 5.730412968009966, + "learning_rate": 1.9108169501362674e-05, + "loss": 1.2727, + "step": 1487 + }, + { + "epoch": 0.8078175895765473, + "grad_norm": 6.47838124684283, + "learning_rate": 1.9106717150005785e-05, + "loss": 1.6491, + "step": 1488 + }, + { + "epoch": 0.8083604777415853, + "grad_norm": 5.311235841870027, + "learning_rate": 1.910526367231561e-05, + "loss": 0.8382, + "step": 1489 + }, + { + "epoch": 0.8089033659066233, + "grad_norm": 5.43072146168114, + "learning_rate": 1.9103809068471914e-05, + "loss": 1.3026, + "step": 1490 + }, + { + "epoch": 0.8094462540716613, + "grad_norm": 4.46156225654375, + "learning_rate": 1.9102353338654597e-05, + "loss": 0.8071, + "step": 1491 + }, + { + "epoch": 0.8099891422366993, + "grad_norm": 5.297520452299967, + "learning_rate": 1.9100896483043714e-05, + "loss": 1.0625, + "step": 1492 + }, + { + "epoch": 0.8105320304017373, + "grad_norm": 6.736735698357962, + "learning_rate": 1.909943850181945e-05, + "loss": 1.6497, + "step": 1493 + }, + { + "epoch": 0.8110749185667753, + "grad_norm": 4.8675910913183955, + "learning_rate": 1.9097979395162132e-05, + "loss": 1.0822, + "step": 1494 + }, + { + "epoch": 0.8116178067318133, + "grad_norm": 4.5489525578867305, + "learning_rate": 1.909651916325222e-05, + "loss": 0.7908, + "step": 1495 + }, + { + "epoch": 0.8121606948968513, + "grad_norm": 6.002121574887706, + "learning_rate": 1.909505780627032e-05, + "loss": 0.5178, + "step": 1496 + }, + { + "epoch": 0.8127035830618893, + "grad_norm": 4.34564271493731, + "learning_rate": 1.9093595324397175e-05, + "loss": 0.9818, + "step": 1497 + }, + { + "epoch": 0.8132464712269273, + "grad_norm": 5.0382562486474525, + "learning_rate": 1.9092131717813668e-05, + "loss": 1.0997, + "step": 1498 + }, + { + "epoch": 0.8137893593919653, + "grad_norm": 6.8289191133749, + "learning_rate": 1.909066698670082e-05, + "loss": 1.0046, + "step": 1499 + }, + { + "epoch": 0.8143322475570033, + "grad_norm": 6.6968746410007585, + "learning_rate": 1.908920113123979e-05, + "loss": 1.5191, + "step": 1500 + }, + { + "epoch": 0.8148751357220413, + "grad_norm": 5.548259576405749, + "learning_rate": 1.9087734151611877e-05, + "loss": 1.3272, + "step": 1501 + }, + { + "epoch": 0.8154180238870793, + "grad_norm": 5.684602861026239, + "learning_rate": 1.9086266047998522e-05, + "loss": 1.0698, + "step": 1502 + }, + { + "epoch": 0.8159609120521173, + "grad_norm": 6.832627276636624, + "learning_rate": 1.90847968205813e-05, + "loss": 1.5796, + "step": 1503 + }, + { + "epoch": 0.8165038002171553, + "grad_norm": 5.605724679057613, + "learning_rate": 1.908332646954193e-05, + "loss": 1.1709, + "step": 1504 + }, + { + "epoch": 0.8170466883821933, + "grad_norm": 4.607345782065147, + "learning_rate": 1.908185499506226e-05, + "loss": 0.8686, + "step": 1505 + }, + { + "epoch": 0.8175895765472313, + "grad_norm": 4.687252800873217, + "learning_rate": 1.9080382397324296e-05, + "loss": 0.8881, + "step": 1506 + }, + { + "epoch": 0.8181324647122693, + "grad_norm": 6.357714385620256, + "learning_rate": 1.907890867651016e-05, + "loss": 1.1948, + "step": 1507 + }, + { + "epoch": 0.8186753528773073, + "grad_norm": 4.8608334992124425, + "learning_rate": 1.9077433832802135e-05, + "loss": 1.0311, + "step": 1508 + }, + { + "epoch": 0.8192182410423453, + "grad_norm": 5.948864135911491, + "learning_rate": 1.9075957866382623e-05, + "loss": 1.1314, + "step": 1509 + }, + { + "epoch": 0.8197611292073833, + "grad_norm": 4.807850657807276, + "learning_rate": 1.9074480777434178e-05, + "loss": 0.9478, + "step": 1510 + }, + { + "epoch": 0.8203040173724213, + "grad_norm": 4.26222909817726, + "learning_rate": 1.9073002566139486e-05, + "loss": 0.8541, + "step": 1511 + }, + { + "epoch": 0.8208469055374593, + "grad_norm": 6.6130092616391005, + "learning_rate": 1.9071523232681382e-05, + "loss": 1.2754, + "step": 1512 + }, + { + "epoch": 0.8213897937024973, + "grad_norm": 5.255155673043404, + "learning_rate": 1.907004277724282e-05, + "loss": 1.0813, + "step": 1513 + }, + { + "epoch": 0.8219326818675353, + "grad_norm": 5.079098182105948, + "learning_rate": 1.9068561200006917e-05, + "loss": 1.0016, + "step": 1514 + }, + { + "epoch": 0.8224755700325733, + "grad_norm": 5.491172879434626, + "learning_rate": 1.906707850115691e-05, + "loss": 1.2884, + "step": 1515 + }, + { + "epoch": 0.8230184581976113, + "grad_norm": 6.718962166599785, + "learning_rate": 1.9065594680876182e-05, + "loss": 1.4973, + "step": 1516 + }, + { + "epoch": 0.8235613463626493, + "grad_norm": 5.348428262646105, + "learning_rate": 1.9064109739348257e-05, + "loss": 1.1113, + "step": 1517 + }, + { + "epoch": 0.8241042345276873, + "grad_norm": 4.604059650726469, + "learning_rate": 1.906262367675679e-05, + "loss": 1.1614, + "step": 1518 + }, + { + "epoch": 0.8246471226927253, + "grad_norm": 5.895860662978225, + "learning_rate": 1.9061136493285586e-05, + "loss": 1.1532, + "step": 1519 + }, + { + "epoch": 0.8251900108577633, + "grad_norm": 5.621847434524929, + "learning_rate": 1.905964818911858e-05, + "loss": 0.9277, + "step": 1520 + }, + { + "epoch": 0.8257328990228013, + "grad_norm": 4.585955492662189, + "learning_rate": 1.9058158764439844e-05, + "loss": 0.7988, + "step": 1521 + }, + { + "epoch": 0.8262757871878393, + "grad_norm": 5.784308925988881, + "learning_rate": 1.9056668219433595e-05, + "loss": 1.6078, + "step": 1522 + }, + { + "epoch": 0.8268186753528773, + "grad_norm": 5.621216253388429, + "learning_rate": 1.905517655428419e-05, + "loss": 1.389, + "step": 1523 + }, + { + "epoch": 0.8273615635179153, + "grad_norm": 5.514208208061458, + "learning_rate": 1.9053683769176115e-05, + "loss": 0.7612, + "step": 1524 + }, + { + "epoch": 0.8279044516829533, + "grad_norm": 4.6572185133859065, + "learning_rate": 1.9052189864294002e-05, + "loss": 0.744, + "step": 1525 + }, + { + "epoch": 0.8284473398479913, + "grad_norm": 5.239632270223703, + "learning_rate": 1.905069483982262e-05, + "loss": 0.939, + "step": 1526 + }, + { + "epoch": 0.8289902280130294, + "grad_norm": 3.7037650901217454, + "learning_rate": 1.9049198695946876e-05, + "loss": 1.0177, + "step": 1527 + }, + { + "epoch": 0.8295331161780674, + "grad_norm": 6.974869726679236, + "learning_rate": 1.9047701432851813e-05, + "loss": 1.2722, + "step": 1528 + }, + { + "epoch": 0.8300760043431054, + "grad_norm": 6.642878483620589, + "learning_rate": 1.904620305072262e-05, + "loss": 1.5369, + "step": 1529 + }, + { + "epoch": 0.8306188925081434, + "grad_norm": 7.99082645392899, + "learning_rate": 1.9044703549744616e-05, + "loss": 1.2245, + "step": 1530 + }, + { + "epoch": 0.8311617806731814, + "grad_norm": 6.5593948883008135, + "learning_rate": 1.904320293010326e-05, + "loss": 1.307, + "step": 1531 + }, + { + "epoch": 0.8317046688382194, + "grad_norm": 4.930764759519961, + "learning_rate": 1.9041701191984155e-05, + "loss": 0.9564, + "step": 1532 + }, + { + "epoch": 0.8322475570032574, + "grad_norm": 3.9621684331427773, + "learning_rate": 1.9040198335573033e-05, + "loss": 0.8153, + "step": 1533 + }, + { + "epoch": 0.8327904451682954, + "grad_norm": 4.859084711241092, + "learning_rate": 1.9038694361055774e-05, + "loss": 1.0967, + "step": 1534 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.9476278172327595, + "learning_rate": 1.903718926861839e-05, + "loss": 1.4743, + "step": 1535 + }, + { + "epoch": 0.8338762214983714, + "grad_norm": 4.312795088472635, + "learning_rate": 1.903568305844704e-05, + "loss": 0.8249, + "step": 1536 + }, + { + "epoch": 0.8344191096634094, + "grad_norm": 6.638027572281857, + "learning_rate": 1.9034175730728e-05, + "loss": 1.6353, + "step": 1537 + }, + { + "epoch": 0.8349619978284474, + "grad_norm": 5.709548912096025, + "learning_rate": 1.9032667285647714e-05, + "loss": 1.8512, + "step": 1538 + }, + { + "epoch": 0.8355048859934854, + "grad_norm": 5.098275426179771, + "learning_rate": 1.9031157723392738e-05, + "loss": 1.0348, + "step": 1539 + }, + { + "epoch": 0.8360477741585234, + "grad_norm": 5.05236269448673, + "learning_rate": 1.9029647044149783e-05, + "loss": 1.108, + "step": 1540 + }, + { + "epoch": 0.8365906623235614, + "grad_norm": 4.674708875536442, + "learning_rate": 1.9028135248105692e-05, + "loss": 1.0453, + "step": 1541 + }, + { + "epoch": 0.8371335504885994, + "grad_norm": 4.717321956389267, + "learning_rate": 1.902662233544744e-05, + "loss": 0.9042, + "step": 1542 + }, + { + "epoch": 0.8376764386536374, + "grad_norm": 6.217040367936212, + "learning_rate": 1.9025108306362158e-05, + "loss": 1.0762, + "step": 1543 + }, + { + "epoch": 0.8382193268186754, + "grad_norm": 5.338744288323046, + "learning_rate": 1.9023593161037094e-05, + "loss": 1.1631, + "step": 1544 + }, + { + "epoch": 0.8387622149837134, + "grad_norm": 4.8224874086687874, + "learning_rate": 1.9022076899659643e-05, + "loss": 1.3907, + "step": 1545 + }, + { + "epoch": 0.8393051031487514, + "grad_norm": 4.699800413179793, + "learning_rate": 1.9020559522417345e-05, + "loss": 0.7682, + "step": 1546 + }, + { + "epoch": 0.8398479913137894, + "grad_norm": 4.271439794022252, + "learning_rate": 1.9019041029497866e-05, + "loss": 0.8475, + "step": 1547 + }, + { + "epoch": 0.8403908794788274, + "grad_norm": 4.240562925498168, + "learning_rate": 1.9017521421089022e-05, + "loss": 0.8201, + "step": 1548 + }, + { + "epoch": 0.8409337676438654, + "grad_norm": 5.24718786534657, + "learning_rate": 1.9016000697378755e-05, + "loss": 1.0728, + "step": 1549 + }, + { + "epoch": 0.8414766558089034, + "grad_norm": 6.492220902503762, + "learning_rate": 1.9014478858555156e-05, + "loss": 0.9432, + "step": 1550 + }, + { + "epoch": 0.8420195439739414, + "grad_norm": 4.917811923935393, + "learning_rate": 1.9012955904806438e-05, + "loss": 1.1672, + "step": 1551 + }, + { + "epoch": 0.8425624321389794, + "grad_norm": 5.66709669173795, + "learning_rate": 1.9011431836320976e-05, + "loss": 1.5058, + "step": 1552 + }, + { + "epoch": 0.8431053203040174, + "grad_norm": 5.658634152413846, + "learning_rate": 1.9009906653287258e-05, + "loss": 1.3653, + "step": 1553 + }, + { + "epoch": 0.8436482084690554, + "grad_norm": 4.956250367213818, + "learning_rate": 1.9008380355893925e-05, + "loss": 0.8309, + "step": 1554 + }, + { + "epoch": 0.8441910966340934, + "grad_norm": 4.329497016180362, + "learning_rate": 1.9006852944329753e-05, + "loss": 1.1141, + "step": 1555 + }, + { + "epoch": 0.8447339847991314, + "grad_norm": 5.717864943326053, + "learning_rate": 1.9005324418783658e-05, + "loss": 1.3274, + "step": 1556 + }, + { + "epoch": 0.8452768729641694, + "grad_norm": 5.657300743895006, + "learning_rate": 1.900379477944468e-05, + "loss": 1.2432, + "step": 1557 + }, + { + "epoch": 0.8458197611292074, + "grad_norm": 5.325761071371062, + "learning_rate": 1.900226402650202e-05, + "loss": 1.3428, + "step": 1558 + }, + { + "epoch": 0.8463626492942454, + "grad_norm": 5.207304704126981, + "learning_rate": 1.9000732160144996e-05, + "loss": 1.455, + "step": 1559 + }, + { + "epoch": 0.8469055374592834, + "grad_norm": 5.248656371435267, + "learning_rate": 1.8999199180563074e-05, + "loss": 0.7851, + "step": 1560 + }, + { + "epoch": 0.8474484256243214, + "grad_norm": 6.92723307445887, + "learning_rate": 1.899766508794585e-05, + "loss": 1.5236, + "step": 1561 + }, + { + "epoch": 0.8479913137893594, + "grad_norm": 4.3434986726191935, + "learning_rate": 1.899612988248307e-05, + "loss": 0.9117, + "step": 1562 + }, + { + "epoch": 0.8485342019543974, + "grad_norm": 5.815800316925401, + "learning_rate": 1.8994593564364612e-05, + "loss": 1.0097, + "step": 1563 + }, + { + "epoch": 0.8490770901194354, + "grad_norm": 5.5494362406379905, + "learning_rate": 1.8993056133780484e-05, + "loss": 1.1829, + "step": 1564 + }, + { + "epoch": 0.8496199782844734, + "grad_norm": 4.323494913553237, + "learning_rate": 1.899151759092084e-05, + "loss": 0.9359, + "step": 1565 + }, + { + "epoch": 0.8501628664495114, + "grad_norm": 4.455911135073797, + "learning_rate": 1.898997793597597e-05, + "loss": 0.9354, + "step": 1566 + }, + { + "epoch": 0.8507057546145494, + "grad_norm": 6.684553981467422, + "learning_rate": 1.8988437169136302e-05, + "loss": 1.3693, + "step": 1567 + }, + { + "epoch": 0.8512486427795874, + "grad_norm": 6.341757024960091, + "learning_rate": 1.89868952905924e-05, + "loss": 0.9441, + "step": 1568 + }, + { + "epoch": 0.8517915309446255, + "grad_norm": 6.017605683279906, + "learning_rate": 1.8985352300534965e-05, + "loss": 1.2747, + "step": 1569 + }, + { + "epoch": 0.8523344191096635, + "grad_norm": 5.501152081837135, + "learning_rate": 1.8983808199154835e-05, + "loss": 0.8414, + "step": 1570 + }, + { + "epoch": 0.8528773072747015, + "grad_norm": 5.075164131477861, + "learning_rate": 1.8982262986642993e-05, + "loss": 1.124, + "step": 1571 + }, + { + "epoch": 0.8534201954397395, + "grad_norm": 5.7750932469486065, + "learning_rate": 1.8980716663190545e-05, + "loss": 1.3968, + "step": 1572 + }, + { + "epoch": 0.8539630836047775, + "grad_norm": 6.967872140444835, + "learning_rate": 1.897916922898875e-05, + "loss": 1.2395, + "step": 1573 + }, + { + "epoch": 0.8545059717698155, + "grad_norm": 5.037804725894728, + "learning_rate": 1.8977620684228994e-05, + "loss": 0.9741, + "step": 1574 + }, + { + "epoch": 0.8550488599348535, + "grad_norm": 4.586000206489282, + "learning_rate": 1.8976071029102802e-05, + "loss": 0.6757, + "step": 1575 + }, + { + "epoch": 0.8555917480998915, + "grad_norm": 4.686417605581938, + "learning_rate": 1.897452026380184e-05, + "loss": 0.8382, + "step": 1576 + }, + { + "epoch": 0.8561346362649295, + "grad_norm": 5.133601528061789, + "learning_rate": 1.8972968388517908e-05, + "loss": 1.2999, + "step": 1577 + }, + { + "epoch": 0.8566775244299675, + "grad_norm": 6.170796021401621, + "learning_rate": 1.8971415403442942e-05, + "loss": 1.1513, + "step": 1578 + }, + { + "epoch": 0.8572204125950055, + "grad_norm": 5.229050623246509, + "learning_rate": 1.8969861308769025e-05, + "loss": 0.9371, + "step": 1579 + }, + { + "epoch": 0.8577633007600435, + "grad_norm": 5.081229057143735, + "learning_rate": 1.8968306104688365e-05, + "loss": 0.9398, + "step": 1580 + }, + { + "epoch": 0.8583061889250815, + "grad_norm": 4.641876540865143, + "learning_rate": 1.896674979139331e-05, + "loss": 1.1401, + "step": 1581 + }, + { + "epoch": 0.8588490770901195, + "grad_norm": 4.149586485324207, + "learning_rate": 1.8965192369076356e-05, + "loss": 0.6768, + "step": 1582 + }, + { + "epoch": 0.8593919652551575, + "grad_norm": 6.422772444923944, + "learning_rate": 1.8963633837930114e-05, + "loss": 1.392, + "step": 1583 + }, + { + "epoch": 0.8599348534201955, + "grad_norm": 4.860635992388028, + "learning_rate": 1.8962074198147357e-05, + "loss": 0.9983, + "step": 1584 + }, + { + "epoch": 0.8604777415852335, + "grad_norm": 4.216877618358879, + "learning_rate": 1.8960513449920982e-05, + "loss": 0.6958, + "step": 1585 + }, + { + "epoch": 0.8610206297502715, + "grad_norm": 5.012340070254638, + "learning_rate": 1.8958951593444017e-05, + "loss": 0.9859, + "step": 1586 + }, + { + "epoch": 0.8615635179153095, + "grad_norm": 5.467321056155664, + "learning_rate": 1.8957388628909644e-05, + "loss": 1.3782, + "step": 1587 + }, + { + "epoch": 0.8621064060803475, + "grad_norm": 5.59589685839463, + "learning_rate": 1.8955824556511168e-05, + "loss": 0.8297, + "step": 1588 + }, + { + "epoch": 0.8626492942453855, + "grad_norm": 6.519624998298618, + "learning_rate": 1.895425937644204e-05, + "loss": 1.7264, + "step": 1589 + }, + { + "epoch": 0.8631921824104235, + "grad_norm": 6.6951270307269795, + "learning_rate": 1.8952693088895837e-05, + "loss": 1.2919, + "step": 1590 + }, + { + "epoch": 0.8637350705754615, + "grad_norm": 4.348741825803296, + "learning_rate": 1.895112569406629e-05, + "loss": 1.0062, + "step": 1591 + }, + { + "epoch": 0.8642779587404995, + "grad_norm": 5.0213162819553565, + "learning_rate": 1.8949557192147243e-05, + "loss": 1.2381, + "step": 1592 + }, + { + "epoch": 0.8648208469055375, + "grad_norm": 6.1637021381056885, + "learning_rate": 1.8947987583332705e-05, + "loss": 1.2561, + "step": 1593 + }, + { + "epoch": 0.8653637350705755, + "grad_norm": 6.638257635344733, + "learning_rate": 1.89464168678168e-05, + "loss": 0.9861, + "step": 1594 + }, + { + "epoch": 0.8659066232356135, + "grad_norm": 4.861237127098975, + "learning_rate": 1.89448450457938e-05, + "loss": 0.9407, + "step": 1595 + }, + { + "epoch": 0.8664495114006515, + "grad_norm": 6.4248430682281565, + "learning_rate": 1.894327211745811e-05, + "loss": 1.3079, + "step": 1596 + }, + { + "epoch": 0.8669923995656895, + "grad_norm": 5.712502208347676, + "learning_rate": 1.8941698083004265e-05, + "loss": 1.4422, + "step": 1597 + }, + { + "epoch": 0.8675352877307275, + "grad_norm": 5.634308307822415, + "learning_rate": 1.8940122942626957e-05, + "loss": 0.7869, + "step": 1598 + }, + { + "epoch": 0.8680781758957655, + "grad_norm": 7.839535783401759, + "learning_rate": 1.893854669652099e-05, + "loss": 1.3397, + "step": 1599 + }, + { + "epoch": 0.8686210640608035, + "grad_norm": 5.156814072355032, + "learning_rate": 1.8936969344881323e-05, + "loss": 1.2541, + "step": 1600 + }, + { + "epoch": 0.8691639522258415, + "grad_norm": 4.912293322601371, + "learning_rate": 1.8935390887903044e-05, + "loss": 0.8418, + "step": 1601 + }, + { + "epoch": 0.8697068403908795, + "grad_norm": 5.436641391280079, + "learning_rate": 1.8933811325781382e-05, + "loss": 1.4456, + "step": 1602 + }, + { + "epoch": 0.8702497285559175, + "grad_norm": 5.221596044730403, + "learning_rate": 1.8932230658711696e-05, + "loss": 0.9626, + "step": 1603 + }, + { + "epoch": 0.8707926167209555, + "grad_norm": 6.103653494209826, + "learning_rate": 1.8930648886889482e-05, + "loss": 1.2338, + "step": 1604 + }, + { + "epoch": 0.8713355048859935, + "grad_norm": 7.003639368124227, + "learning_rate": 1.8929066010510383e-05, + "loss": 1.2216, + "step": 1605 + }, + { + "epoch": 0.8718783930510315, + "grad_norm": 4.911905632269173, + "learning_rate": 1.8927482029770168e-05, + "loss": 0.9049, + "step": 1606 + }, + { + "epoch": 0.8724212812160695, + "grad_norm": 6.079178298711795, + "learning_rate": 1.8925896944864748e-05, + "loss": 1.6408, + "step": 1607 + }, + { + "epoch": 0.8729641693811075, + "grad_norm": 6.045568168649525, + "learning_rate": 1.892431075599017e-05, + "loss": 1.084, + "step": 1608 + }, + { + "epoch": 0.8735070575461455, + "grad_norm": 4.827633470123435, + "learning_rate": 1.892272346334261e-05, + "loss": 0.8121, + "step": 1609 + }, + { + "epoch": 0.8740499457111836, + "grad_norm": 5.314535878915851, + "learning_rate": 1.8921135067118396e-05, + "loss": 1.2657, + "step": 1610 + }, + { + "epoch": 0.8745928338762216, + "grad_norm": 4.786606110240068, + "learning_rate": 1.8919545567513976e-05, + "loss": 1.0534, + "step": 1611 + }, + { + "epoch": 0.8751357220412594, + "grad_norm": 4.075700855026623, + "learning_rate": 1.8917954964725948e-05, + "loss": 0.7757, + "step": 1612 + }, + { + "epoch": 0.8756786102062974, + "grad_norm": 6.286980562376595, + "learning_rate": 1.8916363258951033e-05, + "loss": 0.9782, + "step": 1613 + }, + { + "epoch": 0.8762214983713354, + "grad_norm": 6.745303990081763, + "learning_rate": 1.8914770450386102e-05, + "loss": 0.9488, + "step": 1614 + }, + { + "epoch": 0.8767643865363735, + "grad_norm": 6.2663007227218275, + "learning_rate": 1.8913176539228152e-05, + "loss": 1.9529, + "step": 1615 + }, + { + "epoch": 0.8773072747014115, + "grad_norm": 4.793153881320079, + "learning_rate": 1.8911581525674324e-05, + "loss": 1.0242, + "step": 1616 + }, + { + "epoch": 0.8778501628664495, + "grad_norm": 4.6844946413894, + "learning_rate": 1.890998540992189e-05, + "loss": 0.8565, + "step": 1617 + }, + { + "epoch": 0.8783930510314875, + "grad_norm": 3.993737542716379, + "learning_rate": 1.8908388192168256e-05, + "loss": 0.6107, + "step": 1618 + }, + { + "epoch": 0.8789359391965255, + "grad_norm": 6.099718972513945, + "learning_rate": 1.8906789872610977e-05, + "loss": 0.8043, + "step": 1619 + }, + { + "epoch": 0.8794788273615635, + "grad_norm": 7.113418049135383, + "learning_rate": 1.8905190451447726e-05, + "loss": 1.241, + "step": 1620 + }, + { + "epoch": 0.8800217155266015, + "grad_norm": 4.439254236407197, + "learning_rate": 1.8903589928876337e-05, + "loss": 1.0627, + "step": 1621 + }, + { + "epoch": 0.8805646036916395, + "grad_norm": 5.773948255627926, + "learning_rate": 1.8901988305094746e-05, + "loss": 1.3241, + "step": 1622 + }, + { + "epoch": 0.8811074918566775, + "grad_norm": 6.231870498005159, + "learning_rate": 1.890038558030106e-05, + "loss": 1.5241, + "step": 1623 + }, + { + "epoch": 0.8816503800217155, + "grad_norm": 5.888612961801449, + "learning_rate": 1.8898781754693495e-05, + "loss": 1.109, + "step": 1624 + }, + { + "epoch": 0.8821932681867535, + "grad_norm": 4.880160750019388, + "learning_rate": 1.8897176828470424e-05, + "loss": 1.0124, + "step": 1625 + }, + { + "epoch": 0.8827361563517915, + "grad_norm": 6.365487465081988, + "learning_rate": 1.889557080183034e-05, + "loss": 0.8999, + "step": 1626 + }, + { + "epoch": 0.8832790445168295, + "grad_norm": 4.461166628483842, + "learning_rate": 1.8893963674971883e-05, + "loss": 0.8762, + "step": 1627 + }, + { + "epoch": 0.8838219326818675, + "grad_norm": 5.675449614336963, + "learning_rate": 1.8892355448093825e-05, + "loss": 1.049, + "step": 1628 + }, + { + "epoch": 0.8843648208469055, + "grad_norm": 4.320721965444944, + "learning_rate": 1.8890746121395072e-05, + "loss": 1.1291, + "step": 1629 + }, + { + "epoch": 0.8849077090119435, + "grad_norm": 5.155835351166163, + "learning_rate": 1.8889135695074668e-05, + "loss": 1.1035, + "step": 1630 + }, + { + "epoch": 0.8854505971769815, + "grad_norm": 6.506050711845877, + "learning_rate": 1.8887524169331794e-05, + "loss": 2.1954, + "step": 1631 + }, + { + "epoch": 0.8859934853420195, + "grad_norm": 5.514989348863087, + "learning_rate": 1.8885911544365766e-05, + "loss": 1.0237, + "step": 1632 + }, + { + "epoch": 0.8865363735070575, + "grad_norm": 5.123449587076818, + "learning_rate": 1.8884297820376038e-05, + "loss": 0.9908, + "step": 1633 + }, + { + "epoch": 0.8870792616720955, + "grad_norm": 5.421671567850239, + "learning_rate": 1.8882682997562197e-05, + "loss": 1.7734, + "step": 1634 + }, + { + "epoch": 0.8876221498371335, + "grad_norm": 7.1012436118787186, + "learning_rate": 1.8881067076123963e-05, + "loss": 1.5488, + "step": 1635 + }, + { + "epoch": 0.8881650380021715, + "grad_norm": 5.328780061143714, + "learning_rate": 1.88794500562612e-05, + "loss": 1.3832, + "step": 1636 + }, + { + "epoch": 0.8887079261672095, + "grad_norm": 5.1503624666971595, + "learning_rate": 1.88778319381739e-05, + "loss": 0.8524, + "step": 1637 + }, + { + "epoch": 0.8892508143322475, + "grad_norm": 5.131553278195334, + "learning_rate": 1.88762127220622e-05, + "loss": 0.9769, + "step": 1638 + }, + { + "epoch": 0.8897937024972855, + "grad_norm": 3.1519686008526135, + "learning_rate": 1.8874592408126365e-05, + "loss": 0.472, + "step": 1639 + }, + { + "epoch": 0.8903365906623235, + "grad_norm": 4.164586140054025, + "learning_rate": 1.8872970996566794e-05, + "loss": 0.67, + "step": 1640 + }, + { + "epoch": 0.8908794788273615, + "grad_norm": 7.488659246842809, + "learning_rate": 1.8871348487584028e-05, + "loss": 1.3141, + "step": 1641 + }, + { + "epoch": 0.8914223669923995, + "grad_norm": 5.911298710361154, + "learning_rate": 1.8869724881378743e-05, + "loss": 1.084, + "step": 1642 + }, + { + "epoch": 0.8919652551574375, + "grad_norm": 5.977793625957034, + "learning_rate": 1.886810017815175e-05, + "loss": 0.9273, + "step": 1643 + }, + { + "epoch": 0.8925081433224755, + "grad_norm": 5.2773273690434515, + "learning_rate": 1.8866474378103993e-05, + "loss": 1.2594, + "step": 1644 + }, + { + "epoch": 0.8930510314875135, + "grad_norm": 8.049487951903771, + "learning_rate": 1.8864847481436554e-05, + "loss": 1.0104, + "step": 1645 + }, + { + "epoch": 0.8935939196525515, + "grad_norm": 5.428394606766542, + "learning_rate": 1.886321948835065e-05, + "loss": 0.827, + "step": 1646 + }, + { + "epoch": 0.8941368078175895, + "grad_norm": 5.687133062399138, + "learning_rate": 1.8861590399047635e-05, + "loss": 1.2759, + "step": 1647 + }, + { + "epoch": 0.8946796959826275, + "grad_norm": 5.037591954761082, + "learning_rate": 1.885996021372899e-05, + "loss": 0.8928, + "step": 1648 + }, + { + "epoch": 0.8952225841476655, + "grad_norm": 4.911533421640023, + "learning_rate": 1.8858328932596352e-05, + "loss": 0.705, + "step": 1649 + }, + { + "epoch": 0.8957654723127035, + "grad_norm": 6.054608860222289, + "learning_rate": 1.885669655585147e-05, + "loss": 1.296, + "step": 1650 + }, + { + "epoch": 0.8963083604777415, + "grad_norm": 6.221099773279408, + "learning_rate": 1.8855063083696244e-05, + "loss": 1.5253, + "step": 1651 + }, + { + "epoch": 0.8968512486427795, + "grad_norm": 5.0718329796444035, + "learning_rate": 1.8853428516332702e-05, + "loss": 1.1683, + "step": 1652 + }, + { + "epoch": 0.8973941368078175, + "grad_norm": 5.721985007773526, + "learning_rate": 1.8851792853963015e-05, + "loss": 1.1635, + "step": 1653 + }, + { + "epoch": 0.8979370249728555, + "grad_norm": 7.224045168649458, + "learning_rate": 1.8850156096789473e-05, + "loss": 1.0042, + "step": 1654 + }, + { + "epoch": 0.8984799131378935, + "grad_norm": 4.938883702947416, + "learning_rate": 1.8848518245014526e-05, + "loss": 0.7712, + "step": 1655 + }, + { + "epoch": 0.8990228013029316, + "grad_norm": 5.9261619850877, + "learning_rate": 1.8846879298840735e-05, + "loss": 1.4387, + "step": 1656 + }, + { + "epoch": 0.8995656894679696, + "grad_norm": 5.181216185736725, + "learning_rate": 1.8845239258470817e-05, + "loss": 0.9389, + "step": 1657 + }, + { + "epoch": 0.9001085776330076, + "grad_norm": 5.108249236166185, + "learning_rate": 1.8843598124107608e-05, + "loss": 1.0034, + "step": 1658 + }, + { + "epoch": 0.9006514657980456, + "grad_norm": 4.839799276438571, + "learning_rate": 1.8841955895954088e-05, + "loss": 0.8524, + "step": 1659 + }, + { + "epoch": 0.9011943539630836, + "grad_norm": 5.598018889964705, + "learning_rate": 1.8840312574213372e-05, + "loss": 1.0939, + "step": 1660 + }, + { + "epoch": 0.9017372421281216, + "grad_norm": 5.2496226470868725, + "learning_rate": 1.8838668159088707e-05, + "loss": 1.0812, + "step": 1661 + }, + { + "epoch": 0.9022801302931596, + "grad_norm": 5.2298567821168, + "learning_rate": 1.8837022650783477e-05, + "loss": 0.953, + "step": 1662 + }, + { + "epoch": 0.9028230184581976, + "grad_norm": 4.471282412199049, + "learning_rate": 1.88353760495012e-05, + "loss": 1.2652, + "step": 1663 + }, + { + "epoch": 0.9033659066232356, + "grad_norm": 5.892082884595572, + "learning_rate": 1.8833728355445534e-05, + "loss": 1.1165, + "step": 1664 + }, + { + "epoch": 0.9039087947882736, + "grad_norm": 5.49174731306936, + "learning_rate": 1.8832079568820268e-05, + "loss": 0.9262, + "step": 1665 + }, + { + "epoch": 0.9044516829533116, + "grad_norm": 5.55211658654164, + "learning_rate": 1.883042968982932e-05, + "loss": 1.3112, + "step": 1666 + }, + { + "epoch": 0.9049945711183496, + "grad_norm": 5.597392580982551, + "learning_rate": 1.8828778718676757e-05, + "loss": 0.8511, + "step": 1667 + }, + { + "epoch": 0.9055374592833876, + "grad_norm": 6.183127994440369, + "learning_rate": 1.8827126655566773e-05, + "loss": 1.2784, + "step": 1668 + }, + { + "epoch": 0.9060803474484256, + "grad_norm": 6.991796130593696, + "learning_rate": 1.882547350070369e-05, + "loss": 1.1703, + "step": 1669 + }, + { + "epoch": 0.9066232356134636, + "grad_norm": 5.7794859741652305, + "learning_rate": 1.8823819254291986e-05, + "loss": 1.0974, + "step": 1670 + }, + { + "epoch": 0.9071661237785016, + "grad_norm": 5.388338879165358, + "learning_rate": 1.8822163916536245e-05, + "loss": 1.2087, + "step": 1671 + }, + { + "epoch": 0.9077090119435396, + "grad_norm": 5.072420136637763, + "learning_rate": 1.8820507487641218e-05, + "loss": 0.9244, + "step": 1672 + }, + { + "epoch": 0.9082519001085776, + "grad_norm": 5.9571835605705115, + "learning_rate": 1.8818849967811762e-05, + "loss": 1.3299, + "step": 1673 + }, + { + "epoch": 0.9087947882736156, + "grad_norm": 5.706573186001289, + "learning_rate": 1.8817191357252892e-05, + "loss": 1.1059, + "step": 1674 + }, + { + "epoch": 0.9093376764386536, + "grad_norm": 5.4984127833712435, + "learning_rate": 1.8815531656169737e-05, + "loss": 1.1428, + "step": 1675 + }, + { + "epoch": 0.9098805646036916, + "grad_norm": 4.824306323799433, + "learning_rate": 1.8813870864767582e-05, + "loss": 0.7313, + "step": 1676 + }, + { + "epoch": 0.9104234527687296, + "grad_norm": 6.180183255961245, + "learning_rate": 1.8812208983251828e-05, + "loss": 1.0729, + "step": 1677 + }, + { + "epoch": 0.9109663409337676, + "grad_norm": 6.586598617269493, + "learning_rate": 1.8810546011828024e-05, + "loss": 1.5871, + "step": 1678 + }, + { + "epoch": 0.9115092290988056, + "grad_norm": 5.5611281281711875, + "learning_rate": 1.8808881950701845e-05, + "loss": 0.8133, + "step": 1679 + }, + { + "epoch": 0.9120521172638436, + "grad_norm": 5.214863992160379, + "learning_rate": 1.8807216800079108e-05, + "loss": 1.0437, + "step": 1680 + }, + { + "epoch": 0.9125950054288816, + "grad_norm": 6.631045200232836, + "learning_rate": 1.8805550560165763e-05, + "loss": 1.446, + "step": 1681 + }, + { + "epoch": 0.9131378935939196, + "grad_norm": 6.140822200691373, + "learning_rate": 1.8803883231167887e-05, + "loss": 1.8672, + "step": 1682 + }, + { + "epoch": 0.9136807817589576, + "grad_norm": 4.5606387908479995, + "learning_rate": 1.8802214813291708e-05, + "loss": 1.0662, + "step": 1683 + }, + { + "epoch": 0.9142236699239956, + "grad_norm": 4.166651706834997, + "learning_rate": 1.8800545306743567e-05, + "loss": 0.749, + "step": 1684 + }, + { + "epoch": 0.9147665580890336, + "grad_norm": 4.674238905131916, + "learning_rate": 1.8798874711729957e-05, + "loss": 1.1106, + "step": 1685 + }, + { + "epoch": 0.9153094462540716, + "grad_norm": 5.224600466137879, + "learning_rate": 1.8797203028457497e-05, + "loss": 0.9652, + "step": 1686 + }, + { + "epoch": 0.9158523344191096, + "grad_norm": 5.87706899393542, + "learning_rate": 1.8795530257132947e-05, + "loss": 1.2146, + "step": 1687 + }, + { + "epoch": 0.9163952225841476, + "grad_norm": 7.587853042260599, + "learning_rate": 1.87938563979632e-05, + "loss": 1.9051, + "step": 1688 + }, + { + "epoch": 0.9169381107491856, + "grad_norm": 4.988334377581133, + "learning_rate": 1.8792181451155275e-05, + "loss": 0.8502, + "step": 1689 + }, + { + "epoch": 0.9174809989142236, + "grad_norm": 5.454638378956608, + "learning_rate": 1.8790505416916338e-05, + "loss": 1.0493, + "step": 1690 + }, + { + "epoch": 0.9180238870792616, + "grad_norm": 5.95672873256444, + "learning_rate": 1.878882829545368e-05, + "loss": 1.1296, + "step": 1691 + }, + { + "epoch": 0.9185667752442996, + "grad_norm": 6.921816321230723, + "learning_rate": 1.8787150086974734e-05, + "loss": 1.6388, + "step": 1692 + }, + { + "epoch": 0.9191096634093376, + "grad_norm": 5.423998305707574, + "learning_rate": 1.878547079168706e-05, + "loss": 1.4865, + "step": 1693 + }, + { + "epoch": 0.9196525515743756, + "grad_norm": 4.479238264802495, + "learning_rate": 1.878379040979835e-05, + "loss": 1.0257, + "step": 1694 + }, + { + "epoch": 0.9201954397394136, + "grad_norm": 6.325821606507086, + "learning_rate": 1.8782108941516446e-05, + "loss": 1.1156, + "step": 1695 + }, + { + "epoch": 0.9207383279044516, + "grad_norm": 6.220712298120886, + "learning_rate": 1.8780426387049315e-05, + "loss": 1.375, + "step": 1696 + }, + { + "epoch": 0.9212812160694897, + "grad_norm": 5.813906943099404, + "learning_rate": 1.877874274660505e-05, + "loss": 1.1869, + "step": 1697 + }, + { + "epoch": 0.9218241042345277, + "grad_norm": 5.0160254446208965, + "learning_rate": 1.8777058020391893e-05, + "loss": 1.0443, + "step": 1698 + }, + { + "epoch": 0.9223669923995657, + "grad_norm": 4.273981514908416, + "learning_rate": 1.877537220861821e-05, + "loss": 1.0015, + "step": 1699 + }, + { + "epoch": 0.9229098805646037, + "grad_norm": 4.908796260576538, + "learning_rate": 1.8773685311492513e-05, + "loss": 1.0203, + "step": 1700 + }, + { + "epoch": 0.9234527687296417, + "grad_norm": 5.566794452728763, + "learning_rate": 1.8771997329223425e-05, + "loss": 1.4637, + "step": 1701 + }, + { + "epoch": 0.9239956568946797, + "grad_norm": 4.946565856604512, + "learning_rate": 1.8770308262019733e-05, + "loss": 0.8093, + "step": 1702 + }, + { + "epoch": 0.9245385450597177, + "grad_norm": 4.767566465605265, + "learning_rate": 1.8768618110090334e-05, + "loss": 0.5938, + "step": 1703 + }, + { + "epoch": 0.9250814332247557, + "grad_norm": 6.919371263621627, + "learning_rate": 1.8766926873644272e-05, + "loss": 0.9857, + "step": 1704 + }, + { + "epoch": 0.9256243213897937, + "grad_norm": 3.5297488420709944, + "learning_rate": 1.876523455289072e-05, + "loss": 0.8744, + "step": 1705 + }, + { + "epoch": 0.9261672095548317, + "grad_norm": 4.910386600622229, + "learning_rate": 1.8763541148038994e-05, + "loss": 0.9491, + "step": 1706 + }, + { + "epoch": 0.9267100977198697, + "grad_norm": 5.550892558363115, + "learning_rate": 1.876184665929853e-05, + "loss": 1.2793, + "step": 1707 + }, + { + "epoch": 0.9272529858849077, + "grad_norm": 4.470721855980947, + "learning_rate": 1.8760151086878905e-05, + "loss": 0.9706, + "step": 1708 + }, + { + "epoch": 0.9277958740499457, + "grad_norm": 7.1893201073258926, + "learning_rate": 1.8758454430989833e-05, + "loss": 1.2473, + "step": 1709 + }, + { + "epoch": 0.9283387622149837, + "grad_norm": 5.933304434057961, + "learning_rate": 1.875675669184116e-05, + "loss": 1.4917, + "step": 1710 + }, + { + "epoch": 0.9288816503800217, + "grad_norm": 5.386119916834561, + "learning_rate": 1.8755057869642857e-05, + "loss": 0.9417, + "step": 1711 + }, + { + "epoch": 0.9294245385450597, + "grad_norm": 6.121437468126703, + "learning_rate": 1.875335796460505e-05, + "loss": 1.203, + "step": 1712 + }, + { + "epoch": 0.9299674267100977, + "grad_norm": 5.34957970840911, + "learning_rate": 1.8751656976937974e-05, + "loss": 1.2557, + "step": 1713 + }, + { + "epoch": 0.9305103148751357, + "grad_norm": 7.404283054108375, + "learning_rate": 1.8749954906852023e-05, + "loss": 1.449, + "step": 1714 + }, + { + "epoch": 0.9310532030401737, + "grad_norm": 6.5511839319022585, + "learning_rate": 1.8748251754557696e-05, + "loss": 1.5548, + "step": 1715 + }, + { + "epoch": 0.9315960912052117, + "grad_norm": 5.0254557279034815, + "learning_rate": 1.8746547520265654e-05, + "loss": 0.9777, + "step": 1716 + }, + { + "epoch": 0.9321389793702497, + "grad_norm": 6.385252198006408, + "learning_rate": 1.874484220418667e-05, + "loss": 1.254, + "step": 1717 + }, + { + "epoch": 0.9326818675352877, + "grad_norm": 6.872281478717196, + "learning_rate": 1.874313580653167e-05, + "loss": 1.278, + "step": 1718 + }, + { + "epoch": 0.9332247557003257, + "grad_norm": 8.525842687889295, + "learning_rate": 1.8741428327511696e-05, + "loss": 1.3712, + "step": 1719 + }, + { + "epoch": 0.9337676438653637, + "grad_norm": 7.97420765579221, + "learning_rate": 1.8739719767337933e-05, + "loss": 1.7493, + "step": 1720 + }, + { + "epoch": 0.9343105320304017, + "grad_norm": 6.107109526814546, + "learning_rate": 1.8738010126221705e-05, + "loss": 0.9861, + "step": 1721 + }, + { + "epoch": 0.9348534201954397, + "grad_norm": 6.4666849584489645, + "learning_rate": 1.8736299404374453e-05, + "loss": 0.8041, + "step": 1722 + }, + { + "epoch": 0.9353963083604777, + "grad_norm": 7.489590627219774, + "learning_rate": 1.873458760200777e-05, + "loss": 1.2325, + "step": 1723 + }, + { + "epoch": 0.9359391965255157, + "grad_norm": 7.020182643385788, + "learning_rate": 1.8732874719333373e-05, + "loss": 1.1698, + "step": 1724 + }, + { + "epoch": 0.9364820846905537, + "grad_norm": 6.436449840675483, + "learning_rate": 1.873116075656311e-05, + "loss": 1.084, + "step": 1725 + }, + { + "epoch": 0.9370249728555917, + "grad_norm": 7.27100690341834, + "learning_rate": 1.872944571390897e-05, + "loss": 0.9793, + "step": 1726 + }, + { + "epoch": 0.9375678610206297, + "grad_norm": 6.337617737677888, + "learning_rate": 1.872772959158307e-05, + "loss": 1.0558, + "step": 1727 + }, + { + "epoch": 0.9381107491856677, + "grad_norm": 5.356493873446841, + "learning_rate": 1.8726012389797667e-05, + "loss": 1.1031, + "step": 1728 + }, + { + "epoch": 0.9386536373507057, + "grad_norm": 6.302965546816682, + "learning_rate": 1.8724294108765142e-05, + "loss": 1.1401, + "step": 1729 + }, + { + "epoch": 0.9391965255157437, + "grad_norm": 6.116650787967334, + "learning_rate": 1.872257474869802e-05, + "loss": 1.0848, + "step": 1730 + }, + { + "epoch": 0.9397394136807817, + "grad_norm": 4.3874746967683, + "learning_rate": 1.8720854309808948e-05, + "loss": 0.7463, + "step": 1731 + }, + { + "epoch": 0.9402823018458197, + "grad_norm": 5.00673534942558, + "learning_rate": 1.871913279231072e-05, + "loss": 1.2365, + "step": 1732 + }, + { + "epoch": 0.9408251900108577, + "grad_norm": 5.222845312743783, + "learning_rate": 1.871741019641625e-05, + "loss": 1.062, + "step": 1733 + }, + { + "epoch": 0.9413680781758957, + "grad_norm": 4.624403224726164, + "learning_rate": 1.871568652233859e-05, + "loss": 1.0155, + "step": 1734 + }, + { + "epoch": 0.9419109663409337, + "grad_norm": 5.386412889478532, + "learning_rate": 1.8713961770290936e-05, + "loss": 1.4316, + "step": 1735 + }, + { + "epoch": 0.9424538545059717, + "grad_norm": 6.0899860043030705, + "learning_rate": 1.87122359404866e-05, + "loss": 1.1372, + "step": 1736 + }, + { + "epoch": 0.9429967426710097, + "grad_norm": 5.306362358042622, + "learning_rate": 1.8710509033139037e-05, + "loss": 0.9353, + "step": 1737 + }, + { + "epoch": 0.9435396308360477, + "grad_norm": 6.479420699028343, + "learning_rate": 1.8708781048461832e-05, + "loss": 1.2435, + "step": 1738 + }, + { + "epoch": 0.9440825190010858, + "grad_norm": 6.205746500302448, + "learning_rate": 1.8707051986668712e-05, + "loss": 0.7872, + "step": 1739 + }, + { + "epoch": 0.9446254071661238, + "grad_norm": 4.157411034856641, + "learning_rate": 1.8705321847973523e-05, + "loss": 0.8004, + "step": 1740 + }, + { + "epoch": 0.9451682953311618, + "grad_norm": 5.5814599600578285, + "learning_rate": 1.8703590632590254e-05, + "loss": 0.7384, + "step": 1741 + }, + { + "epoch": 0.9457111834961998, + "grad_norm": 6.040759722497843, + "learning_rate": 1.8701858340733023e-05, + "loss": 1.5941, + "step": 1742 + }, + { + "epoch": 0.9462540716612378, + "grad_norm": 4.592040464655725, + "learning_rate": 1.8700124972616085e-05, + "loss": 1.0662, + "step": 1743 + }, + { + "epoch": 0.9467969598262758, + "grad_norm": 6.077319934267432, + "learning_rate": 1.8698390528453823e-05, + "loss": 1.3726, + "step": 1744 + }, + { + "epoch": 0.9473398479913138, + "grad_norm": 4.255180894541495, + "learning_rate": 1.869665500846076e-05, + "loss": 0.7001, + "step": 1745 + }, + { + "epoch": 0.9478827361563518, + "grad_norm": 6.390593673413393, + "learning_rate": 1.869491841285154e-05, + "loss": 1.2268, + "step": 1746 + }, + { + "epoch": 0.9484256243213898, + "grad_norm": 5.347324129941674, + "learning_rate": 1.8693180741840957e-05, + "loss": 1.2193, + "step": 1747 + }, + { + "epoch": 0.9489685124864278, + "grad_norm": 6.407261723862006, + "learning_rate": 1.8691441995643927e-05, + "loss": 0.868, + "step": 1748 + }, + { + "epoch": 0.9495114006514658, + "grad_norm": 4.400520878629861, + "learning_rate": 1.8689702174475496e-05, + "loss": 0.4679, + "step": 1749 + }, + { + "epoch": 0.9500542888165038, + "grad_norm": 5.6161548427455505, + "learning_rate": 1.8687961278550852e-05, + "loss": 0.8842, + "step": 1750 + }, + { + "epoch": 0.9505971769815418, + "grad_norm": 4.7163538240048, + "learning_rate": 1.8686219308085306e-05, + "loss": 0.9972, + "step": 1751 + }, + { + "epoch": 0.9511400651465798, + "grad_norm": 4.845854095585142, + "learning_rate": 1.8684476263294318e-05, + "loss": 1.1046, + "step": 1752 + }, + { + "epoch": 0.9516829533116178, + "grad_norm": 3.878984502983999, + "learning_rate": 1.8682732144393463e-05, + "loss": 0.7873, + "step": 1753 + }, + { + "epoch": 0.9522258414766558, + "grad_norm": 5.736472248757515, + "learning_rate": 1.8680986951598458e-05, + "loss": 1.2046, + "step": 1754 + }, + { + "epoch": 0.9527687296416938, + "grad_norm": 5.12988223940441, + "learning_rate": 1.867924068512515e-05, + "loss": 0.6293, + "step": 1755 + }, + { + "epoch": 0.9533116178067318, + "grad_norm": 7.822467968073818, + "learning_rate": 1.867749334518952e-05, + "loss": 1.3625, + "step": 1756 + }, + { + "epoch": 0.9538545059717698, + "grad_norm": 5.488388947029871, + "learning_rate": 1.8675744932007687e-05, + "loss": 1.078, + "step": 1757 + }, + { + "epoch": 0.9543973941368078, + "grad_norm": 4.631848438676548, + "learning_rate": 1.8673995445795894e-05, + "loss": 0.69, + "step": 1758 + }, + { + "epoch": 0.9549402823018458, + "grad_norm": 7.078171019993124, + "learning_rate": 1.8672244886770516e-05, + "loss": 1.1036, + "step": 1759 + }, + { + "epoch": 0.9554831704668838, + "grad_norm": 5.9259972832349215, + "learning_rate": 1.8670493255148073e-05, + "loss": 0.7919, + "step": 1760 + }, + { + "epoch": 0.9560260586319218, + "grad_norm": 5.846343833877514, + "learning_rate": 1.8668740551145205e-05, + "loss": 1.1653, + "step": 1761 + }, + { + "epoch": 0.9565689467969598, + "grad_norm": 5.559082125902064, + "learning_rate": 1.8666986774978685e-05, + "loss": 1.4214, + "step": 1762 + }, + { + "epoch": 0.9571118349619978, + "grad_norm": 6.511207981127819, + "learning_rate": 1.8665231926865433e-05, + "loss": 1.3552, + "step": 1763 + }, + { + "epoch": 0.9576547231270358, + "grad_norm": 5.728999474717542, + "learning_rate": 1.8663476007022482e-05, + "loss": 0.7861, + "step": 1764 + }, + { + "epoch": 0.9581976112920738, + "grad_norm": 6.152118564149309, + "learning_rate": 1.8661719015667016e-05, + "loss": 0.8679, + "step": 1765 + }, + { + "epoch": 0.9587404994571118, + "grad_norm": 6.77121947972117, + "learning_rate": 1.8659960953016334e-05, + "loss": 0.8103, + "step": 1766 + }, + { + "epoch": 0.9592833876221498, + "grad_norm": 7.822575173268641, + "learning_rate": 1.865820181928788e-05, + "loss": 1.2558, + "step": 1767 + }, + { + "epoch": 0.9598262757871878, + "grad_norm": 6.389073592671253, + "learning_rate": 1.8656441614699225e-05, + "loss": 1.0314, + "step": 1768 + }, + { + "epoch": 0.9603691639522258, + "grad_norm": 6.141689972496849, + "learning_rate": 1.8654680339468076e-05, + "loss": 1.2452, + "step": 1769 + }, + { + "epoch": 0.9609120521172638, + "grad_norm": 6.934092850250001, + "learning_rate": 1.8652917993812267e-05, + "loss": 1.1186, + "step": 1770 + }, + { + "epoch": 0.9614549402823018, + "grad_norm": 6.129320663785202, + "learning_rate": 1.865115457794977e-05, + "loss": 0.9076, + "step": 1771 + }, + { + "epoch": 0.9619978284473398, + "grad_norm": 7.008261896063745, + "learning_rate": 1.8649390092098693e-05, + "loss": 1.7152, + "step": 1772 + }, + { + "epoch": 0.9625407166123778, + "grad_norm": 6.36833090372477, + "learning_rate": 1.8647624536477255e-05, + "loss": 0.8087, + "step": 1773 + }, + { + "epoch": 0.9630836047774158, + "grad_norm": 4.7595903392838865, + "learning_rate": 1.8645857911303838e-05, + "loss": 0.6374, + "step": 1774 + }, + { + "epoch": 0.9636264929424538, + "grad_norm": 6.410325139984648, + "learning_rate": 1.8644090216796934e-05, + "loss": 1.3611, + "step": 1775 + }, + { + "epoch": 0.9641693811074918, + "grad_norm": 5.900212705071564, + "learning_rate": 1.8642321453175177e-05, + "loss": 0.891, + "step": 1776 + }, + { + "epoch": 0.9647122692725298, + "grad_norm": 6.90562183471938, + "learning_rate": 1.8640551620657326e-05, + "loss": 0.7899, + "step": 1777 + }, + { + "epoch": 0.9652551574375678, + "grad_norm": 6.476066490933977, + "learning_rate": 1.8638780719462278e-05, + "loss": 0.9614, + "step": 1778 + }, + { + "epoch": 0.9657980456026058, + "grad_norm": 6.659586852509603, + "learning_rate": 1.8637008749809065e-05, + "loss": 1.2419, + "step": 1779 + }, + { + "epoch": 0.9663409337676439, + "grad_norm": 6.587754652016677, + "learning_rate": 1.8635235711916847e-05, + "loss": 1.3664, + "step": 1780 + }, + { + "epoch": 0.9668838219326819, + "grad_norm": 6.558578755813245, + "learning_rate": 1.863346160600491e-05, + "loss": 1.1249, + "step": 1781 + }, + { + "epoch": 0.9674267100977199, + "grad_norm": 4.791301716076818, + "learning_rate": 1.8631686432292685e-05, + "loss": 0.6836, + "step": 1782 + }, + { + "epoch": 0.9679695982627579, + "grad_norm": 5.120778292537024, + "learning_rate": 1.862991019099972e-05, + "loss": 0.592, + "step": 1783 + }, + { + "epoch": 0.9685124864277959, + "grad_norm": 8.250755784410368, + "learning_rate": 1.8628132882345713e-05, + "loss": 1.2579, + "step": 1784 + }, + { + "epoch": 0.9690553745928339, + "grad_norm": 5.811263926579046, + "learning_rate": 1.862635450655048e-05, + "loss": 1.2935, + "step": 1785 + }, + { + "epoch": 0.9695982627578719, + "grad_norm": 6.042040633945999, + "learning_rate": 1.862457506383397e-05, + "loss": 0.8784, + "step": 1786 + }, + { + "epoch": 0.9701411509229099, + "grad_norm": 7.115011060082062, + "learning_rate": 1.8622794554416272e-05, + "loss": 1.0717, + "step": 1787 + }, + { + "epoch": 0.9706840390879479, + "grad_norm": 7.520775839533288, + "learning_rate": 1.8621012978517604e-05, + "loss": 1.3468, + "step": 1788 + }, + { + "epoch": 0.9712269272529859, + "grad_norm": 6.949921961354188, + "learning_rate": 1.8619230336358306e-05, + "loss": 1.5609, + "step": 1789 + }, + { + "epoch": 0.9717698154180239, + "grad_norm": 6.396218455762339, + "learning_rate": 1.8617446628158866e-05, + "loss": 1.1807, + "step": 1790 + }, + { + "epoch": 0.9723127035830619, + "grad_norm": 6.314794279037124, + "learning_rate": 1.861566185413989e-05, + "loss": 0.9255, + "step": 1791 + }, + { + "epoch": 0.9728555917480999, + "grad_norm": 5.393752442926871, + "learning_rate": 1.8613876014522128e-05, + "loss": 0.7926, + "step": 1792 + }, + { + "epoch": 0.9733984799131379, + "grad_norm": 5.908313693688933, + "learning_rate": 1.8612089109526453e-05, + "loss": 0.7984, + "step": 1793 + }, + { + "epoch": 0.9739413680781759, + "grad_norm": 4.6952509401002125, + "learning_rate": 1.8610301139373867e-05, + "loss": 1.0344, + "step": 1794 + }, + { + "epoch": 0.9744842562432139, + "grad_norm": 6.81835335105175, + "learning_rate": 1.8608512104285517e-05, + "loss": 1.1532, + "step": 1795 + }, + { + "epoch": 0.9750271444082519, + "grad_norm": 4.159411114912345, + "learning_rate": 1.860672200448267e-05, + "loss": 1.0665, + "step": 1796 + }, + { + "epoch": 0.9755700325732899, + "grad_norm": 4.911404446825004, + "learning_rate": 1.8604930840186726e-05, + "loss": 1.1784, + "step": 1797 + }, + { + "epoch": 0.9761129207383279, + "grad_norm": 5.399453855489055, + "learning_rate": 1.860313861161922e-05, + "loss": 0.8215, + "step": 1798 + }, + { + "epoch": 0.9766558089033659, + "grad_norm": 4.797137239148531, + "learning_rate": 1.860134531900182e-05, + "loss": 0.6405, + "step": 1799 + }, + { + "epoch": 0.9771986970684039, + "grad_norm": 3.937076316066272, + "learning_rate": 1.859955096255633e-05, + "loss": 0.5912, + "step": 1800 + }, + { + "epoch": 0.9777415852334419, + "grad_norm": 5.569110777302129, + "learning_rate": 1.859775554250466e-05, + "loss": 0.8089, + "step": 1801 + }, + { + "epoch": 0.9782844733984799, + "grad_norm": 6.364258901894282, + "learning_rate": 1.859595905906889e-05, + "loss": 1.5181, + "step": 1802 + }, + { + "epoch": 0.9788273615635179, + "grad_norm": 4.995224609073908, + "learning_rate": 1.85941615124712e-05, + "loss": 0.8573, + "step": 1803 + }, + { + "epoch": 0.9793702497285559, + "grad_norm": 5.041992740138782, + "learning_rate": 1.8592362902933918e-05, + "loss": 0.8982, + "step": 1804 + }, + { + "epoch": 0.9799131378935939, + "grad_norm": 5.6054574490953275, + "learning_rate": 1.8590563230679496e-05, + "loss": 0.9285, + "step": 1805 + }, + { + "epoch": 0.9804560260586319, + "grad_norm": 5.269503319178264, + "learning_rate": 1.8588762495930526e-05, + "loss": 1.0963, + "step": 1806 + }, + { + "epoch": 0.9809989142236699, + "grad_norm": 6.184787731593613, + "learning_rate": 1.8586960698909718e-05, + "loss": 1.4005, + "step": 1807 + }, + { + "epoch": 0.9815418023887079, + "grad_norm": 8.50824591479003, + "learning_rate": 1.858515783983993e-05, + "loss": 1.9607, + "step": 1808 + }, + { + "epoch": 0.9820846905537459, + "grad_norm": 5.399464732589782, + "learning_rate": 1.8583353918944134e-05, + "loss": 0.951, + "step": 1809 + }, + { + "epoch": 0.9826275787187839, + "grad_norm": 6.577634699737779, + "learning_rate": 1.8581548936445447e-05, + "loss": 1.5139, + "step": 1810 + }, + { + "epoch": 0.9831704668838219, + "grad_norm": 10.183489163585788, + "learning_rate": 1.8579742892567107e-05, + "loss": 2.0539, + "step": 1811 + }, + { + "epoch": 0.9837133550488599, + "grad_norm": 6.578605968943458, + "learning_rate": 1.8577935787532494e-05, + "loss": 1.1454, + "step": 1812 + }, + { + "epoch": 0.9842562432138979, + "grad_norm": 4.923285404215955, + "learning_rate": 1.8576127621565113e-05, + "loss": 0.8965, + "step": 1813 + }, + { + "epoch": 0.9847991313789359, + "grad_norm": 5.912445932400008, + "learning_rate": 1.85743183948886e-05, + "loss": 1.1311, + "step": 1814 + }, + { + "epoch": 0.9853420195439739, + "grad_norm": 4.5183231221646105, + "learning_rate": 1.8572508107726725e-05, + "loss": 0.7511, + "step": 1815 + }, + { + "epoch": 0.9858849077090119, + "grad_norm": 5.2218543332097935, + "learning_rate": 1.8570696760303378e-05, + "loss": 1.0063, + "step": 1816 + }, + { + "epoch": 0.9864277958740499, + "grad_norm": 7.215748803707923, + "learning_rate": 1.85688843528426e-05, + "loss": 1.3674, + "step": 1817 + }, + { + "epoch": 0.9869706840390879, + "grad_norm": 6.362354074709609, + "learning_rate": 1.8567070885568547e-05, + "loss": 1.1319, + "step": 1818 + }, + { + "epoch": 0.987513572204126, + "grad_norm": 4.680755706450828, + "learning_rate": 1.8565256358705513e-05, + "loss": 0.6613, + "step": 1819 + }, + { + "epoch": 0.988056460369164, + "grad_norm": 4.80241945107321, + "learning_rate": 1.8563440772477922e-05, + "loss": 0.8619, + "step": 1820 + }, + { + "epoch": 0.988599348534202, + "grad_norm": 5.858709563059296, + "learning_rate": 1.856162412711033e-05, + "loss": 0.9432, + "step": 1821 + }, + { + "epoch": 0.98914223669924, + "grad_norm": 6.478119299588854, + "learning_rate": 1.855980642282742e-05, + "loss": 1.1866, + "step": 1822 + }, + { + "epoch": 0.989685124864278, + "grad_norm": 7.081924554034699, + "learning_rate": 1.8557987659854006e-05, + "loss": 1.3657, + "step": 1823 + }, + { + "epoch": 0.990228013029316, + "grad_norm": 5.404410220733361, + "learning_rate": 1.855616783841504e-05, + "loss": 1.1051, + "step": 1824 + }, + { + "epoch": 0.990770901194354, + "grad_norm": 6.742207962837618, + "learning_rate": 1.8554346958735602e-05, + "loss": 1.4667, + "step": 1825 + }, + { + "epoch": 0.991313789359392, + "grad_norm": 6.164178667211635, + "learning_rate": 1.8552525021040895e-05, + "loss": 0.7861, + "step": 1826 + }, + { + "epoch": 0.99185667752443, + "grad_norm": 6.078017237490637, + "learning_rate": 1.8550702025556265e-05, + "loss": 1.3294, + "step": 1827 + }, + { + "epoch": 0.992399565689468, + "grad_norm": 4.7372949568909455, + "learning_rate": 1.8548877972507182e-05, + "loss": 0.9779, + "step": 1828 + }, + { + "epoch": 0.992942453854506, + "grad_norm": 5.25971277777658, + "learning_rate": 1.8547052862119247e-05, + "loss": 1.049, + "step": 1829 + }, + { + "epoch": 0.993485342019544, + "grad_norm": 4.950709970125391, + "learning_rate": 1.854522669461819e-05, + "loss": 0.6665, + "step": 1830 + }, + { + "epoch": 0.994028230184582, + "grad_norm": 5.856428194694733, + "learning_rate": 1.8543399470229876e-05, + "loss": 1.0954, + "step": 1831 + }, + { + "epoch": 0.99457111834962, + "grad_norm": 4.895286613591861, + "learning_rate": 1.85415711891803e-05, + "loss": 0.7069, + "step": 1832 + }, + { + "epoch": 0.995114006514658, + "grad_norm": 5.249228502887773, + "learning_rate": 1.8539741851695586e-05, + "loss": 1.0655, + "step": 1833 + }, + { + "epoch": 0.995656894679696, + "grad_norm": 4.855694799243912, + "learning_rate": 1.8537911458001988e-05, + "loss": 1.1872, + "step": 1834 + }, + { + "epoch": 0.996199782844734, + "grad_norm": 5.180821237913711, + "learning_rate": 1.8536080008325896e-05, + "loss": 0.8281, + "step": 1835 + }, + { + "epoch": 0.996742671009772, + "grad_norm": 5.073254295771654, + "learning_rate": 1.8534247502893823e-05, + "loss": 0.7002, + "step": 1836 + }, + { + "epoch": 0.99728555917481, + "grad_norm": 5.4022604097623255, + "learning_rate": 1.8532413941932416e-05, + "loss": 1.1793, + "step": 1837 + }, + { + "epoch": 0.997828447339848, + "grad_norm": 4.981761324197401, + "learning_rate": 1.8530579325668455e-05, + "loss": 0.8331, + "step": 1838 + }, + { + "epoch": 0.998371335504886, + "grad_norm": 7.325363733761658, + "learning_rate": 1.852874365432885e-05, + "loss": 1.3071, + "step": 1839 + }, + { + "epoch": 0.998914223669924, + "grad_norm": 7.5668251202705505, + "learning_rate": 1.852690692814063e-05, + "loss": 1.4955, + "step": 1840 + }, + { + "epoch": 0.999457111834962, + "grad_norm": 6.706936559053067, + "learning_rate": 1.8525069147330978e-05, + "loss": 1.0286, + "step": 1841 + }, + { + "epoch": 1.0, + "grad_norm": 6.338369283394405, + "learning_rate": 1.8523230312127183e-05, + "loss": 1.0509, + "step": 1842 + }, + { + "epoch": 1.000542888165038, + "grad_norm": 6.119502670416093, + "learning_rate": 1.8521390422756683e-05, + "loss": 1.0805, + "step": 1843 + }, + { + "epoch": 1.001085776330076, + "grad_norm": 5.074420967885639, + "learning_rate": 1.851954947944703e-05, + "loss": 1.0931, + "step": 1844 + }, + { + "epoch": 1.001628664495114, + "grad_norm": 8.194067497007946, + "learning_rate": 1.851770748242592e-05, + "loss": 2.0732, + "step": 1845 + }, + { + "epoch": 1.002171552660152, + "grad_norm": 4.389522929363591, + "learning_rate": 1.8515864431921177e-05, + "loss": 0.8684, + "step": 1846 + }, + { + "epoch": 1.00271444082519, + "grad_norm": 5.720373954211297, + "learning_rate": 1.8514020328160748e-05, + "loss": 1.0223, + "step": 1847 + }, + { + "epoch": 1.003257328990228, + "grad_norm": 5.163678744409467, + "learning_rate": 1.8512175171372713e-05, + "loss": 0.9611, + "step": 1848 + }, + { + "epoch": 1.003800217155266, + "grad_norm": 5.786056891749039, + "learning_rate": 1.8510328961785286e-05, + "loss": 1.3473, + "step": 1849 + }, + { + "epoch": 1.004343105320304, + "grad_norm": 5.9194250471944985, + "learning_rate": 1.850848169962681e-05, + "loss": 0.8218, + "step": 1850 + }, + { + "epoch": 1.004885993485342, + "grad_norm": 5.890241796399565, + "learning_rate": 1.850663338512576e-05, + "loss": 0.842, + "step": 1851 + }, + { + "epoch": 1.00542888165038, + "grad_norm": 6.939341519595904, + "learning_rate": 1.8504784018510732e-05, + "loss": 0.9263, + "step": 1852 + }, + { + "epoch": 1.005971769815418, + "grad_norm": 4.592014594714093, + "learning_rate": 1.850293360001046e-05, + "loss": 1.015, + "step": 1853 + }, + { + "epoch": 1.006514657980456, + "grad_norm": 4.575145990552055, + "learning_rate": 1.8501082129853816e-05, + "loss": 1.0505, + "step": 1854 + }, + { + "epoch": 1.007057546145494, + "grad_norm": 5.999218616951103, + "learning_rate": 1.849922960826978e-05, + "loss": 1.0945, + "step": 1855 + }, + { + "epoch": 1.007600434310532, + "grad_norm": 5.459227552994586, + "learning_rate": 1.8497376035487483e-05, + "loss": 1.0167, + "step": 1856 + }, + { + "epoch": 1.00814332247557, + "grad_norm": 4.785837605715454, + "learning_rate": 1.8495521411736173e-05, + "loss": 1.0559, + "step": 1857 + }, + { + "epoch": 1.008686210640608, + "grad_norm": 5.813312187108195, + "learning_rate": 1.8493665737245236e-05, + "loss": 1.3477, + "step": 1858 + }, + { + "epoch": 1.009229098805646, + "grad_norm": 6.512945450475729, + "learning_rate": 1.8491809012244182e-05, + "loss": 1.4416, + "step": 1859 + }, + { + "epoch": 1.009771986970684, + "grad_norm": 5.599962385405486, + "learning_rate": 1.8489951236962658e-05, + "loss": 0.8961, + "step": 1860 + }, + { + "epoch": 1.010314875135722, + "grad_norm": 4.863989016796961, + "learning_rate": 1.848809241163043e-05, + "loss": 0.639, + "step": 1861 + }, + { + "epoch": 1.01085776330076, + "grad_norm": 5.937471136238718, + "learning_rate": 1.848623253647741e-05, + "loss": 1.0548, + "step": 1862 + }, + { + "epoch": 1.011400651465798, + "grad_norm": 4.75172125273428, + "learning_rate": 1.8484371611733625e-05, + "loss": 0.6493, + "step": 1863 + }, + { + "epoch": 1.011943539630836, + "grad_norm": 5.867263430143688, + "learning_rate": 1.848250963762923e-05, + "loss": 0.8569, + "step": 1864 + }, + { + "epoch": 1.012486427795874, + "grad_norm": 5.932410925593062, + "learning_rate": 1.848064661439453e-05, + "loss": 0.8351, + "step": 1865 + }, + { + "epoch": 1.013029315960912, + "grad_norm": 5.8320858990674, + "learning_rate": 1.847878254225994e-05, + "loss": 0.8916, + "step": 1866 + }, + { + "epoch": 1.01357220412595, + "grad_norm": 5.836721791741835, + "learning_rate": 1.847691742145601e-05, + "loss": 1.1742, + "step": 1867 + }, + { + "epoch": 1.014115092290988, + "grad_norm": 6.806668722890977, + "learning_rate": 1.8475051252213423e-05, + "loss": 1.0714, + "step": 1868 + }, + { + "epoch": 1.014657980456026, + "grad_norm": 7.026347641082525, + "learning_rate": 1.8473184034762992e-05, + "loss": 1.4101, + "step": 1869 + }, + { + "epoch": 1.015200868621064, + "grad_norm": 5.9532414283206005, + "learning_rate": 1.8471315769335657e-05, + "loss": 0.7855, + "step": 1870 + }, + { + "epoch": 1.015743756786102, + "grad_norm": 5.7578706206451855, + "learning_rate": 1.846944645616248e-05, + "loss": 0.7644, + "step": 1871 + }, + { + "epoch": 1.01628664495114, + "grad_norm": 6.340541706204924, + "learning_rate": 1.846757609547467e-05, + "loss": 1.1607, + "step": 1872 + }, + { + "epoch": 1.016829533116178, + "grad_norm": 4.942991579941027, + "learning_rate": 1.8465704687503558e-05, + "loss": 0.7578, + "step": 1873 + }, + { + "epoch": 1.017372421281216, + "grad_norm": 6.42121491667965, + "learning_rate": 1.846383223248059e-05, + "loss": 0.9924, + "step": 1874 + }, + { + "epoch": 1.017915309446254, + "grad_norm": 6.252498990648505, + "learning_rate": 1.8461958730637368e-05, + "loss": 0.811, + "step": 1875 + }, + { + "epoch": 1.018458197611292, + "grad_norm": 5.941336187529878, + "learning_rate": 1.84600841822056e-05, + "loss": 0.772, + "step": 1876 + }, + { + "epoch": 1.01900108577633, + "grad_norm": 5.323895762793836, + "learning_rate": 1.845820858741714e-05, + "loss": 0.8812, + "step": 1877 + }, + { + "epoch": 1.019543973941368, + "grad_norm": 7.2574721612646105, + "learning_rate": 1.845633194650396e-05, + "loss": 1.2759, + "step": 1878 + }, + { + "epoch": 1.020086862106406, + "grad_norm": 4.307176114235153, + "learning_rate": 1.8454454259698165e-05, + "loss": 0.569, + "step": 1879 + }, + { + "epoch": 1.020629750271444, + "grad_norm": 6.504543473682945, + "learning_rate": 1.8452575527231997e-05, + "loss": 0.9355, + "step": 1880 + }, + { + "epoch": 1.021172638436482, + "grad_norm": 5.822065969327671, + "learning_rate": 1.8450695749337816e-05, + "loss": 1.0287, + "step": 1881 + }, + { + "epoch": 1.02171552660152, + "grad_norm": 8.90668319984316, + "learning_rate": 1.8448814926248112e-05, + "loss": 1.5348, + "step": 1882 + }, + { + "epoch": 1.022258414766558, + "grad_norm": 5.233217719773852, + "learning_rate": 1.844693305819552e-05, + "loss": 0.564, + "step": 1883 + }, + { + "epoch": 1.022801302931596, + "grad_norm": 5.806961114054563, + "learning_rate": 1.844505014541278e-05, + "loss": 0.9349, + "step": 1884 + }, + { + "epoch": 1.023344191096634, + "grad_norm": 6.239514346807307, + "learning_rate": 1.8443166188132777e-05, + "loss": 0.9677, + "step": 1885 + }, + { + "epoch": 1.023887079261672, + "grad_norm": 5.596402826374599, + "learning_rate": 1.8441281186588528e-05, + "loss": 0.695, + "step": 1886 + }, + { + "epoch": 1.02442996742671, + "grad_norm": 6.531429104200087, + "learning_rate": 1.8439395141013165e-05, + "loss": 1.1543, + "step": 1887 + }, + { + "epoch": 1.024972855591748, + "grad_norm": 5.129461327017003, + "learning_rate": 1.843750805163996e-05, + "loss": 1.2624, + "step": 1888 + }, + { + "epoch": 1.0255157437567861, + "grad_norm": 6.862993011679228, + "learning_rate": 1.8435619918702318e-05, + "loss": 1.2673, + "step": 1889 + }, + { + "epoch": 1.0260586319218241, + "grad_norm": 5.134671966877655, + "learning_rate": 1.8433730742433755e-05, + "loss": 0.6032, + "step": 1890 + }, + { + "epoch": 1.0266015200868621, + "grad_norm": 6.9199078267443195, + "learning_rate": 1.8431840523067932e-05, + "loss": 1.1255, + "step": 1891 + }, + { + "epoch": 1.0271444082519001, + "grad_norm": 5.808796921918892, + "learning_rate": 1.8429949260838635e-05, + "loss": 0.8488, + "step": 1892 + }, + { + "epoch": 1.0276872964169381, + "grad_norm": 7.102540931353532, + "learning_rate": 1.842805695597978e-05, + "loss": 1.0634, + "step": 1893 + }, + { + "epoch": 1.0282301845819761, + "grad_norm": 9.932212718124791, + "learning_rate": 1.8426163608725403e-05, + "loss": 1.3909, + "step": 1894 + }, + { + "epoch": 1.0287730727470141, + "grad_norm": 6.999752698080143, + "learning_rate": 1.8424269219309686e-05, + "loss": 1.1106, + "step": 1895 + }, + { + "epoch": 1.0293159609120521, + "grad_norm": 6.569873028691542, + "learning_rate": 1.842237378796693e-05, + "loss": 0.9455, + "step": 1896 + }, + { + "epoch": 1.0298588490770901, + "grad_norm": 6.418746090533204, + "learning_rate": 1.8420477314931554e-05, + "loss": 0.934, + "step": 1897 + }, + { + "epoch": 1.0304017372421281, + "grad_norm": 7.7235490740904815, + "learning_rate": 1.8418579800438125e-05, + "loss": 1.1582, + "step": 1898 + }, + { + "epoch": 1.0309446254071661, + "grad_norm": 5.4350436368749895, + "learning_rate": 1.841668124472133e-05, + "loss": 0.7726, + "step": 1899 + }, + { + "epoch": 1.0314875135722041, + "grad_norm": 5.34390520812136, + "learning_rate": 1.8414781648015983e-05, + "loss": 0.7497, + "step": 1900 + }, + { + "epoch": 1.0320304017372421, + "grad_norm": 6.563688266956639, + "learning_rate": 1.841288101055703e-05, + "loss": 1.3352, + "step": 1901 + }, + { + "epoch": 1.0325732899022801, + "grad_norm": 7.710283890330315, + "learning_rate": 1.841097933257955e-05, + "loss": 0.8804, + "step": 1902 + }, + { + "epoch": 1.0331161780673181, + "grad_norm": 5.31500431944041, + "learning_rate": 1.840907661431874e-05, + "loss": 0.8667, + "step": 1903 + }, + { + "epoch": 1.0336590662323561, + "grad_norm": 7.476982099026401, + "learning_rate": 1.840717285600993e-05, + "loss": 0.8317, + "step": 1904 + }, + { + "epoch": 1.0342019543973942, + "grad_norm": 5.1563433597221415, + "learning_rate": 1.840526805788858e-05, + "loss": 1.0108, + "step": 1905 + }, + { + "epoch": 1.0347448425624322, + "grad_norm": 6.411832164370409, + "learning_rate": 1.8403362220190284e-05, + "loss": 0.7778, + "step": 1906 + }, + { + "epoch": 1.0352877307274702, + "grad_norm": 6.688312152010462, + "learning_rate": 1.8401455343150757e-05, + "loss": 1.2438, + "step": 1907 + }, + { + "epoch": 1.0358306188925082, + "grad_norm": 6.412659783246144, + "learning_rate": 1.839954742700584e-05, + "loss": 0.8191, + "step": 1908 + }, + { + "epoch": 1.0363735070575462, + "grad_norm": 6.399618830323376, + "learning_rate": 1.839763847199151e-05, + "loss": 1.4485, + "step": 1909 + }, + { + "epoch": 1.0369163952225842, + "grad_norm": 8.041718562623554, + "learning_rate": 1.8395728478343873e-05, + "loss": 1.0043, + "step": 1910 + }, + { + "epoch": 1.0374592833876222, + "grad_norm": 6.4118854777282905, + "learning_rate": 1.8393817446299152e-05, + "loss": 0.9183, + "step": 1911 + }, + { + "epoch": 1.0380021715526602, + "grad_norm": 6.400834366436549, + "learning_rate": 1.8391905376093717e-05, + "loss": 1.1517, + "step": 1912 + }, + { + "epoch": 1.0385450597176982, + "grad_norm": 6.160364431172616, + "learning_rate": 1.8389992267964046e-05, + "loss": 0.5949, + "step": 1913 + }, + { + "epoch": 1.0390879478827362, + "grad_norm": 7.298564934931577, + "learning_rate": 1.8388078122146763e-05, + "loss": 1.0234, + "step": 1914 + }, + { + "epoch": 1.0396308360477742, + "grad_norm": 6.132796013684569, + "learning_rate": 1.838616293887861e-05, + "loss": 0.9417, + "step": 1915 + }, + { + "epoch": 1.0401737242128122, + "grad_norm": 8.083369310536504, + "learning_rate": 1.8384246718396458e-05, + "loss": 1.2617, + "step": 1916 + }, + { + "epoch": 1.0407166123778502, + "grad_norm": 5.492413210001038, + "learning_rate": 1.8382329460937306e-05, + "loss": 0.7906, + "step": 1917 + }, + { + "epoch": 1.0412595005428882, + "grad_norm": 5.936786319330656, + "learning_rate": 1.838041116673829e-05, + "loss": 0.8254, + "step": 1918 + }, + { + "epoch": 1.0418023887079262, + "grad_norm": 6.637068024744901, + "learning_rate": 1.8378491836036666e-05, + "loss": 1.1779, + "step": 1919 + }, + { + "epoch": 1.0423452768729642, + "grad_norm": 5.481354066687985, + "learning_rate": 1.8376571469069814e-05, + "loss": 0.9586, + "step": 1920 + }, + { + "epoch": 1.0428881650380022, + "grad_norm": 7.669719057080363, + "learning_rate": 1.8374650066075257e-05, + "loss": 1.2058, + "step": 1921 + }, + { + "epoch": 1.0434310532030402, + "grad_norm": 7.387160499244698, + "learning_rate": 1.8372727627290627e-05, + "loss": 1.3861, + "step": 1922 + }, + { + "epoch": 1.0439739413680782, + "grad_norm": 4.975886150156898, + "learning_rate": 1.8370804152953704e-05, + "loss": 0.5849, + "step": 1923 + }, + { + "epoch": 1.0445168295331162, + "grad_norm": 5.413290989228805, + "learning_rate": 1.8368879643302383e-05, + "loss": 0.8425, + "step": 1924 + }, + { + "epoch": 1.0450597176981542, + "grad_norm": 6.564098001301886, + "learning_rate": 1.836695409857469e-05, + "loss": 0.9658, + "step": 1925 + }, + { + "epoch": 1.0456026058631922, + "grad_norm": 8.050721283259312, + "learning_rate": 1.8365027519008774e-05, + "loss": 1.684, + "step": 1926 + }, + { + "epoch": 1.0461454940282302, + "grad_norm": 5.409117554986433, + "learning_rate": 1.836309990484293e-05, + "loss": 0.8247, + "step": 1927 + }, + { + "epoch": 1.0466883821932682, + "grad_norm": 5.108392155413988, + "learning_rate": 1.8361171256315555e-05, + "loss": 0.8244, + "step": 1928 + }, + { + "epoch": 1.0472312703583062, + "grad_norm": 6.806129539780182, + "learning_rate": 1.8359241573665194e-05, + "loss": 1.0678, + "step": 1929 + }, + { + "epoch": 1.0477741585233442, + "grad_norm": 4.43019983774416, + "learning_rate": 1.8357310857130514e-05, + "loss": 0.6213, + "step": 1930 + }, + { + "epoch": 1.0483170466883822, + "grad_norm": 6.194875830854703, + "learning_rate": 1.835537910695031e-05, + "loss": 0.8774, + "step": 1931 + }, + { + "epoch": 1.0488599348534202, + "grad_norm": 4.540451693265499, + "learning_rate": 1.8353446323363496e-05, + "loss": 0.8508, + "step": 1932 + }, + { + "epoch": 1.0494028230184582, + "grad_norm": 5.9491142341914225, + "learning_rate": 1.8351512506609133e-05, + "loss": 1.0139, + "step": 1933 + }, + { + "epoch": 1.0499457111834962, + "grad_norm": 6.436571178648067, + "learning_rate": 1.834957765692639e-05, + "loss": 0.9272, + "step": 1934 + }, + { + "epoch": 1.0504885993485342, + "grad_norm": 5.6538429161280535, + "learning_rate": 1.8347641774554573e-05, + "loss": 0.7834, + "step": 1935 + }, + { + "epoch": 1.0510314875135722, + "grad_norm": 6.481163094620501, + "learning_rate": 1.8345704859733123e-05, + "loss": 1.0916, + "step": 1936 + }, + { + "epoch": 1.0515743756786102, + "grad_norm": 6.165149195118496, + "learning_rate": 1.8343766912701588e-05, + "loss": 1.1054, + "step": 1937 + }, + { + "epoch": 1.0521172638436482, + "grad_norm": 7.778083794032891, + "learning_rate": 1.834182793369967e-05, + "loss": 1.297, + "step": 1938 + }, + { + "epoch": 1.0526601520086862, + "grad_norm": 8.871857523729659, + "learning_rate": 1.8339887922967176e-05, + "loss": 1.5749, + "step": 1939 + }, + { + "epoch": 1.0532030401737242, + "grad_norm": 6.563413115451867, + "learning_rate": 1.8337946880744047e-05, + "loss": 1.0271, + "step": 1940 + }, + { + "epoch": 1.0537459283387622, + "grad_norm": 6.5304770661428275, + "learning_rate": 1.833600480727036e-05, + "loss": 0.9236, + "step": 1941 + }, + { + "epoch": 1.0542888165038002, + "grad_norm": 6.802479270907137, + "learning_rate": 1.8334061702786317e-05, + "loss": 0.9076, + "step": 1942 + }, + { + "epoch": 1.0548317046688382, + "grad_norm": 5.92560525391674, + "learning_rate": 1.833211756753224e-05, + "loss": 0.9601, + "step": 1943 + }, + { + "epoch": 1.0553745928338762, + "grad_norm": 7.436926375817581, + "learning_rate": 1.8330172401748584e-05, + "loss": 1.1758, + "step": 1944 + }, + { + "epoch": 1.0559174809989142, + "grad_norm": 6.369742715679364, + "learning_rate": 1.8328226205675927e-05, + "loss": 0.8424, + "step": 1945 + }, + { + "epoch": 1.0564603691639523, + "grad_norm": 7.60632781901004, + "learning_rate": 1.8326278979554976e-05, + "loss": 1.7815, + "step": 1946 + }, + { + "epoch": 1.0570032573289903, + "grad_norm": 8.499242837541805, + "learning_rate": 1.8324330723626578e-05, + "loss": 1.1341, + "step": 1947 + }, + { + "epoch": 1.0575461454940283, + "grad_norm": 6.469082998862602, + "learning_rate": 1.8322381438131686e-05, + "loss": 0.7798, + "step": 1948 + }, + { + "epoch": 1.0580890336590663, + "grad_norm": 5.923622010018089, + "learning_rate": 1.8320431123311388e-05, + "loss": 0.7971, + "step": 1949 + }, + { + "epoch": 1.0586319218241043, + "grad_norm": 6.617488371294151, + "learning_rate": 1.8318479779406914e-05, + "loss": 1.1551, + "step": 1950 + }, + { + "epoch": 1.0591748099891423, + "grad_norm": 6.813827391520987, + "learning_rate": 1.8316527406659604e-05, + "loss": 1.0234, + "step": 1951 + }, + { + "epoch": 1.0597176981541803, + "grad_norm": 7.580867549521637, + "learning_rate": 1.831457400531093e-05, + "loss": 0.7896, + "step": 1952 + }, + { + "epoch": 1.0602605863192183, + "grad_norm": 5.975929968595557, + "learning_rate": 1.8312619575602486e-05, + "loss": 1.1248, + "step": 1953 + }, + { + "epoch": 1.0608034744842563, + "grad_norm": 6.2878616306091955, + "learning_rate": 1.831066411777601e-05, + "loss": 1.0833, + "step": 1954 + }, + { + "epoch": 1.0613463626492943, + "grad_norm": 6.284963948133933, + "learning_rate": 1.8308707632073345e-05, + "loss": 1.0773, + "step": 1955 + }, + { + "epoch": 1.0618892508143323, + "grad_norm": 4.009278799804313, + "learning_rate": 1.830675011873648e-05, + "loss": 0.4782, + "step": 1956 + }, + { + "epoch": 1.0624321389793703, + "grad_norm": 6.648653088879531, + "learning_rate": 1.8304791578007524e-05, + "loss": 0.9761, + "step": 1957 + }, + { + "epoch": 1.0629750271444083, + "grad_norm": 6.867797713608168, + "learning_rate": 1.830283201012871e-05, + "loss": 0.9315, + "step": 1958 + }, + { + "epoch": 1.0635179153094463, + "grad_norm": 5.989268423819992, + "learning_rate": 1.83008714153424e-05, + "loss": 0.8312, + "step": 1959 + }, + { + "epoch": 1.0640608034744843, + "grad_norm": 6.207518262073172, + "learning_rate": 1.8298909793891083e-05, + "loss": 1.1003, + "step": 1960 + }, + { + "epoch": 1.0646036916395223, + "grad_norm": 5.5435710943195815, + "learning_rate": 1.8296947146017373e-05, + "loss": 0.8012, + "step": 1961 + }, + { + "epoch": 1.0651465798045603, + "grad_norm": 8.000995164961113, + "learning_rate": 1.829498347196402e-05, + "loss": 0.9864, + "step": 1962 + }, + { + "epoch": 1.0656894679695983, + "grad_norm": 6.483090321755992, + "learning_rate": 1.829301877197389e-05, + "loss": 1.2823, + "step": 1963 + }, + { + "epoch": 1.0662323561346363, + "grad_norm": 7.147142303433649, + "learning_rate": 1.8291053046289985e-05, + "loss": 0.9185, + "step": 1964 + }, + { + "epoch": 1.0667752442996743, + "grad_norm": 6.000816189695904, + "learning_rate": 1.828908629515542e-05, + "loss": 1.0284, + "step": 1965 + }, + { + "epoch": 1.0673181324647123, + "grad_norm": 6.188793151434744, + "learning_rate": 1.8287118518813453e-05, + "loss": 0.7041, + "step": 1966 + }, + { + "epoch": 1.0678610206297503, + "grad_norm": 7.353722352349769, + "learning_rate": 1.828514971750746e-05, + "loss": 1.0707, + "step": 1967 + }, + { + "epoch": 1.0684039087947883, + "grad_norm": 7.553378116857428, + "learning_rate": 1.8283179891480944e-05, + "loss": 1.1426, + "step": 1968 + }, + { + "epoch": 1.0689467969598263, + "grad_norm": 7.134594743515001, + "learning_rate": 1.828120904097754e-05, + "loss": 0.9583, + "step": 1969 + }, + { + "epoch": 1.0694896851248643, + "grad_norm": 5.680617024729498, + "learning_rate": 1.8279237166241004e-05, + "loss": 0.9632, + "step": 1970 + }, + { + "epoch": 1.0700325732899023, + "grad_norm": 5.932910663063727, + "learning_rate": 1.8277264267515218e-05, + "loss": 0.9984, + "step": 1971 + }, + { + "epoch": 1.0705754614549403, + "grad_norm": 7.833159555663921, + "learning_rate": 1.8275290345044198e-05, + "loss": 0.8038, + "step": 1972 + }, + { + "epoch": 1.0711183496199783, + "grad_norm": 6.429988603781813, + "learning_rate": 1.8273315399072076e-05, + "loss": 0.9125, + "step": 1973 + }, + { + "epoch": 1.0716612377850163, + "grad_norm": 7.955595430340166, + "learning_rate": 1.827133942984312e-05, + "loss": 0.8697, + "step": 1974 + }, + { + "epoch": 1.0722041259500543, + "grad_norm": 5.983514211221058, + "learning_rate": 1.826936243760172e-05, + "loss": 0.8155, + "step": 1975 + }, + { + "epoch": 1.0727470141150923, + "grad_norm": 6.19913176090901, + "learning_rate": 1.8267384422592398e-05, + "loss": 1.3207, + "step": 1976 + }, + { + "epoch": 1.0732899022801303, + "grad_norm": 6.467400855860047, + "learning_rate": 1.8265405385059792e-05, + "loss": 1.1279, + "step": 1977 + }, + { + "epoch": 1.0738327904451683, + "grad_norm": 6.515150741776156, + "learning_rate": 1.8263425325248675e-05, + "loss": 1.372, + "step": 1978 + }, + { + "epoch": 1.0743756786102063, + "grad_norm": 6.181497786496646, + "learning_rate": 1.8261444243403945e-05, + "loss": 0.8571, + "step": 1979 + }, + { + "epoch": 1.0749185667752443, + "grad_norm": 7.637816753034266, + "learning_rate": 1.8259462139770624e-05, + "loss": 1.4861, + "step": 1980 + }, + { + "epoch": 1.0754614549402823, + "grad_norm": 7.0094523855293485, + "learning_rate": 1.825747901459386e-05, + "loss": 1.0113, + "step": 1981 + }, + { + "epoch": 1.0760043431053203, + "grad_norm": 7.6230550634556655, + "learning_rate": 1.8255494868118933e-05, + "loss": 1.3818, + "step": 1982 + }, + { + "epoch": 1.0765472312703583, + "grad_norm": 6.946643864799888, + "learning_rate": 1.8253509700591242e-05, + "loss": 1.1384, + "step": 1983 + }, + { + "epoch": 1.0770901194353963, + "grad_norm": 7.288235795183878, + "learning_rate": 1.825152351225632e-05, + "loss": 1.1288, + "step": 1984 + }, + { + "epoch": 1.0776330076004343, + "grad_norm": 8.872287367521855, + "learning_rate": 1.8249536303359816e-05, + "loss": 1.6016, + "step": 1985 + }, + { + "epoch": 1.0781758957654723, + "grad_norm": 5.75562973378046, + "learning_rate": 1.8247548074147515e-05, + "loss": 1.1803, + "step": 1986 + }, + { + "epoch": 1.0787187839305103, + "grad_norm": 6.377752945158632, + "learning_rate": 1.824555882486532e-05, + "loss": 0.6991, + "step": 1987 + }, + { + "epoch": 1.0792616720955484, + "grad_norm": 7.178965937765225, + "learning_rate": 1.8243568555759274e-05, + "loss": 0.988, + "step": 1988 + }, + { + "epoch": 1.0798045602605864, + "grad_norm": 6.0869015160234525, + "learning_rate": 1.824157726707553e-05, + "loss": 0.8587, + "step": 1989 + }, + { + "epoch": 1.0803474484256244, + "grad_norm": 5.9113747723637395, + "learning_rate": 1.823958495906037e-05, + "loss": 0.8815, + "step": 1990 + }, + { + "epoch": 1.0808903365906624, + "grad_norm": 7.84172234151508, + "learning_rate": 1.8237591631960218e-05, + "loss": 1.2621, + "step": 1991 + }, + { + "epoch": 1.0814332247557004, + "grad_norm": 7.188505727546616, + "learning_rate": 1.8235597286021597e-05, + "loss": 0.5991, + "step": 1992 + }, + { + "epoch": 1.0819761129207384, + "grad_norm": 6.358900329219305, + "learning_rate": 1.823360192149118e-05, + "loss": 1.3325, + "step": 1993 + }, + { + "epoch": 1.0825190010857764, + "grad_norm": 6.545296131842583, + "learning_rate": 1.8231605538615756e-05, + "loss": 1.1091, + "step": 1994 + }, + { + "epoch": 1.0830618892508144, + "grad_norm": 6.708966367161533, + "learning_rate": 1.8229608137642238e-05, + "loss": 1.1352, + "step": 1995 + }, + { + "epoch": 1.0836047774158524, + "grad_norm": 6.651776068492304, + "learning_rate": 1.822760971881767e-05, + "loss": 0.6818, + "step": 1996 + }, + { + "epoch": 1.0841476655808904, + "grad_norm": 4.745518059898734, + "learning_rate": 1.8225610282389222e-05, + "loss": 0.9403, + "step": 1997 + }, + { + "epoch": 1.0846905537459284, + "grad_norm": 6.526942367542791, + "learning_rate": 1.8223609828604184e-05, + "loss": 1.0543, + "step": 1998 + }, + { + "epoch": 1.0852334419109664, + "grad_norm": 5.331318824902818, + "learning_rate": 1.8221608357709973e-05, + "loss": 0.6626, + "step": 1999 + }, + { + "epoch": 1.0857763300760044, + "grad_norm": 6.640324946305343, + "learning_rate": 1.8219605869954134e-05, + "loss": 1.0166, + "step": 2000 + }, + { + "epoch": 1.0863192182410424, + "grad_norm": 5.684199608008352, + "learning_rate": 1.8217602365584352e-05, + "loss": 0.6796, + "step": 2001 + }, + { + "epoch": 1.0868621064060804, + "grad_norm": 5.429738239817468, + "learning_rate": 1.8215597844848403e-05, + "loss": 0.6783, + "step": 2002 + }, + { + "epoch": 1.0874049945711184, + "grad_norm": 6.1832795338939235, + "learning_rate": 1.821359230799422e-05, + "loss": 1.4244, + "step": 2003 + }, + { + "epoch": 1.0879478827361564, + "grad_norm": 7.118410329483546, + "learning_rate": 1.8211585755269852e-05, + "loss": 1.0274, + "step": 2004 + }, + { + "epoch": 1.0884907709011944, + "grad_norm": 5.5061218005078265, + "learning_rate": 1.820957818692347e-05, + "loss": 0.7143, + "step": 2005 + }, + { + "epoch": 1.0890336590662324, + "grad_norm": 5.518412565715972, + "learning_rate": 1.8207569603203373e-05, + "loss": 1.3158, + "step": 2006 + }, + { + "epoch": 1.0895765472312704, + "grad_norm": 5.47000403426222, + "learning_rate": 1.820556000435799e-05, + "loss": 1.0694, + "step": 2007 + }, + { + "epoch": 1.0901194353963084, + "grad_norm": 7.469961154104622, + "learning_rate": 1.820354939063586e-05, + "loss": 1.0702, + "step": 2008 + }, + { + "epoch": 1.0906623235613464, + "grad_norm": 5.889961089255285, + "learning_rate": 1.8201537762285674e-05, + "loss": 1.0598, + "step": 2009 + }, + { + "epoch": 1.0912052117263844, + "grad_norm": 6.659595389259375, + "learning_rate": 1.8199525119556226e-05, + "loss": 0.8456, + "step": 2010 + }, + { + "epoch": 1.0917480998914224, + "grad_norm": 6.328157257432415, + "learning_rate": 1.8197511462696443e-05, + "loss": 0.9763, + "step": 2011 + }, + { + "epoch": 1.0922909880564604, + "grad_norm": 5.916734142619132, + "learning_rate": 1.8195496791955373e-05, + "loss": 1.1194, + "step": 2012 + }, + { + "epoch": 1.0928338762214984, + "grad_norm": 5.904087624332602, + "learning_rate": 1.8193481107582203e-05, + "loss": 0.756, + "step": 2013 + }, + { + "epoch": 1.0933767643865364, + "grad_norm": 6.284745154848233, + "learning_rate": 1.8191464409826227e-05, + "loss": 0.7931, + "step": 2014 + }, + { + "epoch": 1.0939196525515744, + "grad_norm": 7.838306243079945, + "learning_rate": 1.8189446698936878e-05, + "loss": 1.3635, + "step": 2015 + }, + { + "epoch": 1.0944625407166124, + "grad_norm": 6.3405758352546995, + "learning_rate": 1.818742797516371e-05, + "loss": 0.5365, + "step": 2016 + }, + { + "epoch": 1.0950054288816504, + "grad_norm": 8.315910084823138, + "learning_rate": 1.8185408238756405e-05, + "loss": 1.6592, + "step": 2017 + }, + { + "epoch": 1.0955483170466884, + "grad_norm": 5.205053142209431, + "learning_rate": 1.8183387489964762e-05, + "loss": 0.6416, + "step": 2018 + }, + { + "epoch": 1.0960912052117264, + "grad_norm": 7.018373673519811, + "learning_rate": 1.8181365729038706e-05, + "loss": 1.0814, + "step": 2019 + }, + { + "epoch": 1.0966340933767644, + "grad_norm": 5.705860306884454, + "learning_rate": 1.8179342956228307e-05, + "loss": 0.5806, + "step": 2020 + }, + { + "epoch": 1.0971769815418024, + "grad_norm": 7.077662290720525, + "learning_rate": 1.8177319171783728e-05, + "loss": 0.8969, + "step": 2021 + }, + { + "epoch": 1.0977198697068404, + "grad_norm": 6.540901861387866, + "learning_rate": 1.8175294375955284e-05, + "loss": 0.7982, + "step": 2022 + }, + { + "epoch": 1.0982627578718784, + "grad_norm": 7.283213833491601, + "learning_rate": 1.81732685689934e-05, + "loss": 1.2659, + "step": 2023 + }, + { + "epoch": 1.0988056460369164, + "grad_norm": 7.016589270114781, + "learning_rate": 1.8171241751148633e-05, + "loss": 1.0011, + "step": 2024 + }, + { + "epoch": 1.0993485342019544, + "grad_norm": 5.724817250167612, + "learning_rate": 1.8169213922671666e-05, + "loss": 0.8967, + "step": 2025 + }, + { + "epoch": 1.0998914223669924, + "grad_norm": 7.431207902407107, + "learning_rate": 1.81671850838133e-05, + "loss": 0.9445, + "step": 2026 + }, + { + "epoch": 1.1004343105320304, + "grad_norm": 6.9152342645144955, + "learning_rate": 1.816515523482447e-05, + "loss": 0.8187, + "step": 2027 + }, + { + "epoch": 1.1009771986970684, + "grad_norm": 7.045028624610342, + "learning_rate": 1.816312437595622e-05, + "loss": 0.7323, + "step": 2028 + }, + { + "epoch": 1.1015200868621065, + "grad_norm": 6.63265573559803, + "learning_rate": 1.816109250745974e-05, + "loss": 0.6689, + "step": 2029 + }, + { + "epoch": 1.1020629750271445, + "grad_norm": 8.605675554379312, + "learning_rate": 1.8159059629586333e-05, + "loss": 1.3733, + "step": 2030 + }, + { + "epoch": 1.1026058631921825, + "grad_norm": 7.481808301387036, + "learning_rate": 1.8157025742587426e-05, + "loss": 0.9767, + "step": 2031 + }, + { + "epoch": 1.1031487513572205, + "grad_norm": 7.113214270917786, + "learning_rate": 1.8154990846714575e-05, + "loss": 0.6365, + "step": 2032 + }, + { + "epoch": 1.1036916395222585, + "grad_norm": 7.672152203903941, + "learning_rate": 1.8152954942219462e-05, + "loss": 1.3314, + "step": 2033 + }, + { + "epoch": 1.1042345276872965, + "grad_norm": 7.533257156973554, + "learning_rate": 1.8150918029353885e-05, + "loss": 0.967, + "step": 2034 + }, + { + "epoch": 1.1047774158523345, + "grad_norm": 6.469759228237502, + "learning_rate": 1.8148880108369775e-05, + "loss": 1.348, + "step": 2035 + }, + { + "epoch": 1.1053203040173725, + "grad_norm": 6.620692148711997, + "learning_rate": 1.8146841179519186e-05, + "loss": 0.7538, + "step": 2036 + }, + { + "epoch": 1.1058631921824105, + "grad_norm": 6.974705446587608, + "learning_rate": 1.8144801243054297e-05, + "loss": 1.3664, + "step": 2037 + }, + { + "epoch": 1.1064060803474485, + "grad_norm": 8.047700738112871, + "learning_rate": 1.8142760299227408e-05, + "loss": 1.0466, + "step": 2038 + }, + { + "epoch": 1.1069489685124865, + "grad_norm": 5.469412128245213, + "learning_rate": 1.814071834829095e-05, + "loss": 0.5464, + "step": 2039 + }, + { + "epoch": 1.1074918566775245, + "grad_norm": 5.732997692484992, + "learning_rate": 1.813867539049747e-05, + "loss": 0.7307, + "step": 2040 + }, + { + "epoch": 1.1080347448425625, + "grad_norm": 8.89935072196041, + "learning_rate": 1.8136631426099646e-05, + "loss": 1.8231, + "step": 2041 + }, + { + "epoch": 1.1085776330076005, + "grad_norm": 7.0036806828899465, + "learning_rate": 1.813458645535028e-05, + "loss": 1.063, + "step": 2042 + }, + { + "epoch": 1.1091205211726385, + "grad_norm": 6.013891951698058, + "learning_rate": 1.8132540478502297e-05, + "loss": 0.9836, + "step": 2043 + }, + { + "epoch": 1.1096634093376765, + "grad_norm": 4.234811309290956, + "learning_rate": 1.813049349580875e-05, + "loss": 0.3708, + "step": 2044 + }, + { + "epoch": 1.1102062975027145, + "grad_norm": 6.259459841046158, + "learning_rate": 1.8128445507522806e-05, + "loss": 0.8463, + "step": 2045 + }, + { + "epoch": 1.1107491856677525, + "grad_norm": 7.9568342619649775, + "learning_rate": 1.8126396513897764e-05, + "loss": 1.0563, + "step": 2046 + }, + { + "epoch": 1.1112920738327905, + "grad_norm": 6.937848428371049, + "learning_rate": 1.8124346515187056e-05, + "loss": 0.6359, + "step": 2047 + }, + { + "epoch": 1.1118349619978285, + "grad_norm": 7.672632972051955, + "learning_rate": 1.8122295511644218e-05, + "loss": 0.657, + "step": 2048 + }, + { + "epoch": 1.1123778501628665, + "grad_norm": 6.450214026534184, + "learning_rate": 1.8120243503522924e-05, + "loss": 0.7279, + "step": 2049 + }, + { + "epoch": 1.1129207383279045, + "grad_norm": 6.628216836060006, + "learning_rate": 1.8118190491076978e-05, + "loss": 0.5804, + "step": 2050 + }, + { + "epoch": 1.1134636264929425, + "grad_norm": 6.698074316145226, + "learning_rate": 1.8116136474560288e-05, + "loss": 1.011, + "step": 2051 + }, + { + "epoch": 1.1140065146579805, + "grad_norm": 5.4126983291571795, + "learning_rate": 1.8114081454226905e-05, + "loss": 0.8564, + "step": 2052 + }, + { + "epoch": 1.1145494028230185, + "grad_norm": 8.596797841329312, + "learning_rate": 1.8112025430331e-05, + "loss": 1.1924, + "step": 2053 + }, + { + "epoch": 1.1150922909880565, + "grad_norm": 7.416912053391948, + "learning_rate": 1.8109968403126856e-05, + "loss": 0.9894, + "step": 2054 + }, + { + "epoch": 1.1156351791530945, + "grad_norm": 7.434308789484601, + "learning_rate": 1.8107910372868898e-05, + "loss": 1.1544, + "step": 2055 + }, + { + "epoch": 1.1161780673181325, + "grad_norm": 8.325689788099726, + "learning_rate": 1.8105851339811663e-05, + "loss": 0.9994, + "step": 2056 + }, + { + "epoch": 1.1167209554831705, + "grad_norm": 8.384964960696278, + "learning_rate": 1.8103791304209813e-05, + "loss": 1.1768, + "step": 2057 + }, + { + "epoch": 1.1172638436482085, + "grad_norm": 7.044535285708613, + "learning_rate": 1.810173026631814e-05, + "loss": 1.1801, + "step": 2058 + }, + { + "epoch": 1.1178067318132465, + "grad_norm": 6.665810359686943, + "learning_rate": 1.8099668226391552e-05, + "loss": 0.7865, + "step": 2059 + }, + { + "epoch": 1.1183496199782845, + "grad_norm": 7.395697719004186, + "learning_rate": 1.8097605184685093e-05, + "loss": 1.1357, + "step": 2060 + }, + { + "epoch": 1.1188925081433225, + "grad_norm": 7.1495559268215185, + "learning_rate": 1.809554114145392e-05, + "loss": 1.1507, + "step": 2061 + }, + { + "epoch": 1.1194353963083605, + "grad_norm": 7.877063869804663, + "learning_rate": 1.8093476096953315e-05, + "loss": 1.634, + "step": 2062 + }, + { + "epoch": 1.1199782844733985, + "grad_norm": 6.528060830825324, + "learning_rate": 1.809141005143869e-05, + "loss": 0.697, + "step": 2063 + }, + { + "epoch": 1.1205211726384365, + "grad_norm": 6.5949576624011, + "learning_rate": 1.808934300516557e-05, + "loss": 1.1506, + "step": 2064 + }, + { + "epoch": 1.1210640608034745, + "grad_norm": 6.787348272435354, + "learning_rate": 1.8087274958389612e-05, + "loss": 0.9986, + "step": 2065 + }, + { + "epoch": 1.1216069489685125, + "grad_norm": 6.310955809912108, + "learning_rate": 1.8085205911366602e-05, + "loss": 0.89, + "step": 2066 + }, + { + "epoch": 1.1221498371335505, + "grad_norm": 6.8487503881448895, + "learning_rate": 1.8083135864352442e-05, + "loss": 1.1061, + "step": 2067 + }, + { + "epoch": 1.1226927252985885, + "grad_norm": 5.329494375939551, + "learning_rate": 1.808106481760315e-05, + "loss": 1.0011, + "step": 2068 + }, + { + "epoch": 1.1232356134636265, + "grad_norm": 6.286005313228098, + "learning_rate": 1.8078992771374886e-05, + "loss": 0.8399, + "step": 2069 + }, + { + "epoch": 1.1237785016286646, + "grad_norm": 6.324896503266866, + "learning_rate": 1.8076919725923917e-05, + "loss": 1.2716, + "step": 2070 + }, + { + "epoch": 1.1243213897937026, + "grad_norm": 5.663105266452551, + "learning_rate": 1.8074845681506644e-05, + "loss": 0.9943, + "step": 2071 + }, + { + "epoch": 1.1248642779587406, + "grad_norm": 5.175572461557621, + "learning_rate": 1.807277063837959e-05, + "loss": 0.6319, + "step": 2072 + }, + { + "epoch": 1.1254071661237786, + "grad_norm": 7.297487212372251, + "learning_rate": 1.8070694596799397e-05, + "loss": 0.7533, + "step": 2073 + }, + { + "epoch": 1.1259500542888166, + "grad_norm": 9.046578233072974, + "learning_rate": 1.806861755702283e-05, + "loss": 0.9991, + "step": 2074 + }, + { + "epoch": 1.1264929424538546, + "grad_norm": 4.516511292735119, + "learning_rate": 1.8066539519306786e-05, + "loss": 0.6393, + "step": 2075 + }, + { + "epoch": 1.1270358306188926, + "grad_norm": 5.16623460840236, + "learning_rate": 1.8064460483908283e-05, + "loss": 0.7083, + "step": 2076 + }, + { + "epoch": 1.1275787187839306, + "grad_norm": 10.7323945119629, + "learning_rate": 1.8062380451084445e-05, + "loss": 2.4193, + "step": 2077 + }, + { + "epoch": 1.1281216069489686, + "grad_norm": 6.438876475868993, + "learning_rate": 1.8060299421092554e-05, + "loss": 0.8468, + "step": 2078 + }, + { + "epoch": 1.1286644951140066, + "grad_norm": 8.437791617086905, + "learning_rate": 1.8058217394189976e-05, + "loss": 1.4895, + "step": 2079 + }, + { + "epoch": 1.1292073832790446, + "grad_norm": 5.237528459736948, + "learning_rate": 1.805613437063423e-05, + "loss": 0.643, + "step": 2080 + }, + { + "epoch": 1.1297502714440826, + "grad_norm": 6.010151293762264, + "learning_rate": 1.8054050350682947e-05, + "loss": 0.9001, + "step": 2081 + }, + { + "epoch": 1.1302931596091206, + "grad_norm": 4.816849804237718, + "learning_rate": 1.805196533459388e-05, + "loss": 0.6349, + "step": 2082 + }, + { + "epoch": 1.1308360477741586, + "grad_norm": 6.872496495387926, + "learning_rate": 1.8049879322624906e-05, + "loss": 0.8808, + "step": 2083 + }, + { + "epoch": 1.1313789359391966, + "grad_norm": 5.86067647581003, + "learning_rate": 1.804779231503403e-05, + "loss": 0.6811, + "step": 2084 + }, + { + "epoch": 1.1319218241042346, + "grad_norm": 6.073217718554074, + "learning_rate": 1.8045704312079376e-05, + "loss": 1.0086, + "step": 2085 + }, + { + "epoch": 1.1324647122692726, + "grad_norm": 7.519215180241719, + "learning_rate": 1.804361531401918e-05, + "loss": 0.8815, + "step": 2086 + }, + { + "epoch": 1.1330076004343106, + "grad_norm": 6.49012361077388, + "learning_rate": 1.8041525321111835e-05, + "loss": 1.0129, + "step": 2087 + }, + { + "epoch": 1.1335504885993486, + "grad_norm": 5.938773374319115, + "learning_rate": 1.8039434333615814e-05, + "loss": 0.8485, + "step": 2088 + }, + { + "epoch": 1.1340933767643866, + "grad_norm": 5.607636698742016, + "learning_rate": 1.8037342351789743e-05, + "loss": 0.6209, + "step": 2089 + }, + { + "epoch": 1.1346362649294246, + "grad_norm": 5.522935966563286, + "learning_rate": 1.803524937589236e-05, + "loss": 0.6576, + "step": 2090 + }, + { + "epoch": 1.1351791530944626, + "grad_norm": 8.130974168982206, + "learning_rate": 1.8033155406182533e-05, + "loss": 1.0676, + "step": 2091 + }, + { + "epoch": 1.1357220412595006, + "grad_norm": 6.027276275319545, + "learning_rate": 1.803106044291924e-05, + "loss": 0.9853, + "step": 2092 + }, + { + "epoch": 1.1362649294245386, + "grad_norm": 4.645985748184023, + "learning_rate": 1.8028964486361586e-05, + "loss": 0.7505, + "step": 2093 + }, + { + "epoch": 1.1368078175895766, + "grad_norm": 6.820894579082682, + "learning_rate": 1.8026867536768816e-05, + "loss": 0.9358, + "step": 2094 + }, + { + "epoch": 1.1373507057546146, + "grad_norm": 3.9530727606720872, + "learning_rate": 1.802476959440027e-05, + "loss": 0.7736, + "step": 2095 + }, + { + "epoch": 1.1378935939196526, + "grad_norm": 7.650275740883737, + "learning_rate": 1.8022670659515432e-05, + "loss": 1.4832, + "step": 2096 + }, + { + "epoch": 1.1384364820846906, + "grad_norm": 6.94341164115752, + "learning_rate": 1.80205707323739e-05, + "loss": 1.0336, + "step": 2097 + }, + { + "epoch": 1.1389793702497286, + "grad_norm": 5.608528053150328, + "learning_rate": 1.8018469813235403e-05, + "loss": 0.829, + "step": 2098 + }, + { + "epoch": 1.1395222584147666, + "grad_norm": 5.2191336182398995, + "learning_rate": 1.8016367902359776e-05, + "loss": 0.9708, + "step": 2099 + }, + { + "epoch": 1.1400651465798046, + "grad_norm": 6.773346810380977, + "learning_rate": 1.801426500000699e-05, + "loss": 0.6755, + "step": 2100 + }, + { + "epoch": 1.1406080347448426, + "grad_norm": 7.292992073718992, + "learning_rate": 1.8012161106437137e-05, + "loss": 0.8842, + "step": 2101 + }, + { + "epoch": 1.1411509229098806, + "grad_norm": 6.699434744454233, + "learning_rate": 1.8010056221910427e-05, + "loss": 1.0385, + "step": 2102 + }, + { + "epoch": 1.1416938110749186, + "grad_norm": 6.565299107509569, + "learning_rate": 1.8007950346687198e-05, + "loss": 0.6557, + "step": 2103 + }, + { + "epoch": 1.1422366992399566, + "grad_norm": 7.2741514068627025, + "learning_rate": 1.800584348102791e-05, + "loss": 1.0347, + "step": 2104 + }, + { + "epoch": 1.1427795874049946, + "grad_norm": 6.467241522225306, + "learning_rate": 1.800373562519314e-05, + "loss": 0.8026, + "step": 2105 + }, + { + "epoch": 1.1433224755700326, + "grad_norm": 6.109318081193457, + "learning_rate": 1.800162677944359e-05, + "loss": 0.9182, + "step": 2106 + }, + { + "epoch": 1.1438653637350706, + "grad_norm": 7.136347555199431, + "learning_rate": 1.7999516944040087e-05, + "loss": 1.2743, + "step": 2107 + }, + { + "epoch": 1.1444082519001086, + "grad_norm": 7.791284332349926, + "learning_rate": 1.7997406119243582e-05, + "loss": 0.9738, + "step": 2108 + }, + { + "epoch": 1.1449511400651466, + "grad_norm": 7.315957477694666, + "learning_rate": 1.7995294305315137e-05, + "loss": 1.0674, + "step": 2109 + }, + { + "epoch": 1.1454940282301846, + "grad_norm": 6.305984776426275, + "learning_rate": 1.7993181502515957e-05, + "loss": 0.8204, + "step": 2110 + }, + { + "epoch": 1.1460369163952226, + "grad_norm": 5.824114865699788, + "learning_rate": 1.7991067711107345e-05, + "loss": 0.7953, + "step": 2111 + }, + { + "epoch": 1.1465798045602607, + "grad_norm": 7.4284096250312786, + "learning_rate": 1.798895293135074e-05, + "loss": 0.7349, + "step": 2112 + }, + { + "epoch": 1.1471226927252987, + "grad_norm": 6.415949338549894, + "learning_rate": 1.798683716350771e-05, + "loss": 0.8346, + "step": 2113 + }, + { + "epoch": 1.1476655808903367, + "grad_norm": 6.360883520037482, + "learning_rate": 1.7984720407839925e-05, + "loss": 0.7966, + "step": 2114 + }, + { + "epoch": 1.1482084690553747, + "grad_norm": 6.703196393264975, + "learning_rate": 1.7982602664609198e-05, + "loss": 1.3341, + "step": 2115 + }, + { + "epoch": 1.1487513572204127, + "grad_norm": 8.115384707506347, + "learning_rate": 1.798048393407745e-05, + "loss": 1.0284, + "step": 2116 + }, + { + "epoch": 1.1492942453854507, + "grad_norm": 8.519166176194911, + "learning_rate": 1.797836421650673e-05, + "loss": 1.4075, + "step": 2117 + }, + { + "epoch": 1.1498371335504887, + "grad_norm": 7.120345871314324, + "learning_rate": 1.7976243512159207e-05, + "loss": 1.1933, + "step": 2118 + }, + { + "epoch": 1.1503800217155267, + "grad_norm": 5.887743000051609, + "learning_rate": 1.7974121821297178e-05, + "loss": 0.9143, + "step": 2119 + }, + { + "epoch": 1.1509229098805647, + "grad_norm": 6.264158167802545, + "learning_rate": 1.797199914418305e-05, + "loss": 0.805, + "step": 2120 + }, + { + "epoch": 1.1514657980456027, + "grad_norm": 7.747345696360016, + "learning_rate": 1.7969875481079363e-05, + "loss": 1.1374, + "step": 2121 + }, + { + "epoch": 1.1520086862106407, + "grad_norm": 5.932381639003466, + "learning_rate": 1.7967750832248774e-05, + "loss": 0.7801, + "step": 2122 + }, + { + "epoch": 1.1525515743756787, + "grad_norm": 5.646019326027894, + "learning_rate": 1.7965625197954064e-05, + "loss": 0.5053, + "step": 2123 + }, + { + "epoch": 1.1530944625407167, + "grad_norm": 6.085604242778934, + "learning_rate": 1.7963498578458135e-05, + "loss": 0.6711, + "step": 2124 + }, + { + "epoch": 1.1536373507057547, + "grad_norm": 6.390132028289734, + "learning_rate": 1.7961370974024008e-05, + "loss": 0.9897, + "step": 2125 + }, + { + "epoch": 1.1541802388707927, + "grad_norm": 5.052558465006266, + "learning_rate": 1.795924238491483e-05, + "loss": 0.8742, + "step": 2126 + }, + { + "epoch": 1.1547231270358307, + "grad_norm": 5.86713723099216, + "learning_rate": 1.7957112811393867e-05, + "loss": 0.965, + "step": 2127 + }, + { + "epoch": 1.1552660152008687, + "grad_norm": 6.092033075473227, + "learning_rate": 1.7954982253724515e-05, + "loss": 1.104, + "step": 2128 + }, + { + "epoch": 1.1558089033659067, + "grad_norm": 7.051795930424219, + "learning_rate": 1.7952850712170278e-05, + "loss": 1.0918, + "step": 2129 + }, + { + "epoch": 1.1563517915309447, + "grad_norm": 5.613114568974867, + "learning_rate": 1.7950718186994788e-05, + "loss": 0.6698, + "step": 2130 + }, + { + "epoch": 1.1568946796959827, + "grad_norm": 5.533439816776461, + "learning_rate": 1.7948584678461802e-05, + "loss": 1.4395, + "step": 2131 + }, + { + "epoch": 1.1574375678610207, + "grad_norm": 7.73758559868109, + "learning_rate": 1.7946450186835195e-05, + "loss": 1.0806, + "step": 2132 + }, + { + "epoch": 1.1579804560260587, + "grad_norm": 5.982378084297589, + "learning_rate": 1.7944314712378966e-05, + "loss": 1.0428, + "step": 2133 + }, + { + "epoch": 1.1585233441910967, + "grad_norm": 5.604421383798244, + "learning_rate": 1.7942178255357227e-05, + "loss": 0.9411, + "step": 2134 + }, + { + "epoch": 1.1590662323561347, + "grad_norm": 6.5637069585312435, + "learning_rate": 1.794004081603423e-05, + "loss": 0.8285, + "step": 2135 + }, + { + "epoch": 1.1596091205211727, + "grad_norm": 5.567535422744373, + "learning_rate": 1.7937902394674326e-05, + "loss": 1.0185, + "step": 2136 + }, + { + "epoch": 1.1601520086862107, + "grad_norm": 7.898198443147329, + "learning_rate": 1.7935762991542004e-05, + "loss": 0.8878, + "step": 2137 + }, + { + "epoch": 1.1606948968512487, + "grad_norm": 5.12407458777131, + "learning_rate": 1.7933622606901865e-05, + "loss": 0.9174, + "step": 2138 + }, + { + "epoch": 1.1612377850162867, + "grad_norm": 4.480165604953252, + "learning_rate": 1.7931481241018642e-05, + "loss": 0.9424, + "step": 2139 + }, + { + "epoch": 1.1617806731813247, + "grad_norm": 7.6708161194386495, + "learning_rate": 1.7929338894157173e-05, + "loss": 0.9875, + "step": 2140 + }, + { + "epoch": 1.1623235613463627, + "grad_norm": 9.749916559657342, + "learning_rate": 1.7927195566582435e-05, + "loss": 1.7168, + "step": 2141 + }, + { + "epoch": 1.1628664495114007, + "grad_norm": 5.50355689991458, + "learning_rate": 1.7925051258559516e-05, + "loss": 0.7957, + "step": 2142 + }, + { + "epoch": 1.1634093376764387, + "grad_norm": 5.910752363178181, + "learning_rate": 1.7922905970353627e-05, + "loss": 1.0525, + "step": 2143 + }, + { + "epoch": 1.1639522258414767, + "grad_norm": 6.523454927461212, + "learning_rate": 1.7920759702230098e-05, + "loss": 0.9938, + "step": 2144 + }, + { + "epoch": 1.1644951140065147, + "grad_norm": 6.963057538162332, + "learning_rate": 1.7918612454454387e-05, + "loss": 0.8677, + "step": 2145 + }, + { + "epoch": 1.1650380021715527, + "grad_norm": 6.55011455398606, + "learning_rate": 1.7916464227292067e-05, + "loss": 0.7268, + "step": 2146 + }, + { + "epoch": 1.1655808903365907, + "grad_norm": 6.47957363507043, + "learning_rate": 1.7914315021008836e-05, + "loss": 0.7449, + "step": 2147 + }, + { + "epoch": 1.1661237785016287, + "grad_norm": 6.574809242111077, + "learning_rate": 1.7912164835870506e-05, + "loss": 0.6999, + "step": 2148 + }, + { + "epoch": 1.1666666666666667, + "grad_norm": 8.307297080129953, + "learning_rate": 1.791001367214302e-05, + "loss": 1.0402, + "step": 2149 + }, + { + "epoch": 1.1672095548317047, + "grad_norm": 8.01469917888599, + "learning_rate": 1.7907861530092438e-05, + "loss": 1.1024, + "step": 2150 + }, + { + "epoch": 1.1677524429967427, + "grad_norm": 7.336381076899337, + "learning_rate": 1.7905708409984938e-05, + "loss": 0.9112, + "step": 2151 + }, + { + "epoch": 1.1682953311617807, + "grad_norm": 7.872105686379811, + "learning_rate": 1.7903554312086823e-05, + "loss": 0.9002, + "step": 2152 + }, + { + "epoch": 1.1688382193268188, + "grad_norm": 5.842761146463113, + "learning_rate": 1.7901399236664514e-05, + "loss": 0.9457, + "step": 2153 + }, + { + "epoch": 1.1693811074918568, + "grad_norm": 5.8100183378475005, + "learning_rate": 1.7899243183984552e-05, + "loss": 0.9675, + "step": 2154 + }, + { + "epoch": 1.1699239956568948, + "grad_norm": 5.208689262592845, + "learning_rate": 1.7897086154313604e-05, + "loss": 0.7562, + "step": 2155 + }, + { + "epoch": 1.1704668838219328, + "grad_norm": 7.695958654055186, + "learning_rate": 1.7894928147918458e-05, + "loss": 0.9995, + "step": 2156 + }, + { + "epoch": 1.1710097719869708, + "grad_norm": 5.8108200610880525, + "learning_rate": 1.789276916506601e-05, + "loss": 0.8179, + "step": 2157 + }, + { + "epoch": 1.1715526601520088, + "grad_norm": 5.776078861908261, + "learning_rate": 1.7890609206023298e-05, + "loss": 0.6612, + "step": 2158 + }, + { + "epoch": 1.1720955483170468, + "grad_norm": 6.458218841014727, + "learning_rate": 1.7888448271057463e-05, + "loss": 0.8419, + "step": 2159 + }, + { + "epoch": 1.1726384364820848, + "grad_norm": 7.236506499419343, + "learning_rate": 1.7886286360435772e-05, + "loss": 0.6336, + "step": 2160 + }, + { + "epoch": 1.1731813246471228, + "grad_norm": 7.20552304926134, + "learning_rate": 1.7884123474425614e-05, + "loss": 0.9157, + "step": 2161 + }, + { + "epoch": 1.1737242128121608, + "grad_norm": 8.090329675463652, + "learning_rate": 1.78819596132945e-05, + "loss": 1.6894, + "step": 2162 + }, + { + "epoch": 1.1742671009771988, + "grad_norm": 7.500510561892478, + "learning_rate": 1.787979477731006e-05, + "loss": 0.9469, + "step": 2163 + }, + { + "epoch": 1.1748099891422368, + "grad_norm": 7.6126459328076415, + "learning_rate": 1.7877628966740044e-05, + "loss": 0.9179, + "step": 2164 + }, + { + "epoch": 1.1753528773072748, + "grad_norm": 7.065751172246334, + "learning_rate": 1.787546218185232e-05, + "loss": 0.9649, + "step": 2165 + }, + { + "epoch": 1.1758957654723128, + "grad_norm": 6.923487399525626, + "learning_rate": 1.787329442291488e-05, + "loss": 1.0791, + "step": 2166 + }, + { + "epoch": 1.1764386536373508, + "grad_norm": 7.614292041669914, + "learning_rate": 1.7871125690195843e-05, + "loss": 1.2367, + "step": 2167 + }, + { + "epoch": 1.1769815418023888, + "grad_norm": 8.560468932577576, + "learning_rate": 1.7868955983963434e-05, + "loss": 1.0876, + "step": 2168 + }, + { + "epoch": 1.1775244299674268, + "grad_norm": 6.209836424501219, + "learning_rate": 1.786678530448601e-05, + "loss": 0.6251, + "step": 2169 + }, + { + "epoch": 1.1780673181324648, + "grad_norm": 5.926645140708837, + "learning_rate": 1.7864613652032035e-05, + "loss": 0.5939, + "step": 2170 + }, + { + "epoch": 1.1786102062975028, + "grad_norm": 5.460593849347875, + "learning_rate": 1.7862441026870114e-05, + "loss": 0.8769, + "step": 2171 + }, + { + "epoch": 1.1791530944625408, + "grad_norm": 8.063368903057139, + "learning_rate": 1.7860267429268954e-05, + "loss": 1.1492, + "step": 2172 + }, + { + "epoch": 1.1796959826275788, + "grad_norm": 6.246398264149539, + "learning_rate": 1.7858092859497392e-05, + "loss": 0.9146, + "step": 2173 + }, + { + "epoch": 1.1802388707926168, + "grad_norm": 9.483282452267977, + "learning_rate": 1.7855917317824383e-05, + "loss": 1.3714, + "step": 2174 + }, + { + "epoch": 1.1807817589576548, + "grad_norm": 7.377757446359213, + "learning_rate": 1.7853740804519e-05, + "loss": 0.7914, + "step": 2175 + }, + { + "epoch": 1.1813246471226928, + "grad_norm": 6.555194448443568, + "learning_rate": 1.7851563319850435e-05, + "loss": 1.0025, + "step": 2176 + }, + { + "epoch": 1.1818675352877308, + "grad_norm": 5.968820734002013, + "learning_rate": 1.7849384864088005e-05, + "loss": 1.0449, + "step": 2177 + }, + { + "epoch": 1.1824104234527688, + "grad_norm": 8.050731398373367, + "learning_rate": 1.7847205437501145e-05, + "loss": 1.2944, + "step": 2178 + }, + { + "epoch": 1.1829533116178068, + "grad_norm": 6.479093872062473, + "learning_rate": 1.784502504035941e-05, + "loss": 0.8312, + "step": 2179 + }, + { + "epoch": 1.1834961997828448, + "grad_norm": 9.645172625032446, + "learning_rate": 1.7842843672932473e-05, + "loss": 1.4877, + "step": 2180 + }, + { + "epoch": 1.1840390879478828, + "grad_norm": 7.811721210097083, + "learning_rate": 1.7840661335490133e-05, + "loss": 1.0861, + "step": 2181 + }, + { + "epoch": 1.1845819761129208, + "grad_norm": 5.939133632197409, + "learning_rate": 1.7838478028302303e-05, + "loss": 0.7359, + "step": 2182 + }, + { + "epoch": 1.1851248642779588, + "grad_norm": 7.121597254982603, + "learning_rate": 1.7836293751639017e-05, + "loss": 0.626, + "step": 2183 + }, + { + "epoch": 1.1856677524429968, + "grad_norm": 7.225734005859657, + "learning_rate": 1.783410850577043e-05, + "loss": 1.0651, + "step": 2184 + }, + { + "epoch": 1.1862106406080348, + "grad_norm": 6.3733995711762965, + "learning_rate": 1.783192229096682e-05, + "loss": 1.0399, + "step": 2185 + }, + { + "epoch": 1.1867535287730728, + "grad_norm": 7.215840642904109, + "learning_rate": 1.7829735107498576e-05, + "loss": 1.0958, + "step": 2186 + }, + { + "epoch": 1.1872964169381108, + "grad_norm": 7.2349637750528135, + "learning_rate": 1.7827546955636216e-05, + "loss": 0.8005, + "step": 2187 + }, + { + "epoch": 1.1878393051031488, + "grad_norm": 7.592927799049675, + "learning_rate": 1.7825357835650376e-05, + "loss": 1.0981, + "step": 2188 + }, + { + "epoch": 1.1883821932681868, + "grad_norm": 8.010034188435748, + "learning_rate": 1.7823167747811805e-05, + "loss": 1.3088, + "step": 2189 + }, + { + "epoch": 1.1889250814332248, + "grad_norm": 5.674359355487917, + "learning_rate": 1.7820976692391377e-05, + "loss": 0.8725, + "step": 2190 + }, + { + "epoch": 1.1894679695982628, + "grad_norm": 5.236415159647034, + "learning_rate": 1.781878466966009e-05, + "loss": 0.7162, + "step": 2191 + }, + { + "epoch": 1.1900108577633008, + "grad_norm": 5.677268305787393, + "learning_rate": 1.781659167988905e-05, + "loss": 0.7101, + "step": 2192 + }, + { + "epoch": 1.1905537459283388, + "grad_norm": 6.370730170332413, + "learning_rate": 1.7814397723349496e-05, + "loss": 0.9087, + "step": 2193 + }, + { + "epoch": 1.1910966340933768, + "grad_norm": 6.045624314512159, + "learning_rate": 1.7812202800312776e-05, + "loss": 1.1714, + "step": 2194 + }, + { + "epoch": 1.1916395222584149, + "grad_norm": 10.798495850925244, + "learning_rate": 1.7810006911050366e-05, + "loss": 1.1997, + "step": 2195 + }, + { + "epoch": 1.1921824104234529, + "grad_norm": 5.89010386817935, + "learning_rate": 1.780781005583385e-05, + "loss": 1.0702, + "step": 2196 + }, + { + "epoch": 1.1927252985884909, + "grad_norm": 6.09421937705502, + "learning_rate": 1.7805612234934946e-05, + "loss": 0.825, + "step": 2197 + }, + { + "epoch": 1.1932681867535289, + "grad_norm": 5.582107046729241, + "learning_rate": 1.780341344862548e-05, + "loss": 0.8548, + "step": 2198 + }, + { + "epoch": 1.1938110749185669, + "grad_norm": 7.598179521038406, + "learning_rate": 1.78012136971774e-05, + "loss": 1.0015, + "step": 2199 + }, + { + "epoch": 1.1943539630836049, + "grad_norm": 7.4068666684121585, + "learning_rate": 1.7799012980862777e-05, + "loss": 1.0645, + "step": 2200 + }, + { + "epoch": 1.1948968512486429, + "grad_norm": 6.087565282375561, + "learning_rate": 1.7796811299953796e-05, + "loss": 0.9169, + "step": 2201 + }, + { + "epoch": 1.1954397394136809, + "grad_norm": 6.440325481735819, + "learning_rate": 1.7794608654722772e-05, + "loss": 1.0707, + "step": 2202 + }, + { + "epoch": 1.1959826275787189, + "grad_norm": 7.562903613167714, + "learning_rate": 1.7792405045442125e-05, + "loss": 1.0086, + "step": 2203 + }, + { + "epoch": 1.1965255157437569, + "grad_norm": 6.2377346284888, + "learning_rate": 1.77902004723844e-05, + "loss": 0.6111, + "step": 2204 + }, + { + "epoch": 1.1970684039087949, + "grad_norm": 5.978299124706355, + "learning_rate": 1.7787994935822268e-05, + "loss": 0.5782, + "step": 2205 + }, + { + "epoch": 1.1976112920738329, + "grad_norm": 5.9766366336106715, + "learning_rate": 1.778578843602851e-05, + "loss": 0.8409, + "step": 2206 + }, + { + "epoch": 1.1981541802388709, + "grad_norm": 7.700010929432555, + "learning_rate": 1.7783580973276027e-05, + "loss": 1.091, + "step": 2207 + }, + { + "epoch": 1.1986970684039089, + "grad_norm": 8.043903647823342, + "learning_rate": 1.778137254783785e-05, + "loss": 1.2894, + "step": 2208 + }, + { + "epoch": 1.1992399565689469, + "grad_norm": 8.709529168595635, + "learning_rate": 1.777916315998711e-05, + "loss": 1.5601, + "step": 2209 + }, + { + "epoch": 1.1997828447339849, + "grad_norm": 6.234520197542074, + "learning_rate": 1.7776952809997073e-05, + "loss": 1.0414, + "step": 2210 + }, + { + "epoch": 1.200325732899023, + "grad_norm": 5.8108372705767, + "learning_rate": 1.7774741498141116e-05, + "loss": 0.8811, + "step": 2211 + }, + { + "epoch": 1.200868621064061, + "grad_norm": 6.423388338643525, + "learning_rate": 1.7772529224692744e-05, + "loss": 0.8371, + "step": 2212 + }, + { + "epoch": 1.201411509229099, + "grad_norm": 7.740996825838731, + "learning_rate": 1.7770315989925565e-05, + "loss": 0.9993, + "step": 2213 + }, + { + "epoch": 1.201954397394137, + "grad_norm": 6.33257469290407, + "learning_rate": 1.7768101794113323e-05, + "loss": 0.8563, + "step": 2214 + }, + { + "epoch": 1.202497285559175, + "grad_norm": 8.126954888789392, + "learning_rate": 1.776588663752987e-05, + "loss": 0.9479, + "step": 2215 + }, + { + "epoch": 1.203040173724213, + "grad_norm": 7.781450888823224, + "learning_rate": 1.7763670520449178e-05, + "loss": 1.0559, + "step": 2216 + }, + { + "epoch": 1.203583061889251, + "grad_norm": 6.615763983931515, + "learning_rate": 1.7761453443145348e-05, + "loss": 0.7463, + "step": 2217 + }, + { + "epoch": 1.204125950054289, + "grad_norm": 10.193394757123674, + "learning_rate": 1.7759235405892584e-05, + "loss": 1.3597, + "step": 2218 + }, + { + "epoch": 1.204668838219327, + "grad_norm": 7.365896137130193, + "learning_rate": 1.7757016408965217e-05, + "loss": 0.5822, + "step": 2219 + }, + { + "epoch": 1.205211726384365, + "grad_norm": 8.530190774406401, + "learning_rate": 1.77547964526377e-05, + "loss": 1.0075, + "step": 2220 + }, + { + "epoch": 1.205754614549403, + "grad_norm": 7.143571088845206, + "learning_rate": 1.7752575537184597e-05, + "loss": 0.9395, + "step": 2221 + }, + { + "epoch": 1.206297502714441, + "grad_norm": 7.790375269151535, + "learning_rate": 1.7750353662880595e-05, + "loss": 1.0556, + "step": 2222 + }, + { + "epoch": 1.206840390879479, + "grad_norm": 6.5033471136457255, + "learning_rate": 1.77481308300005e-05, + "loss": 0.7464, + "step": 2223 + }, + { + "epoch": 1.207383279044517, + "grad_norm": 7.06026002468489, + "learning_rate": 1.774590703881924e-05, + "loss": 0.6315, + "step": 2224 + }, + { + "epoch": 1.207926167209555, + "grad_norm": 8.773109114641178, + "learning_rate": 1.774368228961185e-05, + "loss": 0.9884, + "step": 2225 + }, + { + "epoch": 1.208469055374593, + "grad_norm": 7.343364313617542, + "learning_rate": 1.774145658265349e-05, + "loss": 1.1542, + "step": 2226 + }, + { + "epoch": 1.209011943539631, + "grad_norm": 5.953950803365371, + "learning_rate": 1.773922991821944e-05, + "loss": 0.8485, + "step": 2227 + }, + { + "epoch": 1.209554831704669, + "grad_norm": 8.905484324028157, + "learning_rate": 1.7737002296585105e-05, + "loss": 1.0171, + "step": 2228 + }, + { + "epoch": 1.210097719869707, + "grad_norm": 7.158929117808072, + "learning_rate": 1.7734773718025992e-05, + "loss": 0.6134, + "step": 2229 + }, + { + "epoch": 1.210640608034745, + "grad_norm": 8.342029051568577, + "learning_rate": 1.7732544182817737e-05, + "loss": 0.9775, + "step": 2230 + }, + { + "epoch": 1.211183496199783, + "grad_norm": 5.587278578636139, + "learning_rate": 1.7730313691236098e-05, + "loss": 0.9395, + "step": 2231 + }, + { + "epoch": 1.211726384364821, + "grad_norm": 6.1948883535007475, + "learning_rate": 1.7728082243556936e-05, + "loss": 0.9977, + "step": 2232 + }, + { + "epoch": 1.212269272529859, + "grad_norm": 6.095377323439518, + "learning_rate": 1.7725849840056248e-05, + "loss": 0.9141, + "step": 2233 + }, + { + "epoch": 1.212812160694897, + "grad_norm": 8.71522968216411, + "learning_rate": 1.7723616481010137e-05, + "loss": 1.4356, + "step": 2234 + }, + { + "epoch": 1.213355048859935, + "grad_norm": 5.416863970522961, + "learning_rate": 1.7721382166694834e-05, + "loss": 0.6594, + "step": 2235 + }, + { + "epoch": 1.213897937024973, + "grad_norm": 8.046753742934493, + "learning_rate": 1.7719146897386674e-05, + "loss": 1.1187, + "step": 2236 + }, + { + "epoch": 1.214440825190011, + "grad_norm": 6.625446441699582, + "learning_rate": 1.7716910673362123e-05, + "loss": 0.8731, + "step": 2237 + }, + { + "epoch": 1.214983713355049, + "grad_norm": 6.47539429072121, + "learning_rate": 1.7714673494897767e-05, + "loss": 0.9473, + "step": 2238 + }, + { + "epoch": 1.215526601520087, + "grad_norm": 7.458434209050833, + "learning_rate": 1.7712435362270292e-05, + "loss": 0.7844, + "step": 2239 + }, + { + "epoch": 1.216069489685125, + "grad_norm": 6.935847404573923, + "learning_rate": 1.7710196275756524e-05, + "loss": 0.8431, + "step": 2240 + }, + { + "epoch": 1.216612377850163, + "grad_norm": 8.015017144687471, + "learning_rate": 1.770795623563339e-05, + "loss": 1.2329, + "step": 2241 + }, + { + "epoch": 1.217155266015201, + "grad_norm": 6.914919733505818, + "learning_rate": 1.7705715242177944e-05, + "loss": 1.1335, + "step": 2242 + }, + { + "epoch": 1.217698154180239, + "grad_norm": 5.900061304870694, + "learning_rate": 1.7703473295667354e-05, + "loss": 0.7453, + "step": 2243 + }, + { + "epoch": 1.218241042345277, + "grad_norm": 4.959614765477869, + "learning_rate": 1.770123039637891e-05, + "loss": 0.6093, + "step": 2244 + }, + { + "epoch": 1.2187839305103148, + "grad_norm": 8.038465432918807, + "learning_rate": 1.769898654459002e-05, + "loss": 1.0029, + "step": 2245 + }, + { + "epoch": 1.219326818675353, + "grad_norm": 8.08635156220498, + "learning_rate": 1.7696741740578204e-05, + "loss": 1.0839, + "step": 2246 + }, + { + "epoch": 1.2198697068403908, + "grad_norm": 7.985335036181301, + "learning_rate": 1.7694495984621097e-05, + "loss": 1.262, + "step": 2247 + }, + { + "epoch": 1.220412595005429, + "grad_norm": 8.31037377117703, + "learning_rate": 1.769224927699647e-05, + "loss": 1.0265, + "step": 2248 + }, + { + "epoch": 1.2209554831704668, + "grad_norm": 6.241752286447289, + "learning_rate": 1.769000161798219e-05, + "loss": 0.6179, + "step": 2249 + }, + { + "epoch": 1.221498371335505, + "grad_norm": 6.767956890107702, + "learning_rate": 1.7687753007856253e-05, + "loss": 0.9897, + "step": 2250 + }, + { + "epoch": 1.2220412595005428, + "grad_norm": 7.119621854687504, + "learning_rate": 1.7685503446896772e-05, + "loss": 1.0529, + "step": 2251 + }, + { + "epoch": 1.222584147665581, + "grad_norm": 6.37786350129091, + "learning_rate": 1.7683252935381976e-05, + "loss": 0.8575, + "step": 2252 + }, + { + "epoch": 1.2231270358306188, + "grad_norm": 9.280258739154695, + "learning_rate": 1.768100147359021e-05, + "loss": 0.959, + "step": 2253 + }, + { + "epoch": 1.223669923995657, + "grad_norm": 7.172898793975323, + "learning_rate": 1.7678749061799942e-05, + "loss": 1.3745, + "step": 2254 + }, + { + "epoch": 1.2242128121606948, + "grad_norm": 5.3297508420409825, + "learning_rate": 1.7676495700289753e-05, + "loss": 0.6697, + "step": 2255 + }, + { + "epoch": 1.224755700325733, + "grad_norm": 6.518884308754971, + "learning_rate": 1.767424138933834e-05, + "loss": 0.9854, + "step": 2256 + }, + { + "epoch": 1.2252985884907708, + "grad_norm": 7.939498291049385, + "learning_rate": 1.767198612922452e-05, + "loss": 0.8918, + "step": 2257 + }, + { + "epoch": 1.225841476655809, + "grad_norm": 6.309602610605377, + "learning_rate": 1.7669729920227226e-05, + "loss": 0.9939, + "step": 2258 + }, + { + "epoch": 1.2263843648208468, + "grad_norm": 7.462197154770194, + "learning_rate": 1.7667472762625518e-05, + "loss": 1.1299, + "step": 2259 + }, + { + "epoch": 1.226927252985885, + "grad_norm": 6.700031938849371, + "learning_rate": 1.7665214656698555e-05, + "loss": 0.8655, + "step": 2260 + }, + { + "epoch": 1.2274701411509228, + "grad_norm": 6.72124660471967, + "learning_rate": 1.7662955602725627e-05, + "loss": 0.7383, + "step": 2261 + }, + { + "epoch": 1.228013029315961, + "grad_norm": 6.071348581517431, + "learning_rate": 1.7660695600986137e-05, + "loss": 0.8571, + "step": 2262 + }, + { + "epoch": 1.2285559174809988, + "grad_norm": 5.914184858865564, + "learning_rate": 1.7658434651759605e-05, + "loss": 0.7494, + "step": 2263 + }, + { + "epoch": 1.229098805646037, + "grad_norm": 12.685614201985125, + "learning_rate": 1.7656172755325675e-05, + "loss": 1.608, + "step": 2264 + }, + { + "epoch": 1.2296416938110748, + "grad_norm": 7.001237249554534, + "learning_rate": 1.765390991196409e-05, + "loss": 1.0437, + "step": 2265 + }, + { + "epoch": 1.230184581976113, + "grad_norm": 8.232094967642231, + "learning_rate": 1.765164612195473e-05, + "loss": 0.8574, + "step": 2266 + }, + { + "epoch": 1.2307274701411508, + "grad_norm": 6.367083948330669, + "learning_rate": 1.7649381385577582e-05, + "loss": 0.7106, + "step": 2267 + }, + { + "epoch": 1.231270358306189, + "grad_norm": 6.2062819181758195, + "learning_rate": 1.7647115703112756e-05, + "loss": 0.6838, + "step": 2268 + }, + { + "epoch": 1.2318132464712268, + "grad_norm": 6.896096048797596, + "learning_rate": 1.7644849074840472e-05, + "loss": 0.9595, + "step": 2269 + }, + { + "epoch": 1.232356134636265, + "grad_norm": 5.915145487902616, + "learning_rate": 1.7642581501041067e-05, + "loss": 0.5741, + "step": 2270 + }, + { + "epoch": 1.2328990228013028, + "grad_norm": 5.966871119524657, + "learning_rate": 1.7640312981995004e-05, + "loss": 0.8686, + "step": 2271 + }, + { + "epoch": 1.233441910966341, + "grad_norm": 6.835680775363237, + "learning_rate": 1.7638043517982854e-05, + "loss": 0.9532, + "step": 2272 + }, + { + "epoch": 1.2339847991313788, + "grad_norm": 8.228029492129801, + "learning_rate": 1.7635773109285306e-05, + "loss": 1.0834, + "step": 2273 + }, + { + "epoch": 1.234527687296417, + "grad_norm": 8.272927259623488, + "learning_rate": 1.7633501756183174e-05, + "loss": 1.2591, + "step": 2274 + }, + { + "epoch": 1.2350705754614548, + "grad_norm": 6.755748928056399, + "learning_rate": 1.7631229458957377e-05, + "loss": 0.997, + "step": 2275 + }, + { + "epoch": 1.235613463626493, + "grad_norm": 7.335027842642818, + "learning_rate": 1.762895621788896e-05, + "loss": 0.8695, + "step": 2276 + }, + { + "epoch": 1.2361563517915308, + "grad_norm": 5.237971036796901, + "learning_rate": 1.7626682033259077e-05, + "loss": 0.6846, + "step": 2277 + }, + { + "epoch": 1.236699239956569, + "grad_norm": 6.333455514649705, + "learning_rate": 1.7624406905349004e-05, + "loss": 0.7532, + "step": 2278 + }, + { + "epoch": 1.2372421281216068, + "grad_norm": 6.255556726847127, + "learning_rate": 1.7622130834440136e-05, + "loss": 0.7801, + "step": 2279 + }, + { + "epoch": 1.237785016286645, + "grad_norm": 5.602832960919912, + "learning_rate": 1.761985382081398e-05, + "loss": 0.5105, + "step": 2280 + }, + { + "epoch": 1.2383279044516828, + "grad_norm": 5.814619917758832, + "learning_rate": 1.7617575864752155e-05, + "loss": 0.9034, + "step": 2281 + }, + { + "epoch": 1.238870792616721, + "grad_norm": 6.743402665368447, + "learning_rate": 1.7615296966536407e-05, + "loss": 0.7713, + "step": 2282 + }, + { + "epoch": 1.2394136807817588, + "grad_norm": 7.176469994889228, + "learning_rate": 1.7613017126448597e-05, + "loss": 0.957, + "step": 2283 + }, + { + "epoch": 1.239956568946797, + "grad_norm": 8.5427155590311, + "learning_rate": 1.7610736344770693e-05, + "loss": 0.8789, + "step": 2284 + }, + { + "epoch": 1.2404994571118348, + "grad_norm": 6.674205629131359, + "learning_rate": 1.7608454621784786e-05, + "loss": 1.0239, + "step": 2285 + }, + { + "epoch": 1.241042345276873, + "grad_norm": 7.687894188099568, + "learning_rate": 1.7606171957773088e-05, + "loss": 1.3164, + "step": 2286 + }, + { + "epoch": 1.2415852334419109, + "grad_norm": 6.331613648212601, + "learning_rate": 1.760388835301792e-05, + "loss": 0.8047, + "step": 2287 + }, + { + "epoch": 1.242128121606949, + "grad_norm": 7.0776252458357725, + "learning_rate": 1.760160380780172e-05, + "loss": 1.1893, + "step": 2288 + }, + { + "epoch": 1.2426710097719869, + "grad_norm": 5.434346705496786, + "learning_rate": 1.7599318322407047e-05, + "loss": 0.763, + "step": 2289 + }, + { + "epoch": 1.243213897937025, + "grad_norm": 6.0271106768536775, + "learning_rate": 1.759703189711657e-05, + "loss": 1.1129, + "step": 2290 + }, + { + "epoch": 1.2437567861020629, + "grad_norm": 7.286292908201965, + "learning_rate": 1.7594744532213088e-05, + "loss": 1.1136, + "step": 2291 + }, + { + "epoch": 1.244299674267101, + "grad_norm": 6.685086643433707, + "learning_rate": 1.7592456227979488e-05, + "loss": 0.724, + "step": 2292 + }, + { + "epoch": 1.2448425624321389, + "grad_norm": 7.184174802374185, + "learning_rate": 1.7590166984698807e-05, + "loss": 1.1471, + "step": 2293 + }, + { + "epoch": 1.245385450597177, + "grad_norm": 7.787563610771557, + "learning_rate": 1.7587876802654176e-05, + "loss": 1.01, + "step": 2294 + }, + { + "epoch": 1.2459283387622149, + "grad_norm": 7.036440767647583, + "learning_rate": 1.7585585682128848e-05, + "loss": 0.8612, + "step": 2295 + }, + { + "epoch": 1.246471226927253, + "grad_norm": 6.6372154433448705, + "learning_rate": 1.7583293623406192e-05, + "loss": 1.3394, + "step": 2296 + }, + { + "epoch": 1.2470141150922909, + "grad_norm": 7.707986168491473, + "learning_rate": 1.7581000626769697e-05, + "loss": 0.8688, + "step": 2297 + }, + { + "epoch": 1.247557003257329, + "grad_norm": 7.862359368033938, + "learning_rate": 1.757870669250296e-05, + "loss": 1.1999, + "step": 2298 + }, + { + "epoch": 1.2480998914223669, + "grad_norm": 7.942280078814929, + "learning_rate": 1.75764118208897e-05, + "loss": 1.241, + "step": 2299 + }, + { + "epoch": 1.248642779587405, + "grad_norm": 7.258825765708901, + "learning_rate": 1.7574116012213753e-05, + "loss": 0.8336, + "step": 2300 + }, + { + "epoch": 1.2491856677524429, + "grad_norm": 7.0149158036864065, + "learning_rate": 1.7571819266759066e-05, + "loss": 0.8637, + "step": 2301 + }, + { + "epoch": 1.249728555917481, + "grad_norm": 7.069351546856077, + "learning_rate": 1.7569521584809703e-05, + "loss": 0.8146, + "step": 2302 + }, + { + "epoch": 1.250271444082519, + "grad_norm": 4.6424312647112425, + "learning_rate": 1.756722296664985e-05, + "loss": 0.5056, + "step": 2303 + }, + { + "epoch": 1.2508143322475571, + "grad_norm": 7.857412677197525, + "learning_rate": 1.7564923412563797e-05, + "loss": 1.3072, + "step": 2304 + }, + { + "epoch": 1.251357220412595, + "grad_norm": 6.081845076053873, + "learning_rate": 1.7562622922835962e-05, + "loss": 1.0212, + "step": 2305 + }, + { + "epoch": 1.2519001085776331, + "grad_norm": 5.244697868381976, + "learning_rate": 1.7560321497750867e-05, + "loss": 0.9677, + "step": 2306 + }, + { + "epoch": 1.252442996742671, + "grad_norm": 11.027915321544773, + "learning_rate": 1.7558019137593162e-05, + "loss": 0.8984, + "step": 2307 + }, + { + "epoch": 1.2529858849077091, + "grad_norm": 5.955816485468965, + "learning_rate": 1.7555715842647603e-05, + "loss": 0.6282, + "step": 2308 + }, + { + "epoch": 1.253528773072747, + "grad_norm": 8.925991341226974, + "learning_rate": 1.755341161319907e-05, + "loss": 1.2121, + "step": 2309 + }, + { + "epoch": 1.2540716612377851, + "grad_norm": 7.947406679428915, + "learning_rate": 1.7551106449532547e-05, + "loss": 1.0545, + "step": 2310 + }, + { + "epoch": 1.254614549402823, + "grad_norm": 7.329950391674132, + "learning_rate": 1.7548800351933147e-05, + "loss": 1.0585, + "step": 2311 + }, + { + "epoch": 1.2551574375678611, + "grad_norm": 5.0931347466751955, + "learning_rate": 1.7546493320686084e-05, + "loss": 0.5533, + "step": 2312 + }, + { + "epoch": 1.255700325732899, + "grad_norm": 7.125279667319334, + "learning_rate": 1.7544185356076703e-05, + "loss": 1.1011, + "step": 2313 + }, + { + "epoch": 1.2562432138979371, + "grad_norm": 8.060697991979964, + "learning_rate": 1.7541876458390453e-05, + "loss": 0.8732, + "step": 2314 + }, + { + "epoch": 1.256786102062975, + "grad_norm": 9.732797197135078, + "learning_rate": 1.75395666279129e-05, + "loss": 1.5663, + "step": 2315 + }, + { + "epoch": 1.2573289902280131, + "grad_norm": 5.3673655570709435, + "learning_rate": 1.7537255864929733e-05, + "loss": 0.7057, + "step": 2316 + }, + { + "epoch": 1.257871878393051, + "grad_norm": 5.997561584131186, + "learning_rate": 1.7534944169726752e-05, + "loss": 0.8526, + "step": 2317 + }, + { + "epoch": 1.2584147665580891, + "grad_norm": 7.236766116786611, + "learning_rate": 1.753263154258986e-05, + "loss": 0.8921, + "step": 2318 + }, + { + "epoch": 1.258957654723127, + "grad_norm": 7.663917187467161, + "learning_rate": 1.7530317983805096e-05, + "loss": 0.9309, + "step": 2319 + }, + { + "epoch": 1.2595005428881652, + "grad_norm": 5.704193869412544, + "learning_rate": 1.75280034936586e-05, + "loss": 0.6291, + "step": 2320 + }, + { + "epoch": 1.260043431053203, + "grad_norm": 7.365455665485428, + "learning_rate": 1.7525688072436636e-05, + "loss": 0.9946, + "step": 2321 + }, + { + "epoch": 1.2605863192182412, + "grad_norm": 4.8267670121648205, + "learning_rate": 1.752337172042558e-05, + "loss": 0.8254, + "step": 2322 + }, + { + "epoch": 1.261129207383279, + "grad_norm": 7.231116660749922, + "learning_rate": 1.752105443791192e-05, + "loss": 1.3835, + "step": 2323 + }, + { + "epoch": 1.2616720955483172, + "grad_norm": 6.02849395913157, + "learning_rate": 1.7518736225182253e-05, + "loss": 0.6229, + "step": 2324 + }, + { + "epoch": 1.262214983713355, + "grad_norm": 6.458823304206971, + "learning_rate": 1.7516417082523314e-05, + "loss": 0.7707, + "step": 2325 + }, + { + "epoch": 1.2627578718783932, + "grad_norm": 7.163598438660781, + "learning_rate": 1.7514097010221927e-05, + "loss": 1.0007, + "step": 2326 + }, + { + "epoch": 1.263300760043431, + "grad_norm": 6.607394766457111, + "learning_rate": 1.7511776008565043e-05, + "loss": 1.0289, + "step": 2327 + }, + { + "epoch": 1.2638436482084692, + "grad_norm": 8.20658895752514, + "learning_rate": 1.7509454077839736e-05, + "loss": 0.7553, + "step": 2328 + }, + { + "epoch": 1.264386536373507, + "grad_norm": 8.963409997263136, + "learning_rate": 1.750713121833318e-05, + "loss": 0.9558, + "step": 2329 + }, + { + "epoch": 1.2649294245385452, + "grad_norm": 6.463632425962391, + "learning_rate": 1.7504807430332668e-05, + "loss": 0.6553, + "step": 2330 + }, + { + "epoch": 1.265472312703583, + "grad_norm": 6.936173778617146, + "learning_rate": 1.7502482714125612e-05, + "loss": 0.7861, + "step": 2331 + }, + { + "epoch": 1.2660152008686212, + "grad_norm": 6.975529559460763, + "learning_rate": 1.7500157069999536e-05, + "loss": 0.665, + "step": 2332 + }, + { + "epoch": 1.266558089033659, + "grad_norm": 6.768480922885786, + "learning_rate": 1.7497830498242082e-05, + "loss": 1.0163, + "step": 2333 + }, + { + "epoch": 1.2671009771986972, + "grad_norm": 6.357421024100016, + "learning_rate": 1.7495502999141004e-05, + "loss": 0.5897, + "step": 2334 + }, + { + "epoch": 1.267643865363735, + "grad_norm": 7.82539792030223, + "learning_rate": 1.7493174572984168e-05, + "loss": 1.0509, + "step": 2335 + }, + { + "epoch": 1.2681867535287732, + "grad_norm": 7.6523935004894215, + "learning_rate": 1.7490845220059554e-05, + "loss": 0.7603, + "step": 2336 + }, + { + "epoch": 1.268729641693811, + "grad_norm": 6.677360283026087, + "learning_rate": 1.7488514940655267e-05, + "loss": 1.1797, + "step": 2337 + }, + { + "epoch": 1.2692725298588492, + "grad_norm": 7.837606738097467, + "learning_rate": 1.7486183735059517e-05, + "loss": 1.1995, + "step": 2338 + }, + { + "epoch": 1.269815418023887, + "grad_norm": 8.913928805014676, + "learning_rate": 1.7483851603560634e-05, + "loss": 1.6007, + "step": 2339 + }, + { + "epoch": 1.2703583061889252, + "grad_norm": 8.467363918664116, + "learning_rate": 1.7481518546447058e-05, + "loss": 0.9515, + "step": 2340 + }, + { + "epoch": 1.270901194353963, + "grad_norm": 6.758882762498143, + "learning_rate": 1.747918456400734e-05, + "loss": 0.7935, + "step": 2341 + }, + { + "epoch": 1.2714440825190012, + "grad_norm": 5.455764070482765, + "learning_rate": 1.7476849656530158e-05, + "loss": 0.4623, + "step": 2342 + }, + { + "epoch": 1.271986970684039, + "grad_norm": 8.417334291043444, + "learning_rate": 1.7474513824304294e-05, + "loss": 1.1563, + "step": 2343 + }, + { + "epoch": 1.2725298588490772, + "grad_norm": 5.61084844590966, + "learning_rate": 1.7472177067618646e-05, + "loss": 0.8566, + "step": 2344 + }, + { + "epoch": 1.273072747014115, + "grad_norm": 7.726478476188633, + "learning_rate": 1.746983938676223e-05, + "loss": 1.0568, + "step": 2345 + }, + { + "epoch": 1.2736156351791532, + "grad_norm": 8.882013629978706, + "learning_rate": 1.7467500782024177e-05, + "loss": 1.0921, + "step": 2346 + }, + { + "epoch": 1.274158523344191, + "grad_norm": 6.976574380776494, + "learning_rate": 1.746516125369372e-05, + "loss": 1.1222, + "step": 2347 + }, + { + "epoch": 1.2747014115092292, + "grad_norm": 7.641347465211616, + "learning_rate": 1.7462820802060224e-05, + "loss": 1.0672, + "step": 2348 + }, + { + "epoch": 1.275244299674267, + "grad_norm": 6.473243122755322, + "learning_rate": 1.7460479427413156e-05, + "loss": 0.9019, + "step": 2349 + }, + { + "epoch": 1.2757871878393052, + "grad_norm": 6.494958002593054, + "learning_rate": 1.7458137130042105e-05, + "loss": 0.7862, + "step": 2350 + }, + { + "epoch": 1.276330076004343, + "grad_norm": 6.54459648554074, + "learning_rate": 1.7455793910236764e-05, + "loss": 0.9572, + "step": 2351 + }, + { + "epoch": 1.2768729641693812, + "grad_norm": 6.249218254368162, + "learning_rate": 1.7453449768286952e-05, + "loss": 0.9415, + "step": 2352 + }, + { + "epoch": 1.277415852334419, + "grad_norm": 6.335956299458767, + "learning_rate": 1.7451104704482592e-05, + "loss": 0.7647, + "step": 2353 + }, + { + "epoch": 1.2779587404994572, + "grad_norm": 6.296362619304004, + "learning_rate": 1.744875871911373e-05, + "loss": 0.6293, + "step": 2354 + }, + { + "epoch": 1.278501628664495, + "grad_norm": 6.629983181103369, + "learning_rate": 1.7446411812470516e-05, + "loss": 0.7816, + "step": 2355 + }, + { + "epoch": 1.2790445168295332, + "grad_norm": 6.921460108918273, + "learning_rate": 1.744406398484322e-05, + "loss": 0.6373, + "step": 2356 + }, + { + "epoch": 1.279587404994571, + "grad_norm": 9.4421099603369, + "learning_rate": 1.7441715236522228e-05, + "loss": 1.5291, + "step": 2357 + }, + { + "epoch": 1.2801302931596092, + "grad_norm": 7.838644313955947, + "learning_rate": 1.7439365567798035e-05, + "loss": 0.7934, + "step": 2358 + }, + { + "epoch": 1.280673181324647, + "grad_norm": 5.7659414273416, + "learning_rate": 1.743701497896125e-05, + "loss": 0.8358, + "step": 2359 + }, + { + "epoch": 1.2812160694896852, + "grad_norm": 4.731198780322981, + "learning_rate": 1.7434663470302602e-05, + "loss": 0.7357, + "step": 2360 + }, + { + "epoch": 1.281758957654723, + "grad_norm": 8.993227122901354, + "learning_rate": 1.7432311042112926e-05, + "loss": 1.0679, + "step": 2361 + }, + { + "epoch": 1.2823018458197613, + "grad_norm": 7.193753575449645, + "learning_rate": 1.7429957694683175e-05, + "loss": 0.9257, + "step": 2362 + }, + { + "epoch": 1.282844733984799, + "grad_norm": 7.1027186942584635, + "learning_rate": 1.7427603428304416e-05, + "loss": 0.64, + "step": 2363 + }, + { + "epoch": 1.2833876221498373, + "grad_norm": 9.441739719572269, + "learning_rate": 1.7425248243267824e-05, + "loss": 1.3236, + "step": 2364 + }, + { + "epoch": 1.283930510314875, + "grad_norm": 6.478965109994284, + "learning_rate": 1.7422892139864696e-05, + "loss": 0.7726, + "step": 2365 + }, + { + "epoch": 1.2844733984799133, + "grad_norm": 6.372951649068815, + "learning_rate": 1.742053511838644e-05, + "loss": 0.7785, + "step": 2366 + }, + { + "epoch": 1.285016286644951, + "grad_norm": 10.052356612127076, + "learning_rate": 1.7418177179124574e-05, + "loss": 1.7877, + "step": 2367 + }, + { + "epoch": 1.2855591748099893, + "grad_norm": 5.24633801429042, + "learning_rate": 1.741581832237073e-05, + "loss": 0.4195, + "step": 2368 + }, + { + "epoch": 1.286102062975027, + "grad_norm": 7.958362658884473, + "learning_rate": 1.7413458548416656e-05, + "loss": 1.1782, + "step": 2369 + }, + { + "epoch": 1.2866449511400653, + "grad_norm": 7.923031881378813, + "learning_rate": 1.7411097857554216e-05, + "loss": 1.0861, + "step": 2370 + }, + { + "epoch": 1.287187839305103, + "grad_norm": 6.53099588563486, + "learning_rate": 1.7408736250075378e-05, + "loss": 0.8717, + "step": 2371 + }, + { + "epoch": 1.2877307274701413, + "grad_norm": 7.364911313605863, + "learning_rate": 1.740637372627224e-05, + "loss": 0.9341, + "step": 2372 + }, + { + "epoch": 1.288273615635179, + "grad_norm": 8.083622859612968, + "learning_rate": 1.740401028643699e-05, + "loss": 1.1654, + "step": 2373 + }, + { + "epoch": 1.2888165038002173, + "grad_norm": 6.561315990238234, + "learning_rate": 1.7401645930861944e-05, + "loss": 0.8088, + "step": 2374 + }, + { + "epoch": 1.289359391965255, + "grad_norm": 6.926193447339149, + "learning_rate": 1.739928065983954e-05, + "loss": 0.8335, + "step": 2375 + }, + { + "epoch": 1.2899022801302933, + "grad_norm": 7.876730774519387, + "learning_rate": 1.7396914473662306e-05, + "loss": 0.9969, + "step": 2376 + }, + { + "epoch": 1.290445168295331, + "grad_norm": 6.513645878383447, + "learning_rate": 1.73945473726229e-05, + "loss": 0.9302, + "step": 2377 + }, + { + "epoch": 1.2909880564603693, + "grad_norm": 6.441562626276137, + "learning_rate": 1.73921793570141e-05, + "loss": 1.0318, + "step": 2378 + }, + { + "epoch": 1.291530944625407, + "grad_norm": 5.226443422910506, + "learning_rate": 1.7389810427128766e-05, + "loss": 0.6142, + "step": 2379 + }, + { + "epoch": 1.2920738327904453, + "grad_norm": 7.436131751002477, + "learning_rate": 1.7387440583259906e-05, + "loss": 0.9907, + "step": 2380 + }, + { + "epoch": 1.292616720955483, + "grad_norm": 6.335395136223753, + "learning_rate": 1.7385069825700615e-05, + "loss": 0.8431, + "step": 2381 + }, + { + "epoch": 1.2931596091205213, + "grad_norm": 8.93013417703781, + "learning_rate": 1.738269815474412e-05, + "loss": 1.3162, + "step": 2382 + }, + { + "epoch": 1.293702497285559, + "grad_norm": 6.616000665685319, + "learning_rate": 1.7380325570683754e-05, + "loss": 0.81, + "step": 2383 + }, + { + "epoch": 1.2942453854505973, + "grad_norm": 9.126678309827314, + "learning_rate": 1.737795207381296e-05, + "loss": 1.4976, + "step": 2384 + }, + { + "epoch": 1.294788273615635, + "grad_norm": 9.726273416685123, + "learning_rate": 1.737557766442529e-05, + "loss": 1.0284, + "step": 2385 + }, + { + "epoch": 1.2953311617806733, + "grad_norm": 6.207571191669611, + "learning_rate": 1.737320234281442e-05, + "loss": 0.9409, + "step": 2386 + }, + { + "epoch": 1.295874049945711, + "grad_norm": 6.150632184961992, + "learning_rate": 1.7370826109274136e-05, + "loss": 0.7558, + "step": 2387 + }, + { + "epoch": 1.2964169381107493, + "grad_norm": 6.555861715157727, + "learning_rate": 1.736844896409833e-05, + "loss": 1.4217, + "step": 2388 + }, + { + "epoch": 1.296959826275787, + "grad_norm": 7.291607947091272, + "learning_rate": 1.7366070907581015e-05, + "loss": 0.9969, + "step": 2389 + }, + { + "epoch": 1.2975027144408253, + "grad_norm": 7.003873752430508, + "learning_rate": 1.7363691940016307e-05, + "loss": 0.9569, + "step": 2390 + }, + { + "epoch": 1.298045602605863, + "grad_norm": 6.355975398274633, + "learning_rate": 1.7361312061698444e-05, + "loss": 0.9244, + "step": 2391 + }, + { + "epoch": 1.2985884907709013, + "grad_norm": 5.093818720596728, + "learning_rate": 1.7358931272921773e-05, + "loss": 0.5581, + "step": 2392 + }, + { + "epoch": 1.299131378935939, + "grad_norm": 4.532047411652702, + "learning_rate": 1.7356549573980753e-05, + "loss": 0.5153, + "step": 2393 + }, + { + "epoch": 1.2996742671009773, + "grad_norm": 7.378796483754393, + "learning_rate": 1.735416696516996e-05, + "loss": 0.9631, + "step": 2394 + }, + { + "epoch": 1.3002171552660151, + "grad_norm": 6.186232096125507, + "learning_rate": 1.7351783446784075e-05, + "loss": 0.7876, + "step": 2395 + }, + { + "epoch": 1.3007600434310533, + "grad_norm": 7.851387526494335, + "learning_rate": 1.7349399019117897e-05, + "loss": 0.8983, + "step": 2396 + }, + { + "epoch": 1.3013029315960911, + "grad_norm": 7.275506442499215, + "learning_rate": 1.7347013682466335e-05, + "loss": 0.8334, + "step": 2397 + }, + { + "epoch": 1.3018458197611293, + "grad_norm": 6.264072188984264, + "learning_rate": 1.7344627437124407e-05, + "loss": 0.6585, + "step": 2398 + }, + { + "epoch": 1.3023887079261671, + "grad_norm": 7.537073303434296, + "learning_rate": 1.734224028338726e-05, + "loss": 1.1033, + "step": 2399 + }, + { + "epoch": 1.3029315960912053, + "grad_norm": 6.850281208830108, + "learning_rate": 1.7339852221550126e-05, + "loss": 0.8133, + "step": 2400 + }, + { + "epoch": 1.3034744842562431, + "grad_norm": 8.089656198012262, + "learning_rate": 1.7337463251908374e-05, + "loss": 1.1342, + "step": 2401 + }, + { + "epoch": 1.3040173724212814, + "grad_norm": 6.2919884156176105, + "learning_rate": 1.733507337475747e-05, + "loss": 0.6533, + "step": 2402 + }, + { + "epoch": 1.3045602605863191, + "grad_norm": 5.114953151143873, + "learning_rate": 1.7332682590393008e-05, + "loss": 0.7451, + "step": 2403 + }, + { + "epoch": 1.3051031487513574, + "grad_norm": 6.18470588828327, + "learning_rate": 1.733029089911067e-05, + "loss": 1.1299, + "step": 2404 + }, + { + "epoch": 1.3056460369163951, + "grad_norm": 5.950129534147113, + "learning_rate": 1.7327898301206273e-05, + "loss": 0.7329, + "step": 2405 + }, + { + "epoch": 1.3061889250814334, + "grad_norm": 6.293661255114301, + "learning_rate": 1.7325504796975732e-05, + "loss": 0.7535, + "step": 2406 + }, + { + "epoch": 1.3067318132464711, + "grad_norm": 7.682262802579054, + "learning_rate": 1.732311038671509e-05, + "loss": 1.2399, + "step": 2407 + }, + { + "epoch": 1.3072747014115094, + "grad_norm": 7.314299514485556, + "learning_rate": 1.7320715070720478e-05, + "loss": 0.9298, + "step": 2408 + }, + { + "epoch": 1.3078175895765471, + "grad_norm": 8.921126356976833, + "learning_rate": 1.7318318849288158e-05, + "loss": 1.4254, + "step": 2409 + }, + { + "epoch": 1.3083604777415854, + "grad_norm": 7.060288459365394, + "learning_rate": 1.7315921722714503e-05, + "loss": 0.9428, + "step": 2410 + }, + { + "epoch": 1.3089033659066232, + "grad_norm": 8.568251970472561, + "learning_rate": 1.7313523691295988e-05, + "loss": 0.9011, + "step": 2411 + }, + { + "epoch": 1.3094462540716614, + "grad_norm": 6.049363733159396, + "learning_rate": 1.7311124755329206e-05, + "loss": 1.0718, + "step": 2412 + }, + { + "epoch": 1.3099891422366992, + "grad_norm": 8.287809242433424, + "learning_rate": 1.7308724915110864e-05, + "loss": 0.9137, + "step": 2413 + }, + { + "epoch": 1.3105320304017374, + "grad_norm": 8.436233795197621, + "learning_rate": 1.7306324170937774e-05, + "loss": 1.2427, + "step": 2414 + }, + { + "epoch": 1.3110749185667752, + "grad_norm": 7.035710404708695, + "learning_rate": 1.7303922523106863e-05, + "loss": 0.8206, + "step": 2415 + }, + { + "epoch": 1.3116178067318134, + "grad_norm": 5.9728457186472275, + "learning_rate": 1.730151997191518e-05, + "loss": 0.7307, + "step": 2416 + }, + { + "epoch": 1.3121606948968512, + "grad_norm": 6.296494925379207, + "learning_rate": 1.7299116517659865e-05, + "loss": 0.9808, + "step": 2417 + }, + { + "epoch": 1.3127035830618892, + "grad_norm": 7.672949954931739, + "learning_rate": 1.7296712160638192e-05, + "loss": 1.0289, + "step": 2418 + }, + { + "epoch": 1.3132464712269272, + "grad_norm": 5.958844838187463, + "learning_rate": 1.7294306901147525e-05, + "loss": 0.9657, + "step": 2419 + }, + { + "epoch": 1.3137893593919652, + "grad_norm": 6.5874736925476745, + "learning_rate": 1.7291900739485356e-05, + "loss": 0.8701, + "step": 2420 + }, + { + "epoch": 1.3143322475570032, + "grad_norm": 8.273467490572564, + "learning_rate": 1.7289493675949282e-05, + "loss": 0.9353, + "step": 2421 + }, + { + "epoch": 1.3148751357220412, + "grad_norm": 8.115143215098843, + "learning_rate": 1.7287085710837013e-05, + "loss": 0.6241, + "step": 2422 + }, + { + "epoch": 1.3154180238870792, + "grad_norm": 7.004703854660018, + "learning_rate": 1.7284676844446368e-05, + "loss": 0.801, + "step": 2423 + }, + { + "epoch": 1.3159609120521172, + "grad_norm": 6.039976470078682, + "learning_rate": 1.728226707707528e-05, + "loss": 1.085, + "step": 2424 + }, + { + "epoch": 1.3165038002171552, + "grad_norm": 5.318299059101813, + "learning_rate": 1.72798564090218e-05, + "loss": 0.4945, + "step": 2425 + }, + { + "epoch": 1.3170466883821932, + "grad_norm": 6.095578475774351, + "learning_rate": 1.727744484058407e-05, + "loss": 0.7696, + "step": 2426 + }, + { + "epoch": 1.3175895765472312, + "grad_norm": 6.4886080927006695, + "learning_rate": 1.7275032372060368e-05, + "loss": 0.5364, + "step": 2427 + }, + { + "epoch": 1.3181324647122692, + "grad_norm": 8.681669485105683, + "learning_rate": 1.7272619003749066e-05, + "loss": 0.9132, + "step": 2428 + }, + { + "epoch": 1.3186753528773072, + "grad_norm": 6.819144122945338, + "learning_rate": 1.7270204735948653e-05, + "loss": 0.7867, + "step": 2429 + }, + { + "epoch": 1.3192182410423452, + "grad_norm": 6.65338033222052, + "learning_rate": 1.7267789568957734e-05, + "loss": 0.7236, + "step": 2430 + }, + { + "epoch": 1.3197611292073832, + "grad_norm": 6.622862530496735, + "learning_rate": 1.7265373503075014e-05, + "loss": 0.9038, + "step": 2431 + }, + { + "epoch": 1.3203040173724212, + "grad_norm": 6.284216416307323, + "learning_rate": 1.7262956538599323e-05, + "loss": 0.677, + "step": 2432 + }, + { + "epoch": 1.3208469055374592, + "grad_norm": 9.150803536423062, + "learning_rate": 1.7260538675829593e-05, + "loss": 1.7953, + "step": 2433 + }, + { + "epoch": 1.3213897937024972, + "grad_norm": 9.236250071710021, + "learning_rate": 1.7258119915064867e-05, + "loss": 0.8568, + "step": 2434 + }, + { + "epoch": 1.3219326818675352, + "grad_norm": 7.061272817814159, + "learning_rate": 1.72557002566043e-05, + "loss": 0.9355, + "step": 2435 + }, + { + "epoch": 1.3224755700325732, + "grad_norm": 6.786187626307571, + "learning_rate": 1.7253279700747164e-05, + "loss": 0.7707, + "step": 2436 + }, + { + "epoch": 1.3230184581976112, + "grad_norm": 5.592273858747665, + "learning_rate": 1.725085824779283e-05, + "loss": 0.5932, + "step": 2437 + }, + { + "epoch": 1.3235613463626492, + "grad_norm": 8.137496643840409, + "learning_rate": 1.72484358980408e-05, + "loss": 1.0244, + "step": 2438 + }, + { + "epoch": 1.3241042345276872, + "grad_norm": 5.981122138792478, + "learning_rate": 1.724601265179066e-05, + "loss": 0.9807, + "step": 2439 + }, + { + "epoch": 1.3246471226927252, + "grad_norm": 6.912027532729551, + "learning_rate": 1.7243588509342127e-05, + "loss": 0.8673, + "step": 2440 + }, + { + "epoch": 1.3251900108577632, + "grad_norm": 5.836398731450734, + "learning_rate": 1.7241163470995024e-05, + "loss": 0.5738, + "step": 2441 + }, + { + "epoch": 1.3257328990228012, + "grad_norm": 5.0689326588185795, + "learning_rate": 1.7238737537049283e-05, + "loss": 0.4449, + "step": 2442 + }, + { + "epoch": 1.3262757871878392, + "grad_norm": 6.628337387569613, + "learning_rate": 1.7236310707804943e-05, + "loss": 0.7975, + "step": 2443 + }, + { + "epoch": 1.3268186753528772, + "grad_norm": 9.359654484357105, + "learning_rate": 1.7233882983562168e-05, + "loss": 0.8996, + "step": 2444 + }, + { + "epoch": 1.3273615635179152, + "grad_norm": 5.979606493030731, + "learning_rate": 1.723145436462121e-05, + "loss": 0.732, + "step": 2445 + }, + { + "epoch": 1.3279044516829532, + "grad_norm": 6.393853922089353, + "learning_rate": 1.7229024851282453e-05, + "loss": 0.9713, + "step": 2446 + }, + { + "epoch": 1.3284473398479912, + "grad_norm": 9.783386290725945, + "learning_rate": 1.722659444384638e-05, + "loss": 1.2991, + "step": 2447 + }, + { + "epoch": 1.3289902280130292, + "grad_norm": 8.471557152983001, + "learning_rate": 1.722416314261359e-05, + "loss": 0.8517, + "step": 2448 + }, + { + "epoch": 1.3295331161780672, + "grad_norm": 7.20996107804807, + "learning_rate": 1.7221730947884793e-05, + "loss": 1.0892, + "step": 2449 + }, + { + "epoch": 1.3300760043431052, + "grad_norm": 9.97079093211024, + "learning_rate": 1.7219297859960796e-05, + "loss": 0.8866, + "step": 2450 + }, + { + "epoch": 1.3306188925081432, + "grad_norm": 5.88899239355285, + "learning_rate": 1.7216863879142536e-05, + "loss": 0.7286, + "step": 2451 + }, + { + "epoch": 1.3311617806731812, + "grad_norm": 8.317968838525326, + "learning_rate": 1.7214429005731054e-05, + "loss": 0.9374, + "step": 2452 + }, + { + "epoch": 1.3317046688382193, + "grad_norm": 8.278980033154179, + "learning_rate": 1.721199324002749e-05, + "loss": 1.2977, + "step": 2453 + }, + { + "epoch": 1.3322475570032573, + "grad_norm": 7.50073001311901, + "learning_rate": 1.7209556582333106e-05, + "loss": 0.9479, + "step": 2454 + }, + { + "epoch": 1.3327904451682953, + "grad_norm": 9.34140520001984, + "learning_rate": 1.720711903294928e-05, + "loss": 1.1834, + "step": 2455 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 6.155163014555799, + "learning_rate": 1.720468059217748e-05, + "loss": 1.0674, + "step": 2456 + }, + { + "epoch": 1.3338762214983713, + "grad_norm": 8.207152122618293, + "learning_rate": 1.7202241260319305e-05, + "loss": 1.1623, + "step": 2457 + }, + { + "epoch": 1.3344191096634093, + "grad_norm": 8.477127177193351, + "learning_rate": 1.719980103767645e-05, + "loss": 1.0812, + "step": 2458 + }, + { + "epoch": 1.3349619978284473, + "grad_norm": 9.352009366934986, + "learning_rate": 1.7197359924550726e-05, + "loss": 0.9593, + "step": 2459 + }, + { + "epoch": 1.3355048859934853, + "grad_norm": 6.377220189803284, + "learning_rate": 1.719491792124406e-05, + "loss": 0.8569, + "step": 2460 + }, + { + "epoch": 1.3360477741585233, + "grad_norm": 10.372722210417326, + "learning_rate": 1.7192475028058475e-05, + "loss": 0.8778, + "step": 2461 + }, + { + "epoch": 1.3365906623235613, + "grad_norm": 6.943692981514271, + "learning_rate": 1.7190031245296118e-05, + "loss": 0.8143, + "step": 2462 + }, + { + "epoch": 1.3371335504885993, + "grad_norm": 8.53588364120296, + "learning_rate": 1.7187586573259237e-05, + "loss": 1.2608, + "step": 2463 + }, + { + "epoch": 1.3376764386536373, + "grad_norm": 6.85497949710551, + "learning_rate": 1.7185141012250195e-05, + "loss": 0.621, + "step": 2464 + }, + { + "epoch": 1.3382193268186753, + "grad_norm": 9.575890345813113, + "learning_rate": 1.7182694562571458e-05, + "loss": 1.5686, + "step": 2465 + }, + { + "epoch": 1.3387622149837133, + "grad_norm": 8.637844905126164, + "learning_rate": 1.718024722452561e-05, + "loss": 0.8287, + "step": 2466 + }, + { + "epoch": 1.3393051031487513, + "grad_norm": 7.464168837468054, + "learning_rate": 1.7177798998415344e-05, + "loss": 0.9837, + "step": 2467 + }, + { + "epoch": 1.3398479913137893, + "grad_norm": 9.27927006463804, + "learning_rate": 1.7175349884543458e-05, + "loss": 1.3479, + "step": 2468 + }, + { + "epoch": 1.3403908794788273, + "grad_norm": 6.9150180643503765, + "learning_rate": 1.717289988321286e-05, + "loss": 0.7731, + "step": 2469 + }, + { + "epoch": 1.3409337676438653, + "grad_norm": 8.264286643294954, + "learning_rate": 1.7170448994726574e-05, + "loss": 1.1252, + "step": 2470 + }, + { + "epoch": 1.3414766558089033, + "grad_norm": 6.1769106952218715, + "learning_rate": 1.7167997219387728e-05, + "loss": 0.6077, + "step": 2471 + }, + { + "epoch": 1.3420195439739413, + "grad_norm": 6.00046554758586, + "learning_rate": 1.7165544557499563e-05, + "loss": 0.9807, + "step": 2472 + }, + { + "epoch": 1.3425624321389793, + "grad_norm": 5.684026968717126, + "learning_rate": 1.7163091009365427e-05, + "loss": 0.8755, + "step": 2473 + }, + { + "epoch": 1.3431053203040173, + "grad_norm": 8.910694398982464, + "learning_rate": 1.7160636575288777e-05, + "loss": 1.1781, + "step": 2474 + }, + { + "epoch": 1.3436482084690553, + "grad_norm": 6.090059050660724, + "learning_rate": 1.7158181255573184e-05, + "loss": 0.8757, + "step": 2475 + }, + { + "epoch": 1.3441910966340933, + "grad_norm": 5.3693699493967575, + "learning_rate": 1.7155725050522325e-05, + "loss": 0.4111, + "step": 2476 + }, + { + "epoch": 1.3447339847991313, + "grad_norm": 6.4883376045754675, + "learning_rate": 1.7153267960439987e-05, + "loss": 1.1667, + "step": 2477 + }, + { + "epoch": 1.3452768729641693, + "grad_norm": 6.9649527868325585, + "learning_rate": 1.7150809985630065e-05, + "loss": 1.2642, + "step": 2478 + }, + { + "epoch": 1.3458197611292073, + "grad_norm": 7.458967124353787, + "learning_rate": 1.714835112639657e-05, + "loss": 1.1198, + "step": 2479 + }, + { + "epoch": 1.3463626492942453, + "grad_norm": 7.4820540779231015, + "learning_rate": 1.7145891383043613e-05, + "loss": 1.069, + "step": 2480 + }, + { + "epoch": 1.3469055374592833, + "grad_norm": 6.8442083181868165, + "learning_rate": 1.7143430755875422e-05, + "loss": 1.1187, + "step": 2481 + }, + { + "epoch": 1.3474484256243213, + "grad_norm": 8.389465670773259, + "learning_rate": 1.7140969245196332e-05, + "loss": 1.216, + "step": 2482 + }, + { + "epoch": 1.3479913137893593, + "grad_norm": 8.099481746156716, + "learning_rate": 1.713850685131078e-05, + "loss": 1.1365, + "step": 2483 + }, + { + "epoch": 1.3485342019543973, + "grad_norm": 7.005684437567773, + "learning_rate": 1.713604357452333e-05, + "loss": 0.7665, + "step": 2484 + }, + { + "epoch": 1.3490770901194353, + "grad_norm": 6.032904961736961, + "learning_rate": 1.7133579415138634e-05, + "loss": 0.7527, + "step": 2485 + }, + { + "epoch": 1.3496199782844733, + "grad_norm": 7.644359802639278, + "learning_rate": 1.7131114373461466e-05, + "loss": 0.7293, + "step": 2486 + }, + { + "epoch": 1.3501628664495113, + "grad_norm": 9.681466854129468, + "learning_rate": 1.7128648449796706e-05, + "loss": 1.0718, + "step": 2487 + }, + { + "epoch": 1.3507057546145493, + "grad_norm": 7.703820713192501, + "learning_rate": 1.7126181644449348e-05, + "loss": 1.0049, + "step": 2488 + }, + { + "epoch": 1.3512486427795873, + "grad_norm": 8.313943023532504, + "learning_rate": 1.7123713957724482e-05, + "loss": 1.1647, + "step": 2489 + }, + { + "epoch": 1.3517915309446253, + "grad_norm": 6.869318843676937, + "learning_rate": 1.7121245389927326e-05, + "loss": 0.7056, + "step": 2490 + }, + { + "epoch": 1.3523344191096633, + "grad_norm": 8.354366894281007, + "learning_rate": 1.7118775941363186e-05, + "loss": 1.4107, + "step": 2491 + }, + { + "epoch": 1.3528773072747013, + "grad_norm": 7.594996874862703, + "learning_rate": 1.7116305612337493e-05, + "loss": 1.3947, + "step": 2492 + }, + { + "epoch": 1.3534201954397393, + "grad_norm": 5.912640331169303, + "learning_rate": 1.7113834403155782e-05, + "loss": 0.678, + "step": 2493 + }, + { + "epoch": 1.3539630836047774, + "grad_norm": 8.603089873660597, + "learning_rate": 1.7111362314123693e-05, + "loss": 1.1846, + "step": 2494 + }, + { + "epoch": 1.3545059717698154, + "grad_norm": 5.899587064063632, + "learning_rate": 1.710888934554698e-05, + "loss": 0.7241, + "step": 2495 + }, + { + "epoch": 1.3550488599348534, + "grad_norm": 5.1366064982711315, + "learning_rate": 1.7106415497731502e-05, + "loss": 0.6714, + "step": 2496 + }, + { + "epoch": 1.3555917480998914, + "grad_norm": 6.55484035818926, + "learning_rate": 1.710394077098323e-05, + "loss": 1.0163, + "step": 2497 + }, + { + "epoch": 1.3561346362649294, + "grad_norm": 7.060748735507699, + "learning_rate": 1.710146516560824e-05, + "loss": 0.8407, + "step": 2498 + }, + { + "epoch": 1.3566775244299674, + "grad_norm": 7.488417378829399, + "learning_rate": 1.709898868191272e-05, + "loss": 1.0936, + "step": 2499 + }, + { + "epoch": 1.3572204125950054, + "grad_norm": 5.616560879457259, + "learning_rate": 1.7096511320202965e-05, + "loss": 0.6962, + "step": 2500 + }, + { + "epoch": 1.3577633007600434, + "grad_norm": 6.5457002772857775, + "learning_rate": 1.7094033080785384e-05, + "loss": 0.6338, + "step": 2501 + }, + { + "epoch": 1.3583061889250814, + "grad_norm": 5.788382651374094, + "learning_rate": 1.709155396396648e-05, + "loss": 0.9251, + "step": 2502 + }, + { + "epoch": 1.3588490770901194, + "grad_norm": 5.94278314769549, + "learning_rate": 1.7089073970052883e-05, + "loss": 0.5738, + "step": 2503 + }, + { + "epoch": 1.3593919652551574, + "grad_norm": 8.987615230603502, + "learning_rate": 1.7086593099351318e-05, + "loss": 1.1561, + "step": 2504 + }, + { + "epoch": 1.3599348534201954, + "grad_norm": 6.25509019167934, + "learning_rate": 1.7084111352168627e-05, + "loss": 0.5584, + "step": 2505 + }, + { + "epoch": 1.3604777415852334, + "grad_norm": 6.42626784615777, + "learning_rate": 1.708162872881175e-05, + "loss": 0.6385, + "step": 2506 + }, + { + "epoch": 1.3610206297502714, + "grad_norm": 7.342680004921898, + "learning_rate": 1.707914522958775e-05, + "loss": 1.0729, + "step": 2507 + }, + { + "epoch": 1.3615635179153094, + "grad_norm": 6.858864236878158, + "learning_rate": 1.707666085480378e-05, + "loss": 0.8641, + "step": 2508 + }, + { + "epoch": 1.3621064060803474, + "grad_norm": 6.176457712037886, + "learning_rate": 1.707417560476712e-05, + "loss": 0.9472, + "step": 2509 + }, + { + "epoch": 1.3626492942453854, + "grad_norm": 7.346954574146698, + "learning_rate": 1.7071689479785145e-05, + "loss": 1.0253, + "step": 2510 + }, + { + "epoch": 1.3631921824104234, + "grad_norm": 4.9331066104603085, + "learning_rate": 1.7069202480165344e-05, + "loss": 0.7307, + "step": 2511 + }, + { + "epoch": 1.3637350705754614, + "grad_norm": 8.316711829847868, + "learning_rate": 1.7066714606215316e-05, + "loss": 0.8078, + "step": 2512 + }, + { + "epoch": 1.3642779587404994, + "grad_norm": 8.079444940270628, + "learning_rate": 1.706422585824276e-05, + "loss": 1.2922, + "step": 2513 + }, + { + "epoch": 1.3648208469055374, + "grad_norm": 7.503741374806181, + "learning_rate": 1.7061736236555494e-05, + "loss": 0.7943, + "step": 2514 + }, + { + "epoch": 1.3653637350705754, + "grad_norm": 6.4023293864360635, + "learning_rate": 1.7059245741461435e-05, + "loss": 1.0111, + "step": 2515 + }, + { + "epoch": 1.3659066232356134, + "grad_norm": 5.7325920409358595, + "learning_rate": 1.705675437326861e-05, + "loss": 1.1201, + "step": 2516 + }, + { + "epoch": 1.3664495114006514, + "grad_norm": 8.452660578831315, + "learning_rate": 1.705426213228516e-05, + "loss": 0.8818, + "step": 2517 + }, + { + "epoch": 1.3669923995656894, + "grad_norm": 7.8301031574049755, + "learning_rate": 1.7051769018819327e-05, + "loss": 0.8803, + "step": 2518 + }, + { + "epoch": 1.3675352877307274, + "grad_norm": 10.77827779179992, + "learning_rate": 1.704927503317946e-05, + "loss": 1.6667, + "step": 2519 + }, + { + "epoch": 1.3680781758957654, + "grad_norm": 4.579120603442138, + "learning_rate": 1.7046780175674023e-05, + "loss": 0.6101, + "step": 2520 + }, + { + "epoch": 1.3686210640608034, + "grad_norm": 7.443733277623912, + "learning_rate": 1.7044284446611584e-05, + "loss": 0.8174, + "step": 2521 + }, + { + "epoch": 1.3691639522258414, + "grad_norm": 10.860855146774309, + "learning_rate": 1.7041787846300817e-05, + "loss": 1.2849, + "step": 2522 + }, + { + "epoch": 1.3697068403908794, + "grad_norm": 9.311283837684355, + "learning_rate": 1.7039290375050503e-05, + "loss": 1.4923, + "step": 2523 + }, + { + "epoch": 1.3702497285559174, + "grad_norm": 7.171226901795553, + "learning_rate": 1.703679203316954e-05, + "loss": 0.9493, + "step": 2524 + }, + { + "epoch": 1.3707926167209554, + "grad_norm": 8.658053683953712, + "learning_rate": 1.7034292820966923e-05, + "loss": 1.1366, + "step": 2525 + }, + { + "epoch": 1.3713355048859934, + "grad_norm": 6.400723985129287, + "learning_rate": 1.703179273875176e-05, + "loss": 1.0484, + "step": 2526 + }, + { + "epoch": 1.3718783930510314, + "grad_norm": 6.2215780206112274, + "learning_rate": 1.702929178683326e-05, + "loss": 1.1058, + "step": 2527 + }, + { + "epoch": 1.3724212812160694, + "grad_norm": 7.548758658264992, + "learning_rate": 1.7026789965520755e-05, + "loss": 1.094, + "step": 2528 + }, + { + "epoch": 1.3729641693811074, + "grad_norm": 7.250992484079327, + "learning_rate": 1.7024287275123664e-05, + "loss": 0.8424, + "step": 2529 + }, + { + "epoch": 1.3735070575461454, + "grad_norm": 6.664109303085317, + "learning_rate": 1.702178371595153e-05, + "loss": 1.0498, + "step": 2530 + }, + { + "epoch": 1.3740499457111834, + "grad_norm": 6.379180216401721, + "learning_rate": 1.7019279288313994e-05, + "loss": 0.9038, + "step": 2531 + }, + { + "epoch": 1.3745928338762214, + "grad_norm": 7.524913368583668, + "learning_rate": 1.7016773992520806e-05, + "loss": 0.8787, + "step": 2532 + }, + { + "epoch": 1.3751357220412594, + "grad_norm": 5.722778911929788, + "learning_rate": 1.701426782888183e-05, + "loss": 0.8199, + "step": 2533 + }, + { + "epoch": 1.3756786102062974, + "grad_norm": 6.26381521799836, + "learning_rate": 1.701176079770703e-05, + "loss": 0.6989, + "step": 2534 + }, + { + "epoch": 1.3762214983713354, + "grad_norm": 7.2300935887481765, + "learning_rate": 1.700925289930648e-05, + "loss": 1.094, + "step": 2535 + }, + { + "epoch": 1.3767643865363735, + "grad_norm": 7.385614391862118, + "learning_rate": 1.7006744133990358e-05, + "loss": 0.7753, + "step": 2536 + }, + { + "epoch": 1.3773072747014115, + "grad_norm": 5.52011787904346, + "learning_rate": 1.7004234502068952e-05, + "loss": 0.9232, + "step": 2537 + }, + { + "epoch": 1.3778501628664495, + "grad_norm": 6.523028453950784, + "learning_rate": 1.7001724003852668e-05, + "loss": 1.1181, + "step": 2538 + }, + { + "epoch": 1.3783930510314875, + "grad_norm": 6.334090281128193, + "learning_rate": 1.6999212639651995e-05, + "loss": 0.9384, + "step": 2539 + }, + { + "epoch": 1.3789359391965255, + "grad_norm": 7.3687722051890265, + "learning_rate": 1.6996700409777548e-05, + "loss": 0.9678, + "step": 2540 + }, + { + "epoch": 1.3794788273615635, + "grad_norm": 9.49113126498042, + "learning_rate": 1.6994187314540042e-05, + "loss": 0.7875, + "step": 2541 + }, + { + "epoch": 1.3800217155266015, + "grad_norm": 4.885613246464195, + "learning_rate": 1.69916733542503e-05, + "loss": 0.4488, + "step": 2542 + }, + { + "epoch": 1.3805646036916395, + "grad_norm": 7.362521236541499, + "learning_rate": 1.6989158529219262e-05, + "loss": 0.9808, + "step": 2543 + }, + { + "epoch": 1.3811074918566775, + "grad_norm": 8.6488909724118, + "learning_rate": 1.6986642839757953e-05, + "loss": 1.0599, + "step": 2544 + }, + { + "epoch": 1.3816503800217155, + "grad_norm": 6.054541159551389, + "learning_rate": 1.698412628617752e-05, + "loss": 0.8252, + "step": 2545 + }, + { + "epoch": 1.3821932681867535, + "grad_norm": 7.77360294326381, + "learning_rate": 1.698160886878922e-05, + "loss": 1.0718, + "step": 2546 + }, + { + "epoch": 1.3827361563517915, + "grad_norm": 9.493714306510151, + "learning_rate": 1.697909058790441e-05, + "loss": 1.2784, + "step": 2547 + }, + { + "epoch": 1.3832790445168295, + "grad_norm": 7.7307988823000455, + "learning_rate": 1.6976571443834555e-05, + "loss": 1.0308, + "step": 2548 + }, + { + "epoch": 1.3838219326818675, + "grad_norm": 7.149534093427325, + "learning_rate": 1.697405143689122e-05, + "loss": 0.9702, + "step": 2549 + }, + { + "epoch": 1.3843648208469055, + "grad_norm": 6.16619254042592, + "learning_rate": 1.6971530567386087e-05, + "loss": 0.6922, + "step": 2550 + }, + { + "epoch": 1.3849077090119435, + "grad_norm": 6.555255328428672, + "learning_rate": 1.6969008835630947e-05, + "loss": 0.8771, + "step": 2551 + }, + { + "epoch": 1.3854505971769815, + "grad_norm": 7.652524555180243, + "learning_rate": 1.6966486241937685e-05, + "loss": 1.0635, + "step": 2552 + }, + { + "epoch": 1.3859934853420195, + "grad_norm": 9.675328775672968, + "learning_rate": 1.69639627866183e-05, + "loss": 1.1616, + "step": 2553 + }, + { + "epoch": 1.3865363735070575, + "grad_norm": 6.5667057730439256, + "learning_rate": 1.69614384699849e-05, + "loss": 0.7181, + "step": 2554 + }, + { + "epoch": 1.3870792616720955, + "grad_norm": 7.374476144001139, + "learning_rate": 1.6958913292349698e-05, + "loss": 1.123, + "step": 2555 + }, + { + "epoch": 1.3876221498371335, + "grad_norm": 8.873033944832025, + "learning_rate": 1.6956387254025005e-05, + "loss": 1.2651, + "step": 2556 + }, + { + "epoch": 1.3881650380021715, + "grad_norm": 6.877978468812787, + "learning_rate": 1.695386035532325e-05, + "loss": 1.0039, + "step": 2557 + }, + { + "epoch": 1.3887079261672095, + "grad_norm": 4.746919583421645, + "learning_rate": 1.6951332596556966e-05, + "loss": 0.7222, + "step": 2558 + }, + { + "epoch": 1.3892508143322475, + "grad_norm": 5.538693020619686, + "learning_rate": 1.6948803978038787e-05, + "loss": 0.8463, + "step": 2559 + }, + { + "epoch": 1.3897937024972855, + "grad_norm": 6.636448326582328, + "learning_rate": 1.6946274500081455e-05, + "loss": 0.7594, + "step": 2560 + }, + { + "epoch": 1.3903365906623235, + "grad_norm": 8.74658695496399, + "learning_rate": 1.6943744162997825e-05, + "loss": 1.224, + "step": 2561 + }, + { + "epoch": 1.3908794788273615, + "grad_norm": 7.305739264312674, + "learning_rate": 1.694121296710085e-05, + "loss": 1.0755, + "step": 2562 + }, + { + "epoch": 1.3914223669923995, + "grad_norm": 4.6475508332537165, + "learning_rate": 1.693868091270359e-05, + "loss": 0.6127, + "step": 2563 + }, + { + "epoch": 1.3919652551574375, + "grad_norm": 8.484116710765257, + "learning_rate": 1.6936148000119218e-05, + "loss": 1.029, + "step": 2564 + }, + { + "epoch": 1.3925081433224755, + "grad_norm": 6.532276109161297, + "learning_rate": 1.6933614229661008e-05, + "loss": 1.0517, + "step": 2565 + }, + { + "epoch": 1.3930510314875135, + "grad_norm": 5.43723324019315, + "learning_rate": 1.6931079601642336e-05, + "loss": 0.6915, + "step": 2566 + }, + { + "epoch": 1.3935939196525515, + "grad_norm": 7.562306350008547, + "learning_rate": 1.69285441163767e-05, + "loss": 0.5726, + "step": 2567 + }, + { + "epoch": 1.3941368078175895, + "grad_norm": 8.500602954507755, + "learning_rate": 1.6926007774177678e-05, + "loss": 1.0297, + "step": 2568 + }, + { + "epoch": 1.3946796959826275, + "grad_norm": 6.432151204429844, + "learning_rate": 1.6923470575358977e-05, + "loss": 1.0016, + "step": 2569 + }, + { + "epoch": 1.3952225841476655, + "grad_norm": 9.552257208599151, + "learning_rate": 1.6920932520234402e-05, + "loss": 1.0814, + "step": 2570 + }, + { + "epoch": 1.3957654723127035, + "grad_norm": 8.075918082294125, + "learning_rate": 1.6918393609117863e-05, + "loss": 1.1699, + "step": 2571 + }, + { + "epoch": 1.3963083604777415, + "grad_norm": 6.296577538864887, + "learning_rate": 1.6915853842323373e-05, + "loss": 1.0211, + "step": 2572 + }, + { + "epoch": 1.3968512486427795, + "grad_norm": 7.785873884786678, + "learning_rate": 1.691331322016506e-05, + "loss": 0.9225, + "step": 2573 + }, + { + "epoch": 1.3973941368078175, + "grad_norm": 6.768130802749829, + "learning_rate": 1.691077174295715e-05, + "loss": 0.8092, + "step": 2574 + }, + { + "epoch": 1.3979370249728555, + "grad_norm": 5.566346022945414, + "learning_rate": 1.6908229411013977e-05, + "loss": 0.8312, + "step": 2575 + }, + { + "epoch": 1.3984799131378935, + "grad_norm": 7.2860384645139815, + "learning_rate": 1.6905686224649978e-05, + "loss": 0.9357, + "step": 2576 + }, + { + "epoch": 1.3990228013029316, + "grad_norm": 7.072468614016554, + "learning_rate": 1.6903142184179704e-05, + "loss": 0.7362, + "step": 2577 + }, + { + "epoch": 1.3995656894679696, + "grad_norm": 6.421225685489366, + "learning_rate": 1.6900597289917803e-05, + "loss": 0.6828, + "step": 2578 + }, + { + "epoch": 1.4001085776330076, + "grad_norm": 8.167489592155093, + "learning_rate": 1.689805154217903e-05, + "loss": 1.1237, + "step": 2579 + }, + { + "epoch": 1.4006514657980456, + "grad_norm": 11.676833397587332, + "learning_rate": 1.6895504941278246e-05, + "loss": 0.9532, + "step": 2580 + }, + { + "epoch": 1.4011943539630836, + "grad_norm": 6.04293054204543, + "learning_rate": 1.689295748753042e-05, + "loss": 0.6541, + "step": 2581 + }, + { + "epoch": 1.4017372421281216, + "grad_norm": 8.432023560966908, + "learning_rate": 1.6890409181250632e-05, + "loss": 1.0167, + "step": 2582 + }, + { + "epoch": 1.4022801302931596, + "grad_norm": 7.494983940960417, + "learning_rate": 1.688786002275405e-05, + "loss": 0.8913, + "step": 2583 + }, + { + "epoch": 1.4028230184581976, + "grad_norm": 6.475912046078606, + "learning_rate": 1.6885310012355964e-05, + "loss": 0.8599, + "step": 2584 + }, + { + "epoch": 1.4033659066232356, + "grad_norm": 5.145369710998311, + "learning_rate": 1.6882759150371765e-05, + "loss": 0.8079, + "step": 2585 + }, + { + "epoch": 1.4039087947882736, + "grad_norm": 6.1855027756804155, + "learning_rate": 1.688020743711694e-05, + "loss": 0.8046, + "step": 2586 + }, + { + "epoch": 1.4044516829533116, + "grad_norm": 8.409570707749053, + "learning_rate": 1.68776548729071e-05, + "loss": 0.75, + "step": 2587 + }, + { + "epoch": 1.4049945711183496, + "grad_norm": 6.150666283054011, + "learning_rate": 1.687510145805794e-05, + "loss": 0.8339, + "step": 2588 + }, + { + "epoch": 1.4055374592833876, + "grad_norm": 5.4082205122977625, + "learning_rate": 1.6872547192885272e-05, + "loss": 0.7878, + "step": 2589 + }, + { + "epoch": 1.4060803474484256, + "grad_norm": 7.011727228130495, + "learning_rate": 1.686999207770502e-05, + "loss": 0.6415, + "step": 2590 + }, + { + "epoch": 1.4066232356134636, + "grad_norm": 7.384524419917869, + "learning_rate": 1.6867436112833193e-05, + "loss": 0.8489, + "step": 2591 + }, + { + "epoch": 1.4071661237785016, + "grad_norm": 6.846659366372164, + "learning_rate": 1.6864879298585925e-05, + "loss": 1.128, + "step": 2592 + }, + { + "epoch": 1.4077090119435396, + "grad_norm": 6.142948047083128, + "learning_rate": 1.6862321635279444e-05, + "loss": 0.7886, + "step": 2593 + }, + { + "epoch": 1.4082519001085776, + "grad_norm": 10.23557881641297, + "learning_rate": 1.6859763123230086e-05, + "loss": 0.8323, + "step": 2594 + }, + { + "epoch": 1.4087947882736156, + "grad_norm": 6.995135696669204, + "learning_rate": 1.6857203762754294e-05, + "loss": 1.0283, + "step": 2595 + }, + { + "epoch": 1.4093376764386536, + "grad_norm": 7.709014920487576, + "learning_rate": 1.685464355416861e-05, + "loss": 0.8586, + "step": 2596 + }, + { + "epoch": 1.4098805646036916, + "grad_norm": 8.153794053960306, + "learning_rate": 1.6852082497789684e-05, + "loss": 1.1797, + "step": 2597 + }, + { + "epoch": 1.4104234527687296, + "grad_norm": 7.721286185320884, + "learning_rate": 1.684952059393428e-05, + "loss": 0.9196, + "step": 2598 + }, + { + "epoch": 1.4109663409337676, + "grad_norm": 8.024349951964256, + "learning_rate": 1.684695784291925e-05, + "loss": 0.7769, + "step": 2599 + }, + { + "epoch": 1.4115092290988056, + "grad_norm": 8.135395328882872, + "learning_rate": 1.684439424506156e-05, + "loss": 1.1642, + "step": 2600 + }, + { + "epoch": 1.4120521172638436, + "grad_norm": 6.057801451822921, + "learning_rate": 1.684182980067828e-05, + "loss": 0.7352, + "step": 2601 + }, + { + "epoch": 1.4125950054288816, + "grad_norm": 7.808755828672067, + "learning_rate": 1.683926451008659e-05, + "loss": 0.9788, + "step": 2602 + }, + { + "epoch": 1.4131378935939196, + "grad_norm": 7.426450515083944, + "learning_rate": 1.6836698373603765e-05, + "loss": 0.7696, + "step": 2603 + }, + { + "epoch": 1.4136807817589576, + "grad_norm": 9.382792907740685, + "learning_rate": 1.6834131391547187e-05, + "loss": 1.0872, + "step": 2604 + }, + { + "epoch": 1.4142236699239956, + "grad_norm": 6.171700145892427, + "learning_rate": 1.6831563564234347e-05, + "loss": 0.7233, + "step": 2605 + }, + { + "epoch": 1.4147665580890336, + "grad_norm": 11.270930151197227, + "learning_rate": 1.682899489198284e-05, + "loss": 0.9739, + "step": 2606 + }, + { + "epoch": 1.4153094462540716, + "grad_norm": 9.050544540368543, + "learning_rate": 1.6826425375110357e-05, + "loss": 0.9436, + "step": 2607 + }, + { + "epoch": 1.4158523344191096, + "grad_norm": 6.969930349987723, + "learning_rate": 1.6823855013934705e-05, + "loss": 0.7632, + "step": 2608 + }, + { + "epoch": 1.4163952225841476, + "grad_norm": 5.791480821515424, + "learning_rate": 1.682128380877379e-05, + "loss": 0.8202, + "step": 2609 + }, + { + "epoch": 1.4169381107491856, + "grad_norm": 9.905757175497309, + "learning_rate": 1.6818711759945623e-05, + "loss": 1.2161, + "step": 2610 + }, + { + "epoch": 1.4174809989142236, + "grad_norm": 6.082684402096193, + "learning_rate": 1.6816138867768318e-05, + "loss": 0.7245, + "step": 2611 + }, + { + "epoch": 1.4180238870792616, + "grad_norm": 10.15499139340414, + "learning_rate": 1.6813565132560092e-05, + "loss": 1.1514, + "step": 2612 + }, + { + "epoch": 1.4185667752442996, + "grad_norm": 6.978513243855861, + "learning_rate": 1.6810990554639276e-05, + "loss": 0.5331, + "step": 2613 + }, + { + "epoch": 1.4191096634093376, + "grad_norm": 8.037739371135425, + "learning_rate": 1.6808415134324288e-05, + "loss": 0.7089, + "step": 2614 + }, + { + "epoch": 1.4196525515743756, + "grad_norm": 6.925791024229076, + "learning_rate": 1.6805838871933664e-05, + "loss": 0.8981, + "step": 2615 + }, + { + "epoch": 1.4201954397394136, + "grad_norm": 6.352165521713976, + "learning_rate": 1.6803261767786048e-05, + "loss": 0.5644, + "step": 2616 + }, + { + "epoch": 1.4207383279044516, + "grad_norm": 10.267497627922179, + "learning_rate": 1.680068382220017e-05, + "loss": 0.9607, + "step": 2617 + }, + { + "epoch": 1.4212812160694897, + "grad_norm": 7.696468630335302, + "learning_rate": 1.679810503549488e-05, + "loss": 0.729, + "step": 2618 + }, + { + "epoch": 1.4218241042345277, + "grad_norm": 6.868981955235214, + "learning_rate": 1.679552540798912e-05, + "loss": 0.9152, + "step": 2619 + }, + { + "epoch": 1.4223669923995657, + "grad_norm": 8.909790381587262, + "learning_rate": 1.6792944940001952e-05, + "loss": 0.9335, + "step": 2620 + }, + { + "epoch": 1.4229098805646037, + "grad_norm": 9.62415635421133, + "learning_rate": 1.6790363631852524e-05, + "loss": 1.1012, + "step": 2621 + }, + { + "epoch": 1.4234527687296417, + "grad_norm": 7.307385111665146, + "learning_rate": 1.67877814838601e-05, + "loss": 0.6963, + "step": 2622 + }, + { + "epoch": 1.4239956568946797, + "grad_norm": 9.018642559605246, + "learning_rate": 1.678519849634405e-05, + "loss": 1.0612, + "step": 2623 + }, + { + "epoch": 1.4245385450597177, + "grad_norm": 6.929244675286548, + "learning_rate": 1.6782614669623827e-05, + "loss": 0.9462, + "step": 2624 + }, + { + "epoch": 1.4250814332247557, + "grad_norm": 8.577419731292203, + "learning_rate": 1.6780030004019016e-05, + "loss": 1.112, + "step": 2625 + }, + { + "epoch": 1.4256243213897937, + "grad_norm": 7.125694710248634, + "learning_rate": 1.677744449984929e-05, + "loss": 0.851, + "step": 2626 + }, + { + "epoch": 1.4261672095548317, + "grad_norm": 9.488635353984538, + "learning_rate": 1.6774858157434425e-05, + "loss": 1.3064, + "step": 2627 + }, + { + "epoch": 1.4267100977198697, + "grad_norm": 7.500971938576216, + "learning_rate": 1.6772270977094307e-05, + "loss": 1.3675, + "step": 2628 + }, + { + "epoch": 1.4272529858849077, + "grad_norm": 6.979855088071626, + "learning_rate": 1.676968295914892e-05, + "loss": 0.8896, + "step": 2629 + }, + { + "epoch": 1.4277958740499457, + "grad_norm": 7.795843207695432, + "learning_rate": 1.6767094103918357e-05, + "loss": 0.7758, + "step": 2630 + }, + { + "epoch": 1.4283387622149837, + "grad_norm": 6.131884468994572, + "learning_rate": 1.6764504411722806e-05, + "loss": 0.7744, + "step": 2631 + }, + { + "epoch": 1.4288816503800217, + "grad_norm": 10.33693974618067, + "learning_rate": 1.676191388288257e-05, + "loss": 1.5364, + "step": 2632 + }, + { + "epoch": 1.4294245385450597, + "grad_norm": 10.005426546410897, + "learning_rate": 1.6759322517718048e-05, + "loss": 0.9713, + "step": 2633 + }, + { + "epoch": 1.4299674267100977, + "grad_norm": 9.480236119835816, + "learning_rate": 1.6756730316549745e-05, + "loss": 1.2573, + "step": 2634 + }, + { + "epoch": 1.4305103148751357, + "grad_norm": 10.874101086577593, + "learning_rate": 1.675413727969827e-05, + "loss": 1.0308, + "step": 2635 + }, + { + "epoch": 1.4310532030401737, + "grad_norm": 5.34984499382639, + "learning_rate": 1.675154340748433e-05, + "loss": 0.6013, + "step": 2636 + }, + { + "epoch": 1.4315960912052117, + "grad_norm": 7.464386702227026, + "learning_rate": 1.674894870022874e-05, + "loss": 0.9918, + "step": 2637 + }, + { + "epoch": 1.4321389793702497, + "grad_norm": 8.280480153626241, + "learning_rate": 1.674635315825242e-05, + "loss": 0.8745, + "step": 2638 + }, + { + "epoch": 1.4326818675352877, + "grad_norm": 6.169575731939732, + "learning_rate": 1.6743756781876385e-05, + "loss": 0.6984, + "step": 2639 + }, + { + "epoch": 1.4332247557003257, + "grad_norm": 5.390920772379431, + "learning_rate": 1.6741159571421768e-05, + "loss": 0.8886, + "step": 2640 + }, + { + "epoch": 1.4337676438653637, + "grad_norm": 7.003336357340687, + "learning_rate": 1.6738561527209792e-05, + "loss": 1.0335, + "step": 2641 + }, + { + "epoch": 1.4343105320304017, + "grad_norm": 7.722166856340297, + "learning_rate": 1.6735962649561784e-05, + "loss": 0.9804, + "step": 2642 + }, + { + "epoch": 1.4348534201954397, + "grad_norm": 7.328242752900802, + "learning_rate": 1.673336293879918e-05, + "loss": 0.8675, + "step": 2643 + }, + { + "epoch": 1.4353963083604777, + "grad_norm": 7.344577484351048, + "learning_rate": 1.6730762395243515e-05, + "loss": 1.2518, + "step": 2644 + }, + { + "epoch": 1.4359391965255157, + "grad_norm": 7.560893408391802, + "learning_rate": 1.6728161019216433e-05, + "loss": 1.0196, + "step": 2645 + }, + { + "epoch": 1.4364820846905537, + "grad_norm": 7.509821569775923, + "learning_rate": 1.6725558811039674e-05, + "loss": 1.1112, + "step": 2646 + }, + { + "epoch": 1.4370249728555917, + "grad_norm": 7.672412677248201, + "learning_rate": 1.672295577103508e-05, + "loss": 0.9075, + "step": 2647 + }, + { + "epoch": 1.4375678610206297, + "grad_norm": 8.562389123203614, + "learning_rate": 1.67203518995246e-05, + "loss": 1.2123, + "step": 2648 + }, + { + "epoch": 1.4381107491856677, + "grad_norm": 8.172008773566091, + "learning_rate": 1.671774719683029e-05, + "loss": 0.8064, + "step": 2649 + }, + { + "epoch": 1.4386536373507057, + "grad_norm": 7.5145061670971405, + "learning_rate": 1.6715141663274297e-05, + "loss": 1.3513, + "step": 2650 + }, + { + "epoch": 1.4391965255157437, + "grad_norm": 5.675981880534199, + "learning_rate": 1.6712535299178883e-05, + "loss": 0.5029, + "step": 2651 + }, + { + "epoch": 1.4397394136807817, + "grad_norm": 8.195158801834715, + "learning_rate": 1.6709928104866403e-05, + "loss": 0.9414, + "step": 2652 + }, + { + "epoch": 1.4402823018458197, + "grad_norm": 7.374510239535514, + "learning_rate": 1.6707320080659322e-05, + "loss": 0.8103, + "step": 2653 + }, + { + "epoch": 1.4408251900108577, + "grad_norm": 5.703232221762084, + "learning_rate": 1.6704711226880204e-05, + "loss": 1.1646, + "step": 2654 + }, + { + "epoch": 1.4413680781758957, + "grad_norm": 6.981749095787225, + "learning_rate": 1.6702101543851714e-05, + "loss": 0.7174, + "step": 2655 + }, + { + "epoch": 1.4419109663409337, + "grad_norm": 6.456866833246808, + "learning_rate": 1.6699491031896625e-05, + "loss": 0.9795, + "step": 2656 + }, + { + "epoch": 1.4424538545059717, + "grad_norm": 10.193228844116039, + "learning_rate": 1.6696879691337807e-05, + "loss": 1.1288, + "step": 2657 + }, + { + "epoch": 1.4429967426710097, + "grad_norm": 9.920534268729181, + "learning_rate": 1.6694267522498237e-05, + "loss": 0.8379, + "step": 2658 + }, + { + "epoch": 1.4435396308360477, + "grad_norm": 6.153776792682404, + "learning_rate": 1.669165452570099e-05, + "loss": 0.8818, + "step": 2659 + }, + { + "epoch": 1.4440825190010858, + "grad_norm": 5.650871346064418, + "learning_rate": 1.6689040701269245e-05, + "loss": 0.732, + "step": 2660 + }, + { + "epoch": 1.4446254071661238, + "grad_norm": 7.829649689315363, + "learning_rate": 1.668642604952629e-05, + "loss": 0.8119, + "step": 2661 + }, + { + "epoch": 1.4451682953311618, + "grad_norm": 5.959350361834891, + "learning_rate": 1.6683810570795498e-05, + "loss": 0.6829, + "step": 2662 + }, + { + "epoch": 1.4457111834961998, + "grad_norm": 8.76687690585514, + "learning_rate": 1.6681194265400365e-05, + "loss": 1.4888, + "step": 2663 + }, + { + "epoch": 1.4462540716612378, + "grad_norm": 6.129763847493397, + "learning_rate": 1.6678577133664476e-05, + "loss": 0.8372, + "step": 2664 + }, + { + "epoch": 1.4467969598262758, + "grad_norm": 6.529235674005995, + "learning_rate": 1.6675959175911527e-05, + "loss": 0.9655, + "step": 2665 + }, + { + "epoch": 1.4473398479913138, + "grad_norm": 7.741547370588863, + "learning_rate": 1.6673340392465304e-05, + "loss": 1.0772, + "step": 2666 + }, + { + "epoch": 1.4478827361563518, + "grad_norm": 6.674042204787818, + "learning_rate": 1.6670720783649706e-05, + "loss": 0.9681, + "step": 2667 + }, + { + "epoch": 1.4484256243213898, + "grad_norm": 7.920578917847642, + "learning_rate": 1.666810034978873e-05, + "loss": 1.1068, + "step": 2668 + }, + { + "epoch": 1.4489685124864278, + "grad_norm": 7.249023283604718, + "learning_rate": 1.6665479091206476e-05, + "loss": 1.1025, + "step": 2669 + }, + { + "epoch": 1.4495114006514658, + "grad_norm": 6.407638163624444, + "learning_rate": 1.6662857008227145e-05, + "loss": 0.8236, + "step": 2670 + }, + { + "epoch": 1.4500542888165038, + "grad_norm": 7.289336395940614, + "learning_rate": 1.6660234101175036e-05, + "loss": 0.9386, + "step": 2671 + }, + { + "epoch": 1.4505971769815418, + "grad_norm": 6.973959594036107, + "learning_rate": 1.665761037037456e-05, + "loss": 0.8435, + "step": 2672 + }, + { + "epoch": 1.4511400651465798, + "grad_norm": 6.971928597160326, + "learning_rate": 1.665498581615023e-05, + "loss": 0.9065, + "step": 2673 + }, + { + "epoch": 1.4516829533116178, + "grad_norm": 6.1608850694063735, + "learning_rate": 1.665236043882664e-05, + "loss": 0.9289, + "step": 2674 + }, + { + "epoch": 1.4522258414766558, + "grad_norm": 6.62873967155635, + "learning_rate": 1.6649734238728512e-05, + "loss": 0.7797, + "step": 2675 + }, + { + "epoch": 1.4527687296416938, + "grad_norm": 7.089586063125632, + "learning_rate": 1.6647107216180655e-05, + "loss": 0.6306, + "step": 2676 + }, + { + "epoch": 1.4533116178067318, + "grad_norm": 10.986990416571043, + "learning_rate": 1.6644479371507985e-05, + "loss": 0.8764, + "step": 2677 + }, + { + "epoch": 1.4538545059717698, + "grad_norm": 5.865852703001031, + "learning_rate": 1.664185070503551e-05, + "loss": 0.9106, + "step": 2678 + }, + { + "epoch": 1.4543973941368078, + "grad_norm": 7.303323841843, + "learning_rate": 1.663922121708836e-05, + "loss": 0.8188, + "step": 2679 + }, + { + "epoch": 1.4549402823018458, + "grad_norm": 5.715280365147904, + "learning_rate": 1.663659090799175e-05, + "loss": 1.0573, + "step": 2680 + }, + { + "epoch": 1.4554831704668838, + "grad_norm": 7.084313061900368, + "learning_rate": 1.6633959778070992e-05, + "loss": 1.0749, + "step": 2681 + }, + { + "epoch": 1.4560260586319218, + "grad_norm": 7.902311581251679, + "learning_rate": 1.6631327827651524e-05, + "loss": 0.8677, + "step": 2682 + }, + { + "epoch": 1.4565689467969598, + "grad_norm": 8.1594384957164, + "learning_rate": 1.6628695057058855e-05, + "loss": 0.9999, + "step": 2683 + }, + { + "epoch": 1.4571118349619978, + "grad_norm": 5.580141751242927, + "learning_rate": 1.6626061466618623e-05, + "loss": 0.6387, + "step": 2684 + }, + { + "epoch": 1.4576547231270358, + "grad_norm": 8.145754034329466, + "learning_rate": 1.6623427056656544e-05, + "loss": 0.8984, + "step": 2685 + }, + { + "epoch": 1.4581976112920738, + "grad_norm": 6.453637820501138, + "learning_rate": 1.6620791827498454e-05, + "loss": 0.6922, + "step": 2686 + }, + { + "epoch": 1.4587404994571118, + "grad_norm": 7.3511834666776075, + "learning_rate": 1.6618155779470275e-05, + "loss": 0.8462, + "step": 2687 + }, + { + "epoch": 1.4592833876221498, + "grad_norm": 7.962874909840381, + "learning_rate": 1.6615518912898043e-05, + "loss": 0.9639, + "step": 2688 + }, + { + "epoch": 1.4598262757871878, + "grad_norm": 6.975602073560851, + "learning_rate": 1.6612881228107886e-05, + "loss": 0.9364, + "step": 2689 + }, + { + "epoch": 1.4603691639522258, + "grad_norm": 6.916712660881488, + "learning_rate": 1.6610242725426044e-05, + "loss": 0.8033, + "step": 2690 + }, + { + "epoch": 1.4609120521172638, + "grad_norm": 6.268208953592866, + "learning_rate": 1.6607603405178842e-05, + "loss": 0.7542, + "step": 2691 + }, + { + "epoch": 1.4614549402823018, + "grad_norm": 9.047138051843492, + "learning_rate": 1.660496326769272e-05, + "loss": 1.5956, + "step": 2692 + }, + { + "epoch": 1.4619978284473398, + "grad_norm": 7.228556514923675, + "learning_rate": 1.6602322313294216e-05, + "loss": 0.8461, + "step": 2693 + }, + { + "epoch": 1.4625407166123778, + "grad_norm": 6.195288408674038, + "learning_rate": 1.659968054230997e-05, + "loss": 0.9851, + "step": 2694 + }, + { + "epoch": 1.4630836047774158, + "grad_norm": 7.413116367868534, + "learning_rate": 1.6597037955066713e-05, + "loss": 0.9693, + "step": 2695 + }, + { + "epoch": 1.4636264929424538, + "grad_norm": 7.014031390984585, + "learning_rate": 1.6594394551891288e-05, + "loss": 1.0539, + "step": 2696 + }, + { + "epoch": 1.4641693811074918, + "grad_norm": 6.8196013668965225, + "learning_rate": 1.6591750333110634e-05, + "loss": 0.712, + "step": 2697 + }, + { + "epoch": 1.4647122692725298, + "grad_norm": 8.160690900976675, + "learning_rate": 1.658910529905179e-05, + "loss": 0.6551, + "step": 2698 + }, + { + "epoch": 1.4652551574375678, + "grad_norm": 5.174705288640052, + "learning_rate": 1.6586459450041906e-05, + "loss": 0.8285, + "step": 2699 + }, + { + "epoch": 1.4657980456026058, + "grad_norm": 7.029482771999071, + "learning_rate": 1.6583812786408216e-05, + "loss": 0.923, + "step": 2700 + }, + { + "epoch": 1.4663409337676439, + "grad_norm": 7.8877107275923946, + "learning_rate": 1.658116530847807e-05, + "loss": 1.1915, + "step": 2701 + }, + { + "epoch": 1.4668838219326819, + "grad_norm": 9.269305925663183, + "learning_rate": 1.657851701657891e-05, + "loss": 1.098, + "step": 2702 + }, + { + "epoch": 1.4674267100977199, + "grad_norm": 10.478546362437942, + "learning_rate": 1.657586791103828e-05, + "loss": 1.2416, + "step": 2703 + }, + { + "epoch": 1.4679695982627579, + "grad_norm": 8.207786471433751, + "learning_rate": 1.6573217992183826e-05, + "loss": 0.8658, + "step": 2704 + }, + { + "epoch": 1.4685124864277959, + "grad_norm": 7.862118410069405, + "learning_rate": 1.6570567260343294e-05, + "loss": 0.8714, + "step": 2705 + }, + { + "epoch": 1.4690553745928339, + "grad_norm": 7.744957075186954, + "learning_rate": 1.6567915715844534e-05, + "loss": 0.9374, + "step": 2706 + }, + { + "epoch": 1.4695982627578719, + "grad_norm": 6.97424297060321, + "learning_rate": 1.6565263359015488e-05, + "loss": 0.7183, + "step": 2707 + }, + { + "epoch": 1.4701411509229099, + "grad_norm": 7.349900880491086, + "learning_rate": 1.6562610190184206e-05, + "loss": 0.6066, + "step": 2708 + }, + { + "epoch": 1.4706840390879479, + "grad_norm": 6.982564531951699, + "learning_rate": 1.655995620967884e-05, + "loss": 1.2218, + "step": 2709 + }, + { + "epoch": 1.4712269272529859, + "grad_norm": 7.572476514419708, + "learning_rate": 1.6557301417827632e-05, + "loss": 0.89, + "step": 2710 + }, + { + "epoch": 1.4717698154180239, + "grad_norm": 8.654751087453503, + "learning_rate": 1.6554645814958932e-05, + "loss": 0.7188, + "step": 2711 + }, + { + "epoch": 1.4723127035830619, + "grad_norm": 6.282711052010393, + "learning_rate": 1.6551989401401196e-05, + "loss": 0.7586, + "step": 2712 + }, + { + "epoch": 1.4728555917480999, + "grad_norm": 8.732262593936317, + "learning_rate": 1.6549332177482966e-05, + "loss": 1.0178, + "step": 2713 + }, + { + "epoch": 1.4733984799131379, + "grad_norm": 6.989470691687101, + "learning_rate": 1.6546674143532895e-05, + "loss": 0.6981, + "step": 2714 + }, + { + "epoch": 1.4739413680781759, + "grad_norm": 7.179372740296565, + "learning_rate": 1.6544015299879734e-05, + "loss": 0.7617, + "step": 2715 + }, + { + "epoch": 1.4744842562432139, + "grad_norm": 9.117294639273645, + "learning_rate": 1.6541355646852327e-05, + "loss": 0.8822, + "step": 2716 + }, + { + "epoch": 1.475027144408252, + "grad_norm": 7.875727317127435, + "learning_rate": 1.653869518477963e-05, + "loss": 0.9149, + "step": 2717 + }, + { + "epoch": 1.47557003257329, + "grad_norm": 7.926910105501724, + "learning_rate": 1.6536033913990687e-05, + "loss": 0.8797, + "step": 2718 + }, + { + "epoch": 1.476112920738328, + "grad_norm": 5.14878192977047, + "learning_rate": 1.6533371834814657e-05, + "loss": 0.4491, + "step": 2719 + }, + { + "epoch": 1.476655808903366, + "grad_norm": 5.327596982546826, + "learning_rate": 1.6530708947580785e-05, + "loss": 0.4335, + "step": 2720 + }, + { + "epoch": 1.477198697068404, + "grad_norm": 9.158336314582998, + "learning_rate": 1.6528045252618423e-05, + "loss": 1.1208, + "step": 2721 + }, + { + "epoch": 1.477741585233442, + "grad_norm": 10.630652314970348, + "learning_rate": 1.6525380750257022e-05, + "loss": 1.097, + "step": 2722 + }, + { + "epoch": 1.47828447339848, + "grad_norm": 8.97220182564547, + "learning_rate": 1.652271544082613e-05, + "loss": 0.7838, + "step": 2723 + }, + { + "epoch": 1.478827361563518, + "grad_norm": 10.609251261516084, + "learning_rate": 1.652004932465539e-05, + "loss": 0.8861, + "step": 2724 + }, + { + "epoch": 1.479370249728556, + "grad_norm": 7.303801635012744, + "learning_rate": 1.6517382402074563e-05, + "loss": 0.5437, + "step": 2725 + }, + { + "epoch": 1.479913137893594, + "grad_norm": 11.770584674671793, + "learning_rate": 1.651471467341349e-05, + "loss": 1.4947, + "step": 2726 + }, + { + "epoch": 1.480456026058632, + "grad_norm": 6.824299484966605, + "learning_rate": 1.6512046139002128e-05, + "loss": 0.7144, + "step": 2727 + }, + { + "epoch": 1.48099891422367, + "grad_norm": 7.595224046166877, + "learning_rate": 1.650937679917052e-05, + "loss": 1.0259, + "step": 2728 + }, + { + "epoch": 1.481541802388708, + "grad_norm": 8.979875171863954, + "learning_rate": 1.6506706654248813e-05, + "loss": 1.0951, + "step": 2729 + }, + { + "epoch": 1.482084690553746, + "grad_norm": 7.700209331252355, + "learning_rate": 1.650403570456726e-05, + "loss": 1.0228, + "step": 2730 + }, + { + "epoch": 1.482627578718784, + "grad_norm": 9.090566362046857, + "learning_rate": 1.65013639504562e-05, + "loss": 1.0199, + "step": 2731 + }, + { + "epoch": 1.483170466883822, + "grad_norm": 6.644791741009562, + "learning_rate": 1.6498691392246088e-05, + "loss": 0.5478, + "step": 2732 + }, + { + "epoch": 1.48371335504886, + "grad_norm": 7.067718277939746, + "learning_rate": 1.6496018030267467e-05, + "loss": 0.864, + "step": 2733 + }, + { + "epoch": 1.484256243213898, + "grad_norm": 8.017737461642762, + "learning_rate": 1.6493343864850984e-05, + "loss": 1.1688, + "step": 2734 + }, + { + "epoch": 1.484799131378936, + "grad_norm": 6.479930584435578, + "learning_rate": 1.6490668896327382e-05, + "loss": 0.7227, + "step": 2735 + }, + { + "epoch": 1.485342019543974, + "grad_norm": 8.392289588009671, + "learning_rate": 1.6487993125027504e-05, + "loss": 0.7213, + "step": 2736 + }, + { + "epoch": 1.485884907709012, + "grad_norm": 7.817891248616158, + "learning_rate": 1.64853165512823e-05, + "loss": 0.7521, + "step": 2737 + }, + { + "epoch": 1.48642779587405, + "grad_norm": 8.824885297850575, + "learning_rate": 1.6482639175422804e-05, + "loss": 1.016, + "step": 2738 + }, + { + "epoch": 1.486970684039088, + "grad_norm": 11.48655873033237, + "learning_rate": 1.6479960997780165e-05, + "loss": 0.9814, + "step": 2739 + }, + { + "epoch": 1.487513572204126, + "grad_norm": 10.88221555712033, + "learning_rate": 1.6477282018685628e-05, + "loss": 1.464, + "step": 2740 + }, + { + "epoch": 1.488056460369164, + "grad_norm": 8.701247869952772, + "learning_rate": 1.6474602238470524e-05, + "loss": 0.918, + "step": 2741 + }, + { + "epoch": 1.488599348534202, + "grad_norm": 6.4327269926453505, + "learning_rate": 1.6471921657466294e-05, + "loss": 0.7949, + "step": 2742 + }, + { + "epoch": 1.48914223669924, + "grad_norm": 7.592315420975657, + "learning_rate": 1.6469240276004477e-05, + "loss": 0.9485, + "step": 2743 + }, + { + "epoch": 1.489685124864278, + "grad_norm": 7.682698616258038, + "learning_rate": 1.6466558094416717e-05, + "loss": 0.8439, + "step": 2744 + }, + { + "epoch": 1.490228013029316, + "grad_norm": 6.127626205001148, + "learning_rate": 1.6463875113034743e-05, + "loss": 0.8652, + "step": 2745 + }, + { + "epoch": 1.490770901194354, + "grad_norm": 9.65248236696942, + "learning_rate": 1.6461191332190397e-05, + "loss": 1.6278, + "step": 2746 + }, + { + "epoch": 1.491313789359392, + "grad_norm": 8.46309710838681, + "learning_rate": 1.6458506752215603e-05, + "loss": 1.5697, + "step": 2747 + }, + { + "epoch": 1.49185667752443, + "grad_norm": 6.578253971678976, + "learning_rate": 1.6455821373442407e-05, + "loss": 0.8836, + "step": 2748 + }, + { + "epoch": 1.492399565689468, + "grad_norm": 5.927853369454149, + "learning_rate": 1.645313519620293e-05, + "loss": 0.7182, + "step": 2749 + }, + { + "epoch": 1.492942453854506, + "grad_norm": 7.725891960915116, + "learning_rate": 1.645044822082941e-05, + "loss": 1.1794, + "step": 2750 + }, + { + "epoch": 1.493485342019544, + "grad_norm": 8.975220195142995, + "learning_rate": 1.644776044765417e-05, + "loss": 0.8909, + "step": 2751 + }, + { + "epoch": 1.494028230184582, + "grad_norm": 7.34643792387778, + "learning_rate": 1.6445071877009643e-05, + "loss": 1.0047, + "step": 2752 + }, + { + "epoch": 1.49457111834962, + "grad_norm": 7.561064198594243, + "learning_rate": 1.6442382509228355e-05, + "loss": 0.7533, + "step": 2753 + }, + { + "epoch": 1.495114006514658, + "grad_norm": 9.327085390245353, + "learning_rate": 1.6439692344642933e-05, + "loss": 0.8048, + "step": 2754 + }, + { + "epoch": 1.495656894679696, + "grad_norm": 8.662401438640156, + "learning_rate": 1.6437001383586095e-05, + "loss": 1.2513, + "step": 2755 + }, + { + "epoch": 1.496199782844734, + "grad_norm": 7.683989985812991, + "learning_rate": 1.6434309626390667e-05, + "loss": 0.8005, + "step": 2756 + }, + { + "epoch": 1.496742671009772, + "grad_norm": 6.295413547214335, + "learning_rate": 1.6431617073389574e-05, + "loss": 0.8778, + "step": 2757 + }, + { + "epoch": 1.49728555917481, + "grad_norm": 10.390849259618786, + "learning_rate": 1.6428923724915825e-05, + "loss": 1.6234, + "step": 2758 + }, + { + "epoch": 1.497828447339848, + "grad_norm": 6.815996692584369, + "learning_rate": 1.6426229581302545e-05, + "loss": 0.8142, + "step": 2759 + }, + { + "epoch": 1.498371335504886, + "grad_norm": 7.203388917066917, + "learning_rate": 1.642353464288295e-05, + "loss": 1.1273, + "step": 2760 + }, + { + "epoch": 1.498914223669924, + "grad_norm": 6.45381693158517, + "learning_rate": 1.6420838909990356e-05, + "loss": 0.7989, + "step": 2761 + }, + { + "epoch": 1.499457111834962, + "grad_norm": 7.135095705156932, + "learning_rate": 1.6418142382958167e-05, + "loss": 0.6753, + "step": 2762 + }, + { + "epoch": 1.5, + "grad_norm": 8.263302638252904, + "learning_rate": 1.64154450621199e-05, + "loss": 1.0357, + "step": 2763 + }, + { + "epoch": 1.500542888165038, + "grad_norm": 8.19537793615059, + "learning_rate": 1.6412746947809165e-05, + "loss": 0.8135, + "step": 2764 + }, + { + "epoch": 1.501085776330076, + "grad_norm": 5.084112201229359, + "learning_rate": 1.6410048040359665e-05, + "loss": 0.3466, + "step": 2765 + }, + { + "epoch": 1.501628664495114, + "grad_norm": 6.621661501175659, + "learning_rate": 1.6407348340105208e-05, + "loss": 1.1468, + "step": 2766 + }, + { + "epoch": 1.502171552660152, + "grad_norm": 7.5339066223692495, + "learning_rate": 1.6404647847379696e-05, + "loss": 0.8808, + "step": 2767 + }, + { + "epoch": 1.50271444082519, + "grad_norm": 6.501647846893153, + "learning_rate": 1.6401946562517134e-05, + "loss": 0.7589, + "step": 2768 + }, + { + "epoch": 1.503257328990228, + "grad_norm": 8.29356168583173, + "learning_rate": 1.6399244485851614e-05, + "loss": 1.3969, + "step": 2769 + }, + { + "epoch": 1.503800217155266, + "grad_norm": 6.518146966456392, + "learning_rate": 1.6396541617717337e-05, + "loss": 0.7951, + "step": 2770 + }, + { + "epoch": 1.504343105320304, + "grad_norm": 6.51134736256057, + "learning_rate": 1.63938379584486e-05, + "loss": 0.9073, + "step": 2771 + }, + { + "epoch": 1.504885993485342, + "grad_norm": 6.682459044312825, + "learning_rate": 1.6391133508379797e-05, + "loss": 0.733, + "step": 2772 + }, + { + "epoch": 1.50542888165038, + "grad_norm": 6.390687464891663, + "learning_rate": 1.638842826784541e-05, + "loss": 0.6934, + "step": 2773 + }, + { + "epoch": 1.505971769815418, + "grad_norm": 8.130197298030478, + "learning_rate": 1.6385722237180038e-05, + "loss": 0.8567, + "step": 2774 + }, + { + "epoch": 1.506514657980456, + "grad_norm": 7.959636836831885, + "learning_rate": 1.6383015416718356e-05, + "loss": 0.9926, + "step": 2775 + }, + { + "epoch": 1.507057546145494, + "grad_norm": 7.732366564125716, + "learning_rate": 1.638030780679516e-05, + "loss": 1.2228, + "step": 2776 + }, + { + "epoch": 1.507600434310532, + "grad_norm": 4.633651796670195, + "learning_rate": 1.6377599407745324e-05, + "loss": 0.4633, + "step": 2777 + }, + { + "epoch": 1.50814332247557, + "grad_norm": 7.016748827947926, + "learning_rate": 1.6374890219903828e-05, + "loss": 0.9593, + "step": 2778 + }, + { + "epoch": 1.508686210640608, + "grad_norm": 8.428067394567327, + "learning_rate": 1.637218024360575e-05, + "loss": 1.0493, + "step": 2779 + }, + { + "epoch": 1.509229098805646, + "grad_norm": 8.297567044395613, + "learning_rate": 1.6369469479186266e-05, + "loss": 1.2335, + "step": 2780 + }, + { + "epoch": 1.509771986970684, + "grad_norm": 5.8755950768313205, + "learning_rate": 1.6366757926980643e-05, + "loss": 0.6556, + "step": 2781 + }, + { + "epoch": 1.510314875135722, + "grad_norm": 6.837923725489964, + "learning_rate": 1.6364045587324254e-05, + "loss": 0.7031, + "step": 2782 + }, + { + "epoch": 1.51085776330076, + "grad_norm": 6.718777411432131, + "learning_rate": 1.6361332460552565e-05, + "loss": 0.8982, + "step": 2783 + }, + { + "epoch": 1.511400651465798, + "grad_norm": 8.788667184000067, + "learning_rate": 1.6358618547001137e-05, + "loss": 0.8548, + "step": 2784 + }, + { + "epoch": 1.511943539630836, + "grad_norm": 6.559897969720367, + "learning_rate": 1.635590384700563e-05, + "loss": 0.9627, + "step": 2785 + }, + { + "epoch": 1.512486427795874, + "grad_norm": 6.538968751920423, + "learning_rate": 1.635318836090181e-05, + "loss": 1.0731, + "step": 2786 + }, + { + "epoch": 1.513029315960912, + "grad_norm": 8.149646454407613, + "learning_rate": 1.6350472089025523e-05, + "loss": 0.9898, + "step": 2787 + }, + { + "epoch": 1.51357220412595, + "grad_norm": 8.544281989077668, + "learning_rate": 1.6347755031712734e-05, + "loss": 1.1125, + "step": 2788 + }, + { + "epoch": 1.514115092290988, + "grad_norm": 6.930736082842491, + "learning_rate": 1.634503718929948e-05, + "loss": 0.7617, + "step": 2789 + }, + { + "epoch": 1.514657980456026, + "grad_norm": 8.527931500643358, + "learning_rate": 1.6342318562121916e-05, + "loss": 1.2179, + "step": 2790 + }, + { + "epoch": 1.515200868621064, + "grad_norm": 7.509610271006853, + "learning_rate": 1.6339599150516283e-05, + "loss": 0.908, + "step": 2791 + }, + { + "epoch": 1.515743756786102, + "grad_norm": 6.0902820346277045, + "learning_rate": 1.6336878954818926e-05, + "loss": 0.829, + "step": 2792 + }, + { + "epoch": 1.51628664495114, + "grad_norm": 7.3274444131512935, + "learning_rate": 1.6334157975366278e-05, + "loss": 0.9667, + "step": 2793 + }, + { + "epoch": 1.516829533116178, + "grad_norm": 7.753641450951333, + "learning_rate": 1.633143621249488e-05, + "loss": 0.7961, + "step": 2794 + }, + { + "epoch": 1.517372421281216, + "grad_norm": 14.786266283108317, + "learning_rate": 1.6328713666541357e-05, + "loss": 1.4119, + "step": 2795 + }, + { + "epoch": 1.517915309446254, + "grad_norm": 6.1388168060810875, + "learning_rate": 1.632599033784244e-05, + "loss": 0.7726, + "step": 2796 + }, + { + "epoch": 1.518458197611292, + "grad_norm": 9.134190897524304, + "learning_rate": 1.632326622673496e-05, + "loss": 1.208, + "step": 2797 + }, + { + "epoch": 1.51900108577633, + "grad_norm": 6.898277731717493, + "learning_rate": 1.632054133355583e-05, + "loss": 0.6929, + "step": 2798 + }, + { + "epoch": 1.519543973941368, + "grad_norm": 6.90513711625334, + "learning_rate": 1.631781565864208e-05, + "loss": 0.7279, + "step": 2799 + }, + { + "epoch": 1.520086862106406, + "grad_norm": 8.332173251791465, + "learning_rate": 1.6315089202330817e-05, + "loss": 0.987, + "step": 2800 + }, + { + "epoch": 1.520629750271444, + "grad_norm": 6.783078863499027, + "learning_rate": 1.631236196495926e-05, + "loss": 0.6793, + "step": 2801 + }, + { + "epoch": 1.521172638436482, + "grad_norm": 7.841204229725534, + "learning_rate": 1.6309633946864712e-05, + "loss": 0.832, + "step": 2802 + }, + { + "epoch": 1.52171552660152, + "grad_norm": 6.4931904342190405, + "learning_rate": 1.630690514838458e-05, + "loss": 0.6662, + "step": 2803 + }, + { + "epoch": 1.522258414766558, + "grad_norm": 8.50187086749661, + "learning_rate": 1.6304175569856368e-05, + "loss": 0.8359, + "step": 2804 + }, + { + "epoch": 1.522801302931596, + "grad_norm": 6.196174007418195, + "learning_rate": 1.6301445211617676e-05, + "loss": 0.8437, + "step": 2805 + }, + { + "epoch": 1.523344191096634, + "grad_norm": 8.232047070633076, + "learning_rate": 1.6298714074006196e-05, + "loss": 0.9134, + "step": 2806 + }, + { + "epoch": 1.523887079261672, + "grad_norm": 6.594428796101092, + "learning_rate": 1.629598215735972e-05, + "loss": 0.6703, + "step": 2807 + }, + { + "epoch": 1.52442996742671, + "grad_norm": 7.634546429927066, + "learning_rate": 1.629324946201614e-05, + "loss": 0.8208, + "step": 2808 + }, + { + "epoch": 1.524972855591748, + "grad_norm": 8.225245989940067, + "learning_rate": 1.6290515988313432e-05, + "loss": 1.0002, + "step": 2809 + }, + { + "epoch": 1.5255157437567861, + "grad_norm": 8.334512060802671, + "learning_rate": 1.628778173658968e-05, + "loss": 0.8883, + "step": 2810 + }, + { + "epoch": 1.5260586319218241, + "grad_norm": 7.733244019009635, + "learning_rate": 1.6285046707183068e-05, + "loss": 0.7155, + "step": 2811 + }, + { + "epoch": 1.5266015200868621, + "grad_norm": 9.607081900344873, + "learning_rate": 1.628231090043186e-05, + "loss": 0.8321, + "step": 2812 + }, + { + "epoch": 1.5271444082519001, + "grad_norm": 8.390046905361096, + "learning_rate": 1.6279574316674426e-05, + "loss": 0.8818, + "step": 2813 + }, + { + "epoch": 1.5276872964169381, + "grad_norm": 7.608191285481891, + "learning_rate": 1.6276836956249235e-05, + "loss": 0.7382, + "step": 2814 + }, + { + "epoch": 1.5282301845819761, + "grad_norm": 8.998269112751723, + "learning_rate": 1.6274098819494844e-05, + "loss": 0.9285, + "step": 2815 + }, + { + "epoch": 1.5287730727470141, + "grad_norm": 5.877268783539307, + "learning_rate": 1.627135990674991e-05, + "loss": 0.8434, + "step": 2816 + }, + { + "epoch": 1.5293159609120521, + "grad_norm": 6.956730702284698, + "learning_rate": 1.6268620218353188e-05, + "loss": 0.5463, + "step": 2817 + }, + { + "epoch": 1.5298588490770901, + "grad_norm": 6.516430093047641, + "learning_rate": 1.626587975464353e-05, + "loss": 0.4886, + "step": 2818 + }, + { + "epoch": 1.5304017372421281, + "grad_norm": 6.751333342823053, + "learning_rate": 1.626313851595987e-05, + "loss": 0.6279, + "step": 2819 + }, + { + "epoch": 1.5309446254071661, + "grad_norm": 8.120808986494415, + "learning_rate": 1.6260396502641264e-05, + "loss": 0.8192, + "step": 2820 + }, + { + "epoch": 1.5314875135722041, + "grad_norm": 7.570228095805488, + "learning_rate": 1.6257653715026837e-05, + "loss": 0.5813, + "step": 2821 + }, + { + "epoch": 1.5320304017372421, + "grad_norm": 8.436330907807918, + "learning_rate": 1.625491015345583e-05, + "loss": 0.9115, + "step": 2822 + }, + { + "epoch": 1.5325732899022801, + "grad_norm": 8.384048301051982, + "learning_rate": 1.6252165818267564e-05, + "loss": 1.0131, + "step": 2823 + }, + { + "epoch": 1.5331161780673181, + "grad_norm": 6.665532827172479, + "learning_rate": 1.6249420709801462e-05, + "loss": 0.5454, + "step": 2824 + }, + { + "epoch": 1.5336590662323561, + "grad_norm": 7.096056294730765, + "learning_rate": 1.624667482839705e-05, + "loss": 1.043, + "step": 2825 + }, + { + "epoch": 1.5342019543973942, + "grad_norm": 5.384161272714602, + "learning_rate": 1.6243928174393935e-05, + "loss": 0.8345, + "step": 2826 + }, + { + "epoch": 1.5347448425624322, + "grad_norm": 8.398527018753486, + "learning_rate": 1.6241180748131834e-05, + "loss": 0.6911, + "step": 2827 + }, + { + "epoch": 1.5352877307274702, + "grad_norm": 6.880924484125559, + "learning_rate": 1.6238432549950552e-05, + "loss": 0.6039, + "step": 2828 + }, + { + "epoch": 1.5358306188925082, + "grad_norm": 9.23669884568822, + "learning_rate": 1.623568358018999e-05, + "loss": 1.1145, + "step": 2829 + }, + { + "epoch": 1.5363735070575462, + "grad_norm": 9.838376476241027, + "learning_rate": 1.6232933839190146e-05, + "loss": 1.0435, + "step": 2830 + }, + { + "epoch": 1.5369163952225842, + "grad_norm": 8.64475205930121, + "learning_rate": 1.6230183327291108e-05, + "loss": 1.0886, + "step": 2831 + }, + { + "epoch": 1.5374592833876222, + "grad_norm": 8.446847258487077, + "learning_rate": 1.6227432044833072e-05, + "loss": 0.8922, + "step": 2832 + }, + { + "epoch": 1.5380021715526602, + "grad_norm": 7.373724851347831, + "learning_rate": 1.622467999215631e-05, + "loss": 0.7823, + "step": 2833 + }, + { + "epoch": 1.5385450597176982, + "grad_norm": 9.661226436708493, + "learning_rate": 1.622192716960121e-05, + "loss": 1.2651, + "step": 2834 + }, + { + "epoch": 1.5390879478827362, + "grad_norm": 9.316115050127708, + "learning_rate": 1.6219173577508237e-05, + "loss": 0.897, + "step": 2835 + }, + { + "epoch": 1.5396308360477742, + "grad_norm": 12.011440258586251, + "learning_rate": 1.621641921621797e-05, + "loss": 0.868, + "step": 2836 + }, + { + "epoch": 1.5401737242128122, + "grad_norm": 7.737580132404622, + "learning_rate": 1.6213664086071058e-05, + "loss": 1.0727, + "step": 2837 + }, + { + "epoch": 1.5407166123778502, + "grad_norm": 6.696935187657473, + "learning_rate": 1.6210908187408275e-05, + "loss": 0.7404, + "step": 2838 + }, + { + "epoch": 1.5412595005428882, + "grad_norm": 6.460846604904407, + "learning_rate": 1.6208151520570465e-05, + "loss": 0.5892, + "step": 2839 + }, + { + "epoch": 1.5418023887079262, + "grad_norm": 8.244871763303658, + "learning_rate": 1.6205394085898586e-05, + "loss": 1.0692, + "step": 2840 + }, + { + "epoch": 1.5423452768729642, + "grad_norm": 6.140143281113326, + "learning_rate": 1.620263588373367e-05, + "loss": 0.7648, + "step": 2841 + }, + { + "epoch": 1.5428881650380022, + "grad_norm": 6.3369705303866235, + "learning_rate": 1.619987691441687e-05, + "loss": 0.6665, + "step": 2842 + }, + { + "epoch": 1.5434310532030402, + "grad_norm": 8.864356355873657, + "learning_rate": 1.6197117178289405e-05, + "loss": 0.9855, + "step": 2843 + }, + { + "epoch": 1.5439739413680782, + "grad_norm": 8.487808912288623, + "learning_rate": 1.6194356675692614e-05, + "loss": 1.3813, + "step": 2844 + }, + { + "epoch": 1.5445168295331162, + "grad_norm": 6.5782618837581435, + "learning_rate": 1.619159540696792e-05, + "loss": 1.1941, + "step": 2845 + }, + { + "epoch": 1.5450597176981542, + "grad_norm": 8.668580644259926, + "learning_rate": 1.6188833372456833e-05, + "loss": 1.1044, + "step": 2846 + }, + { + "epoch": 1.5456026058631922, + "grad_norm": 8.677506532765584, + "learning_rate": 1.6186070572500972e-05, + "loss": 1.3984, + "step": 2847 + }, + { + "epoch": 1.5461454940282302, + "grad_norm": 6.822422879140709, + "learning_rate": 1.6183307007442046e-05, + "loss": 0.6903, + "step": 2848 + }, + { + "epoch": 1.5466883821932682, + "grad_norm": 8.480828664812217, + "learning_rate": 1.6180542677621852e-05, + "loss": 1.2466, + "step": 2849 + }, + { + "epoch": 1.5472312703583062, + "grad_norm": 7.804300661419004, + "learning_rate": 1.617777758338229e-05, + "loss": 0.8833, + "step": 2850 + }, + { + "epoch": 1.5477741585233442, + "grad_norm": 6.418809144902363, + "learning_rate": 1.617501172506535e-05, + "loss": 0.9174, + "step": 2851 + }, + { + "epoch": 1.5483170466883822, + "grad_norm": 6.627083812958508, + "learning_rate": 1.617224510301312e-05, + "loss": 0.7867, + "step": 2852 + }, + { + "epoch": 1.5488599348534202, + "grad_norm": 7.792114190076133, + "learning_rate": 1.616947771756778e-05, + "loss": 0.6242, + "step": 2853 + }, + { + "epoch": 1.5494028230184582, + "grad_norm": 6.965108246318713, + "learning_rate": 1.6166709569071598e-05, + "loss": 0.5844, + "step": 2854 + }, + { + "epoch": 1.5499457111834962, + "grad_norm": 6.9966373703206655, + "learning_rate": 1.616394065786695e-05, + "loss": 1.1688, + "step": 2855 + }, + { + "epoch": 1.5504885993485342, + "grad_norm": 8.466603441198636, + "learning_rate": 1.6161170984296298e-05, + "loss": 0.9506, + "step": 2856 + }, + { + "epoch": 1.5510314875135722, + "grad_norm": 6.361127259105106, + "learning_rate": 1.61584005487022e-05, + "loss": 0.5573, + "step": 2857 + }, + { + "epoch": 1.5515743756786102, + "grad_norm": 6.137955961198825, + "learning_rate": 1.6155629351427306e-05, + "loss": 1.0474, + "step": 2858 + }, + { + "epoch": 1.5521172638436482, + "grad_norm": 6.262211837195136, + "learning_rate": 1.6152857392814367e-05, + "loss": 0.6786, + "step": 2859 + }, + { + "epoch": 1.5526601520086862, + "grad_norm": 9.3980797837583, + "learning_rate": 1.6150084673206214e-05, + "loss": 1.2893, + "step": 2860 + }, + { + "epoch": 1.5532030401737242, + "grad_norm": 9.600537200283373, + "learning_rate": 1.614731119294579e-05, + "loss": 0.8668, + "step": 2861 + }, + { + "epoch": 1.5537459283387622, + "grad_norm": 8.65304941049683, + "learning_rate": 1.614453695237612e-05, + "loss": 1.0377, + "step": 2862 + }, + { + "epoch": 1.5542888165038002, + "grad_norm": 6.075747510405264, + "learning_rate": 1.6141761951840327e-05, + "loss": 0.8203, + "step": 2863 + }, + { + "epoch": 1.5548317046688382, + "grad_norm": 6.578770531031146, + "learning_rate": 1.6138986191681626e-05, + "loss": 0.6808, + "step": 2864 + }, + { + "epoch": 1.5553745928338762, + "grad_norm": 9.206515322081769, + "learning_rate": 1.6136209672243332e-05, + "loss": 0.9111, + "step": 2865 + }, + { + "epoch": 1.5559174809989142, + "grad_norm": 8.829723041453656, + "learning_rate": 1.613343239386884e-05, + "loss": 1.7392, + "step": 2866 + }, + { + "epoch": 1.5564603691639523, + "grad_norm": 9.489367814866748, + "learning_rate": 1.613065435690166e-05, + "loss": 0.9186, + "step": 2867 + }, + { + "epoch": 1.5570032573289903, + "grad_norm": 7.7196736231335175, + "learning_rate": 1.6127875561685376e-05, + "loss": 0.769, + "step": 2868 + }, + { + "epoch": 1.5575461454940283, + "grad_norm": 7.668511006453606, + "learning_rate": 1.6125096008563677e-05, + "loss": 0.7572, + "step": 2869 + }, + { + "epoch": 1.5580890336590663, + "grad_norm": 8.31551289254719, + "learning_rate": 1.6122315697880343e-05, + "loss": 1.0657, + "step": 2870 + }, + { + "epoch": 1.5586319218241043, + "grad_norm": 11.828690633927348, + "learning_rate": 1.6119534629979244e-05, + "loss": 1.0095, + "step": 2871 + }, + { + "epoch": 1.5591748099891423, + "grad_norm": 7.137823181609677, + "learning_rate": 1.611675280520435e-05, + "loss": 0.7246, + "step": 2872 + }, + { + "epoch": 1.5597176981541803, + "grad_norm": 5.051313327297567, + "learning_rate": 1.611397022389972e-05, + "loss": 0.5905, + "step": 2873 + }, + { + "epoch": 1.5602605863192183, + "grad_norm": 8.165805991671068, + "learning_rate": 1.6111186886409504e-05, + "loss": 1.0678, + "step": 2874 + }, + { + "epoch": 1.5608034744842563, + "grad_norm": 11.597649604311457, + "learning_rate": 1.6108402793077957e-05, + "loss": 1.4836, + "step": 2875 + }, + { + "epoch": 1.5613463626492943, + "grad_norm": 7.953203184026488, + "learning_rate": 1.610561794424942e-05, + "loss": 0.9216, + "step": 2876 + }, + { + "epoch": 1.5618892508143323, + "grad_norm": 7.397930684647979, + "learning_rate": 1.6102832340268322e-05, + "loss": 0.6688, + "step": 2877 + }, + { + "epoch": 1.5624321389793703, + "grad_norm": 8.999084750858366, + "learning_rate": 1.6100045981479195e-05, + "loss": 1.254, + "step": 2878 + }, + { + "epoch": 1.5629750271444083, + "grad_norm": 7.445594920489225, + "learning_rate": 1.6097258868226658e-05, + "loss": 0.6406, + "step": 2879 + }, + { + "epoch": 1.5635179153094463, + "grad_norm": 10.469355169581702, + "learning_rate": 1.609447100085543e-05, + "loss": 0.7928, + "step": 2880 + }, + { + "epoch": 1.5640608034744843, + "grad_norm": 8.853634937458189, + "learning_rate": 1.6091682379710313e-05, + "loss": 1.0788, + "step": 2881 + }, + { + "epoch": 1.5646036916395223, + "grad_norm": 7.909620213759866, + "learning_rate": 1.6088893005136206e-05, + "loss": 1.011, + "step": 2882 + }, + { + "epoch": 1.5651465798045603, + "grad_norm": 10.379252728708625, + "learning_rate": 1.6086102877478117e-05, + "loss": 0.8675, + "step": 2883 + }, + { + "epoch": 1.5656894679695983, + "grad_norm": 8.143888785034273, + "learning_rate": 1.6083311997081116e-05, + "loss": 0.8935, + "step": 2884 + }, + { + "epoch": 1.5662323561346363, + "grad_norm": 8.577520797942114, + "learning_rate": 1.6080520364290396e-05, + "loss": 0.6763, + "step": 2885 + }, + { + "epoch": 1.5667752442996743, + "grad_norm": 9.226094142154105, + "learning_rate": 1.6077727979451228e-05, + "loss": 0.6679, + "step": 2886 + }, + { + "epoch": 1.5673181324647123, + "grad_norm": 11.197536431412058, + "learning_rate": 1.607493484290897e-05, + "loss": 1.0525, + "step": 2887 + }, + { + "epoch": 1.5678610206297503, + "grad_norm": 8.43012225245798, + "learning_rate": 1.6072140955009093e-05, + "loss": 1.0317, + "step": 2888 + }, + { + "epoch": 1.5684039087947883, + "grad_norm": 6.69358812320454, + "learning_rate": 1.606934631609715e-05, + "loss": 0.7703, + "step": 2889 + }, + { + "epoch": 1.5689467969598263, + "grad_norm": 9.36327291463899, + "learning_rate": 1.6066550926518776e-05, + "loss": 1.0844, + "step": 2890 + }, + { + "epoch": 1.5694896851248643, + "grad_norm": 8.40518653624583, + "learning_rate": 1.6063754786619716e-05, + "loss": 0.6863, + "step": 2891 + }, + { + "epoch": 1.5700325732899023, + "grad_norm": 9.552486022227878, + "learning_rate": 1.60609578967458e-05, + "loss": 1.0342, + "step": 2892 + }, + { + "epoch": 1.5705754614549403, + "grad_norm": 6.76445711513327, + "learning_rate": 1.6058160257242953e-05, + "loss": 0.8777, + "step": 2893 + }, + { + "epoch": 1.5711183496199783, + "grad_norm": 9.040602694092291, + "learning_rate": 1.6055361868457188e-05, + "loss": 0.9634, + "step": 2894 + }, + { + "epoch": 1.5716612377850163, + "grad_norm": 7.00278619777266, + "learning_rate": 1.6052562730734614e-05, + "loss": 0.8995, + "step": 2895 + }, + { + "epoch": 1.5722041259500543, + "grad_norm": 8.533327622490077, + "learning_rate": 1.604976284442144e-05, + "loss": 1.0896, + "step": 2896 + }, + { + "epoch": 1.5727470141150923, + "grad_norm": 9.895218231842426, + "learning_rate": 1.6046962209863953e-05, + "loss": 0.9309, + "step": 2897 + }, + { + "epoch": 1.5732899022801303, + "grad_norm": 5.945823958044379, + "learning_rate": 1.604416082740854e-05, + "loss": 0.6512, + "step": 2898 + }, + { + "epoch": 1.5738327904451683, + "grad_norm": 10.822429137471252, + "learning_rate": 1.6041358697401687e-05, + "loss": 1.0744, + "step": 2899 + }, + { + "epoch": 1.5743756786102063, + "grad_norm": 7.728147258511819, + "learning_rate": 1.603855582018996e-05, + "loss": 0.9566, + "step": 2900 + }, + { + "epoch": 1.5749185667752443, + "grad_norm": 8.643748761880973, + "learning_rate": 1.603575219612003e-05, + "loss": 0.9188, + "step": 2901 + }, + { + "epoch": 1.5754614549402823, + "grad_norm": 9.357572097357403, + "learning_rate": 1.603294782553864e-05, + "loss": 0.9045, + "step": 2902 + }, + { + "epoch": 1.5760043431053203, + "grad_norm": 6.76207345266047, + "learning_rate": 1.6030142708792653e-05, + "loss": 0.7658, + "step": 2903 + }, + { + "epoch": 1.5765472312703583, + "grad_norm": 6.47603683733189, + "learning_rate": 1.6027336846229005e-05, + "loss": 0.5406, + "step": 2904 + }, + { + "epoch": 1.5770901194353963, + "grad_norm": 6.030033648190992, + "learning_rate": 1.602453023819473e-05, + "loss": 0.5141, + "step": 2905 + }, + { + "epoch": 1.5776330076004343, + "grad_norm": 8.344821553681937, + "learning_rate": 1.6021722885036954e-05, + "loss": 1.255, + "step": 2906 + }, + { + "epoch": 1.5781758957654723, + "grad_norm": 7.2392259468937885, + "learning_rate": 1.601891478710289e-05, + "loss": 0.8358, + "step": 2907 + }, + { + "epoch": 1.5787187839305103, + "grad_norm": 8.011046574978268, + "learning_rate": 1.6016105944739856e-05, + "loss": 1.0316, + "step": 2908 + }, + { + "epoch": 1.5792616720955484, + "grad_norm": 6.609140115647256, + "learning_rate": 1.601329635829525e-05, + "loss": 0.7324, + "step": 2909 + }, + { + "epoch": 1.5798045602605864, + "grad_norm": 6.672864395642082, + "learning_rate": 1.6010486028116568e-05, + "loss": 0.6386, + "step": 2910 + }, + { + "epoch": 1.5803474484256244, + "grad_norm": 7.270447246601273, + "learning_rate": 1.600767495455139e-05, + "loss": 0.6699, + "step": 2911 + }, + { + "epoch": 1.5808903365906624, + "grad_norm": 9.7186017734922, + "learning_rate": 1.6004863137947405e-05, + "loss": 0.9604, + "step": 2912 + }, + { + "epoch": 1.5814332247557004, + "grad_norm": 7.18305646407285, + "learning_rate": 1.6002050578652374e-05, + "loss": 0.7286, + "step": 2913 + }, + { + "epoch": 1.5819761129207384, + "grad_norm": 8.171408196684215, + "learning_rate": 1.5999237277014162e-05, + "loss": 0.9121, + "step": 2914 + }, + { + "epoch": 1.5825190010857764, + "grad_norm": 7.710673421134947, + "learning_rate": 1.599642323338072e-05, + "loss": 0.7287, + "step": 2915 + }, + { + "epoch": 1.5830618892508144, + "grad_norm": 8.264934044661224, + "learning_rate": 1.5993608448100095e-05, + "loss": 0.6537, + "step": 2916 + }, + { + "epoch": 1.5836047774158524, + "grad_norm": 7.874914961028911, + "learning_rate": 1.599079292152043e-05, + "loss": 0.6511, + "step": 2917 + }, + { + "epoch": 1.5841476655808904, + "grad_norm": 7.399378371785218, + "learning_rate": 1.5987976653989945e-05, + "loss": 0.6236, + "step": 2918 + }, + { + "epoch": 1.5846905537459284, + "grad_norm": 7.837764322959507, + "learning_rate": 1.5985159645856966e-05, + "loss": 0.8184, + "step": 2919 + }, + { + "epoch": 1.5852334419109664, + "grad_norm": 8.64272103945922, + "learning_rate": 1.5982341897469903e-05, + "loss": 0.9182, + "step": 2920 + }, + { + "epoch": 1.5857763300760044, + "grad_norm": 9.994158036669248, + "learning_rate": 1.5979523409177254e-05, + "loss": 1.3466, + "step": 2921 + }, + { + "epoch": 1.5863192182410424, + "grad_norm": 6.671268632879107, + "learning_rate": 1.5976704181327626e-05, + "loss": 0.7506, + "step": 2922 + }, + { + "epoch": 1.5868621064060804, + "grad_norm": 8.161895609948308, + "learning_rate": 1.59738842142697e-05, + "loss": 0.9683, + "step": 2923 + }, + { + "epoch": 1.5874049945711184, + "grad_norm": 7.236619510875819, + "learning_rate": 1.597106350835225e-05, + "loss": 0.798, + "step": 2924 + }, + { + "epoch": 1.5879478827361564, + "grad_norm": 10.263209688443808, + "learning_rate": 1.5968242063924152e-05, + "loss": 1.0799, + "step": 2925 + }, + { + "epoch": 1.5884907709011944, + "grad_norm": 9.433308260958457, + "learning_rate": 1.596541988133436e-05, + "loss": 0.8755, + "step": 2926 + }, + { + "epoch": 1.5890336590662324, + "grad_norm": 9.48995441422385, + "learning_rate": 1.5962596960931927e-05, + "loss": 1.4712, + "step": 2927 + }, + { + "epoch": 1.5895765472312704, + "grad_norm": 7.317297215230638, + "learning_rate": 1.5959773303066005e-05, + "loss": 1.0421, + "step": 2928 + }, + { + "epoch": 1.5901194353963084, + "grad_norm": 7.491426794362114, + "learning_rate": 1.595694890808582e-05, + "loss": 0.9468, + "step": 2929 + }, + { + "epoch": 1.5906623235613464, + "grad_norm": 7.320440852081977, + "learning_rate": 1.5954123776340702e-05, + "loss": 0.7459, + "step": 2930 + }, + { + "epoch": 1.5912052117263844, + "grad_norm": 8.334158307527314, + "learning_rate": 1.5951297908180062e-05, + "loss": 1.3182, + "step": 2931 + }, + { + "epoch": 1.5917480998914224, + "grad_norm": 7.02613629768153, + "learning_rate": 1.5948471303953418e-05, + "loss": 0.919, + "step": 2932 + }, + { + "epoch": 1.5922909880564604, + "grad_norm": 6.904443487786306, + "learning_rate": 1.594564396401036e-05, + "loss": 0.5737, + "step": 2933 + }, + { + "epoch": 1.5928338762214984, + "grad_norm": 8.041576521446402, + "learning_rate": 1.594281588870058e-05, + "loss": 0.7191, + "step": 2934 + }, + { + "epoch": 1.5933767643865364, + "grad_norm": 8.557472225459021, + "learning_rate": 1.5939987078373856e-05, + "loss": 0.8936, + "step": 2935 + }, + { + "epoch": 1.5939196525515744, + "grad_norm": 10.051674589973098, + "learning_rate": 1.5937157533380065e-05, + "loss": 1.1777, + "step": 2936 + }, + { + "epoch": 1.5944625407166124, + "grad_norm": 7.1580018473739235, + "learning_rate": 1.5934327254069167e-05, + "loss": 0.8628, + "step": 2937 + }, + { + "epoch": 1.5950054288816504, + "grad_norm": 8.689338133459966, + "learning_rate": 1.593149624079122e-05, + "loss": 1.1877, + "step": 2938 + }, + { + "epoch": 1.5955483170466884, + "grad_norm": 8.40904043481009, + "learning_rate": 1.5928664493896364e-05, + "loss": 0.9815, + "step": 2939 + }, + { + "epoch": 1.5960912052117264, + "grad_norm": 7.778305354024487, + "learning_rate": 1.5925832013734832e-05, + "loss": 0.8282, + "step": 2940 + }, + { + "epoch": 1.5966340933767644, + "grad_norm": 7.074281482828989, + "learning_rate": 1.5922998800656956e-05, + "loss": 0.8347, + "step": 2941 + }, + { + "epoch": 1.5971769815418024, + "grad_norm": 9.188840060350078, + "learning_rate": 1.5920164855013145e-05, + "loss": 0.8566, + "step": 2942 + }, + { + "epoch": 1.5977198697068404, + "grad_norm": 7.021116417649334, + "learning_rate": 1.591733017715391e-05, + "loss": 0.8367, + "step": 2943 + }, + { + "epoch": 1.5982627578718784, + "grad_norm": 8.039041768789302, + "learning_rate": 1.5914494767429846e-05, + "loss": 0.9889, + "step": 2944 + }, + { + "epoch": 1.5988056460369164, + "grad_norm": 6.252701385804125, + "learning_rate": 1.5911658626191645e-05, + "loss": 1.0478, + "step": 2945 + }, + { + "epoch": 1.5993485342019544, + "grad_norm": 8.223409826369243, + "learning_rate": 1.5908821753790083e-05, + "loss": 1.2458, + "step": 2946 + }, + { + "epoch": 1.5998914223669924, + "grad_norm": 8.972690403999096, + "learning_rate": 1.590598415057603e-05, + "loss": 0.8471, + "step": 2947 + }, + { + "epoch": 1.6004343105320304, + "grad_norm": 8.163032020074416, + "learning_rate": 1.5903145816900445e-05, + "loss": 0.8213, + "step": 2948 + }, + { + "epoch": 1.6009771986970684, + "grad_norm": 5.667868413023322, + "learning_rate": 1.5900306753114375e-05, + "loss": 0.8426, + "step": 2949 + }, + { + "epoch": 1.6015200868621065, + "grad_norm": 6.960744978365833, + "learning_rate": 1.5897466959568967e-05, + "loss": 0.7224, + "step": 2950 + }, + { + "epoch": 1.6020629750271445, + "grad_norm": 7.648210267430006, + "learning_rate": 1.589462643661544e-05, + "loss": 0.7451, + "step": 2951 + }, + { + "epoch": 1.6026058631921825, + "grad_norm": 5.9568596208122875, + "learning_rate": 1.5891785184605123e-05, + "loss": 0.5252, + "step": 2952 + }, + { + "epoch": 1.6031487513572205, + "grad_norm": 6.41676218105273, + "learning_rate": 1.5888943203889427e-05, + "loss": 0.6502, + "step": 2953 + }, + { + "epoch": 1.6036916395222585, + "grad_norm": 7.704871607753619, + "learning_rate": 1.5886100494819846e-05, + "loss": 1.1762, + "step": 2954 + }, + { + "epoch": 1.6042345276872965, + "grad_norm": 7.656219757403813, + "learning_rate": 1.5883257057747975e-05, + "loss": 0.8076, + "step": 2955 + }, + { + "epoch": 1.6047774158523345, + "grad_norm": 6.897030921112048, + "learning_rate": 1.58804128930255e-05, + "loss": 0.883, + "step": 2956 + }, + { + "epoch": 1.6053203040173725, + "grad_norm": 6.718724904599124, + "learning_rate": 1.5877568001004182e-05, + "loss": 0.8533, + "step": 2957 + }, + { + "epoch": 1.6058631921824105, + "grad_norm": 8.458861423777194, + "learning_rate": 1.5874722382035887e-05, + "loss": 1.2017, + "step": 2958 + }, + { + "epoch": 1.6064060803474485, + "grad_norm": 9.16622725543938, + "learning_rate": 1.5871876036472565e-05, + "loss": 1.0074, + "step": 2959 + }, + { + "epoch": 1.6069489685124865, + "grad_norm": 7.053786350768386, + "learning_rate": 1.5869028964666254e-05, + "loss": 0.8193, + "step": 2960 + }, + { + "epoch": 1.6074918566775245, + "grad_norm": 7.670976634760575, + "learning_rate": 1.5866181166969088e-05, + "loss": 0.8331, + "step": 2961 + }, + { + "epoch": 1.6080347448425625, + "grad_norm": 9.942186867470006, + "learning_rate": 1.586333264373329e-05, + "loss": 0.8672, + "step": 2962 + }, + { + "epoch": 1.6085776330076005, + "grad_norm": 7.744463789862724, + "learning_rate": 1.586048339531116e-05, + "loss": 0.8859, + "step": 2963 + }, + { + "epoch": 1.6091205211726385, + "grad_norm": 7.181553947846623, + "learning_rate": 1.5857633422055104e-05, + "loss": 1.0376, + "step": 2964 + }, + { + "epoch": 1.6096634093376765, + "grad_norm": 7.184586780377298, + "learning_rate": 1.5854782724317616e-05, + "loss": 0.694, + "step": 2965 + }, + { + "epoch": 1.6102062975027145, + "grad_norm": 6.9154974167720855, + "learning_rate": 1.5851931302451262e-05, + "loss": 0.8573, + "step": 2966 + }, + { + "epoch": 1.6107491856677525, + "grad_norm": 5.056944401416877, + "learning_rate": 1.5849079156808726e-05, + "loss": 0.4972, + "step": 2967 + }, + { + "epoch": 1.6112920738327905, + "grad_norm": 6.98635366168859, + "learning_rate": 1.584622628774275e-05, + "loss": 0.9002, + "step": 2968 + }, + { + "epoch": 1.6118349619978285, + "grad_norm": 9.22418255248469, + "learning_rate": 1.5843372695606196e-05, + "loss": 1.4143, + "step": 2969 + }, + { + "epoch": 1.6123778501628665, + "grad_norm": 9.49766654897752, + "learning_rate": 1.584051838075199e-05, + "loss": 1.0145, + "step": 2970 + }, + { + "epoch": 1.6129207383279045, + "grad_norm": 9.681135461557052, + "learning_rate": 1.5837663343533166e-05, + "loss": 0.8841, + "step": 2971 + }, + { + "epoch": 1.6134636264929425, + "grad_norm": 6.00661489729409, + "learning_rate": 1.583480758430283e-05, + "loss": 0.677, + "step": 2972 + }, + { + "epoch": 1.6140065146579805, + "grad_norm": 7.843885336692957, + "learning_rate": 1.5831951103414194e-05, + "loss": 0.4729, + "step": 2973 + }, + { + "epoch": 1.6145494028230185, + "grad_norm": 9.391971866087287, + "learning_rate": 1.5829093901220557e-05, + "loss": 1.0833, + "step": 2974 + }, + { + "epoch": 1.6150922909880565, + "grad_norm": 7.202123614778415, + "learning_rate": 1.582623597807529e-05, + "loss": 0.9755, + "step": 2975 + }, + { + "epoch": 1.6156351791530945, + "grad_norm": 8.050976157508988, + "learning_rate": 1.5823377334331875e-05, + "loss": 0.8981, + "step": 2976 + }, + { + "epoch": 1.6161780673181325, + "grad_norm": 6.409986462897389, + "learning_rate": 1.5820517970343867e-05, + "loss": 0.3994, + "step": 2977 + }, + { + "epoch": 1.6167209554831705, + "grad_norm": 6.001133896600598, + "learning_rate": 1.581765788646492e-05, + "loss": 0.4701, + "step": 2978 + }, + { + "epoch": 1.6172638436482085, + "grad_norm": 12.858193971885504, + "learning_rate": 1.581479708304878e-05, + "loss": 1.4327, + "step": 2979 + }, + { + "epoch": 1.6178067318132465, + "grad_norm": 9.559044098763156, + "learning_rate": 1.5811935560449262e-05, + "loss": 0.9058, + "step": 2980 + }, + { + "epoch": 1.6183496199782845, + "grad_norm": 8.041859101089647, + "learning_rate": 1.5809073319020293e-05, + "loss": 0.6251, + "step": 2981 + }, + { + "epoch": 1.6188925081433225, + "grad_norm": 7.413714357196275, + "learning_rate": 1.580621035911588e-05, + "loss": 0.7887, + "step": 2982 + }, + { + "epoch": 1.6194353963083605, + "grad_norm": 11.941805543277479, + "learning_rate": 1.5803346681090113e-05, + "loss": 1.2547, + "step": 2983 + }, + { + "epoch": 1.6199782844733985, + "grad_norm": 5.728003242952451, + "learning_rate": 1.580048228529718e-05, + "loss": 0.4941, + "step": 2984 + }, + { + "epoch": 1.6205211726384365, + "grad_norm": 8.67398306592389, + "learning_rate": 1.5797617172091354e-05, + "loss": 0.7758, + "step": 2985 + }, + { + "epoch": 1.6210640608034745, + "grad_norm": 6.198163559825324, + "learning_rate": 1.5794751341826996e-05, + "loss": 0.4601, + "step": 2986 + }, + { + "epoch": 1.6216069489685125, + "grad_norm": 8.257469675141675, + "learning_rate": 1.5791884794858557e-05, + "loss": 0.8679, + "step": 2987 + }, + { + "epoch": 1.6221498371335505, + "grad_norm": 12.953868836673164, + "learning_rate": 1.5789017531540575e-05, + "loss": 1.572, + "step": 2988 + }, + { + "epoch": 1.6226927252985885, + "grad_norm": 9.232767459675694, + "learning_rate": 1.5786149552227682e-05, + "loss": 0.7511, + "step": 2989 + }, + { + "epoch": 1.6232356134636265, + "grad_norm": 7.465499623540017, + "learning_rate": 1.5783280857274586e-05, + "loss": 0.5754, + "step": 2990 + }, + { + "epoch": 1.6237785016286646, + "grad_norm": 5.087952971939767, + "learning_rate": 1.5780411447036097e-05, + "loss": 0.532, + "step": 2991 + }, + { + "epoch": 1.6243213897937026, + "grad_norm": 11.064786965172743, + "learning_rate": 1.577754132186711e-05, + "loss": 0.7195, + "step": 2992 + }, + { + "epoch": 1.6248642779587406, + "grad_norm": 10.268274657097892, + "learning_rate": 1.57746704821226e-05, + "loss": 1.1258, + "step": 2993 + }, + { + "epoch": 1.6254071661237783, + "grad_norm": 7.650078703788778, + "learning_rate": 1.5771798928157645e-05, + "loss": 0.5949, + "step": 2994 + }, + { + "epoch": 1.6259500542888166, + "grad_norm": 9.330939436907839, + "learning_rate": 1.5768926660327396e-05, + "loss": 0.9243, + "step": 2995 + }, + { + "epoch": 1.6264929424538543, + "grad_norm": 11.566902464205237, + "learning_rate": 1.576605367898711e-05, + "loss": 1.5588, + "step": 2996 + }, + { + "epoch": 1.6270358306188926, + "grad_norm": 9.91827217636312, + "learning_rate": 1.576317998449211e-05, + "loss": 1.1418, + "step": 2997 + }, + { + "epoch": 1.6275787187839303, + "grad_norm": 5.842910545051031, + "learning_rate": 1.5760305577197824e-05, + "loss": 0.5064, + "step": 2998 + }, + { + "epoch": 1.6281216069489686, + "grad_norm": 7.20637691224012, + "learning_rate": 1.5757430457459765e-05, + "loss": 0.8129, + "step": 2999 + }, + { + "epoch": 1.6286644951140063, + "grad_norm": 7.5338669003432965, + "learning_rate": 1.5754554625633535e-05, + "loss": 0.7244, + "step": 3000 + }, + { + "epoch": 1.6292073832790446, + "grad_norm": 7.489311206471124, + "learning_rate": 1.5751678082074813e-05, + "loss": 1.1409, + "step": 3001 + }, + { + "epoch": 1.6297502714440824, + "grad_norm": 7.26299222123934, + "learning_rate": 1.574880082713938e-05, + "loss": 0.7061, + "step": 3002 + }, + { + "epoch": 1.6302931596091206, + "grad_norm": 6.292246912680897, + "learning_rate": 1.5745922861183095e-05, + "loss": 0.6383, + "step": 3003 + }, + { + "epoch": 1.6308360477741584, + "grad_norm": 7.483086031387996, + "learning_rate": 1.574304418456192e-05, + "loss": 0.8194, + "step": 3004 + }, + { + "epoch": 1.6313789359391966, + "grad_norm": 8.062451557552306, + "learning_rate": 1.5740164797631882e-05, + "loss": 0.9474, + "step": 3005 + }, + { + "epoch": 1.6319218241042344, + "grad_norm": 7.62904254588456, + "learning_rate": 1.5737284700749116e-05, + "loss": 1.2324, + "step": 3006 + }, + { + "epoch": 1.6324647122692726, + "grad_norm": 8.533361938490982, + "learning_rate": 1.573440389426983e-05, + "loss": 0.6182, + "step": 3007 + }, + { + "epoch": 1.6330076004343104, + "grad_norm": 7.0451736765954855, + "learning_rate": 1.5731522378550337e-05, + "loss": 0.7464, + "step": 3008 + }, + { + "epoch": 1.6335504885993486, + "grad_norm": 8.121336222871497, + "learning_rate": 1.572864015394702e-05, + "loss": 1.1087, + "step": 3009 + }, + { + "epoch": 1.6340933767643864, + "grad_norm": 7.2112540180194955, + "learning_rate": 1.5725757220816356e-05, + "loss": 0.7756, + "step": 3010 + }, + { + "epoch": 1.6346362649294246, + "grad_norm": 8.092755806422884, + "learning_rate": 1.5722873579514915e-05, + "loss": 1.1184, + "step": 3011 + }, + { + "epoch": 1.6351791530944624, + "grad_norm": 7.485157971844339, + "learning_rate": 1.5719989230399347e-05, + "loss": 0.6839, + "step": 3012 + }, + { + "epoch": 1.6357220412595006, + "grad_norm": 8.516980380341119, + "learning_rate": 1.5717104173826397e-05, + "loss": 0.948, + "step": 3013 + }, + { + "epoch": 1.6362649294245384, + "grad_norm": 7.602547351124631, + "learning_rate": 1.571421841015289e-05, + "loss": 0.8479, + "step": 3014 + }, + { + "epoch": 1.6368078175895766, + "grad_norm": 7.0470960880026245, + "learning_rate": 1.5711331939735744e-05, + "loss": 0.9014, + "step": 3015 + }, + { + "epoch": 1.6373507057546144, + "grad_norm": 6.760810209806174, + "learning_rate": 1.570844476293196e-05, + "loss": 0.9316, + "step": 3016 + }, + { + "epoch": 1.6378935939196526, + "grad_norm": 8.02053578699558, + "learning_rate": 1.570555688009863e-05, + "loss": 0.7837, + "step": 3017 + }, + { + "epoch": 1.6384364820846904, + "grad_norm": 4.624578610614632, + "learning_rate": 1.5702668291592936e-05, + "loss": 0.4706, + "step": 3018 + }, + { + "epoch": 1.6389793702497286, + "grad_norm": 8.011712265192347, + "learning_rate": 1.569977899777213e-05, + "loss": 1.0338, + "step": 3019 + }, + { + "epoch": 1.6395222584147664, + "grad_norm": 7.773921575637797, + "learning_rate": 1.569688899899358e-05, + "loss": 0.7512, + "step": 3020 + }, + { + "epoch": 1.6400651465798046, + "grad_norm": 12.19433508686472, + "learning_rate": 1.569399829561472e-05, + "loss": 1.3502, + "step": 3021 + }, + { + "epoch": 1.6406080347448424, + "grad_norm": 11.319107865931453, + "learning_rate": 1.569110688799307e-05, + "loss": 0.9996, + "step": 3022 + }, + { + "epoch": 1.6411509229098806, + "grad_norm": 7.1911299711947585, + "learning_rate": 1.5688214776486255e-05, + "loss": 0.9037, + "step": 3023 + }, + { + "epoch": 1.6416938110749184, + "grad_norm": 8.326843836043436, + "learning_rate": 1.5685321961451968e-05, + "loss": 1.1179, + "step": 3024 + }, + { + "epoch": 1.6422366992399566, + "grad_norm": 8.257534801333424, + "learning_rate": 1.5682428443248002e-05, + "loss": 0.7197, + "step": 3025 + }, + { + "epoch": 1.6427795874049944, + "grad_norm": 6.669062512754087, + "learning_rate": 1.567953422223223e-05, + "loss": 0.6292, + "step": 3026 + }, + { + "epoch": 1.6433224755700326, + "grad_norm": 7.99477616152165, + "learning_rate": 1.567663929876261e-05, + "loss": 0.926, + "step": 3027 + }, + { + "epoch": 1.6438653637350704, + "grad_norm": 5.3717184947082615, + "learning_rate": 1.56737436731972e-05, + "loss": 0.5378, + "step": 3028 + }, + { + "epoch": 1.6444082519001086, + "grad_norm": 6.522826762176589, + "learning_rate": 1.5670847345894125e-05, + "loss": 0.7393, + "step": 3029 + }, + { + "epoch": 1.6449511400651464, + "grad_norm": 9.503371738808127, + "learning_rate": 1.5667950317211612e-05, + "loss": 1.2727, + "step": 3030 + }, + { + "epoch": 1.6454940282301846, + "grad_norm": 8.208077816222845, + "learning_rate": 1.5665052587507974e-05, + "loss": 0.7032, + "step": 3031 + }, + { + "epoch": 1.6460369163952224, + "grad_norm": 9.101101018529372, + "learning_rate": 1.56621541571416e-05, + "loss": 0.6983, + "step": 3032 + }, + { + "epoch": 1.6465798045602607, + "grad_norm": 7.254188682380924, + "learning_rate": 1.565925502647098e-05, + "loss": 0.7196, + "step": 3033 + }, + { + "epoch": 1.6471226927252984, + "grad_norm": 7.837018641144827, + "learning_rate": 1.5656355195854676e-05, + "loss": 1.2035, + "step": 3034 + }, + { + "epoch": 1.6476655808903367, + "grad_norm": 8.949191823273495, + "learning_rate": 1.5653454665651344e-05, + "loss": 1.0169, + "step": 3035 + }, + { + "epoch": 1.6482084690553744, + "grad_norm": 8.166317934132502, + "learning_rate": 1.5650553436219732e-05, + "loss": 0.672, + "step": 3036 + }, + { + "epoch": 1.6487513572204127, + "grad_norm": 6.443983188038021, + "learning_rate": 1.564765150791866e-05, + "loss": 0.5552, + "step": 3037 + }, + { + "epoch": 1.6492942453854504, + "grad_norm": 10.403060348162544, + "learning_rate": 1.5644748881107057e-05, + "loss": 1.0709, + "step": 3038 + }, + { + "epoch": 1.6498371335504887, + "grad_norm": 8.319421686790951, + "learning_rate": 1.564184555614391e-05, + "loss": 0.7923, + "step": 3039 + }, + { + "epoch": 1.6503800217155264, + "grad_norm": 9.65474819234395, + "learning_rate": 1.5638941533388318e-05, + "loss": 0.8519, + "step": 3040 + }, + { + "epoch": 1.6509229098805647, + "grad_norm": 8.635959327913778, + "learning_rate": 1.5636036813199445e-05, + "loss": 0.8447, + "step": 3041 + }, + { + "epoch": 1.6514657980456025, + "grad_norm": 7.674155625456285, + "learning_rate": 1.563313139593656e-05, + "loss": 0.7205, + "step": 3042 + }, + { + "epoch": 1.6520086862106407, + "grad_norm": 7.249478120616782, + "learning_rate": 1.5630225281959003e-05, + "loss": 0.6632, + "step": 3043 + }, + { + "epoch": 1.6525515743756785, + "grad_norm": 8.90384172612564, + "learning_rate": 1.5627318471626208e-05, + "loss": 0.8243, + "step": 3044 + }, + { + "epoch": 1.6530944625407167, + "grad_norm": 8.274462585020828, + "learning_rate": 1.5624410965297703e-05, + "loss": 0.7381, + "step": 3045 + }, + { + "epoch": 1.6536373507057545, + "grad_norm": 7.266215459792981, + "learning_rate": 1.562150276333308e-05, + "loss": 0.4989, + "step": 3046 + }, + { + "epoch": 1.6541802388707927, + "grad_norm": 4.816150056433193, + "learning_rate": 1.5618593866092036e-05, + "loss": 0.5382, + "step": 3047 + }, + { + "epoch": 1.6547231270358305, + "grad_norm": 6.835417235425346, + "learning_rate": 1.561568427393435e-05, + "loss": 1.044, + "step": 3048 + }, + { + "epoch": 1.6552660152008687, + "grad_norm": 7.2539344971377435, + "learning_rate": 1.5612773987219885e-05, + "loss": 0.5901, + "step": 3049 + }, + { + "epoch": 1.6558089033659065, + "grad_norm": 11.003646663954257, + "learning_rate": 1.5609863006308586e-05, + "loss": 0.9407, + "step": 3050 + }, + { + "epoch": 1.6563517915309447, + "grad_norm": 8.268114443948154, + "learning_rate": 1.560695133156049e-05, + "loss": 1.0164, + "step": 3051 + }, + { + "epoch": 1.6568946796959825, + "grad_norm": 7.824779774533612, + "learning_rate": 1.5604038963335716e-05, + "loss": 0.7111, + "step": 3052 + }, + { + "epoch": 1.6574375678610207, + "grad_norm": 8.542948361290183, + "learning_rate": 1.560112590199447e-05, + "loss": 0.6501, + "step": 3053 + }, + { + "epoch": 1.6579804560260585, + "grad_norm": 7.85824626614946, + "learning_rate": 1.5598212147897047e-05, + "loss": 0.752, + "step": 3054 + }, + { + "epoch": 1.6585233441910967, + "grad_norm": 6.588621135611242, + "learning_rate": 1.559529770140382e-05, + "loss": 0.507, + "step": 3055 + }, + { + "epoch": 1.6590662323561345, + "grad_norm": 6.418764525989286, + "learning_rate": 1.559238256287526e-05, + "loss": 0.5947, + "step": 3056 + }, + { + "epoch": 1.6596091205211727, + "grad_norm": 7.681071515192806, + "learning_rate": 1.5589466732671913e-05, + "loss": 0.8854, + "step": 3057 + }, + { + "epoch": 1.6601520086862105, + "grad_norm": 5.918230957225734, + "learning_rate": 1.558655021115441e-05, + "loss": 0.5546, + "step": 3058 + }, + { + "epoch": 1.6606948968512487, + "grad_norm": 10.669803193569846, + "learning_rate": 1.5583632998683475e-05, + "loss": 1.1548, + "step": 3059 + }, + { + "epoch": 1.6612377850162865, + "grad_norm": 9.271579004183549, + "learning_rate": 1.558071509561991e-05, + "loss": 0.8049, + "step": 3060 + }, + { + "epoch": 1.6617806731813247, + "grad_norm": 8.337085337830533, + "learning_rate": 1.557779650232461e-05, + "loss": 0.954, + "step": 3061 + }, + { + "epoch": 1.6623235613463625, + "grad_norm": 11.145075141347627, + "learning_rate": 1.5574877219158543e-05, + "loss": 1.1592, + "step": 3062 + }, + { + "epoch": 1.6628664495114007, + "grad_norm": 13.953049521423079, + "learning_rate": 1.557195724648278e-05, + "loss": 1.5699, + "step": 3063 + }, + { + "epoch": 1.6634093376764385, + "grad_norm": 8.654584332416588, + "learning_rate": 1.5569036584658466e-05, + "loss": 1.0185, + "step": 3064 + }, + { + "epoch": 1.6639522258414767, + "grad_norm": 7.520260829640502, + "learning_rate": 1.556611523404683e-05, + "loss": 0.7898, + "step": 3065 + }, + { + "epoch": 1.6644951140065145, + "grad_norm": 6.7373805901465875, + "learning_rate": 1.5563193195009188e-05, + "loss": 0.8392, + "step": 3066 + }, + { + "epoch": 1.6650380021715527, + "grad_norm": 6.732598633253056, + "learning_rate": 1.556027046790695e-05, + "loss": 0.6672, + "step": 3067 + }, + { + "epoch": 1.6655808903365905, + "grad_norm": 9.217725100287737, + "learning_rate": 1.55573470531016e-05, + "loss": 0.9874, + "step": 3068 + }, + { + "epoch": 1.6661237785016287, + "grad_norm": 6.963043776606882, + "learning_rate": 1.5554422950954706e-05, + "loss": 0.812, + "step": 3069 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 6.449074540999346, + "learning_rate": 1.5551498161827928e-05, + "loss": 0.5035, + "step": 3070 + }, + { + "epoch": 1.6672095548317047, + "grad_norm": 7.793693453975787, + "learning_rate": 1.554857268608301e-05, + "loss": 0.7974, + "step": 3071 + }, + { + "epoch": 1.6677524429967425, + "grad_norm": 10.124707886106192, + "learning_rate": 1.554564652408178e-05, + "loss": 1.4946, + "step": 3072 + }, + { + "epoch": 1.6682953311617807, + "grad_norm": 5.481775138355068, + "learning_rate": 1.5542719676186153e-05, + "loss": 0.502, + "step": 3073 + }, + { + "epoch": 1.6688382193268185, + "grad_norm": 6.055974601989459, + "learning_rate": 1.5539792142758114e-05, + "loss": 0.7463, + "step": 3074 + }, + { + "epoch": 1.6693811074918568, + "grad_norm": 7.452121005302326, + "learning_rate": 1.5536863924159762e-05, + "loss": 0.7282, + "step": 3075 + }, + { + "epoch": 1.6699239956568945, + "grad_norm": 8.991453230181806, + "learning_rate": 1.5533935020753252e-05, + "loss": 0.9327, + "step": 3076 + }, + { + "epoch": 1.6704668838219328, + "grad_norm": 8.138196256677595, + "learning_rate": 1.5531005432900838e-05, + "loss": 0.658, + "step": 3077 + }, + { + "epoch": 1.6710097719869705, + "grad_norm": 6.674313540160648, + "learning_rate": 1.552807516096486e-05, + "loss": 0.569, + "step": 3078 + }, + { + "epoch": 1.6715526601520088, + "grad_norm": 7.2303438904479105, + "learning_rate": 1.5525144205307734e-05, + "loss": 0.5186, + "step": 3079 + }, + { + "epoch": 1.6720955483170465, + "grad_norm": 6.872945037718312, + "learning_rate": 1.5522212566291966e-05, + "loss": 0.7282, + "step": 3080 + }, + { + "epoch": 1.6726384364820848, + "grad_norm": 7.913543579080126, + "learning_rate": 1.5519280244280146e-05, + "loss": 0.7015, + "step": 3081 + }, + { + "epoch": 1.6731813246471225, + "grad_norm": 6.144062932600708, + "learning_rate": 1.551634723963495e-05, + "loss": 0.7829, + "step": 3082 + }, + { + "epoch": 1.6737242128121608, + "grad_norm": 7.111349951384809, + "learning_rate": 1.5513413552719137e-05, + "loss": 0.6904, + "step": 3083 + }, + { + "epoch": 1.6742671009771986, + "grad_norm": 10.453360973997349, + "learning_rate": 1.5510479183895545e-05, + "loss": 1.4584, + "step": 3084 + }, + { + "epoch": 1.6748099891422368, + "grad_norm": 13.976555611119428, + "learning_rate": 1.5507544133527108e-05, + "loss": 1.542, + "step": 3085 + }, + { + "epoch": 1.6753528773072746, + "grad_norm": 5.894489690173581, + "learning_rate": 1.5504608401976835e-05, + "loss": 0.6792, + "step": 3086 + }, + { + "epoch": 1.6758957654723128, + "grad_norm": 8.937162315762803, + "learning_rate": 1.550167198960782e-05, + "loss": 0.8243, + "step": 3087 + }, + { + "epoch": 1.6764386536373506, + "grad_norm": 9.787723535687237, + "learning_rate": 1.549873489678325e-05, + "loss": 1.0071, + "step": 3088 + }, + { + "epoch": 1.6769815418023888, + "grad_norm": 6.0371827348978755, + "learning_rate": 1.5495797123866378e-05, + "loss": 0.6407, + "step": 3089 + }, + { + "epoch": 1.6775244299674266, + "grad_norm": 5.997673025227573, + "learning_rate": 1.5492858671220563e-05, + "loss": 0.7129, + "step": 3090 + }, + { + "epoch": 1.6780673181324648, + "grad_norm": 11.08813931574187, + "learning_rate": 1.5489919539209232e-05, + "loss": 0.8435, + "step": 3091 + }, + { + "epoch": 1.6786102062975026, + "grad_norm": 8.730951161091724, + "learning_rate": 1.54869797281959e-05, + "loss": 0.9159, + "step": 3092 + }, + { + "epoch": 1.6791530944625408, + "grad_norm": 7.293119927853279, + "learning_rate": 1.5484039238544173e-05, + "loss": 1.1001, + "step": 3093 + }, + { + "epoch": 1.6796959826275786, + "grad_norm": 7.091810627268002, + "learning_rate": 1.5481098070617734e-05, + "loss": 0.747, + "step": 3094 + }, + { + "epoch": 1.6802388707926168, + "grad_norm": 9.079124871383598, + "learning_rate": 1.547815622478035e-05, + "loss": 1.2292, + "step": 3095 + }, + { + "epoch": 1.6807817589576546, + "grad_norm": 7.238161650182196, + "learning_rate": 1.5475213701395867e-05, + "loss": 0.6657, + "step": 3096 + }, + { + "epoch": 1.6813246471226928, + "grad_norm": 9.385885763398004, + "learning_rate": 1.5472270500828236e-05, + "loss": 1.305, + "step": 3097 + }, + { + "epoch": 1.6818675352877306, + "grad_norm": 8.98902696621929, + "learning_rate": 1.5469326623441463e-05, + "loss": 0.8641, + "step": 3098 + }, + { + "epoch": 1.6824104234527688, + "grad_norm": 6.519773416264212, + "learning_rate": 1.5466382069599656e-05, + "loss": 0.4567, + "step": 3099 + }, + { + "epoch": 1.6829533116178066, + "grad_norm": 6.775917495491388, + "learning_rate": 1.5463436839667007e-05, + "loss": 0.6874, + "step": 3100 + }, + { + "epoch": 1.6834961997828448, + "grad_norm": 7.82569199258335, + "learning_rate": 1.5460490934007776e-05, + "loss": 0.819, + "step": 3101 + }, + { + "epoch": 1.6840390879478826, + "grad_norm": 6.885216490352139, + "learning_rate": 1.5457544352986326e-05, + "loss": 0.7891, + "step": 3102 + }, + { + "epoch": 1.6845819761129208, + "grad_norm": 6.733247427994312, + "learning_rate": 1.5454597096967093e-05, + "loss": 0.6664, + "step": 3103 + }, + { + "epoch": 1.6851248642779586, + "grad_norm": 6.96769902934235, + "learning_rate": 1.5451649166314598e-05, + "loss": 0.6578, + "step": 3104 + }, + { + "epoch": 1.6856677524429968, + "grad_norm": 6.941713803652074, + "learning_rate": 1.5448700561393444e-05, + "loss": 0.5855, + "step": 3105 + }, + { + "epoch": 1.6862106406080346, + "grad_norm": 8.203921934361416, + "learning_rate": 1.5445751282568324e-05, + "loss": 0.9843, + "step": 3106 + }, + { + "epoch": 1.6867535287730728, + "grad_norm": 8.822160140984163, + "learning_rate": 1.5442801330204004e-05, + "loss": 1.1003, + "step": 3107 + }, + { + "epoch": 1.6872964169381106, + "grad_norm": 8.330851912949297, + "learning_rate": 1.5439850704665338e-05, + "loss": 1.0992, + "step": 3108 + }, + { + "epoch": 1.6878393051031488, + "grad_norm": 6.552171568470756, + "learning_rate": 1.543689940631727e-05, + "loss": 0.5434, + "step": 3109 + }, + { + "epoch": 1.6883821932681866, + "grad_norm": 5.772049074992556, + "learning_rate": 1.5433947435524822e-05, + "loss": 0.7515, + "step": 3110 + }, + { + "epoch": 1.6889250814332248, + "grad_norm": 9.575440058135785, + "learning_rate": 1.543099479265309e-05, + "loss": 0.9691, + "step": 3111 + }, + { + "epoch": 1.6894679695982626, + "grad_norm": 6.008162314259628, + "learning_rate": 1.5428041478067263e-05, + "loss": 0.6137, + "step": 3112 + }, + { + "epoch": 1.6900108577633008, + "grad_norm": 8.993171311918207, + "learning_rate": 1.542508749213262e-05, + "loss": 0.7608, + "step": 3113 + }, + { + "epoch": 1.6905537459283386, + "grad_norm": 8.042583538721042, + "learning_rate": 1.542213283521451e-05, + "loss": 0.8188, + "step": 3114 + }, + { + "epoch": 1.6910966340933768, + "grad_norm": 9.290959640283983, + "learning_rate": 1.541917750767837e-05, + "loss": 1.1737, + "step": 3115 + }, + { + "epoch": 1.6916395222584146, + "grad_norm": 6.3883435583657535, + "learning_rate": 1.5416221509889718e-05, + "loss": 0.4376, + "step": 3116 + }, + { + "epoch": 1.6921824104234529, + "grad_norm": 8.943225081508276, + "learning_rate": 1.541326484221416e-05, + "loss": 0.9983, + "step": 3117 + }, + { + "epoch": 1.6927252985884906, + "grad_norm": 7.583355336491315, + "learning_rate": 1.541030750501737e-05, + "loss": 0.6078, + "step": 3118 + }, + { + "epoch": 1.6932681867535289, + "grad_norm": 7.318011543277784, + "learning_rate": 1.5407349498665133e-05, + "loss": 0.745, + "step": 3119 + }, + { + "epoch": 1.6938110749185666, + "grad_norm": 6.212465291912028, + "learning_rate": 1.5404390823523287e-05, + "loss": 0.4739, + "step": 3120 + }, + { + "epoch": 1.6943539630836049, + "grad_norm": 8.223585888775384, + "learning_rate": 1.5401431479957775e-05, + "loss": 1.1819, + "step": 3121 + }, + { + "epoch": 1.6948968512486426, + "grad_norm": 7.957232407413796, + "learning_rate": 1.5398471468334605e-05, + "loss": 0.7071, + "step": 3122 + }, + { + "epoch": 1.6954397394136809, + "grad_norm": 5.52597775233848, + "learning_rate": 1.5395510789019884e-05, + "loss": 0.6171, + "step": 3123 + }, + { + "epoch": 1.6959826275787186, + "grad_norm": 8.015216509333372, + "learning_rate": 1.5392549442379785e-05, + "loss": 0.9051, + "step": 3124 + }, + { + "epoch": 1.6965255157437569, + "grad_norm": 8.330470177699917, + "learning_rate": 1.538958742878058e-05, + "loss": 0.5037, + "step": 3125 + }, + { + "epoch": 1.6970684039087947, + "grad_norm": 6.1063862074204245, + "learning_rate": 1.538662474858861e-05, + "loss": 0.6509, + "step": 3126 + }, + { + "epoch": 1.6976112920738329, + "grad_norm": 8.11378323370098, + "learning_rate": 1.5383661402170308e-05, + "loss": 1.0356, + "step": 3127 + }, + { + "epoch": 1.6981541802388707, + "grad_norm": 5.995828293111682, + "learning_rate": 1.5380697389892185e-05, + "loss": 0.5297, + "step": 3128 + }, + { + "epoch": 1.6986970684039089, + "grad_norm": 6.962108385144619, + "learning_rate": 1.537773271212083e-05, + "loss": 0.8711, + "step": 3129 + }, + { + "epoch": 1.6992399565689467, + "grad_norm": 8.69611973474616, + "learning_rate": 1.5374767369222922e-05, + "loss": 0.911, + "step": 3130 + }, + { + "epoch": 1.6997828447339849, + "grad_norm": 11.047832664292853, + "learning_rate": 1.5371801361565223e-05, + "loss": 0.8499, + "step": 3131 + }, + { + "epoch": 1.7003257328990227, + "grad_norm": 10.490928296052365, + "learning_rate": 1.5368834689514568e-05, + "loss": 0.7617, + "step": 3132 + }, + { + "epoch": 1.700868621064061, + "grad_norm": 8.723324926351989, + "learning_rate": 1.536586735343788e-05, + "loss": 0.7158, + "step": 3133 + }, + { + "epoch": 1.7014115092290987, + "grad_norm": 7.780646168464086, + "learning_rate": 1.536289935370217e-05, + "loss": 0.6452, + "step": 3134 + }, + { + "epoch": 1.701954397394137, + "grad_norm": 8.037748179528911, + "learning_rate": 1.5359930690674518e-05, + "loss": 0.8368, + "step": 3135 + }, + { + "epoch": 1.7024972855591747, + "grad_norm": 7.539822669872962, + "learning_rate": 1.5356961364722096e-05, + "loss": 0.7224, + "step": 3136 + }, + { + "epoch": 1.703040173724213, + "grad_norm": 10.929660828744602, + "learning_rate": 1.5353991376212155e-05, + "loss": 1.2459, + "step": 3137 + }, + { + "epoch": 1.7035830618892507, + "grad_norm": 11.118929306326729, + "learning_rate": 1.5351020725512028e-05, + "loss": 1.2769, + "step": 3138 + }, + { + "epoch": 1.704125950054289, + "grad_norm": 8.820097629927263, + "learning_rate": 1.534804941298913e-05, + "loss": 1.0722, + "step": 3139 + }, + { + "epoch": 1.7046688382193267, + "grad_norm": 8.79991079654506, + "learning_rate": 1.5345077439010956e-05, + "loss": 0.7596, + "step": 3140 + }, + { + "epoch": 1.705211726384365, + "grad_norm": 7.5009896900263096, + "learning_rate": 1.5342104803945087e-05, + "loss": 0.6457, + "step": 3141 + }, + { + "epoch": 1.7057546145494027, + "grad_norm": 9.190121352339093, + "learning_rate": 1.533913150815918e-05, + "loss": 0.7152, + "step": 3142 + }, + { + "epoch": 1.706297502714441, + "grad_norm": 7.209026897095485, + "learning_rate": 1.5336157552020977e-05, + "loss": 0.534, + "step": 3143 + }, + { + "epoch": 1.7068403908794787, + "grad_norm": 8.752247270964865, + "learning_rate": 1.5333182935898306e-05, + "loss": 1.0953, + "step": 3144 + }, + { + "epoch": 1.707383279044517, + "grad_norm": 8.714563086288697, + "learning_rate": 1.5330207660159068e-05, + "loss": 0.8016, + "step": 3145 + }, + { + "epoch": 1.7079261672095547, + "grad_norm": 9.540320185420228, + "learning_rate": 1.5327231725171255e-05, + "loss": 1.2679, + "step": 3146 + }, + { + "epoch": 1.708469055374593, + "grad_norm": 8.091263210809633, + "learning_rate": 1.532425513130293e-05, + "loss": 0.7521, + "step": 3147 + }, + { + "epoch": 1.7090119435396307, + "grad_norm": 10.062104841108379, + "learning_rate": 1.5321277878922246e-05, + "loss": 1.0696, + "step": 3148 + }, + { + "epoch": 1.709554831704669, + "grad_norm": 8.184716628635337, + "learning_rate": 1.531829996839743e-05, + "loss": 0.643, + "step": 3149 + }, + { + "epoch": 1.7100977198697067, + "grad_norm": 11.130045309860368, + "learning_rate": 1.53153214000968e-05, + "loss": 1.2565, + "step": 3150 + }, + { + "epoch": 1.710640608034745, + "grad_norm": 8.856470164658607, + "learning_rate": 1.5312342174388746e-05, + "loss": 0.8376, + "step": 3151 + }, + { + "epoch": 1.7111834961997827, + "grad_norm": 11.958863336707164, + "learning_rate": 1.5309362291641747e-05, + "loss": 0.978, + "step": 3152 + }, + { + "epoch": 1.711726384364821, + "grad_norm": 10.369392130776264, + "learning_rate": 1.5306381752224357e-05, + "loss": 1.1822, + "step": 3153 + }, + { + "epoch": 1.7122692725298587, + "grad_norm": 7.346196971786586, + "learning_rate": 1.5303400556505213e-05, + "loss": 0.9284, + "step": 3154 + }, + { + "epoch": 1.712812160694897, + "grad_norm": 6.652184209108613, + "learning_rate": 1.5300418704853042e-05, + "loss": 0.5787, + "step": 3155 + }, + { + "epoch": 1.7133550488599347, + "grad_norm": 9.745597378440513, + "learning_rate": 1.5297436197636634e-05, + "loss": 1.0194, + "step": 3156 + }, + { + "epoch": 1.713897937024973, + "grad_norm": 6.17200933356452, + "learning_rate": 1.5294453035224874e-05, + "loss": 0.6315, + "step": 3157 + }, + { + "epoch": 1.7144408251900107, + "grad_norm": 8.373044157038668, + "learning_rate": 1.5291469217986724e-05, + "loss": 0.9127, + "step": 3158 + }, + { + "epoch": 1.714983713355049, + "grad_norm": 9.811623309823384, + "learning_rate": 1.5288484746291227e-05, + "loss": 0.7791, + "step": 3159 + }, + { + "epoch": 1.7155266015200867, + "grad_norm": 9.264521529810194, + "learning_rate": 1.5285499620507513e-05, + "loss": 0.6764, + "step": 3160 + }, + { + "epoch": 1.716069489685125, + "grad_norm": 8.688562646594697, + "learning_rate": 1.5282513841004777e-05, + "loss": 1.0202, + "step": 3161 + }, + { + "epoch": 1.7166123778501627, + "grad_norm": 8.522452395448054, + "learning_rate": 1.527952740815231e-05, + "loss": 1.0043, + "step": 3162 + }, + { + "epoch": 1.717155266015201, + "grad_norm": 6.330008909603884, + "learning_rate": 1.527654032231948e-05, + "loss": 0.6651, + "step": 3163 + }, + { + "epoch": 1.7176981541802387, + "grad_norm": 9.314495623946685, + "learning_rate": 1.5273552583875736e-05, + "loss": 0.8312, + "step": 3164 + }, + { + "epoch": 1.718241042345277, + "grad_norm": 9.157480690732486, + "learning_rate": 1.52705641931906e-05, + "loss": 1.0017, + "step": 3165 + }, + { + "epoch": 1.7187839305103148, + "grad_norm": 8.300251884517943, + "learning_rate": 1.5267575150633687e-05, + "loss": 0.9929, + "step": 3166 + }, + { + "epoch": 1.719326818675353, + "grad_norm": 8.328644452020212, + "learning_rate": 1.5264585456574684e-05, + "loss": 0.8105, + "step": 3167 + }, + { + "epoch": 1.7198697068403908, + "grad_norm": 9.499775808882656, + "learning_rate": 1.526159511138336e-05, + "loss": 0.8593, + "step": 3168 + }, + { + "epoch": 1.720412595005429, + "grad_norm": 7.684886832360556, + "learning_rate": 1.5258604115429567e-05, + "loss": 0.5919, + "step": 3169 + }, + { + "epoch": 1.7209554831704668, + "grad_norm": 8.666869336691208, + "learning_rate": 1.5255612469083239e-05, + "loss": 0.6537, + "step": 3170 + }, + { + "epoch": 1.721498371335505, + "grad_norm": 11.871617471990923, + "learning_rate": 1.5252620172714378e-05, + "loss": 1.3826, + "step": 3171 + }, + { + "epoch": 1.7220412595005428, + "grad_norm": 6.088995555337644, + "learning_rate": 1.5249627226693089e-05, + "loss": 0.5085, + "step": 3172 + }, + { + "epoch": 1.722584147665581, + "grad_norm": 9.255760547122362, + "learning_rate": 1.5246633631389536e-05, + "loss": 0.8823, + "step": 3173 + }, + { + "epoch": 1.7231270358306188, + "grad_norm": 9.91965748862537, + "learning_rate": 1.5243639387173974e-05, + "loss": 0.8752, + "step": 3174 + }, + { + "epoch": 1.723669923995657, + "grad_norm": 6.397561552165108, + "learning_rate": 1.5240644494416734e-05, + "loss": 0.6673, + "step": 3175 + }, + { + "epoch": 1.7242128121606948, + "grad_norm": 7.066817871358342, + "learning_rate": 1.523764895348823e-05, + "loss": 1.041, + "step": 3176 + }, + { + "epoch": 1.724755700325733, + "grad_norm": 8.282830238008364, + "learning_rate": 1.5234652764758959e-05, + "loss": 0.9521, + "step": 3177 + }, + { + "epoch": 1.7252985884907708, + "grad_norm": 7.611547937947443, + "learning_rate": 1.523165592859949e-05, + "loss": 0.7182, + "step": 3178 + }, + { + "epoch": 1.725841476655809, + "grad_norm": 8.068698205768472, + "learning_rate": 1.5228658445380475e-05, + "loss": 0.9078, + "step": 3179 + }, + { + "epoch": 1.7263843648208468, + "grad_norm": 8.731772184014808, + "learning_rate": 1.5225660315472652e-05, + "loss": 1.2148, + "step": 3180 + }, + { + "epoch": 1.726927252985885, + "grad_norm": 7.785982141134894, + "learning_rate": 1.5222661539246832e-05, + "loss": 0.8653, + "step": 3181 + }, + { + "epoch": 1.7274701411509228, + "grad_norm": 7.6267653051373285, + "learning_rate": 1.521966211707391e-05, + "loss": 0.7454, + "step": 3182 + }, + { + "epoch": 1.728013029315961, + "grad_norm": 7.440711922406893, + "learning_rate": 1.521666204932486e-05, + "loss": 0.6576, + "step": 3183 + }, + { + "epoch": 1.7285559174809988, + "grad_norm": 7.69035147644333, + "learning_rate": 1.521366133637073e-05, + "loss": 0.8269, + "step": 3184 + }, + { + "epoch": 1.729098805646037, + "grad_norm": 8.387844627258719, + "learning_rate": 1.5210659978582662e-05, + "loss": 1.1049, + "step": 3185 + }, + { + "epoch": 1.7296416938110748, + "grad_norm": 9.08085381726355, + "learning_rate": 1.5207657976331862e-05, + "loss": 0.97, + "step": 3186 + }, + { + "epoch": 1.730184581976113, + "grad_norm": 7.621613071814458, + "learning_rate": 1.5204655329989617e-05, + "loss": 0.7113, + "step": 3187 + }, + { + "epoch": 1.7307274701411508, + "grad_norm": 6.743838306403499, + "learning_rate": 1.5201652039927313e-05, + "loss": 0.7307, + "step": 3188 + }, + { + "epoch": 1.731270358306189, + "grad_norm": 9.602658786805504, + "learning_rate": 1.5198648106516392e-05, + "loss": 1.0515, + "step": 3189 + }, + { + "epoch": 1.7318132464712268, + "grad_norm": 8.153765987898234, + "learning_rate": 1.5195643530128387e-05, + "loss": 1.0879, + "step": 3190 + }, + { + "epoch": 1.732356134636265, + "grad_norm": 9.745244927474419, + "learning_rate": 1.519263831113491e-05, + "loss": 0.9542, + "step": 3191 + }, + { + "epoch": 1.7328990228013028, + "grad_norm": 8.908425729214914, + "learning_rate": 1.5189632449907654e-05, + "loss": 0.8936, + "step": 3192 + }, + { + "epoch": 1.733441910966341, + "grad_norm": 7.409065356665995, + "learning_rate": 1.5186625946818382e-05, + "loss": 0.7645, + "step": 3193 + }, + { + "epoch": 1.7339847991313788, + "grad_norm": 7.958784484679066, + "learning_rate": 1.5183618802238949e-05, + "loss": 0.8512, + "step": 3194 + }, + { + "epoch": 1.734527687296417, + "grad_norm": 5.808082449761269, + "learning_rate": 1.5180611016541278e-05, + "loss": 0.7241, + "step": 3195 + }, + { + "epoch": 1.7350705754614548, + "grad_norm": 8.88976614296984, + "learning_rate": 1.5177602590097382e-05, + "loss": 1.0506, + "step": 3196 + }, + { + "epoch": 1.735613463626493, + "grad_norm": 8.197925091983134, + "learning_rate": 1.5174593523279346e-05, + "loss": 0.7959, + "step": 3197 + }, + { + "epoch": 1.7361563517915308, + "grad_norm": 10.2752921216035, + "learning_rate": 1.5171583816459334e-05, + "loss": 1.2859, + "step": 3198 + }, + { + "epoch": 1.736699239956569, + "grad_norm": 9.43498245507836, + "learning_rate": 1.5168573470009596e-05, + "loss": 1.0459, + "step": 3199 + }, + { + "epoch": 1.7372421281216068, + "grad_norm": 9.033995794822129, + "learning_rate": 1.516556248430245e-05, + "loss": 1.0003, + "step": 3200 + }, + { + "epoch": 1.737785016286645, + "grad_norm": 9.577836032406582, + "learning_rate": 1.5162550859710306e-05, + "loss": 1.2394, + "step": 3201 + }, + { + "epoch": 1.7383279044516828, + "grad_norm": 5.2063230201693775, + "learning_rate": 1.5159538596605642e-05, + "loss": 0.3625, + "step": 3202 + }, + { + "epoch": 1.738870792616721, + "grad_norm": 7.741298457702819, + "learning_rate": 1.5156525695361022e-05, + "loss": 0.6744, + "step": 3203 + }, + { + "epoch": 1.7394136807817588, + "grad_norm": 8.718382461609222, + "learning_rate": 1.5153512156349086e-05, + "loss": 1.0423, + "step": 3204 + }, + { + "epoch": 1.739956568946797, + "grad_norm": 6.706656901306365, + "learning_rate": 1.5150497979942556e-05, + "loss": 0.5121, + "step": 3205 + }, + { + "epoch": 1.7404994571118348, + "grad_norm": 9.934397785927217, + "learning_rate": 1.5147483166514225e-05, + "loss": 1.5269, + "step": 3206 + }, + { + "epoch": 1.741042345276873, + "grad_norm": 8.069224944733923, + "learning_rate": 1.514446771643697e-05, + "loss": 0.7801, + "step": 3207 + }, + { + "epoch": 1.7415852334419109, + "grad_norm": 9.193161021474104, + "learning_rate": 1.514145163008375e-05, + "loss": 1.2144, + "step": 3208 + }, + { + "epoch": 1.742128121606949, + "grad_norm": 6.5647281089901, + "learning_rate": 1.5138434907827599e-05, + "loss": 0.6692, + "step": 3209 + }, + { + "epoch": 1.7426710097719869, + "grad_norm": 9.387942500806554, + "learning_rate": 1.513541755004163e-05, + "loss": 1.1655, + "step": 3210 + }, + { + "epoch": 1.743213897937025, + "grad_norm": 6.936398185140981, + "learning_rate": 1.5132399557099031e-05, + "loss": 0.7761, + "step": 3211 + }, + { + "epoch": 1.7437567861020629, + "grad_norm": 8.699280554503527, + "learning_rate": 1.5129380929373076e-05, + "loss": 1.2891, + "step": 3212 + }, + { + "epoch": 1.744299674267101, + "grad_norm": 11.932192326740894, + "learning_rate": 1.5126361667237116e-05, + "loss": 0.9868, + "step": 3213 + }, + { + "epoch": 1.7448425624321389, + "grad_norm": 8.456694582402795, + "learning_rate": 1.512334177106457e-05, + "loss": 1.199, + "step": 3214 + }, + { + "epoch": 1.745385450597177, + "grad_norm": 8.284979161281417, + "learning_rate": 1.512032124122895e-05, + "loss": 0.658, + "step": 3215 + }, + { + "epoch": 1.7459283387622149, + "grad_norm": 8.490755188379255, + "learning_rate": 1.5117300078103841e-05, + "loss": 0.6974, + "step": 3216 + }, + { + "epoch": 1.746471226927253, + "grad_norm": 14.621603534181931, + "learning_rate": 1.5114278282062898e-05, + "loss": 1.4386, + "step": 3217 + }, + { + "epoch": 1.7470141150922909, + "grad_norm": 8.35853855400652, + "learning_rate": 1.511125585347987e-05, + "loss": 0.7991, + "step": 3218 + }, + { + "epoch": 1.747557003257329, + "grad_norm": 10.14192231306983, + "learning_rate": 1.5108232792728567e-05, + "loss": 0.9942, + "step": 3219 + }, + { + "epoch": 1.7480998914223669, + "grad_norm": 6.003113555943506, + "learning_rate": 1.5105209100182893e-05, + "loss": 0.5614, + "step": 3220 + }, + { + "epoch": 1.748642779587405, + "grad_norm": 7.420823995167257, + "learning_rate": 1.5102184776216824e-05, + "loss": 0.6918, + "step": 3221 + }, + { + "epoch": 1.7491856677524429, + "grad_norm": 7.3724702610256, + "learning_rate": 1.5099159821204406e-05, + "loss": 0.815, + "step": 3222 + }, + { + "epoch": 1.749728555917481, + "grad_norm": 8.539056209113738, + "learning_rate": 1.509613423551978e-05, + "loss": 1.0859, + "step": 3223 + }, + { + "epoch": 1.750271444082519, + "grad_norm": 10.518635534148203, + "learning_rate": 1.5093108019537147e-05, + "loss": 1.1873, + "step": 3224 + }, + { + "epoch": 1.7508143322475571, + "grad_norm": 8.049010262043616, + "learning_rate": 1.5090081173630795e-05, + "loss": 0.971, + "step": 3225 + }, + { + "epoch": 1.751357220412595, + "grad_norm": 7.874724964543097, + "learning_rate": 1.5087053698175092e-05, + "loss": 1.06, + "step": 3226 + }, + { + "epoch": 1.7519001085776331, + "grad_norm": 8.018813791724389, + "learning_rate": 1.508402559354448e-05, + "loss": 0.7192, + "step": 3227 + }, + { + "epoch": 1.752442996742671, + "grad_norm": 11.748757549078345, + "learning_rate": 1.508099686011348e-05, + "loss": 1.4476, + "step": 3228 + }, + { + "epoch": 1.7529858849077091, + "grad_norm": 7.885344513181394, + "learning_rate": 1.507796749825669e-05, + "loss": 0.7493, + "step": 3229 + }, + { + "epoch": 1.753528773072747, + "grad_norm": 6.808213775155858, + "learning_rate": 1.507493750834879e-05, + "loss": 0.6646, + "step": 3230 + }, + { + "epoch": 1.7540716612377851, + "grad_norm": 6.304489454276083, + "learning_rate": 1.5071906890764527e-05, + "loss": 0.7375, + "step": 3231 + }, + { + "epoch": 1.754614549402823, + "grad_norm": 8.217636823407403, + "learning_rate": 1.5068875645878739e-05, + "loss": 1.0069, + "step": 3232 + }, + { + "epoch": 1.7551574375678611, + "grad_norm": 10.706707510999644, + "learning_rate": 1.5065843774066329e-05, + "loss": 0.689, + "step": 3233 + }, + { + "epoch": 1.755700325732899, + "grad_norm": 7.600175930285094, + "learning_rate": 1.5062811275702291e-05, + "loss": 0.8645, + "step": 3234 + }, + { + "epoch": 1.7562432138979371, + "grad_norm": 8.463001355349205, + "learning_rate": 1.5059778151161684e-05, + "loss": 0.7185, + "step": 3235 + }, + { + "epoch": 1.756786102062975, + "grad_norm": 5.834156873881953, + "learning_rate": 1.5056744400819651e-05, + "loss": 0.6486, + "step": 3236 + }, + { + "epoch": 1.7573289902280131, + "grad_norm": 7.152665898997485, + "learning_rate": 1.5053710025051411e-05, + "loss": 0.9116, + "step": 3237 + }, + { + "epoch": 1.757871878393051, + "grad_norm": 8.037482464894001, + "learning_rate": 1.5050675024232262e-05, + "loss": 0.8231, + "step": 3238 + }, + { + "epoch": 1.7584147665580891, + "grad_norm": 10.416359782719013, + "learning_rate": 1.5047639398737573e-05, + "loss": 1.2555, + "step": 3239 + }, + { + "epoch": 1.758957654723127, + "grad_norm": 6.789336798229921, + "learning_rate": 1.50446031489428e-05, + "loss": 0.7295, + "step": 3240 + }, + { + "epoch": 1.7595005428881652, + "grad_norm": 9.927450801451293, + "learning_rate": 1.5041566275223472e-05, + "loss": 1.5417, + "step": 3241 + }, + { + "epoch": 1.760043431053203, + "grad_norm": 7.040644062299268, + "learning_rate": 1.5038528777955188e-05, + "loss": 0.4495, + "step": 3242 + }, + { + "epoch": 1.7605863192182412, + "grad_norm": 8.848733486454524, + "learning_rate": 1.5035490657513638e-05, + "loss": 1.0123, + "step": 3243 + }, + { + "epoch": 1.761129207383279, + "grad_norm": 8.109736914130819, + "learning_rate": 1.503245191427458e-05, + "loss": 0.8984, + "step": 3244 + }, + { + "epoch": 1.7616720955483172, + "grad_norm": 7.210785707597705, + "learning_rate": 1.5029412548613845e-05, + "loss": 0.6947, + "step": 3245 + }, + { + "epoch": 1.762214983713355, + "grad_norm": 10.368251730052483, + "learning_rate": 1.502637256090735e-05, + "loss": 0.8413, + "step": 3246 + }, + { + "epoch": 1.7627578718783932, + "grad_norm": 6.770955437649876, + "learning_rate": 1.5023331951531086e-05, + "loss": 0.6386, + "step": 3247 + }, + { + "epoch": 1.763300760043431, + "grad_norm": 10.818384833705863, + "learning_rate": 1.5020290720861122e-05, + "loss": 1.2544, + "step": 3248 + }, + { + "epoch": 1.7638436482084692, + "grad_norm": 7.9104385792951195, + "learning_rate": 1.50172488692736e-05, + "loss": 0.6538, + "step": 3249 + }, + { + "epoch": 1.764386536373507, + "grad_norm": 9.565689306609565, + "learning_rate": 1.5014206397144742e-05, + "loss": 1.1232, + "step": 3250 + }, + { + "epoch": 1.7649294245385452, + "grad_norm": 11.651444088862487, + "learning_rate": 1.5011163304850844e-05, + "loss": 1.2054, + "step": 3251 + }, + { + "epoch": 1.765472312703583, + "grad_norm": 10.326063989708704, + "learning_rate": 1.500811959276828e-05, + "loss": 0.9441, + "step": 3252 + }, + { + "epoch": 1.7660152008686212, + "grad_norm": 8.0529225996154, + "learning_rate": 1.5005075261273506e-05, + "loss": 0.9431, + "step": 3253 + }, + { + "epoch": 1.766558089033659, + "grad_norm": 7.391183931496241, + "learning_rate": 1.500203031074305e-05, + "loss": 0.8619, + "step": 3254 + }, + { + "epoch": 1.7671009771986972, + "grad_norm": 6.892445188811786, + "learning_rate": 1.4998984741553508e-05, + "loss": 0.7039, + "step": 3255 + }, + { + "epoch": 1.767643865363735, + "grad_norm": 6.541771639151763, + "learning_rate": 1.4995938554081568e-05, + "loss": 0.6359, + "step": 3256 + }, + { + "epoch": 1.7681867535287732, + "grad_norm": 7.690176361123399, + "learning_rate": 1.4992891748703985e-05, + "loss": 0.8998, + "step": 3257 + }, + { + "epoch": 1.768729641693811, + "grad_norm": 6.497301879748706, + "learning_rate": 1.498984432579759e-05, + "loss": 0.5743, + "step": 3258 + }, + { + "epoch": 1.7692725298588492, + "grad_norm": 6.656723951695686, + "learning_rate": 1.4986796285739298e-05, + "loss": 0.7002, + "step": 3259 + }, + { + "epoch": 1.769815418023887, + "grad_norm": 7.9129397206513685, + "learning_rate": 1.4983747628906095e-05, + "loss": 0.9239, + "step": 3260 + }, + { + "epoch": 1.7703583061889252, + "grad_norm": 4.544544755639526, + "learning_rate": 1.4980698355675043e-05, + "loss": 0.3431, + "step": 3261 + }, + { + "epoch": 1.770901194353963, + "grad_norm": 8.243743590575315, + "learning_rate": 1.4977648466423278e-05, + "loss": 1.1386, + "step": 3262 + }, + { + "epoch": 1.7714440825190012, + "grad_norm": 9.404041237838763, + "learning_rate": 1.4974597961528021e-05, + "loss": 0.9087, + "step": 3263 + }, + { + "epoch": 1.771986970684039, + "grad_norm": 8.534670509056685, + "learning_rate": 1.4971546841366556e-05, + "loss": 0.8769, + "step": 3264 + }, + { + "epoch": 1.7725298588490772, + "grad_norm": 9.442139570892468, + "learning_rate": 1.4968495106316254e-05, + "loss": 0.6942, + "step": 3265 + }, + { + "epoch": 1.773072747014115, + "grad_norm": 8.449365139837813, + "learning_rate": 1.496544275675456e-05, + "loss": 1.0743, + "step": 3266 + }, + { + "epoch": 1.7736156351791532, + "grad_norm": 9.461545898435599, + "learning_rate": 1.4962389793058994e-05, + "loss": 0.7473, + "step": 3267 + }, + { + "epoch": 1.774158523344191, + "grad_norm": 11.871672085414737, + "learning_rate": 1.4959336215607147e-05, + "loss": 1.5433, + "step": 3268 + }, + { + "epoch": 1.7747014115092292, + "grad_norm": 6.417686676489462, + "learning_rate": 1.4956282024776693e-05, + "loss": 0.5006, + "step": 3269 + }, + { + "epoch": 1.775244299674267, + "grad_norm": 10.49161190534651, + "learning_rate": 1.4953227220945381e-05, + "loss": 0.9918, + "step": 3270 + }, + { + "epoch": 1.7757871878393052, + "grad_norm": 9.298743913365943, + "learning_rate": 1.4950171804491031e-05, + "loss": 1.3149, + "step": 3271 + }, + { + "epoch": 1.776330076004343, + "grad_norm": 7.775473503246929, + "learning_rate": 1.4947115775791541e-05, + "loss": 0.6873, + "step": 3272 + }, + { + "epoch": 1.7768729641693812, + "grad_norm": 6.75591368497042, + "learning_rate": 1.4944059135224891e-05, + "loss": 0.7799, + "step": 3273 + }, + { + "epoch": 1.777415852334419, + "grad_norm": 8.234180005996857, + "learning_rate": 1.4941001883169124e-05, + "loss": 1.0393, + "step": 3274 + }, + { + "epoch": 1.7779587404994572, + "grad_norm": 8.419005181659063, + "learning_rate": 1.4937944020002371e-05, + "loss": 0.6798, + "step": 3275 + }, + { + "epoch": 1.778501628664495, + "grad_norm": 7.0982600255490205, + "learning_rate": 1.4934885546102833e-05, + "loss": 0.4936, + "step": 3276 + }, + { + "epoch": 1.7790445168295332, + "grad_norm": 8.404626943584326, + "learning_rate": 1.4931826461848785e-05, + "loss": 1.0507, + "step": 3277 + }, + { + "epoch": 1.779587404994571, + "grad_norm": 5.861882830901275, + "learning_rate": 1.4928766767618576e-05, + "loss": 0.6785, + "step": 3278 + }, + { + "epoch": 1.7801302931596092, + "grad_norm": 8.114055102204162, + "learning_rate": 1.4925706463790642e-05, + "loss": 0.765, + "step": 3279 + }, + { + "epoch": 1.780673181324647, + "grad_norm": 10.033269670592238, + "learning_rate": 1.492264555074348e-05, + "loss": 0.9163, + "step": 3280 + }, + { + "epoch": 1.7812160694896852, + "grad_norm": 7.040755628619449, + "learning_rate": 1.4919584028855671e-05, + "loss": 0.5962, + "step": 3281 + }, + { + "epoch": 1.781758957654723, + "grad_norm": 12.093987616559675, + "learning_rate": 1.4916521898505872e-05, + "loss": 0.8999, + "step": 3282 + }, + { + "epoch": 1.7823018458197613, + "grad_norm": 9.400200287574112, + "learning_rate": 1.4913459160072805e-05, + "loss": 1.0203, + "step": 3283 + }, + { + "epoch": 1.782844733984799, + "grad_norm": 8.605701710844945, + "learning_rate": 1.4910395813935279e-05, + "loss": 0.7874, + "step": 3284 + }, + { + "epoch": 1.7833876221498373, + "grad_norm": 10.279849472300688, + "learning_rate": 1.4907331860472174e-05, + "loss": 1.4306, + "step": 3285 + }, + { + "epoch": 1.783930510314875, + "grad_norm": 10.582881297295359, + "learning_rate": 1.4904267300062443e-05, + "loss": 0.6536, + "step": 3286 + }, + { + "epoch": 1.7844733984799133, + "grad_norm": 6.960885594125151, + "learning_rate": 1.4901202133085115e-05, + "loss": 0.7256, + "step": 3287 + }, + { + "epoch": 1.785016286644951, + "grad_norm": 9.088589913670967, + "learning_rate": 1.4898136359919298e-05, + "loss": 0.9299, + "step": 3288 + }, + { + "epoch": 1.7855591748099893, + "grad_norm": 8.922355254191798, + "learning_rate": 1.4895069980944168e-05, + "loss": 0.8159, + "step": 3289 + }, + { + "epoch": 1.786102062975027, + "grad_norm": 9.459478678151763, + "learning_rate": 1.4892002996538982e-05, + "loss": 1.132, + "step": 3290 + }, + { + "epoch": 1.7866449511400653, + "grad_norm": 9.746458814865415, + "learning_rate": 1.4888935407083065e-05, + "loss": 0.9874, + "step": 3291 + }, + { + "epoch": 1.787187839305103, + "grad_norm": 8.13065950431162, + "learning_rate": 1.4885867212955827e-05, + "loss": 0.8013, + "step": 3292 + }, + { + "epoch": 1.7877307274701413, + "grad_norm": 8.469962208509646, + "learning_rate": 1.4882798414536749e-05, + "loss": 1.0408, + "step": 3293 + }, + { + "epoch": 1.788273615635179, + "grad_norm": 9.018264603768202, + "learning_rate": 1.4879729012205379e-05, + "loss": 0.8324, + "step": 3294 + }, + { + "epoch": 1.7888165038002173, + "grad_norm": 8.934092671110381, + "learning_rate": 1.4876659006341347e-05, + "loss": 0.8842, + "step": 3295 + }, + { + "epoch": 1.789359391965255, + "grad_norm": 7.612464888658943, + "learning_rate": 1.4873588397324356e-05, + "loss": 0.9356, + "step": 3296 + }, + { + "epoch": 1.7899022801302933, + "grad_norm": 7.303452266512586, + "learning_rate": 1.4870517185534184e-05, + "loss": 1.0291, + "step": 3297 + }, + { + "epoch": 1.790445168295331, + "grad_norm": 9.805172223903986, + "learning_rate": 1.4867445371350687e-05, + "loss": 1.0243, + "step": 3298 + }, + { + "epoch": 1.7909880564603693, + "grad_norm": 7.742224062288237, + "learning_rate": 1.4864372955153785e-05, + "loss": 0.7173, + "step": 3299 + }, + { + "epoch": 1.791530944625407, + "grad_norm": 9.174714938761847, + "learning_rate": 1.4861299937323485e-05, + "loss": 0.9721, + "step": 3300 + }, + { + "epoch": 1.7920738327904453, + "grad_norm": 8.116980821679304, + "learning_rate": 1.4858226318239864e-05, + "loss": 0.7776, + "step": 3301 + }, + { + "epoch": 1.792616720955483, + "grad_norm": 5.885386861657401, + "learning_rate": 1.4855152098283068e-05, + "loss": 0.3833, + "step": 3302 + }, + { + "epoch": 1.7931596091205213, + "grad_norm": 6.804177263707199, + "learning_rate": 1.4852077277833319e-05, + "loss": 0.7972, + "step": 3303 + }, + { + "epoch": 1.793702497285559, + "grad_norm": 8.874804659745028, + "learning_rate": 1.4849001857270922e-05, + "loss": 1.063, + "step": 3304 + }, + { + "epoch": 1.7942453854505973, + "grad_norm": 10.231899453600786, + "learning_rate": 1.4845925836976247e-05, + "loss": 1.6617, + "step": 3305 + }, + { + "epoch": 1.794788273615635, + "grad_norm": 9.067079935451718, + "learning_rate": 1.4842849217329742e-05, + "loss": 1.1062, + "step": 3306 + }, + { + "epoch": 1.7953311617806733, + "grad_norm": 9.5261315220339, + "learning_rate": 1.4839771998711925e-05, + "loss": 1.5991, + "step": 3307 + }, + { + "epoch": 1.795874049945711, + "grad_norm": 9.53442504374803, + "learning_rate": 1.4836694181503397e-05, + "loss": 0.9267, + "step": 3308 + }, + { + "epoch": 1.7964169381107493, + "grad_norm": 8.231497735162955, + "learning_rate": 1.4833615766084821e-05, + "loss": 0.9802, + "step": 3309 + }, + { + "epoch": 1.796959826275787, + "grad_norm": 7.2171054452276815, + "learning_rate": 1.4830536752836944e-05, + "loss": 0.6116, + "step": 3310 + }, + { + "epoch": 1.7975027144408253, + "grad_norm": 8.412637332813022, + "learning_rate": 1.4827457142140583e-05, + "loss": 0.7482, + "step": 3311 + }, + { + "epoch": 1.798045602605863, + "grad_norm": 6.24239912297862, + "learning_rate": 1.482437693437663e-05, + "loss": 0.4333, + "step": 3312 + }, + { + "epoch": 1.7985884907709013, + "grad_norm": 9.706212881828902, + "learning_rate": 1.4821296129926049e-05, + "loss": 1.3862, + "step": 3313 + }, + { + "epoch": 1.799131378935939, + "grad_norm": 8.275124350006633, + "learning_rate": 1.4818214729169878e-05, + "loss": 0.812, + "step": 3314 + }, + { + "epoch": 1.7996742671009773, + "grad_norm": 7.0810692617964275, + "learning_rate": 1.4815132732489228e-05, + "loss": 0.8919, + "step": 3315 + }, + { + "epoch": 1.8002171552660151, + "grad_norm": 7.225811411306636, + "learning_rate": 1.4812050140265288e-05, + "loss": 0.7512, + "step": 3316 + }, + { + "epoch": 1.8007600434310533, + "grad_norm": 6.458874811050634, + "learning_rate": 1.4808966952879318e-05, + "loss": 0.6914, + "step": 3317 + }, + { + "epoch": 1.8013029315960911, + "grad_norm": 6.537680447766501, + "learning_rate": 1.4805883170712653e-05, + "loss": 0.5511, + "step": 3318 + }, + { + "epoch": 1.8018458197611293, + "grad_norm": 8.023777979063889, + "learning_rate": 1.4802798794146695e-05, + "loss": 0.6814, + "step": 3319 + }, + { + "epoch": 1.8023887079261671, + "grad_norm": 13.062126505855263, + "learning_rate": 1.4799713823562932e-05, + "loss": 1.0526, + "step": 3320 + }, + { + "epoch": 1.8029315960912053, + "grad_norm": 8.173027381817493, + "learning_rate": 1.479662825934291e-05, + "loss": 0.8939, + "step": 3321 + }, + { + "epoch": 1.8034744842562431, + "grad_norm": 7.465360921787222, + "learning_rate": 1.4793542101868261e-05, + "loss": 0.5965, + "step": 3322 + }, + { + "epoch": 1.8040173724212814, + "grad_norm": 6.5098081549027835, + "learning_rate": 1.4790455351520684e-05, + "loss": 0.5963, + "step": 3323 + }, + { + "epoch": 1.8045602605863191, + "grad_norm": 11.071700718231854, + "learning_rate": 1.4787368008681956e-05, + "loss": 0.9728, + "step": 3324 + }, + { + "epoch": 1.8051031487513574, + "grad_norm": 9.251168051579464, + "learning_rate": 1.4784280073733924e-05, + "loss": 0.8816, + "step": 3325 + }, + { + "epoch": 1.8056460369163951, + "grad_norm": 7.876824650963893, + "learning_rate": 1.4781191547058505e-05, + "loss": 1.0365, + "step": 3326 + }, + { + "epoch": 1.8061889250814334, + "grad_norm": 7.984254709859244, + "learning_rate": 1.4778102429037696e-05, + "loss": 0.7252, + "step": 3327 + }, + { + "epoch": 1.8067318132464711, + "grad_norm": 8.162664097932598, + "learning_rate": 1.4775012720053563e-05, + "loss": 0.8297, + "step": 3328 + }, + { + "epoch": 1.8072747014115094, + "grad_norm": 6.631733087479191, + "learning_rate": 1.4771922420488245e-05, + "loss": 0.6962, + "step": 3329 + }, + { + "epoch": 1.8078175895765471, + "grad_norm": 8.80250228426254, + "learning_rate": 1.476883153072396e-05, + "loss": 1.0576, + "step": 3330 + }, + { + "epoch": 1.8083604777415854, + "grad_norm": 10.637063152950418, + "learning_rate": 1.476574005114299e-05, + "loss": 0.8866, + "step": 3331 + }, + { + "epoch": 1.8089033659066232, + "grad_norm": 9.58332512337787, + "learning_rate": 1.4762647982127696e-05, + "loss": 1.6148, + "step": 3332 + }, + { + "epoch": 1.8094462540716614, + "grad_norm": 7.826567769007031, + "learning_rate": 1.4759555324060508e-05, + "loss": 0.9256, + "step": 3333 + }, + { + "epoch": 1.8099891422366992, + "grad_norm": 9.445307614493151, + "learning_rate": 1.4756462077323931e-05, + "loss": 1.1389, + "step": 3334 + }, + { + "epoch": 1.8105320304017374, + "grad_norm": 7.504505604214132, + "learning_rate": 1.4753368242300543e-05, + "loss": 1.1333, + "step": 3335 + }, + { + "epoch": 1.8110749185667752, + "grad_norm": 6.996150668976886, + "learning_rate": 1.4750273819372996e-05, + "loss": 0.6721, + "step": 3336 + }, + { + "epoch": 1.8116178067318134, + "grad_norm": 7.593185686378431, + "learning_rate": 1.4747178808924012e-05, + "loss": 0.6856, + "step": 3337 + }, + { + "epoch": 1.8121606948968512, + "grad_norm": 7.277948846492979, + "learning_rate": 1.474408321133639e-05, + "loss": 0.5487, + "step": 3338 + }, + { + "epoch": 1.8127035830618894, + "grad_norm": 9.500801329454344, + "learning_rate": 1.474098702699299e-05, + "loss": 1.063, + "step": 3339 + }, + { + "epoch": 1.8132464712269272, + "grad_norm": 7.732920511283457, + "learning_rate": 1.4737890256276763e-05, + "loss": 0.8875, + "step": 3340 + }, + { + "epoch": 1.8137893593919654, + "grad_norm": 9.56546309643246, + "learning_rate": 1.4734792899570716e-05, + "loss": 0.7724, + "step": 3341 + }, + { + "epoch": 1.8143322475570032, + "grad_norm": 7.735046028758256, + "learning_rate": 1.4731694957257938e-05, + "loss": 0.8498, + "step": 3342 + }, + { + "epoch": 1.8148751357220414, + "grad_norm": 9.077413763001829, + "learning_rate": 1.4728596429721587e-05, + "loss": 1.1604, + "step": 3343 + }, + { + "epoch": 1.8154180238870792, + "grad_norm": 8.33530084886159, + "learning_rate": 1.4725497317344892e-05, + "loss": 1.0126, + "step": 3344 + }, + { + "epoch": 1.8159609120521174, + "grad_norm": 11.938749223535789, + "learning_rate": 1.4722397620511158e-05, + "loss": 0.9538, + "step": 3345 + }, + { + "epoch": 1.8165038002171552, + "grad_norm": 4.946988727080576, + "learning_rate": 1.471929733960376e-05, + "loss": 0.3007, + "step": 3346 + }, + { + "epoch": 1.8170466883821934, + "grad_norm": 6.934881579248824, + "learning_rate": 1.4716196475006145e-05, + "loss": 0.5286, + "step": 3347 + }, + { + "epoch": 1.8175895765472312, + "grad_norm": 12.504257989709076, + "learning_rate": 1.4713095027101833e-05, + "loss": 0.9948, + "step": 3348 + }, + { + "epoch": 1.8181324647122694, + "grad_norm": 6.299717892798866, + "learning_rate": 1.4709992996274416e-05, + "loss": 0.6503, + "step": 3349 + }, + { + "epoch": 1.8186753528773072, + "grad_norm": 8.48347618555589, + "learning_rate": 1.4706890382907562e-05, + "loss": 0.7195, + "step": 3350 + }, + { + "epoch": 1.8192182410423454, + "grad_norm": 7.992565784639017, + "learning_rate": 1.4703787187385002e-05, + "loss": 0.7349, + "step": 3351 + }, + { + "epoch": 1.8197611292073832, + "grad_norm": 7.567197193038979, + "learning_rate": 1.4700683410090546e-05, + "loss": 0.4719, + "step": 3352 + }, + { + "epoch": 1.8203040173724214, + "grad_norm": 7.276389893333727, + "learning_rate": 1.4697579051408073e-05, + "loss": 0.93, + "step": 3353 + }, + { + "epoch": 1.8208469055374592, + "grad_norm": 8.13251629417801, + "learning_rate": 1.4694474111721537e-05, + "loss": 0.9865, + "step": 3354 + }, + { + "epoch": 1.8213897937024974, + "grad_norm": 6.788010338042944, + "learning_rate": 1.469136859141496e-05, + "loss": 0.5671, + "step": 3355 + }, + { + "epoch": 1.8219326818675352, + "grad_norm": 6.83194614525194, + "learning_rate": 1.4688262490872438e-05, + "loss": 0.5486, + "step": 3356 + }, + { + "epoch": 1.8224755700325734, + "grad_norm": 12.201391760355081, + "learning_rate": 1.4685155810478143e-05, + "loss": 0.9178, + "step": 3357 + }, + { + "epoch": 1.8230184581976112, + "grad_norm": 7.955977866907339, + "learning_rate": 1.4682048550616306e-05, + "loss": 0.9671, + "step": 3358 + }, + { + "epoch": 1.8235613463626494, + "grad_norm": 9.699881747283202, + "learning_rate": 1.4678940711671245e-05, + "loss": 0.7699, + "step": 3359 + }, + { + "epoch": 1.8241042345276872, + "grad_norm": 7.703855587772515, + "learning_rate": 1.4675832294027339e-05, + "loss": 0.6471, + "step": 3360 + }, + { + "epoch": 1.8246471226927254, + "grad_norm": 11.230217201340418, + "learning_rate": 1.4672723298069043e-05, + "loss": 1.1256, + "step": 3361 + }, + { + "epoch": 1.8251900108577632, + "grad_norm": 11.964156332533788, + "learning_rate": 1.4669613724180883e-05, + "loss": 1.1829, + "step": 3362 + }, + { + "epoch": 1.8257328990228014, + "grad_norm": 8.221701868139569, + "learning_rate": 1.4666503572747452e-05, + "loss": 0.6004, + "step": 3363 + }, + { + "epoch": 1.8262757871878392, + "grad_norm": 6.179895131215982, + "learning_rate": 1.4663392844153426e-05, + "loss": 0.6923, + "step": 3364 + }, + { + "epoch": 1.8268186753528775, + "grad_norm": 8.333959237389427, + "learning_rate": 1.4660281538783538e-05, + "loss": 0.8785, + "step": 3365 + }, + { + "epoch": 1.8273615635179152, + "grad_norm": 9.742872979143373, + "learning_rate": 1.4657169657022603e-05, + "loss": 1.0129, + "step": 3366 + }, + { + "epoch": 1.8279044516829535, + "grad_norm": 8.019435507225802, + "learning_rate": 1.46540571992555e-05, + "loss": 1.0615, + "step": 3367 + }, + { + "epoch": 1.8284473398479912, + "grad_norm": 5.872559756186675, + "learning_rate": 1.4650944165867187e-05, + "loss": 0.6391, + "step": 3368 + }, + { + "epoch": 1.8289902280130295, + "grad_norm": 7.878073300625923, + "learning_rate": 1.4647830557242685e-05, + "loss": 1.0317, + "step": 3369 + }, + { + "epoch": 1.8295331161780672, + "grad_norm": 10.290400619980408, + "learning_rate": 1.4644716373767096e-05, + "loss": 1.3108, + "step": 3370 + }, + { + "epoch": 1.8300760043431055, + "grad_norm": 7.655417729944032, + "learning_rate": 1.4641601615825582e-05, + "loss": 0.8434, + "step": 3371 + }, + { + "epoch": 1.8306188925081432, + "grad_norm": 8.894875963607664, + "learning_rate": 1.4638486283803384e-05, + "loss": 1.4973, + "step": 3372 + }, + { + "epoch": 1.8311617806731815, + "grad_norm": 7.70346849256546, + "learning_rate": 1.4635370378085809e-05, + "loss": 0.9202, + "step": 3373 + }, + { + "epoch": 1.8317046688382193, + "grad_norm": 7.428334777443758, + "learning_rate": 1.4632253899058233e-05, + "loss": 1.005, + "step": 3374 + }, + { + "epoch": 1.8322475570032575, + "grad_norm": 6.905061230055496, + "learning_rate": 1.4629136847106118e-05, + "loss": 0.7045, + "step": 3375 + }, + { + "epoch": 1.8327904451682953, + "grad_norm": 9.092059147160219, + "learning_rate": 1.4626019222614977e-05, + "loss": 1.1478, + "step": 3376 + }, + { + "epoch": 1.8333333333333335, + "grad_norm": 8.171417716573657, + "learning_rate": 1.4622901025970406e-05, + "loss": 0.8098, + "step": 3377 + }, + { + "epoch": 1.8338762214983713, + "grad_norm": 8.935159598593566, + "learning_rate": 1.461978225755807e-05, + "loss": 0.9121, + "step": 3378 + }, + { + "epoch": 1.8344191096634095, + "grad_norm": 9.059802045574934, + "learning_rate": 1.46166629177637e-05, + "loss": 0.952, + "step": 3379 + }, + { + "epoch": 1.8349619978284473, + "grad_norm": 6.404641837192718, + "learning_rate": 1.4613543006973103e-05, + "loss": 0.7178, + "step": 3380 + }, + { + "epoch": 1.8355048859934855, + "grad_norm": 7.714640842582964, + "learning_rate": 1.4610422525572155e-05, + "loss": 0.7053, + "step": 3381 + }, + { + "epoch": 1.8360477741585233, + "grad_norm": 9.232659815202982, + "learning_rate": 1.46073014739468e-05, + "loss": 0.7707, + "step": 3382 + }, + { + "epoch": 1.8365906623235615, + "grad_norm": 7.597273066110282, + "learning_rate": 1.4604179852483056e-05, + "loss": 0.9593, + "step": 3383 + }, + { + "epoch": 1.8371335504885993, + "grad_norm": 7.246167686662127, + "learning_rate": 1.460105766156701e-05, + "loss": 0.8489, + "step": 3384 + }, + { + "epoch": 1.8376764386536375, + "grad_norm": 9.724005322694314, + "learning_rate": 1.4597934901584818e-05, + "loss": 1.4838, + "step": 3385 + }, + { + "epoch": 1.8382193268186753, + "grad_norm": 5.623806840885622, + "learning_rate": 1.459481157292271e-05, + "loss": 0.4218, + "step": 3386 + }, + { + "epoch": 1.8387622149837135, + "grad_norm": 10.505590233471066, + "learning_rate": 1.4591687675966984e-05, + "loss": 1.3634, + "step": 3387 + }, + { + "epoch": 1.8393051031487513, + "grad_norm": 7.225825527188378, + "learning_rate": 1.458856321110401e-05, + "loss": 1.0878, + "step": 3388 + }, + { + "epoch": 1.8398479913137895, + "grad_norm": 6.92346369662784, + "learning_rate": 1.4585438178720221e-05, + "loss": 0.8579, + "step": 3389 + }, + { + "epoch": 1.8403908794788273, + "grad_norm": 6.78903317106508, + "learning_rate": 1.4582312579202134e-05, + "loss": 0.9986, + "step": 3390 + }, + { + "epoch": 1.8409337676438655, + "grad_norm": 7.493115074862596, + "learning_rate": 1.4579186412936323e-05, + "loss": 0.6403, + "step": 3391 + }, + { + "epoch": 1.8414766558089033, + "grad_norm": 7.292140639462925, + "learning_rate": 1.457605968030944e-05, + "loss": 0.7705, + "step": 3392 + }, + { + "epoch": 1.8420195439739415, + "grad_norm": 8.413656799764597, + "learning_rate": 1.4572932381708198e-05, + "loss": 0.7362, + "step": 3393 + }, + { + "epoch": 1.8425624321389793, + "grad_norm": 8.092404119538978, + "learning_rate": 1.4569804517519392e-05, + "loss": 0.8419, + "step": 3394 + }, + { + "epoch": 1.8431053203040175, + "grad_norm": 8.746713556642208, + "learning_rate": 1.456667608812988e-05, + "loss": 0.8992, + "step": 3395 + }, + { + "epoch": 1.8436482084690553, + "grad_norm": 9.315011907817567, + "learning_rate": 1.4563547093926595e-05, + "loss": 1.1811, + "step": 3396 + }, + { + "epoch": 1.8441910966340935, + "grad_norm": 7.921580843874425, + "learning_rate": 1.4560417535296529e-05, + "loss": 0.9896, + "step": 3397 + }, + { + "epoch": 1.8447339847991313, + "grad_norm": 5.324625495868631, + "learning_rate": 1.4557287412626755e-05, + "loss": 0.5757, + "step": 3398 + }, + { + "epoch": 1.8452768729641695, + "grad_norm": 7.061326556489923, + "learning_rate": 1.4554156726304411e-05, + "loss": 0.7016, + "step": 3399 + }, + { + "epoch": 1.8458197611292073, + "grad_norm": 7.833004564597817, + "learning_rate": 1.4551025476716704e-05, + "loss": 0.6112, + "step": 3400 + }, + { + "epoch": 1.8463626492942455, + "grad_norm": 6.762019290951184, + "learning_rate": 1.4547893664250912e-05, + "loss": 0.6197, + "step": 3401 + }, + { + "epoch": 1.8469055374592833, + "grad_norm": 8.02135944193365, + "learning_rate": 1.4544761289294384e-05, + "loss": 0.7665, + "step": 3402 + }, + { + "epoch": 1.8474484256243215, + "grad_norm": 9.145981041987138, + "learning_rate": 1.4541628352234538e-05, + "loss": 1.4491, + "step": 3403 + }, + { + "epoch": 1.8479913137893593, + "grad_norm": 10.127072377919164, + "learning_rate": 1.4538494853458858e-05, + "loss": 1.1929, + "step": 3404 + }, + { + "epoch": 1.8485342019543975, + "grad_norm": 9.041637256167727, + "learning_rate": 1.45353607933549e-05, + "loss": 0.8471, + "step": 3405 + }, + { + "epoch": 1.8490770901194353, + "grad_norm": 7.08038124220055, + "learning_rate": 1.453222617231029e-05, + "loss": 0.5152, + "step": 3406 + }, + { + "epoch": 1.8496199782844736, + "grad_norm": 8.460108981189723, + "learning_rate": 1.4529090990712726e-05, + "loss": 0.9495, + "step": 3407 + }, + { + "epoch": 1.8501628664495113, + "grad_norm": 8.939702203617836, + "learning_rate": 1.452595524894997e-05, + "loss": 1.0959, + "step": 3408 + }, + { + "epoch": 1.8507057546145496, + "grad_norm": 7.5616663713048595, + "learning_rate": 1.4522818947409855e-05, + "loss": 1.0391, + "step": 3409 + }, + { + "epoch": 1.8512486427795873, + "grad_norm": 8.75531450227851, + "learning_rate": 1.4519682086480287e-05, + "loss": 0.7298, + "step": 3410 + }, + { + "epoch": 1.8517915309446256, + "grad_norm": 7.003753183958329, + "learning_rate": 1.4516544666549233e-05, + "loss": 0.4788, + "step": 3411 + }, + { + "epoch": 1.8523344191096633, + "grad_norm": 10.952441700747523, + "learning_rate": 1.4513406688004734e-05, + "loss": 1.0903, + "step": 3412 + }, + { + "epoch": 1.8528773072747016, + "grad_norm": 8.61536651212902, + "learning_rate": 1.4510268151234903e-05, + "loss": 0.6924, + "step": 3413 + }, + { + "epoch": 1.8534201954397393, + "grad_norm": 9.237257682023538, + "learning_rate": 1.4507129056627922e-05, + "loss": 0.8686, + "step": 3414 + }, + { + "epoch": 1.8539630836047776, + "grad_norm": 7.792081331119547, + "learning_rate": 1.4503989404572031e-05, + "loss": 0.9818, + "step": 3415 + }, + { + "epoch": 1.8545059717698154, + "grad_norm": 8.622336937881878, + "learning_rate": 1.4500849195455557e-05, + "loss": 0.9444, + "step": 3416 + }, + { + "epoch": 1.8550488599348536, + "grad_norm": 7.607437519589793, + "learning_rate": 1.4497708429666882e-05, + "loss": 0.846, + "step": 3417 + }, + { + "epoch": 1.8555917480998914, + "grad_norm": 8.358779281223866, + "learning_rate": 1.4494567107594457e-05, + "loss": 0.7125, + "step": 3418 + }, + { + "epoch": 1.8561346362649296, + "grad_norm": 6.8176708905901044, + "learning_rate": 1.449142522962681e-05, + "loss": 0.8921, + "step": 3419 + }, + { + "epoch": 1.8566775244299674, + "grad_norm": 8.931750092418099, + "learning_rate": 1.448828279615253e-05, + "loss": 0.7989, + "step": 3420 + }, + { + "epoch": 1.8572204125950056, + "grad_norm": 7.481066786036915, + "learning_rate": 1.4485139807560281e-05, + "loss": 0.6642, + "step": 3421 + }, + { + "epoch": 1.8577633007600434, + "grad_norm": 7.460924029721227, + "learning_rate": 1.4481996264238796e-05, + "loss": 0.6509, + "step": 3422 + }, + { + "epoch": 1.8583061889250816, + "grad_norm": 7.224784834236484, + "learning_rate": 1.4478852166576869e-05, + "loss": 0.7244, + "step": 3423 + }, + { + "epoch": 1.8588490770901194, + "grad_norm": 5.912102755785491, + "learning_rate": 1.4475707514963367e-05, + "loss": 0.5713, + "step": 3424 + }, + { + "epoch": 1.8593919652551576, + "grad_norm": 6.994884495015334, + "learning_rate": 1.4472562309787224e-05, + "loss": 0.4875, + "step": 3425 + }, + { + "epoch": 1.8599348534201954, + "grad_norm": 6.38845062016643, + "learning_rate": 1.4469416551437445e-05, + "loss": 0.4708, + "step": 3426 + }, + { + "epoch": 1.8604777415852336, + "grad_norm": 11.250341421225603, + "learning_rate": 1.4466270240303109e-05, + "loss": 1.4338, + "step": 3427 + }, + { + "epoch": 1.8610206297502714, + "grad_norm": 9.828853213376544, + "learning_rate": 1.4463123376773348e-05, + "loss": 0.7307, + "step": 3428 + }, + { + "epoch": 1.8615635179153096, + "grad_norm": 7.96498248613748, + "learning_rate": 1.4459975961237375e-05, + "loss": 0.6727, + "step": 3429 + }, + { + "epoch": 1.8621064060803474, + "grad_norm": 9.736354023562699, + "learning_rate": 1.4456827994084473e-05, + "loss": 0.8091, + "step": 3430 + }, + { + "epoch": 1.8626492942453856, + "grad_norm": 10.317215804231997, + "learning_rate": 1.4453679475703974e-05, + "loss": 1.0997, + "step": 3431 + }, + { + "epoch": 1.8631921824104234, + "grad_norm": 10.00211430988322, + "learning_rate": 1.4450530406485301e-05, + "loss": 1.4007, + "step": 3432 + }, + { + "epoch": 1.8637350705754616, + "grad_norm": 10.665872978467544, + "learning_rate": 1.4447380786817934e-05, + "loss": 0.9221, + "step": 3433 + }, + { + "epoch": 1.8642779587404994, + "grad_norm": 8.8489121199226, + "learning_rate": 1.4444230617091424e-05, + "loss": 0.7745, + "step": 3434 + }, + { + "epoch": 1.8648208469055376, + "grad_norm": 7.083707820329958, + "learning_rate": 1.444107989769539e-05, + "loss": 0.7152, + "step": 3435 + }, + { + "epoch": 1.8653637350705754, + "grad_norm": 7.496522090143239, + "learning_rate": 1.4437928629019515e-05, + "loss": 0.5901, + "step": 3436 + }, + { + "epoch": 1.8659066232356136, + "grad_norm": 10.576310319834935, + "learning_rate": 1.4434776811453552e-05, + "loss": 0.7752, + "step": 3437 + }, + { + "epoch": 1.8664495114006514, + "grad_norm": 8.686730112665717, + "learning_rate": 1.4431624445387327e-05, + "loss": 0.8748, + "step": 3438 + }, + { + "epoch": 1.8669923995656896, + "grad_norm": 9.435491336462407, + "learning_rate": 1.4428471531210726e-05, + "loss": 1.1954, + "step": 3439 + }, + { + "epoch": 1.8675352877307274, + "grad_norm": 9.012159097359817, + "learning_rate": 1.4425318069313712e-05, + "loss": 0.7353, + "step": 3440 + }, + { + "epoch": 1.8680781758957656, + "grad_norm": 9.617409218513528, + "learning_rate": 1.4422164060086304e-05, + "loss": 0.6611, + "step": 3441 + }, + { + "epoch": 1.8686210640608034, + "grad_norm": 9.612516235298202, + "learning_rate": 1.4419009503918598e-05, + "loss": 0.8101, + "step": 3442 + }, + { + "epoch": 1.8691639522258416, + "grad_norm": 6.542519560792006, + "learning_rate": 1.4415854401200751e-05, + "loss": 0.473, + "step": 3443 + }, + { + "epoch": 1.8697068403908794, + "grad_norm": 10.696166776054064, + "learning_rate": 1.4412698752322998e-05, + "loss": 1.0001, + "step": 3444 + }, + { + "epoch": 1.8702497285559176, + "grad_norm": 10.087499378994824, + "learning_rate": 1.4409542557675625e-05, + "loss": 0.9323, + "step": 3445 + }, + { + "epoch": 1.8707926167209554, + "grad_norm": 7.829085764895678, + "learning_rate": 1.4406385817649002e-05, + "loss": 0.8624, + "step": 3446 + }, + { + "epoch": 1.8713355048859937, + "grad_norm": 11.554825371147222, + "learning_rate": 1.4403228532633561e-05, + "loss": 0.7735, + "step": 3447 + }, + { + "epoch": 1.8718783930510314, + "grad_norm": 9.43390979006707, + "learning_rate": 1.4400070703019797e-05, + "loss": 0.9128, + "step": 3448 + }, + { + "epoch": 1.8724212812160697, + "grad_norm": 11.851717520188132, + "learning_rate": 1.4396912329198276e-05, + "loss": 1.2379, + "step": 3449 + }, + { + "epoch": 1.8729641693811074, + "grad_norm": 8.408242915866566, + "learning_rate": 1.4393753411559632e-05, + "loss": 0.7549, + "step": 3450 + }, + { + "epoch": 1.8735070575461457, + "grad_norm": 8.144856314123112, + "learning_rate": 1.4390593950494557e-05, + "loss": 0.7281, + "step": 3451 + }, + { + "epoch": 1.8740499457111834, + "grad_norm": 8.202846690293642, + "learning_rate": 1.438743394639383e-05, + "loss": 0.6773, + "step": 3452 + }, + { + "epoch": 1.8745928338762217, + "grad_norm": 9.423562458268494, + "learning_rate": 1.438427339964828e-05, + "loss": 1.1, + "step": 3453 + }, + { + "epoch": 1.8751357220412594, + "grad_norm": 7.196483568403804, + "learning_rate": 1.4381112310648806e-05, + "loss": 0.5809, + "step": 3454 + }, + { + "epoch": 1.8756786102062974, + "grad_norm": 7.26770042201738, + "learning_rate": 1.4377950679786382e-05, + "loss": 0.8053, + "step": 3455 + }, + { + "epoch": 1.8762214983713354, + "grad_norm": 9.472969211327861, + "learning_rate": 1.4374788507452039e-05, + "loss": 0.7998, + "step": 3456 + }, + { + "epoch": 1.8767643865363735, + "grad_norm": 9.514618174868874, + "learning_rate": 1.437162579403688e-05, + "loss": 0.9572, + "step": 3457 + }, + { + "epoch": 1.8773072747014115, + "grad_norm": 9.363034597703797, + "learning_rate": 1.4368462539932073e-05, + "loss": 0.8528, + "step": 3458 + }, + { + "epoch": 1.8778501628664495, + "grad_norm": 7.113386807438533, + "learning_rate": 1.4365298745528863e-05, + "loss": 0.733, + "step": 3459 + }, + { + "epoch": 1.8783930510314875, + "grad_norm": 9.09654372206602, + "learning_rate": 1.4362134411218545e-05, + "loss": 0.7779, + "step": 3460 + }, + { + "epoch": 1.8789359391965255, + "grad_norm": 7.449242636613164, + "learning_rate": 1.4358969537392489e-05, + "loss": 0.899, + "step": 3461 + }, + { + "epoch": 1.8794788273615635, + "grad_norm": 10.755480560042127, + "learning_rate": 1.4355804124442133e-05, + "loss": 1.099, + "step": 3462 + }, + { + "epoch": 1.8800217155266015, + "grad_norm": 9.091218073930861, + "learning_rate": 1.4352638172758984e-05, + "loss": 0.7489, + "step": 3463 + }, + { + "epoch": 1.8805646036916395, + "grad_norm": 9.701260384390375, + "learning_rate": 1.4349471682734603e-05, + "loss": 0.8495, + "step": 3464 + }, + { + "epoch": 1.8811074918566775, + "grad_norm": 9.008809447502522, + "learning_rate": 1.4346304654760637e-05, + "loss": 0.8984, + "step": 3465 + }, + { + "epoch": 1.8816503800217155, + "grad_norm": 8.176404536031864, + "learning_rate": 1.4343137089228783e-05, + "loss": 0.8278, + "step": 3466 + }, + { + "epoch": 1.8821932681867535, + "grad_norm": 9.197307290475337, + "learning_rate": 1.4339968986530813e-05, + "loss": 0.7142, + "step": 3467 + }, + { + "epoch": 1.8827361563517915, + "grad_norm": 5.505228455971892, + "learning_rate": 1.433680034705856e-05, + "loss": 0.4204, + "step": 3468 + }, + { + "epoch": 1.8832790445168295, + "grad_norm": 9.014964705534542, + "learning_rate": 1.433363117120393e-05, + "loss": 0.8988, + "step": 3469 + }, + { + "epoch": 1.8838219326818675, + "grad_norm": 7.290687267145188, + "learning_rate": 1.4330461459358889e-05, + "loss": 0.7683, + "step": 3470 + }, + { + "epoch": 1.8843648208469055, + "grad_norm": 10.834879009670003, + "learning_rate": 1.4327291211915473e-05, + "loss": 1.0446, + "step": 3471 + }, + { + "epoch": 1.8849077090119435, + "grad_norm": 8.240452054600638, + "learning_rate": 1.4324120429265781e-05, + "loss": 0.893, + "step": 3472 + }, + { + "epoch": 1.8854505971769815, + "grad_norm": 8.17919491677486, + "learning_rate": 1.4320949111801987e-05, + "loss": 0.9846, + "step": 3473 + }, + { + "epoch": 1.8859934853420195, + "grad_norm": 7.693481880154859, + "learning_rate": 1.431777725991632e-05, + "loss": 0.6054, + "step": 3474 + }, + { + "epoch": 1.8865363735070575, + "grad_norm": 7.929110152467141, + "learning_rate": 1.4314604874001076e-05, + "loss": 0.7299, + "step": 3475 + }, + { + "epoch": 1.8870792616720955, + "grad_norm": 9.650029930461764, + "learning_rate": 1.4311431954448629e-05, + "loss": 1.0855, + "step": 3476 + }, + { + "epoch": 1.8876221498371335, + "grad_norm": 8.487868906333745, + "learning_rate": 1.4308258501651404e-05, + "loss": 0.7117, + "step": 3477 + }, + { + "epoch": 1.8881650380021715, + "grad_norm": 7.111157851383161, + "learning_rate": 1.43050845160019e-05, + "loss": 0.6477, + "step": 3478 + }, + { + "epoch": 1.8887079261672095, + "grad_norm": 8.2908796123436, + "learning_rate": 1.4301909997892684e-05, + "loss": 0.9819, + "step": 3479 + }, + { + "epoch": 1.8892508143322475, + "grad_norm": 10.322900044501562, + "learning_rate": 1.4298734947716385e-05, + "loss": 0.9368, + "step": 3480 + }, + { + "epoch": 1.8897937024972855, + "grad_norm": 6.902927389625658, + "learning_rate": 1.4295559365865694e-05, + "loss": 0.6201, + "step": 3481 + }, + { + "epoch": 1.8903365906623235, + "grad_norm": 8.909574060437844, + "learning_rate": 1.4292383252733375e-05, + "loss": 0.9393, + "step": 3482 + }, + { + "epoch": 1.8908794788273615, + "grad_norm": 7.643883869480783, + "learning_rate": 1.4289206608712252e-05, + "loss": 0.9349, + "step": 3483 + }, + { + "epoch": 1.8914223669923995, + "grad_norm": 7.103928943926094, + "learning_rate": 1.428602943419522e-05, + "loss": 0.6511, + "step": 3484 + }, + { + "epoch": 1.8919652551574375, + "grad_norm": 8.843983286039231, + "learning_rate": 1.428285172957524e-05, + "loss": 0.6958, + "step": 3485 + }, + { + "epoch": 1.8925081433224755, + "grad_norm": 5.044738479793456, + "learning_rate": 1.427967349524533e-05, + "loss": 0.59, + "step": 3486 + }, + { + "epoch": 1.8930510314875135, + "grad_norm": 6.9998626206641195, + "learning_rate": 1.4276494731598582e-05, + "loss": 0.5973, + "step": 3487 + }, + { + "epoch": 1.8935939196525515, + "grad_norm": 8.316471360458495, + "learning_rate": 1.4273315439028151e-05, + "loss": 0.8941, + "step": 3488 + }, + { + "epoch": 1.8941368078175895, + "grad_norm": 12.713883241844957, + "learning_rate": 1.4270135617927254e-05, + "loss": 1.4921, + "step": 3489 + }, + { + "epoch": 1.8946796959826275, + "grad_norm": 10.043253133675226, + "learning_rate": 1.426695526868918e-05, + "loss": 1.0242, + "step": 3490 + }, + { + "epoch": 1.8952225841476655, + "grad_norm": 11.142761788047633, + "learning_rate": 1.4263774391707274e-05, + "loss": 1.265, + "step": 3491 + }, + { + "epoch": 1.8957654723127035, + "grad_norm": 7.12450280478247, + "learning_rate": 1.4260592987374962e-05, + "loss": 0.484, + "step": 3492 + }, + { + "epoch": 1.8963083604777415, + "grad_norm": 7.017938163026924, + "learning_rate": 1.4257411056085712e-05, + "loss": 0.6549, + "step": 3493 + }, + { + "epoch": 1.8968512486427795, + "grad_norm": 7.643680244898419, + "learning_rate": 1.4254228598233082e-05, + "loss": 0.614, + "step": 3494 + }, + { + "epoch": 1.8973941368078175, + "grad_norm": 9.484247292353862, + "learning_rate": 1.4251045614210678e-05, + "loss": 0.9667, + "step": 3495 + }, + { + "epoch": 1.8979370249728555, + "grad_norm": 10.846455465412243, + "learning_rate": 1.4247862104412175e-05, + "loss": 1.4425, + "step": 3496 + }, + { + "epoch": 1.8984799131378935, + "grad_norm": 7.211983265810251, + "learning_rate": 1.4244678069231319e-05, + "loss": 0.7508, + "step": 3497 + }, + { + "epoch": 1.8990228013029316, + "grad_norm": 8.283019454355765, + "learning_rate": 1.4241493509061912e-05, + "loss": 0.9747, + "step": 3498 + }, + { + "epoch": 1.8995656894679696, + "grad_norm": 6.678762476220025, + "learning_rate": 1.4238308424297833e-05, + "loss": 0.4702, + "step": 3499 + }, + { + "epoch": 1.9001085776330076, + "grad_norm": 8.750314729983545, + "learning_rate": 1.423512281533301e-05, + "loss": 0.7636, + "step": 3500 + }, + { + "epoch": 1.9006514657980456, + "grad_norm": 7.304249105438763, + "learning_rate": 1.4231936682561446e-05, + "loss": 0.7198, + "step": 3501 + }, + { + "epoch": 1.9011943539630836, + "grad_norm": 7.676960806676722, + "learning_rate": 1.4228750026377212e-05, + "loss": 0.8639, + "step": 3502 + }, + { + "epoch": 1.9017372421281216, + "grad_norm": 10.327339388528987, + "learning_rate": 1.4225562847174431e-05, + "loss": 0.9386, + "step": 3503 + }, + { + "epoch": 1.9022801302931596, + "grad_norm": 8.828430779713369, + "learning_rate": 1.4222375145347304e-05, + "loss": 1.0821, + "step": 3504 + }, + { + "epoch": 1.9028230184581976, + "grad_norm": 8.099300889556567, + "learning_rate": 1.421918692129009e-05, + "loss": 0.6251, + "step": 3505 + }, + { + "epoch": 1.9033659066232356, + "grad_norm": 9.197832742387652, + "learning_rate": 1.4215998175397115e-05, + "loss": 1.1794, + "step": 3506 + }, + { + "epoch": 1.9039087947882736, + "grad_norm": 9.305579433655693, + "learning_rate": 1.4212808908062763e-05, + "loss": 1.3447, + "step": 3507 + }, + { + "epoch": 1.9044516829533116, + "grad_norm": 6.897677914233707, + "learning_rate": 1.4209619119681497e-05, + "loss": 0.5178, + "step": 3508 + }, + { + "epoch": 1.9049945711183496, + "grad_norm": 9.359746343598495, + "learning_rate": 1.420642881064782e-05, + "loss": 0.985, + "step": 3509 + }, + { + "epoch": 1.9055374592833876, + "grad_norm": 5.855418324834113, + "learning_rate": 1.4203237981356331e-05, + "loss": 0.5144, + "step": 3510 + }, + { + "epoch": 1.9060803474484256, + "grad_norm": 7.408991833255532, + "learning_rate": 1.4200046632201665e-05, + "loss": 0.7358, + "step": 3511 + }, + { + "epoch": 1.9066232356134636, + "grad_norm": 9.024544122531433, + "learning_rate": 1.419685476357854e-05, + "loss": 0.778, + "step": 3512 + }, + { + "epoch": 1.9071661237785016, + "grad_norm": 7.874409137592207, + "learning_rate": 1.419366237588173e-05, + "loss": 0.8703, + "step": 3513 + }, + { + "epoch": 1.9077090119435396, + "grad_norm": 8.221027989379657, + "learning_rate": 1.4190469469506073e-05, + "loss": 0.7436, + "step": 3514 + }, + { + "epoch": 1.9082519001085776, + "grad_norm": 10.797659549408706, + "learning_rate": 1.4187276044846473e-05, + "loss": 1.0635, + "step": 3515 + }, + { + "epoch": 1.9087947882736156, + "grad_norm": 9.718910477411724, + "learning_rate": 1.4184082102297896e-05, + "loss": 0.8802, + "step": 3516 + }, + { + "epoch": 1.9093376764386536, + "grad_norm": 7.365463419520298, + "learning_rate": 1.4180887642255376e-05, + "loss": 0.5919, + "step": 3517 + }, + { + "epoch": 1.9098805646036916, + "grad_norm": 6.487376521266865, + "learning_rate": 1.4177692665114014e-05, + "loss": 0.7098, + "step": 3518 + }, + { + "epoch": 1.9104234527687296, + "grad_norm": 8.328541010595174, + "learning_rate": 1.4174497171268962e-05, + "loss": 0.6124, + "step": 3519 + }, + { + "epoch": 1.9109663409337676, + "grad_norm": 7.888322073346588, + "learning_rate": 1.4171301161115447e-05, + "loss": 0.827, + "step": 3520 + }, + { + "epoch": 1.9115092290988056, + "grad_norm": 10.407728423373431, + "learning_rate": 1.4168104635048756e-05, + "loss": 1.0511, + "step": 3521 + }, + { + "epoch": 1.9120521172638436, + "grad_norm": 8.154984984105669, + "learning_rate": 1.4164907593464239e-05, + "loss": 0.8156, + "step": 3522 + }, + { + "epoch": 1.9125950054288816, + "grad_norm": 9.142522776167024, + "learning_rate": 1.4161710036757314e-05, + "loss": 0.9046, + "step": 3523 + }, + { + "epoch": 1.9131378935939196, + "grad_norm": 8.610209396948733, + "learning_rate": 1.415851196532346e-05, + "loss": 0.5673, + "step": 3524 + }, + { + "epoch": 1.9136807817589576, + "grad_norm": 8.072698611315632, + "learning_rate": 1.415531337955822e-05, + "loss": 0.7375, + "step": 3525 + }, + { + "epoch": 1.9142236699239956, + "grad_norm": 9.878888535422037, + "learning_rate": 1.4152114279857197e-05, + "loss": 0.9677, + "step": 3526 + }, + { + "epoch": 1.9147665580890336, + "grad_norm": 8.524331697287977, + "learning_rate": 1.4148914666616062e-05, + "loss": 0.7617, + "step": 3527 + }, + { + "epoch": 1.9153094462540716, + "grad_norm": 8.04939882692512, + "learning_rate": 1.4145714540230549e-05, + "loss": 0.8767, + "step": 3528 + }, + { + "epoch": 1.9158523344191096, + "grad_norm": 10.69782500673993, + "learning_rate": 1.4142513901096453e-05, + "loss": 0.6357, + "step": 3529 + }, + { + "epoch": 1.9163952225841476, + "grad_norm": 7.279779735048647, + "learning_rate": 1.4139312749609637e-05, + "loss": 0.6474, + "step": 3530 + }, + { + "epoch": 1.9169381107491856, + "grad_norm": 7.290241222497209, + "learning_rate": 1.4136111086166024e-05, + "loss": 0.6929, + "step": 3531 + }, + { + "epoch": 1.9174809989142236, + "grad_norm": 8.3613074350218, + "learning_rate": 1.4132908911161598e-05, + "loss": 0.8199, + "step": 3532 + }, + { + "epoch": 1.9180238870792616, + "grad_norm": 9.015830183265928, + "learning_rate": 1.4129706224992413e-05, + "loss": 0.9811, + "step": 3533 + }, + { + "epoch": 1.9185667752442996, + "grad_norm": 9.621458967761185, + "learning_rate": 1.4126503028054579e-05, + "loss": 0.7498, + "step": 3534 + }, + { + "epoch": 1.9191096634093376, + "grad_norm": 11.105255757566027, + "learning_rate": 1.412329932074427e-05, + "loss": 0.7587, + "step": 3535 + }, + { + "epoch": 1.9196525515743756, + "grad_norm": 8.599145695443788, + "learning_rate": 1.4120095103457734e-05, + "loss": 0.934, + "step": 3536 + }, + { + "epoch": 1.9201954397394136, + "grad_norm": 7.727889210926642, + "learning_rate": 1.4116890376591268e-05, + "loss": 0.8305, + "step": 3537 + }, + { + "epoch": 1.9207383279044516, + "grad_norm": 12.029529213228301, + "learning_rate": 1.4113685140541242e-05, + "loss": 1.494, + "step": 3538 + }, + { + "epoch": 1.9212812160694897, + "grad_norm": 7.404838471492053, + "learning_rate": 1.411047939570408e-05, + "loss": 0.5168, + "step": 3539 + }, + { + "epoch": 1.9218241042345277, + "grad_norm": 7.540024472508733, + "learning_rate": 1.4107273142476272e-05, + "loss": 0.7162, + "step": 3540 + }, + { + "epoch": 1.9223669923995657, + "grad_norm": 7.885769945154571, + "learning_rate": 1.4104066381254378e-05, + "loss": 0.5964, + "step": 3541 + }, + { + "epoch": 1.9229098805646037, + "grad_norm": 8.287652724014645, + "learning_rate": 1.4100859112435013e-05, + "loss": 0.6175, + "step": 3542 + }, + { + "epoch": 1.9234527687296417, + "grad_norm": 8.335133777216624, + "learning_rate": 1.4097651336414857e-05, + "loss": 0.7163, + "step": 3543 + }, + { + "epoch": 1.9239956568946797, + "grad_norm": 7.765701865468595, + "learning_rate": 1.4094443053590652e-05, + "loss": 0.4534, + "step": 3544 + }, + { + "epoch": 1.9245385450597177, + "grad_norm": 7.716709219714201, + "learning_rate": 1.4091234264359206e-05, + "loss": 0.4425, + "step": 3545 + }, + { + "epoch": 1.9250814332247557, + "grad_norm": 11.979255681165771, + "learning_rate": 1.4088024969117387e-05, + "loss": 1.0249, + "step": 3546 + }, + { + "epoch": 1.9256243213897937, + "grad_norm": 9.424753754574079, + "learning_rate": 1.4084815168262123e-05, + "loss": 0.9663, + "step": 3547 + }, + { + "epoch": 1.9261672095548317, + "grad_norm": 10.101408847774994, + "learning_rate": 1.4081604862190407e-05, + "loss": 0.8136, + "step": 3548 + }, + { + "epoch": 1.9267100977198697, + "grad_norm": 11.755751498543427, + "learning_rate": 1.4078394051299298e-05, + "loss": 0.9451, + "step": 3549 + }, + { + "epoch": 1.9272529858849077, + "grad_norm": 12.251136923575924, + "learning_rate": 1.4075182735985913e-05, + "loss": 1.5716, + "step": 3550 + }, + { + "epoch": 1.9277958740499457, + "grad_norm": 10.087922707091717, + "learning_rate": 1.4071970916647432e-05, + "loss": 0.7026, + "step": 3551 + }, + { + "epoch": 1.9283387622149837, + "grad_norm": 8.507872704310849, + "learning_rate": 1.40687585936811e-05, + "loss": 0.7016, + "step": 3552 + }, + { + "epoch": 1.9288816503800217, + "grad_norm": 8.98138340766033, + "learning_rate": 1.4065545767484218e-05, + "loss": 0.6834, + "step": 3553 + }, + { + "epoch": 1.9294245385450597, + "grad_norm": 7.1085330003672045, + "learning_rate": 1.4062332438454156e-05, + "loss": 0.6813, + "step": 3554 + }, + { + "epoch": 1.9299674267100977, + "grad_norm": 13.499346110936985, + "learning_rate": 1.4059118606988345e-05, + "loss": 1.0068, + "step": 3555 + }, + { + "epoch": 1.9305103148751357, + "grad_norm": 7.698407375447584, + "learning_rate": 1.4055904273484275e-05, + "loss": 0.8418, + "step": 3556 + }, + { + "epoch": 1.9310532030401737, + "grad_norm": 8.424294783097983, + "learning_rate": 1.40526894383395e-05, + "loss": 0.8087, + "step": 3557 + }, + { + "epoch": 1.9315960912052117, + "grad_norm": 8.962780671623113, + "learning_rate": 1.4049474101951639e-05, + "loss": 0.8644, + "step": 3558 + }, + { + "epoch": 1.9321389793702497, + "grad_norm": 8.580581507909985, + "learning_rate": 1.4046258264718363e-05, + "loss": 1.0325, + "step": 3559 + }, + { + "epoch": 1.9326818675352877, + "grad_norm": 10.417878574145533, + "learning_rate": 1.4043041927037418e-05, + "loss": 0.7705, + "step": 3560 + }, + { + "epoch": 1.9332247557003257, + "grad_norm": 8.250628073154727, + "learning_rate": 1.4039825089306603e-05, + "loss": 0.794, + "step": 3561 + }, + { + "epoch": 1.9337676438653637, + "grad_norm": 8.228275233396532, + "learning_rate": 1.4036607751923783e-05, + "loss": 0.8759, + "step": 3562 + }, + { + "epoch": 1.9343105320304017, + "grad_norm": 7.760702739983814, + "learning_rate": 1.4033389915286884e-05, + "loss": 0.6458, + "step": 3563 + }, + { + "epoch": 1.9348534201954397, + "grad_norm": 6.508974748032192, + "learning_rate": 1.403017157979389e-05, + "loss": 0.6348, + "step": 3564 + }, + { + "epoch": 1.9353963083604777, + "grad_norm": 11.570824985629717, + "learning_rate": 1.4026952745842853e-05, + "loss": 1.0935, + "step": 3565 + }, + { + "epoch": 1.9359391965255157, + "grad_norm": 9.901046454739335, + "learning_rate": 1.4023733413831881e-05, + "loss": 0.7458, + "step": 3566 + }, + { + "epoch": 1.9364820846905537, + "grad_norm": 9.103951624002182, + "learning_rate": 1.402051358415915e-05, + "loss": 0.8253, + "step": 3567 + }, + { + "epoch": 1.9370249728555917, + "grad_norm": 9.293062225974824, + "learning_rate": 1.4017293257222887e-05, + "loss": 0.5801, + "step": 3568 + }, + { + "epoch": 1.9375678610206297, + "grad_norm": 7.180133735690273, + "learning_rate": 1.4014072433421398e-05, + "loss": 0.6063, + "step": 3569 + }, + { + "epoch": 1.9381107491856677, + "grad_norm": 9.028614205338394, + "learning_rate": 1.4010851113153028e-05, + "loss": 0.9412, + "step": 3570 + }, + { + "epoch": 1.9386536373507057, + "grad_norm": 9.51192022384654, + "learning_rate": 1.4007629296816202e-05, + "loss": 0.6459, + "step": 3571 + }, + { + "epoch": 1.9391965255157437, + "grad_norm": 9.883078787028563, + "learning_rate": 1.4004406984809396e-05, + "loss": 1.0324, + "step": 3572 + }, + { + "epoch": 1.9397394136807817, + "grad_norm": 7.433318187249177, + "learning_rate": 1.4001184177531154e-05, + "loss": 0.6152, + "step": 3573 + }, + { + "epoch": 1.9402823018458197, + "grad_norm": 10.165800890508075, + "learning_rate": 1.399796087538007e-05, + "loss": 0.8751, + "step": 3574 + }, + { + "epoch": 1.9408251900108577, + "grad_norm": 7.668138219396951, + "learning_rate": 1.3994737078754819e-05, + "loss": 0.43, + "step": 3575 + }, + { + "epoch": 1.9413680781758957, + "grad_norm": 8.929591410847367, + "learning_rate": 1.3991512788054115e-05, + "loss": 0.6173, + "step": 3576 + }, + { + "epoch": 1.9419109663409337, + "grad_norm": 8.069899029370944, + "learning_rate": 1.3988288003676755e-05, + "loss": 0.6782, + "step": 3577 + }, + { + "epoch": 1.9424538545059717, + "grad_norm": 7.635775018643444, + "learning_rate": 1.3985062726021574e-05, + "loss": 0.7503, + "step": 3578 + }, + { + "epoch": 1.9429967426710097, + "grad_norm": 11.649932328973133, + "learning_rate": 1.3981836955487485e-05, + "loss": 0.9154, + "step": 3579 + }, + { + "epoch": 1.9435396308360477, + "grad_norm": 9.11865472044189, + "learning_rate": 1.397861069247345e-05, + "loss": 0.8799, + "step": 3580 + }, + { + "epoch": 1.9440825190010858, + "grad_norm": 7.628563459317163, + "learning_rate": 1.3975383937378508e-05, + "loss": 0.7199, + "step": 3581 + }, + { + "epoch": 1.9446254071661238, + "grad_norm": 10.640127088045624, + "learning_rate": 1.3972156690601747e-05, + "loss": 0.7885, + "step": 3582 + }, + { + "epoch": 1.9451682953311618, + "grad_norm": 10.192406733466123, + "learning_rate": 1.3968928952542313e-05, + "loss": 0.6036, + "step": 3583 + }, + { + "epoch": 1.9457111834961998, + "grad_norm": 12.020287450972278, + "learning_rate": 1.396570072359942e-05, + "loss": 0.9477, + "step": 3584 + }, + { + "epoch": 1.9462540716612378, + "grad_norm": 8.855072111194623, + "learning_rate": 1.3962472004172343e-05, + "loss": 0.4384, + "step": 3585 + }, + { + "epoch": 1.9467969598262758, + "grad_norm": 7.173236338345466, + "learning_rate": 1.3959242794660412e-05, + "loss": 0.4969, + "step": 3586 + }, + { + "epoch": 1.9473398479913138, + "grad_norm": 6.505005692892076, + "learning_rate": 1.3956013095463024e-05, + "loss": 0.4585, + "step": 3587 + }, + { + "epoch": 1.9478827361563518, + "grad_norm": 7.041517208688192, + "learning_rate": 1.395278290697963e-05, + "loss": 0.5439, + "step": 3588 + }, + { + "epoch": 1.9484256243213898, + "grad_norm": 9.730465359550896, + "learning_rate": 1.3949552229609746e-05, + "loss": 0.8865, + "step": 3589 + }, + { + "epoch": 1.9489685124864278, + "grad_norm": 7.73678668778, + "learning_rate": 1.3946321063752948e-05, + "loss": 0.5094, + "step": 3590 + }, + { + "epoch": 1.9495114006514658, + "grad_norm": 7.095011334397053, + "learning_rate": 1.3943089409808872e-05, + "loss": 0.5224, + "step": 3591 + }, + { + "epoch": 1.9500542888165038, + "grad_norm": 8.086388468165458, + "learning_rate": 1.393985726817721e-05, + "loss": 0.7798, + "step": 3592 + }, + { + "epoch": 1.9505971769815418, + "grad_norm": 10.262604564335476, + "learning_rate": 1.3936624639257726e-05, + "loss": 1.2595, + "step": 3593 + }, + { + "epoch": 1.9511400651465798, + "grad_norm": 8.535907967530573, + "learning_rate": 1.393339152345023e-05, + "loss": 0.8205, + "step": 3594 + }, + { + "epoch": 1.9516829533116178, + "grad_norm": 7.864477340620307, + "learning_rate": 1.3930157921154601e-05, + "loss": 0.5967, + "step": 3595 + }, + { + "epoch": 1.9522258414766558, + "grad_norm": 6.739275201436143, + "learning_rate": 1.392692383277078e-05, + "loss": 0.5836, + "step": 3596 + }, + { + "epoch": 1.9527687296416938, + "grad_norm": 9.533970605538912, + "learning_rate": 1.392368925869876e-05, + "loss": 0.5761, + "step": 3597 + }, + { + "epoch": 1.9533116178067318, + "grad_norm": 11.376343280359167, + "learning_rate": 1.3920454199338598e-05, + "loss": 0.7596, + "step": 3598 + }, + { + "epoch": 1.9538545059717698, + "grad_norm": 7.280453649201749, + "learning_rate": 1.391721865509041e-05, + "loss": 0.4949, + "step": 3599 + }, + { + "epoch": 1.9543973941368078, + "grad_norm": 7.734190007799969, + "learning_rate": 1.391398262635438e-05, + "loss": 0.9177, + "step": 3600 + }, + { + "epoch": 1.9549402823018458, + "grad_norm": 13.106124045444247, + "learning_rate": 1.3910746113530738e-05, + "loss": 1.1884, + "step": 3601 + }, + { + "epoch": 1.9554831704668838, + "grad_norm": 11.868912051178873, + "learning_rate": 1.3907509117019783e-05, + "loss": 0.9245, + "step": 3602 + }, + { + "epoch": 1.9560260586319218, + "grad_norm": 10.307498071834987, + "learning_rate": 1.3904271637221876e-05, + "loss": 1.0802, + "step": 3603 + }, + { + "epoch": 1.9565689467969598, + "grad_norm": 8.308697157532139, + "learning_rate": 1.390103367453743e-05, + "loss": 0.4979, + "step": 3604 + }, + { + "epoch": 1.9571118349619978, + "grad_norm": 8.447604525372762, + "learning_rate": 1.3897795229366919e-05, + "loss": 0.8942, + "step": 3605 + }, + { + "epoch": 1.9576547231270358, + "grad_norm": 7.864013969756676, + "learning_rate": 1.3894556302110883e-05, + "loss": 0.642, + "step": 3606 + }, + { + "epoch": 1.9581976112920738, + "grad_norm": 7.903822544650853, + "learning_rate": 1.389131689316992e-05, + "loss": 0.8251, + "step": 3607 + }, + { + "epoch": 1.9587404994571118, + "grad_norm": 9.744275276467558, + "learning_rate": 1.3888077002944678e-05, + "loss": 1.0052, + "step": 3608 + }, + { + "epoch": 1.9592833876221498, + "grad_norm": 7.731164092689308, + "learning_rate": 1.3884836631835877e-05, + "loss": 0.6883, + "step": 3609 + }, + { + "epoch": 1.9598262757871878, + "grad_norm": 8.849291385162568, + "learning_rate": 1.3881595780244288e-05, + "loss": 0.9985, + "step": 3610 + }, + { + "epoch": 1.9603691639522258, + "grad_norm": 9.531929458494986, + "learning_rate": 1.3878354448570748e-05, + "loss": 0.7643, + "step": 3611 + }, + { + "epoch": 1.9609120521172638, + "grad_norm": 10.32443732773501, + "learning_rate": 1.3875112637216145e-05, + "loss": 0.848, + "step": 3612 + }, + { + "epoch": 1.9614549402823018, + "grad_norm": 7.417155617136626, + "learning_rate": 1.387187034658144e-05, + "loss": 0.5682, + "step": 3613 + }, + { + "epoch": 1.9619978284473398, + "grad_norm": 14.072831276291689, + "learning_rate": 1.386862757706764e-05, + "loss": 0.8279, + "step": 3614 + }, + { + "epoch": 1.9625407166123778, + "grad_norm": 7.764451928370161, + "learning_rate": 1.3865384329075812e-05, + "loss": 0.6687, + "step": 3615 + }, + { + "epoch": 1.9630836047774158, + "grad_norm": 8.686094567618806, + "learning_rate": 1.3862140603007095e-05, + "loss": 1.1034, + "step": 3616 + }, + { + "epoch": 1.9636264929424538, + "grad_norm": 10.753407337858437, + "learning_rate": 1.3858896399262669e-05, + "loss": 1.1692, + "step": 3617 + }, + { + "epoch": 1.9641693811074918, + "grad_norm": 9.096005648867607, + "learning_rate": 1.3855651718243786e-05, + "loss": 0.6191, + "step": 3618 + }, + { + "epoch": 1.9647122692725298, + "grad_norm": 10.387647049254962, + "learning_rate": 1.3852406560351752e-05, + "loss": 1.3573, + "step": 3619 + }, + { + "epoch": 1.9652551574375678, + "grad_norm": 7.173958455599766, + "learning_rate": 1.3849160925987936e-05, + "loss": 0.8106, + "step": 3620 + }, + { + "epoch": 1.9657980456026058, + "grad_norm": 9.434903281870886, + "learning_rate": 1.3845914815553765e-05, + "loss": 0.9581, + "step": 3621 + }, + { + "epoch": 1.9663409337676439, + "grad_norm": 7.671221453670278, + "learning_rate": 1.3842668229450717e-05, + "loss": 0.6989, + "step": 3622 + }, + { + "epoch": 1.9668838219326819, + "grad_norm": 7.360534692664803, + "learning_rate": 1.3839421168080338e-05, + "loss": 0.4912, + "step": 3623 + }, + { + "epoch": 1.9674267100977199, + "grad_norm": 6.853527063108637, + "learning_rate": 1.3836173631844231e-05, + "loss": 0.7485, + "step": 3624 + }, + { + "epoch": 1.9679695982627579, + "grad_norm": 8.619792011062271, + "learning_rate": 1.3832925621144057e-05, + "loss": 0.712, + "step": 3625 + }, + { + "epoch": 1.9685124864277959, + "grad_norm": 8.062990147577946, + "learning_rate": 1.3829677136381532e-05, + "loss": 0.7126, + "step": 3626 + }, + { + "epoch": 1.9690553745928339, + "grad_norm": 6.35400375341867, + "learning_rate": 1.3826428177958433e-05, + "loss": 0.5262, + "step": 3627 + }, + { + "epoch": 1.9695982627578719, + "grad_norm": 7.322594621519839, + "learning_rate": 1.3823178746276603e-05, + "loss": 0.6644, + "step": 3628 + }, + { + "epoch": 1.9701411509229099, + "grad_norm": 9.976906946653425, + "learning_rate": 1.3819928841737929e-05, + "loss": 0.7411, + "step": 3629 + }, + { + "epoch": 1.9706840390879479, + "grad_norm": 11.052922199516026, + "learning_rate": 1.3816678464744368e-05, + "loss": 0.9998, + "step": 3630 + }, + { + "epoch": 1.9712269272529859, + "grad_norm": 7.695988647358978, + "learning_rate": 1.381342761569793e-05, + "loss": 0.6805, + "step": 3631 + }, + { + "epoch": 1.9717698154180239, + "grad_norm": 10.387946538947885, + "learning_rate": 1.381017629500069e-05, + "loss": 0.8954, + "step": 3632 + }, + { + "epoch": 1.9723127035830619, + "grad_norm": 8.215553841008688, + "learning_rate": 1.380692450305477e-05, + "loss": 1.0488, + "step": 3633 + }, + { + "epoch": 1.9728555917480999, + "grad_norm": 8.550898170705977, + "learning_rate": 1.3803672240262364e-05, + "loss": 0.9301, + "step": 3634 + }, + { + "epoch": 1.9733984799131379, + "grad_norm": 9.978410770263547, + "learning_rate": 1.380041950702571e-05, + "loss": 0.9725, + "step": 3635 + }, + { + "epoch": 1.9739413680781759, + "grad_norm": 10.211160158432088, + "learning_rate": 1.3797166303747119e-05, + "loss": 1.1789, + "step": 3636 + }, + { + "epoch": 1.9744842562432139, + "grad_norm": 7.47091491709439, + "learning_rate": 1.3793912630828942e-05, + "loss": 1.2311, + "step": 3637 + }, + { + "epoch": 1.975027144408252, + "grad_norm": 9.966534074490724, + "learning_rate": 1.3790658488673607e-05, + "loss": 1.0057, + "step": 3638 + }, + { + "epoch": 1.97557003257329, + "grad_norm": 9.341517091004942, + "learning_rate": 1.378740387768359e-05, + "loss": 0.7708, + "step": 3639 + }, + { + "epoch": 1.976112920738328, + "grad_norm": 10.066755071795829, + "learning_rate": 1.3784148798261422e-05, + "loss": 1.3146, + "step": 3640 + }, + { + "epoch": 1.976655808903366, + "grad_norm": 8.306959980081418, + "learning_rate": 1.3780893250809705e-05, + "loss": 0.6926, + "step": 3641 + }, + { + "epoch": 1.977198697068404, + "grad_norm": 5.980548702977229, + "learning_rate": 1.377763723573108e-05, + "loss": 0.6776, + "step": 3642 + }, + { + "epoch": 1.977741585233442, + "grad_norm": 10.796897781225315, + "learning_rate": 1.3774380753428263e-05, + "loss": 1.0422, + "step": 3643 + }, + { + "epoch": 1.97828447339848, + "grad_norm": 9.650854393967004, + "learning_rate": 1.3771123804304018e-05, + "loss": 1.2566, + "step": 3644 + }, + { + "epoch": 1.978827361563518, + "grad_norm": 6.718663747121732, + "learning_rate": 1.3767866388761168e-05, + "loss": 0.5018, + "step": 3645 + }, + { + "epoch": 1.979370249728556, + "grad_norm": 8.345918705516796, + "learning_rate": 1.3764608507202604e-05, + "loss": 0.8501, + "step": 3646 + }, + { + "epoch": 1.979913137893594, + "grad_norm": 11.065999609001787, + "learning_rate": 1.3761350160031258e-05, + "loss": 1.2627, + "step": 3647 + }, + { + "epoch": 1.980456026058632, + "grad_norm": 8.567448073304528, + "learning_rate": 1.3758091347650126e-05, + "loss": 0.4477, + "step": 3648 + }, + { + "epoch": 1.98099891422367, + "grad_norm": 11.930236219703469, + "learning_rate": 1.3754832070462269e-05, + "loss": 1.5386, + "step": 3649 + }, + { + "epoch": 1.981541802388708, + "grad_norm": 7.334841489157127, + "learning_rate": 1.3751572328870797e-05, + "loss": 0.5759, + "step": 3650 + }, + { + "epoch": 1.982084690553746, + "grad_norm": 6.392865090419794, + "learning_rate": 1.3748312123278879e-05, + "loss": 0.6078, + "step": 3651 + }, + { + "epoch": 1.982627578718784, + "grad_norm": 8.598388892264492, + "learning_rate": 1.3745051454089744e-05, + "loss": 0.7682, + "step": 3652 + }, + { + "epoch": 1.983170466883822, + "grad_norm": 9.310542957266444, + "learning_rate": 1.3741790321706678e-05, + "loss": 0.9833, + "step": 3653 + }, + { + "epoch": 1.98371335504886, + "grad_norm": 8.917570196452148, + "learning_rate": 1.3738528726533021e-05, + "loss": 0.5915, + "step": 3654 + }, + { + "epoch": 1.984256243213898, + "grad_norm": 9.503202436394638, + "learning_rate": 1.3735266668972174e-05, + "loss": 0.9347, + "step": 3655 + }, + { + "epoch": 1.984799131378936, + "grad_norm": 9.187024367480415, + "learning_rate": 1.3732004149427592e-05, + "loss": 0.6669, + "step": 3656 + }, + { + "epoch": 1.985342019543974, + "grad_norm": 10.497812893691144, + "learning_rate": 1.3728741168302785e-05, + "loss": 1.0663, + "step": 3657 + }, + { + "epoch": 1.985884907709012, + "grad_norm": 9.646030688042543, + "learning_rate": 1.3725477726001332e-05, + "loss": 0.9913, + "step": 3658 + }, + { + "epoch": 1.98642779587405, + "grad_norm": 9.475431343889321, + "learning_rate": 1.3722213822926855e-05, + "loss": 0.8377, + "step": 3659 + }, + { + "epoch": 1.986970684039088, + "grad_norm": 9.449591675048769, + "learning_rate": 1.371894945948304e-05, + "loss": 0.8763, + "step": 3660 + }, + { + "epoch": 1.987513572204126, + "grad_norm": 7.3204142422253025, + "learning_rate": 1.3715684636073628e-05, + "loss": 0.7384, + "step": 3661 + }, + { + "epoch": 1.988056460369164, + "grad_norm": 10.339298014193695, + "learning_rate": 1.371241935310242e-05, + "loss": 1.0206, + "step": 3662 + }, + { + "epoch": 1.988599348534202, + "grad_norm": 7.083819102334954, + "learning_rate": 1.3709153610973266e-05, + "loss": 0.6905, + "step": 3663 + }, + { + "epoch": 1.98914223669924, + "grad_norm": 6.318629816327554, + "learning_rate": 1.3705887410090085e-05, + "loss": 0.7946, + "step": 3664 + }, + { + "epoch": 1.989685124864278, + "grad_norm": 7.109392074745231, + "learning_rate": 1.3702620750856843e-05, + "loss": 0.7125, + "step": 3665 + }, + { + "epoch": 1.990228013029316, + "grad_norm": 6.7832401368423545, + "learning_rate": 1.3699353633677565e-05, + "loss": 0.6091, + "step": 3666 + }, + { + "epoch": 1.990770901194354, + "grad_norm": 9.301471508167843, + "learning_rate": 1.3696086058956333e-05, + "loss": 0.9335, + "step": 3667 + }, + { + "epoch": 1.991313789359392, + "grad_norm": 8.167349191349723, + "learning_rate": 1.3692818027097288e-05, + "loss": 0.7148, + "step": 3668 + }, + { + "epoch": 1.99185667752443, + "grad_norm": 7.582331201930941, + "learning_rate": 1.3689549538504622e-05, + "loss": 0.5952, + "step": 3669 + }, + { + "epoch": 1.992399565689468, + "grad_norm": 8.743965167551632, + "learning_rate": 1.3686280593582588e-05, + "loss": 0.9268, + "step": 3670 + }, + { + "epoch": 1.992942453854506, + "grad_norm": 10.291738284427778, + "learning_rate": 1.3683011192735496e-05, + "loss": 0.9934, + "step": 3671 + }, + { + "epoch": 1.993485342019544, + "grad_norm": 9.808936601876956, + "learning_rate": 1.3679741336367711e-05, + "loss": 1.0506, + "step": 3672 + }, + { + "epoch": 1.994028230184582, + "grad_norm": 10.002691567142412, + "learning_rate": 1.3676471024883654e-05, + "loss": 1.1439, + "step": 3673 + }, + { + "epoch": 1.99457111834962, + "grad_norm": 7.558103827510508, + "learning_rate": 1.36732002586878e-05, + "loss": 1.0059, + "step": 3674 + }, + { + "epoch": 1.995114006514658, + "grad_norm": 7.958890420546252, + "learning_rate": 1.3669929038184684e-05, + "loss": 1.0301, + "step": 3675 + }, + { + "epoch": 1.995656894679696, + "grad_norm": 8.042682701022132, + "learning_rate": 1.3666657363778895e-05, + "loss": 0.6881, + "step": 3676 + }, + { + "epoch": 1.996199782844734, + "grad_norm": 7.54271762136155, + "learning_rate": 1.366338523587508e-05, + "loss": 0.5227, + "step": 3677 + }, + { + "epoch": 1.996742671009772, + "grad_norm": 9.060117135659972, + "learning_rate": 1.3660112654877939e-05, + "loss": 0.9847, + "step": 3678 + }, + { + "epoch": 1.99728555917481, + "grad_norm": 11.00270077841202, + "learning_rate": 1.3656839621192233e-05, + "loss": 1.0918, + "step": 3679 + }, + { + "epoch": 1.997828447339848, + "grad_norm": 9.156434005073692, + "learning_rate": 1.3653566135222774e-05, + "loss": 0.8104, + "step": 3680 + }, + { + "epoch": 1.998371335504886, + "grad_norm": 8.405413514743108, + "learning_rate": 1.3650292197374433e-05, + "loss": 0.9414, + "step": 3681 + }, + { + "epoch": 1.998914223669924, + "grad_norm": 7.9594907003419495, + "learning_rate": 1.3647017808052135e-05, + "loss": 0.9257, + "step": 3682 + }, + { + "epoch": 1.999457111834962, + "grad_norm": 9.821777823228977, + "learning_rate": 1.3643742967660859e-05, + "loss": 0.8787, + "step": 3683 + }, + { + "epoch": 2.0, + "grad_norm": 8.766507481527489, + "learning_rate": 1.3640467676605648e-05, + "loss": 0.8472, + "step": 3684 + }, + { + "epoch": 2.000542888165038, + "grad_norm": 9.493942149196295, + "learning_rate": 1.3637191935291596e-05, + "loss": 0.8147, + "step": 3685 + }, + { + "epoch": 2.001085776330076, + "grad_norm": 9.782638508357579, + "learning_rate": 1.3633915744123844e-05, + "loss": 0.8788, + "step": 3686 + }, + { + "epoch": 2.001628664495114, + "grad_norm": 8.46210638839138, + "learning_rate": 1.3630639103507604e-05, + "loss": 1.078, + "step": 3687 + }, + { + "epoch": 2.002171552660152, + "grad_norm": 7.345658753018104, + "learning_rate": 1.3627362013848134e-05, + "loss": 0.6838, + "step": 3688 + }, + { + "epoch": 2.00271444082519, + "grad_norm": 7.329182601936008, + "learning_rate": 1.3624084475550743e-05, + "loss": 0.7249, + "step": 3689 + }, + { + "epoch": 2.003257328990228, + "grad_norm": 9.89510466995638, + "learning_rate": 1.3620806489020813e-05, + "loss": 0.8018, + "step": 3690 + }, + { + "epoch": 2.003800217155266, + "grad_norm": 6.836942603575798, + "learning_rate": 1.3617528054663764e-05, + "loss": 0.6645, + "step": 3691 + }, + { + "epoch": 2.004343105320304, + "grad_norm": 8.842045073299985, + "learning_rate": 1.3614249172885081e-05, + "loss": 0.5143, + "step": 3692 + }, + { + "epoch": 2.004885993485342, + "grad_norm": 7.671200527248645, + "learning_rate": 1.36109698440903e-05, + "loss": 0.8038, + "step": 3693 + }, + { + "epoch": 2.00542888165038, + "grad_norm": 9.544464428119456, + "learning_rate": 1.3607690068685013e-05, + "loss": 0.6396, + "step": 3694 + }, + { + "epoch": 2.005971769815418, + "grad_norm": 8.537202930183216, + "learning_rate": 1.3604409847074868e-05, + "loss": 0.7104, + "step": 3695 + }, + { + "epoch": 2.006514657980456, + "grad_norm": 7.121091005738386, + "learning_rate": 1.3601129179665572e-05, + "loss": 0.4337, + "step": 3696 + }, + { + "epoch": 2.007057546145494, + "grad_norm": 9.740494539210806, + "learning_rate": 1.3597848066862875e-05, + "loss": 0.9114, + "step": 3697 + }, + { + "epoch": 2.007600434310532, + "grad_norm": 8.100951122902112, + "learning_rate": 1.3594566509072599e-05, + "loss": 0.722, + "step": 3698 + }, + { + "epoch": 2.00814332247557, + "grad_norm": 6.959347625774498, + "learning_rate": 1.3591284506700606e-05, + "loss": 0.5429, + "step": 3699 + }, + { + "epoch": 2.008686210640608, + "grad_norm": 9.993493206941414, + "learning_rate": 1.3588002060152822e-05, + "loss": 0.8251, + "step": 3700 + }, + { + "epoch": 2.009229098805646, + "grad_norm": 9.966526343412045, + "learning_rate": 1.3584719169835226e-05, + "loss": 0.4847, + "step": 3701 + }, + { + "epoch": 2.009771986970684, + "grad_norm": 7.888503709599375, + "learning_rate": 1.3581435836153847e-05, + "loss": 0.7941, + "step": 3702 + }, + { + "epoch": 2.010314875135722, + "grad_norm": 9.057972693252164, + "learning_rate": 1.3578152059514778e-05, + "loss": 0.8963, + "step": 3703 + }, + { + "epoch": 2.01085776330076, + "grad_norm": 10.717910066603423, + "learning_rate": 1.3574867840324157e-05, + "loss": 0.6485, + "step": 3704 + }, + { + "epoch": 2.011400651465798, + "grad_norm": 7.257931643777971, + "learning_rate": 1.3571583178988188e-05, + "loss": 0.4569, + "step": 3705 + }, + { + "epoch": 2.011943539630836, + "grad_norm": 16.546104792345954, + "learning_rate": 1.3568298075913119e-05, + "loss": 1.0473, + "step": 3706 + }, + { + "epoch": 2.012486427795874, + "grad_norm": 9.882729561546991, + "learning_rate": 1.3565012531505252e-05, + "loss": 0.9729, + "step": 3707 + }, + { + "epoch": 2.013029315960912, + "grad_norm": 7.060037163975289, + "learning_rate": 1.3561726546170956e-05, + "loss": 0.5935, + "step": 3708 + }, + { + "epoch": 2.01357220412595, + "grad_norm": 7.081117948590212, + "learning_rate": 1.3558440120316644e-05, + "loss": 0.6676, + "step": 3709 + }, + { + "epoch": 2.014115092290988, + "grad_norm": 10.048606586375149, + "learning_rate": 1.3555153254348788e-05, + "loss": 0.6307, + "step": 3710 + }, + { + "epoch": 2.014657980456026, + "grad_norm": 7.720688031890385, + "learning_rate": 1.3551865948673912e-05, + "loss": 0.7882, + "step": 3711 + }, + { + "epoch": 2.015200868621064, + "grad_norm": 7.644495073412038, + "learning_rate": 1.3548578203698592e-05, + "loss": 0.8904, + "step": 3712 + }, + { + "epoch": 2.015743756786102, + "grad_norm": 10.13216917020047, + "learning_rate": 1.3545290019829466e-05, + "loss": 0.7832, + "step": 3713 + }, + { + "epoch": 2.01628664495114, + "grad_norm": 6.321686691568546, + "learning_rate": 1.3542001397473219e-05, + "loss": 0.5682, + "step": 3714 + }, + { + "epoch": 2.016829533116178, + "grad_norm": 7.966895819853924, + "learning_rate": 1.3538712337036594e-05, + "loss": 0.7704, + "step": 3715 + }, + { + "epoch": 2.017372421281216, + "grad_norm": 8.827713158118389, + "learning_rate": 1.3535422838926389e-05, + "loss": 0.6741, + "step": 3716 + }, + { + "epoch": 2.017915309446254, + "grad_norm": 7.404288938912365, + "learning_rate": 1.3532132903549453e-05, + "loss": 0.5325, + "step": 3717 + }, + { + "epoch": 2.018458197611292, + "grad_norm": 9.869324924245625, + "learning_rate": 1.352884253131269e-05, + "loss": 0.7326, + "step": 3718 + }, + { + "epoch": 2.01900108577633, + "grad_norm": 7.824783976445888, + "learning_rate": 1.3525551722623056e-05, + "loss": 0.4683, + "step": 3719 + }, + { + "epoch": 2.019543973941368, + "grad_norm": 10.516375854488775, + "learning_rate": 1.3522260477887566e-05, + "loss": 0.839, + "step": 3720 + }, + { + "epoch": 2.020086862106406, + "grad_norm": 10.698678918246005, + "learning_rate": 1.3518968797513288e-05, + "loss": 1.0513, + "step": 3721 + }, + { + "epoch": 2.020629750271444, + "grad_norm": 9.430167728870957, + "learning_rate": 1.351567668190734e-05, + "loss": 0.6696, + "step": 3722 + }, + { + "epoch": 2.021172638436482, + "grad_norm": 9.619270238605017, + "learning_rate": 1.3512384131476897e-05, + "loss": 0.4608, + "step": 3723 + }, + { + "epoch": 2.02171552660152, + "grad_norm": 6.9328636261250525, + "learning_rate": 1.350909114662919e-05, + "loss": 0.3952, + "step": 3724 + }, + { + "epoch": 2.022258414766558, + "grad_norm": 11.694555068035188, + "learning_rate": 1.3505797727771493e-05, + "loss": 0.7526, + "step": 3725 + }, + { + "epoch": 2.022801302931596, + "grad_norm": 7.417427417334792, + "learning_rate": 1.3502503875311149e-05, + "loss": 0.5239, + "step": 3726 + }, + { + "epoch": 2.023344191096634, + "grad_norm": 10.99931861711161, + "learning_rate": 1.349920958965554e-05, + "loss": 0.5504, + "step": 3727 + }, + { + "epoch": 2.023887079261672, + "grad_norm": 9.418315087755559, + "learning_rate": 1.3495914871212113e-05, + "loss": 0.6791, + "step": 3728 + }, + { + "epoch": 2.02442996742671, + "grad_norm": 12.589472190346669, + "learning_rate": 1.3492619720388363e-05, + "loss": 0.815, + "step": 3729 + }, + { + "epoch": 2.024972855591748, + "grad_norm": 6.60691672927324, + "learning_rate": 1.348932413759184e-05, + "loss": 0.4112, + "step": 3730 + }, + { + "epoch": 2.025515743756786, + "grad_norm": 10.970405529985833, + "learning_rate": 1.3486028123230145e-05, + "loss": 0.6164, + "step": 3731 + }, + { + "epoch": 2.026058631921824, + "grad_norm": 9.419473243249211, + "learning_rate": 1.3482731677710938e-05, + "loss": 0.829, + "step": 3732 + }, + { + "epoch": 2.026601520086862, + "grad_norm": 8.809487998193466, + "learning_rate": 1.3479434801441925e-05, + "loss": 0.6386, + "step": 3733 + }, + { + "epoch": 2.0271444082519, + "grad_norm": 8.586869223296738, + "learning_rate": 1.3476137494830872e-05, + "loss": 0.4586, + "step": 3734 + }, + { + "epoch": 2.027687296416938, + "grad_norm": 10.792985384487755, + "learning_rate": 1.3472839758285595e-05, + "loss": 0.68, + "step": 3735 + }, + { + "epoch": 2.028230184581976, + "grad_norm": 11.494014730485675, + "learning_rate": 1.346954159221396e-05, + "loss": 0.9465, + "step": 3736 + }, + { + "epoch": 2.028773072747014, + "grad_norm": 9.047474116723855, + "learning_rate": 1.3466242997023891e-05, + "loss": 0.637, + "step": 3737 + }, + { + "epoch": 2.029315960912052, + "grad_norm": 7.32591412143244, + "learning_rate": 1.3462943973123362e-05, + "loss": 0.5071, + "step": 3738 + }, + { + "epoch": 2.02985884907709, + "grad_norm": 8.462579835722188, + "learning_rate": 1.3459644520920405e-05, + "loss": 0.5568, + "step": 3739 + }, + { + "epoch": 2.030401737242128, + "grad_norm": 11.667085873745426, + "learning_rate": 1.34563446408231e-05, + "loss": 0.579, + "step": 3740 + }, + { + "epoch": 2.030944625407166, + "grad_norm": 6.778347309469529, + "learning_rate": 1.3453044333239577e-05, + "loss": 0.3399, + "step": 3741 + }, + { + "epoch": 2.031487513572204, + "grad_norm": 7.280157509234148, + "learning_rate": 1.3449743598578033e-05, + "loss": 0.4236, + "step": 3742 + }, + { + "epoch": 2.032030401737242, + "grad_norm": 10.902366947251025, + "learning_rate": 1.34464424372467e-05, + "loss": 0.7984, + "step": 3743 + }, + { + "epoch": 2.03257328990228, + "grad_norm": 7.729938792603154, + "learning_rate": 1.344314084965388e-05, + "loss": 0.5115, + "step": 3744 + }, + { + "epoch": 2.033116178067318, + "grad_norm": 9.034690624481534, + "learning_rate": 1.3439838836207905e-05, + "loss": 0.3528, + "step": 3745 + }, + { + "epoch": 2.033659066232356, + "grad_norm": 10.813371264470662, + "learning_rate": 1.3436536397317183e-05, + "loss": 0.4972, + "step": 3746 + }, + { + "epoch": 2.034201954397394, + "grad_norm": 12.116428074423327, + "learning_rate": 1.343323353339016e-05, + "loss": 0.9414, + "step": 3747 + }, + { + "epoch": 2.034744842562432, + "grad_norm": 11.714995187138708, + "learning_rate": 1.3429930244835343e-05, + "loss": 0.7842, + "step": 3748 + }, + { + "epoch": 2.03528773072747, + "grad_norm": 9.956102297397003, + "learning_rate": 1.3426626532061287e-05, + "loss": 0.7304, + "step": 3749 + }, + { + "epoch": 2.035830618892508, + "grad_norm": 10.670014393137768, + "learning_rate": 1.34233223954766e-05, + "loss": 0.746, + "step": 3750 + }, + { + "epoch": 2.036373507057546, + "grad_norm": 10.266437034084714, + "learning_rate": 1.3420017835489945e-05, + "loss": 1.0871, + "step": 3751 + }, + { + "epoch": 2.036916395222584, + "grad_norm": 11.75190588784515, + "learning_rate": 1.3416712852510033e-05, + "loss": 0.7606, + "step": 3752 + }, + { + "epoch": 2.037459283387622, + "grad_norm": 11.54288048263274, + "learning_rate": 1.3413407446945627e-05, + "loss": 0.5887, + "step": 3753 + }, + { + "epoch": 2.03800217155266, + "grad_norm": 18.002644323614724, + "learning_rate": 1.3410101619205552e-05, + "loss": 1.0593, + "step": 3754 + }, + { + "epoch": 2.038545059717698, + "grad_norm": 8.97759172274721, + "learning_rate": 1.3406795369698671e-05, + "loss": 0.7194, + "step": 3755 + }, + { + "epoch": 2.039087947882736, + "grad_norm": 8.793507308104033, + "learning_rate": 1.3403488698833912e-05, + "loss": 0.4891, + "step": 3756 + }, + { + "epoch": 2.039630836047774, + "grad_norm": 9.263725202719241, + "learning_rate": 1.3400181607020243e-05, + "loss": 0.8258, + "step": 3757 + }, + { + "epoch": 2.040173724212812, + "grad_norm": 10.568214203997627, + "learning_rate": 1.3396874094666694e-05, + "loss": 0.5841, + "step": 3758 + }, + { + "epoch": 2.04071661237785, + "grad_norm": 10.237711379932081, + "learning_rate": 1.3393566162182346e-05, + "loss": 0.6901, + "step": 3759 + }, + { + "epoch": 2.041259500542888, + "grad_norm": 11.264856346485661, + "learning_rate": 1.3390257809976322e-05, + "loss": 0.8029, + "step": 3760 + }, + { + "epoch": 2.041802388707926, + "grad_norm": 7.504173750209935, + "learning_rate": 1.3386949038457813e-05, + "loss": 0.3739, + "step": 3761 + }, + { + "epoch": 2.042345276872964, + "grad_norm": 8.099570741006275, + "learning_rate": 1.3383639848036044e-05, + "loss": 0.7457, + "step": 3762 + }, + { + "epoch": 2.042888165038002, + "grad_norm": 9.87563463161881, + "learning_rate": 1.3380330239120313e-05, + "loss": 0.632, + "step": 3763 + }, + { + "epoch": 2.04343105320304, + "grad_norm": 9.304965814712034, + "learning_rate": 1.3377020212119946e-05, + "loss": 0.657, + "step": 3764 + }, + { + "epoch": 2.043973941368078, + "grad_norm": 7.738549848691211, + "learning_rate": 1.3373709767444339e-05, + "loss": 0.4198, + "step": 3765 + }, + { + "epoch": 2.044516829533116, + "grad_norm": 11.805824611419341, + "learning_rate": 1.3370398905502928e-05, + "loss": 0.8173, + "step": 3766 + }, + { + "epoch": 2.045059717698154, + "grad_norm": 10.425380865754251, + "learning_rate": 1.3367087626705211e-05, + "loss": 0.7149, + "step": 3767 + }, + { + "epoch": 2.045602605863192, + "grad_norm": 11.884153232685467, + "learning_rate": 1.336377593146073e-05, + "loss": 0.723, + "step": 3768 + }, + { + "epoch": 2.04614549402823, + "grad_norm": 9.663581488842127, + "learning_rate": 1.336046382017908e-05, + "loss": 0.6457, + "step": 3769 + }, + { + "epoch": 2.046688382193268, + "grad_norm": 9.184284700313228, + "learning_rate": 1.335715129326991e-05, + "loss": 0.6645, + "step": 3770 + }, + { + "epoch": 2.047231270358306, + "grad_norm": 7.398657595562234, + "learning_rate": 1.3353838351142915e-05, + "loss": 0.7288, + "step": 3771 + }, + { + "epoch": 2.047774158523344, + "grad_norm": 10.416946087073072, + "learning_rate": 1.335052499420785e-05, + "loss": 0.4974, + "step": 3772 + }, + { + "epoch": 2.048317046688382, + "grad_norm": 8.676282075018099, + "learning_rate": 1.3347211222874514e-05, + "loss": 0.7128, + "step": 3773 + }, + { + "epoch": 2.04885993485342, + "grad_norm": 10.339910800835435, + "learning_rate": 1.3343897037552758e-05, + "loss": 0.6674, + "step": 3774 + }, + { + "epoch": 2.049402823018458, + "grad_norm": 10.976308424624598, + "learning_rate": 1.3340582438652488e-05, + "loss": 0.7683, + "step": 3775 + }, + { + "epoch": 2.049945711183496, + "grad_norm": 11.874329810000683, + "learning_rate": 1.3337267426583658e-05, + "loss": 1.2199, + "step": 3776 + }, + { + "epoch": 2.050488599348534, + "grad_norm": 9.306473243935873, + "learning_rate": 1.3333952001756272e-05, + "loss": 0.8681, + "step": 3777 + }, + { + "epoch": 2.0510314875135722, + "grad_norm": 9.083116427983501, + "learning_rate": 1.333063616458039e-05, + "loss": 0.7512, + "step": 3778 + }, + { + "epoch": 2.05157437567861, + "grad_norm": 5.762711234835306, + "learning_rate": 1.3327319915466119e-05, + "loss": 0.3384, + "step": 3779 + }, + { + "epoch": 2.0521172638436482, + "grad_norm": 7.431272084519616, + "learning_rate": 1.332400325482362e-05, + "loss": 0.5188, + "step": 3780 + }, + { + "epoch": 2.052660152008686, + "grad_norm": 5.736965263179954, + "learning_rate": 1.3320686183063096e-05, + "loss": 0.3476, + "step": 3781 + }, + { + "epoch": 2.0532030401737242, + "grad_norm": 7.130870597882685, + "learning_rate": 1.3317368700594815e-05, + "loss": 0.7929, + "step": 3782 + }, + { + "epoch": 2.053745928338762, + "grad_norm": 7.713657404988891, + "learning_rate": 1.3314050807829088e-05, + "loss": 0.422, + "step": 3783 + }, + { + "epoch": 2.0542888165038002, + "grad_norm": 6.383542293487279, + "learning_rate": 1.3310732505176276e-05, + "loss": 0.568, + "step": 3784 + }, + { + "epoch": 2.054831704668838, + "grad_norm": 12.326747003957852, + "learning_rate": 1.3307413793046787e-05, + "loss": 0.8061, + "step": 3785 + }, + { + "epoch": 2.0553745928338762, + "grad_norm": 10.668098962086818, + "learning_rate": 1.330409467185109e-05, + "loss": 1.4344, + "step": 3786 + }, + { + "epoch": 2.055917480998914, + "grad_norm": 9.844537720560512, + "learning_rate": 1.3300775141999698e-05, + "loss": 1.0089, + "step": 3787 + }, + { + "epoch": 2.0564603691639523, + "grad_norm": 9.714672703692647, + "learning_rate": 1.3297455203903176e-05, + "loss": 0.6069, + "step": 3788 + }, + { + "epoch": 2.05700325732899, + "grad_norm": 10.27882170416886, + "learning_rate": 1.3294134857972139e-05, + "loss": 0.7279, + "step": 3789 + }, + { + "epoch": 2.0575461454940283, + "grad_norm": 10.99914755541959, + "learning_rate": 1.3290814104617253e-05, + "loss": 0.6787, + "step": 3790 + }, + { + "epoch": 2.058089033659066, + "grad_norm": 13.627823006472678, + "learning_rate": 1.3287492944249234e-05, + "loss": 0.9052, + "step": 3791 + }, + { + "epoch": 2.0586319218241043, + "grad_norm": 7.734783593910536, + "learning_rate": 1.3284171377278849e-05, + "loss": 0.7141, + "step": 3792 + }, + { + "epoch": 2.059174809989142, + "grad_norm": 9.247896244835857, + "learning_rate": 1.3280849404116913e-05, + "loss": 0.6071, + "step": 3793 + }, + { + "epoch": 2.0597176981541803, + "grad_norm": 9.004013607687236, + "learning_rate": 1.3277527025174295e-05, + "loss": 0.5887, + "step": 3794 + }, + { + "epoch": 2.060260586319218, + "grad_norm": 9.62029412051341, + "learning_rate": 1.3274204240861908e-05, + "loss": 0.8847, + "step": 3795 + }, + { + "epoch": 2.0608034744842563, + "grad_norm": 7.500348953035017, + "learning_rate": 1.3270881051590725e-05, + "loss": 0.4769, + "step": 3796 + }, + { + "epoch": 2.061346362649294, + "grad_norm": 10.153144580789224, + "learning_rate": 1.326755745777176e-05, + "loss": 0.645, + "step": 3797 + }, + { + "epoch": 2.0618892508143323, + "grad_norm": 6.813155627292939, + "learning_rate": 1.326423345981608e-05, + "loss": 0.6711, + "step": 3798 + }, + { + "epoch": 2.06243213897937, + "grad_norm": 9.362898780009592, + "learning_rate": 1.32609090581348e-05, + "loss": 0.5464, + "step": 3799 + }, + { + "epoch": 2.0629750271444083, + "grad_norm": 8.540638734313639, + "learning_rate": 1.3257584253139096e-05, + "loss": 0.6523, + "step": 3800 + }, + { + "epoch": 2.063517915309446, + "grad_norm": 9.000753471311711, + "learning_rate": 1.3254259045240176e-05, + "loss": 0.5362, + "step": 3801 + }, + { + "epoch": 2.0640608034744843, + "grad_norm": 10.275669325762056, + "learning_rate": 1.3250933434849316e-05, + "loss": 0.6061, + "step": 3802 + }, + { + "epoch": 2.064603691639522, + "grad_norm": 11.615294601299743, + "learning_rate": 1.3247607422377823e-05, + "loss": 0.673, + "step": 3803 + }, + { + "epoch": 2.0651465798045603, + "grad_norm": 10.316289286330115, + "learning_rate": 1.324428100823707e-05, + "loss": 0.7997, + "step": 3804 + }, + { + "epoch": 2.065689467969598, + "grad_norm": 14.555740784089037, + "learning_rate": 1.3240954192838472e-05, + "loss": 0.9957, + "step": 3805 + }, + { + "epoch": 2.0662323561346363, + "grad_norm": 7.674873455902871, + "learning_rate": 1.323762697659349e-05, + "loss": 0.4771, + "step": 3806 + }, + { + "epoch": 2.066775244299674, + "grad_norm": 8.530555441214027, + "learning_rate": 1.3234299359913647e-05, + "loss": 0.8666, + "step": 3807 + }, + { + "epoch": 2.0673181324647123, + "grad_norm": 6.463731480876421, + "learning_rate": 1.3230971343210503e-05, + "loss": 0.5742, + "step": 3808 + }, + { + "epoch": 2.06786102062975, + "grad_norm": 8.074017879990592, + "learning_rate": 1.3227642926895676e-05, + "loss": 0.3637, + "step": 3809 + }, + { + "epoch": 2.0684039087947883, + "grad_norm": 9.989752857060338, + "learning_rate": 1.3224314111380828e-05, + "loss": 0.7389, + "step": 3810 + }, + { + "epoch": 2.068946796959826, + "grad_norm": 12.892065053280843, + "learning_rate": 1.3220984897077669e-05, + "loss": 1.2959, + "step": 3811 + }, + { + "epoch": 2.0694896851248643, + "grad_norm": 7.497956491794191, + "learning_rate": 1.3217655284397965e-05, + "loss": 0.4384, + "step": 3812 + }, + { + "epoch": 2.070032573289902, + "grad_norm": 14.719876212901973, + "learning_rate": 1.3214325273753528e-05, + "loss": 1.3054, + "step": 3813 + }, + { + "epoch": 2.0705754614549403, + "grad_norm": 10.767865599283661, + "learning_rate": 1.3210994865556219e-05, + "loss": 0.9961, + "step": 3814 + }, + { + "epoch": 2.071118349619978, + "grad_norm": 9.198818655810543, + "learning_rate": 1.3207664060217946e-05, + "loss": 0.5916, + "step": 3815 + }, + { + "epoch": 2.0716612377850163, + "grad_norm": 9.598428080121797, + "learning_rate": 1.320433285815067e-05, + "loss": 0.7907, + "step": 3816 + }, + { + "epoch": 2.072204125950054, + "grad_norm": 9.48389213525687, + "learning_rate": 1.32010012597664e-05, + "loss": 0.8385, + "step": 3817 + }, + { + "epoch": 2.0727470141150923, + "grad_norm": 7.562927002528845, + "learning_rate": 1.3197669265477191e-05, + "loss": 0.6748, + "step": 3818 + }, + { + "epoch": 2.07328990228013, + "grad_norm": 7.5241368226931185, + "learning_rate": 1.319433687569515e-05, + "loss": 0.5605, + "step": 3819 + }, + { + "epoch": 2.0738327904451683, + "grad_norm": 7.952800933557159, + "learning_rate": 1.3191004090832436e-05, + "loss": 0.6624, + "step": 3820 + }, + { + "epoch": 2.074375678610206, + "grad_norm": 8.02814530035842, + "learning_rate": 1.318767091130125e-05, + "loss": 0.488, + "step": 3821 + }, + { + "epoch": 2.0749185667752443, + "grad_norm": 7.000417608550192, + "learning_rate": 1.3184337337513849e-05, + "loss": 0.7194, + "step": 3822 + }, + { + "epoch": 2.075461454940282, + "grad_norm": 8.965861432524546, + "learning_rate": 1.3181003369882527e-05, + "loss": 0.4823, + "step": 3823 + }, + { + "epoch": 2.0760043431053203, + "grad_norm": 8.717885349674289, + "learning_rate": 1.3177669008819635e-05, + "loss": 0.6482, + "step": 3824 + }, + { + "epoch": 2.076547231270358, + "grad_norm": 9.217767101253152, + "learning_rate": 1.317433425473758e-05, + "loss": 0.5276, + "step": 3825 + }, + { + "epoch": 2.0770901194353963, + "grad_norm": 8.84305940258517, + "learning_rate": 1.3170999108048804e-05, + "loss": 0.6579, + "step": 3826 + }, + { + "epoch": 2.077633007600434, + "grad_norm": 7.501762934992334, + "learning_rate": 1.3167663569165803e-05, + "loss": 0.6633, + "step": 3827 + }, + { + "epoch": 2.0781758957654723, + "grad_norm": 7.815925996775423, + "learning_rate": 1.3164327638501126e-05, + "loss": 0.6881, + "step": 3828 + }, + { + "epoch": 2.07871878393051, + "grad_norm": 11.789396888417095, + "learning_rate": 1.3160991316467362e-05, + "loss": 0.5761, + "step": 3829 + }, + { + "epoch": 2.0792616720955484, + "grad_norm": 8.450719528350264, + "learning_rate": 1.3157654603477155e-05, + "loss": 0.7263, + "step": 3830 + }, + { + "epoch": 2.079804560260586, + "grad_norm": 9.544428310813693, + "learning_rate": 1.315431749994319e-05, + "loss": 0.6408, + "step": 3831 + }, + { + "epoch": 2.0803474484256244, + "grad_norm": 7.454950880630089, + "learning_rate": 1.3150980006278219e-05, + "loss": 0.4425, + "step": 3832 + }, + { + "epoch": 2.080890336590662, + "grad_norm": 11.462923993924347, + "learning_rate": 1.314764212289501e-05, + "loss": 0.5349, + "step": 3833 + }, + { + "epoch": 2.0814332247557004, + "grad_norm": 8.900231872199269, + "learning_rate": 1.314430385020641e-05, + "loss": 0.8022, + "step": 3834 + }, + { + "epoch": 2.081976112920738, + "grad_norm": 7.035528648373333, + "learning_rate": 1.3140965188625299e-05, + "loss": 0.6513, + "step": 3835 + }, + { + "epoch": 2.0825190010857764, + "grad_norm": 12.327229999200162, + "learning_rate": 1.3137626138564606e-05, + "loss": 1.0623, + "step": 3836 + }, + { + "epoch": 2.083061889250814, + "grad_norm": 8.942692654829559, + "learning_rate": 1.3134286700437308e-05, + "loss": 0.4689, + "step": 3837 + }, + { + "epoch": 2.0836047774158524, + "grad_norm": 11.092455380734007, + "learning_rate": 1.313094687465644e-05, + "loss": 0.6808, + "step": 3838 + }, + { + "epoch": 2.08414766558089, + "grad_norm": 8.918018917563646, + "learning_rate": 1.3127606661635075e-05, + "loss": 0.523, + "step": 3839 + }, + { + "epoch": 2.0846905537459284, + "grad_norm": 13.878014157853947, + "learning_rate": 1.312426606178633e-05, + "loss": 1.0231, + "step": 3840 + }, + { + "epoch": 2.085233441910966, + "grad_norm": 10.432596711564004, + "learning_rate": 1.3120925075523379e-05, + "loss": 0.8011, + "step": 3841 + }, + { + "epoch": 2.0857763300760044, + "grad_norm": 8.126214858686065, + "learning_rate": 1.3117583703259445e-05, + "loss": 0.5272, + "step": 3842 + }, + { + "epoch": 2.086319218241042, + "grad_norm": 9.596541824389169, + "learning_rate": 1.3114241945407783e-05, + "loss": 0.6585, + "step": 3843 + }, + { + "epoch": 2.0868621064060804, + "grad_norm": 10.438282441705661, + "learning_rate": 1.3110899802381718e-05, + "loss": 0.8474, + "step": 3844 + }, + { + "epoch": 2.087404994571118, + "grad_norm": 11.98868629313521, + "learning_rate": 1.3107557274594607e-05, + "loss": 1.0492, + "step": 3845 + }, + { + "epoch": 2.0879478827361564, + "grad_norm": 9.734289423556795, + "learning_rate": 1.310421436245986e-05, + "loss": 0.6573, + "step": 3846 + }, + { + "epoch": 2.088490770901194, + "grad_norm": 10.196913105458549, + "learning_rate": 1.310087106639093e-05, + "loss": 0.8964, + "step": 3847 + }, + { + "epoch": 2.0890336590662324, + "grad_norm": 9.155222462665467, + "learning_rate": 1.3097527386801327e-05, + "loss": 0.6668, + "step": 3848 + }, + { + "epoch": 2.08957654723127, + "grad_norm": 8.512351137441192, + "learning_rate": 1.3094183324104602e-05, + "loss": 0.6084, + "step": 3849 + }, + { + "epoch": 2.0901194353963084, + "grad_norm": 7.62701351090082, + "learning_rate": 1.3090838878714349e-05, + "loss": 0.4199, + "step": 3850 + }, + { + "epoch": 2.090662323561346, + "grad_norm": 8.57862015121728, + "learning_rate": 1.3087494051044218e-05, + "loss": 0.3694, + "step": 3851 + }, + { + "epoch": 2.0912052117263844, + "grad_norm": 11.407572249271688, + "learning_rate": 1.3084148841507904e-05, + "loss": 0.8861, + "step": 3852 + }, + { + "epoch": 2.091748099891422, + "grad_norm": 9.297247466948287, + "learning_rate": 1.3080803250519142e-05, + "loss": 0.8432, + "step": 3853 + }, + { + "epoch": 2.0922909880564604, + "grad_norm": 7.909660106539073, + "learning_rate": 1.3077457278491728e-05, + "loss": 0.5103, + "step": 3854 + }, + { + "epoch": 2.092833876221498, + "grad_norm": 11.013011688454522, + "learning_rate": 1.3074110925839491e-05, + "loss": 1.0117, + "step": 3855 + }, + { + "epoch": 2.0933767643865364, + "grad_norm": 10.249008356283914, + "learning_rate": 1.3070764192976315e-05, + "loss": 0.5248, + "step": 3856 + }, + { + "epoch": 2.093919652551574, + "grad_norm": 8.240438304204865, + "learning_rate": 1.306741708031613e-05, + "loss": 0.6811, + "step": 3857 + }, + { + "epoch": 2.0944625407166124, + "grad_norm": 9.542880797186081, + "learning_rate": 1.3064069588272913e-05, + "loss": 0.4746, + "step": 3858 + }, + { + "epoch": 2.09500542888165, + "grad_norm": 9.251214264026915, + "learning_rate": 1.3060721717260685e-05, + "loss": 0.6405, + "step": 3859 + }, + { + "epoch": 2.0955483170466884, + "grad_norm": 12.366199541130733, + "learning_rate": 1.3057373467693515e-05, + "loss": 0.8378, + "step": 3860 + }, + { + "epoch": 2.096091205211726, + "grad_norm": 10.566968498265949, + "learning_rate": 1.3054024839985526e-05, + "loss": 0.5975, + "step": 3861 + }, + { + "epoch": 2.0966340933767644, + "grad_norm": 9.58134406953771, + "learning_rate": 1.3050675834550872e-05, + "loss": 0.6846, + "step": 3862 + }, + { + "epoch": 2.097176981541802, + "grad_norm": 8.437576527492594, + "learning_rate": 1.3047326451803772e-05, + "loss": 0.6301, + "step": 3863 + }, + { + "epoch": 2.0977198697068404, + "grad_norm": 8.68957421381677, + "learning_rate": 1.304397669215848e-05, + "loss": 0.6121, + "step": 3864 + }, + { + "epoch": 2.098262757871878, + "grad_norm": 12.281098753982102, + "learning_rate": 1.30406265560293e-05, + "loss": 1.174, + "step": 3865 + }, + { + "epoch": 2.0988056460369164, + "grad_norm": 12.64390360725308, + "learning_rate": 1.303727604383058e-05, + "loss": 0.727, + "step": 3866 + }, + { + "epoch": 2.099348534201954, + "grad_norm": 8.982715547555616, + "learning_rate": 1.3033925155976718e-05, + "loss": 0.4946, + "step": 3867 + }, + { + "epoch": 2.0998914223669924, + "grad_norm": 8.77924533690129, + "learning_rate": 1.303057389288216e-05, + "loss": 0.5027, + "step": 3868 + }, + { + "epoch": 2.1004343105320302, + "grad_norm": 12.246059108799978, + "learning_rate": 1.302722225496139e-05, + "loss": 0.6522, + "step": 3869 + }, + { + "epoch": 2.1009771986970684, + "grad_norm": 13.95610327170378, + "learning_rate": 1.3023870242628944e-05, + "loss": 1.043, + "step": 3870 + }, + { + "epoch": 2.1015200868621062, + "grad_norm": 10.355142745873641, + "learning_rate": 1.3020517856299413e-05, + "loss": 0.9167, + "step": 3871 + }, + { + "epoch": 2.1020629750271445, + "grad_norm": 8.43332484970558, + "learning_rate": 1.3017165096387419e-05, + "loss": 0.5952, + "step": 3872 + }, + { + "epoch": 2.1026058631921822, + "grad_norm": 9.224204311836482, + "learning_rate": 1.3013811963307634e-05, + "loss": 0.5631, + "step": 3873 + }, + { + "epoch": 2.1031487513572205, + "grad_norm": 9.719137418718985, + "learning_rate": 1.3010458457474784e-05, + "loss": 0.9293, + "step": 3874 + }, + { + "epoch": 2.1036916395222582, + "grad_norm": 8.234100299546002, + "learning_rate": 1.3007104579303636e-05, + "loss": 0.7041, + "step": 3875 + }, + { + "epoch": 2.1042345276872965, + "grad_norm": 11.0180135357339, + "learning_rate": 1.3003750329208995e-05, + "loss": 0.6068, + "step": 3876 + }, + { + "epoch": 2.1047774158523342, + "grad_norm": 8.790070650912016, + "learning_rate": 1.300039570760573e-05, + "loss": 0.4956, + "step": 3877 + }, + { + "epoch": 2.1053203040173725, + "grad_norm": 8.43754180724615, + "learning_rate": 1.2997040714908742e-05, + "loss": 0.3965, + "step": 3878 + }, + { + "epoch": 2.1058631921824102, + "grad_norm": 11.815520632897917, + "learning_rate": 1.299368535153298e-05, + "loss": 0.5963, + "step": 3879 + }, + { + "epoch": 2.1064060803474485, + "grad_norm": 14.168765746771772, + "learning_rate": 1.2990329617893445e-05, + "loss": 0.8111, + "step": 3880 + }, + { + "epoch": 2.1069489685124863, + "grad_norm": 5.706418305089391, + "learning_rate": 1.2986973514405176e-05, + "loss": 0.41, + "step": 3881 + }, + { + "epoch": 2.1074918566775245, + "grad_norm": 10.870432547178469, + "learning_rate": 1.2983617041483261e-05, + "loss": 0.7805, + "step": 3882 + }, + { + "epoch": 2.1080347448425623, + "grad_norm": 9.440648088759033, + "learning_rate": 1.2980260199542838e-05, + "loss": 0.6047, + "step": 3883 + }, + { + "epoch": 2.1085776330076005, + "grad_norm": 12.145529985557056, + "learning_rate": 1.2976902988999081e-05, + "loss": 0.8366, + "step": 3884 + }, + { + "epoch": 2.1091205211726383, + "grad_norm": 9.17886258135476, + "learning_rate": 1.2973545410267218e-05, + "loss": 0.6402, + "step": 3885 + }, + { + "epoch": 2.1096634093376765, + "grad_norm": 9.613840291682566, + "learning_rate": 1.297018746376252e-05, + "loss": 0.8175, + "step": 3886 + }, + { + "epoch": 2.1102062975027143, + "grad_norm": 12.898243756363112, + "learning_rate": 1.2966829149900304e-05, + "loss": 0.8514, + "step": 3887 + }, + { + "epoch": 2.1107491856677525, + "grad_norm": 8.825284004643853, + "learning_rate": 1.2963470469095928e-05, + "loss": 0.4508, + "step": 3888 + }, + { + "epoch": 2.1112920738327903, + "grad_norm": 9.953568896257451, + "learning_rate": 1.2960111421764803e-05, + "loss": 0.5657, + "step": 3889 + }, + { + "epoch": 2.1118349619978285, + "grad_norm": 8.988870757227582, + "learning_rate": 1.2956752008322378e-05, + "loss": 0.7242, + "step": 3890 + }, + { + "epoch": 2.1123778501628663, + "grad_norm": 9.68770492429417, + "learning_rate": 1.2953392229184156e-05, + "loss": 0.9403, + "step": 3891 + }, + { + "epoch": 2.1129207383279045, + "grad_norm": 11.903522492661367, + "learning_rate": 1.2950032084765674e-05, + "loss": 0.8743, + "step": 3892 + }, + { + "epoch": 2.1134636264929423, + "grad_norm": 12.24198259147124, + "learning_rate": 1.294667157548252e-05, + "loss": 0.9861, + "step": 3893 + }, + { + "epoch": 2.1140065146579805, + "grad_norm": 13.105368254945933, + "learning_rate": 1.2943310701750331e-05, + "loss": 0.8815, + "step": 3894 + }, + { + "epoch": 2.1145494028230183, + "grad_norm": 9.322992215038694, + "learning_rate": 1.2939949463984782e-05, + "loss": 0.4986, + "step": 3895 + }, + { + "epoch": 2.1150922909880565, + "grad_norm": 7.59330739725767, + "learning_rate": 1.29365878626016e-05, + "loss": 0.4343, + "step": 3896 + }, + { + "epoch": 2.1156351791530943, + "grad_norm": 9.782670203398764, + "learning_rate": 1.293322589801655e-05, + "loss": 0.4843, + "step": 3897 + }, + { + "epoch": 2.1161780673181325, + "grad_norm": 7.130500958222279, + "learning_rate": 1.2929863570645446e-05, + "loss": 0.3784, + "step": 3898 + }, + { + "epoch": 2.1167209554831703, + "grad_norm": 7.336707110452157, + "learning_rate": 1.2926500880904147e-05, + "loss": 0.4938, + "step": 3899 + }, + { + "epoch": 2.1172638436482085, + "grad_norm": 7.047478722817777, + "learning_rate": 1.2923137829208555e-05, + "loss": 0.3002, + "step": 3900 + }, + { + "epoch": 2.1178067318132463, + "grad_norm": 9.533093539252569, + "learning_rate": 1.2919774415974616e-05, + "loss": 0.711, + "step": 3901 + }, + { + "epoch": 2.1183496199782845, + "grad_norm": 8.468755965185016, + "learning_rate": 1.2916410641618324e-05, + "loss": 0.5491, + "step": 3902 + }, + { + "epoch": 2.1188925081433223, + "grad_norm": 10.351769959990103, + "learning_rate": 1.2913046506555715e-05, + "loss": 0.6269, + "step": 3903 + }, + { + "epoch": 2.1194353963083605, + "grad_norm": 10.2212813809041, + "learning_rate": 1.2909682011202875e-05, + "loss": 0.7249, + "step": 3904 + }, + { + "epoch": 2.1199782844733983, + "grad_norm": 9.344205468199693, + "learning_rate": 1.2906317155975922e-05, + "loss": 0.6277, + "step": 3905 + }, + { + "epoch": 2.1205211726384365, + "grad_norm": 8.704990474328149, + "learning_rate": 1.2902951941291035e-05, + "loss": 0.6317, + "step": 3906 + }, + { + "epoch": 2.1210640608034743, + "grad_norm": 11.644285794605782, + "learning_rate": 1.2899586367564422e-05, + "loss": 0.8585, + "step": 3907 + }, + { + "epoch": 2.1216069489685125, + "grad_norm": 8.462034849045626, + "learning_rate": 1.2896220435212347e-05, + "loss": 0.5174, + "step": 3908 + }, + { + "epoch": 2.1221498371335503, + "grad_norm": 10.023149647380059, + "learning_rate": 1.2892854144651112e-05, + "loss": 0.6044, + "step": 3909 + }, + { + "epoch": 2.1226927252985885, + "grad_norm": 9.437366874691593, + "learning_rate": 1.2889487496297068e-05, + "loss": 0.5081, + "step": 3910 + }, + { + "epoch": 2.1232356134636263, + "grad_norm": 11.06235798573597, + "learning_rate": 1.2886120490566607e-05, + "loss": 0.7385, + "step": 3911 + }, + { + "epoch": 2.1237785016286646, + "grad_norm": 9.45153797141243, + "learning_rate": 1.288275312787616e-05, + "loss": 0.6165, + "step": 3912 + }, + { + "epoch": 2.1243213897937023, + "grad_norm": 8.907808417684745, + "learning_rate": 1.2879385408642214e-05, + "loss": 0.3956, + "step": 3913 + }, + { + "epoch": 2.1248642779587406, + "grad_norm": 12.022235830361943, + "learning_rate": 1.2876017333281288e-05, + "loss": 0.7492, + "step": 3914 + }, + { + "epoch": 2.1254071661237783, + "grad_norm": 10.18229463165741, + "learning_rate": 1.2872648902209957e-05, + "loss": 1.0188, + "step": 3915 + }, + { + "epoch": 2.1259500542888166, + "grad_norm": 8.068860605172112, + "learning_rate": 1.2869280115844831e-05, + "loss": 0.4258, + "step": 3916 + }, + { + "epoch": 2.1264929424538543, + "grad_norm": 7.300141006448243, + "learning_rate": 1.2865910974602568e-05, + "loss": 0.414, + "step": 3917 + }, + { + "epoch": 2.1270358306188926, + "grad_norm": 9.447874959421453, + "learning_rate": 1.2862541478899868e-05, + "loss": 0.6323, + "step": 3918 + }, + { + "epoch": 2.1275787187839303, + "grad_norm": 11.845175678539553, + "learning_rate": 1.2859171629153476e-05, + "loss": 0.8539, + "step": 3919 + }, + { + "epoch": 2.1281216069489686, + "grad_norm": 8.29306868561595, + "learning_rate": 1.2855801425780179e-05, + "loss": 0.7127, + "step": 3920 + }, + { + "epoch": 2.1286644951140063, + "grad_norm": 8.040765832896788, + "learning_rate": 1.285243086919681e-05, + "loss": 0.495, + "step": 3921 + }, + { + "epoch": 2.1292073832790446, + "grad_norm": 13.042470321623425, + "learning_rate": 1.2849059959820248e-05, + "loss": 0.616, + "step": 3922 + }, + { + "epoch": 2.1297502714440824, + "grad_norm": 8.585317997471494, + "learning_rate": 1.2845688698067406e-05, + "loss": 0.5599, + "step": 3923 + }, + { + "epoch": 2.1302931596091206, + "grad_norm": 9.356896896219936, + "learning_rate": 1.2842317084355251e-05, + "loss": 0.605, + "step": 3924 + }, + { + "epoch": 2.1308360477741584, + "grad_norm": 17.057557558601207, + "learning_rate": 1.283894511910079e-05, + "loss": 0.9396, + "step": 3925 + }, + { + "epoch": 2.1313789359391966, + "grad_norm": 10.659876178353805, + "learning_rate": 1.2835572802721072e-05, + "loss": 0.765, + "step": 3926 + }, + { + "epoch": 2.1319218241042344, + "grad_norm": 9.903600273604221, + "learning_rate": 1.2832200135633191e-05, + "loss": 0.4745, + "step": 3927 + }, + { + "epoch": 2.1324647122692726, + "grad_norm": 12.075983241730247, + "learning_rate": 1.2828827118254279e-05, + "loss": 0.8613, + "step": 3928 + }, + { + "epoch": 2.1330076004343104, + "grad_norm": 10.13381566988187, + "learning_rate": 1.2825453751001526e-05, + "loss": 0.7342, + "step": 3929 + }, + { + "epoch": 2.1335504885993486, + "grad_norm": 12.671225163451464, + "learning_rate": 1.282208003429215e-05, + "loss": 0.7765, + "step": 3930 + }, + { + "epoch": 2.1340933767643864, + "grad_norm": 10.503393267034513, + "learning_rate": 1.2818705968543417e-05, + "loss": 0.7042, + "step": 3931 + }, + { + "epoch": 2.1346362649294246, + "grad_norm": 7.8720699552641635, + "learning_rate": 1.2815331554172634e-05, + "loss": 0.4393, + "step": 3932 + }, + { + "epoch": 2.1351791530944624, + "grad_norm": 14.057115653429742, + "learning_rate": 1.281195679159716e-05, + "loss": 0.8676, + "step": 3933 + }, + { + "epoch": 2.1357220412595006, + "grad_norm": 12.006814229853843, + "learning_rate": 1.2808581681234387e-05, + "loss": 0.8764, + "step": 3934 + }, + { + "epoch": 2.1362649294245384, + "grad_norm": 11.511042628784308, + "learning_rate": 1.2805206223501756e-05, + "loss": 0.6034, + "step": 3935 + }, + { + "epoch": 2.1368078175895766, + "grad_norm": 6.922027915468186, + "learning_rate": 1.2801830418816749e-05, + "loss": 0.4303, + "step": 3936 + }, + { + "epoch": 2.1373507057546144, + "grad_norm": 8.583211919241508, + "learning_rate": 1.2798454267596892e-05, + "loss": 0.6059, + "step": 3937 + }, + { + "epoch": 2.1378935939196526, + "grad_norm": 10.400540758946061, + "learning_rate": 1.2795077770259749e-05, + "loss": 0.7327, + "step": 3938 + }, + { + "epoch": 2.1384364820846904, + "grad_norm": 11.253848098269156, + "learning_rate": 1.2791700927222932e-05, + "loss": 0.7247, + "step": 3939 + }, + { + "epoch": 2.1389793702497286, + "grad_norm": 10.650957738366092, + "learning_rate": 1.2788323738904098e-05, + "loss": 0.8589, + "step": 3940 + }, + { + "epoch": 2.1395222584147664, + "grad_norm": 7.284559079353519, + "learning_rate": 1.2784946205720936e-05, + "loss": 0.3361, + "step": 3941 + }, + { + "epoch": 2.1400651465798046, + "grad_norm": 10.532878125436458, + "learning_rate": 1.2781568328091192e-05, + "loss": 0.8466, + "step": 3942 + }, + { + "epoch": 2.1406080347448424, + "grad_norm": 16.5717179039462, + "learning_rate": 1.2778190106432643e-05, + "loss": 1.1453, + "step": 3943 + }, + { + "epoch": 2.1411509229098806, + "grad_norm": 8.591001506273539, + "learning_rate": 1.2774811541163114e-05, + "loss": 0.5379, + "step": 3944 + }, + { + "epoch": 2.1416938110749184, + "grad_norm": 11.173192436567485, + "learning_rate": 1.2771432632700471e-05, + "loss": 1.0715, + "step": 3945 + }, + { + "epoch": 2.1422366992399566, + "grad_norm": 11.381968938636014, + "learning_rate": 1.2768053381462625e-05, + "loss": 0.8903, + "step": 3946 + }, + { + "epoch": 2.1427795874049944, + "grad_norm": 9.847481766271917, + "learning_rate": 1.2764673787867519e-05, + "loss": 0.5874, + "step": 3947 + }, + { + "epoch": 2.1433224755700326, + "grad_norm": 9.183549346287741, + "learning_rate": 1.2761293852333156e-05, + "loss": 0.6745, + "step": 3948 + }, + { + "epoch": 2.1438653637350704, + "grad_norm": 9.610890909648008, + "learning_rate": 1.2757913575277572e-05, + "loss": 0.9781, + "step": 3949 + }, + { + "epoch": 2.1444082519001086, + "grad_norm": 9.814885217738354, + "learning_rate": 1.275453295711884e-05, + "loss": 0.9977, + "step": 3950 + }, + { + "epoch": 2.1449511400651464, + "grad_norm": 10.984041832783662, + "learning_rate": 1.275115199827508e-05, + "loss": 0.7822, + "step": 3951 + }, + { + "epoch": 2.1454940282301846, + "grad_norm": 9.634528803708669, + "learning_rate": 1.2747770699164457e-05, + "loss": 1.1719, + "step": 3952 + }, + { + "epoch": 2.1460369163952224, + "grad_norm": 9.767304548901404, + "learning_rate": 1.2744389060205173e-05, + "loss": 0.6284, + "step": 3953 + }, + { + "epoch": 2.1465798045602607, + "grad_norm": 10.373642481495464, + "learning_rate": 1.2741007081815478e-05, + "loss": 0.6742, + "step": 3954 + }, + { + "epoch": 2.1471226927252984, + "grad_norm": 10.390187789918086, + "learning_rate": 1.2737624764413659e-05, + "loss": 0.6647, + "step": 3955 + }, + { + "epoch": 2.1476655808903367, + "grad_norm": 10.617840626931814, + "learning_rate": 1.2734242108418044e-05, + "loss": 0.877, + "step": 3956 + }, + { + "epoch": 2.1482084690553744, + "grad_norm": 9.898468473630299, + "learning_rate": 1.273085911424701e-05, + "loss": 0.7335, + "step": 3957 + }, + { + "epoch": 2.1487513572204127, + "grad_norm": 10.409230729626815, + "learning_rate": 1.2727475782318966e-05, + "loss": 0.7169, + "step": 3958 + }, + { + "epoch": 2.1492942453854504, + "grad_norm": 10.425915655294723, + "learning_rate": 1.2724092113052372e-05, + "loss": 0.6578, + "step": 3959 + }, + { + "epoch": 2.1498371335504887, + "grad_norm": 6.3375473265246995, + "learning_rate": 1.2720708106865722e-05, + "loss": 0.3471, + "step": 3960 + }, + { + "epoch": 2.1503800217155264, + "grad_norm": 9.396081485808933, + "learning_rate": 1.2717323764177559e-05, + "loss": 0.5232, + "step": 3961 + }, + { + "epoch": 2.1509229098805647, + "grad_norm": 13.419504298012779, + "learning_rate": 1.2713939085406461e-05, + "loss": 1.8096, + "step": 3962 + }, + { + "epoch": 2.1514657980456025, + "grad_norm": 9.635837991132963, + "learning_rate": 1.2710554070971053e-05, + "loss": 0.5949, + "step": 3963 + }, + { + "epoch": 2.1520086862106407, + "grad_norm": 10.111061783908927, + "learning_rate": 1.2707168721289995e-05, + "loss": 0.8731, + "step": 3964 + }, + { + "epoch": 2.1525515743756785, + "grad_norm": 8.256139809775819, + "learning_rate": 1.2703783036781998e-05, + "loss": 0.726, + "step": 3965 + }, + { + "epoch": 2.1530944625407167, + "grad_norm": 8.385511442711055, + "learning_rate": 1.2700397017865802e-05, + "loss": 0.6209, + "step": 3966 + }, + { + "epoch": 2.1536373507057545, + "grad_norm": 8.556891915186654, + "learning_rate": 1.2697010664960201e-05, + "loss": 0.6746, + "step": 3967 + }, + { + "epoch": 2.1541802388707927, + "grad_norm": 8.219611855039698, + "learning_rate": 1.2693623978484022e-05, + "loss": 0.7542, + "step": 3968 + }, + { + "epoch": 2.1547231270358305, + "grad_norm": 9.425031325511556, + "learning_rate": 1.2690236958856139e-05, + "loss": 0.6735, + "step": 3969 + }, + { + "epoch": 2.1552660152008687, + "grad_norm": 9.21346190912602, + "learning_rate": 1.2686849606495461e-05, + "loss": 0.5311, + "step": 3970 + }, + { + "epoch": 2.1558089033659065, + "grad_norm": 12.261438418537729, + "learning_rate": 1.268346192182094e-05, + "loss": 0.8556, + "step": 3971 + }, + { + "epoch": 2.1563517915309447, + "grad_norm": 8.630907790930149, + "learning_rate": 1.268007390525157e-05, + "loss": 0.7789, + "step": 3972 + }, + { + "epoch": 2.1568946796959825, + "grad_norm": 10.654367663061162, + "learning_rate": 1.267668555720639e-05, + "loss": 1.1708, + "step": 3973 + }, + { + "epoch": 2.1574375678610207, + "grad_norm": 15.113883332023352, + "learning_rate": 1.2673296878104472e-05, + "loss": 1.2273, + "step": 3974 + }, + { + "epoch": 2.1579804560260585, + "grad_norm": 10.205014286438582, + "learning_rate": 1.266990786836494e-05, + "loss": 1.2711, + "step": 3975 + }, + { + "epoch": 2.1585233441910967, + "grad_norm": 8.32550465973343, + "learning_rate": 1.2666518528406944e-05, + "loss": 0.6387, + "step": 3976 + }, + { + "epoch": 2.1590662323561345, + "grad_norm": 8.80250087611585, + "learning_rate": 1.2663128858649686e-05, + "loss": 0.5669, + "step": 3977 + }, + { + "epoch": 2.1596091205211727, + "grad_norm": 7.977255679001254, + "learning_rate": 1.2659738859512411e-05, + "loss": 0.4443, + "step": 3978 + }, + { + "epoch": 2.1601520086862105, + "grad_norm": 6.086622373481402, + "learning_rate": 1.265634853141439e-05, + "loss": 0.3029, + "step": 3979 + }, + { + "epoch": 2.1606948968512487, + "grad_norm": 16.46377636281654, + "learning_rate": 1.2652957874774953e-05, + "loss": 1.09, + "step": 3980 + }, + { + "epoch": 2.1612377850162865, + "grad_norm": 10.805039141935024, + "learning_rate": 1.2649566890013456e-05, + "loss": 0.672, + "step": 3981 + }, + { + "epoch": 2.1617806731813247, + "grad_norm": 10.731407400272554, + "learning_rate": 1.2646175577549303e-05, + "loss": 0.6618, + "step": 3982 + }, + { + "epoch": 2.1623235613463625, + "grad_norm": 9.572301578766334, + "learning_rate": 1.2642783937801937e-05, + "loss": 0.6608, + "step": 3983 + }, + { + "epoch": 2.1628664495114007, + "grad_norm": 8.415534879772189, + "learning_rate": 1.2639391971190842e-05, + "loss": 0.5314, + "step": 3984 + }, + { + "epoch": 2.1634093376764385, + "grad_norm": 8.290688286613944, + "learning_rate": 1.263599967813554e-05, + "loss": 0.6736, + "step": 3985 + }, + { + "epoch": 2.1639522258414767, + "grad_norm": 8.92112434155926, + "learning_rate": 1.2632607059055597e-05, + "loss": 0.5572, + "step": 3986 + }, + { + "epoch": 2.1644951140065145, + "grad_norm": 9.82202342503724, + "learning_rate": 1.2629214114370618e-05, + "loss": 0.7838, + "step": 3987 + }, + { + "epoch": 2.1650380021715527, + "grad_norm": 7.20619661733841, + "learning_rate": 1.262582084450025e-05, + "loss": 0.5577, + "step": 3988 + }, + { + "epoch": 2.1655808903365905, + "grad_norm": 8.640482408220107, + "learning_rate": 1.2622427249864172e-05, + "loss": 0.819, + "step": 3989 + }, + { + "epoch": 2.1661237785016287, + "grad_norm": 9.245400712044432, + "learning_rate": 1.2619033330882114e-05, + "loss": 0.6009, + "step": 3990 + }, + { + "epoch": 2.1666666666666665, + "grad_norm": 8.891547209594252, + "learning_rate": 1.2615639087973835e-05, + "loss": 0.6085, + "step": 3991 + }, + { + "epoch": 2.1672095548317047, + "grad_norm": 12.196798951360252, + "learning_rate": 1.2612244521559148e-05, + "loss": 0.745, + "step": 3992 + }, + { + "epoch": 2.1677524429967425, + "grad_norm": 9.610928616001752, + "learning_rate": 1.2608849632057895e-05, + "loss": 0.7709, + "step": 3993 + }, + { + "epoch": 2.1682953311617807, + "grad_norm": 9.454740121591225, + "learning_rate": 1.2605454419889962e-05, + "loss": 0.6455, + "step": 3994 + }, + { + "epoch": 2.1688382193268185, + "grad_norm": 13.245987794421069, + "learning_rate": 1.2602058885475273e-05, + "loss": 0.7104, + "step": 3995 + }, + { + "epoch": 2.1693811074918568, + "grad_norm": 10.51068899063252, + "learning_rate": 1.2598663029233794e-05, + "loss": 0.5747, + "step": 3996 + }, + { + "epoch": 2.1699239956568945, + "grad_norm": 10.589584490390159, + "learning_rate": 1.2595266851585532e-05, + "loss": 0.7878, + "step": 3997 + }, + { + "epoch": 2.1704668838219328, + "grad_norm": 7.818930212111502, + "learning_rate": 1.2591870352950528e-05, + "loss": 0.5179, + "step": 3998 + }, + { + "epoch": 2.1710097719869705, + "grad_norm": 8.222376749362372, + "learning_rate": 1.2588473533748868e-05, + "loss": 0.4722, + "step": 3999 + }, + { + "epoch": 2.1715526601520088, + "grad_norm": 12.451645756942625, + "learning_rate": 1.2585076394400675e-05, + "loss": 0.7742, + "step": 4000 + }, + { + "epoch": 2.1720955483170465, + "grad_norm": 11.33368393512842, + "learning_rate": 1.2581678935326117e-05, + "loss": 0.3945, + "step": 4001 + }, + { + "epoch": 2.1726384364820848, + "grad_norm": 10.914311466768305, + "learning_rate": 1.2578281156945389e-05, + "loss": 0.6345, + "step": 4002 + }, + { + "epoch": 2.1731813246471225, + "grad_norm": 12.867661375339019, + "learning_rate": 1.2574883059678742e-05, + "loss": 1.1569, + "step": 4003 + }, + { + "epoch": 2.1737242128121608, + "grad_norm": 11.4420969393128, + "learning_rate": 1.2571484643946452e-05, + "loss": 0.4927, + "step": 4004 + }, + { + "epoch": 2.1742671009771986, + "grad_norm": 10.234468502706436, + "learning_rate": 1.2568085910168842e-05, + "loss": 0.463, + "step": 4005 + }, + { + "epoch": 2.1748099891422368, + "grad_norm": 7.649699643480241, + "learning_rate": 1.2564686858766275e-05, + "loss": 0.3981, + "step": 4006 + }, + { + "epoch": 2.1753528773072746, + "grad_norm": 9.503268181536553, + "learning_rate": 1.2561287490159151e-05, + "loss": 0.4162, + "step": 4007 + }, + { + "epoch": 2.175895765472313, + "grad_norm": 8.81726258724681, + "learning_rate": 1.2557887804767907e-05, + "loss": 0.92, + "step": 4008 + }, + { + "epoch": 2.1764386536373506, + "grad_norm": 9.625823227183314, + "learning_rate": 1.2554487803013027e-05, + "loss": 0.7527, + "step": 4009 + }, + { + "epoch": 2.176981541802389, + "grad_norm": 11.452435192265499, + "learning_rate": 1.2551087485315017e-05, + "loss": 0.8611, + "step": 4010 + }, + { + "epoch": 2.1775244299674266, + "grad_norm": 9.123700040373722, + "learning_rate": 1.2547686852094445e-05, + "loss": 0.475, + "step": 4011 + }, + { + "epoch": 2.178067318132465, + "grad_norm": 9.306092992521913, + "learning_rate": 1.2544285903771902e-05, + "loss": 0.6063, + "step": 4012 + }, + { + "epoch": 2.1786102062975026, + "grad_norm": 14.283804972037375, + "learning_rate": 1.2540884640768022e-05, + "loss": 0.9142, + "step": 4013 + }, + { + "epoch": 2.179153094462541, + "grad_norm": 14.299740055929474, + "learning_rate": 1.2537483063503483e-05, + "loss": 0.8124, + "step": 4014 + }, + { + "epoch": 2.1796959826275786, + "grad_norm": 9.637221226214642, + "learning_rate": 1.2534081172398993e-05, + "loss": 0.6278, + "step": 4015 + }, + { + "epoch": 2.180238870792617, + "grad_norm": 6.674062415604907, + "learning_rate": 1.2530678967875304e-05, + "loss": 0.5078, + "step": 4016 + }, + { + "epoch": 2.1807817589576546, + "grad_norm": 11.68250334366375, + "learning_rate": 1.252727645035321e-05, + "loss": 0.7393, + "step": 4017 + }, + { + "epoch": 2.181324647122693, + "grad_norm": 9.642668530719417, + "learning_rate": 1.2523873620253535e-05, + "loss": 0.6218, + "step": 4018 + }, + { + "epoch": 2.1818675352877306, + "grad_norm": 9.955103488106618, + "learning_rate": 1.2520470477997146e-05, + "loss": 0.899, + "step": 4019 + }, + { + "epoch": 2.182410423452769, + "grad_norm": 13.184397199493825, + "learning_rate": 1.2517067024004955e-05, + "loss": 0.8437, + "step": 4020 + }, + { + "epoch": 2.1829533116178066, + "grad_norm": 11.142111068408965, + "learning_rate": 1.2513663258697901e-05, + "loss": 0.6064, + "step": 4021 + }, + { + "epoch": 2.183496199782845, + "grad_norm": 9.336503366824324, + "learning_rate": 1.251025918249697e-05, + "loss": 0.6649, + "step": 4022 + }, + { + "epoch": 2.1840390879478826, + "grad_norm": 11.35476662379509, + "learning_rate": 1.250685479582318e-05, + "loss": 0.8573, + "step": 4023 + }, + { + "epoch": 2.184581976112921, + "grad_norm": 9.049806197407708, + "learning_rate": 1.2503450099097594e-05, + "loss": 0.3839, + "step": 4024 + }, + { + "epoch": 2.1851248642779586, + "grad_norm": 12.739046308083323, + "learning_rate": 1.250004509274131e-05, + "loss": 0.9344, + "step": 4025 + }, + { + "epoch": 2.185667752442997, + "grad_norm": 7.539200963039068, + "learning_rate": 1.2496639777175465e-05, + "loss": 0.3691, + "step": 4026 + }, + { + "epoch": 2.1862106406080346, + "grad_norm": 10.659038181144831, + "learning_rate": 1.2493234152821234e-05, + "loss": 0.9462, + "step": 4027 + }, + { + "epoch": 2.186753528773073, + "grad_norm": 8.047186696422612, + "learning_rate": 1.2489828220099831e-05, + "loss": 0.4974, + "step": 4028 + }, + { + "epoch": 2.1872964169381106, + "grad_norm": 11.048564474051437, + "learning_rate": 1.2486421979432503e-05, + "loss": 0.6221, + "step": 4029 + }, + { + "epoch": 2.187839305103149, + "grad_norm": 9.808559304060044, + "learning_rate": 1.2483015431240542e-05, + "loss": 0.5221, + "step": 4030 + }, + { + "epoch": 2.1883821932681866, + "grad_norm": 7.720481766274275, + "learning_rate": 1.2479608575945274e-05, + "loss": 0.5248, + "step": 4031 + }, + { + "epoch": 2.188925081433225, + "grad_norm": 7.791803772948659, + "learning_rate": 1.2476201413968068e-05, + "loss": 0.5313, + "step": 4032 + }, + { + "epoch": 2.1894679695982626, + "grad_norm": 8.280464830936529, + "learning_rate": 1.2472793945730323e-05, + "loss": 0.5155, + "step": 4033 + }, + { + "epoch": 2.190010857763301, + "grad_norm": 6.500045291722395, + "learning_rate": 1.2469386171653483e-05, + "loss": 0.6057, + "step": 4034 + }, + { + "epoch": 2.1905537459283386, + "grad_norm": 6.735518823264981, + "learning_rate": 1.2465978092159025e-05, + "loss": 0.5207, + "step": 4035 + }, + { + "epoch": 2.191096634093377, + "grad_norm": 12.977966385021237, + "learning_rate": 1.2462569707668468e-05, + "loss": 1.0472, + "step": 4036 + }, + { + "epoch": 2.1916395222584146, + "grad_norm": 12.805295393279883, + "learning_rate": 1.245916101860336e-05, + "loss": 0.7486, + "step": 4037 + }, + { + "epoch": 2.192182410423453, + "grad_norm": 11.04531935760714, + "learning_rate": 1.2455752025385304e-05, + "loss": 0.6107, + "step": 4038 + }, + { + "epoch": 2.1927252985884906, + "grad_norm": 12.12096270238078, + "learning_rate": 1.245234272843592e-05, + "loss": 0.6264, + "step": 4039 + }, + { + "epoch": 2.193268186753529, + "grad_norm": 14.202650747934204, + "learning_rate": 1.244893312817688e-05, + "loss": 0.7965, + "step": 4040 + }, + { + "epoch": 2.1938110749185666, + "grad_norm": 10.409453866661186, + "learning_rate": 1.2445523225029887e-05, + "loss": 0.6012, + "step": 4041 + }, + { + "epoch": 2.194353963083605, + "grad_norm": 8.68329891299571, + "learning_rate": 1.2442113019416683e-05, + "loss": 0.4393, + "step": 4042 + }, + { + "epoch": 2.1948968512486426, + "grad_norm": 11.77894618341781, + "learning_rate": 1.2438702511759049e-05, + "loss": 0.7838, + "step": 4043 + }, + { + "epoch": 2.195439739413681, + "grad_norm": 8.31374036003515, + "learning_rate": 1.2435291702478802e-05, + "loss": 0.6043, + "step": 4044 + }, + { + "epoch": 2.1959826275787186, + "grad_norm": 10.219603175661236, + "learning_rate": 1.2431880591997799e-05, + "loss": 0.9292, + "step": 4045 + }, + { + "epoch": 2.196525515743757, + "grad_norm": 10.662478372576002, + "learning_rate": 1.2428469180737923e-05, + "loss": 0.5951, + "step": 4046 + }, + { + "epoch": 2.1970684039087947, + "grad_norm": 15.414471746340716, + "learning_rate": 1.2425057469121113e-05, + "loss": 1.0091, + "step": 4047 + }, + { + "epoch": 2.197611292073833, + "grad_norm": 8.553095532374297, + "learning_rate": 1.242164545756933e-05, + "loss": 0.4494, + "step": 4048 + }, + { + "epoch": 2.1981541802388707, + "grad_norm": 10.520590758759942, + "learning_rate": 1.2418233146504575e-05, + "loss": 0.5277, + "step": 4049 + }, + { + "epoch": 2.198697068403909, + "grad_norm": 10.792986797196212, + "learning_rate": 1.241482053634889e-05, + "loss": 0.7036, + "step": 4050 + }, + { + "epoch": 2.1992399565689467, + "grad_norm": 10.425002068049416, + "learning_rate": 1.2411407627524354e-05, + "loss": 0.8283, + "step": 4051 + }, + { + "epoch": 2.199782844733985, + "grad_norm": 10.421667203553737, + "learning_rate": 1.240799442045308e-05, + "loss": 0.8689, + "step": 4052 + }, + { + "epoch": 2.2003257328990227, + "grad_norm": 9.866413466645723, + "learning_rate": 1.2404580915557217e-05, + "loss": 0.517, + "step": 4053 + }, + { + "epoch": 2.200868621064061, + "grad_norm": 9.806755744401848, + "learning_rate": 1.2401167113258954e-05, + "loss": 0.7176, + "step": 4054 + }, + { + "epoch": 2.2014115092290987, + "grad_norm": 12.103200009209074, + "learning_rate": 1.2397753013980516e-05, + "loss": 0.8563, + "step": 4055 + }, + { + "epoch": 2.201954397394137, + "grad_norm": 9.310089002310463, + "learning_rate": 1.2394338618144162e-05, + "loss": 0.5478, + "step": 4056 + }, + { + "epoch": 2.2024972855591747, + "grad_norm": 8.147499111803365, + "learning_rate": 1.2390923926172194e-05, + "loss": 0.4616, + "step": 4057 + }, + { + "epoch": 2.203040173724213, + "grad_norm": 10.363627124726912, + "learning_rate": 1.2387508938486945e-05, + "loss": 0.6039, + "step": 4058 + }, + { + "epoch": 2.2035830618892507, + "grad_norm": 11.09580557172493, + "learning_rate": 1.2384093655510785e-05, + "loss": 0.6772, + "step": 4059 + }, + { + "epoch": 2.204125950054289, + "grad_norm": 10.762880100606296, + "learning_rate": 1.238067807766612e-05, + "loss": 0.8104, + "step": 4060 + }, + { + "epoch": 2.2046688382193267, + "grad_norm": 10.590539714326667, + "learning_rate": 1.2377262205375398e-05, + "loss": 0.6766, + "step": 4061 + }, + { + "epoch": 2.205211726384365, + "grad_norm": 10.060689580140192, + "learning_rate": 1.2373846039061095e-05, + "loss": 1.0282, + "step": 4062 + }, + { + "epoch": 2.2057546145494027, + "grad_norm": 7.612672548635284, + "learning_rate": 1.237042957914573e-05, + "loss": 0.449, + "step": 4063 + }, + { + "epoch": 2.206297502714441, + "grad_norm": 12.239451929254052, + "learning_rate": 1.2367012826051861e-05, + "loss": 0.5385, + "step": 4064 + }, + { + "epoch": 2.2068403908794787, + "grad_norm": 14.199356139948698, + "learning_rate": 1.236359578020207e-05, + "loss": 0.7867, + "step": 4065 + }, + { + "epoch": 2.207383279044517, + "grad_norm": 8.357180617953833, + "learning_rate": 1.2360178442018989e-05, + "loss": 0.6803, + "step": 4066 + }, + { + "epoch": 2.2079261672095547, + "grad_norm": 9.203584379654275, + "learning_rate": 1.2356760811925277e-05, + "loss": 0.677, + "step": 4067 + }, + { + "epoch": 2.208469055374593, + "grad_norm": 8.063224350440471, + "learning_rate": 1.2353342890343626e-05, + "loss": 0.644, + "step": 4068 + }, + { + "epoch": 2.2090119435396307, + "grad_norm": 9.352085551954278, + "learning_rate": 1.2349924677696781e-05, + "loss": 0.9888, + "step": 4069 + }, + { + "epoch": 2.209554831704669, + "grad_norm": 10.54696840423902, + "learning_rate": 1.2346506174407505e-05, + "loss": 0.5895, + "step": 4070 + }, + { + "epoch": 2.2100977198697067, + "grad_norm": 8.568183164428444, + "learning_rate": 1.2343087380898604e-05, + "loss": 0.6287, + "step": 4071 + }, + { + "epoch": 2.210640608034745, + "grad_norm": 11.191021916852304, + "learning_rate": 1.2339668297592924e-05, + "loss": 0.8216, + "step": 4072 + }, + { + "epoch": 2.2111834961997827, + "grad_norm": 7.67795379285754, + "learning_rate": 1.233624892491334e-05, + "loss": 0.423, + "step": 4073 + }, + { + "epoch": 2.211726384364821, + "grad_norm": 8.108373133947762, + "learning_rate": 1.2332829263282764e-05, + "loss": 0.6548, + "step": 4074 + }, + { + "epoch": 2.2122692725298587, + "grad_norm": 9.0962841262343, + "learning_rate": 1.2329409313124146e-05, + "loss": 0.5506, + "step": 4075 + }, + { + "epoch": 2.212812160694897, + "grad_norm": 8.85949097192624, + "learning_rate": 1.2325989074860472e-05, + "loss": 0.5258, + "step": 4076 + }, + { + "epoch": 2.2133550488599347, + "grad_norm": 7.051914022367066, + "learning_rate": 1.2322568548914764e-05, + "loss": 0.5242, + "step": 4077 + }, + { + "epoch": 2.213897937024973, + "grad_norm": 8.287257693111446, + "learning_rate": 1.2319147735710076e-05, + "loss": 0.4769, + "step": 4078 + }, + { + "epoch": 2.2144408251900107, + "grad_norm": 9.473240892632637, + "learning_rate": 1.2315726635669498e-05, + "loss": 1.0853, + "step": 4079 + }, + { + "epoch": 2.214983713355049, + "grad_norm": 10.460874830183508, + "learning_rate": 1.231230524921616e-05, + "loss": 0.8907, + "step": 4080 + }, + { + "epoch": 2.2155266015200867, + "grad_norm": 9.17618869023913, + "learning_rate": 1.2308883576773223e-05, + "loss": 0.6355, + "step": 4081 + }, + { + "epoch": 2.216069489685125, + "grad_norm": 10.389188046047401, + "learning_rate": 1.2305461618763886e-05, + "loss": 0.9455, + "step": 4082 + }, + { + "epoch": 2.2166123778501627, + "grad_norm": 8.41220072964392, + "learning_rate": 1.230203937561138e-05, + "loss": 0.5574, + "step": 4083 + }, + { + "epoch": 2.217155266015201, + "grad_norm": 8.193062738430084, + "learning_rate": 1.2298616847738978e-05, + "loss": 0.548, + "step": 4084 + }, + { + "epoch": 2.2176981541802387, + "grad_norm": 8.55082220874788, + "learning_rate": 1.2295194035569979e-05, + "loss": 0.6087, + "step": 4085 + }, + { + "epoch": 2.218241042345277, + "grad_norm": 10.482752217531827, + "learning_rate": 1.2291770939527725e-05, + "loss": 0.5404, + "step": 4086 + }, + { + "epoch": 2.2187839305103148, + "grad_norm": 9.786686689665721, + "learning_rate": 1.2288347560035587e-05, + "loss": 0.6795, + "step": 4087 + }, + { + "epoch": 2.219326818675353, + "grad_norm": 9.696178175990593, + "learning_rate": 1.2284923897516978e-05, + "loss": 1.2474, + "step": 4088 + }, + { + "epoch": 2.2198697068403908, + "grad_norm": 9.135935376096977, + "learning_rate": 1.2281499952395336e-05, + "loss": 0.6746, + "step": 4089 + }, + { + "epoch": 2.220412595005429, + "grad_norm": 10.341327194770198, + "learning_rate": 1.2278075725094147e-05, + "loss": 0.6501, + "step": 4090 + }, + { + "epoch": 2.2209554831704668, + "grad_norm": 11.750926749777816, + "learning_rate": 1.2274651216036921e-05, + "loss": 0.6421, + "step": 4091 + }, + { + "epoch": 2.221498371335505, + "grad_norm": 15.264249369051182, + "learning_rate": 1.2271226425647207e-05, + "loss": 0.9053, + "step": 4092 + }, + { + "epoch": 2.2220412595005428, + "grad_norm": 10.996946312108877, + "learning_rate": 1.2267801354348589e-05, + "loss": 0.6776, + "step": 4093 + }, + { + "epoch": 2.222584147665581, + "grad_norm": 9.01852590664764, + "learning_rate": 1.2264376002564687e-05, + "loss": 0.7342, + "step": 4094 + }, + { + "epoch": 2.2231270358306188, + "grad_norm": 9.776717004980412, + "learning_rate": 1.2260950370719146e-05, + "loss": 0.7353, + "step": 4095 + }, + { + "epoch": 2.223669923995657, + "grad_norm": 12.928031483014212, + "learning_rate": 1.2257524459235666e-05, + "loss": 0.8649, + "step": 4096 + }, + { + "epoch": 2.2242128121606948, + "grad_norm": 9.549545586005916, + "learning_rate": 1.2254098268537962e-05, + "loss": 0.8056, + "step": 4097 + }, + { + "epoch": 2.224755700325733, + "grad_norm": 10.729492841694832, + "learning_rate": 1.2250671799049791e-05, + "loss": 0.5817, + "step": 4098 + }, + { + "epoch": 2.225298588490771, + "grad_norm": 14.13032585380651, + "learning_rate": 1.2247245051194944e-05, + "loss": 0.5998, + "step": 4099 + }, + { + "epoch": 2.225841476655809, + "grad_norm": 9.143101156244217, + "learning_rate": 1.2243818025397247e-05, + "loss": 0.5633, + "step": 4100 + }, + { + "epoch": 2.226384364820847, + "grad_norm": 9.010721857216959, + "learning_rate": 1.224039072208056e-05, + "loss": 0.5855, + "step": 4101 + }, + { + "epoch": 2.226927252985885, + "grad_norm": 11.257457693848512, + "learning_rate": 1.2236963141668778e-05, + "loss": 0.5672, + "step": 4102 + }, + { + "epoch": 2.227470141150923, + "grad_norm": 9.238231248800124, + "learning_rate": 1.2233535284585831e-05, + "loss": 0.4633, + "step": 4103 + }, + { + "epoch": 2.228013029315961, + "grad_norm": 8.93957445803232, + "learning_rate": 1.223010715125568e-05, + "loss": 0.8019, + "step": 4104 + }, + { + "epoch": 2.228555917480999, + "grad_norm": 10.869791394117428, + "learning_rate": 1.2226678742102322e-05, + "loss": 0.4235, + "step": 4105 + }, + { + "epoch": 2.229098805646037, + "grad_norm": 11.71960282550904, + "learning_rate": 1.2223250057549789e-05, + "loss": 0.8825, + "step": 4106 + }, + { + "epoch": 2.229641693811075, + "grad_norm": 9.133591574750724, + "learning_rate": 1.2219821098022146e-05, + "loss": 0.456, + "step": 4107 + }, + { + "epoch": 2.230184581976113, + "grad_norm": 7.713253727965175, + "learning_rate": 1.2216391863943492e-05, + "loss": 0.407, + "step": 4108 + }, + { + "epoch": 2.230727470141151, + "grad_norm": 8.351720726424757, + "learning_rate": 1.2212962355737959e-05, + "loss": 0.6003, + "step": 4109 + }, + { + "epoch": 2.231270358306189, + "grad_norm": 12.012727761361187, + "learning_rate": 1.2209532573829714e-05, + "loss": 0.731, + "step": 4110 + }, + { + "epoch": 2.231813246471227, + "grad_norm": 12.902339895008394, + "learning_rate": 1.220610251864296e-05, + "loss": 0.9873, + "step": 4111 + }, + { + "epoch": 2.232356134636265, + "grad_norm": 8.935941303438996, + "learning_rate": 1.220267219060193e-05, + "loss": 0.7549, + "step": 4112 + }, + { + "epoch": 2.232899022801303, + "grad_norm": 9.69521468611751, + "learning_rate": 1.2199241590130892e-05, + "loss": 0.6099, + "step": 4113 + }, + { + "epoch": 2.233441910966341, + "grad_norm": 7.437030925496973, + "learning_rate": 1.2195810717654148e-05, + "loss": 0.4743, + "step": 4114 + }, + { + "epoch": 2.233984799131379, + "grad_norm": 7.435823614079823, + "learning_rate": 1.2192379573596036e-05, + "loss": 0.5019, + "step": 4115 + }, + { + "epoch": 2.234527687296417, + "grad_norm": 7.563032718583443, + "learning_rate": 1.2188948158380927e-05, + "loss": 0.7176, + "step": 4116 + }, + { + "epoch": 2.235070575461455, + "grad_norm": 10.449779515276235, + "learning_rate": 1.2185516472433218e-05, + "loss": 0.7921, + "step": 4117 + }, + { + "epoch": 2.235613463626493, + "grad_norm": 9.088046167756731, + "learning_rate": 1.2182084516177351e-05, + "loss": 0.811, + "step": 4118 + }, + { + "epoch": 2.236156351791531, + "grad_norm": 9.664269830378425, + "learning_rate": 1.217865229003779e-05, + "loss": 0.8024, + "step": 4119 + }, + { + "epoch": 2.236699239956569, + "grad_norm": 8.055354430405984, + "learning_rate": 1.2175219794439044e-05, + "loss": 0.5747, + "step": 4120 + }, + { + "epoch": 2.237242128121607, + "grad_norm": 6.428342674450092, + "learning_rate": 1.2171787029805646e-05, + "loss": 0.5673, + "step": 4121 + }, + { + "epoch": 2.237785016286645, + "grad_norm": 7.502074776894542, + "learning_rate": 1.2168353996562167e-05, + "loss": 0.566, + "step": 4122 + }, + { + "epoch": 2.238327904451683, + "grad_norm": 10.380110541434405, + "learning_rate": 1.2164920695133212e-05, + "loss": 1.0609, + "step": 4123 + }, + { + "epoch": 2.238870792616721, + "grad_norm": 13.199317832444493, + "learning_rate": 1.2161487125943413e-05, + "loss": 0.7662, + "step": 4124 + }, + { + "epoch": 2.239413680781759, + "grad_norm": 9.067564902950002, + "learning_rate": 1.2158053289417439e-05, + "loss": 0.6325, + "step": 4125 + }, + { + "epoch": 2.239956568946797, + "grad_norm": 8.074724230493178, + "learning_rate": 1.215461918598e-05, + "loss": 0.6854, + "step": 4126 + }, + { + "epoch": 2.240499457111835, + "grad_norm": 7.916509466850638, + "learning_rate": 1.2151184816055822e-05, + "loss": 0.3852, + "step": 4127 + }, + { + "epoch": 2.241042345276873, + "grad_norm": 8.647089691838135, + "learning_rate": 1.214775018006968e-05, + "loss": 0.6291, + "step": 4128 + }, + { + "epoch": 2.241585233441911, + "grad_norm": 9.238872667328287, + "learning_rate": 1.214431527844637e-05, + "loss": 0.6117, + "step": 4129 + }, + { + "epoch": 2.242128121606949, + "grad_norm": 12.537117099920934, + "learning_rate": 1.214088011161073e-05, + "loss": 0.745, + "step": 4130 + }, + { + "epoch": 2.242671009771987, + "grad_norm": 9.99529796933279, + "learning_rate": 1.2137444679987627e-05, + "loss": 0.612, + "step": 4131 + }, + { + "epoch": 2.243213897937025, + "grad_norm": 6.673816908475189, + "learning_rate": 1.2134008984001959e-05, + "loss": 0.3219, + "step": 4132 + }, + { + "epoch": 2.243756786102063, + "grad_norm": 13.936566760747976, + "learning_rate": 1.2130573024078656e-05, + "loss": 0.8689, + "step": 4133 + }, + { + "epoch": 2.244299674267101, + "grad_norm": 7.800578548383095, + "learning_rate": 1.2127136800642689e-05, + "loss": 0.3804, + "step": 4134 + }, + { + "epoch": 2.244842562432139, + "grad_norm": 9.283785411998215, + "learning_rate": 1.2123700314119052e-05, + "loss": 0.699, + "step": 4135 + }, + { + "epoch": 2.245385450597177, + "grad_norm": 8.613596910195628, + "learning_rate": 1.2120263564932775e-05, + "loss": 0.7155, + "step": 4136 + }, + { + "epoch": 2.245928338762215, + "grad_norm": 8.522662049603209, + "learning_rate": 1.2116826553508923e-05, + "loss": 0.5705, + "step": 4137 + }, + { + "epoch": 2.246471226927253, + "grad_norm": 9.023598184433123, + "learning_rate": 1.2113389280272587e-05, + "loss": 0.4392, + "step": 4138 + }, + { + "epoch": 2.247014115092291, + "grad_norm": 10.959276458237868, + "learning_rate": 1.2109951745648894e-05, + "loss": 0.7827, + "step": 4139 + }, + { + "epoch": 2.247557003257329, + "grad_norm": 11.660193747264707, + "learning_rate": 1.210651395006301e-05, + "loss": 0.9824, + "step": 4140 + }, + { + "epoch": 2.248099891422367, + "grad_norm": 18.121219443467986, + "learning_rate": 1.2103075893940122e-05, + "loss": 1.2339, + "step": 4141 + }, + { + "epoch": 2.248642779587405, + "grad_norm": 15.059306090767514, + "learning_rate": 1.2099637577705457e-05, + "loss": 0.7434, + "step": 4142 + }, + { + "epoch": 2.249185667752443, + "grad_norm": 10.999940106944315, + "learning_rate": 1.2096199001784268e-05, + "loss": 0.6422, + "step": 4143 + }, + { + "epoch": 2.249728555917481, + "grad_norm": 6.704480729179984, + "learning_rate": 1.2092760166601848e-05, + "loss": 0.4829, + "step": 4144 + }, + { + "epoch": 2.250271444082519, + "grad_norm": 11.849496386091012, + "learning_rate": 1.2089321072583512e-05, + "loss": 0.8969, + "step": 4145 + }, + { + "epoch": 2.250814332247557, + "grad_norm": 8.596596068581539, + "learning_rate": 1.2085881720154617e-05, + "loss": 0.6148, + "step": 4146 + }, + { + "epoch": 2.251357220412595, + "grad_norm": 7.50181808267645, + "learning_rate": 1.2082442109740548e-05, + "loss": 0.5499, + "step": 4147 + }, + { + "epoch": 2.251900108577633, + "grad_norm": 10.178327453067736, + "learning_rate": 1.207900224176672e-05, + "loss": 0.5582, + "step": 4148 + }, + { + "epoch": 2.252442996742671, + "grad_norm": 10.74128468394649, + "learning_rate": 1.2075562116658581e-05, + "loss": 0.8023, + "step": 4149 + }, + { + "epoch": 2.252985884907709, + "grad_norm": 10.195557872174708, + "learning_rate": 1.207212173484161e-05, + "loss": 0.6092, + "step": 4150 + }, + { + "epoch": 2.253528773072747, + "grad_norm": 9.843275181764184, + "learning_rate": 1.2068681096741322e-05, + "loss": 0.8451, + "step": 4151 + }, + { + "epoch": 2.254071661237785, + "grad_norm": 7.859122806337064, + "learning_rate": 1.2065240202783259e-05, + "loss": 0.6158, + "step": 4152 + }, + { + "epoch": 2.254614549402823, + "grad_norm": 11.4233860908698, + "learning_rate": 1.2061799053392994e-05, + "loss": 0.6552, + "step": 4153 + }, + { + "epoch": 2.255157437567861, + "grad_norm": 7.339673441521232, + "learning_rate": 1.2058357648996138e-05, + "loss": 0.5103, + "step": 4154 + }, + { + "epoch": 2.255700325732899, + "grad_norm": 11.603888679814519, + "learning_rate": 1.2054915990018326e-05, + "loss": 0.8117, + "step": 4155 + }, + { + "epoch": 2.256243213897937, + "grad_norm": 9.757430713637703, + "learning_rate": 1.2051474076885232e-05, + "loss": 0.6942, + "step": 4156 + }, + { + "epoch": 2.256786102062975, + "grad_norm": 10.286384353132442, + "learning_rate": 1.2048031910022553e-05, + "loss": 0.8958, + "step": 4157 + }, + { + "epoch": 2.257328990228013, + "grad_norm": 7.6945320206947185, + "learning_rate": 1.2044589489856023e-05, + "loss": 0.4508, + "step": 4158 + }, + { + "epoch": 2.257871878393051, + "grad_norm": 9.546205961324182, + "learning_rate": 1.2041146816811406e-05, + "loss": 0.6889, + "step": 4159 + }, + { + "epoch": 2.258414766558089, + "grad_norm": 6.440627468414342, + "learning_rate": 1.2037703891314498e-05, + "loss": 0.5146, + "step": 4160 + }, + { + "epoch": 2.258957654723127, + "grad_norm": 8.769926030437105, + "learning_rate": 1.2034260713791124e-05, + "loss": 0.539, + "step": 4161 + }, + { + "epoch": 2.259500542888165, + "grad_norm": 8.385311572142552, + "learning_rate": 1.2030817284667145e-05, + "loss": 0.6471, + "step": 4162 + }, + { + "epoch": 2.260043431053203, + "grad_norm": 10.687000132395355, + "learning_rate": 1.2027373604368446e-05, + "loss": 0.4047, + "step": 4163 + }, + { + "epoch": 2.260586319218241, + "grad_norm": 9.22667613364525, + "learning_rate": 1.2023929673320952e-05, + "loss": 0.6039, + "step": 4164 + }, + { + "epoch": 2.261129207383279, + "grad_norm": 8.822764152626888, + "learning_rate": 1.2020485491950607e-05, + "loss": 0.5333, + "step": 4165 + }, + { + "epoch": 2.261672095548317, + "grad_norm": 8.505680622817092, + "learning_rate": 1.2017041060683394e-05, + "loss": 0.6679, + "step": 4166 + }, + { + "epoch": 2.262214983713355, + "grad_norm": 12.980914094830798, + "learning_rate": 1.2013596379945331e-05, + "loss": 0.7265, + "step": 4167 + }, + { + "epoch": 2.262757871878393, + "grad_norm": 12.278007785084847, + "learning_rate": 1.201015145016246e-05, + "loss": 0.7175, + "step": 4168 + }, + { + "epoch": 2.263300760043431, + "grad_norm": 8.871985541095814, + "learning_rate": 1.2006706271760851e-05, + "loss": 0.4937, + "step": 4169 + }, + { + "epoch": 2.263843648208469, + "grad_norm": 10.070308144714824, + "learning_rate": 1.2003260845166613e-05, + "loss": 0.5994, + "step": 4170 + }, + { + "epoch": 2.264386536373507, + "grad_norm": 8.846376302708403, + "learning_rate": 1.1999815170805882e-05, + "loss": 0.5795, + "step": 4171 + }, + { + "epoch": 2.264929424538545, + "grad_norm": 11.984661783852376, + "learning_rate": 1.199636924910482e-05, + "loss": 1.1718, + "step": 4172 + }, + { + "epoch": 2.265472312703583, + "grad_norm": 10.035244756765978, + "learning_rate": 1.1992923080489629e-05, + "loss": 0.8101, + "step": 4173 + }, + { + "epoch": 2.266015200868621, + "grad_norm": 7.429999917349929, + "learning_rate": 1.1989476665386537e-05, + "loss": 0.4121, + "step": 4174 + }, + { + "epoch": 2.266558089033659, + "grad_norm": 8.464944728590373, + "learning_rate": 1.1986030004221802e-05, + "loss": 0.5553, + "step": 4175 + }, + { + "epoch": 2.267100977198697, + "grad_norm": 11.414355640267063, + "learning_rate": 1.1982583097421707e-05, + "loss": 0.8705, + "step": 4176 + }, + { + "epoch": 2.267643865363735, + "grad_norm": 10.094383752442985, + "learning_rate": 1.1979135945412576e-05, + "loss": 0.5512, + "step": 4177 + }, + { + "epoch": 2.268186753528773, + "grad_norm": 10.131657295101125, + "learning_rate": 1.1975688548620756e-05, + "loss": 0.5678, + "step": 4178 + }, + { + "epoch": 2.268729641693811, + "grad_norm": 11.675191701181866, + "learning_rate": 1.197224090747263e-05, + "loss": 0.5967, + "step": 4179 + }, + { + "epoch": 2.269272529858849, + "grad_norm": 9.879570041096912, + "learning_rate": 1.1968793022394603e-05, + "loss": 1.0245, + "step": 4180 + }, + { + "epoch": 2.269815418023887, + "grad_norm": 12.255119190489571, + "learning_rate": 1.1965344893813117e-05, + "loss": 0.9963, + "step": 4181 + }, + { + "epoch": 2.270358306188925, + "grad_norm": 8.870271458097042, + "learning_rate": 1.1961896522154643e-05, + "loss": 0.3658, + "step": 4182 + }, + { + "epoch": 2.270901194353963, + "grad_norm": 9.50086592138889, + "learning_rate": 1.1958447907845684e-05, + "loss": 0.588, + "step": 4183 + }, + { + "epoch": 2.271444082519001, + "grad_norm": 8.92062913157286, + "learning_rate": 1.1954999051312762e-05, + "loss": 0.6038, + "step": 4184 + }, + { + "epoch": 2.271986970684039, + "grad_norm": 11.511359098427171, + "learning_rate": 1.1951549952982442e-05, + "loss": 0.7837, + "step": 4185 + }, + { + "epoch": 2.272529858849077, + "grad_norm": 7.7230189119336075, + "learning_rate": 1.1948100613281315e-05, + "loss": 0.5235, + "step": 4186 + }, + { + "epoch": 2.273072747014115, + "grad_norm": 8.85303744288786, + "learning_rate": 1.1944651032635997e-05, + "loss": 0.5194, + "step": 4187 + }, + { + "epoch": 2.273615635179153, + "grad_norm": 9.931552012214697, + "learning_rate": 1.1941201211473142e-05, + "loss": 0.6419, + "step": 4188 + }, + { + "epoch": 2.274158523344191, + "grad_norm": 11.635637385284054, + "learning_rate": 1.1937751150219428e-05, + "loss": 0.7948, + "step": 4189 + }, + { + "epoch": 2.274701411509229, + "grad_norm": 7.783494952264655, + "learning_rate": 1.1934300849301561e-05, + "loss": 0.568, + "step": 4190 + }, + { + "epoch": 2.275244299674267, + "grad_norm": 7.979443655042872, + "learning_rate": 1.1930850309146283e-05, + "loss": 0.531, + "step": 4191 + }, + { + "epoch": 2.2757871878393052, + "grad_norm": 12.081321074807914, + "learning_rate": 1.1927399530180359e-05, + "loss": 0.9917, + "step": 4192 + }, + { + "epoch": 2.276330076004343, + "grad_norm": 9.376910406413142, + "learning_rate": 1.1923948512830595e-05, + "loss": 0.6587, + "step": 4193 + }, + { + "epoch": 2.2768729641693812, + "grad_norm": 9.576589478578342, + "learning_rate": 1.1920497257523809e-05, + "loss": 0.5703, + "step": 4194 + }, + { + "epoch": 2.277415852334419, + "grad_norm": 9.129784891039465, + "learning_rate": 1.1917045764686863e-05, + "loss": 0.668, + "step": 4195 + }, + { + "epoch": 2.2779587404994572, + "grad_norm": 11.723388678968943, + "learning_rate": 1.191359403474664e-05, + "loss": 0.6293, + "step": 4196 + }, + { + "epoch": 2.278501628664495, + "grad_norm": 10.662032565840782, + "learning_rate": 1.1910142068130058e-05, + "loss": 1.0251, + "step": 4197 + }, + { + "epoch": 2.2790445168295332, + "grad_norm": 5.976705265828664, + "learning_rate": 1.1906689865264058e-05, + "loss": 0.3892, + "step": 4198 + }, + { + "epoch": 2.279587404994571, + "grad_norm": 9.77428172632025, + "learning_rate": 1.190323742657562e-05, + "loss": 0.8023, + "step": 4199 + }, + { + "epoch": 2.2801302931596092, + "grad_norm": 9.701826347676784, + "learning_rate": 1.1899784752491743e-05, + "loss": 0.6668, + "step": 4200 + }, + { + "epoch": 2.280673181324647, + "grad_norm": 10.939483572398276, + "learning_rate": 1.1896331843439459e-05, + "loss": 0.8751, + "step": 4201 + }, + { + "epoch": 2.2812160694896852, + "grad_norm": 12.656192710489524, + "learning_rate": 1.1892878699845832e-05, + "loss": 0.7874, + "step": 4202 + }, + { + "epoch": 2.281758957654723, + "grad_norm": 8.821175958313784, + "learning_rate": 1.1889425322137948e-05, + "loss": 0.7599, + "step": 4203 + }, + { + "epoch": 2.2823018458197613, + "grad_norm": 11.535196460357659, + "learning_rate": 1.188597171074293e-05, + "loss": 0.7448, + "step": 4204 + }, + { + "epoch": 2.282844733984799, + "grad_norm": 9.221866003570257, + "learning_rate": 1.1882517866087926e-05, + "loss": 0.7176, + "step": 4205 + }, + { + "epoch": 2.2833876221498373, + "grad_norm": 10.791249990779047, + "learning_rate": 1.1879063788600113e-05, + "loss": 0.6905, + "step": 4206 + }, + { + "epoch": 2.283930510314875, + "grad_norm": 11.953403018554845, + "learning_rate": 1.1875609478706694e-05, + "loss": 0.7006, + "step": 4207 + }, + { + "epoch": 2.2844733984799133, + "grad_norm": 10.830659589867402, + "learning_rate": 1.1872154936834905e-05, + "loss": 0.6872, + "step": 4208 + }, + { + "epoch": 2.285016286644951, + "grad_norm": 10.8147041656128, + "learning_rate": 1.186870016341201e-05, + "loss": 0.6429, + "step": 4209 + }, + { + "epoch": 2.2855591748099893, + "grad_norm": 10.70508214004579, + "learning_rate": 1.18652451588653e-05, + "loss": 0.9063, + "step": 4210 + }, + { + "epoch": 2.286102062975027, + "grad_norm": 8.333893553672915, + "learning_rate": 1.1861789923622099e-05, + "loss": 0.5245, + "step": 4211 + }, + { + "epoch": 2.2866449511400653, + "grad_norm": 9.62415991156623, + "learning_rate": 1.1858334458109747e-05, + "loss": 0.734, + "step": 4212 + }, + { + "epoch": 2.287187839305103, + "grad_norm": 15.893278058161348, + "learning_rate": 1.1854878762755633e-05, + "loss": 0.898, + "step": 4213 + }, + { + "epoch": 2.2877307274701413, + "grad_norm": 8.12705480823925, + "learning_rate": 1.1851422837987159e-05, + "loss": 0.5586, + "step": 4214 + }, + { + "epoch": 2.288273615635179, + "grad_norm": 19.440766463719637, + "learning_rate": 1.1847966684231754e-05, + "loss": 1.0309, + "step": 4215 + }, + { + "epoch": 2.2888165038002173, + "grad_norm": 10.590183829770972, + "learning_rate": 1.1844510301916883e-05, + "loss": 0.8545, + "step": 4216 + }, + { + "epoch": 2.289359391965255, + "grad_norm": 9.825370158957117, + "learning_rate": 1.184105369147004e-05, + "loss": 0.7426, + "step": 4217 + }, + { + "epoch": 2.2899022801302933, + "grad_norm": 10.341219416998992, + "learning_rate": 1.1837596853318742e-05, + "loss": 0.8584, + "step": 4218 + }, + { + "epoch": 2.290445168295331, + "grad_norm": 11.121322184848584, + "learning_rate": 1.1834139787890538e-05, + "loss": 0.6598, + "step": 4219 + }, + { + "epoch": 2.2909880564603693, + "grad_norm": 9.600608599739235, + "learning_rate": 1.1830682495613e-05, + "loss": 0.5885, + "step": 4220 + }, + { + "epoch": 2.291530944625407, + "grad_norm": 10.517374576113433, + "learning_rate": 1.1827224976913732e-05, + "loss": 0.751, + "step": 4221 + }, + { + "epoch": 2.2920738327904453, + "grad_norm": 9.666522143402908, + "learning_rate": 1.1823767232220369e-05, + "loss": 0.7602, + "step": 4222 + }, + { + "epoch": 2.292616720955483, + "grad_norm": 9.390811834354404, + "learning_rate": 1.1820309261960567e-05, + "loss": 0.6187, + "step": 4223 + }, + { + "epoch": 2.2931596091205213, + "grad_norm": 8.772653626817414, + "learning_rate": 1.1816851066562017e-05, + "loss": 0.5221, + "step": 4224 + }, + { + "epoch": 2.293702497285559, + "grad_norm": 11.957278593216312, + "learning_rate": 1.1813392646452429e-05, + "loss": 0.9231, + "step": 4225 + }, + { + "epoch": 2.2942453854505973, + "grad_norm": 9.58191568158383, + "learning_rate": 1.1809934002059547e-05, + "loss": 0.6747, + "step": 4226 + }, + { + "epoch": 2.294788273615635, + "grad_norm": 8.242154702381846, + "learning_rate": 1.1806475133811146e-05, + "loss": 0.4053, + "step": 4227 + }, + { + "epoch": 2.2953311617806733, + "grad_norm": 8.83768826569311, + "learning_rate": 1.1803016042135021e-05, + "loss": 0.4177, + "step": 4228 + }, + { + "epoch": 2.295874049945711, + "grad_norm": 7.496802017426867, + "learning_rate": 1.1799556727458996e-05, + "loss": 0.6402, + "step": 4229 + }, + { + "epoch": 2.2964169381107493, + "grad_norm": 10.15598151517949, + "learning_rate": 1.1796097190210927e-05, + "loss": 0.6301, + "step": 4230 + }, + { + "epoch": 2.296959826275787, + "grad_norm": 9.100517578174307, + "learning_rate": 1.1792637430818695e-05, + "loss": 0.6059, + "step": 4231 + }, + { + "epoch": 2.2975027144408253, + "grad_norm": 9.341766570492698, + "learning_rate": 1.178917744971021e-05, + "loss": 0.7462, + "step": 4232 + }, + { + "epoch": 2.298045602605863, + "grad_norm": 9.790098160117006, + "learning_rate": 1.1785717247313407e-05, + "loss": 0.7422, + "step": 4233 + }, + { + "epoch": 2.2985884907709013, + "grad_norm": 11.785065190423747, + "learning_rate": 1.1782256824056249e-05, + "loss": 0.8544, + "step": 4234 + }, + { + "epoch": 2.299131378935939, + "grad_norm": 11.165339576639699, + "learning_rate": 1.1778796180366722e-05, + "loss": 0.6492, + "step": 4235 + }, + { + "epoch": 2.2996742671009773, + "grad_norm": 9.026810304485423, + "learning_rate": 1.1775335316672854e-05, + "loss": 0.751, + "step": 4236 + }, + { + "epoch": 2.300217155266015, + "grad_norm": 10.085898624762818, + "learning_rate": 1.1771874233402682e-05, + "loss": 0.6052, + "step": 4237 + }, + { + "epoch": 2.3007600434310533, + "grad_norm": 9.302471978496024, + "learning_rate": 1.176841293098428e-05, + "loss": 0.7269, + "step": 4238 + }, + { + "epoch": 2.301302931596091, + "grad_norm": 11.406933536282311, + "learning_rate": 1.176495140984575e-05, + "loss": 0.7287, + "step": 4239 + }, + { + "epoch": 2.3018458197611293, + "grad_norm": 8.172755818720033, + "learning_rate": 1.1761489670415218e-05, + "loss": 0.4692, + "step": 4240 + }, + { + "epoch": 2.302388707926167, + "grad_norm": 11.729275499093248, + "learning_rate": 1.1758027713120833e-05, + "loss": 0.8301, + "step": 4241 + }, + { + "epoch": 2.3029315960912053, + "grad_norm": 7.31457707003035, + "learning_rate": 1.1754565538390782e-05, + "loss": 0.5372, + "step": 4242 + }, + { + "epoch": 2.303474484256243, + "grad_norm": 10.15517896282683, + "learning_rate": 1.1751103146653265e-05, + "loss": 0.8744, + "step": 4243 + }, + { + "epoch": 2.3040173724212814, + "grad_norm": 7.780368081203029, + "learning_rate": 1.1747640538336525e-05, + "loss": 0.4743, + "step": 4244 + }, + { + "epoch": 2.304560260586319, + "grad_norm": 7.814816774582812, + "learning_rate": 1.1744177713868814e-05, + "loss": 0.628, + "step": 4245 + }, + { + "epoch": 2.3051031487513574, + "grad_norm": 8.531245398152228, + "learning_rate": 1.1740714673678425e-05, + "loss": 0.4419, + "step": 4246 + }, + { + "epoch": 2.305646036916395, + "grad_norm": 8.091385303458143, + "learning_rate": 1.1737251418193669e-05, + "loss": 0.6249, + "step": 4247 + }, + { + "epoch": 2.3061889250814334, + "grad_norm": 10.253051982681455, + "learning_rate": 1.1733787947842892e-05, + "loss": 0.8892, + "step": 4248 + }, + { + "epoch": 2.306731813246471, + "grad_norm": 8.95406502670429, + "learning_rate": 1.1730324263054453e-05, + "loss": 0.7452, + "step": 4249 + }, + { + "epoch": 2.3072747014115094, + "grad_norm": 10.491159901998326, + "learning_rate": 1.1726860364256756e-05, + "loss": 0.4995, + "step": 4250 + }, + { + "epoch": 2.307817589576547, + "grad_norm": 9.241135791509704, + "learning_rate": 1.1723396251878214e-05, + "loss": 0.7167, + "step": 4251 + }, + { + "epoch": 2.3083604777415854, + "grad_norm": 9.466621171091207, + "learning_rate": 1.1719931926347279e-05, + "loss": 0.5849, + "step": 4252 + }, + { + "epoch": 2.308903365906623, + "grad_norm": 10.933415330757905, + "learning_rate": 1.1716467388092423e-05, + "loss": 1.0411, + "step": 4253 + }, + { + "epoch": 2.3094462540716614, + "grad_norm": 10.009779873858768, + "learning_rate": 1.1713002637542146e-05, + "loss": 0.5456, + "step": 4254 + }, + { + "epoch": 2.309989142236699, + "grad_norm": 11.678038797864954, + "learning_rate": 1.1709537675124966e-05, + "loss": 0.5537, + "step": 4255 + }, + { + "epoch": 2.3105320304017374, + "grad_norm": 11.578035258644414, + "learning_rate": 1.1706072501269448e-05, + "loss": 0.7804, + "step": 4256 + }, + { + "epoch": 2.311074918566775, + "grad_norm": 8.678307181037598, + "learning_rate": 1.1702607116404163e-05, + "loss": 0.5378, + "step": 4257 + }, + { + "epoch": 2.3116178067318134, + "grad_norm": 8.487670920129334, + "learning_rate": 1.1699141520957716e-05, + "loss": 0.558, + "step": 4258 + }, + { + "epoch": 2.312160694896851, + "grad_norm": 10.92371973791016, + "learning_rate": 1.1695675715358737e-05, + "loss": 0.6615, + "step": 4259 + }, + { + "epoch": 2.3127035830618894, + "grad_norm": 9.360985487621843, + "learning_rate": 1.169220970003588e-05, + "loss": 0.5896, + "step": 4260 + }, + { + "epoch": 2.313246471226927, + "grad_norm": 10.382914656511591, + "learning_rate": 1.1688743475417836e-05, + "loss": 0.8869, + "step": 4261 + }, + { + "epoch": 2.3137893593919654, + "grad_norm": 10.27408683638607, + "learning_rate": 1.1685277041933303e-05, + "loss": 1.0929, + "step": 4262 + }, + { + "epoch": 2.314332247557003, + "grad_norm": 10.738574683790112, + "learning_rate": 1.168181040001102e-05, + "loss": 0.845, + "step": 4263 + }, + { + "epoch": 2.3148751357220414, + "grad_norm": 11.783082834204276, + "learning_rate": 1.1678343550079751e-05, + "loss": 0.8358, + "step": 4264 + }, + { + "epoch": 2.315418023887079, + "grad_norm": 9.277577078134165, + "learning_rate": 1.1674876492568272e-05, + "loss": 0.4774, + "step": 4265 + }, + { + "epoch": 2.3159609120521174, + "grad_norm": 10.232983577395474, + "learning_rate": 1.1671409227905401e-05, + "loss": 0.6358, + "step": 4266 + }, + { + "epoch": 2.316503800217155, + "grad_norm": 7.971516257815676, + "learning_rate": 1.1667941756519973e-05, + "loss": 0.5724, + "step": 4267 + }, + { + "epoch": 2.3170466883821934, + "grad_norm": 6.657326238165826, + "learning_rate": 1.1664474078840848e-05, + "loss": 0.3377, + "step": 4268 + }, + { + "epoch": 2.317589576547231, + "grad_norm": 9.831416988328497, + "learning_rate": 1.1661006195296918e-05, + "loss": 0.7115, + "step": 4269 + }, + { + "epoch": 2.3181324647122694, + "grad_norm": 7.2473297381695225, + "learning_rate": 1.1657538106317094e-05, + "loss": 0.4828, + "step": 4270 + }, + { + "epoch": 2.318675352877307, + "grad_norm": 11.474212650892378, + "learning_rate": 1.1654069812330315e-05, + "loss": 0.66, + "step": 4271 + }, + { + "epoch": 2.3192182410423454, + "grad_norm": 9.354650178037069, + "learning_rate": 1.1650601313765547e-05, + "loss": 0.6651, + "step": 4272 + }, + { + "epoch": 2.319761129207383, + "grad_norm": 8.945160340560053, + "learning_rate": 1.1647132611051779e-05, + "loss": 0.6422, + "step": 4273 + }, + { + "epoch": 2.3203040173724214, + "grad_norm": 8.790681463696455, + "learning_rate": 1.164366370461802e-05, + "loss": 0.6387, + "step": 4274 + }, + { + "epoch": 2.320846905537459, + "grad_norm": 8.950872212457588, + "learning_rate": 1.1640194594893317e-05, + "loss": 0.4455, + "step": 4275 + }, + { + "epoch": 2.3213897937024974, + "grad_norm": 9.879748686678086, + "learning_rate": 1.1636725282306733e-05, + "loss": 0.5539, + "step": 4276 + }, + { + "epoch": 2.321932681867535, + "grad_norm": 8.351885716181597, + "learning_rate": 1.1633255767287356e-05, + "loss": 0.4273, + "step": 4277 + }, + { + "epoch": 2.3224755700325734, + "grad_norm": 12.165689696691171, + "learning_rate": 1.1629786050264304e-05, + "loss": 0.6877, + "step": 4278 + }, + { + "epoch": 2.323018458197611, + "grad_norm": 10.70545900286579, + "learning_rate": 1.1626316131666714e-05, + "loss": 0.6667, + "step": 4279 + }, + { + "epoch": 2.3235613463626494, + "grad_norm": 7.521776321190906, + "learning_rate": 1.1622846011923756e-05, + "loss": 0.6538, + "step": 4280 + }, + { + "epoch": 2.324104234527687, + "grad_norm": 10.429691195494106, + "learning_rate": 1.1619375691464614e-05, + "loss": 0.7918, + "step": 4281 + }, + { + "epoch": 2.3246471226927254, + "grad_norm": 13.40241359482942, + "learning_rate": 1.1615905170718507e-05, + "loss": 0.8492, + "step": 4282 + }, + { + "epoch": 2.3251900108577632, + "grad_norm": 9.180074174181755, + "learning_rate": 1.1612434450114674e-05, + "loss": 0.9197, + "step": 4283 + }, + { + "epoch": 2.3257328990228014, + "grad_norm": 8.662388876238891, + "learning_rate": 1.1608963530082375e-05, + "loss": 0.4911, + "step": 4284 + }, + { + "epoch": 2.3262757871878392, + "grad_norm": 12.915716152146215, + "learning_rate": 1.1605492411050907e-05, + "loss": 1.2342, + "step": 4285 + }, + { + "epoch": 2.3268186753528775, + "grad_norm": 15.064512853420336, + "learning_rate": 1.1602021093449577e-05, + "loss": 0.9441, + "step": 4286 + }, + { + "epoch": 2.3273615635179152, + "grad_norm": 8.152078540931088, + "learning_rate": 1.159854957770772e-05, + "loss": 0.4407, + "step": 4287 + }, + { + "epoch": 2.3279044516829535, + "grad_norm": 13.440334362031807, + "learning_rate": 1.1595077864254708e-05, + "loss": 1.0186, + "step": 4288 + }, + { + "epoch": 2.3284473398479912, + "grad_norm": 14.960404464650807, + "learning_rate": 1.1591605953519925e-05, + "loss": 0.854, + "step": 4289 + }, + { + "epoch": 2.3289902280130295, + "grad_norm": 10.871106066518651, + "learning_rate": 1.1588133845932779e-05, + "loss": 0.8522, + "step": 4290 + }, + { + "epoch": 2.3295331161780672, + "grad_norm": 9.218798486570465, + "learning_rate": 1.1584661541922706e-05, + "loss": 0.5529, + "step": 4291 + }, + { + "epoch": 2.3300760043431055, + "grad_norm": 9.644924391242567, + "learning_rate": 1.158118904191917e-05, + "loss": 0.619, + "step": 4292 + }, + { + "epoch": 2.3306188925081432, + "grad_norm": 7.573505575520159, + "learning_rate": 1.1577716346351655e-05, + "loss": 0.5931, + "step": 4293 + }, + { + "epoch": 2.3311617806731815, + "grad_norm": 11.909869183875244, + "learning_rate": 1.1574243455649666e-05, + "loss": 0.8922, + "step": 4294 + }, + { + "epoch": 2.3317046688382193, + "grad_norm": 13.963856423636752, + "learning_rate": 1.1570770370242737e-05, + "loss": 0.9243, + "step": 4295 + }, + { + "epoch": 2.3322475570032575, + "grad_norm": 8.732100079298183, + "learning_rate": 1.1567297090560425e-05, + "loss": 0.5075, + "step": 4296 + }, + { + "epoch": 2.3327904451682953, + "grad_norm": 8.823180159543773, + "learning_rate": 1.1563823617032313e-05, + "loss": 0.5471, + "step": 4297 + }, + { + "epoch": 2.3333333333333335, + "grad_norm": 8.566225161651246, + "learning_rate": 1.1560349950088001e-05, + "loss": 0.5719, + "step": 4298 + }, + { + "epoch": 2.3338762214983713, + "grad_norm": 10.549198011052681, + "learning_rate": 1.155687609015712e-05, + "loss": 0.4775, + "step": 4299 + }, + { + "epoch": 2.3344191096634095, + "grad_norm": 9.79276879932382, + "learning_rate": 1.1553402037669326e-05, + "loss": 0.5429, + "step": 4300 + }, + { + "epoch": 2.3349619978284473, + "grad_norm": 12.065070131397079, + "learning_rate": 1.154992779305429e-05, + "loss": 0.9434, + "step": 4301 + }, + { + "epoch": 2.3355048859934855, + "grad_norm": 11.007790011549671, + "learning_rate": 1.1546453356741714e-05, + "loss": 0.6199, + "step": 4302 + }, + { + "epoch": 2.3360477741585233, + "grad_norm": 12.466762411503309, + "learning_rate": 1.1542978729161326e-05, + "loss": 0.7941, + "step": 4303 + }, + { + "epoch": 2.3365906623235615, + "grad_norm": 11.55197462123976, + "learning_rate": 1.1539503910742866e-05, + "loss": 0.6766, + "step": 4304 + }, + { + "epoch": 2.3371335504885993, + "grad_norm": 8.462143561329682, + "learning_rate": 1.153602890191611e-05, + "loss": 0.5183, + "step": 4305 + }, + { + "epoch": 2.3376764386536375, + "grad_norm": 13.818285051654057, + "learning_rate": 1.1532553703110852e-05, + "loss": 0.8363, + "step": 4306 + }, + { + "epoch": 2.3382193268186753, + "grad_norm": 8.676120343621662, + "learning_rate": 1.1529078314756905e-05, + "loss": 0.6976, + "step": 4307 + }, + { + "epoch": 2.3387622149837135, + "grad_norm": 10.024443385726268, + "learning_rate": 1.1525602737284119e-05, + "loss": 0.4879, + "step": 4308 + }, + { + "epoch": 2.3393051031487513, + "grad_norm": 11.965265440206402, + "learning_rate": 1.1522126971122355e-05, + "loss": 0.8655, + "step": 4309 + }, + { + "epoch": 2.3398479913137895, + "grad_norm": 9.503916256652449, + "learning_rate": 1.15186510167015e-05, + "loss": 0.8966, + "step": 4310 + }, + { + "epoch": 2.3403908794788273, + "grad_norm": 11.942859003506952, + "learning_rate": 1.1515174874451466e-05, + "loss": 0.7769, + "step": 4311 + }, + { + "epoch": 2.3409337676438655, + "grad_norm": 9.50234633554061, + "learning_rate": 1.1511698544802193e-05, + "loss": 0.552, + "step": 4312 + }, + { + "epoch": 2.3414766558089033, + "grad_norm": 8.495987313736284, + "learning_rate": 1.1508222028183632e-05, + "loss": 0.8566, + "step": 4313 + }, + { + "epoch": 2.3420195439739415, + "grad_norm": 11.854978708479615, + "learning_rate": 1.1504745325025766e-05, + "loss": 0.6736, + "step": 4314 + }, + { + "epoch": 2.3425624321389793, + "grad_norm": 11.169148481350696, + "learning_rate": 1.1501268435758602e-05, + "loss": 0.8075, + "step": 4315 + }, + { + "epoch": 2.3431053203040175, + "grad_norm": 11.983506203130691, + "learning_rate": 1.1497791360812163e-05, + "loss": 0.6407, + "step": 4316 + }, + { + "epoch": 2.3436482084690553, + "grad_norm": 8.507852970222672, + "learning_rate": 1.1494314100616502e-05, + "loss": 0.6247, + "step": 4317 + }, + { + "epoch": 2.3441910966340935, + "grad_norm": 19.964294777139003, + "learning_rate": 1.1490836655601692e-05, + "loss": 0.9219, + "step": 4318 + }, + { + "epoch": 2.3447339847991313, + "grad_norm": 8.202074828225685, + "learning_rate": 1.1487359026197828e-05, + "loss": 0.4887, + "step": 4319 + }, + { + "epoch": 2.3452768729641695, + "grad_norm": 10.22404099258637, + "learning_rate": 1.1483881212835029e-05, + "loss": 0.4028, + "step": 4320 + }, + { + "epoch": 2.3458197611292073, + "grad_norm": 8.30587529218613, + "learning_rate": 1.1480403215943436e-05, + "loss": 0.4047, + "step": 4321 + }, + { + "epoch": 2.3463626492942455, + "grad_norm": 7.121924947407594, + "learning_rate": 1.1476925035953217e-05, + "loss": 0.4922, + "step": 4322 + }, + { + "epoch": 2.3469055374592833, + "grad_norm": 10.58793763221932, + "learning_rate": 1.1473446673294552e-05, + "loss": 0.9691, + "step": 4323 + }, + { + "epoch": 2.3474484256243215, + "grad_norm": 11.51770882137354, + "learning_rate": 1.1469968128397656e-05, + "loss": 0.6713, + "step": 4324 + }, + { + "epoch": 2.3479913137893593, + "grad_norm": 9.960125063324698, + "learning_rate": 1.146648940169276e-05, + "loss": 0.8266, + "step": 4325 + }, + { + "epoch": 2.3485342019543975, + "grad_norm": 14.574982915069748, + "learning_rate": 1.1463010493610114e-05, + "loss": 1.4129, + "step": 4326 + }, + { + "epoch": 2.3490770901194353, + "grad_norm": 7.9866843222052, + "learning_rate": 1.145953140458e-05, + "loss": 0.6368, + "step": 4327 + }, + { + "epoch": 2.3496199782844736, + "grad_norm": 14.879386143441858, + "learning_rate": 1.1456052135032717e-05, + "loss": 0.579, + "step": 4328 + }, + { + "epoch": 2.3501628664495113, + "grad_norm": 9.951154243493171, + "learning_rate": 1.1452572685398586e-05, + "loss": 0.6989, + "step": 4329 + }, + { + "epoch": 2.3507057546145496, + "grad_norm": 8.242511521892395, + "learning_rate": 1.144909305610795e-05, + "loss": 0.5213, + "step": 4330 + }, + { + "epoch": 2.3512486427795873, + "grad_norm": 10.259159837692009, + "learning_rate": 1.1445613247591174e-05, + "loss": 0.9849, + "step": 4331 + }, + { + "epoch": 2.3517915309446256, + "grad_norm": 7.97479592417325, + "learning_rate": 1.1442133260278648e-05, + "loss": 0.5867, + "step": 4332 + }, + { + "epoch": 2.3523344191096633, + "grad_norm": 8.493125656661944, + "learning_rate": 1.143865309460078e-05, + "loss": 0.9095, + "step": 4333 + }, + { + "epoch": 2.3528773072747016, + "grad_norm": 12.49067508129187, + "learning_rate": 1.1435172750988009e-05, + "loss": 0.8484, + "step": 4334 + }, + { + "epoch": 2.3534201954397393, + "grad_norm": 11.134649233778442, + "learning_rate": 1.1431692229870782e-05, + "loss": 0.8422, + "step": 4335 + }, + { + "epoch": 2.3539630836047776, + "grad_norm": 6.99064506687852, + "learning_rate": 1.1428211531679576e-05, + "loss": 0.5388, + "step": 4336 + }, + { + "epoch": 2.3545059717698154, + "grad_norm": 7.798255243454909, + "learning_rate": 1.1424730656844895e-05, + "loss": 0.593, + "step": 4337 + }, + { + "epoch": 2.3550488599348536, + "grad_norm": 8.178641043550261, + "learning_rate": 1.1421249605797255e-05, + "loss": 0.4778, + "step": 4338 + }, + { + "epoch": 2.3555917480998914, + "grad_norm": 8.917024765042555, + "learning_rate": 1.1417768378967196e-05, + "loss": 0.5782, + "step": 4339 + }, + { + "epoch": 2.3561346362649296, + "grad_norm": 9.742472427719518, + "learning_rate": 1.1414286976785286e-05, + "loss": 0.8709, + "step": 4340 + }, + { + "epoch": 2.3566775244299674, + "grad_norm": 10.08457903035175, + "learning_rate": 1.141080539968211e-05, + "loss": 0.5582, + "step": 4341 + }, + { + "epoch": 2.3572204125950056, + "grad_norm": 10.74958408516018, + "learning_rate": 1.1407323648088275e-05, + "loss": 0.7498, + "step": 4342 + }, + { + "epoch": 2.3577633007600434, + "grad_norm": 12.356830179446728, + "learning_rate": 1.1403841722434405e-05, + "loss": 0.9357, + "step": 4343 + }, + { + "epoch": 2.3583061889250816, + "grad_norm": 8.220220873503408, + "learning_rate": 1.1400359623151153e-05, + "loss": 0.5554, + "step": 4344 + }, + { + "epoch": 2.3588490770901194, + "grad_norm": 10.916598567149965, + "learning_rate": 1.1396877350669194e-05, + "loss": 0.8239, + "step": 4345 + }, + { + "epoch": 2.3593919652551576, + "grad_norm": 10.51236366107464, + "learning_rate": 1.1393394905419215e-05, + "loss": 0.5945, + "step": 4346 + }, + { + "epoch": 2.3599348534201954, + "grad_norm": 10.534153676139686, + "learning_rate": 1.1389912287831935e-05, + "loss": 0.575, + "step": 4347 + }, + { + "epoch": 2.3604777415852336, + "grad_norm": 10.06758846687261, + "learning_rate": 1.1386429498338088e-05, + "loss": 0.702, + "step": 4348 + }, + { + "epoch": 2.3610206297502714, + "grad_norm": 9.282335539362101, + "learning_rate": 1.1382946537368432e-05, + "loss": 0.5645, + "step": 4349 + }, + { + "epoch": 2.3615635179153096, + "grad_norm": 8.137108078430895, + "learning_rate": 1.1379463405353741e-05, + "loss": 0.5057, + "step": 4350 + }, + { + "epoch": 2.3621064060803474, + "grad_norm": 10.651520296934317, + "learning_rate": 1.137598010272482e-05, + "loss": 0.762, + "step": 4351 + }, + { + "epoch": 2.3626492942453856, + "grad_norm": 10.001083727438939, + "learning_rate": 1.1372496629912486e-05, + "loss": 0.7255, + "step": 4352 + }, + { + "epoch": 2.3631921824104234, + "grad_norm": 12.180927670825834, + "learning_rate": 1.1369012987347582e-05, + "loss": 0.7295, + "step": 4353 + }, + { + "epoch": 2.3637350705754616, + "grad_norm": 10.697210588401258, + "learning_rate": 1.136552917546097e-05, + "loss": 0.5159, + "step": 4354 + }, + { + "epoch": 2.3642779587404994, + "grad_norm": 13.551389197850602, + "learning_rate": 1.1362045194683534e-05, + "loss": 0.9265, + "step": 4355 + }, + { + "epoch": 2.3648208469055376, + "grad_norm": 12.38857946888456, + "learning_rate": 1.1358561045446175e-05, + "loss": 0.7577, + "step": 4356 + }, + { + "epoch": 2.3653637350705754, + "grad_norm": 10.138108604403497, + "learning_rate": 1.1355076728179824e-05, + "loss": 0.5328, + "step": 4357 + }, + { + "epoch": 2.3659066232356136, + "grad_norm": 9.823785539850283, + "learning_rate": 1.1351592243315416e-05, + "loss": 0.6116, + "step": 4358 + }, + { + "epoch": 2.3664495114006514, + "grad_norm": 8.238880239030014, + "learning_rate": 1.1348107591283932e-05, + "loss": 0.3863, + "step": 4359 + }, + { + "epoch": 2.3669923995656896, + "grad_norm": 11.34031175123055, + "learning_rate": 1.1344622772516352e-05, + "loss": 0.9601, + "step": 4360 + }, + { + "epoch": 2.3675352877307274, + "grad_norm": 11.270983899241404, + "learning_rate": 1.1341137787443683e-05, + "loss": 0.7598, + "step": 4361 + }, + { + "epoch": 2.3680781758957656, + "grad_norm": 10.649130568340542, + "learning_rate": 1.1337652636496959e-05, + "loss": 0.6547, + "step": 4362 + }, + { + "epoch": 2.3686210640608034, + "grad_norm": 9.462979596192575, + "learning_rate": 1.1334167320107224e-05, + "loss": 0.5247, + "step": 4363 + }, + { + "epoch": 2.3691639522258416, + "grad_norm": 11.496313370817235, + "learning_rate": 1.1330681838705544e-05, + "loss": 0.5465, + "step": 4364 + }, + { + "epoch": 2.3697068403908794, + "grad_norm": 10.125563730301518, + "learning_rate": 1.1327196192723016e-05, + "loss": 0.5687, + "step": 4365 + }, + { + "epoch": 2.3702497285559176, + "grad_norm": 14.511657264928447, + "learning_rate": 1.1323710382590747e-05, + "loss": 0.7274, + "step": 4366 + }, + { + "epoch": 2.3707926167209554, + "grad_norm": 10.08856219742556, + "learning_rate": 1.1320224408739868e-05, + "loss": 0.4633, + "step": 4367 + }, + { + "epoch": 2.3713355048859937, + "grad_norm": 9.943874647753937, + "learning_rate": 1.131673827160153e-05, + "loss": 0.7042, + "step": 4368 + }, + { + "epoch": 2.3718783930510314, + "grad_norm": 10.550654748070723, + "learning_rate": 1.1313251971606906e-05, + "loss": 0.6336, + "step": 4369 + }, + { + "epoch": 2.3724212812160697, + "grad_norm": 14.84394771322204, + "learning_rate": 1.1309765509187182e-05, + "loss": 0.6591, + "step": 4370 + }, + { + "epoch": 2.3729641693811074, + "grad_norm": 10.962774929698096, + "learning_rate": 1.1306278884773576e-05, + "loss": 0.9648, + "step": 4371 + }, + { + "epoch": 2.3735070575461457, + "grad_norm": 9.063486875880555, + "learning_rate": 1.1302792098797312e-05, + "loss": 0.4618, + "step": 4372 + }, + { + "epoch": 2.3740499457111834, + "grad_norm": 10.458070080982525, + "learning_rate": 1.1299305151689648e-05, + "loss": 0.6216, + "step": 4373 + }, + { + "epoch": 2.3745928338762217, + "grad_norm": 10.041421519621146, + "learning_rate": 1.1295818043881852e-05, + "loss": 0.6642, + "step": 4374 + }, + { + "epoch": 2.3751357220412594, + "grad_norm": 8.817266669804289, + "learning_rate": 1.1292330775805214e-05, + "loss": 0.4725, + "step": 4375 + }, + { + "epoch": 2.3756786102062977, + "grad_norm": 12.823614653687773, + "learning_rate": 1.1288843347891045e-05, + "loss": 0.553, + "step": 4376 + }, + { + "epoch": 2.3762214983713354, + "grad_norm": 9.281557160359569, + "learning_rate": 1.1285355760570677e-05, + "loss": 0.6436, + "step": 4377 + }, + { + "epoch": 2.3767643865363737, + "grad_norm": 13.234297666234115, + "learning_rate": 1.1281868014275458e-05, + "loss": 0.7665, + "step": 4378 + }, + { + "epoch": 2.3773072747014115, + "grad_norm": 13.013112559574942, + "learning_rate": 1.1278380109436764e-05, + "loss": 0.9087, + "step": 4379 + }, + { + "epoch": 2.3778501628664497, + "grad_norm": 9.413433369666237, + "learning_rate": 1.1274892046485979e-05, + "loss": 0.5248, + "step": 4380 + }, + { + "epoch": 2.3783930510314875, + "grad_norm": 10.865873661981059, + "learning_rate": 1.1271403825854516e-05, + "loss": 0.7398, + "step": 4381 + }, + { + "epoch": 2.3789359391965257, + "grad_norm": 8.516596816721957, + "learning_rate": 1.12679154479738e-05, + "loss": 0.4652, + "step": 4382 + }, + { + "epoch": 2.3794788273615635, + "grad_norm": 7.9464803374429085, + "learning_rate": 1.1264426913275277e-05, + "loss": 0.6952, + "step": 4383 + }, + { + "epoch": 2.3800217155266017, + "grad_norm": 9.171877282099603, + "learning_rate": 1.1260938222190423e-05, + "loss": 0.7534, + "step": 4384 + }, + { + "epoch": 2.3805646036916395, + "grad_norm": 10.833526250938174, + "learning_rate": 1.1257449375150717e-05, + "loss": 0.7966, + "step": 4385 + }, + { + "epoch": 2.3811074918566777, + "grad_norm": 7.087015821603363, + "learning_rate": 1.1253960372587669e-05, + "loss": 0.5145, + "step": 4386 + }, + { + "epoch": 2.3816503800217155, + "grad_norm": 9.355325600419727, + "learning_rate": 1.1250471214932805e-05, + "loss": 0.6084, + "step": 4387 + }, + { + "epoch": 2.3821932681867537, + "grad_norm": 11.679312988029317, + "learning_rate": 1.1246981902617666e-05, + "loss": 0.7734, + "step": 4388 + }, + { + "epoch": 2.3827361563517915, + "grad_norm": 11.696985760230557, + "learning_rate": 1.124349243607382e-05, + "loss": 0.7843, + "step": 4389 + }, + { + "epoch": 2.3832790445168297, + "grad_norm": 11.045261072625825, + "learning_rate": 1.1240002815732847e-05, + "loss": 0.8102, + "step": 4390 + }, + { + "epoch": 2.3838219326818675, + "grad_norm": 13.853716237422205, + "learning_rate": 1.123651304202635e-05, + "loss": 0.6548, + "step": 4391 + }, + { + "epoch": 2.3843648208469057, + "grad_norm": 9.256284103603257, + "learning_rate": 1.1233023115385949e-05, + "loss": 0.5769, + "step": 4392 + }, + { + "epoch": 2.3849077090119435, + "grad_norm": 11.719289709502613, + "learning_rate": 1.1229533036243283e-05, + "loss": 0.9445, + "step": 4393 + }, + { + "epoch": 2.3854505971769817, + "grad_norm": 13.181528127455582, + "learning_rate": 1.1226042805030014e-05, + "loss": 0.78, + "step": 4394 + }, + { + "epoch": 2.3859934853420195, + "grad_norm": 10.610212858388932, + "learning_rate": 1.1222552422177813e-05, + "loss": 0.9325, + "step": 4395 + }, + { + "epoch": 2.3865363735070577, + "grad_norm": 8.750069163521957, + "learning_rate": 1.1219061888118381e-05, + "loss": 0.5576, + "step": 4396 + }, + { + "epoch": 2.3870792616720955, + "grad_norm": 11.697241415039246, + "learning_rate": 1.121557120328343e-05, + "loss": 0.6042, + "step": 4397 + }, + { + "epoch": 2.3876221498371337, + "grad_norm": 14.150909202475516, + "learning_rate": 1.1212080368104699e-05, + "loss": 0.8888, + "step": 4398 + }, + { + "epoch": 2.3881650380021715, + "grad_norm": 8.45338495407854, + "learning_rate": 1.1208589383013934e-05, + "loss": 0.5394, + "step": 4399 + }, + { + "epoch": 2.3887079261672097, + "grad_norm": 8.581855639244504, + "learning_rate": 1.120509824844291e-05, + "loss": 0.4231, + "step": 4400 + }, + { + "epoch": 2.3892508143322475, + "grad_norm": 9.723553293714806, + "learning_rate": 1.1201606964823415e-05, + "loss": 0.5184, + "step": 4401 + }, + { + "epoch": 2.3897937024972857, + "grad_norm": 11.437338247112033, + "learning_rate": 1.1198115532587254e-05, + "loss": 0.9033, + "step": 4402 + }, + { + "epoch": 2.3903365906623235, + "grad_norm": 11.185644591472458, + "learning_rate": 1.1194623952166252e-05, + "loss": 0.7607, + "step": 4403 + }, + { + "epoch": 2.3908794788273617, + "grad_norm": 11.125764861054968, + "learning_rate": 1.1191132223992258e-05, + "loss": 0.7844, + "step": 4404 + }, + { + "epoch": 2.3914223669923995, + "grad_norm": 10.915426846928037, + "learning_rate": 1.1187640348497131e-05, + "loss": 0.5862, + "step": 4405 + }, + { + "epoch": 2.3919652551574377, + "grad_norm": 8.032417531111763, + "learning_rate": 1.1184148326112753e-05, + "loss": 0.5992, + "step": 4406 + }, + { + "epoch": 2.3925081433224755, + "grad_norm": 10.322050440175012, + "learning_rate": 1.1180656157271026e-05, + "loss": 0.6275, + "step": 4407 + }, + { + "epoch": 2.3930510314875137, + "grad_norm": 5.756218508975568, + "learning_rate": 1.117716384240386e-05, + "loss": 0.3605, + "step": 4408 + }, + { + "epoch": 2.3935939196525515, + "grad_norm": 13.072320066260373, + "learning_rate": 1.1173671381943197e-05, + "loss": 1.1771, + "step": 4409 + }, + { + "epoch": 2.3941368078175898, + "grad_norm": 8.390483377503248, + "learning_rate": 1.1170178776320982e-05, + "loss": 0.484, + "step": 4410 + }, + { + "epoch": 2.3946796959826275, + "grad_norm": 9.168734569918955, + "learning_rate": 1.1166686025969196e-05, + "loss": 0.4637, + "step": 4411 + }, + { + "epoch": 2.3952225841476658, + "grad_norm": 11.945726197992448, + "learning_rate": 1.116319313131982e-05, + "loss": 0.6501, + "step": 4412 + }, + { + "epoch": 2.3957654723127035, + "grad_norm": 9.925140956274975, + "learning_rate": 1.1159700092804866e-05, + "loss": 0.5583, + "step": 4413 + }, + { + "epoch": 2.3963083604777418, + "grad_norm": 10.0377125619045, + "learning_rate": 1.1156206910856353e-05, + "loss": 0.5783, + "step": 4414 + }, + { + "epoch": 2.3968512486427795, + "grad_norm": 13.878494510161136, + "learning_rate": 1.115271358590633e-05, + "loss": 0.7029, + "step": 4415 + }, + { + "epoch": 2.3973941368078178, + "grad_norm": 8.414419089776398, + "learning_rate": 1.114922011838685e-05, + "loss": 0.4398, + "step": 4416 + }, + { + "epoch": 2.3979370249728555, + "grad_norm": 10.455474671185716, + "learning_rate": 1.1145726508729994e-05, + "loss": 0.5462, + "step": 4417 + }, + { + "epoch": 2.3984799131378938, + "grad_norm": 8.80305307358664, + "learning_rate": 1.1142232757367861e-05, + "loss": 0.3885, + "step": 4418 + }, + { + "epoch": 2.3990228013029316, + "grad_norm": 9.381551586854757, + "learning_rate": 1.1138738864732559e-05, + "loss": 0.6646, + "step": 4419 + }, + { + "epoch": 2.3995656894679698, + "grad_norm": 9.409317053539013, + "learning_rate": 1.113524483125622e-05, + "loss": 0.7772, + "step": 4420 + }, + { + "epoch": 2.4001085776330076, + "grad_norm": 7.414798924114551, + "learning_rate": 1.113175065737099e-05, + "loss": 0.4508, + "step": 4421 + }, + { + "epoch": 2.400651465798046, + "grad_norm": 10.931216387643186, + "learning_rate": 1.1128256343509032e-05, + "loss": 0.6434, + "step": 4422 + }, + { + "epoch": 2.4011943539630836, + "grad_norm": 9.811349462515302, + "learning_rate": 1.1124761890102535e-05, + "loss": 0.5104, + "step": 4423 + }, + { + "epoch": 2.401737242128122, + "grad_norm": 10.856233458965448, + "learning_rate": 1.1121267297583694e-05, + "loss": 0.6535, + "step": 4424 + }, + { + "epoch": 2.4022801302931596, + "grad_norm": 10.2119482085421, + "learning_rate": 1.1117772566384725e-05, + "loss": 0.593, + "step": 4425 + }, + { + "epoch": 2.402823018458198, + "grad_norm": 11.327455944037217, + "learning_rate": 1.1114277696937865e-05, + "loss": 0.9199, + "step": 4426 + }, + { + "epoch": 2.4033659066232356, + "grad_norm": 10.64004734338448, + "learning_rate": 1.1110782689675365e-05, + "loss": 0.793, + "step": 4427 + }, + { + "epoch": 2.403908794788274, + "grad_norm": 5.71867532259649, + "learning_rate": 1.110728754502949e-05, + "loss": 0.2726, + "step": 4428 + }, + { + "epoch": 2.4044516829533116, + "grad_norm": 10.33820400695821, + "learning_rate": 1.1103792263432525e-05, + "loss": 0.787, + "step": 4429 + }, + { + "epoch": 2.40499457111835, + "grad_norm": 10.02381386485252, + "learning_rate": 1.1100296845316777e-05, + "loss": 0.7701, + "step": 4430 + }, + { + "epoch": 2.4055374592833876, + "grad_norm": 12.111852512562214, + "learning_rate": 1.1096801291114561e-05, + "loss": 0.6123, + "step": 4431 + }, + { + "epoch": 2.406080347448426, + "grad_norm": 8.432777542688768, + "learning_rate": 1.1093305601258213e-05, + "loss": 0.4817, + "step": 4432 + }, + { + "epoch": 2.4066232356134636, + "grad_norm": 12.648769211741241, + "learning_rate": 1.1089809776180084e-05, + "loss": 0.7328, + "step": 4433 + }, + { + "epoch": 2.407166123778502, + "grad_norm": 9.260727895387426, + "learning_rate": 1.1086313816312546e-05, + "loss": 0.2951, + "step": 4434 + }, + { + "epoch": 2.4077090119435396, + "grad_norm": 7.801968230318713, + "learning_rate": 1.1082817722087981e-05, + "loss": 0.6204, + "step": 4435 + }, + { + "epoch": 2.408251900108578, + "grad_norm": 18.28104514211005, + "learning_rate": 1.10793214939388e-05, + "loss": 1.2717, + "step": 4436 + }, + { + "epoch": 2.4087947882736156, + "grad_norm": 8.074911353186037, + "learning_rate": 1.1075825132297414e-05, + "loss": 0.391, + "step": 4437 + }, + { + "epoch": 2.409337676438654, + "grad_norm": 10.223302145592744, + "learning_rate": 1.1072328637596259e-05, + "loss": 0.8764, + "step": 4438 + }, + { + "epoch": 2.4098805646036916, + "grad_norm": 8.909676311441572, + "learning_rate": 1.1068832010267792e-05, + "loss": 0.5866, + "step": 4439 + }, + { + "epoch": 2.41042345276873, + "grad_norm": 8.270875830864977, + "learning_rate": 1.1065335250744478e-05, + "loss": 0.4935, + "step": 4440 + }, + { + "epoch": 2.4109663409337676, + "grad_norm": 8.125694150069458, + "learning_rate": 1.1061838359458801e-05, + "loss": 0.5855, + "step": 4441 + }, + { + "epoch": 2.411509229098806, + "grad_norm": 9.00158598137614, + "learning_rate": 1.1058341336843264e-05, + "loss": 0.4464, + "step": 4442 + }, + { + "epoch": 2.4120521172638436, + "grad_norm": 10.568261973334824, + "learning_rate": 1.1054844183330385e-05, + "loss": 0.844, + "step": 4443 + }, + { + "epoch": 2.412595005428882, + "grad_norm": 10.29648068473729, + "learning_rate": 1.1051346899352693e-05, + "loss": 0.7083, + "step": 4444 + }, + { + "epoch": 2.4131378935939196, + "grad_norm": 10.658053637486233, + "learning_rate": 1.1047849485342744e-05, + "loss": 0.917, + "step": 4445 + }, + { + "epoch": 2.413680781758958, + "grad_norm": 10.85927844636501, + "learning_rate": 1.1044351941733097e-05, + "loss": 0.7737, + "step": 4446 + }, + { + "epoch": 2.4142236699239956, + "grad_norm": 13.105614857394716, + "learning_rate": 1.1040854268956343e-05, + "loss": 0.7626, + "step": 4447 + }, + { + "epoch": 2.414766558089034, + "grad_norm": 12.088913830608213, + "learning_rate": 1.1037356467445071e-05, + "loss": 0.7439, + "step": 4448 + }, + { + "epoch": 2.4153094462540716, + "grad_norm": 13.524031933483466, + "learning_rate": 1.1033858537631894e-05, + "loss": 0.6172, + "step": 4449 + }, + { + "epoch": 2.41585233441911, + "grad_norm": 12.25037908572634, + "learning_rate": 1.1030360479949452e-05, + "loss": 0.7959, + "step": 4450 + }, + { + "epoch": 2.4163952225841476, + "grad_norm": 10.26766943217781, + "learning_rate": 1.102686229483038e-05, + "loss": 0.8928, + "step": 4451 + }, + { + "epoch": 2.416938110749186, + "grad_norm": 10.66247714207056, + "learning_rate": 1.1023363982707345e-05, + "loss": 0.8498, + "step": 4452 + }, + { + "epoch": 2.4174809989142236, + "grad_norm": 9.37397925289073, + "learning_rate": 1.101986554401302e-05, + "loss": 0.5633, + "step": 4453 + }, + { + "epoch": 2.418023887079262, + "grad_norm": 8.656197915403546, + "learning_rate": 1.10163669791801e-05, + "loss": 0.6403, + "step": 4454 + }, + { + "epoch": 2.4185667752442996, + "grad_norm": 10.199878069425685, + "learning_rate": 1.101286828864129e-05, + "loss": 0.7139, + "step": 4455 + }, + { + "epoch": 2.419109663409338, + "grad_norm": 8.176125775602177, + "learning_rate": 1.1009369472829316e-05, + "loss": 0.5322, + "step": 4456 + }, + { + "epoch": 2.4196525515743756, + "grad_norm": 9.441655350085792, + "learning_rate": 1.1005870532176918e-05, + "loss": 0.8574, + "step": 4457 + }, + { + "epoch": 2.420195439739414, + "grad_norm": 10.487844747641686, + "learning_rate": 1.100237146711685e-05, + "loss": 0.7399, + "step": 4458 + }, + { + "epoch": 2.4207383279044516, + "grad_norm": 8.456961156540515, + "learning_rate": 1.0998872278081883e-05, + "loss": 0.5795, + "step": 4459 + }, + { + "epoch": 2.42128121606949, + "grad_norm": 10.247575624097415, + "learning_rate": 1.0995372965504798e-05, + "loss": 0.7048, + "step": 4460 + }, + { + "epoch": 2.4218241042345277, + "grad_norm": 8.127194271974078, + "learning_rate": 1.09918735298184e-05, + "loss": 0.6339, + "step": 4461 + }, + { + "epoch": 2.422366992399566, + "grad_norm": 11.080151666019892, + "learning_rate": 1.0988373971455502e-05, + "loss": 0.6, + "step": 4462 + }, + { + "epoch": 2.4229098805646037, + "grad_norm": 9.249791609753641, + "learning_rate": 1.0984874290848935e-05, + "loss": 0.5823, + "step": 4463 + }, + { + "epoch": 2.423452768729642, + "grad_norm": 8.673439204574109, + "learning_rate": 1.0981374488431546e-05, + "loss": 0.4486, + "step": 4464 + }, + { + "epoch": 2.4239956568946797, + "grad_norm": 8.530441262511525, + "learning_rate": 1.0977874564636199e-05, + "loss": 0.4685, + "step": 4465 + }, + { + "epoch": 2.424538545059718, + "grad_norm": 11.975099667402777, + "learning_rate": 1.0974374519895765e-05, + "loss": 0.5608, + "step": 4466 + }, + { + "epoch": 2.4250814332247557, + "grad_norm": 12.28650683165383, + "learning_rate": 1.097087435464314e-05, + "loss": 0.6428, + "step": 4467 + }, + { + "epoch": 2.425624321389794, + "grad_norm": 11.443904338187624, + "learning_rate": 1.0967374069311224e-05, + "loss": 0.5093, + "step": 4468 + }, + { + "epoch": 2.4261672095548317, + "grad_norm": 12.11642038793139, + "learning_rate": 1.0963873664332944e-05, + "loss": 0.7573, + "step": 4469 + }, + { + "epoch": 2.42671009771987, + "grad_norm": 12.716855768902285, + "learning_rate": 1.0960373140141232e-05, + "loss": 1.0742, + "step": 4470 + }, + { + "epoch": 2.4272529858849077, + "grad_norm": 8.060615842022353, + "learning_rate": 1.095687249716904e-05, + "loss": 0.5299, + "step": 4471 + }, + { + "epoch": 2.427795874049946, + "grad_norm": 10.663178608260651, + "learning_rate": 1.0953371735849329e-05, + "loss": 0.6575, + "step": 4472 + }, + { + "epoch": 2.4283387622149837, + "grad_norm": 11.883215571850693, + "learning_rate": 1.0949870856615088e-05, + "loss": 0.7516, + "step": 4473 + }, + { + "epoch": 2.428881650380022, + "grad_norm": 8.725361323128087, + "learning_rate": 1.0946369859899299e-05, + "loss": 0.5543, + "step": 4474 + }, + { + "epoch": 2.4294245385450597, + "grad_norm": 7.964297132373328, + "learning_rate": 1.0942868746134983e-05, + "loss": 0.3799, + "step": 4475 + }, + { + "epoch": 2.429967426710098, + "grad_norm": 11.129935043446988, + "learning_rate": 1.0939367515755155e-05, + "loss": 0.5274, + "step": 4476 + }, + { + "epoch": 2.4305103148751357, + "grad_norm": 8.84506268400313, + "learning_rate": 1.0935866169192858e-05, + "loss": 0.6314, + "step": 4477 + }, + { + "epoch": 2.431053203040174, + "grad_norm": 9.898346545614544, + "learning_rate": 1.0932364706881138e-05, + "loss": 0.4733, + "step": 4478 + }, + { + "epoch": 2.4315960912052117, + "grad_norm": 12.708528067939865, + "learning_rate": 1.092886312925307e-05, + "loss": 1.1591, + "step": 4479 + }, + { + "epoch": 2.43213897937025, + "grad_norm": 8.852913949596426, + "learning_rate": 1.0925361436741726e-05, + "loss": 0.4934, + "step": 4480 + }, + { + "epoch": 2.4326818675352877, + "grad_norm": 9.976618553528468, + "learning_rate": 1.0921859629780205e-05, + "loss": 0.5726, + "step": 4481 + }, + { + "epoch": 2.433224755700326, + "grad_norm": 10.591185923462703, + "learning_rate": 1.0918357708801617e-05, + "loss": 0.9099, + "step": 4482 + }, + { + "epoch": 2.4337676438653637, + "grad_norm": 9.782319605062739, + "learning_rate": 1.0914855674239083e-05, + "loss": 0.7016, + "step": 4483 + }, + { + "epoch": 2.434310532030402, + "grad_norm": 7.962337485760134, + "learning_rate": 1.0911353526525741e-05, + "loss": 0.3899, + "step": 4484 + }, + { + "epoch": 2.4348534201954397, + "grad_norm": 10.926665567713723, + "learning_rate": 1.0907851266094742e-05, + "loss": 1.0416, + "step": 4485 + }, + { + "epoch": 2.435396308360478, + "grad_norm": 10.774706653267238, + "learning_rate": 1.0904348893379252e-05, + "loss": 0.8539, + "step": 4486 + }, + { + "epoch": 2.4359391965255157, + "grad_norm": 9.962550119554402, + "learning_rate": 1.0900846408812449e-05, + "loss": 0.6202, + "step": 4487 + }, + { + "epoch": 2.436482084690554, + "grad_norm": 7.425101710467455, + "learning_rate": 1.0897343812827528e-05, + "loss": 0.4138, + "step": 4488 + }, + { + "epoch": 2.4370249728555917, + "grad_norm": 11.159990654051901, + "learning_rate": 1.0893841105857695e-05, + "loss": 0.7927, + "step": 4489 + }, + { + "epoch": 2.4375678610206295, + "grad_norm": 8.584412414000917, + "learning_rate": 1.0890338288336168e-05, + "loss": 0.4596, + "step": 4490 + }, + { + "epoch": 2.4381107491856677, + "grad_norm": 13.182969196780135, + "learning_rate": 1.0886835360696182e-05, + "loss": 0.7052, + "step": 4491 + }, + { + "epoch": 2.438653637350706, + "grad_norm": 12.643144928400368, + "learning_rate": 1.0883332323370986e-05, + "loss": 0.7273, + "step": 4492 + }, + { + "epoch": 2.4391965255157437, + "grad_norm": 11.758017246786887, + "learning_rate": 1.087982917679384e-05, + "loss": 0.6389, + "step": 4493 + }, + { + "epoch": 2.4397394136807815, + "grad_norm": 10.332831381965816, + "learning_rate": 1.0876325921398016e-05, + "loss": 0.49, + "step": 4494 + }, + { + "epoch": 2.4402823018458197, + "grad_norm": 12.79909493300538, + "learning_rate": 1.0872822557616811e-05, + "loss": 1.027, + "step": 4495 + }, + { + "epoch": 2.440825190010858, + "grad_norm": 10.225359261113429, + "learning_rate": 1.0869319085883518e-05, + "loss": 0.6319, + "step": 4496 + }, + { + "epoch": 2.4413680781758957, + "grad_norm": 10.600390533226815, + "learning_rate": 1.0865815506631456e-05, + "loss": 0.5513, + "step": 4497 + }, + { + "epoch": 2.4419109663409335, + "grad_norm": 9.069709981087057, + "learning_rate": 1.0862311820293951e-05, + "loss": 0.4358, + "step": 4498 + }, + { + "epoch": 2.4424538545059717, + "grad_norm": 10.228930326876846, + "learning_rate": 1.0858808027304346e-05, + "loss": 1.0149, + "step": 4499 + }, + { + "epoch": 2.44299674267101, + "grad_norm": 6.981580424693794, + "learning_rate": 1.0855304128095997e-05, + "loss": 0.3823, + "step": 4500 + }, + { + "epoch": 2.4435396308360477, + "grad_norm": 12.347328486710586, + "learning_rate": 1.085180012310227e-05, + "loss": 0.8202, + "step": 4501 + }, + { + "epoch": 2.4440825190010855, + "grad_norm": 7.794475194787575, + "learning_rate": 1.0848296012756545e-05, + "loss": 0.4972, + "step": 4502 + }, + { + "epoch": 2.4446254071661238, + "grad_norm": 8.640639888495457, + "learning_rate": 1.0844791797492218e-05, + "loss": 0.4174, + "step": 4503 + }, + { + "epoch": 2.445168295331162, + "grad_norm": 9.363787480425108, + "learning_rate": 1.0841287477742695e-05, + "loss": 0.5783, + "step": 4504 + }, + { + "epoch": 2.4457111834961998, + "grad_norm": 9.455569437628846, + "learning_rate": 1.0837783053941396e-05, + "loss": 0.5611, + "step": 4505 + }, + { + "epoch": 2.4462540716612375, + "grad_norm": 7.587727857256139, + "learning_rate": 1.083427852652175e-05, + "loss": 0.4129, + "step": 4506 + }, + { + "epoch": 2.4467969598262758, + "grad_norm": 11.579317957056839, + "learning_rate": 1.083077389591721e-05, + "loss": 0.6628, + "step": 4507 + }, + { + "epoch": 2.447339847991314, + "grad_norm": 9.130000079525795, + "learning_rate": 1.0827269162561229e-05, + "loss": 0.4626, + "step": 4508 + }, + { + "epoch": 2.4478827361563518, + "grad_norm": 8.791608822057356, + "learning_rate": 1.0823764326887283e-05, + "loss": 0.7441, + "step": 4509 + }, + { + "epoch": 2.4484256243213895, + "grad_norm": 8.698379661142646, + "learning_rate": 1.0820259389328852e-05, + "loss": 0.4924, + "step": 4510 + }, + { + "epoch": 2.4489685124864278, + "grad_norm": 12.811407852860862, + "learning_rate": 1.0816754350319427e-05, + "loss": 0.9829, + "step": 4511 + }, + { + "epoch": 2.449511400651466, + "grad_norm": 11.588653893201888, + "learning_rate": 1.0813249210292525e-05, + "loss": 0.7218, + "step": 4512 + }, + { + "epoch": 2.450054288816504, + "grad_norm": 12.334703169006005, + "learning_rate": 1.0809743969681665e-05, + "loss": 0.8737, + "step": 4513 + }, + { + "epoch": 2.4505971769815416, + "grad_norm": 12.583313163331177, + "learning_rate": 1.080623862892038e-05, + "loss": 0.7226, + "step": 4514 + }, + { + "epoch": 2.45114006514658, + "grad_norm": 12.11103490873811, + "learning_rate": 1.0802733188442216e-05, + "loss": 0.7867, + "step": 4515 + }, + { + "epoch": 2.451682953311618, + "grad_norm": 11.222788743446143, + "learning_rate": 1.0799227648680732e-05, + "loss": 0.5457, + "step": 4516 + }, + { + "epoch": 2.452225841476656, + "grad_norm": 7.685503695913846, + "learning_rate": 1.07957220100695e-05, + "loss": 0.6483, + "step": 4517 + }, + { + "epoch": 2.4527687296416936, + "grad_norm": 8.407197016108878, + "learning_rate": 1.0792216273042099e-05, + "loss": 0.4895, + "step": 4518 + }, + { + "epoch": 2.453311617806732, + "grad_norm": 9.426708870733849, + "learning_rate": 1.0788710438032131e-05, + "loss": 1.0719, + "step": 4519 + }, + { + "epoch": 2.45385450597177, + "grad_norm": 13.638918388289357, + "learning_rate": 1.0785204505473197e-05, + "loss": 0.8828, + "step": 4520 + }, + { + "epoch": 2.454397394136808, + "grad_norm": 12.478693804182608, + "learning_rate": 1.0781698475798917e-05, + "loss": 0.6527, + "step": 4521 + }, + { + "epoch": 2.4549402823018456, + "grad_norm": 10.20424250017949, + "learning_rate": 1.0778192349442923e-05, + "loss": 0.9303, + "step": 4522 + }, + { + "epoch": 2.455483170466884, + "grad_norm": 12.175885097025716, + "learning_rate": 1.0774686126838863e-05, + "loss": 0.6391, + "step": 4523 + }, + { + "epoch": 2.456026058631922, + "grad_norm": 10.568732307533892, + "learning_rate": 1.0771179808420385e-05, + "loss": 0.5524, + "step": 4524 + }, + { + "epoch": 2.45656894679696, + "grad_norm": 7.314447097275954, + "learning_rate": 1.0767673394621159e-05, + "loss": 0.4976, + "step": 4525 + }, + { + "epoch": 2.4571118349619976, + "grad_norm": 11.397811881243696, + "learning_rate": 1.0764166885874862e-05, + "loss": 0.4918, + "step": 4526 + }, + { + "epoch": 2.457654723127036, + "grad_norm": 9.691957338181856, + "learning_rate": 1.0760660282615191e-05, + "loss": 0.6725, + "step": 4527 + }, + { + "epoch": 2.458197611292074, + "grad_norm": 7.510379855172465, + "learning_rate": 1.0757153585275844e-05, + "loss": 0.4684, + "step": 4528 + }, + { + "epoch": 2.458740499457112, + "grad_norm": 7.751570484202345, + "learning_rate": 1.0753646794290534e-05, + "loss": 0.5495, + "step": 4529 + }, + { + "epoch": 2.4592833876221496, + "grad_norm": 7.390182612219503, + "learning_rate": 1.0750139910092987e-05, + "loss": 0.4538, + "step": 4530 + }, + { + "epoch": 2.459826275787188, + "grad_norm": 10.180132351193016, + "learning_rate": 1.074663293311694e-05, + "loss": 0.5579, + "step": 4531 + }, + { + "epoch": 2.460369163952226, + "grad_norm": 8.909847677295557, + "learning_rate": 1.074312586379614e-05, + "loss": 0.5023, + "step": 4532 + }, + { + "epoch": 2.460912052117264, + "grad_norm": 10.091297535262147, + "learning_rate": 1.073961870256435e-05, + "loss": 0.7068, + "step": 4533 + }, + { + "epoch": 2.4614549402823016, + "grad_norm": 11.778234555569338, + "learning_rate": 1.0736111449855341e-05, + "loss": 0.7687, + "step": 4534 + }, + { + "epoch": 2.46199782844734, + "grad_norm": 11.871444812620142, + "learning_rate": 1.0732604106102895e-05, + "loss": 0.8333, + "step": 4535 + }, + { + "epoch": 2.462540716612378, + "grad_norm": 6.164790402663993, + "learning_rate": 1.0729096671740806e-05, + "loss": 0.3601, + "step": 4536 + }, + { + "epoch": 2.463083604777416, + "grad_norm": 9.524060445013603, + "learning_rate": 1.0725589147202877e-05, + "loss": 0.6859, + "step": 4537 + }, + { + "epoch": 2.4636264929424536, + "grad_norm": 11.266171306725775, + "learning_rate": 1.0722081532922925e-05, + "loss": 0.7459, + "step": 4538 + }, + { + "epoch": 2.464169381107492, + "grad_norm": 8.945835373858307, + "learning_rate": 1.0718573829334782e-05, + "loss": 0.3787, + "step": 4539 + }, + { + "epoch": 2.46471226927253, + "grad_norm": 6.838806611638338, + "learning_rate": 1.071506603687228e-05, + "loss": 0.4995, + "step": 4540 + }, + { + "epoch": 2.465255157437568, + "grad_norm": 11.246624755560871, + "learning_rate": 1.071155815596927e-05, + "loss": 0.8699, + "step": 4541 + }, + { + "epoch": 2.4657980456026056, + "grad_norm": 7.324342918236285, + "learning_rate": 1.0708050187059615e-05, + "loss": 0.4778, + "step": 4542 + }, + { + "epoch": 2.466340933767644, + "grad_norm": 10.063475303955434, + "learning_rate": 1.0704542130577185e-05, + "loss": 0.5774, + "step": 4543 + }, + { + "epoch": 2.466883821932682, + "grad_norm": 12.644825932835714, + "learning_rate": 1.070103398695586e-05, + "loss": 0.8158, + "step": 4544 + }, + { + "epoch": 2.46742671009772, + "grad_norm": 10.332286780133, + "learning_rate": 1.0697525756629531e-05, + "loss": 0.6518, + "step": 4545 + }, + { + "epoch": 2.4679695982627576, + "grad_norm": 9.98899797230979, + "learning_rate": 1.069401744003211e-05, + "loss": 0.6167, + "step": 4546 + }, + { + "epoch": 2.468512486427796, + "grad_norm": 13.445015438013797, + "learning_rate": 1.0690509037597502e-05, + "loss": 0.8229, + "step": 4547 + }, + { + "epoch": 2.469055374592834, + "grad_norm": 13.637667335089724, + "learning_rate": 1.0687000549759641e-05, + "loss": 0.7186, + "step": 4548 + }, + { + "epoch": 2.469598262757872, + "grad_norm": 10.021629994296763, + "learning_rate": 1.0683491976952454e-05, + "loss": 0.6209, + "step": 4549 + }, + { + "epoch": 2.4701411509229096, + "grad_norm": 10.510874896443969, + "learning_rate": 1.067998331960989e-05, + "loss": 0.9009, + "step": 4550 + }, + { + "epoch": 2.470684039087948, + "grad_norm": 9.53841539843291, + "learning_rate": 1.0676474578165903e-05, + "loss": 0.7187, + "step": 4551 + }, + { + "epoch": 2.471226927252986, + "grad_norm": 11.693787891306421, + "learning_rate": 1.0672965753054466e-05, + "loss": 0.5568, + "step": 4552 + }, + { + "epoch": 2.471769815418024, + "grad_norm": 11.523811853237843, + "learning_rate": 1.0669456844709549e-05, + "loss": 0.8465, + "step": 4553 + }, + { + "epoch": 2.4723127035830617, + "grad_norm": 9.165552336797202, + "learning_rate": 1.0665947853565144e-05, + "loss": 0.7148, + "step": 4554 + }, + { + "epoch": 2.4728555917481, + "grad_norm": 10.390907422337587, + "learning_rate": 1.0662438780055247e-05, + "loss": 0.5271, + "step": 4555 + }, + { + "epoch": 2.473398479913138, + "grad_norm": 10.514660404488053, + "learning_rate": 1.0658929624613866e-05, + "loss": 0.7972, + "step": 4556 + }, + { + "epoch": 2.473941368078176, + "grad_norm": 8.804604897853023, + "learning_rate": 1.065542038767502e-05, + "loss": 0.4391, + "step": 4557 + }, + { + "epoch": 2.4744842562432137, + "grad_norm": 12.58202482779834, + "learning_rate": 1.0651911069672733e-05, + "loss": 0.601, + "step": 4558 + }, + { + "epoch": 2.475027144408252, + "grad_norm": 8.510520469826693, + "learning_rate": 1.0648401671041048e-05, + "loss": 0.6638, + "step": 4559 + }, + { + "epoch": 2.47557003257329, + "grad_norm": 10.132731434552191, + "learning_rate": 1.0644892192214012e-05, + "loss": 0.547, + "step": 4560 + }, + { + "epoch": 2.476112920738328, + "grad_norm": 9.486912806484277, + "learning_rate": 1.0641382633625678e-05, + "loss": 0.6903, + "step": 4561 + }, + { + "epoch": 2.4766558089033657, + "grad_norm": 8.656529900389968, + "learning_rate": 1.0637872995710122e-05, + "loss": 0.6445, + "step": 4562 + }, + { + "epoch": 2.477198697068404, + "grad_norm": 8.377471382601675, + "learning_rate": 1.0634363278901413e-05, + "loss": 0.475, + "step": 4563 + }, + { + "epoch": 2.477741585233442, + "grad_norm": 11.941789514685023, + "learning_rate": 1.0630853483633643e-05, + "loss": 0.8705, + "step": 4564 + }, + { + "epoch": 2.47828447339848, + "grad_norm": 8.551632836218364, + "learning_rate": 1.0627343610340912e-05, + "loss": 0.3577, + "step": 4565 + }, + { + "epoch": 2.4788273615635177, + "grad_norm": 10.220220103336683, + "learning_rate": 1.0623833659457322e-05, + "loss": 0.7002, + "step": 4566 + }, + { + "epoch": 2.479370249728556, + "grad_norm": 10.783849224421678, + "learning_rate": 1.0620323631416994e-05, + "loss": 0.6969, + "step": 4567 + }, + { + "epoch": 2.479913137893594, + "grad_norm": 9.018283325771, + "learning_rate": 1.0616813526654047e-05, + "loss": 0.6844, + "step": 4568 + }, + { + "epoch": 2.480456026058632, + "grad_norm": 9.570551013210757, + "learning_rate": 1.061330334560262e-05, + "loss": 0.5063, + "step": 4569 + }, + { + "epoch": 2.4809989142236697, + "grad_norm": 10.106289919073877, + "learning_rate": 1.0609793088696856e-05, + "loss": 0.6131, + "step": 4570 + }, + { + "epoch": 2.481541802388708, + "grad_norm": 10.842801269455231, + "learning_rate": 1.060628275637091e-05, + "loss": 0.6637, + "step": 4571 + }, + { + "epoch": 2.482084690553746, + "grad_norm": 11.953941591792438, + "learning_rate": 1.0602772349058952e-05, + "loss": 0.7631, + "step": 4572 + }, + { + "epoch": 2.482627578718784, + "grad_norm": 8.194683405468492, + "learning_rate": 1.0599261867195146e-05, + "loss": 0.5936, + "step": 4573 + }, + { + "epoch": 2.4831704668838217, + "grad_norm": 10.216659522745752, + "learning_rate": 1.0595751311213678e-05, + "loss": 0.7017, + "step": 4574 + }, + { + "epoch": 2.48371335504886, + "grad_norm": 11.0782985653065, + "learning_rate": 1.0592240681548738e-05, + "loss": 0.7285, + "step": 4575 + }, + { + "epoch": 2.484256243213898, + "grad_norm": 7.740854250685081, + "learning_rate": 1.0588729978634527e-05, + "loss": 0.4544, + "step": 4576 + }, + { + "epoch": 2.484799131378936, + "grad_norm": 6.9863990778515745, + "learning_rate": 1.0585219202905255e-05, + "loss": 0.2831, + "step": 4577 + }, + { + "epoch": 2.4853420195439737, + "grad_norm": 9.687327462597597, + "learning_rate": 1.0581708354795136e-05, + "loss": 0.5028, + "step": 4578 + }, + { + "epoch": 2.485884907709012, + "grad_norm": 8.009180785791076, + "learning_rate": 1.0578197434738405e-05, + "loss": 0.342, + "step": 4579 + }, + { + "epoch": 2.48642779587405, + "grad_norm": 8.101662777133065, + "learning_rate": 1.057468644316929e-05, + "loss": 0.4146, + "step": 4580 + }, + { + "epoch": 2.486970684039088, + "grad_norm": 9.370936142076715, + "learning_rate": 1.0571175380522041e-05, + "loss": 0.6006, + "step": 4581 + }, + { + "epoch": 2.4875135722041257, + "grad_norm": 11.164770907813724, + "learning_rate": 1.0567664247230911e-05, + "loss": 0.6787, + "step": 4582 + }, + { + "epoch": 2.488056460369164, + "grad_norm": 12.573930877103525, + "learning_rate": 1.056415304373016e-05, + "loss": 0.8051, + "step": 4583 + }, + { + "epoch": 2.488599348534202, + "grad_norm": 8.395060833285095, + "learning_rate": 1.0560641770454063e-05, + "loss": 0.5596, + "step": 4584 + }, + { + "epoch": 2.48914223669924, + "grad_norm": 11.015478890741145, + "learning_rate": 1.0557130427836899e-05, + "loss": 0.5442, + "step": 4585 + }, + { + "epoch": 2.4896851248642777, + "grad_norm": 9.420432906299153, + "learning_rate": 1.0553619016312955e-05, + "loss": 0.4917, + "step": 4586 + }, + { + "epoch": 2.490228013029316, + "grad_norm": 11.026358328990858, + "learning_rate": 1.0550107536316532e-05, + "loss": 0.4291, + "step": 4587 + }, + { + "epoch": 2.490770901194354, + "grad_norm": 16.545112067555284, + "learning_rate": 1.0546595988281928e-05, + "loss": 0.9417, + "step": 4588 + }, + { + "epoch": 2.491313789359392, + "grad_norm": 7.252757680637398, + "learning_rate": 1.0543084372643462e-05, + "loss": 0.7908, + "step": 4589 + }, + { + "epoch": 2.4918566775244297, + "grad_norm": 7.598562688636616, + "learning_rate": 1.0539572689835455e-05, + "loss": 0.3229, + "step": 4590 + }, + { + "epoch": 2.492399565689468, + "grad_norm": 12.449529704914275, + "learning_rate": 1.0536060940292238e-05, + "loss": 0.7373, + "step": 4591 + }, + { + "epoch": 2.492942453854506, + "grad_norm": 11.706159828307792, + "learning_rate": 1.0532549124448151e-05, + "loss": 0.6525, + "step": 4592 + }, + { + "epoch": 2.493485342019544, + "grad_norm": 8.000677691767947, + "learning_rate": 1.0529037242737538e-05, + "loss": 0.3613, + "step": 4593 + }, + { + "epoch": 2.4940282301845818, + "grad_norm": 9.529463393752437, + "learning_rate": 1.0525525295594755e-05, + "loss": 0.5921, + "step": 4594 + }, + { + "epoch": 2.49457111834962, + "grad_norm": 11.71769497271457, + "learning_rate": 1.0522013283454169e-05, + "loss": 0.5507, + "step": 4595 + }, + { + "epoch": 2.495114006514658, + "grad_norm": 8.546908524296436, + "learning_rate": 1.0518501206750145e-05, + "loss": 0.4564, + "step": 4596 + }, + { + "epoch": 2.495656894679696, + "grad_norm": 14.968617219210655, + "learning_rate": 1.0514989065917068e-05, + "loss": 0.5593, + "step": 4597 + }, + { + "epoch": 2.4961997828447338, + "grad_norm": 13.0682164313941, + "learning_rate": 1.051147686138932e-05, + "loss": 0.8812, + "step": 4598 + }, + { + "epoch": 2.496742671009772, + "grad_norm": 15.01014358585534, + "learning_rate": 1.05079645936013e-05, + "loss": 1.0756, + "step": 4599 + }, + { + "epoch": 2.49728555917481, + "grad_norm": 7.598240677419264, + "learning_rate": 1.0504452262987407e-05, + "loss": 0.3936, + "step": 4600 + }, + { + "epoch": 2.497828447339848, + "grad_norm": 10.39263281879832, + "learning_rate": 1.0500939869982058e-05, + "loss": 0.4944, + "step": 4601 + }, + { + "epoch": 2.4983713355048858, + "grad_norm": 12.104880157441473, + "learning_rate": 1.0497427415019665e-05, + "loss": 0.7498, + "step": 4602 + }, + { + "epoch": 2.498914223669924, + "grad_norm": 9.18522035093318, + "learning_rate": 1.0493914898534656e-05, + "loss": 0.6921, + "step": 4603 + }, + { + "epoch": 2.499457111834962, + "grad_norm": 15.812637613525052, + "learning_rate": 1.0490402320961466e-05, + "loss": 0.937, + "step": 4604 + }, + { + "epoch": 2.5, + "grad_norm": 11.400921058050775, + "learning_rate": 1.0486889682734535e-05, + "loss": 0.6453, + "step": 4605 + }, + { + "epoch": 2.500542888165038, + "grad_norm": 8.270129268701748, + "learning_rate": 1.0483376984288314e-05, + "loss": 0.3921, + "step": 4606 + }, + { + "epoch": 2.501085776330076, + "grad_norm": 15.017103371329522, + "learning_rate": 1.0479864226057258e-05, + "loss": 1.0397, + "step": 4607 + }, + { + "epoch": 2.5016286644951142, + "grad_norm": 8.346121650934089, + "learning_rate": 1.047635140847583e-05, + "loss": 0.3872, + "step": 4608 + }, + { + "epoch": 2.502171552660152, + "grad_norm": 14.17791420098929, + "learning_rate": 1.0472838531978495e-05, + "loss": 0.6932, + "step": 4609 + }, + { + "epoch": 2.50271444082519, + "grad_norm": 12.940166215438651, + "learning_rate": 1.046932559699974e-05, + "loss": 0.7045, + "step": 4610 + }, + { + "epoch": 2.503257328990228, + "grad_norm": 8.154813733233013, + "learning_rate": 1.0465812603974048e-05, + "loss": 0.4225, + "step": 4611 + }, + { + "epoch": 2.5038002171552662, + "grad_norm": 13.356750387494248, + "learning_rate": 1.0462299553335911e-05, + "loss": 0.7951, + "step": 4612 + }, + { + "epoch": 2.504343105320304, + "grad_norm": 8.683404141286427, + "learning_rate": 1.0458786445519827e-05, + "loss": 0.5871, + "step": 4613 + }, + { + "epoch": 2.504885993485342, + "grad_norm": 5.750758875120564, + "learning_rate": 1.0455273280960306e-05, + "loss": 0.3815, + "step": 4614 + }, + { + "epoch": 2.50542888165038, + "grad_norm": 9.328766968136149, + "learning_rate": 1.0451760060091857e-05, + "loss": 0.6451, + "step": 4615 + }, + { + "epoch": 2.5059717698154182, + "grad_norm": 9.704349389930364, + "learning_rate": 1.0448246783349002e-05, + "loss": 0.5136, + "step": 4616 + }, + { + "epoch": 2.506514657980456, + "grad_norm": 10.438938086465326, + "learning_rate": 1.0444733451166277e-05, + "loss": 0.6278, + "step": 4617 + }, + { + "epoch": 2.507057546145494, + "grad_norm": 11.019760442005925, + "learning_rate": 1.0441220063978208e-05, + "loss": 0.7477, + "step": 4618 + }, + { + "epoch": 2.507600434310532, + "grad_norm": 11.738067010881544, + "learning_rate": 1.0437706622219337e-05, + "loss": 0.5768, + "step": 4619 + }, + { + "epoch": 2.5081433224755703, + "grad_norm": 9.601242936349621, + "learning_rate": 1.0434193126324215e-05, + "loss": 0.4236, + "step": 4620 + }, + { + "epoch": 2.508686210640608, + "grad_norm": 8.041097216565952, + "learning_rate": 1.0430679576727392e-05, + "loss": 0.494, + "step": 4621 + }, + { + "epoch": 2.509229098805646, + "grad_norm": 11.144286090694623, + "learning_rate": 1.0427165973863434e-05, + "loss": 0.7882, + "step": 4622 + }, + { + "epoch": 2.509771986970684, + "grad_norm": 14.369854225910128, + "learning_rate": 1.0423652318166905e-05, + "loss": 0.7112, + "step": 4623 + }, + { + "epoch": 2.5103148751357223, + "grad_norm": 9.688103222900846, + "learning_rate": 1.0420138610072386e-05, + "loss": 0.4564, + "step": 4624 + }, + { + "epoch": 2.51085776330076, + "grad_norm": 8.2914031160658, + "learning_rate": 1.0416624850014454e-05, + "loss": 0.4106, + "step": 4625 + }, + { + "epoch": 2.511400651465798, + "grad_norm": 13.145683230031937, + "learning_rate": 1.0413111038427696e-05, + "loss": 0.905, + "step": 4626 + }, + { + "epoch": 2.511943539630836, + "grad_norm": 10.544391990984453, + "learning_rate": 1.0409597175746709e-05, + "loss": 0.5791, + "step": 4627 + }, + { + "epoch": 2.5124864277958743, + "grad_norm": 8.830789770145405, + "learning_rate": 1.0406083262406084e-05, + "loss": 0.4806, + "step": 4628 + }, + { + "epoch": 2.513029315960912, + "grad_norm": 9.186020814962909, + "learning_rate": 1.0402569298840437e-05, + "loss": 0.7227, + "step": 4629 + }, + { + "epoch": 2.51357220412595, + "grad_norm": 8.395511535015437, + "learning_rate": 1.0399055285484378e-05, + "loss": 0.5372, + "step": 4630 + }, + { + "epoch": 2.514115092290988, + "grad_norm": 9.969741133621088, + "learning_rate": 1.0395541222772524e-05, + "loss": 0.5151, + "step": 4631 + }, + { + "epoch": 2.5146579804560263, + "grad_norm": 11.64674593889149, + "learning_rate": 1.0392027111139505e-05, + "loss": 0.6159, + "step": 4632 + }, + { + "epoch": 2.515200868621064, + "grad_norm": 12.194306814356107, + "learning_rate": 1.0388512951019945e-05, + "loss": 0.6977, + "step": 4633 + }, + { + "epoch": 2.515743756786102, + "grad_norm": 9.063328798639105, + "learning_rate": 1.0384998742848485e-05, + "loss": 0.6518, + "step": 4634 + }, + { + "epoch": 2.51628664495114, + "grad_norm": 8.709419510130173, + "learning_rate": 1.0381484487059764e-05, + "loss": 0.5571, + "step": 4635 + }, + { + "epoch": 2.5168295331161783, + "grad_norm": 10.875178380627338, + "learning_rate": 1.0377970184088438e-05, + "loss": 0.8384, + "step": 4636 + }, + { + "epoch": 2.517372421281216, + "grad_norm": 8.42478873312803, + "learning_rate": 1.0374455834369155e-05, + "loss": 0.6209, + "step": 4637 + }, + { + "epoch": 2.517915309446254, + "grad_norm": 13.875461781038107, + "learning_rate": 1.0370941438336577e-05, + "loss": 0.8787, + "step": 4638 + }, + { + "epoch": 2.518458197611292, + "grad_norm": 11.372446219871874, + "learning_rate": 1.036742699642537e-05, + "loss": 0.569, + "step": 4639 + }, + { + "epoch": 2.5190010857763303, + "grad_norm": 10.725098253853016, + "learning_rate": 1.036391250907021e-05, + "loss": 0.8431, + "step": 4640 + }, + { + "epoch": 2.519543973941368, + "grad_norm": 8.268642654673096, + "learning_rate": 1.0360397976705767e-05, + "loss": 0.5833, + "step": 4641 + }, + { + "epoch": 2.520086862106406, + "grad_norm": 10.560802722537995, + "learning_rate": 1.0356883399766731e-05, + "loss": 0.7495, + "step": 4642 + }, + { + "epoch": 2.520629750271444, + "grad_norm": 11.789754393456334, + "learning_rate": 1.0353368778687784e-05, + "loss": 0.644, + "step": 4643 + }, + { + "epoch": 2.5211726384364823, + "grad_norm": 10.236138559776574, + "learning_rate": 1.0349854113903625e-05, + "loss": 0.6544, + "step": 4644 + }, + { + "epoch": 2.52171552660152, + "grad_norm": 8.552964039446971, + "learning_rate": 1.0346339405848953e-05, + "loss": 0.5093, + "step": 4645 + }, + { + "epoch": 2.522258414766558, + "grad_norm": 8.301860195209034, + "learning_rate": 1.0342824654958471e-05, + "loss": 0.6285, + "step": 4646 + }, + { + "epoch": 2.522801302931596, + "grad_norm": 9.238558119037732, + "learning_rate": 1.0339309861666889e-05, + "loss": 0.6519, + "step": 4647 + }, + { + "epoch": 2.5233441910966343, + "grad_norm": 9.840812219132841, + "learning_rate": 1.0335795026408922e-05, + "loss": 0.5874, + "step": 4648 + }, + { + "epoch": 2.523887079261672, + "grad_norm": 8.108461441633077, + "learning_rate": 1.0332280149619291e-05, + "loss": 0.5793, + "step": 4649 + }, + { + "epoch": 2.52442996742671, + "grad_norm": 9.446694020377457, + "learning_rate": 1.032876523173272e-05, + "loss": 0.607, + "step": 4650 + }, + { + "epoch": 2.524972855591748, + "grad_norm": 11.88681485670853, + "learning_rate": 1.0325250273183944e-05, + "loss": 1.0745, + "step": 4651 + }, + { + "epoch": 2.5255157437567863, + "grad_norm": 8.347193432264572, + "learning_rate": 1.0321735274407694e-05, + "loss": 0.8431, + "step": 4652 + }, + { + "epoch": 2.526058631921824, + "grad_norm": 9.639244837832804, + "learning_rate": 1.0318220235838715e-05, + "loss": 0.7827, + "step": 4653 + }, + { + "epoch": 2.526601520086862, + "grad_norm": 12.239750504764533, + "learning_rate": 1.0314705157911744e-05, + "loss": 0.6735, + "step": 4654 + }, + { + "epoch": 2.5271444082519, + "grad_norm": 15.826940901953726, + "learning_rate": 1.0311190041061543e-05, + "loss": 0.9182, + "step": 4655 + }, + { + "epoch": 2.5276872964169383, + "grad_norm": 11.015508331511823, + "learning_rate": 1.0307674885722863e-05, + "loss": 0.7834, + "step": 4656 + }, + { + "epoch": 2.528230184581976, + "grad_norm": 11.163753697057714, + "learning_rate": 1.030415969233046e-05, + "loss": 0.6613, + "step": 4657 + }, + { + "epoch": 2.528773072747014, + "grad_norm": 8.747908439433298, + "learning_rate": 1.03006444613191e-05, + "loss": 0.6974, + "step": 4658 + }, + { + "epoch": 2.529315960912052, + "grad_norm": 8.370456762868022, + "learning_rate": 1.0297129193123555e-05, + "loss": 0.5343, + "step": 4659 + }, + { + "epoch": 2.5298588490770904, + "grad_norm": 13.46061984772866, + "learning_rate": 1.0293613888178597e-05, + "loss": 1.3054, + "step": 4660 + }, + { + "epoch": 2.530401737242128, + "grad_norm": 10.948007938187232, + "learning_rate": 1.0290098546919007e-05, + "loss": 0.7028, + "step": 4661 + }, + { + "epoch": 2.530944625407166, + "grad_norm": 8.833095994203507, + "learning_rate": 1.0286583169779568e-05, + "loss": 0.7663, + "step": 4662 + }, + { + "epoch": 2.531487513572204, + "grad_norm": 10.283493007620729, + "learning_rate": 1.0283067757195062e-05, + "loss": 0.7804, + "step": 4663 + }, + { + "epoch": 2.5320304017372424, + "grad_norm": 13.55508996688301, + "learning_rate": 1.0279552309600288e-05, + "loss": 0.656, + "step": 4664 + }, + { + "epoch": 2.53257328990228, + "grad_norm": 8.738397704085456, + "learning_rate": 1.0276036827430036e-05, + "loss": 0.4902, + "step": 4665 + }, + { + "epoch": 2.533116178067318, + "grad_norm": 10.260011622402263, + "learning_rate": 1.0272521311119113e-05, + "loss": 0.4798, + "step": 4666 + }, + { + "epoch": 2.533659066232356, + "grad_norm": 10.743076748133591, + "learning_rate": 1.0269005761102315e-05, + "loss": 0.9671, + "step": 4667 + }, + { + "epoch": 2.5342019543973944, + "grad_norm": 11.846945314251228, + "learning_rate": 1.0265490177814458e-05, + "loss": 0.7168, + "step": 4668 + }, + { + "epoch": 2.534744842562432, + "grad_norm": 9.129703791341635, + "learning_rate": 1.0261974561690356e-05, + "loss": 0.6008, + "step": 4669 + }, + { + "epoch": 2.53528773072747, + "grad_norm": 8.464074733631888, + "learning_rate": 1.025845891316482e-05, + "loss": 0.4327, + "step": 4670 + }, + { + "epoch": 2.535830618892508, + "grad_norm": 10.894552922917367, + "learning_rate": 1.0254943232672671e-05, + "loss": 0.5627, + "step": 4671 + }, + { + "epoch": 2.5363735070575464, + "grad_norm": 12.365454725760225, + "learning_rate": 1.025142752064874e-05, + "loss": 0.606, + "step": 4672 + }, + { + "epoch": 2.536916395222584, + "grad_norm": 10.066952200044973, + "learning_rate": 1.0247911777527852e-05, + "loss": 0.7503, + "step": 4673 + }, + { + "epoch": 2.537459283387622, + "grad_norm": 8.386860314534095, + "learning_rate": 1.0244396003744836e-05, + "loss": 0.7663, + "step": 4674 + }, + { + "epoch": 2.53800217155266, + "grad_norm": 9.309667750090998, + "learning_rate": 1.024088019973454e-05, + "loss": 0.5658, + "step": 4675 + }, + { + "epoch": 2.5385450597176984, + "grad_norm": 13.530278392237364, + "learning_rate": 1.0237364365931794e-05, + "loss": 0.799, + "step": 4676 + }, + { + "epoch": 2.539087947882736, + "grad_norm": 11.344034257453236, + "learning_rate": 1.0233848502771444e-05, + "loss": 0.5931, + "step": 4677 + }, + { + "epoch": 2.539630836047774, + "grad_norm": 9.26197757297852, + "learning_rate": 1.0230332610688338e-05, + "loss": 0.5419, + "step": 4678 + }, + { + "epoch": 2.540173724212812, + "grad_norm": 12.620486559012823, + "learning_rate": 1.022681669011733e-05, + "loss": 0.6746, + "step": 4679 + }, + { + "epoch": 2.5407166123778504, + "grad_norm": 13.640092752737832, + "learning_rate": 1.0223300741493268e-05, + "loss": 0.8806, + "step": 4680 + }, + { + "epoch": 2.541259500542888, + "grad_norm": 10.435820181571753, + "learning_rate": 1.0219784765251018e-05, + "loss": 0.6876, + "step": 4681 + }, + { + "epoch": 2.541802388707926, + "grad_norm": 11.624678093656978, + "learning_rate": 1.0216268761825436e-05, + "loss": 0.7377, + "step": 4682 + }, + { + "epoch": 2.542345276872964, + "grad_norm": 10.361224010713974, + "learning_rate": 1.021275273165139e-05, + "loss": 0.5911, + "step": 4683 + }, + { + "epoch": 2.5428881650380024, + "grad_norm": 6.869744257762391, + "learning_rate": 1.0209236675163747e-05, + "loss": 0.4204, + "step": 4684 + }, + { + "epoch": 2.54343105320304, + "grad_norm": 16.790885412103638, + "learning_rate": 1.020572059279738e-05, + "loss": 0.8837, + "step": 4685 + }, + { + "epoch": 2.543973941368078, + "grad_norm": 9.226978433564653, + "learning_rate": 1.0202204484987159e-05, + "loss": 0.4686, + "step": 4686 + }, + { + "epoch": 2.544516829533116, + "grad_norm": 7.391909229375343, + "learning_rate": 1.019868835216797e-05, + "loss": 0.6996, + "step": 4687 + }, + { + "epoch": 2.5450597176981544, + "grad_norm": 9.103089555285077, + "learning_rate": 1.0195172194774684e-05, + "loss": 0.7243, + "step": 4688 + }, + { + "epoch": 2.545602605863192, + "grad_norm": 7.27509790187114, + "learning_rate": 1.0191656013242192e-05, + "loss": 0.4612, + "step": 4689 + }, + { + "epoch": 2.54614549402823, + "grad_norm": 6.939999239823949, + "learning_rate": 1.0188139808005379e-05, + "loss": 0.3483, + "step": 4690 + }, + { + "epoch": 2.546688382193268, + "grad_norm": 9.036119044447904, + "learning_rate": 1.0184623579499133e-05, + "loss": 0.5822, + "step": 4691 + }, + { + "epoch": 2.5472312703583064, + "grad_norm": 15.325233133671158, + "learning_rate": 1.018110732815835e-05, + "loss": 0.9375, + "step": 4692 + }, + { + "epoch": 2.547774158523344, + "grad_norm": 13.335480222530155, + "learning_rate": 1.017759105441792e-05, + "loss": 0.8165, + "step": 4693 + }, + { + "epoch": 2.548317046688382, + "grad_norm": 10.097959894991517, + "learning_rate": 1.0174074758712751e-05, + "loss": 0.7773, + "step": 4694 + }, + { + "epoch": 2.54885993485342, + "grad_norm": 9.7484700487623, + "learning_rate": 1.0170558441477738e-05, + "loss": 0.7805, + "step": 4695 + }, + { + "epoch": 2.5494028230184584, + "grad_norm": 9.687849482731272, + "learning_rate": 1.0167042103147784e-05, + "loss": 0.6675, + "step": 4696 + }, + { + "epoch": 2.549945711183496, + "grad_norm": 12.578144763595915, + "learning_rate": 1.0163525744157796e-05, + "loss": 0.7757, + "step": 4697 + }, + { + "epoch": 2.550488599348534, + "grad_norm": 9.03070780679602, + "learning_rate": 1.0160009364942684e-05, + "loss": 0.5321, + "step": 4698 + }, + { + "epoch": 2.5510314875135722, + "grad_norm": 10.632448946778698, + "learning_rate": 1.0156492965937357e-05, + "loss": 0.5589, + "step": 4699 + }, + { + "epoch": 2.5515743756786105, + "grad_norm": 10.053170332804227, + "learning_rate": 1.015297654757673e-05, + "loss": 0.4974, + "step": 4700 + }, + { + "epoch": 2.5521172638436482, + "grad_norm": 9.833516973903698, + "learning_rate": 1.0149460110295722e-05, + "loss": 0.6022, + "step": 4701 + }, + { + "epoch": 2.552660152008686, + "grad_norm": 16.704821665027783, + "learning_rate": 1.0145943654529248e-05, + "loss": 0.776, + "step": 4702 + }, + { + "epoch": 2.5532030401737242, + "grad_norm": 12.589948540180313, + "learning_rate": 1.0142427180712231e-05, + "loss": 0.8549, + "step": 4703 + }, + { + "epoch": 2.5537459283387625, + "grad_norm": 10.414828597120934, + "learning_rate": 1.0138910689279596e-05, + "loss": 0.7209, + "step": 4704 + }, + { + "epoch": 2.5542888165038002, + "grad_norm": 10.158714848928406, + "learning_rate": 1.0135394180666261e-05, + "loss": 0.6835, + "step": 4705 + }, + { + "epoch": 2.554831704668838, + "grad_norm": 8.255885711809917, + "learning_rate": 1.013187765530716e-05, + "loss": 0.4542, + "step": 4706 + }, + { + "epoch": 2.5553745928338762, + "grad_norm": 10.386471672795576, + "learning_rate": 1.012836111363722e-05, + "loss": 0.9478, + "step": 4707 + }, + { + "epoch": 2.5559174809989145, + "grad_norm": 11.828466617801858, + "learning_rate": 1.0124844556091374e-05, + "loss": 0.7975, + "step": 4708 + }, + { + "epoch": 2.5564603691639523, + "grad_norm": 10.742130491540307, + "learning_rate": 1.0121327983104555e-05, + "loss": 0.9712, + "step": 4709 + }, + { + "epoch": 2.55700325732899, + "grad_norm": 13.314461670532042, + "learning_rate": 1.0117811395111695e-05, + "loss": 0.9351, + "step": 4710 + }, + { + "epoch": 2.5575461454940283, + "grad_norm": 10.15371491897671, + "learning_rate": 1.0114294792547733e-05, + "loss": 0.6056, + "step": 4711 + }, + { + "epoch": 2.5580890336590665, + "grad_norm": 8.317356068473407, + "learning_rate": 1.011077817584761e-05, + "loss": 0.6411, + "step": 4712 + }, + { + "epoch": 2.5586319218241043, + "grad_norm": 11.215419343504372, + "learning_rate": 1.0107261545446267e-05, + "loss": 0.5411, + "step": 4713 + }, + { + "epoch": 2.559174809989142, + "grad_norm": 10.343171151673559, + "learning_rate": 1.0103744901778645e-05, + "loss": 0.5395, + "step": 4714 + }, + { + "epoch": 2.5597176981541803, + "grad_norm": 14.089833480392985, + "learning_rate": 1.0100228245279688e-05, + "loss": 0.9855, + "step": 4715 + }, + { + "epoch": 2.5602605863192185, + "grad_norm": 14.836069456421422, + "learning_rate": 1.0096711576384344e-05, + "loss": 0.8219, + "step": 4716 + }, + { + "epoch": 2.5608034744842563, + "grad_norm": 16.933747543735457, + "learning_rate": 1.0093194895527556e-05, + "loss": 0.9341, + "step": 4717 + }, + { + "epoch": 2.561346362649294, + "grad_norm": 13.574017237717102, + "learning_rate": 1.008967820314427e-05, + "loss": 1.1551, + "step": 4718 + }, + { + "epoch": 2.5618892508143323, + "grad_norm": 8.877246747020687, + "learning_rate": 1.0086161499669447e-05, + "loss": 0.6341, + "step": 4719 + }, + { + "epoch": 2.5624321389793705, + "grad_norm": 9.391996392121968, + "learning_rate": 1.0082644785538031e-05, + "loss": 0.5285, + "step": 4720 + }, + { + "epoch": 2.5629750271444083, + "grad_norm": 10.876272607307573, + "learning_rate": 1.0079128061184977e-05, + "loss": 0.5678, + "step": 4721 + }, + { + "epoch": 2.563517915309446, + "grad_norm": 12.516236640792583, + "learning_rate": 1.0075611327045236e-05, + "loss": 0.8382, + "step": 4722 + }, + { + "epoch": 2.5640608034744843, + "grad_norm": 6.395711230996481, + "learning_rate": 1.0072094583553769e-05, + "loss": 0.3785, + "step": 4723 + }, + { + "epoch": 2.5646036916395225, + "grad_norm": 8.671727273867468, + "learning_rate": 1.0068577831145526e-05, + "loss": 0.5039, + "step": 4724 + }, + { + "epoch": 2.5651465798045603, + "grad_norm": 12.77911107984442, + "learning_rate": 1.0065061070255469e-05, + "loss": 0.7647, + "step": 4725 + }, + { + "epoch": 2.565689467969598, + "grad_norm": 9.108473588917152, + "learning_rate": 1.0061544301318556e-05, + "loss": 0.798, + "step": 4726 + }, + { + "epoch": 2.5662323561346363, + "grad_norm": 8.776410877174934, + "learning_rate": 1.0058027524769742e-05, + "loss": 0.6388, + "step": 4727 + }, + { + "epoch": 2.5667752442996745, + "grad_norm": 10.392346204038525, + "learning_rate": 1.0054510741043995e-05, + "loss": 0.8178, + "step": 4728 + }, + { + "epoch": 2.5673181324647123, + "grad_norm": 8.611264791773579, + "learning_rate": 1.005099395057627e-05, + "loss": 0.5022, + "step": 4729 + }, + { + "epoch": 2.56786102062975, + "grad_norm": 8.228466341722912, + "learning_rate": 1.0047477153801534e-05, + "loss": 0.6572, + "step": 4730 + }, + { + "epoch": 2.5684039087947883, + "grad_norm": 8.410912812963819, + "learning_rate": 1.0043960351154745e-05, + "loss": 0.4526, + "step": 4731 + }, + { + "epoch": 2.5689467969598265, + "grad_norm": 8.24917306612946, + "learning_rate": 1.0040443543070872e-05, + "loss": 0.5492, + "step": 4732 + }, + { + "epoch": 2.5694896851248643, + "grad_norm": 7.578877723490418, + "learning_rate": 1.0036926729984878e-05, + "loss": 0.5482, + "step": 4733 + }, + { + "epoch": 2.570032573289902, + "grad_norm": 8.464719447395414, + "learning_rate": 1.003340991233173e-05, + "loss": 0.5236, + "step": 4734 + }, + { + "epoch": 2.5705754614549403, + "grad_norm": 10.297009632330814, + "learning_rate": 1.0029893090546385e-05, + "loss": 0.6747, + "step": 4735 + }, + { + "epoch": 2.5711183496199785, + "grad_norm": 9.895183624345862, + "learning_rate": 1.0026376265063818e-05, + "loss": 0.7036, + "step": 4736 + }, + { + "epoch": 2.5716612377850163, + "grad_norm": 16.034084078355107, + "learning_rate": 1.0022859436318988e-05, + "loss": 0.8583, + "step": 4737 + }, + { + "epoch": 2.572204125950054, + "grad_norm": 10.881395091863821, + "learning_rate": 1.0019342604746871e-05, + "loss": 0.7214, + "step": 4738 + }, + { + "epoch": 2.5727470141150923, + "grad_norm": 9.051906164481618, + "learning_rate": 1.0015825770782428e-05, + "loss": 0.36, + "step": 4739 + }, + { + "epoch": 2.5732899022801305, + "grad_norm": 11.71081741433651, + "learning_rate": 1.001230893486063e-05, + "loss": 0.8467, + "step": 4740 + }, + { + "epoch": 2.5738327904451683, + "grad_norm": 8.4727788666991, + "learning_rate": 1.0008792097416442e-05, + "loss": 0.5312, + "step": 4741 + }, + { + "epoch": 2.574375678610206, + "grad_norm": 10.308791039753308, + "learning_rate": 1.0005275258884835e-05, + "loss": 0.7132, + "step": 4742 + }, + { + "epoch": 2.5749185667752443, + "grad_norm": 13.591326014197278, + "learning_rate": 1.0001758419700773e-05, + "loss": 0.7087, + "step": 4743 + }, + { + "epoch": 2.5754614549402826, + "grad_norm": 9.610601406060551, + "learning_rate": 9.998241580299229e-06, + "loss": 0.6708, + "step": 4744 + }, + { + "epoch": 2.5760043431053203, + "grad_norm": 10.04740706892737, + "learning_rate": 9.994724741115169e-06, + "loss": 0.6644, + "step": 4745 + }, + { + "epoch": 2.576547231270358, + "grad_norm": 7.760395046863199, + "learning_rate": 9.991207902583558e-06, + "loss": 0.4361, + "step": 4746 + }, + { + "epoch": 2.5770901194353963, + "grad_norm": 9.097495186854097, + "learning_rate": 9.987691065139373e-06, + "loss": 0.5613, + "step": 4747 + }, + { + "epoch": 2.5776330076004346, + "grad_norm": 9.584167359092984, + "learning_rate": 9.984174229217572e-06, + "loss": 0.6408, + "step": 4748 + }, + { + "epoch": 2.5781758957654723, + "grad_norm": 10.329246788617224, + "learning_rate": 9.980657395253132e-06, + "loss": 0.6776, + "step": 4749 + }, + { + "epoch": 2.57871878393051, + "grad_norm": 10.000264678013584, + "learning_rate": 9.977140563681015e-06, + "loss": 0.7489, + "step": 4750 + }, + { + "epoch": 2.5792616720955484, + "grad_norm": 6.80240569067885, + "learning_rate": 9.973623734936185e-06, + "loss": 0.4639, + "step": 4751 + }, + { + "epoch": 2.5798045602605866, + "grad_norm": 7.578175941165148, + "learning_rate": 9.97010690945362e-06, + "loss": 0.38, + "step": 4752 + }, + { + "epoch": 2.5803474484256244, + "grad_norm": 12.48681779678407, + "learning_rate": 9.966590087668274e-06, + "loss": 0.8046, + "step": 4753 + }, + { + "epoch": 2.580890336590662, + "grad_norm": 13.31142746407252, + "learning_rate": 9.963073270015126e-06, + "loss": 0.7671, + "step": 4754 + }, + { + "epoch": 2.5814332247557004, + "grad_norm": 8.987640426318515, + "learning_rate": 9.959556456929128e-06, + "loss": 0.4691, + "step": 4755 + }, + { + "epoch": 2.5819761129207386, + "grad_norm": 9.551368066054666, + "learning_rate": 9.956039648845257e-06, + "loss": 0.6207, + "step": 4756 + }, + { + "epoch": 2.5825190010857764, + "grad_norm": 17.060876787385865, + "learning_rate": 9.952522846198468e-06, + "loss": 0.9276, + "step": 4757 + }, + { + "epoch": 2.583061889250814, + "grad_norm": 9.760247190255836, + "learning_rate": 9.949006049423731e-06, + "loss": 0.539, + "step": 4758 + }, + { + "epoch": 2.5836047774158524, + "grad_norm": 14.929770896239804, + "learning_rate": 9.94548925895601e-06, + "loss": 1.0896, + "step": 4759 + }, + { + "epoch": 2.5841476655808906, + "grad_norm": 14.902014520576888, + "learning_rate": 9.94197247523026e-06, + "loss": 1.0207, + "step": 4760 + }, + { + "epoch": 2.5846905537459284, + "grad_norm": 9.630889650448959, + "learning_rate": 9.93845569868145e-06, + "loss": 0.6322, + "step": 4761 + }, + { + "epoch": 2.585233441910966, + "grad_norm": 6.534627620002312, + "learning_rate": 9.934938929744535e-06, + "loss": 0.2627, + "step": 4762 + }, + { + "epoch": 2.5857763300760044, + "grad_norm": 10.701147725996508, + "learning_rate": 9.931422168854476e-06, + "loss": 0.6789, + "step": 4763 + }, + { + "epoch": 2.5863192182410426, + "grad_norm": 9.539331065006966, + "learning_rate": 9.927905416446233e-06, + "loss": 0.7109, + "step": 4764 + }, + { + "epoch": 2.5868621064060804, + "grad_norm": 9.979058002114003, + "learning_rate": 9.924388672954766e-06, + "loss": 0.6497, + "step": 4765 + }, + { + "epoch": 2.587404994571118, + "grad_norm": 9.638452627150874, + "learning_rate": 9.920871938815024e-06, + "loss": 0.6695, + "step": 4766 + }, + { + "epoch": 2.5879478827361564, + "grad_norm": 12.326422327522844, + "learning_rate": 9.91735521446197e-06, + "loss": 0.6354, + "step": 4767 + }, + { + "epoch": 2.5884907709011946, + "grad_norm": 8.571541287165285, + "learning_rate": 9.913838500330553e-06, + "loss": 0.3939, + "step": 4768 + }, + { + "epoch": 2.5890336590662324, + "grad_norm": 9.709310867338898, + "learning_rate": 9.910321796855732e-06, + "loss": 0.5866, + "step": 4769 + }, + { + "epoch": 2.58957654723127, + "grad_norm": 8.48642736575235, + "learning_rate": 9.90680510447245e-06, + "loss": 0.8167, + "step": 4770 + }, + { + "epoch": 2.5901194353963084, + "grad_norm": 9.153780615106163, + "learning_rate": 9.90328842361566e-06, + "loss": 0.6551, + "step": 4771 + }, + { + "epoch": 2.5906623235613466, + "grad_norm": 9.331431330636146, + "learning_rate": 9.899771754720315e-06, + "loss": 0.598, + "step": 4772 + }, + { + "epoch": 2.5912052117263844, + "grad_norm": 10.71941498109364, + "learning_rate": 9.896255098221357e-06, + "loss": 0.6996, + "step": 4773 + }, + { + "epoch": 2.591748099891422, + "grad_norm": 11.415708336876241, + "learning_rate": 9.892738454553736e-06, + "loss": 0.8247, + "step": 4774 + }, + { + "epoch": 2.5922909880564604, + "grad_norm": 9.526809613609492, + "learning_rate": 9.889221824152391e-06, + "loss": 0.5088, + "step": 4775 + }, + { + "epoch": 2.5928338762214986, + "grad_norm": 8.911159845652058, + "learning_rate": 9.885705207452268e-06, + "loss": 0.3377, + "step": 4776 + }, + { + "epoch": 2.5933767643865364, + "grad_norm": 13.451504414556261, + "learning_rate": 9.882188604888307e-06, + "loss": 0.9247, + "step": 4777 + }, + { + "epoch": 2.593919652551574, + "grad_norm": 10.251509263282227, + "learning_rate": 9.87867201689545e-06, + "loss": 0.4384, + "step": 4778 + }, + { + "epoch": 2.5944625407166124, + "grad_norm": 9.955424126921221, + "learning_rate": 9.875155443908631e-06, + "loss": 0.6318, + "step": 4779 + }, + { + "epoch": 2.5950054288816506, + "grad_norm": 8.584868222198555, + "learning_rate": 9.871638886362782e-06, + "loss": 0.5148, + "step": 4780 + }, + { + "epoch": 2.5955483170466884, + "grad_norm": 16.858447088677025, + "learning_rate": 9.868122344692846e-06, + "loss": 1.2564, + "step": 4781 + }, + { + "epoch": 2.596091205211726, + "grad_norm": 9.7339544501354, + "learning_rate": 9.86460581933374e-06, + "loss": 0.7158, + "step": 4782 + }, + { + "epoch": 2.5966340933767644, + "grad_norm": 11.03551156903236, + "learning_rate": 9.861089310720409e-06, + "loss": 0.5136, + "step": 4783 + }, + { + "epoch": 2.5971769815418027, + "grad_norm": 11.173351784755749, + "learning_rate": 9.857572819287768e-06, + "loss": 0.6439, + "step": 4784 + }, + { + "epoch": 2.5977198697068404, + "grad_norm": 14.318584084756125, + "learning_rate": 9.854056345470754e-06, + "loss": 1.2827, + "step": 4785 + }, + { + "epoch": 2.598262757871878, + "grad_norm": 11.132635136137834, + "learning_rate": 9.850539889704278e-06, + "loss": 0.499, + "step": 4786 + }, + { + "epoch": 2.5988056460369164, + "grad_norm": 11.710993594625656, + "learning_rate": 9.847023452423272e-06, + "loss": 0.466, + "step": 4787 + }, + { + "epoch": 2.5993485342019547, + "grad_norm": 12.05431003500401, + "learning_rate": 9.843507034062646e-06, + "loss": 1.002, + "step": 4788 + }, + { + "epoch": 2.5998914223669924, + "grad_norm": 12.651344078635795, + "learning_rate": 9.83999063505732e-06, + "loss": 0.6906, + "step": 4789 + }, + { + "epoch": 2.6004343105320302, + "grad_norm": 11.592009871908612, + "learning_rate": 9.83647425584221e-06, + "loss": 0.908, + "step": 4790 + }, + { + "epoch": 2.6009771986970684, + "grad_norm": 11.164823794029719, + "learning_rate": 9.83295789685222e-06, + "loss": 1.1006, + "step": 4791 + }, + { + "epoch": 2.6015200868621067, + "grad_norm": 18.55453736068596, + "learning_rate": 9.829441558522267e-06, + "loss": 0.6774, + "step": 4792 + }, + { + "epoch": 2.6020629750271445, + "grad_norm": 9.004957503789628, + "learning_rate": 9.825925241287249e-06, + "loss": 0.612, + "step": 4793 + }, + { + "epoch": 2.6026058631921822, + "grad_norm": 15.331324393349687, + "learning_rate": 9.822408945582081e-06, + "loss": 0.588, + "step": 4794 + }, + { + "epoch": 2.6031487513572205, + "grad_norm": 11.30698173094918, + "learning_rate": 9.818892671841653e-06, + "loss": 0.8413, + "step": 4795 + }, + { + "epoch": 2.6036916395222587, + "grad_norm": 10.970237049386885, + "learning_rate": 9.815376420500869e-06, + "loss": 0.5687, + "step": 4796 + }, + { + "epoch": 2.6042345276872965, + "grad_norm": 12.062898313733237, + "learning_rate": 9.811860191994623e-06, + "loss": 0.88, + "step": 4797 + }, + { + "epoch": 2.6047774158523342, + "grad_norm": 13.760783812579094, + "learning_rate": 9.808343986757811e-06, + "loss": 0.7319, + "step": 4798 + }, + { + "epoch": 2.6053203040173725, + "grad_norm": 10.142158775926303, + "learning_rate": 9.80482780522532e-06, + "loss": 0.5404, + "step": 4799 + }, + { + "epoch": 2.6058631921824107, + "grad_norm": 14.235611160734313, + "learning_rate": 9.801311647832035e-06, + "loss": 0.6849, + "step": 4800 + }, + { + "epoch": 2.6064060803474485, + "grad_norm": 14.33484497964673, + "learning_rate": 9.797795515012843e-06, + "loss": 1.0677, + "step": 4801 + }, + { + "epoch": 2.6069489685124863, + "grad_norm": 12.238534475065178, + "learning_rate": 9.794279407202624e-06, + "loss": 0.8279, + "step": 4802 + }, + { + "epoch": 2.6074918566775245, + "grad_norm": 13.51950414574984, + "learning_rate": 9.790763324836256e-06, + "loss": 0.8372, + "step": 4803 + }, + { + "epoch": 2.6080347448425627, + "grad_norm": 9.360760584245645, + "learning_rate": 9.78724726834861e-06, + "loss": 0.666, + "step": 4804 + }, + { + "epoch": 2.6085776330076005, + "grad_norm": 10.793352407977682, + "learning_rate": 9.783731238174566e-06, + "loss": 0.8308, + "step": 4805 + }, + { + "epoch": 2.6091205211726383, + "grad_norm": 13.862491939748782, + "learning_rate": 9.780215234748984e-06, + "loss": 0.8018, + "step": 4806 + }, + { + "epoch": 2.6096634093376765, + "grad_norm": 8.019652931555422, + "learning_rate": 9.776699258506734e-06, + "loss": 0.3801, + "step": 4807 + }, + { + "epoch": 2.6102062975027147, + "grad_norm": 10.161487802712038, + "learning_rate": 9.773183309882675e-06, + "loss": 0.6723, + "step": 4808 + }, + { + "epoch": 2.6107491856677525, + "grad_norm": 9.64703000959591, + "learning_rate": 9.769667389311664e-06, + "loss": 0.6294, + "step": 4809 + }, + { + "epoch": 2.6112920738327903, + "grad_norm": 9.792927662736604, + "learning_rate": 9.766151497228561e-06, + "loss": 0.4596, + "step": 4810 + }, + { + "epoch": 2.6118349619978285, + "grad_norm": 10.336511067045754, + "learning_rate": 9.76263563406821e-06, + "loss": 0.6494, + "step": 4811 + }, + { + "epoch": 2.6123778501628667, + "grad_norm": 12.121823172890695, + "learning_rate": 9.759119800265464e-06, + "loss": 1.0394, + "step": 4812 + }, + { + "epoch": 2.6129207383279045, + "grad_norm": 10.891668605270631, + "learning_rate": 9.755603996255164e-06, + "loss": 0.6212, + "step": 4813 + }, + { + "epoch": 2.6134636264929423, + "grad_norm": 9.599793914114507, + "learning_rate": 9.752088222472151e-06, + "loss": 0.6182, + "step": 4814 + }, + { + "epoch": 2.6140065146579805, + "grad_norm": 10.448317247236231, + "learning_rate": 9.748572479351263e-06, + "loss": 0.6779, + "step": 4815 + }, + { + "epoch": 2.6145494028230187, + "grad_norm": 12.703205416876392, + "learning_rate": 9.74505676732733e-06, + "loss": 0.8093, + "step": 4816 + }, + { + "epoch": 2.6150922909880565, + "grad_norm": 10.14294888595903, + "learning_rate": 9.741541086835182e-06, + "loss": 0.6493, + "step": 4817 + }, + { + "epoch": 2.6156351791530943, + "grad_norm": 11.455318908063175, + "learning_rate": 9.738025438309649e-06, + "loss": 0.9375, + "step": 4818 + }, + { + "epoch": 2.6161780673181325, + "grad_norm": 8.90202557317791, + "learning_rate": 9.734509822185545e-06, + "loss": 0.5978, + "step": 4819 + }, + { + "epoch": 2.6167209554831707, + "grad_norm": 10.38484629007874, + "learning_rate": 9.730994238897687e-06, + "loss": 0.9772, + "step": 4820 + }, + { + "epoch": 2.6172638436482085, + "grad_norm": 10.650781833239714, + "learning_rate": 9.72747868888089e-06, + "loss": 0.7354, + "step": 4821 + }, + { + "epoch": 2.6178067318132463, + "grad_norm": 11.462972783418381, + "learning_rate": 9.723963172569964e-06, + "loss": 0.8748, + "step": 4822 + }, + { + "epoch": 2.6183496199782845, + "grad_norm": 7.4804704893307505, + "learning_rate": 9.720447690399716e-06, + "loss": 0.4853, + "step": 4823 + }, + { + "epoch": 2.6188925081433228, + "grad_norm": 8.032622743372611, + "learning_rate": 9.716932242804938e-06, + "loss": 0.6484, + "step": 4824 + }, + { + "epoch": 2.6194353963083605, + "grad_norm": 10.141962204854094, + "learning_rate": 9.713416830220436e-06, + "loss": 0.669, + "step": 4825 + }, + { + "epoch": 2.6199782844733983, + "grad_norm": 11.48618310533414, + "learning_rate": 9.709901453080993e-06, + "loss": 0.6279, + "step": 4826 + }, + { + "epoch": 2.6205211726384365, + "grad_norm": 9.789449860925128, + "learning_rate": 9.706386111821406e-06, + "loss": 0.6072, + "step": 4827 + }, + { + "epoch": 2.6210640608034748, + "grad_norm": 8.10222084311329, + "learning_rate": 9.702870806876448e-06, + "loss": 0.6476, + "step": 4828 + }, + { + "epoch": 2.6216069489685125, + "grad_norm": 10.30951278757575, + "learning_rate": 9.699355538680902e-06, + "loss": 0.4691, + "step": 4829 + }, + { + "epoch": 2.6221498371335503, + "grad_norm": 10.960835467770405, + "learning_rate": 9.695840307669547e-06, + "loss": 0.7962, + "step": 4830 + }, + { + "epoch": 2.6226927252985885, + "grad_norm": 6.691408537531251, + "learning_rate": 9.69232511427714e-06, + "loss": 0.3644, + "step": 4831 + }, + { + "epoch": 2.6232356134636268, + "grad_norm": 7.019904615334962, + "learning_rate": 9.68880995893846e-06, + "loss": 0.4594, + "step": 4832 + }, + { + "epoch": 2.6237785016286646, + "grad_norm": 11.780644240657415, + "learning_rate": 9.685294842088256e-06, + "loss": 1.0625, + "step": 4833 + }, + { + "epoch": 2.6243213897937023, + "grad_norm": 7.34603375318287, + "learning_rate": 9.681779764161289e-06, + "loss": 0.3502, + "step": 4834 + }, + { + "epoch": 2.6248642779587406, + "grad_norm": 9.787656886442962, + "learning_rate": 9.678264725592306e-06, + "loss": 0.5054, + "step": 4835 + }, + { + "epoch": 2.6254071661237783, + "grad_norm": 9.687484513116981, + "learning_rate": 9.674749726816058e-06, + "loss": 0.8678, + "step": 4836 + }, + { + "epoch": 2.6259500542888166, + "grad_norm": 7.620468702427845, + "learning_rate": 9.671234768267284e-06, + "loss": 0.4672, + "step": 4837 + }, + { + "epoch": 2.6264929424538543, + "grad_norm": 14.051025442281276, + "learning_rate": 9.667719850380712e-06, + "loss": 0.6462, + "step": 4838 + }, + { + "epoch": 2.6270358306188926, + "grad_norm": 11.579748119704858, + "learning_rate": 9.664204973591081e-06, + "loss": 0.8856, + "step": 4839 + }, + { + "epoch": 2.6275787187839303, + "grad_norm": 8.348102354642592, + "learning_rate": 9.660690138333114e-06, + "loss": 0.5623, + "step": 4840 + }, + { + "epoch": 2.6281216069489686, + "grad_norm": 9.351950984929493, + "learning_rate": 9.657175345041532e-06, + "loss": 0.5809, + "step": 4841 + }, + { + "epoch": 2.6286644951140063, + "grad_norm": 12.421448876984805, + "learning_rate": 9.653660594151047e-06, + "loss": 0.7831, + "step": 4842 + }, + { + "epoch": 2.6292073832790446, + "grad_norm": 11.981212578199576, + "learning_rate": 9.650145886096376e-06, + "loss": 0.8767, + "step": 4843 + }, + { + "epoch": 2.6297502714440824, + "grad_norm": 10.659578351567276, + "learning_rate": 9.646631221312216e-06, + "loss": 0.7313, + "step": 4844 + }, + { + "epoch": 2.6302931596091206, + "grad_norm": 12.480207661204222, + "learning_rate": 9.643116600233274e-06, + "loss": 0.7412, + "step": 4845 + }, + { + "epoch": 2.6308360477741584, + "grad_norm": 9.131072461148838, + "learning_rate": 9.639602023294233e-06, + "loss": 0.5518, + "step": 4846 + }, + { + "epoch": 2.6313789359391966, + "grad_norm": 7.489089256780018, + "learning_rate": 9.636087490929793e-06, + "loss": 0.4321, + "step": 4847 + }, + { + "epoch": 2.6319218241042344, + "grad_norm": 12.12610683915495, + "learning_rate": 9.632573003574634e-06, + "loss": 0.8265, + "step": 4848 + }, + { + "epoch": 2.6324647122692726, + "grad_norm": 7.965586818160769, + "learning_rate": 9.629058561663426e-06, + "loss": 0.373, + "step": 4849 + }, + { + "epoch": 2.6330076004343104, + "grad_norm": 9.348866579305472, + "learning_rate": 9.62554416563085e-06, + "loss": 0.4274, + "step": 4850 + }, + { + "epoch": 2.6335504885993486, + "grad_norm": 10.13050086693858, + "learning_rate": 9.622029815911566e-06, + "loss": 0.5276, + "step": 4851 + }, + { + "epoch": 2.6340933767643864, + "grad_norm": 10.19180911105006, + "learning_rate": 9.61851551294024e-06, + "loss": 0.5871, + "step": 4852 + }, + { + "epoch": 2.6346362649294246, + "grad_norm": 11.184621464328854, + "learning_rate": 9.61500125715152e-06, + "loss": 0.6047, + "step": 4853 + }, + { + "epoch": 2.6351791530944624, + "grad_norm": 14.484355041392742, + "learning_rate": 9.611487048980058e-06, + "loss": 0.8628, + "step": 4854 + }, + { + "epoch": 2.6357220412595006, + "grad_norm": 13.070986825901516, + "learning_rate": 9.607972888860497e-06, + "loss": 0.7793, + "step": 4855 + }, + { + "epoch": 2.6362649294245384, + "grad_norm": 9.640708751060075, + "learning_rate": 9.604458777227477e-06, + "loss": 0.6016, + "step": 4856 + }, + { + "epoch": 2.6368078175895766, + "grad_norm": 11.05523237140924, + "learning_rate": 9.600944714515627e-06, + "loss": 0.5999, + "step": 4857 + }, + { + "epoch": 2.6373507057546144, + "grad_norm": 13.240640169530616, + "learning_rate": 9.597430701159565e-06, + "loss": 0.6779, + "step": 4858 + }, + { + "epoch": 2.6378935939196526, + "grad_norm": 10.878893821343203, + "learning_rate": 9.593916737593919e-06, + "loss": 0.6119, + "step": 4859 + }, + { + "epoch": 2.6384364820846904, + "grad_norm": 10.820526634571284, + "learning_rate": 9.590402824253295e-06, + "loss": 0.839, + "step": 4860 + }, + { + "epoch": 2.6389793702497286, + "grad_norm": 12.938750622065529, + "learning_rate": 9.586888961572307e-06, + "loss": 0.517, + "step": 4861 + }, + { + "epoch": 2.6395222584147664, + "grad_norm": 9.725430374953143, + "learning_rate": 9.583375149985547e-06, + "loss": 0.7604, + "step": 4862 + }, + { + "epoch": 2.6400651465798046, + "grad_norm": 11.840312760690757, + "learning_rate": 9.579861389927615e-06, + "loss": 0.7614, + "step": 4863 + }, + { + "epoch": 2.6406080347448424, + "grad_norm": 11.047185391206618, + "learning_rate": 9.576347681833093e-06, + "loss": 0.6351, + "step": 4864 + }, + { + "epoch": 2.6411509229098806, + "grad_norm": 9.986349266550256, + "learning_rate": 9.572834026136571e-06, + "loss": 0.5549, + "step": 4865 + }, + { + "epoch": 2.6416938110749184, + "grad_norm": 14.455847384105025, + "learning_rate": 9.56932042327261e-06, + "loss": 0.8206, + "step": 4866 + }, + { + "epoch": 2.6422366992399566, + "grad_norm": 10.377988559117977, + "learning_rate": 9.56580687367579e-06, + "loss": 0.5275, + "step": 4867 + }, + { + "epoch": 2.6427795874049944, + "grad_norm": 10.71910613137521, + "learning_rate": 9.562293377780668e-06, + "loss": 0.6958, + "step": 4868 + }, + { + "epoch": 2.6433224755700326, + "grad_norm": 8.635819291033194, + "learning_rate": 9.558779936021795e-06, + "loss": 0.3946, + "step": 4869 + }, + { + "epoch": 2.6438653637350704, + "grad_norm": 16.60173101533996, + "learning_rate": 9.555266548833728e-06, + "loss": 1.066, + "step": 4870 + }, + { + "epoch": 2.6444082519001086, + "grad_norm": 14.726517961222527, + "learning_rate": 9.551753216650998e-06, + "loss": 1.1356, + "step": 4871 + }, + { + "epoch": 2.6449511400651464, + "grad_norm": 11.123184991281361, + "learning_rate": 9.548239939908146e-06, + "loss": 0.757, + "step": 4872 + }, + { + "epoch": 2.6454940282301846, + "grad_norm": 11.666285426731331, + "learning_rate": 9.544726719039699e-06, + "loss": 0.7109, + "step": 4873 + }, + { + "epoch": 2.6460369163952224, + "grad_norm": 9.657892010966115, + "learning_rate": 9.541213554480176e-06, + "loss": 0.6191, + "step": 4874 + }, + { + "epoch": 2.6465798045602607, + "grad_norm": 9.984832056257446, + "learning_rate": 9.53770044666409e-06, + "loss": 0.8649, + "step": 4875 + }, + { + "epoch": 2.6471226927252984, + "grad_norm": 9.143852173958573, + "learning_rate": 9.534187396025955e-06, + "loss": 0.5282, + "step": 4876 + }, + { + "epoch": 2.6476655808903367, + "grad_norm": 8.129441981336987, + "learning_rate": 9.530674403000266e-06, + "loss": 0.3783, + "step": 4877 + }, + { + "epoch": 2.6482084690553744, + "grad_norm": 10.147175407433437, + "learning_rate": 9.527161468021508e-06, + "loss": 0.6676, + "step": 4878 + }, + { + "epoch": 2.6487513572204127, + "grad_norm": 11.015226101188722, + "learning_rate": 9.523648591524176e-06, + "loss": 0.5218, + "step": 4879 + }, + { + "epoch": 2.6492942453854504, + "grad_norm": 9.374140703922862, + "learning_rate": 9.520135773942743e-06, + "loss": 0.6754, + "step": 4880 + }, + { + "epoch": 2.6498371335504887, + "grad_norm": 8.84122429163583, + "learning_rate": 9.516623015711688e-06, + "loss": 0.5236, + "step": 4881 + }, + { + "epoch": 2.6503800217155264, + "grad_norm": 10.659047956662723, + "learning_rate": 9.513110317265463e-06, + "loss": 0.4742, + "step": 4882 + }, + { + "epoch": 2.6509229098805647, + "grad_norm": 8.798273983585764, + "learning_rate": 9.509597679038536e-06, + "loss": 0.4949, + "step": 4883 + }, + { + "epoch": 2.6514657980456025, + "grad_norm": 14.827321458575321, + "learning_rate": 9.506085101465344e-06, + "loss": 0.967, + "step": 4884 + }, + { + "epoch": 2.6520086862106407, + "grad_norm": 13.212543627120823, + "learning_rate": 9.502572584980338e-06, + "loss": 0.904, + "step": 4885 + }, + { + "epoch": 2.6525515743756785, + "grad_norm": 12.151700192108333, + "learning_rate": 9.499060130017947e-06, + "loss": 0.6512, + "step": 4886 + }, + { + "epoch": 2.6530944625407167, + "grad_norm": 9.051653631541132, + "learning_rate": 9.495547737012594e-06, + "loss": 0.5264, + "step": 4887 + }, + { + "epoch": 2.6536373507057545, + "grad_norm": 11.286333388065833, + "learning_rate": 9.492035406398706e-06, + "loss": 0.5864, + "step": 4888 + }, + { + "epoch": 2.6541802388707927, + "grad_norm": 8.303712468523708, + "learning_rate": 9.488523138610684e-06, + "loss": 0.5887, + "step": 4889 + }, + { + "epoch": 2.6547231270358305, + "grad_norm": 8.762643235318839, + "learning_rate": 9.485010934082939e-06, + "loss": 0.5079, + "step": 4890 + }, + { + "epoch": 2.6552660152008687, + "grad_norm": 13.211351989906888, + "learning_rate": 9.481498793249859e-06, + "loss": 0.8983, + "step": 4891 + }, + { + "epoch": 2.6558089033659065, + "grad_norm": 11.087558391384889, + "learning_rate": 9.477986716545834e-06, + "loss": 1.1671, + "step": 4892 + }, + { + "epoch": 2.6563517915309447, + "grad_norm": 6.825801048730381, + "learning_rate": 9.474474704405245e-06, + "loss": 0.3664, + "step": 4893 + }, + { + "epoch": 2.6568946796959825, + "grad_norm": 6.583440546159338, + "learning_rate": 9.470962757262465e-06, + "loss": 0.5464, + "step": 4894 + }, + { + "epoch": 2.6574375678610207, + "grad_norm": 8.987234989797157, + "learning_rate": 9.46745087555185e-06, + "loss": 0.5954, + "step": 4895 + }, + { + "epoch": 2.6579804560260585, + "grad_norm": 10.699367568639158, + "learning_rate": 9.463939059707763e-06, + "loss": 0.632, + "step": 4896 + }, + { + "epoch": 2.6585233441910967, + "grad_norm": 10.246298042329498, + "learning_rate": 9.460427310164548e-06, + "loss": 0.6766, + "step": 4897 + }, + { + "epoch": 2.6590662323561345, + "grad_norm": 12.300177238674, + "learning_rate": 9.456915627356542e-06, + "loss": 0.7277, + "step": 4898 + }, + { + "epoch": 2.6596091205211727, + "grad_norm": 9.893644041111745, + "learning_rate": 9.453404011718077e-06, + "loss": 0.644, + "step": 4899 + }, + { + "epoch": 2.6601520086862105, + "grad_norm": 10.80727748895235, + "learning_rate": 9.449892463683471e-06, + "loss": 0.5844, + "step": 4900 + }, + { + "epoch": 2.6606948968512487, + "grad_norm": 8.350836469669698, + "learning_rate": 9.446380983687046e-06, + "loss": 0.4745, + "step": 4901 + }, + { + "epoch": 2.6612377850162865, + "grad_norm": 10.697471057638213, + "learning_rate": 9.442869572163101e-06, + "loss": 0.5749, + "step": 4902 + }, + { + "epoch": 2.6617806731813247, + "grad_norm": 7.582465634030485, + "learning_rate": 9.43935822954594e-06, + "loss": 0.4481, + "step": 4903 + }, + { + "epoch": 2.6623235613463625, + "grad_norm": 10.01801356389796, + "learning_rate": 9.435846956269841e-06, + "loss": 0.5817, + "step": 4904 + }, + { + "epoch": 2.6628664495114007, + "grad_norm": 9.383471785643394, + "learning_rate": 9.432335752769092e-06, + "loss": 0.526, + "step": 4905 + }, + { + "epoch": 2.6634093376764385, + "grad_norm": 9.555536887097055, + "learning_rate": 9.428824619477964e-06, + "loss": 0.5056, + "step": 4906 + }, + { + "epoch": 2.6639522258414767, + "grad_norm": 7.636054626405326, + "learning_rate": 9.425313556830713e-06, + "loss": 0.4383, + "step": 4907 + }, + { + "epoch": 2.6644951140065145, + "grad_norm": 9.842725681691716, + "learning_rate": 9.421802565261602e-06, + "loss": 0.548, + "step": 4908 + }, + { + "epoch": 2.6650380021715527, + "grad_norm": 12.891942531084513, + "learning_rate": 9.418291645204865e-06, + "loss": 0.6489, + "step": 4909 + }, + { + "epoch": 2.6655808903365905, + "grad_norm": 15.298821348423894, + "learning_rate": 9.41478079709475e-06, + "loss": 0.8286, + "step": 4910 + }, + { + "epoch": 2.6661237785016287, + "grad_norm": 10.80698538838323, + "learning_rate": 9.411270021365475e-06, + "loss": 0.7704, + "step": 4911 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 10.837498285149955, + "learning_rate": 9.407759318451264e-06, + "loss": 0.5384, + "step": 4912 + }, + { + "epoch": 2.6672095548317047, + "grad_norm": 11.666848847579812, + "learning_rate": 9.404248688786322e-06, + "loss": 0.6402, + "step": 4913 + }, + { + "epoch": 2.6677524429967425, + "grad_norm": 12.038932116915525, + "learning_rate": 9.400738132804856e-06, + "loss": 0.559, + "step": 4914 + }, + { + "epoch": 2.6682953311617807, + "grad_norm": 13.411348889811451, + "learning_rate": 9.397227650941048e-06, + "loss": 0.9587, + "step": 4915 + }, + { + "epoch": 2.6688382193268185, + "grad_norm": 8.79354274091956, + "learning_rate": 9.393717243629091e-06, + "loss": 0.5627, + "step": 4916 + }, + { + "epoch": 2.6693811074918568, + "grad_norm": 8.956388294197298, + "learning_rate": 9.390206911303148e-06, + "loss": 0.5694, + "step": 4917 + }, + { + "epoch": 2.6699239956568945, + "grad_norm": 11.296789676569125, + "learning_rate": 9.386696654397384e-06, + "loss": 0.8356, + "step": 4918 + }, + { + "epoch": 2.6704668838219328, + "grad_norm": 8.778967068719545, + "learning_rate": 9.38318647334596e-06, + "loss": 0.5283, + "step": 4919 + }, + { + "epoch": 2.6710097719869705, + "grad_norm": 10.041165955923448, + "learning_rate": 9.379676368583011e-06, + "loss": 0.7492, + "step": 4920 + }, + { + "epoch": 2.6715526601520088, + "grad_norm": 7.152774382953927, + "learning_rate": 9.376166340542682e-06, + "loss": 0.4563, + "step": 4921 + }, + { + "epoch": 2.6720955483170465, + "grad_norm": 10.189040304576451, + "learning_rate": 9.37265638965909e-06, + "loss": 0.5869, + "step": 4922 + }, + { + "epoch": 2.6726384364820848, + "grad_norm": 8.47728562416046, + "learning_rate": 9.369146516366359e-06, + "loss": 0.6168, + "step": 4923 + }, + { + "epoch": 2.6731813246471225, + "grad_norm": 9.47866332824828, + "learning_rate": 9.365636721098588e-06, + "loss": 0.5233, + "step": 4924 + }, + { + "epoch": 2.6737242128121608, + "grad_norm": 11.264475439058582, + "learning_rate": 9.362127004289883e-06, + "loss": 0.693, + "step": 4925 + }, + { + "epoch": 2.6742671009771986, + "grad_norm": 11.495205201206721, + "learning_rate": 9.358617366374327e-06, + "loss": 0.587, + "step": 4926 + }, + { + "epoch": 2.6748099891422368, + "grad_norm": 10.081706772004408, + "learning_rate": 9.355107807785993e-06, + "loss": 0.65, + "step": 4927 + }, + { + "epoch": 2.6753528773072746, + "grad_norm": 8.912371373305913, + "learning_rate": 9.351598328958957e-06, + "loss": 0.6475, + "step": 4928 + }, + { + "epoch": 2.675895765472313, + "grad_norm": 8.434539074965906, + "learning_rate": 9.348088930327269e-06, + "loss": 0.4321, + "step": 4929 + }, + { + "epoch": 2.6764386536373506, + "grad_norm": 10.356652945785703, + "learning_rate": 9.344579612324984e-06, + "loss": 0.4243, + "step": 4930 + }, + { + "epoch": 2.676981541802389, + "grad_norm": 10.975361657654028, + "learning_rate": 9.341070375386134e-06, + "loss": 0.6042, + "step": 4931 + }, + { + "epoch": 2.6775244299674266, + "grad_norm": 10.912795228418068, + "learning_rate": 9.337561219944756e-06, + "loss": 0.6655, + "step": 4932 + }, + { + "epoch": 2.678067318132465, + "grad_norm": 9.852947274867201, + "learning_rate": 9.334052146434857e-06, + "loss": 0.7696, + "step": 4933 + }, + { + "epoch": 2.6786102062975026, + "grad_norm": 7.323342130507832, + "learning_rate": 9.330543155290454e-06, + "loss": 0.3453, + "step": 4934 + }, + { + "epoch": 2.679153094462541, + "grad_norm": 7.335485317101141, + "learning_rate": 9.32703424694554e-06, + "loss": 0.4291, + "step": 4935 + }, + { + "epoch": 2.6796959826275786, + "grad_norm": 12.700387251668076, + "learning_rate": 9.3235254218341e-06, + "loss": 0.7156, + "step": 4936 + }, + { + "epoch": 2.680238870792617, + "grad_norm": 10.828431765432423, + "learning_rate": 9.320016680390115e-06, + "loss": 0.7293, + "step": 4937 + }, + { + "epoch": 2.6807817589576546, + "grad_norm": 9.224454671275938, + "learning_rate": 9.316508023047548e-06, + "loss": 0.4157, + "step": 4938 + }, + { + "epoch": 2.681324647122693, + "grad_norm": 13.095737517917204, + "learning_rate": 9.312999450240362e-06, + "loss": 0.7928, + "step": 4939 + }, + { + "epoch": 2.6818675352877306, + "grad_norm": 10.338830489426515, + "learning_rate": 9.309490962402498e-06, + "loss": 0.6826, + "step": 4940 + }, + { + "epoch": 2.682410423452769, + "grad_norm": 11.650643384237476, + "learning_rate": 9.305982559967894e-06, + "loss": 0.6709, + "step": 4941 + }, + { + "epoch": 2.6829533116178066, + "grad_norm": 9.153121673948778, + "learning_rate": 9.302474243370469e-06, + "loss": 0.4998, + "step": 4942 + }, + { + "epoch": 2.683496199782845, + "grad_norm": 11.760747525125, + "learning_rate": 9.298966013044144e-06, + "loss": 1.0389, + "step": 4943 + }, + { + "epoch": 2.6840390879478826, + "grad_norm": 13.657927099528628, + "learning_rate": 9.295457869422818e-06, + "loss": 0.9807, + "step": 4944 + }, + { + "epoch": 2.684581976112921, + "grad_norm": 12.726310962899635, + "learning_rate": 9.291949812940387e-06, + "loss": 0.6495, + "step": 4945 + }, + { + "epoch": 2.6851248642779586, + "grad_norm": 11.439120086220523, + "learning_rate": 9.288441844030735e-06, + "loss": 0.6616, + "step": 4946 + }, + { + "epoch": 2.685667752442997, + "grad_norm": 10.384831951074661, + "learning_rate": 9.284933963127724e-06, + "loss": 0.536, + "step": 4947 + }, + { + "epoch": 2.6862106406080346, + "grad_norm": 8.803102808116847, + "learning_rate": 9.281426170665225e-06, + "loss": 0.6621, + "step": 4948 + }, + { + "epoch": 2.686753528773073, + "grad_norm": 14.516361098495818, + "learning_rate": 9.277918467077077e-06, + "loss": 1.0579, + "step": 4949 + }, + { + "epoch": 2.6872964169381106, + "grad_norm": 11.08498885907847, + "learning_rate": 9.274410852797126e-06, + "loss": 0.463, + "step": 4950 + }, + { + "epoch": 2.687839305103149, + "grad_norm": 10.827937520590183, + "learning_rate": 9.270903328259195e-06, + "loss": 0.5015, + "step": 4951 + }, + { + "epoch": 2.6883821932681866, + "grad_norm": 11.572829724801146, + "learning_rate": 9.267395893897107e-06, + "loss": 0.663, + "step": 4952 + }, + { + "epoch": 2.688925081433225, + "grad_norm": 9.3016488903525, + "learning_rate": 9.26388855014466e-06, + "loss": 0.6876, + "step": 4953 + }, + { + "epoch": 2.6894679695982626, + "grad_norm": 17.324535315422423, + "learning_rate": 9.260381297435652e-06, + "loss": 1.1741, + "step": 4954 + }, + { + "epoch": 2.690010857763301, + "grad_norm": 11.465003825461562, + "learning_rate": 9.256874136203864e-06, + "loss": 0.6867, + "step": 4955 + }, + { + "epoch": 2.6905537459283386, + "grad_norm": 8.846935307480056, + "learning_rate": 9.253367066883063e-06, + "loss": 0.4518, + "step": 4956 + }, + { + "epoch": 2.691096634093377, + "grad_norm": 11.007513424878196, + "learning_rate": 9.249860089907018e-06, + "loss": 0.7987, + "step": 4957 + }, + { + "epoch": 2.6916395222584146, + "grad_norm": 12.297659996678936, + "learning_rate": 9.246353205709468e-06, + "loss": 0.588, + "step": 4958 + }, + { + "epoch": 2.692182410423453, + "grad_norm": 10.549862062131211, + "learning_rate": 9.24284641472416e-06, + "loss": 0.737, + "step": 4959 + }, + { + "epoch": 2.6927252985884906, + "grad_norm": 10.712460113916158, + "learning_rate": 9.23933971738481e-06, + "loss": 0.6092, + "step": 4960 + }, + { + "epoch": 2.693268186753529, + "grad_norm": 14.104109743551868, + "learning_rate": 9.235833114125141e-06, + "loss": 1.008, + "step": 4961 + }, + { + "epoch": 2.6938110749185666, + "grad_norm": 7.755916952898445, + "learning_rate": 9.232326605378843e-06, + "loss": 0.4708, + "step": 4962 + }, + { + "epoch": 2.694353963083605, + "grad_norm": 11.677016459785008, + "learning_rate": 9.228820191579618e-06, + "loss": 0.7099, + "step": 4963 + }, + { + "epoch": 2.6948968512486426, + "grad_norm": 10.070884129580872, + "learning_rate": 9.225313873161139e-06, + "loss": 0.6546, + "step": 4964 + }, + { + "epoch": 2.695439739413681, + "grad_norm": 8.108710814711873, + "learning_rate": 9.221807650557078e-06, + "loss": 0.6102, + "step": 4965 + }, + { + "epoch": 2.6959826275787186, + "grad_norm": 10.279519502018594, + "learning_rate": 9.218301524201087e-06, + "loss": 0.5783, + "step": 4966 + }, + { + "epoch": 2.696525515743757, + "grad_norm": 11.882105089319197, + "learning_rate": 9.214795494526806e-06, + "loss": 1.1489, + "step": 4967 + }, + { + "epoch": 2.6970684039087947, + "grad_norm": 9.767493447568013, + "learning_rate": 9.211289561967872e-06, + "loss": 0.6051, + "step": 4968 + }, + { + "epoch": 2.697611292073833, + "grad_norm": 11.690415242614575, + "learning_rate": 9.207783726957903e-06, + "loss": 0.592, + "step": 4969 + }, + { + "epoch": 2.6981541802388707, + "grad_norm": 7.427254360689941, + "learning_rate": 9.204277989930502e-06, + "loss": 0.4672, + "step": 4970 + }, + { + "epoch": 2.698697068403909, + "grad_norm": 9.41305388818184, + "learning_rate": 9.200772351319266e-06, + "loss": 0.515, + "step": 4971 + }, + { + "epoch": 2.6992399565689467, + "grad_norm": 11.02585368985108, + "learning_rate": 9.197266811557787e-06, + "loss": 0.7867, + "step": 4972 + }, + { + "epoch": 2.699782844733985, + "grad_norm": 10.462642998204657, + "learning_rate": 9.193761371079622e-06, + "loss": 0.6689, + "step": 4973 + }, + { + "epoch": 2.7003257328990227, + "grad_norm": 9.964585703249949, + "learning_rate": 9.190256030318339e-06, + "loss": 0.6676, + "step": 4974 + }, + { + "epoch": 2.700868621064061, + "grad_norm": 12.383650152937822, + "learning_rate": 9.186750789707478e-06, + "loss": 0.9124, + "step": 4975 + }, + { + "epoch": 2.7014115092290987, + "grad_norm": 7.428651587888086, + "learning_rate": 9.183245649680574e-06, + "loss": 0.6241, + "step": 4976 + }, + { + "epoch": 2.701954397394137, + "grad_norm": 8.171702461332401, + "learning_rate": 9.179740610671155e-06, + "loss": 0.468, + "step": 4977 + }, + { + "epoch": 2.7024972855591747, + "grad_norm": 15.630545262276021, + "learning_rate": 9.176235673112719e-06, + "loss": 1.4177, + "step": 4978 + }, + { + "epoch": 2.703040173724213, + "grad_norm": 9.04535779717474, + "learning_rate": 9.172730837438774e-06, + "loss": 0.504, + "step": 4979 + }, + { + "epoch": 2.7035830618892507, + "grad_norm": 9.263908210238581, + "learning_rate": 9.169226104082792e-06, + "loss": 0.4043, + "step": 4980 + }, + { + "epoch": 2.704125950054289, + "grad_norm": 10.022418469090024, + "learning_rate": 9.165721473478253e-06, + "loss": 0.6372, + "step": 4981 + }, + { + "epoch": 2.7046688382193267, + "grad_norm": 12.924869609819282, + "learning_rate": 9.16221694605861e-06, + "loss": 0.9433, + "step": 4982 + }, + { + "epoch": 2.705211726384365, + "grad_norm": 8.79512920135937, + "learning_rate": 9.158712522257309e-06, + "loss": 0.4725, + "step": 4983 + }, + { + "epoch": 2.7057546145494027, + "grad_norm": 11.753381756239166, + "learning_rate": 9.155208202507789e-06, + "loss": 0.7395, + "step": 4984 + }, + { + "epoch": 2.706297502714441, + "grad_norm": 6.454949124720575, + "learning_rate": 9.151703987243459e-06, + "loss": 0.3108, + "step": 4985 + }, + { + "epoch": 2.7068403908794787, + "grad_norm": 15.244281036714051, + "learning_rate": 9.148199876897737e-06, + "loss": 0.802, + "step": 4986 + }, + { + "epoch": 2.707383279044517, + "grad_norm": 11.093235618680396, + "learning_rate": 9.144695871904005e-06, + "loss": 0.6252, + "step": 4987 + }, + { + "epoch": 2.7079261672095547, + "grad_norm": 10.703835311171614, + "learning_rate": 9.141191972695655e-06, + "loss": 0.6642, + "step": 4988 + }, + { + "epoch": 2.708469055374593, + "grad_norm": 11.327820449600674, + "learning_rate": 9.137688179706049e-06, + "loss": 1.1613, + "step": 4989 + }, + { + "epoch": 2.7090119435396307, + "grad_norm": 7.527861958807383, + "learning_rate": 9.134184493368548e-06, + "loss": 0.4991, + "step": 4990 + }, + { + "epoch": 2.709554831704669, + "grad_norm": 10.199776895453677, + "learning_rate": 9.130680914116484e-06, + "loss": 0.977, + "step": 4991 + }, + { + "epoch": 2.7100977198697067, + "grad_norm": 9.778998506495286, + "learning_rate": 9.127177442383192e-06, + "loss": 0.6729, + "step": 4992 + }, + { + "epoch": 2.710640608034745, + "grad_norm": 10.831437132716289, + "learning_rate": 9.123674078601984e-06, + "loss": 0.6252, + "step": 4993 + }, + { + "epoch": 2.7111834961997827, + "grad_norm": 9.722012246624114, + "learning_rate": 9.120170823206165e-06, + "loss": 0.6169, + "step": 4994 + }, + { + "epoch": 2.711726384364821, + "grad_norm": 9.696836781330795, + "learning_rate": 9.116667676629019e-06, + "loss": 0.7824, + "step": 4995 + }, + { + "epoch": 2.7122692725298587, + "grad_norm": 8.532030770830737, + "learning_rate": 9.11316463930382e-06, + "loss": 0.7649, + "step": 4996 + }, + { + "epoch": 2.712812160694897, + "grad_norm": 10.155581062168379, + "learning_rate": 9.109661711663837e-06, + "loss": 0.5581, + "step": 4997 + }, + { + "epoch": 2.7133550488599347, + "grad_norm": 11.913092610845165, + "learning_rate": 9.106158894142307e-06, + "loss": 0.6256, + "step": 4998 + }, + { + "epoch": 2.713897937024973, + "grad_norm": 10.326179923874385, + "learning_rate": 9.102656187172475e-06, + "loss": 0.8703, + "step": 4999 + }, + { + "epoch": 2.7144408251900107, + "grad_norm": 10.51344451860866, + "learning_rate": 9.099153591187553e-06, + "loss": 0.8281, + "step": 5000 + }, + { + "epoch": 2.714983713355049, + "grad_norm": 9.101512118782772, + "learning_rate": 9.09565110662075e-06, + "loss": 0.4127, + "step": 5001 + }, + { + "epoch": 2.7155266015200867, + "grad_norm": 14.912595390103256, + "learning_rate": 9.092148733905257e-06, + "loss": 0.828, + "step": 5002 + }, + { + "epoch": 2.716069489685125, + "grad_norm": 10.567445107841705, + "learning_rate": 9.088646473474262e-06, + "loss": 0.8542, + "step": 5003 + }, + { + "epoch": 2.7166123778501627, + "grad_norm": 11.727344925695448, + "learning_rate": 9.085144325760922e-06, + "loss": 0.7593, + "step": 5004 + }, + { + "epoch": 2.717155266015201, + "grad_norm": 10.964390265146204, + "learning_rate": 9.081642291198387e-06, + "loss": 0.696, + "step": 5005 + }, + { + "epoch": 2.7176981541802387, + "grad_norm": 8.594728435704301, + "learning_rate": 9.0781403702198e-06, + "loss": 0.3582, + "step": 5006 + }, + { + "epoch": 2.718241042345277, + "grad_norm": 10.662428429321805, + "learning_rate": 9.074638563258279e-06, + "loss": 0.6213, + "step": 5007 + }, + { + "epoch": 2.7187839305103148, + "grad_norm": 13.853700163509458, + "learning_rate": 9.071136870746934e-06, + "loss": 0.653, + "step": 5008 + }, + { + "epoch": 2.719326818675353, + "grad_norm": 9.421600748659138, + "learning_rate": 9.067635293118862e-06, + "loss": 0.4653, + "step": 5009 + }, + { + "epoch": 2.7198697068403908, + "grad_norm": 15.40532102462592, + "learning_rate": 9.064133830807147e-06, + "loss": 0.8502, + "step": 5010 + }, + { + "epoch": 2.720412595005429, + "grad_norm": 14.246891389640892, + "learning_rate": 9.060632484244845e-06, + "loss": 0.711, + "step": 5011 + }, + { + "epoch": 2.7209554831704668, + "grad_norm": 13.657358501095473, + "learning_rate": 9.057131253865022e-06, + "loss": 1.0586, + "step": 5012 + }, + { + "epoch": 2.721498371335505, + "grad_norm": 11.08278909637767, + "learning_rate": 9.053630140100701e-06, + "loss": 0.5658, + "step": 5013 + }, + { + "epoch": 2.7220412595005428, + "grad_norm": 11.922765678431421, + "learning_rate": 9.050129143384917e-06, + "loss": 0.6929, + "step": 5014 + }, + { + "epoch": 2.722584147665581, + "grad_norm": 10.93390692730677, + "learning_rate": 9.046628264150674e-06, + "loss": 1.417, + "step": 5015 + }, + { + "epoch": 2.7231270358306188, + "grad_norm": 12.86011863530373, + "learning_rate": 9.043127502830964e-06, + "loss": 1.0364, + "step": 5016 + }, + { + "epoch": 2.723669923995657, + "grad_norm": 7.848749292562499, + "learning_rate": 9.039626859858773e-06, + "loss": 0.6696, + "step": 5017 + }, + { + "epoch": 2.7242128121606948, + "grad_norm": 11.212923870670298, + "learning_rate": 9.036126335667059e-06, + "loss": 0.5751, + "step": 5018 + }, + { + "epoch": 2.724755700325733, + "grad_norm": 9.666644723437658, + "learning_rate": 9.032625930688781e-06, + "loss": 0.5699, + "step": 5019 + }, + { + "epoch": 2.725298588490771, + "grad_norm": 8.860936490743164, + "learning_rate": 9.029125645356864e-06, + "loss": 0.4262, + "step": 5020 + }, + { + "epoch": 2.725841476655809, + "grad_norm": 8.597880980516297, + "learning_rate": 9.025625480104238e-06, + "loss": 0.5646, + "step": 5021 + }, + { + "epoch": 2.726384364820847, + "grad_norm": 11.334362425216222, + "learning_rate": 9.022125435363803e-06, + "loss": 0.7424, + "step": 5022 + }, + { + "epoch": 2.726927252985885, + "grad_norm": 10.551777615550284, + "learning_rate": 9.018625511568456e-06, + "loss": 0.4961, + "step": 5023 + }, + { + "epoch": 2.727470141150923, + "grad_norm": 9.061655577456406, + "learning_rate": 9.015125709151069e-06, + "loss": 0.6466, + "step": 5024 + }, + { + "epoch": 2.728013029315961, + "grad_norm": 12.51368822693772, + "learning_rate": 9.011626028544502e-06, + "loss": 0.8461, + "step": 5025 + }, + { + "epoch": 2.728555917480999, + "grad_norm": 9.972383117450178, + "learning_rate": 9.008126470181605e-06, + "loss": 0.7941, + "step": 5026 + }, + { + "epoch": 2.729098805646037, + "grad_norm": 9.372092186313578, + "learning_rate": 9.004627034495204e-06, + "loss": 0.5291, + "step": 5027 + }, + { + "epoch": 2.729641693811075, + "grad_norm": 11.649348346812042, + "learning_rate": 9.00112772191812e-06, + "loss": 0.658, + "step": 5028 + }, + { + "epoch": 2.730184581976113, + "grad_norm": 9.4217070731629, + "learning_rate": 8.997628532883149e-06, + "loss": 0.7391, + "step": 5029 + }, + { + "epoch": 2.730727470141151, + "grad_norm": 9.294886195505088, + "learning_rate": 8.994129467823083e-06, + "loss": 0.7955, + "step": 5030 + }, + { + "epoch": 2.731270358306189, + "grad_norm": 9.046521616766688, + "learning_rate": 8.990630527170684e-06, + "loss": 0.6163, + "step": 5031 + }, + { + "epoch": 2.731813246471227, + "grad_norm": 13.078763070358596, + "learning_rate": 8.987131711358714e-06, + "loss": 0.6352, + "step": 5032 + }, + { + "epoch": 2.732356134636265, + "grad_norm": 8.372819066707777, + "learning_rate": 8.983633020819906e-06, + "loss": 0.5139, + "step": 5033 + }, + { + "epoch": 2.732899022801303, + "grad_norm": 7.908976790945395, + "learning_rate": 8.980134455986982e-06, + "loss": 0.4263, + "step": 5034 + }, + { + "epoch": 2.733441910966341, + "grad_norm": 9.508847688216756, + "learning_rate": 8.97663601729266e-06, + "loss": 0.614, + "step": 5035 + }, + { + "epoch": 2.733984799131379, + "grad_norm": 9.135340041370506, + "learning_rate": 8.973137705169621e-06, + "loss": 0.5803, + "step": 5036 + }, + { + "epoch": 2.734527687296417, + "grad_norm": 10.900299047123902, + "learning_rate": 8.969639520050553e-06, + "loss": 0.4833, + "step": 5037 + }, + { + "epoch": 2.735070575461455, + "grad_norm": 9.243890767443602, + "learning_rate": 8.966141462368106e-06, + "loss": 0.4405, + "step": 5038 + }, + { + "epoch": 2.735613463626493, + "grad_norm": 11.372121051541363, + "learning_rate": 8.962643532554934e-06, + "loss": 0.6618, + "step": 5039 + }, + { + "epoch": 2.736156351791531, + "grad_norm": 11.318259460716959, + "learning_rate": 8.95914573104366e-06, + "loss": 1.0367, + "step": 5040 + }, + { + "epoch": 2.736699239956569, + "grad_norm": 9.055918919260264, + "learning_rate": 8.955648058266904e-06, + "loss": 0.5463, + "step": 5041 + }, + { + "epoch": 2.737242128121607, + "grad_norm": 12.452665937187591, + "learning_rate": 8.952150514657258e-06, + "loss": 0.9772, + "step": 5042 + }, + { + "epoch": 2.737785016286645, + "grad_norm": 11.532813733829784, + "learning_rate": 8.94865310064731e-06, + "loss": 0.6372, + "step": 5043 + }, + { + "epoch": 2.738327904451683, + "grad_norm": 9.800145767151978, + "learning_rate": 8.945155816669622e-06, + "loss": 0.6384, + "step": 5044 + }, + { + "epoch": 2.738870792616721, + "grad_norm": 13.619886494303222, + "learning_rate": 8.94165866315674e-06, + "loss": 0.7295, + "step": 5045 + }, + { + "epoch": 2.739413680781759, + "grad_norm": 13.097331078127143, + "learning_rate": 8.938161640541202e-06, + "loss": 1.1075, + "step": 5046 + }, + { + "epoch": 2.739956568946797, + "grad_norm": 17.458162898454454, + "learning_rate": 8.934664749255524e-06, + "loss": 1.0174, + "step": 5047 + }, + { + "epoch": 2.740499457111835, + "grad_norm": 7.51493556495456, + "learning_rate": 8.931167989732212e-06, + "loss": 0.6407, + "step": 5048 + }, + { + "epoch": 2.741042345276873, + "grad_norm": 11.717405055058808, + "learning_rate": 8.927671362403741e-06, + "loss": 0.6899, + "step": 5049 + }, + { + "epoch": 2.741585233441911, + "grad_norm": 11.433736258508228, + "learning_rate": 8.924174867702591e-06, + "loss": 0.8481, + "step": 5050 + }, + { + "epoch": 2.742128121606949, + "grad_norm": 11.800039466500984, + "learning_rate": 8.920678506061202e-06, + "loss": 0.6044, + "step": 5051 + }, + { + "epoch": 2.742671009771987, + "grad_norm": 8.499003916350551, + "learning_rate": 8.91718227791202e-06, + "loss": 0.475, + "step": 5052 + }, + { + "epoch": 2.743213897937025, + "grad_norm": 10.586652955325345, + "learning_rate": 8.913686183687459e-06, + "loss": 0.5853, + "step": 5053 + }, + { + "epoch": 2.743756786102063, + "grad_norm": 14.066921486210767, + "learning_rate": 8.910190223819919e-06, + "loss": 0.8164, + "step": 5054 + }, + { + "epoch": 2.744299674267101, + "grad_norm": 9.77990149419258, + "learning_rate": 8.906694398741792e-06, + "loss": 0.6053, + "step": 5055 + }, + { + "epoch": 2.744842562432139, + "grad_norm": 15.328290145831401, + "learning_rate": 8.903198708885442e-06, + "loss": 1.3555, + "step": 5056 + }, + { + "epoch": 2.745385450597177, + "grad_norm": 11.626317894758118, + "learning_rate": 8.899703154683228e-06, + "loss": 0.6692, + "step": 5057 + }, + { + "epoch": 2.745928338762215, + "grad_norm": 11.566102899846406, + "learning_rate": 8.896207736567476e-06, + "loss": 0.6589, + "step": 5058 + }, + { + "epoch": 2.746471226927253, + "grad_norm": 9.521589192518187, + "learning_rate": 8.892712454970512e-06, + "loss": 0.7559, + "step": 5059 + }, + { + "epoch": 2.747014115092291, + "grad_norm": 8.719647496284242, + "learning_rate": 8.889217310324636e-06, + "loss": 0.4123, + "step": 5060 + }, + { + "epoch": 2.747557003257329, + "grad_norm": 6.709750992678039, + "learning_rate": 8.885722303062136e-06, + "loss": 0.3492, + "step": 5061 + }, + { + "epoch": 2.748099891422367, + "grad_norm": 11.273080989021693, + "learning_rate": 8.882227433615275e-06, + "loss": 0.7416, + "step": 5062 + }, + { + "epoch": 2.748642779587405, + "grad_norm": 13.864388014548679, + "learning_rate": 8.87873270241631e-06, + "loss": 0.8197, + "step": 5063 + }, + { + "epoch": 2.749185667752443, + "grad_norm": 9.492558149681301, + "learning_rate": 8.875238109897468e-06, + "loss": 0.6517, + "step": 5064 + }, + { + "epoch": 2.749728555917481, + "grad_norm": 10.634433093222732, + "learning_rate": 8.87174365649097e-06, + "loss": 0.5431, + "step": 5065 + }, + { + "epoch": 2.750271444082519, + "grad_norm": 8.469745230774542, + "learning_rate": 8.868249342629015e-06, + "loss": 0.5189, + "step": 5066 + }, + { + "epoch": 2.750814332247557, + "grad_norm": 8.182617911488233, + "learning_rate": 8.864755168743783e-06, + "loss": 0.4322, + "step": 5067 + }, + { + "epoch": 2.751357220412595, + "grad_norm": 7.6335690540689685, + "learning_rate": 8.861261135267444e-06, + "loss": 0.5268, + "step": 5068 + }, + { + "epoch": 2.751900108577633, + "grad_norm": 11.52493444073719, + "learning_rate": 8.85776724263214e-06, + "loss": 0.9757, + "step": 5069 + }, + { + "epoch": 2.752442996742671, + "grad_norm": 9.591556669015672, + "learning_rate": 8.854273491270008e-06, + "loss": 0.5527, + "step": 5070 + }, + { + "epoch": 2.752985884907709, + "grad_norm": 11.178077098634855, + "learning_rate": 8.850779881613151e-06, + "loss": 0.581, + "step": 5071 + }, + { + "epoch": 2.753528773072747, + "grad_norm": 8.921814933147093, + "learning_rate": 8.847286414093673e-06, + "loss": 0.5604, + "step": 5072 + }, + { + "epoch": 2.754071661237785, + "grad_norm": 11.887605361931898, + "learning_rate": 8.84379308914365e-06, + "loss": 0.6055, + "step": 5073 + }, + { + "epoch": 2.754614549402823, + "grad_norm": 14.12418845494581, + "learning_rate": 8.840299907195137e-06, + "loss": 0.9614, + "step": 5074 + }, + { + "epoch": 2.755157437567861, + "grad_norm": 10.094252933928091, + "learning_rate": 8.836806868680185e-06, + "loss": 0.6836, + "step": 5075 + }, + { + "epoch": 2.755700325732899, + "grad_norm": 10.048762466683991, + "learning_rate": 8.833313974030807e-06, + "loss": 0.5722, + "step": 5076 + }, + { + "epoch": 2.756243213897937, + "grad_norm": 12.51909069586841, + "learning_rate": 8.829821223679022e-06, + "loss": 1.2238, + "step": 5077 + }, + { + "epoch": 2.756786102062975, + "grad_norm": 9.50434691504725, + "learning_rate": 8.826328618056808e-06, + "loss": 0.6393, + "step": 5078 + }, + { + "epoch": 2.757328990228013, + "grad_norm": 10.737059961404412, + "learning_rate": 8.822836157596141e-06, + "loss": 0.7401, + "step": 5079 + }, + { + "epoch": 2.757871878393051, + "grad_norm": 9.16480609854369, + "learning_rate": 8.819343842728976e-06, + "loss": 0.5289, + "step": 5080 + }, + { + "epoch": 2.758414766558089, + "grad_norm": 10.849970138344764, + "learning_rate": 8.815851673887248e-06, + "loss": 0.6157, + "step": 5081 + }, + { + "epoch": 2.758957654723127, + "grad_norm": 8.609887529945185, + "learning_rate": 8.812359651502872e-06, + "loss": 0.5961, + "step": 5082 + }, + { + "epoch": 2.759500542888165, + "grad_norm": 10.765964781119916, + "learning_rate": 8.808867776007745e-06, + "loss": 1.1942, + "step": 5083 + }, + { + "epoch": 2.760043431053203, + "grad_norm": 10.466671227569803, + "learning_rate": 8.80537604783375e-06, + "loss": 0.7117, + "step": 5084 + }, + { + "epoch": 2.760586319218241, + "grad_norm": 10.079228354003783, + "learning_rate": 8.801884467412747e-06, + "loss": 0.4238, + "step": 5085 + }, + { + "epoch": 2.761129207383279, + "grad_norm": 11.422859396303908, + "learning_rate": 8.798393035176588e-06, + "loss": 0.8841, + "step": 5086 + }, + { + "epoch": 2.761672095548317, + "grad_norm": 9.736302153055926, + "learning_rate": 8.79490175155709e-06, + "loss": 0.6414, + "step": 5087 + }, + { + "epoch": 2.762214983713355, + "grad_norm": 8.513897139101958, + "learning_rate": 8.791410616986067e-06, + "loss": 0.3052, + "step": 5088 + }, + { + "epoch": 2.762757871878393, + "grad_norm": 10.101572440599828, + "learning_rate": 8.787919631895301e-06, + "loss": 0.5305, + "step": 5089 + }, + { + "epoch": 2.763300760043431, + "grad_norm": 12.17319240146357, + "learning_rate": 8.784428796716571e-06, + "loss": 0.6585, + "step": 5090 + }, + { + "epoch": 2.763843648208469, + "grad_norm": 10.930394499045825, + "learning_rate": 8.78093811188162e-06, + "loss": 0.5984, + "step": 5091 + }, + { + "epoch": 2.764386536373507, + "grad_norm": 9.550475415502788, + "learning_rate": 8.77744757782219e-06, + "loss": 0.5634, + "step": 5092 + }, + { + "epoch": 2.764929424538545, + "grad_norm": 12.890085749205962, + "learning_rate": 8.773957194969993e-06, + "loss": 0.9274, + "step": 5093 + }, + { + "epoch": 2.765472312703583, + "grad_norm": 12.38000962026181, + "learning_rate": 8.77046696375672e-06, + "loss": 0.6988, + "step": 5094 + }, + { + "epoch": 2.766015200868621, + "grad_norm": 9.523692226826551, + "learning_rate": 8.766976884614056e-06, + "loss": 0.4856, + "step": 5095 + }, + { + "epoch": 2.766558089033659, + "grad_norm": 9.26829606100349, + "learning_rate": 8.763486957973652e-06, + "loss": 0.6863, + "step": 5096 + }, + { + "epoch": 2.767100977198697, + "grad_norm": 11.227870751984687, + "learning_rate": 8.759997184267155e-06, + "loss": 0.6344, + "step": 5097 + }, + { + "epoch": 2.767643865363735, + "grad_norm": 11.653609442391305, + "learning_rate": 8.756507563926182e-06, + "loss": 0.5884, + "step": 5098 + }, + { + "epoch": 2.768186753528773, + "grad_norm": 11.59192980688709, + "learning_rate": 8.753018097382336e-06, + "loss": 0.8214, + "step": 5099 + }, + { + "epoch": 2.768729641693811, + "grad_norm": 8.668298021688548, + "learning_rate": 8.749528785067196e-06, + "loss": 0.6487, + "step": 5100 + }, + { + "epoch": 2.769272529858849, + "grad_norm": 5.456725122843202, + "learning_rate": 8.746039627412333e-06, + "loss": 0.2868, + "step": 5101 + }, + { + "epoch": 2.769815418023887, + "grad_norm": 12.29265212176352, + "learning_rate": 8.742550624849288e-06, + "loss": 0.512, + "step": 5102 + }, + { + "epoch": 2.770358306188925, + "grad_norm": 14.125650271042089, + "learning_rate": 8.73906177780958e-06, + "loss": 0.8903, + "step": 5103 + }, + { + "epoch": 2.770901194353963, + "grad_norm": 12.046023026774584, + "learning_rate": 8.735573086724725e-06, + "loss": 0.801, + "step": 5104 + }, + { + "epoch": 2.771444082519001, + "grad_norm": 10.153944591634119, + "learning_rate": 8.732084552026203e-06, + "loss": 0.6888, + "step": 5105 + }, + { + "epoch": 2.771986970684039, + "grad_norm": 10.581737293617769, + "learning_rate": 8.72859617414549e-06, + "loss": 0.7775, + "step": 5106 + }, + { + "epoch": 2.772529858849077, + "grad_norm": 8.304371234513637, + "learning_rate": 8.725107953514021e-06, + "loss": 0.4704, + "step": 5107 + }, + { + "epoch": 2.773072747014115, + "grad_norm": 12.579288268640527, + "learning_rate": 8.72161989056324e-06, + "loss": 0.5861, + "step": 5108 + }, + { + "epoch": 2.773615635179153, + "grad_norm": 10.68813978428288, + "learning_rate": 8.718131985724542e-06, + "loss": 0.6367, + "step": 5109 + }, + { + "epoch": 2.774158523344191, + "grad_norm": 8.921584649335454, + "learning_rate": 8.714644239429326e-06, + "loss": 0.5597, + "step": 5110 + }, + { + "epoch": 2.774701411509229, + "grad_norm": 10.473739328784887, + "learning_rate": 8.711156652108957e-06, + "loss": 0.8587, + "step": 5111 + }, + { + "epoch": 2.775244299674267, + "grad_norm": 10.905504725358476, + "learning_rate": 8.70766922419479e-06, + "loss": 0.5729, + "step": 5112 + }, + { + "epoch": 2.7757871878393052, + "grad_norm": 7.88310879226585, + "learning_rate": 8.704181956118153e-06, + "loss": 0.4998, + "step": 5113 + }, + { + "epoch": 2.776330076004343, + "grad_norm": 8.595679686197345, + "learning_rate": 8.700694848310354e-06, + "loss": 0.699, + "step": 5114 + }, + { + "epoch": 2.7768729641693812, + "grad_norm": 9.784504057993846, + "learning_rate": 8.697207901202691e-06, + "loss": 0.6509, + "step": 5115 + }, + { + "epoch": 2.777415852334419, + "grad_norm": 9.537890155200703, + "learning_rate": 8.693721115226427e-06, + "loss": 0.4484, + "step": 5116 + }, + { + "epoch": 2.7779587404994572, + "grad_norm": 10.110968424420188, + "learning_rate": 8.69023449081282e-06, + "loss": 0.6746, + "step": 5117 + }, + { + "epoch": 2.778501628664495, + "grad_norm": 14.837427287717714, + "learning_rate": 8.686748028393096e-06, + "loss": 1.3601, + "step": 5118 + }, + { + "epoch": 2.7790445168295332, + "grad_norm": 11.553555350476039, + "learning_rate": 8.683261728398472e-06, + "loss": 0.6054, + "step": 5119 + }, + { + "epoch": 2.779587404994571, + "grad_norm": 12.684609984344696, + "learning_rate": 8.679775591260132e-06, + "loss": 0.947, + "step": 5120 + }, + { + "epoch": 2.7801302931596092, + "grad_norm": 9.50144977278699, + "learning_rate": 8.676289617409256e-06, + "loss": 0.6401, + "step": 5121 + }, + { + "epoch": 2.780673181324647, + "grad_norm": 9.588585115620386, + "learning_rate": 8.672803807276988e-06, + "loss": 0.7214, + "step": 5122 + }, + { + "epoch": 2.7812160694896852, + "grad_norm": 14.597044499431716, + "learning_rate": 8.66931816129446e-06, + "loss": 1.0037, + "step": 5123 + }, + { + "epoch": 2.781758957654723, + "grad_norm": 10.552996924508232, + "learning_rate": 8.665832679892783e-06, + "loss": 0.5097, + "step": 5124 + }, + { + "epoch": 2.7823018458197613, + "grad_norm": 8.368460915764464, + "learning_rate": 8.662347363503043e-06, + "loss": 0.5289, + "step": 5125 + }, + { + "epoch": 2.782844733984799, + "grad_norm": 7.412774968306893, + "learning_rate": 8.658862212556318e-06, + "loss": 0.4117, + "step": 5126 + }, + { + "epoch": 2.7833876221498373, + "grad_norm": 7.965326482678092, + "learning_rate": 8.655377227483648e-06, + "loss": 0.4492, + "step": 5127 + }, + { + "epoch": 2.783930510314875, + "grad_norm": 10.16388699920939, + "learning_rate": 8.65189240871607e-06, + "loss": 0.6847, + "step": 5128 + }, + { + "epoch": 2.7844733984799133, + "grad_norm": 8.606167058304848, + "learning_rate": 8.648407756684582e-06, + "loss": 0.5952, + "step": 5129 + }, + { + "epoch": 2.785016286644951, + "grad_norm": 16.346182786192927, + "learning_rate": 8.64492327182018e-06, + "loss": 1.0626, + "step": 5130 + }, + { + "epoch": 2.7855591748099893, + "grad_norm": 9.397161435588243, + "learning_rate": 8.64143895455383e-06, + "loss": 0.6416, + "step": 5131 + }, + { + "epoch": 2.786102062975027, + "grad_norm": 10.215541582389136, + "learning_rate": 8.63795480531647e-06, + "loss": 0.5311, + "step": 5132 + }, + { + "epoch": 2.7866449511400653, + "grad_norm": 13.362954088182768, + "learning_rate": 8.634470824539035e-06, + "loss": 1.0249, + "step": 5133 + }, + { + "epoch": 2.787187839305103, + "grad_norm": 13.843076908870096, + "learning_rate": 8.630987012652421e-06, + "loss": 0.767, + "step": 5134 + }, + { + "epoch": 2.7877307274701413, + "grad_norm": 10.581731921492887, + "learning_rate": 8.627503370087519e-06, + "loss": 0.6027, + "step": 5135 + }, + { + "epoch": 2.788273615635179, + "grad_norm": 11.161057087658808, + "learning_rate": 8.624019897275184e-06, + "loss": 0.8193, + "step": 5136 + }, + { + "epoch": 2.7888165038002173, + "grad_norm": 7.972214509867566, + "learning_rate": 8.620536594646262e-06, + "loss": 0.4776, + "step": 5137 + }, + { + "epoch": 2.789359391965255, + "grad_norm": 12.57635432127455, + "learning_rate": 8.61705346263157e-06, + "loss": 0.7194, + "step": 5138 + }, + { + "epoch": 2.7899022801302933, + "grad_norm": 7.077115343788877, + "learning_rate": 8.613570501661915e-06, + "loss": 0.3571, + "step": 5139 + }, + { + "epoch": 2.790445168295331, + "grad_norm": 13.158625298591346, + "learning_rate": 8.610087712168065e-06, + "loss": 0.5872, + "step": 5140 + }, + { + "epoch": 2.7909880564603693, + "grad_norm": 8.652483474203775, + "learning_rate": 8.606605094580788e-06, + "loss": 0.624, + "step": 5141 + }, + { + "epoch": 2.791530944625407, + "grad_norm": 8.73276991998644, + "learning_rate": 8.603122649330811e-06, + "loss": 0.4358, + "step": 5142 + }, + { + "epoch": 2.7920738327904453, + "grad_norm": 12.749713905257105, + "learning_rate": 8.599640376848849e-06, + "loss": 0.7574, + "step": 5143 + }, + { + "epoch": 2.792616720955483, + "grad_norm": 12.054416741196698, + "learning_rate": 8.5961582775656e-06, + "loss": 0.886, + "step": 5144 + }, + { + "epoch": 2.7931596091205213, + "grad_norm": 10.958538797565106, + "learning_rate": 8.592676351911728e-06, + "loss": 0.7312, + "step": 5145 + }, + { + "epoch": 2.793702497285559, + "grad_norm": 10.124234819891349, + "learning_rate": 8.589194600317894e-06, + "loss": 0.5059, + "step": 5146 + }, + { + "epoch": 2.7942453854505973, + "grad_norm": 12.84934263063866, + "learning_rate": 8.585713023214715e-06, + "loss": 0.7016, + "step": 5147 + }, + { + "epoch": 2.794788273615635, + "grad_norm": 11.859965090765057, + "learning_rate": 8.582231621032807e-06, + "loss": 0.7158, + "step": 5148 + }, + { + "epoch": 2.7953311617806733, + "grad_norm": 11.57843340583913, + "learning_rate": 8.578750394202749e-06, + "loss": 0.7103, + "step": 5149 + }, + { + "epoch": 2.795874049945711, + "grad_norm": 7.143782607145388, + "learning_rate": 8.575269343155108e-06, + "loss": 0.5275, + "step": 5150 + }, + { + "epoch": 2.7964169381107493, + "grad_norm": 12.095108158973952, + "learning_rate": 8.571788468320427e-06, + "loss": 0.7308, + "step": 5151 + }, + { + "epoch": 2.796959826275787, + "grad_norm": 11.153004271537196, + "learning_rate": 8.568307770129223e-06, + "loss": 0.7576, + "step": 5152 + }, + { + "epoch": 2.7975027144408253, + "grad_norm": 8.326263454089252, + "learning_rate": 8.564827249011998e-06, + "loss": 0.408, + "step": 5153 + }, + { + "epoch": 2.798045602605863, + "grad_norm": 13.46902871500607, + "learning_rate": 8.561346905399221e-06, + "loss": 0.6624, + "step": 5154 + }, + { + "epoch": 2.7985884907709013, + "grad_norm": 8.925567749596054, + "learning_rate": 8.557866739721356e-06, + "loss": 0.4557, + "step": 5155 + }, + { + "epoch": 2.799131378935939, + "grad_norm": 11.700305884529763, + "learning_rate": 8.554386752408827e-06, + "loss": 0.6818, + "step": 5156 + }, + { + "epoch": 2.7996742671009773, + "grad_norm": 9.49767460620272, + "learning_rate": 8.550906943892054e-06, + "loss": 0.7256, + "step": 5157 + }, + { + "epoch": 2.800217155266015, + "grad_norm": 11.749650388906772, + "learning_rate": 8.547427314601416e-06, + "loss": 0.6933, + "step": 5158 + }, + { + "epoch": 2.8007600434310533, + "grad_norm": 11.830065328109, + "learning_rate": 8.543947864967286e-06, + "loss": 0.4282, + "step": 5159 + }, + { + "epoch": 2.801302931596091, + "grad_norm": 12.11100989733807, + "learning_rate": 8.540468595419999e-06, + "loss": 0.3617, + "step": 5160 + }, + { + "epoch": 2.8018458197611293, + "grad_norm": 9.923764364092053, + "learning_rate": 8.536989506389889e-06, + "loss": 0.6469, + "step": 5161 + }, + { + "epoch": 2.802388707926167, + "grad_norm": 10.555923377692814, + "learning_rate": 8.533510598307244e-06, + "loss": 0.9672, + "step": 5162 + }, + { + "epoch": 2.8029315960912053, + "grad_norm": 12.544866122081839, + "learning_rate": 8.530031871602345e-06, + "loss": 0.6885, + "step": 5163 + }, + { + "epoch": 2.803474484256243, + "grad_norm": 11.680102099467787, + "learning_rate": 8.526553326705452e-06, + "loss": 0.5571, + "step": 5164 + }, + { + "epoch": 2.8040173724212814, + "grad_norm": 11.769615311559438, + "learning_rate": 8.523074964046785e-06, + "loss": 0.6176, + "step": 5165 + }, + { + "epoch": 2.804560260586319, + "grad_norm": 10.249168095085986, + "learning_rate": 8.519596784056567e-06, + "loss": 0.7974, + "step": 5166 + }, + { + "epoch": 2.8051031487513574, + "grad_norm": 7.602995989085923, + "learning_rate": 8.516118787164973e-06, + "loss": 0.4132, + "step": 5167 + }, + { + "epoch": 2.805646036916395, + "grad_norm": 9.32078805379704, + "learning_rate": 8.512640973802175e-06, + "loss": 0.6492, + "step": 5168 + }, + { + "epoch": 2.8061889250814334, + "grad_norm": 13.80437475122657, + "learning_rate": 8.50916334439831e-06, + "loss": 0.6663, + "step": 5169 + }, + { + "epoch": 2.806731813246471, + "grad_norm": 9.121579447999185, + "learning_rate": 8.5056858993835e-06, + "loss": 0.4546, + "step": 5170 + }, + { + "epoch": 2.8072747014115094, + "grad_norm": 8.073186873408805, + "learning_rate": 8.502208639187842e-06, + "loss": 0.4467, + "step": 5171 + }, + { + "epoch": 2.807817589576547, + "grad_norm": 10.498644561860338, + "learning_rate": 8.498731564241403e-06, + "loss": 0.4907, + "step": 5172 + }, + { + "epoch": 2.8083604777415854, + "grad_norm": 9.027361205852865, + "learning_rate": 8.495254674974239e-06, + "loss": 0.4581, + "step": 5173 + }, + { + "epoch": 2.808903365906623, + "grad_norm": 9.692775527105594, + "learning_rate": 8.491777971816372e-06, + "loss": 0.5898, + "step": 5174 + }, + { + "epoch": 2.8094462540716614, + "grad_norm": 9.171182553280271, + "learning_rate": 8.48830145519781e-06, + "loss": 0.4129, + "step": 5175 + }, + { + "epoch": 2.809989142236699, + "grad_norm": 11.635096708230561, + "learning_rate": 8.484825125548532e-06, + "loss": 0.7632, + "step": 5176 + }, + { + "epoch": 2.8105320304017374, + "grad_norm": 12.18174333972076, + "learning_rate": 8.481348983298503e-06, + "loss": 0.6158, + "step": 5177 + }, + { + "epoch": 2.811074918566775, + "grad_norm": 8.880576943532684, + "learning_rate": 8.477873028877645e-06, + "loss": 0.374, + "step": 5178 + }, + { + "epoch": 2.8116178067318134, + "grad_norm": 11.735757011642823, + "learning_rate": 8.474397262715884e-06, + "loss": 0.6772, + "step": 5179 + }, + { + "epoch": 2.812160694896851, + "grad_norm": 9.50535973508058, + "learning_rate": 8.470921685243098e-06, + "loss": 0.6021, + "step": 5180 + }, + { + "epoch": 2.8127035830618894, + "grad_norm": 11.780091573431722, + "learning_rate": 8.467446296889151e-06, + "loss": 0.6944, + "step": 5181 + }, + { + "epoch": 2.813246471226927, + "grad_norm": 10.177875512109438, + "learning_rate": 8.463971098083896e-06, + "loss": 0.4312, + "step": 5182 + }, + { + "epoch": 2.8137893593919654, + "grad_norm": 16.588608629478436, + "learning_rate": 8.460496089257136e-06, + "loss": 0.818, + "step": 5183 + }, + { + "epoch": 2.814332247557003, + "grad_norm": 12.416295723031903, + "learning_rate": 8.45702127083868e-06, + "loss": 0.5531, + "step": 5184 + }, + { + "epoch": 2.8148751357220414, + "grad_norm": 13.657036134609482, + "learning_rate": 8.453546643258286e-06, + "loss": 0.7343, + "step": 5185 + }, + { + "epoch": 2.815418023887079, + "grad_norm": 9.833034529955514, + "learning_rate": 8.450072206945715e-06, + "loss": 0.6592, + "step": 5186 + }, + { + "epoch": 2.8159609120521174, + "grad_norm": 10.036112921946827, + "learning_rate": 8.446597962330675e-06, + "loss": 0.8127, + "step": 5187 + }, + { + "epoch": 2.816503800217155, + "grad_norm": 12.901123274951528, + "learning_rate": 8.443123909842882e-06, + "loss": 1.0507, + "step": 5188 + }, + { + "epoch": 2.8170466883821934, + "grad_norm": 10.692152694001132, + "learning_rate": 8.439650049911999e-06, + "loss": 0.5602, + "step": 5189 + }, + { + "epoch": 2.817589576547231, + "grad_norm": 19.731345695292617, + "learning_rate": 8.436176382967692e-06, + "loss": 1.0473, + "step": 5190 + }, + { + "epoch": 2.8181324647122694, + "grad_norm": 14.567179107047348, + "learning_rate": 8.432702909439579e-06, + "loss": 0.6359, + "step": 5191 + }, + { + "epoch": 2.818675352877307, + "grad_norm": 13.853200529175792, + "learning_rate": 8.429229629757266e-06, + "loss": 0.8277, + "step": 5192 + }, + { + "epoch": 2.8192182410423454, + "grad_norm": 9.811701731637578, + "learning_rate": 8.425756544350338e-06, + "loss": 0.533, + "step": 5193 + }, + { + "epoch": 2.819761129207383, + "grad_norm": 11.486229326022617, + "learning_rate": 8.422283653648348e-06, + "loss": 0.9078, + "step": 5194 + }, + { + "epoch": 2.8203040173724214, + "grad_norm": 11.086905618662822, + "learning_rate": 8.418810958080832e-06, + "loss": 0.7101, + "step": 5195 + }, + { + "epoch": 2.820846905537459, + "grad_norm": 8.065915758013317, + "learning_rate": 8.415338458077293e-06, + "loss": 0.5744, + "step": 5196 + }, + { + "epoch": 2.8213897937024974, + "grad_norm": 9.063098353131359, + "learning_rate": 8.411866154067224e-06, + "loss": 0.5008, + "step": 5197 + }, + { + "epoch": 2.821932681867535, + "grad_norm": 12.457981789036095, + "learning_rate": 8.408394046480077e-06, + "loss": 0.6377, + "step": 5198 + }, + { + "epoch": 2.8224755700325734, + "grad_norm": 10.806638216082748, + "learning_rate": 8.404922135745295e-06, + "loss": 0.6137, + "step": 5199 + }, + { + "epoch": 2.823018458197611, + "grad_norm": 10.78339089292669, + "learning_rate": 8.401450422292281e-06, + "loss": 0.7984, + "step": 5200 + }, + { + "epoch": 2.8235613463626494, + "grad_norm": 9.574355357184858, + "learning_rate": 8.397978906550429e-06, + "loss": 0.6307, + "step": 5201 + }, + { + "epoch": 2.824104234527687, + "grad_norm": 9.457238410125305, + "learning_rate": 8.3945075889491e-06, + "loss": 0.6782, + "step": 5202 + }, + { + "epoch": 2.8246471226927254, + "grad_norm": 15.132207636686802, + "learning_rate": 8.391036469917626e-06, + "loss": 0.7843, + "step": 5203 + }, + { + "epoch": 2.8251900108577632, + "grad_norm": 12.425847301998207, + "learning_rate": 8.387565549885331e-06, + "loss": 0.9953, + "step": 5204 + }, + { + "epoch": 2.8257328990228014, + "grad_norm": 9.948899666849417, + "learning_rate": 8.384094829281495e-06, + "loss": 0.6601, + "step": 5205 + }, + { + "epoch": 2.8262757871878392, + "grad_norm": 9.412819573011207, + "learning_rate": 8.38062430853539e-06, + "loss": 0.3747, + "step": 5206 + }, + { + "epoch": 2.8268186753528775, + "grad_norm": 8.137925073552552, + "learning_rate": 8.377153988076247e-06, + "loss": 0.3818, + "step": 5207 + }, + { + "epoch": 2.8273615635179152, + "grad_norm": 8.994244424095461, + "learning_rate": 8.373683868333287e-06, + "loss": 0.796, + "step": 5208 + }, + { + "epoch": 2.8279044516829535, + "grad_norm": 7.570203929319876, + "learning_rate": 8.370213949735696e-06, + "loss": 0.5587, + "step": 5209 + }, + { + "epoch": 2.8284473398479912, + "grad_norm": 13.134200586671845, + "learning_rate": 8.366744232712645e-06, + "loss": 0.8274, + "step": 5210 + }, + { + "epoch": 2.8289902280130295, + "grad_norm": 9.833443120274195, + "learning_rate": 8.363274717693272e-06, + "loss": 0.7984, + "step": 5211 + }, + { + "epoch": 2.8295331161780672, + "grad_norm": 11.623259569780537, + "learning_rate": 8.359805405106685e-06, + "loss": 0.7305, + "step": 5212 + }, + { + "epoch": 2.8300760043431055, + "grad_norm": 10.494299983717905, + "learning_rate": 8.356336295381981e-06, + "loss": 0.6299, + "step": 5213 + }, + { + "epoch": 2.8306188925081432, + "grad_norm": 7.662506144803812, + "learning_rate": 8.352867388948223e-06, + "loss": 0.4432, + "step": 5214 + }, + { + "epoch": 2.8311617806731815, + "grad_norm": 11.250218590721435, + "learning_rate": 8.349398686234455e-06, + "loss": 0.5805, + "step": 5215 + }, + { + "epoch": 2.8317046688382193, + "grad_norm": 12.63080713053447, + "learning_rate": 8.345930187669685e-06, + "loss": 0.7236, + "step": 5216 + }, + { + "epoch": 2.8322475570032575, + "grad_norm": 12.902831533172366, + "learning_rate": 8.342461893682908e-06, + "loss": 0.6462, + "step": 5217 + }, + { + "epoch": 2.8327904451682953, + "grad_norm": 8.440334577273788, + "learning_rate": 8.338993804703082e-06, + "loss": 0.6557, + "step": 5218 + }, + { + "epoch": 2.8333333333333335, + "grad_norm": 8.946600771683327, + "learning_rate": 8.335525921159155e-06, + "loss": 0.4964, + "step": 5219 + }, + { + "epoch": 2.8338762214983713, + "grad_norm": 11.427720200073038, + "learning_rate": 8.332058243480032e-06, + "loss": 0.5832, + "step": 5220 + }, + { + "epoch": 2.8344191096634095, + "grad_norm": 10.663320470599537, + "learning_rate": 8.328590772094602e-06, + "loss": 0.385, + "step": 5221 + }, + { + "epoch": 2.8349619978284473, + "grad_norm": 10.091874752226794, + "learning_rate": 8.325123507431732e-06, + "loss": 0.8347, + "step": 5222 + }, + { + "epoch": 2.8355048859934855, + "grad_norm": 11.844959588854952, + "learning_rate": 8.321656449920252e-06, + "loss": 0.5577, + "step": 5223 + }, + { + "epoch": 2.8360477741585233, + "grad_norm": 10.895554793526724, + "learning_rate": 8.318189599988982e-06, + "loss": 0.7961, + "step": 5224 + }, + { + "epoch": 2.8365906623235615, + "grad_norm": 15.457894886801157, + "learning_rate": 8.314722958066699e-06, + "loss": 0.9784, + "step": 5225 + }, + { + "epoch": 2.8371335504885993, + "grad_norm": 9.287014113188532, + "learning_rate": 8.311256524582167e-06, + "loss": 0.5947, + "step": 5226 + }, + { + "epoch": 2.8376764386536375, + "grad_norm": 12.953808801302275, + "learning_rate": 8.307790299964119e-06, + "loss": 0.6073, + "step": 5227 + }, + { + "epoch": 2.8382193268186753, + "grad_norm": 12.53202947557571, + "learning_rate": 8.304324284641268e-06, + "loss": 0.8199, + "step": 5228 + }, + { + "epoch": 2.8387622149837135, + "grad_norm": 10.62038819495681, + "learning_rate": 8.300858479042291e-06, + "loss": 0.488, + "step": 5229 + }, + { + "epoch": 2.8393051031487513, + "grad_norm": 10.914681492949684, + "learning_rate": 8.29739288359584e-06, + "loss": 0.6159, + "step": 5230 + }, + { + "epoch": 2.8398479913137895, + "grad_norm": 9.514892864934003, + "learning_rate": 8.293927498730556e-06, + "loss": 0.8757, + "step": 5231 + }, + { + "epoch": 2.8403908794788273, + "grad_norm": 10.004218198768447, + "learning_rate": 8.290462324875036e-06, + "loss": 0.4543, + "step": 5232 + }, + { + "epoch": 2.8409337676438655, + "grad_norm": 7.955044675492623, + "learning_rate": 8.286997362457859e-06, + "loss": 0.3227, + "step": 5233 + }, + { + "epoch": 2.8414766558089033, + "grad_norm": 5.674531662978047, + "learning_rate": 8.283532611907577e-06, + "loss": 0.2674, + "step": 5234 + }, + { + "epoch": 2.8420195439739415, + "grad_norm": 10.825015665199876, + "learning_rate": 8.280068073652723e-06, + "loss": 0.7373, + "step": 5235 + }, + { + "epoch": 2.8425624321389793, + "grad_norm": 11.937523350465815, + "learning_rate": 8.276603748121786e-06, + "loss": 0.6019, + "step": 5236 + }, + { + "epoch": 2.8431053203040175, + "grad_norm": 11.660316243586271, + "learning_rate": 8.273139635743249e-06, + "loss": 0.8063, + "step": 5237 + }, + { + "epoch": 2.8436482084690553, + "grad_norm": 10.978384303044228, + "learning_rate": 8.269675736945547e-06, + "loss": 0.6178, + "step": 5238 + }, + { + "epoch": 2.8441910966340935, + "grad_norm": 10.489917172678325, + "learning_rate": 8.266212052157113e-06, + "loss": 0.4382, + "step": 5239 + }, + { + "epoch": 2.8447339847991313, + "grad_norm": 9.430697797795169, + "learning_rate": 8.262748581806336e-06, + "loss": 0.4811, + "step": 5240 + }, + { + "epoch": 2.8452768729641695, + "grad_norm": 13.552717272826865, + "learning_rate": 8.259285326321579e-06, + "loss": 0.6866, + "step": 5241 + }, + { + "epoch": 2.8458197611292073, + "grad_norm": 12.883094545932988, + "learning_rate": 8.25582228613119e-06, + "loss": 0.9637, + "step": 5242 + }, + { + "epoch": 2.8463626492942455, + "grad_norm": 18.04770906924292, + "learning_rate": 8.25235946166348e-06, + "loss": 0.801, + "step": 5243 + }, + { + "epoch": 2.8469055374592833, + "grad_norm": 12.482559439953985, + "learning_rate": 8.248896853346738e-06, + "loss": 0.7968, + "step": 5244 + }, + { + "epoch": 2.8474484256243215, + "grad_norm": 10.728211626455765, + "learning_rate": 8.245434461609221e-06, + "loss": 0.5192, + "step": 5245 + }, + { + "epoch": 2.8479913137893593, + "grad_norm": 10.793312471383818, + "learning_rate": 8.241972286879168e-06, + "loss": 0.6121, + "step": 5246 + }, + { + "epoch": 2.8485342019543975, + "grad_norm": 12.822804552881774, + "learning_rate": 8.238510329584782e-06, + "loss": 0.7807, + "step": 5247 + }, + { + "epoch": 2.8490770901194353, + "grad_norm": 14.208687960885756, + "learning_rate": 8.23504859015425e-06, + "loss": 0.73, + "step": 5248 + }, + { + "epoch": 2.8496199782844736, + "grad_norm": 13.561299495458854, + "learning_rate": 8.231587069015723e-06, + "loss": 1.269, + "step": 5249 + }, + { + "epoch": 2.8501628664495113, + "grad_norm": 12.542912517352129, + "learning_rate": 8.22812576659732e-06, + "loss": 0.5758, + "step": 5250 + }, + { + "epoch": 2.8507057546145496, + "grad_norm": 9.184884000564415, + "learning_rate": 8.22466468332715e-06, + "loss": 0.5323, + "step": 5251 + }, + { + "epoch": 2.8512486427795873, + "grad_norm": 10.234457404263107, + "learning_rate": 8.22120381963328e-06, + "loss": 0.5455, + "step": 5252 + }, + { + "epoch": 2.8517915309446256, + "grad_norm": 13.327787107696029, + "learning_rate": 8.217743175943756e-06, + "loss": 0.6097, + "step": 5253 + }, + { + "epoch": 2.8523344191096633, + "grad_norm": 10.375306748712186, + "learning_rate": 8.214282752686595e-06, + "loss": 0.9622, + "step": 5254 + }, + { + "epoch": 2.8528773072747016, + "grad_norm": 10.445896680603816, + "learning_rate": 8.210822550289794e-06, + "loss": 0.6347, + "step": 5255 + }, + { + "epoch": 2.8534201954397393, + "grad_norm": 8.375912125960554, + "learning_rate": 8.207362569181305e-06, + "loss": 0.419, + "step": 5256 + }, + { + "epoch": 2.8539630836047776, + "grad_norm": 12.786819256474223, + "learning_rate": 8.203902809789078e-06, + "loss": 1.0155, + "step": 5257 + }, + { + "epoch": 2.8545059717698154, + "grad_norm": 9.916362057542294, + "learning_rate": 8.200443272541007e-06, + "loss": 0.5501, + "step": 5258 + }, + { + "epoch": 2.8550488599348536, + "grad_norm": 12.303081851739867, + "learning_rate": 8.196983957864984e-06, + "loss": 0.5965, + "step": 5259 + }, + { + "epoch": 2.8555917480998914, + "grad_norm": 11.992761200744544, + "learning_rate": 8.19352486618886e-06, + "loss": 0.4804, + "step": 5260 + }, + { + "epoch": 2.8561346362649296, + "grad_norm": 11.342631954988665, + "learning_rate": 8.190065997940455e-06, + "loss": 0.6241, + "step": 5261 + }, + { + "epoch": 2.8566775244299674, + "grad_norm": 14.06120317718869, + "learning_rate": 8.186607353547578e-06, + "loss": 1.3883, + "step": 5262 + }, + { + "epoch": 2.8572204125950056, + "grad_norm": 9.985876763230308, + "learning_rate": 8.183148933437988e-06, + "loss": 0.6843, + "step": 5263 + }, + { + "epoch": 2.8577633007600434, + "grad_norm": 13.658693433537799, + "learning_rate": 8.179690738039436e-06, + "loss": 1.027, + "step": 5264 + }, + { + "epoch": 2.8583061889250816, + "grad_norm": 13.070549937452688, + "learning_rate": 8.176232767779634e-06, + "loss": 0.9171, + "step": 5265 + }, + { + "epoch": 2.8588490770901194, + "grad_norm": 10.202999590557772, + "learning_rate": 8.17277502308627e-06, + "loss": 0.6412, + "step": 5266 + }, + { + "epoch": 2.8593919652551576, + "grad_norm": 11.358682840396053, + "learning_rate": 8.169317504387002e-06, + "loss": 0.8102, + "step": 5267 + }, + { + "epoch": 2.8599348534201954, + "grad_norm": 12.795930497437475, + "learning_rate": 8.165860212109467e-06, + "loss": 0.7885, + "step": 5268 + }, + { + "epoch": 2.8604777415852336, + "grad_norm": 12.540977748153804, + "learning_rate": 8.162403146681263e-06, + "loss": 0.453, + "step": 5269 + }, + { + "epoch": 2.8610206297502714, + "grad_norm": 9.835742357069565, + "learning_rate": 8.158946308529964e-06, + "loss": 0.553, + "step": 5270 + }, + { + "epoch": 2.8615635179153096, + "grad_norm": 8.768245847600197, + "learning_rate": 8.15548969808312e-06, + "loss": 0.6084, + "step": 5271 + }, + { + "epoch": 2.8621064060803474, + "grad_norm": 11.27517155767251, + "learning_rate": 8.152033315768248e-06, + "loss": 0.708, + "step": 5272 + }, + { + "epoch": 2.8626492942453856, + "grad_norm": 10.175542619256326, + "learning_rate": 8.148577162012848e-06, + "loss": 0.567, + "step": 5273 + }, + { + "epoch": 2.8631921824104234, + "grad_norm": 13.49575371946775, + "learning_rate": 8.145121237244367e-06, + "loss": 0.7178, + "step": 5274 + }, + { + "epoch": 2.8637350705754616, + "grad_norm": 12.848120779354371, + "learning_rate": 8.141665541890254e-06, + "loss": 0.7601, + "step": 5275 + }, + { + "epoch": 2.8642779587404994, + "grad_norm": 9.660459046348308, + "learning_rate": 8.138210076377903e-06, + "loss": 0.5867, + "step": 5276 + }, + { + "epoch": 2.8648208469055376, + "grad_norm": 12.014506663974018, + "learning_rate": 8.134754841134704e-06, + "loss": 0.6157, + "step": 5277 + }, + { + "epoch": 2.8653637350705754, + "grad_norm": 9.006817793845983, + "learning_rate": 8.131299836587994e-06, + "loss": 0.417, + "step": 5278 + }, + { + "epoch": 2.8659066232356136, + "grad_norm": 14.075391102924492, + "learning_rate": 8.127845063165097e-06, + "loss": 0.9453, + "step": 5279 + }, + { + "epoch": 2.8664495114006514, + "grad_norm": 11.384071845529691, + "learning_rate": 8.124390521293311e-06, + "loss": 0.757, + "step": 5280 + }, + { + "epoch": 2.8669923995656896, + "grad_norm": 11.412348591705651, + "learning_rate": 8.12093621139989e-06, + "loss": 0.7149, + "step": 5281 + }, + { + "epoch": 2.8675352877307274, + "grad_norm": 10.533279510688539, + "learning_rate": 8.117482133912078e-06, + "loss": 0.4981, + "step": 5282 + }, + { + "epoch": 2.8680781758957656, + "grad_norm": 10.800216526898991, + "learning_rate": 8.114028289257071e-06, + "loss": 0.6033, + "step": 5283 + }, + { + "epoch": 2.8686210640608034, + "grad_norm": 10.834462944873792, + "learning_rate": 8.110574677862054e-06, + "loss": 0.5554, + "step": 5284 + }, + { + "epoch": 2.8691639522258416, + "grad_norm": 13.02642177423764, + "learning_rate": 8.10712130015417e-06, + "loss": 0.7303, + "step": 5285 + }, + { + "epoch": 2.8697068403908794, + "grad_norm": 13.210484866681897, + "learning_rate": 8.103668156560543e-06, + "loss": 0.9579, + "step": 5286 + }, + { + "epoch": 2.8702497285559176, + "grad_norm": 7.866855822388311, + "learning_rate": 8.100215247508259e-06, + "loss": 0.5878, + "step": 5287 + }, + { + "epoch": 2.8707926167209554, + "grad_norm": 10.857793278838274, + "learning_rate": 8.096762573424384e-06, + "loss": 0.6734, + "step": 5288 + }, + { + "epoch": 2.8713355048859937, + "grad_norm": 11.519175454268337, + "learning_rate": 8.093310134735945e-06, + "loss": 0.7203, + "step": 5289 + }, + { + "epoch": 2.8718783930510314, + "grad_norm": 12.16192132511212, + "learning_rate": 8.089857931869947e-06, + "loss": 0.5887, + "step": 5290 + }, + { + "epoch": 2.8724212812160697, + "grad_norm": 8.545914428475232, + "learning_rate": 8.086405965253364e-06, + "loss": 0.677, + "step": 5291 + }, + { + "epoch": 2.8729641693811074, + "grad_norm": 12.463412130388328, + "learning_rate": 8.08295423531314e-06, + "loss": 0.8954, + "step": 5292 + }, + { + "epoch": 2.8735070575461457, + "grad_norm": 10.296132022290973, + "learning_rate": 8.079502742476195e-06, + "loss": 0.6582, + "step": 5293 + }, + { + "epoch": 2.8740499457111834, + "grad_norm": 9.170071716678386, + "learning_rate": 8.076051487169407e-06, + "loss": 0.5406, + "step": 5294 + }, + { + "epoch": 2.8745928338762217, + "grad_norm": 14.43641149914591, + "learning_rate": 8.072600469819643e-06, + "loss": 0.7459, + "step": 5295 + }, + { + "epoch": 2.8751357220412594, + "grad_norm": 9.705510058137687, + "learning_rate": 8.069149690853719e-06, + "loss": 0.6341, + "step": 5296 + }, + { + "epoch": 2.8756786102062977, + "grad_norm": 11.720382904656155, + "learning_rate": 8.065699150698442e-06, + "loss": 0.6016, + "step": 5297 + }, + { + "epoch": 2.8762214983713354, + "grad_norm": 14.838955909579008, + "learning_rate": 8.062248849780578e-06, + "loss": 0.8757, + "step": 5298 + }, + { + "epoch": 2.8767643865363732, + "grad_norm": 10.594386765695665, + "learning_rate": 8.05879878852686e-06, + "loss": 0.4864, + "step": 5299 + }, + { + "epoch": 2.8773072747014115, + "grad_norm": 9.83553313698661, + "learning_rate": 8.055348967364007e-06, + "loss": 0.5116, + "step": 5300 + }, + { + "epoch": 2.8778501628664497, + "grad_norm": 12.500251504139728, + "learning_rate": 8.05189938671869e-06, + "loss": 0.891, + "step": 5301 + }, + { + "epoch": 2.8783930510314875, + "grad_norm": 8.988571871064325, + "learning_rate": 8.048450047017563e-06, + "loss": 0.3699, + "step": 5302 + }, + { + "epoch": 2.8789359391965252, + "grad_norm": 10.893701604379793, + "learning_rate": 8.045000948687242e-06, + "loss": 0.5933, + "step": 5303 + }, + { + "epoch": 2.8794788273615635, + "grad_norm": 10.474072412748654, + "learning_rate": 8.041552092154321e-06, + "loss": 0.8787, + "step": 5304 + }, + { + "epoch": 2.8800217155266017, + "grad_norm": 8.555744151752272, + "learning_rate": 8.038103477845357e-06, + "loss": 0.3793, + "step": 5305 + }, + { + "epoch": 2.8805646036916395, + "grad_norm": 9.792363571958322, + "learning_rate": 8.034655106186884e-06, + "loss": 0.5347, + "step": 5306 + }, + { + "epoch": 2.8811074918566772, + "grad_norm": 8.810561156197549, + "learning_rate": 8.031206977605399e-06, + "loss": 0.5128, + "step": 5307 + }, + { + "epoch": 2.8816503800217155, + "grad_norm": 12.383361133227218, + "learning_rate": 8.027759092527374e-06, + "loss": 0.7834, + "step": 5308 + }, + { + "epoch": 2.8821932681867537, + "grad_norm": 9.51185389588206, + "learning_rate": 8.024311451379247e-06, + "loss": 0.4661, + "step": 5309 + }, + { + "epoch": 2.8827361563517915, + "grad_norm": 10.564107506597352, + "learning_rate": 8.020864054587426e-06, + "loss": 0.8199, + "step": 5310 + }, + { + "epoch": 2.8832790445168293, + "grad_norm": 14.579672874726727, + "learning_rate": 8.017416902578296e-06, + "loss": 1.1114, + "step": 5311 + }, + { + "epoch": 2.8838219326818675, + "grad_norm": 10.739540875321001, + "learning_rate": 8.013969995778201e-06, + "loss": 0.5384, + "step": 5312 + }, + { + "epoch": 2.8843648208469057, + "grad_norm": 11.955074827477672, + "learning_rate": 8.010523334613466e-06, + "loss": 0.8475, + "step": 5313 + }, + { + "epoch": 2.8849077090119435, + "grad_norm": 12.842414101367579, + "learning_rate": 8.007076919510371e-06, + "loss": 0.7096, + "step": 5314 + }, + { + "epoch": 2.8854505971769813, + "grad_norm": 9.17766217916239, + "learning_rate": 8.003630750895183e-06, + "loss": 0.4673, + "step": 5315 + }, + { + "epoch": 2.8859934853420195, + "grad_norm": 13.581449358552742, + "learning_rate": 8.000184829194121e-06, + "loss": 0.7501, + "step": 5316 + }, + { + "epoch": 2.8865363735070577, + "grad_norm": 6.157129192897698, + "learning_rate": 7.99673915483339e-06, + "loss": 0.2521, + "step": 5317 + }, + { + "epoch": 2.8870792616720955, + "grad_norm": 9.179432551374715, + "learning_rate": 7.993293728239154e-06, + "loss": 0.5682, + "step": 5318 + }, + { + "epoch": 2.8876221498371333, + "grad_norm": 8.458089104581813, + "learning_rate": 7.989848549837544e-06, + "loss": 0.7131, + "step": 5319 + }, + { + "epoch": 2.8881650380021715, + "grad_norm": 11.039294062728864, + "learning_rate": 7.986403620054674e-06, + "loss": 0.5421, + "step": 5320 + }, + { + "epoch": 2.8887079261672097, + "grad_norm": 8.296966992960693, + "learning_rate": 7.982958939316607e-06, + "loss": 0.5809, + "step": 5321 + }, + { + "epoch": 2.8892508143322475, + "grad_norm": 9.862357699114755, + "learning_rate": 7.979514508049398e-06, + "loss": 0.7253, + "step": 5322 + }, + { + "epoch": 2.8897937024972853, + "grad_norm": 10.60481802559805, + "learning_rate": 7.976070326679053e-06, + "loss": 0.5593, + "step": 5323 + }, + { + "epoch": 2.8903365906623235, + "grad_norm": 12.452210665582081, + "learning_rate": 7.972626395631556e-06, + "loss": 0.9987, + "step": 5324 + }, + { + "epoch": 2.8908794788273617, + "grad_norm": 8.161513982505882, + "learning_rate": 7.969182715332855e-06, + "loss": 0.5242, + "step": 5325 + }, + { + "epoch": 2.8914223669923995, + "grad_norm": 9.073053317988423, + "learning_rate": 7.965739286208878e-06, + "loss": 0.5015, + "step": 5326 + }, + { + "epoch": 2.8919652551574373, + "grad_norm": 8.632672314975157, + "learning_rate": 7.962296108685507e-06, + "loss": 0.4398, + "step": 5327 + }, + { + "epoch": 2.8925081433224755, + "grad_norm": 12.024378540292089, + "learning_rate": 7.958853183188597e-06, + "loss": 0.7876, + "step": 5328 + }, + { + "epoch": 2.8930510314875137, + "grad_norm": 9.943181911450774, + "learning_rate": 7.955410510143982e-06, + "loss": 0.6408, + "step": 5329 + }, + { + "epoch": 2.8935939196525515, + "grad_norm": 11.580662527673274, + "learning_rate": 7.951968089977449e-06, + "loss": 0.6484, + "step": 5330 + }, + { + "epoch": 2.8941368078175893, + "grad_norm": 10.477418271397871, + "learning_rate": 7.948525923114773e-06, + "loss": 0.7197, + "step": 5331 + }, + { + "epoch": 2.8946796959826275, + "grad_norm": 8.959473056544534, + "learning_rate": 7.945084009981674e-06, + "loss": 0.5344, + "step": 5332 + }, + { + "epoch": 2.8952225841476658, + "grad_norm": 7.339268198721023, + "learning_rate": 7.941642351003867e-06, + "loss": 0.4373, + "step": 5333 + }, + { + "epoch": 2.8957654723127035, + "grad_norm": 11.710220973199451, + "learning_rate": 7.938200946607008e-06, + "loss": 0.7354, + "step": 5334 + }, + { + "epoch": 2.8963083604777413, + "grad_norm": 10.548034804633037, + "learning_rate": 7.934759797216744e-06, + "loss": 1.0912, + "step": 5335 + }, + { + "epoch": 2.8968512486427795, + "grad_norm": 7.427458411514844, + "learning_rate": 7.93131890325868e-06, + "loss": 0.4515, + "step": 5336 + }, + { + "epoch": 2.8973941368078178, + "grad_norm": 7.5336162811756475, + "learning_rate": 7.927878265158391e-06, + "loss": 0.4106, + "step": 5337 + }, + { + "epoch": 2.8979370249728555, + "grad_norm": 9.845844831585431, + "learning_rate": 7.924437883341424e-06, + "loss": 0.394, + "step": 5338 + }, + { + "epoch": 2.8984799131378933, + "grad_norm": 11.91533381613983, + "learning_rate": 7.920997758233282e-06, + "loss": 0.5863, + "step": 5339 + }, + { + "epoch": 2.8990228013029316, + "grad_norm": 7.229828747210386, + "learning_rate": 7.917557890259457e-06, + "loss": 0.5663, + "step": 5340 + }, + { + "epoch": 2.8995656894679698, + "grad_norm": 11.378354320604084, + "learning_rate": 7.914118279845385e-06, + "loss": 0.7622, + "step": 5341 + }, + { + "epoch": 2.9001085776330076, + "grad_norm": 7.591103634044025, + "learning_rate": 7.91067892741649e-06, + "loss": 0.3933, + "step": 5342 + }, + { + "epoch": 2.9006514657980453, + "grad_norm": 12.040725731955204, + "learning_rate": 7.907239833398154e-06, + "loss": 0.7143, + "step": 5343 + }, + { + "epoch": 2.9011943539630836, + "grad_norm": 14.022224272834306, + "learning_rate": 7.903800998215735e-06, + "loss": 0.7389, + "step": 5344 + }, + { + "epoch": 2.901737242128122, + "grad_norm": 9.316170523577583, + "learning_rate": 7.900362422294545e-06, + "loss": 0.6061, + "step": 5345 + }, + { + "epoch": 2.9022801302931596, + "grad_norm": 8.20692587330132, + "learning_rate": 7.896924106059882e-06, + "loss": 0.4652, + "step": 5346 + }, + { + "epoch": 2.9028230184581973, + "grad_norm": 12.086544028526333, + "learning_rate": 7.893486049936993e-06, + "loss": 0.6579, + "step": 5347 + }, + { + "epoch": 2.9033659066232356, + "grad_norm": 11.562435941032952, + "learning_rate": 7.89004825435111e-06, + "loss": 0.6619, + "step": 5348 + }, + { + "epoch": 2.903908794788274, + "grad_norm": 9.535770499152292, + "learning_rate": 7.88661071972742e-06, + "loss": 0.4646, + "step": 5349 + }, + { + "epoch": 2.9044516829533116, + "grad_norm": 9.216711023311518, + "learning_rate": 7.88317344649108e-06, + "loss": 0.8006, + "step": 5350 + }, + { + "epoch": 2.9049945711183494, + "grad_norm": 9.577156917577227, + "learning_rate": 7.879736435067228e-06, + "loss": 0.7352, + "step": 5351 + }, + { + "epoch": 2.9055374592833876, + "grad_norm": 7.559386359727195, + "learning_rate": 7.87629968588095e-06, + "loss": 0.3841, + "step": 5352 + }, + { + "epoch": 2.906080347448426, + "grad_norm": 13.704096139083338, + "learning_rate": 7.872863199357315e-06, + "loss": 0.992, + "step": 5353 + }, + { + "epoch": 2.9066232356134636, + "grad_norm": 9.988539472899605, + "learning_rate": 7.869426975921344e-06, + "loss": 0.455, + "step": 5354 + }, + { + "epoch": 2.9071661237785014, + "grad_norm": 11.253082081483269, + "learning_rate": 7.865991015998043e-06, + "loss": 0.8226, + "step": 5355 + }, + { + "epoch": 2.9077090119435396, + "grad_norm": 10.526429101674665, + "learning_rate": 7.862555320012373e-06, + "loss": 0.5357, + "step": 5356 + }, + { + "epoch": 2.908251900108578, + "grad_norm": 9.646031093569851, + "learning_rate": 7.859119888389271e-06, + "loss": 0.4366, + "step": 5357 + }, + { + "epoch": 2.9087947882736156, + "grad_norm": 11.695549963620051, + "learning_rate": 7.855684721553633e-06, + "loss": 0.5834, + "step": 5358 + }, + { + "epoch": 2.9093376764386534, + "grad_norm": 13.315809341009361, + "learning_rate": 7.852249819930324e-06, + "loss": 1.0645, + "step": 5359 + }, + { + "epoch": 2.9098805646036916, + "grad_norm": 12.349221876871807, + "learning_rate": 7.848815183944183e-06, + "loss": 0.577, + "step": 5360 + }, + { + "epoch": 2.91042345276873, + "grad_norm": 16.63305773978975, + "learning_rate": 7.845380814020004e-06, + "loss": 0.9412, + "step": 5361 + }, + { + "epoch": 2.9109663409337676, + "grad_norm": 8.450548593743887, + "learning_rate": 7.841946710582563e-06, + "loss": 0.4955, + "step": 5362 + }, + { + "epoch": 2.9115092290988054, + "grad_norm": 11.465063371310775, + "learning_rate": 7.838512874056589e-06, + "loss": 0.6399, + "step": 5363 + }, + { + "epoch": 2.9120521172638436, + "grad_norm": 8.712777448735437, + "learning_rate": 7.835079304866793e-06, + "loss": 0.309, + "step": 5364 + }, + { + "epoch": 2.912595005428882, + "grad_norm": 9.44188587266759, + "learning_rate": 7.831646003437835e-06, + "loss": 0.435, + "step": 5365 + }, + { + "epoch": 2.9131378935939196, + "grad_norm": 12.094488103222616, + "learning_rate": 7.828212970194357e-06, + "loss": 0.7245, + "step": 5366 + }, + { + "epoch": 2.9136807817589574, + "grad_norm": 11.219679589625631, + "learning_rate": 7.82478020556096e-06, + "loss": 0.8716, + "step": 5367 + }, + { + "epoch": 2.9142236699239956, + "grad_norm": 10.091805149105353, + "learning_rate": 7.821347709962211e-06, + "loss": 0.553, + "step": 5368 + }, + { + "epoch": 2.914766558089034, + "grad_norm": 12.047544009676127, + "learning_rate": 7.817915483822654e-06, + "loss": 0.7283, + "step": 5369 + }, + { + "epoch": 2.9153094462540716, + "grad_norm": 10.891198338724942, + "learning_rate": 7.814483527566783e-06, + "loss": 0.6233, + "step": 5370 + }, + { + "epoch": 2.9158523344191094, + "grad_norm": 8.372829160348498, + "learning_rate": 7.811051841619077e-06, + "loss": 0.3781, + "step": 5371 + }, + { + "epoch": 2.9163952225841476, + "grad_norm": 7.8341866326850464, + "learning_rate": 7.807620426403964e-06, + "loss": 0.3455, + "step": 5372 + }, + { + "epoch": 2.916938110749186, + "grad_norm": 11.836979275353194, + "learning_rate": 7.804189282345855e-06, + "loss": 0.5573, + "step": 5373 + }, + { + "epoch": 2.9174809989142236, + "grad_norm": 8.195935708953526, + "learning_rate": 7.800758409869111e-06, + "loss": 0.4219, + "step": 5374 + }, + { + "epoch": 2.9180238870792614, + "grad_norm": 12.93514826751349, + "learning_rate": 7.797327809398074e-06, + "loss": 1.0471, + "step": 5375 + }, + { + "epoch": 2.9185667752442996, + "grad_norm": 9.384641490219252, + "learning_rate": 7.793897481357047e-06, + "loss": 0.5004, + "step": 5376 + }, + { + "epoch": 2.919109663409338, + "grad_norm": 10.40559875162228, + "learning_rate": 7.79046742617029e-06, + "loss": 0.5434, + "step": 5377 + }, + { + "epoch": 2.9196525515743756, + "grad_norm": 10.11339680443781, + "learning_rate": 7.787037644262048e-06, + "loss": 0.6003, + "step": 5378 + }, + { + "epoch": 2.9201954397394134, + "grad_norm": 12.37057409853331, + "learning_rate": 7.783608136056512e-06, + "loss": 0.685, + "step": 5379 + }, + { + "epoch": 2.9207383279044516, + "grad_norm": 11.589913705537441, + "learning_rate": 7.780178901977857e-06, + "loss": 0.7646, + "step": 5380 + }, + { + "epoch": 2.92128121606949, + "grad_norm": 12.462452174896484, + "learning_rate": 7.77674994245021e-06, + "loss": 0.6442, + "step": 5381 + }, + { + "epoch": 2.9218241042345277, + "grad_norm": 9.567483949166586, + "learning_rate": 7.77332125789768e-06, + "loss": 0.4645, + "step": 5382 + }, + { + "epoch": 2.9223669923995654, + "grad_norm": 11.079212789248007, + "learning_rate": 7.76989284874432e-06, + "loss": 0.7285, + "step": 5383 + }, + { + "epoch": 2.9229098805646037, + "grad_norm": 14.661322701022721, + "learning_rate": 7.76646471541417e-06, + "loss": 0.7481, + "step": 5384 + }, + { + "epoch": 2.923452768729642, + "grad_norm": 13.968723101600672, + "learning_rate": 7.763036858331222e-06, + "loss": 0.6632, + "step": 5385 + }, + { + "epoch": 2.9239956568946797, + "grad_norm": 12.145893524297147, + "learning_rate": 7.759609277919442e-06, + "loss": 0.8389, + "step": 5386 + }, + { + "epoch": 2.9245385450597174, + "grad_norm": 10.052846347427483, + "learning_rate": 7.756181974602757e-06, + "loss": 0.5351, + "step": 5387 + }, + { + "epoch": 2.9250814332247557, + "grad_norm": 11.030654536959574, + "learning_rate": 7.75275494880506e-06, + "loss": 0.6469, + "step": 5388 + }, + { + "epoch": 2.925624321389794, + "grad_norm": 13.670311382356902, + "learning_rate": 7.749328200950215e-06, + "loss": 0.984, + "step": 5389 + }, + { + "epoch": 2.9261672095548317, + "grad_norm": 10.129392847361466, + "learning_rate": 7.745901731462041e-06, + "loss": 0.8149, + "step": 5390 + }, + { + "epoch": 2.9267100977198695, + "grad_norm": 7.867905327161546, + "learning_rate": 7.742475540764339e-06, + "loss": 0.4957, + "step": 5391 + }, + { + "epoch": 2.9272529858849077, + "grad_norm": 8.467285225937513, + "learning_rate": 7.739049629280854e-06, + "loss": 0.579, + "step": 5392 + }, + { + "epoch": 2.927795874049946, + "grad_norm": 8.215885752516279, + "learning_rate": 7.735623997435318e-06, + "loss": 0.4073, + "step": 5393 + }, + { + "epoch": 2.9283387622149837, + "grad_norm": 8.51797242171681, + "learning_rate": 7.732198645651413e-06, + "loss": 0.63, + "step": 5394 + }, + { + "epoch": 2.9288816503800215, + "grad_norm": 15.898718458126094, + "learning_rate": 7.728773574352795e-06, + "loss": 0.7468, + "step": 5395 + }, + { + "epoch": 2.9294245385450597, + "grad_norm": 11.412421000549452, + "learning_rate": 7.725348783963084e-06, + "loss": 0.9522, + "step": 5396 + }, + { + "epoch": 2.929967426710098, + "grad_norm": 9.428473061604137, + "learning_rate": 7.721924274905855e-06, + "loss": 0.5388, + "step": 5397 + }, + { + "epoch": 2.9305103148751357, + "grad_norm": 10.083367258939811, + "learning_rate": 7.718500047604667e-06, + "loss": 0.6605, + "step": 5398 + }, + { + "epoch": 2.9310532030401735, + "grad_norm": 12.85805147473796, + "learning_rate": 7.715076102483026e-06, + "loss": 0.7407, + "step": 5399 + }, + { + "epoch": 2.9315960912052117, + "grad_norm": 9.866769374539361, + "learning_rate": 7.711652439964415e-06, + "loss": 0.6007, + "step": 5400 + }, + { + "epoch": 2.93213897937025, + "grad_norm": 12.367813318049626, + "learning_rate": 7.708229060472277e-06, + "loss": 0.5462, + "step": 5401 + }, + { + "epoch": 2.9326818675352877, + "grad_norm": 13.515093387837554, + "learning_rate": 7.704805964430023e-06, + "loss": 1.0205, + "step": 5402 + }, + { + "epoch": 2.9332247557003255, + "grad_norm": 10.667525494454107, + "learning_rate": 7.701383152261022e-06, + "loss": 0.7873, + "step": 5403 + }, + { + "epoch": 2.9337676438653637, + "grad_norm": 13.024983806475014, + "learning_rate": 7.697960624388621e-06, + "loss": 0.6388, + "step": 5404 + }, + { + "epoch": 2.934310532030402, + "grad_norm": 7.885758920671933, + "learning_rate": 7.694538381236115e-06, + "loss": 0.44, + "step": 5405 + }, + { + "epoch": 2.9348534201954397, + "grad_norm": 9.862140672970398, + "learning_rate": 7.69111642322678e-06, + "loss": 0.4065, + "step": 5406 + }, + { + "epoch": 2.9353963083604775, + "grad_norm": 11.605538896695489, + "learning_rate": 7.687694750783844e-06, + "loss": 0.6574, + "step": 5407 + }, + { + "epoch": 2.9359391965255157, + "grad_norm": 9.384935873492674, + "learning_rate": 7.684273364330505e-06, + "loss": 0.573, + "step": 5408 + }, + { + "epoch": 2.936482084690554, + "grad_norm": 7.506351724565119, + "learning_rate": 7.68085226428993e-06, + "loss": 0.4421, + "step": 5409 + }, + { + "epoch": 2.9370249728555917, + "grad_norm": 9.783370092521682, + "learning_rate": 7.677431451085238e-06, + "loss": 0.49, + "step": 5410 + }, + { + "epoch": 2.9375678610206295, + "grad_norm": 10.612737595311591, + "learning_rate": 7.674010925139533e-06, + "loss": 0.6855, + "step": 5411 + }, + { + "epoch": 2.9381107491856677, + "grad_norm": 10.77127188283207, + "learning_rate": 7.670590686875856e-06, + "loss": 0.4701, + "step": 5412 + }, + { + "epoch": 2.938653637350706, + "grad_norm": 9.009892989770995, + "learning_rate": 7.66717073671724e-06, + "loss": 0.3802, + "step": 5413 + }, + { + "epoch": 2.9391965255157437, + "grad_norm": 9.718396629861862, + "learning_rate": 7.663751075086662e-06, + "loss": 0.6433, + "step": 5414 + }, + { + "epoch": 2.9397394136807815, + "grad_norm": 8.544955706387817, + "learning_rate": 7.66033170240708e-06, + "loss": 0.3449, + "step": 5415 + }, + { + "epoch": 2.9402823018458197, + "grad_norm": 12.174166457717448, + "learning_rate": 7.656912619101401e-06, + "loss": 0.4685, + "step": 5416 + }, + { + "epoch": 2.940825190010858, + "grad_norm": 8.829471274310071, + "learning_rate": 7.6534938255925e-06, + "loss": 0.6213, + "step": 5417 + }, + { + "epoch": 2.9413680781758957, + "grad_norm": 11.089620150396728, + "learning_rate": 7.650075322303222e-06, + "loss": 0.5682, + "step": 5418 + }, + { + "epoch": 2.9419109663409335, + "grad_norm": 7.099888540087664, + "learning_rate": 7.646657109656375e-06, + "loss": 0.2937, + "step": 5419 + }, + { + "epoch": 2.9424538545059717, + "grad_norm": 9.8313977098716, + "learning_rate": 7.64323918807473e-06, + "loss": 0.6224, + "step": 5420 + }, + { + "epoch": 2.94299674267101, + "grad_norm": 14.892485492602605, + "learning_rate": 7.639821557981013e-06, + "loss": 1.0678, + "step": 5421 + }, + { + "epoch": 2.9435396308360477, + "grad_norm": 8.648910922847081, + "learning_rate": 7.636404219797931e-06, + "loss": 0.3605, + "step": 5422 + }, + { + "epoch": 2.9440825190010855, + "grad_norm": 12.093378946161085, + "learning_rate": 7.63298717394814e-06, + "loss": 0.9133, + "step": 5423 + }, + { + "epoch": 2.9446254071661238, + "grad_norm": 9.302461570715176, + "learning_rate": 7.629570420854271e-06, + "loss": 0.5305, + "step": 5424 + }, + { + "epoch": 2.945168295331162, + "grad_norm": 10.492242594764365, + "learning_rate": 7.626153960938909e-06, + "loss": 0.6007, + "step": 5425 + }, + { + "epoch": 2.9457111834961998, + "grad_norm": 11.212789857802488, + "learning_rate": 7.622737794624605e-06, + "loss": 0.4854, + "step": 5426 + }, + { + "epoch": 2.9462540716612375, + "grad_norm": 11.607518266622822, + "learning_rate": 7.619321922333884e-06, + "loss": 0.5543, + "step": 5427 + }, + { + "epoch": 2.9467969598262758, + "grad_norm": 11.268089998182159, + "learning_rate": 7.615906344489219e-06, + "loss": 0.6998, + "step": 5428 + }, + { + "epoch": 2.947339847991314, + "grad_norm": 12.383131636281371, + "learning_rate": 7.61249106151306e-06, + "loss": 0.8179, + "step": 5429 + }, + { + "epoch": 2.9478827361563518, + "grad_norm": 11.603914807740104, + "learning_rate": 7.609076073827808e-06, + "loss": 0.9006, + "step": 5430 + }, + { + "epoch": 2.9484256243213895, + "grad_norm": 9.674307468707637, + "learning_rate": 7.6056613818558415e-06, + "loss": 0.6638, + "step": 5431 + }, + { + "epoch": 2.9489685124864278, + "grad_norm": 13.055177059547185, + "learning_rate": 7.602246986019487e-06, + "loss": 0.9642, + "step": 5432 + }, + { + "epoch": 2.949511400651466, + "grad_norm": 11.00057075160912, + "learning_rate": 7.598832886741049e-06, + "loss": 0.4843, + "step": 5433 + }, + { + "epoch": 2.950054288816504, + "grad_norm": 11.15214171419394, + "learning_rate": 7.5954190844427834e-06, + "loss": 0.6722, + "step": 5434 + }, + { + "epoch": 2.9505971769815416, + "grad_norm": 8.91145682769725, + "learning_rate": 7.592005579546924e-06, + "loss": 0.6258, + "step": 5435 + }, + { + "epoch": 2.95114006514658, + "grad_norm": 12.685207726915342, + "learning_rate": 7.588592372475651e-06, + "loss": 0.8481, + "step": 5436 + }, + { + "epoch": 2.951682953311618, + "grad_norm": 13.10934163141668, + "learning_rate": 7.585179463651112e-06, + "loss": 0.6655, + "step": 5437 + }, + { + "epoch": 2.952225841476656, + "grad_norm": 10.993826657178786, + "learning_rate": 7.5817668534954295e-06, + "loss": 0.5687, + "step": 5438 + }, + { + "epoch": 2.9527687296416936, + "grad_norm": 8.424743777393193, + "learning_rate": 7.578354542430673e-06, + "loss": 0.5485, + "step": 5439 + }, + { + "epoch": 2.953311617806732, + "grad_norm": 14.730970588587084, + "learning_rate": 7.574942530878891e-06, + "loss": 0.8759, + "step": 5440 + }, + { + "epoch": 2.95385450597177, + "grad_norm": 8.667781572678388, + "learning_rate": 7.571530819262077e-06, + "loss": 0.3704, + "step": 5441 + }, + { + "epoch": 2.954397394136808, + "grad_norm": 6.739834901590041, + "learning_rate": 7.568119408002206e-06, + "loss": 0.3619, + "step": 5442 + }, + { + "epoch": 2.9549402823018456, + "grad_norm": 5.9221212094491795, + "learning_rate": 7.564708297521197e-06, + "loss": 0.3177, + "step": 5443 + }, + { + "epoch": 2.955483170466884, + "grad_norm": 9.008687743468037, + "learning_rate": 7.561297488240953e-06, + "loss": 0.5756, + "step": 5444 + }, + { + "epoch": 2.956026058631922, + "grad_norm": 10.49570531884228, + "learning_rate": 7.55788698058332e-06, + "loss": 0.609, + "step": 5445 + }, + { + "epoch": 2.95656894679696, + "grad_norm": 11.053313913817204, + "learning_rate": 7.554476774970115e-06, + "loss": 0.6893, + "step": 5446 + }, + { + "epoch": 2.9571118349619976, + "grad_norm": 10.747800237741414, + "learning_rate": 7.551066871823125e-06, + "loss": 0.6361, + "step": 5447 + }, + { + "epoch": 2.957654723127036, + "grad_norm": 12.158436457559148, + "learning_rate": 7.547657271564083e-06, + "loss": 0.6427, + "step": 5448 + }, + { + "epoch": 2.958197611292074, + "grad_norm": 11.021084054397756, + "learning_rate": 7.544247974614701e-06, + "loss": 0.4399, + "step": 5449 + }, + { + "epoch": 2.958740499457112, + "grad_norm": 8.337942068125022, + "learning_rate": 7.540838981396641e-06, + "loss": 0.3781, + "step": 5450 + }, + { + "epoch": 2.9592833876221496, + "grad_norm": 9.86638573791189, + "learning_rate": 7.537430292331536e-06, + "loss": 0.4257, + "step": 5451 + }, + { + "epoch": 2.959826275787188, + "grad_norm": 11.512553822962575, + "learning_rate": 7.534021907840975e-06, + "loss": 0.64, + "step": 5452 + }, + { + "epoch": 2.960369163952226, + "grad_norm": 10.540338627981921, + "learning_rate": 7.530613828346519e-06, + "loss": 0.6669, + "step": 5453 + }, + { + "epoch": 2.960912052117264, + "grad_norm": 12.092382555989715, + "learning_rate": 7.5272060542696765e-06, + "loss": 0.6346, + "step": 5454 + }, + { + "epoch": 2.9614549402823016, + "grad_norm": 11.429168527550765, + "learning_rate": 7.5237985860319344e-06, + "loss": 0.7454, + "step": 5455 + }, + { + "epoch": 2.96199782844734, + "grad_norm": 10.331080624952307, + "learning_rate": 7.52039142405473e-06, + "loss": 0.5472, + "step": 5456 + }, + { + "epoch": 2.962540716612378, + "grad_norm": 16.484391160691512, + "learning_rate": 7.516984568759461e-06, + "loss": 1.3039, + "step": 5457 + }, + { + "epoch": 2.963083604777416, + "grad_norm": 12.853254391289898, + "learning_rate": 7.5135780205675e-06, + "loss": 0.6846, + "step": 5458 + }, + { + "epoch": 2.9636264929424536, + "grad_norm": 12.062646480747443, + "learning_rate": 7.510171779900171e-06, + "loss": 0.5656, + "step": 5459 + }, + { + "epoch": 2.964169381107492, + "grad_norm": 8.21924446991787, + "learning_rate": 7.506765847178768e-06, + "loss": 0.3401, + "step": 5460 + }, + { + "epoch": 2.96471226927253, + "grad_norm": 11.116098821110214, + "learning_rate": 7.503360222824535e-06, + "loss": 0.4931, + "step": 5461 + }, + { + "epoch": 2.965255157437568, + "grad_norm": 9.759573410672228, + "learning_rate": 7.499954907258693e-06, + "loss": 0.5363, + "step": 5462 + }, + { + "epoch": 2.9657980456026056, + "grad_norm": 15.195018271421278, + "learning_rate": 7.496549900902408e-06, + "loss": 0.5349, + "step": 5463 + }, + { + "epoch": 2.966340933767644, + "grad_norm": 11.555768004386108, + "learning_rate": 7.493145204176823e-06, + "loss": 0.8273, + "step": 5464 + }, + { + "epoch": 2.966883821932682, + "grad_norm": 12.516961882115927, + "learning_rate": 7.4897408175030366e-06, + "loss": 0.7826, + "step": 5465 + }, + { + "epoch": 2.96742671009772, + "grad_norm": 11.67450620684471, + "learning_rate": 7.486336741302103e-06, + "loss": 0.5689, + "step": 5466 + }, + { + "epoch": 2.9679695982627576, + "grad_norm": 13.075852566862391, + "learning_rate": 7.48293297599505e-06, + "loss": 0.8355, + "step": 5467 + }, + { + "epoch": 2.968512486427796, + "grad_norm": 8.943584587537828, + "learning_rate": 7.479529522002855e-06, + "loss": 0.3921, + "step": 5468 + }, + { + "epoch": 2.969055374592834, + "grad_norm": 10.91179454178022, + "learning_rate": 7.476126379746471e-06, + "loss": 0.7276, + "step": 5469 + }, + { + "epoch": 2.969598262757872, + "grad_norm": 11.79182407811329, + "learning_rate": 7.472723549646793e-06, + "loss": 1.012, + "step": 5470 + }, + { + "epoch": 2.9701411509229096, + "grad_norm": 12.476871332438998, + "learning_rate": 7.469321032124697e-06, + "loss": 0.7994, + "step": 5471 + }, + { + "epoch": 2.970684039087948, + "grad_norm": 14.350758882559228, + "learning_rate": 7.465918827601008e-06, + "loss": 0.8477, + "step": 5472 + }, + { + "epoch": 2.971226927252986, + "grad_norm": 13.79963097595528, + "learning_rate": 7.46251693649652e-06, + "loss": 1.0897, + "step": 5473 + }, + { + "epoch": 2.971769815418024, + "grad_norm": 9.061867840853571, + "learning_rate": 7.459115359231977e-06, + "loss": 0.2972, + "step": 5474 + }, + { + "epoch": 2.9723127035830617, + "grad_norm": 11.193898477701971, + "learning_rate": 7.455714096228102e-06, + "loss": 0.9238, + "step": 5475 + }, + { + "epoch": 2.9728555917481, + "grad_norm": 13.156649778942683, + "learning_rate": 7.452313147905559e-06, + "loss": 0.9945, + "step": 5476 + }, + { + "epoch": 2.973398479913138, + "grad_norm": 10.673119691620487, + "learning_rate": 7.448912514684985e-06, + "loss": 0.6296, + "step": 5477 + }, + { + "epoch": 2.973941368078176, + "grad_norm": 13.208883224921669, + "learning_rate": 7.44551219698698e-06, + "loss": 0.5018, + "step": 5478 + }, + { + "epoch": 2.9744842562432137, + "grad_norm": 10.477928943580013, + "learning_rate": 7.442112195232093e-06, + "loss": 0.4782, + "step": 5479 + }, + { + "epoch": 2.975027144408252, + "grad_norm": 10.052094162546567, + "learning_rate": 7.438712509840852e-06, + "loss": 0.5661, + "step": 5480 + }, + { + "epoch": 2.97557003257329, + "grad_norm": 14.921523553916886, + "learning_rate": 7.435313141233724e-06, + "loss": 1.3182, + "step": 5481 + }, + { + "epoch": 2.976112920738328, + "grad_norm": 9.3079131640488, + "learning_rate": 7.43191408983116e-06, + "loss": 0.5672, + "step": 5482 + }, + { + "epoch": 2.9766558089033657, + "grad_norm": 8.76037752811841, + "learning_rate": 7.428515356053551e-06, + "loss": 0.6026, + "step": 5483 + }, + { + "epoch": 2.977198697068404, + "grad_norm": 11.037798492427186, + "learning_rate": 7.425116940321262e-06, + "loss": 0.7053, + "step": 5484 + }, + { + "epoch": 2.977741585233442, + "grad_norm": 12.329608171901635, + "learning_rate": 7.421718843054615e-06, + "loss": 0.6576, + "step": 5485 + }, + { + "epoch": 2.97828447339848, + "grad_norm": 14.664205153544405, + "learning_rate": 7.4183210646738875e-06, + "loss": 0.7546, + "step": 5486 + }, + { + "epoch": 2.9788273615635177, + "grad_norm": 16.28337626285995, + "learning_rate": 7.414923605599329e-06, + "loss": 1.2774, + "step": 5487 + }, + { + "epoch": 2.979370249728556, + "grad_norm": 11.936294560502736, + "learning_rate": 7.411526466251135e-06, + "loss": 0.9397, + "step": 5488 + }, + { + "epoch": 2.979913137893594, + "grad_norm": 15.443918342772207, + "learning_rate": 7.408129647049474e-06, + "loss": 0.7591, + "step": 5489 + }, + { + "epoch": 2.980456026058632, + "grad_norm": 9.303363471188012, + "learning_rate": 7.404733148414471e-06, + "loss": 0.4917, + "step": 5490 + }, + { + "epoch": 2.9809989142236697, + "grad_norm": 16.111589693524394, + "learning_rate": 7.401336970766207e-06, + "loss": 0.8724, + "step": 5491 + }, + { + "epoch": 2.981541802388708, + "grad_norm": 15.913107093535366, + "learning_rate": 7.397941114524727e-06, + "loss": 1.0915, + "step": 5492 + }, + { + "epoch": 2.982084690553746, + "grad_norm": 12.432387434616404, + "learning_rate": 7.3945455801100404e-06, + "loss": 1.1143, + "step": 5493 + }, + { + "epoch": 2.982627578718784, + "grad_norm": 11.65910861056599, + "learning_rate": 7.39115036794211e-06, + "loss": 0.7012, + "step": 5494 + }, + { + "epoch": 2.9831704668838217, + "grad_norm": 10.645886415744744, + "learning_rate": 7.387755478440855e-06, + "loss": 0.5327, + "step": 5495 + }, + { + "epoch": 2.98371335504886, + "grad_norm": 11.340492191845808, + "learning_rate": 7.384360912026167e-06, + "loss": 0.8393, + "step": 5496 + }, + { + "epoch": 2.984256243213898, + "grad_norm": 11.222712079595647, + "learning_rate": 7.38096666911789e-06, + "loss": 0.4731, + "step": 5497 + }, + { + "epoch": 2.984799131378936, + "grad_norm": 14.06199638971423, + "learning_rate": 7.377572750135833e-06, + "loss": 1.0083, + "step": 5498 + }, + { + "epoch": 2.9853420195439737, + "grad_norm": 11.609516661749161, + "learning_rate": 7.374179155499752e-06, + "loss": 0.5397, + "step": 5499 + }, + { + "epoch": 2.985884907709012, + "grad_norm": 11.809076211376697, + "learning_rate": 7.370785885629384e-06, + "loss": 0.5726, + "step": 5500 + }, + { + "epoch": 2.98642779587405, + "grad_norm": 12.767099228571903, + "learning_rate": 7.367392940944403e-06, + "loss": 0.8919, + "step": 5501 + }, + { + "epoch": 2.986970684039088, + "grad_norm": 8.511059646511987, + "learning_rate": 7.364000321864464e-06, + "loss": 0.4431, + "step": 5502 + }, + { + "epoch": 2.9875135722041257, + "grad_norm": 14.166198872023001, + "learning_rate": 7.360608028809161e-06, + "loss": 0.8709, + "step": 5503 + }, + { + "epoch": 2.988056460369164, + "grad_norm": 11.774867588892974, + "learning_rate": 7.357216062198066e-06, + "loss": 0.4857, + "step": 5504 + }, + { + "epoch": 2.988599348534202, + "grad_norm": 9.622067683785492, + "learning_rate": 7.353824422450702e-06, + "loss": 0.4385, + "step": 5505 + }, + { + "epoch": 2.98914223669924, + "grad_norm": 9.959533869172905, + "learning_rate": 7.3504331099865474e-06, + "loss": 0.6041, + "step": 5506 + }, + { + "epoch": 2.9896851248642777, + "grad_norm": 12.611515654376284, + "learning_rate": 7.347042125225052e-06, + "loss": 0.5667, + "step": 5507 + }, + { + "epoch": 2.990228013029316, + "grad_norm": 16.346151732068346, + "learning_rate": 7.343651468585611e-06, + "loss": 0.8017, + "step": 5508 + }, + { + "epoch": 2.990770901194354, + "grad_norm": 14.039933672423732, + "learning_rate": 7.340261140487593e-06, + "loss": 0.6905, + "step": 5509 + }, + { + "epoch": 2.991313789359392, + "grad_norm": 9.68037751098424, + "learning_rate": 7.336871141350313e-06, + "loss": 0.7553, + "step": 5510 + }, + { + "epoch": 2.9918566775244297, + "grad_norm": 11.6044953546391, + "learning_rate": 7.333481471593058e-06, + "loss": 0.7525, + "step": 5511 + }, + { + "epoch": 2.992399565689468, + "grad_norm": 10.257246759089158, + "learning_rate": 7.330092131635061e-06, + "loss": 0.5479, + "step": 5512 + }, + { + "epoch": 2.992942453854506, + "grad_norm": 9.37947259585779, + "learning_rate": 7.326703121895528e-06, + "loss": 0.6117, + "step": 5513 + }, + { + "epoch": 2.993485342019544, + "grad_norm": 10.344183596900065, + "learning_rate": 7.323314442793614e-06, + "loss": 0.5074, + "step": 5514 + }, + { + "epoch": 2.9940282301845818, + "grad_norm": 8.940259405187144, + "learning_rate": 7.319926094748433e-06, + "loss": 0.2863, + "step": 5515 + }, + { + "epoch": 2.99457111834962, + "grad_norm": 10.920924886109798, + "learning_rate": 7.316538078179065e-06, + "loss": 0.7706, + "step": 5516 + }, + { + "epoch": 2.995114006514658, + "grad_norm": 4.8493484123114445, + "learning_rate": 7.3131503935045424e-06, + "loss": 0.2066, + "step": 5517 + }, + { + "epoch": 2.995656894679696, + "grad_norm": 11.385458894978902, + "learning_rate": 7.309763041143865e-06, + "loss": 0.5706, + "step": 5518 + }, + { + "epoch": 2.9961997828447338, + "grad_norm": 8.263697834815542, + "learning_rate": 7.306376021515977e-06, + "loss": 0.667, + "step": 5519 + }, + { + "epoch": 2.996742671009772, + "grad_norm": 11.30709409496411, + "learning_rate": 7.302989335039801e-06, + "loss": 0.5522, + "step": 5520 + }, + { + "epoch": 2.99728555917481, + "grad_norm": 9.914253293704078, + "learning_rate": 7.299602982134199e-06, + "loss": 0.6109, + "step": 5521 + }, + { + "epoch": 2.997828447339848, + "grad_norm": 15.471798238376472, + "learning_rate": 7.2962169632180055e-06, + "loss": 0.8321, + "step": 5522 + }, + { + "epoch": 2.9983713355048858, + "grad_norm": 10.956384342902831, + "learning_rate": 7.292831278710007e-06, + "loss": 0.5791, + "step": 5523 + }, + { + "epoch": 2.998914223669924, + "grad_norm": 8.8527583386149, + "learning_rate": 7.289445929028951e-06, + "loss": 0.5807, + "step": 5524 + }, + { + "epoch": 2.999457111834962, + "grad_norm": 10.520949488289421, + "learning_rate": 7.286060914593544e-06, + "loss": 0.6861, + "step": 5525 + }, + { + "epoch": 3.0, + "grad_norm": 11.327416720286067, + "learning_rate": 7.282676235822444e-06, + "loss": 0.4657, + "step": 5526 + }, + { + "epoch": 3.000542888165038, + "grad_norm": 8.97355367199935, + "learning_rate": 7.279291893134283e-06, + "loss": 0.2931, + "step": 5527 + }, + { + "epoch": 3.001085776330076, + "grad_norm": 8.169786294777076, + "learning_rate": 7.275907886947632e-06, + "loss": 0.3065, + "step": 5528 + }, + { + "epoch": 3.001628664495114, + "grad_norm": 9.498228913363004, + "learning_rate": 7.272524217681036e-06, + "loss": 0.3704, + "step": 5529 + }, + { + "epoch": 3.002171552660152, + "grad_norm": 11.241888839923089, + "learning_rate": 7.269140885752992e-06, + "loss": 0.5917, + "step": 5530 + }, + { + "epoch": 3.00271444082519, + "grad_norm": 8.725213535150235, + "learning_rate": 7.26575789158196e-06, + "loss": 0.4699, + "step": 5531 + }, + { + "epoch": 3.003257328990228, + "grad_norm": 7.545852057856303, + "learning_rate": 7.262375235586343e-06, + "loss": 0.5126, + "step": 5532 + }, + { + "epoch": 3.003800217155266, + "grad_norm": 11.581979137257708, + "learning_rate": 7.2589929181845255e-06, + "loss": 0.4424, + "step": 5533 + }, + { + "epoch": 3.004343105320304, + "grad_norm": 9.70311467595886, + "learning_rate": 7.255610939794831e-06, + "loss": 0.4074, + "step": 5534 + }, + { + "epoch": 3.004885993485342, + "grad_norm": 7.016594112536396, + "learning_rate": 7.252229300835545e-06, + "loss": 0.2748, + "step": 5535 + }, + { + "epoch": 3.00542888165038, + "grad_norm": 11.131101403236986, + "learning_rate": 7.248848001724924e-06, + "loss": 0.4696, + "step": 5536 + }, + { + "epoch": 3.005971769815418, + "grad_norm": 9.402705826901306, + "learning_rate": 7.2454670428811625e-06, + "loss": 0.4554, + "step": 5537 + }, + { + "epoch": 3.006514657980456, + "grad_norm": 10.04584435406275, + "learning_rate": 7.242086424722432e-06, + "loss": 0.5017, + "step": 5538 + }, + { + "epoch": 3.007057546145494, + "grad_norm": 9.814079675077302, + "learning_rate": 7.238706147666843e-06, + "loss": 0.3366, + "step": 5539 + }, + { + "epoch": 3.007600434310532, + "grad_norm": 8.970687114442066, + "learning_rate": 7.235326212132483e-06, + "loss": 0.5007, + "step": 5540 + }, + { + "epoch": 3.00814332247557, + "grad_norm": 10.57886281786491, + "learning_rate": 7.23194661853738e-06, + "loss": 0.3796, + "step": 5541 + }, + { + "epoch": 3.008686210640608, + "grad_norm": 10.781975432954557, + "learning_rate": 7.228567367299532e-06, + "loss": 0.5567, + "step": 5542 + }, + { + "epoch": 3.009229098805646, + "grad_norm": 11.394439293433617, + "learning_rate": 7.2251884588368916e-06, + "loss": 0.521, + "step": 5543 + }, + { + "epoch": 3.009771986970684, + "grad_norm": 10.408091638796, + "learning_rate": 7.22180989356736e-06, + "loss": 0.4805, + "step": 5544 + }, + { + "epoch": 3.010314875135722, + "grad_norm": 11.389146042667898, + "learning_rate": 7.218431671908812e-06, + "loss": 0.49, + "step": 5545 + }, + { + "epoch": 3.01085776330076, + "grad_norm": 12.57922467984807, + "learning_rate": 7.215053794279066e-06, + "loss": 0.5101, + "step": 5546 + }, + { + "epoch": 3.011400651465798, + "grad_norm": 14.591941806963382, + "learning_rate": 7.211676261095905e-06, + "loss": 0.6942, + "step": 5547 + }, + { + "epoch": 3.011943539630836, + "grad_norm": 12.696865069952242, + "learning_rate": 7.20829907277707e-06, + "loss": 0.6091, + "step": 5548 + }, + { + "epoch": 3.012486427795874, + "grad_norm": 8.118652375371086, + "learning_rate": 7.204922229740255e-06, + "loss": 0.346, + "step": 5549 + }, + { + "epoch": 3.013029315960912, + "grad_norm": 10.942362893265422, + "learning_rate": 7.2015457324031105e-06, + "loss": 0.6571, + "step": 5550 + }, + { + "epoch": 3.01357220412595, + "grad_norm": 10.204677993952515, + "learning_rate": 7.198169581183254e-06, + "loss": 0.5563, + "step": 5551 + }, + { + "epoch": 3.014115092290988, + "grad_norm": 9.634479609863043, + "learning_rate": 7.194793776498245e-06, + "loss": 0.5768, + "step": 5552 + }, + { + "epoch": 3.014657980456026, + "grad_norm": 9.591830793887732, + "learning_rate": 7.191418318765616e-06, + "loss": 0.3283, + "step": 5553 + }, + { + "epoch": 3.015200868621064, + "grad_norm": 13.201493691800454, + "learning_rate": 7.188043208402844e-06, + "loss": 0.6985, + "step": 5554 + }, + { + "epoch": 3.015743756786102, + "grad_norm": 8.863552231556248, + "learning_rate": 7.184668445827367e-06, + "loss": 0.569, + "step": 5555 + }, + { + "epoch": 3.01628664495114, + "grad_norm": 16.6519621896966, + "learning_rate": 7.181294031456589e-06, + "loss": 0.5339, + "step": 5556 + }, + { + "epoch": 3.016829533116178, + "grad_norm": 10.16813004001019, + "learning_rate": 7.177919965707853e-06, + "loss": 0.4042, + "step": 5557 + }, + { + "epoch": 3.017372421281216, + "grad_norm": 9.578956705487652, + "learning_rate": 7.174546248998477e-06, + "loss": 0.454, + "step": 5558 + }, + { + "epoch": 3.017915309446254, + "grad_norm": 12.884511869394336, + "learning_rate": 7.1711728817457204e-06, + "loss": 0.769, + "step": 5559 + }, + { + "epoch": 3.018458197611292, + "grad_norm": 9.438716836740053, + "learning_rate": 7.1677998643668124e-06, + "loss": 0.3148, + "step": 5560 + }, + { + "epoch": 3.01900108577633, + "grad_norm": 10.176350005774413, + "learning_rate": 7.164427197278931e-06, + "loss": 0.4634, + "step": 5561 + }, + { + "epoch": 3.019543973941368, + "grad_norm": 13.390457057860667, + "learning_rate": 7.161054880899212e-06, + "loss": 0.8095, + "step": 5562 + }, + { + "epoch": 3.020086862106406, + "grad_norm": 15.075886108978443, + "learning_rate": 7.157682915644754e-06, + "loss": 0.4937, + "step": 5563 + }, + { + "epoch": 3.020629750271444, + "grad_norm": 10.999178702680004, + "learning_rate": 7.154311301932597e-06, + "loss": 0.6033, + "step": 5564 + }, + { + "epoch": 3.021172638436482, + "grad_norm": 11.230531078526122, + "learning_rate": 7.150940040179759e-06, + "loss": 0.4776, + "step": 5565 + }, + { + "epoch": 3.02171552660152, + "grad_norm": 10.476946027170913, + "learning_rate": 7.147569130803193e-06, + "loss": 0.3336, + "step": 5566 + }, + { + "epoch": 3.022258414766558, + "grad_norm": 11.753790547662213, + "learning_rate": 7.144198574219824e-06, + "loss": 0.5004, + "step": 5567 + }, + { + "epoch": 3.022801302931596, + "grad_norm": 8.64883980987418, + "learning_rate": 7.140828370846525e-06, + "loss": 0.3042, + "step": 5568 + }, + { + "epoch": 3.023344191096634, + "grad_norm": 13.493035832393069, + "learning_rate": 7.137458521100136e-06, + "loss": 0.5849, + "step": 5569 + }, + { + "epoch": 3.023887079261672, + "grad_norm": 13.029908881843049, + "learning_rate": 7.1340890253974324e-06, + "loss": 0.583, + "step": 5570 + }, + { + "epoch": 3.02442996742671, + "grad_norm": 11.852047436985602, + "learning_rate": 7.130719884155173e-06, + "loss": 0.4575, + "step": 5571 + }, + { + "epoch": 3.024972855591748, + "grad_norm": 13.476829762238244, + "learning_rate": 7.127351097790045e-06, + "loss": 0.5722, + "step": 5572 + }, + { + "epoch": 3.025515743756786, + "grad_norm": 11.710899279093313, + "learning_rate": 7.123982666718716e-06, + "loss": 0.5369, + "step": 5573 + }, + { + "epoch": 3.026058631921824, + "grad_norm": 12.982691203975197, + "learning_rate": 7.120614591357793e-06, + "loss": 0.4587, + "step": 5574 + }, + { + "epoch": 3.026601520086862, + "grad_norm": 13.494356693487626, + "learning_rate": 7.117246872123843e-06, + "loss": 0.5962, + "step": 5575 + }, + { + "epoch": 3.0271444082519, + "grad_norm": 9.838797633608028, + "learning_rate": 7.113879509433399e-06, + "loss": 0.3549, + "step": 5576 + }, + { + "epoch": 3.027687296416938, + "grad_norm": 15.816616895711462, + "learning_rate": 7.110512503702933e-06, + "loss": 0.6669, + "step": 5577 + }, + { + "epoch": 3.028230184581976, + "grad_norm": 11.513602729284349, + "learning_rate": 7.1071458553488904e-06, + "loss": 0.3573, + "step": 5578 + }, + { + "epoch": 3.028773072747014, + "grad_norm": 14.040376133978413, + "learning_rate": 7.103779564787654e-06, + "loss": 0.5898, + "step": 5579 + }, + { + "epoch": 3.029315960912052, + "grad_norm": 12.839234013609035, + "learning_rate": 7.10041363243558e-06, + "loss": 0.525, + "step": 5580 + }, + { + "epoch": 3.02985884907709, + "grad_norm": 12.773102345996788, + "learning_rate": 7.097048058708966e-06, + "loss": 0.7338, + "step": 5581 + }, + { + "epoch": 3.030401737242128, + "grad_norm": 9.757116728754855, + "learning_rate": 7.093682844024079e-06, + "loss": 0.4107, + "step": 5582 + }, + { + "epoch": 3.030944625407166, + "grad_norm": 13.284417042689952, + "learning_rate": 7.090317988797131e-06, + "loss": 0.4909, + "step": 5583 + }, + { + "epoch": 3.031487513572204, + "grad_norm": 12.858923373950708, + "learning_rate": 7.086953493444286e-06, + "loss": 0.4164, + "step": 5584 + }, + { + "epoch": 3.032030401737242, + "grad_norm": 11.460500988764892, + "learning_rate": 7.083589358381681e-06, + "loss": 0.6339, + "step": 5585 + }, + { + "epoch": 3.03257328990228, + "grad_norm": 11.76507179913234, + "learning_rate": 7.0802255840253865e-06, + "loss": 0.33, + "step": 5586 + }, + { + "epoch": 3.033116178067318, + "grad_norm": 14.990076792803963, + "learning_rate": 7.076862170791449e-06, + "loss": 0.5647, + "step": 5587 + }, + { + "epoch": 3.033659066232356, + "grad_norm": 12.953159557836509, + "learning_rate": 7.0734991190958545e-06, + "loss": 0.5641, + "step": 5588 + }, + { + "epoch": 3.034201954397394, + "grad_norm": 9.746562255040114, + "learning_rate": 7.070136429354557e-06, + "loss": 0.5755, + "step": 5589 + }, + { + "epoch": 3.034744842562432, + "grad_norm": 14.867296485205143, + "learning_rate": 7.066774101983452e-06, + "loss": 0.4915, + "step": 5590 + }, + { + "epoch": 3.03528773072747, + "grad_norm": 12.725313171891754, + "learning_rate": 7.063412137398402e-06, + "loss": 0.541, + "step": 5591 + }, + { + "epoch": 3.035830618892508, + "grad_norm": 11.023036821532774, + "learning_rate": 7.060050536015219e-06, + "loss": 0.3524, + "step": 5592 + }, + { + "epoch": 3.036373507057546, + "grad_norm": 9.747447295285967, + "learning_rate": 7.05668929824967e-06, + "loss": 0.4588, + "step": 5593 + }, + { + "epoch": 3.036916395222584, + "grad_norm": 12.477427859321583, + "learning_rate": 7.0533284245174826e-06, + "loss": 0.382, + "step": 5594 + }, + { + "epoch": 3.037459283387622, + "grad_norm": 9.62366763242101, + "learning_rate": 7.049967915234329e-06, + "loss": 0.2984, + "step": 5595 + }, + { + "epoch": 3.03800217155266, + "grad_norm": 13.368913470296512, + "learning_rate": 7.046607770815849e-06, + "loss": 0.6245, + "step": 5596 + }, + { + "epoch": 3.038545059717698, + "grad_norm": 10.380221393303437, + "learning_rate": 7.043247991677622e-06, + "loss": 0.4542, + "step": 5597 + }, + { + "epoch": 3.039087947882736, + "grad_norm": 9.778914279269939, + "learning_rate": 7.039888578235201e-06, + "loss": 0.3281, + "step": 5598 + }, + { + "epoch": 3.039630836047774, + "grad_norm": 7.360801727275878, + "learning_rate": 7.036529530904073e-06, + "loss": 0.2668, + "step": 5599 + }, + { + "epoch": 3.040173724212812, + "grad_norm": 7.962043569210736, + "learning_rate": 7.033170850099699e-06, + "loss": 0.2437, + "step": 5600 + }, + { + "epoch": 3.04071661237785, + "grad_norm": 13.434858118835157, + "learning_rate": 7.029812536237479e-06, + "loss": 0.5657, + "step": 5601 + }, + { + "epoch": 3.041259500542888, + "grad_norm": 12.720111358562143, + "learning_rate": 7.026454589732784e-06, + "loss": 0.5799, + "step": 5602 + }, + { + "epoch": 3.041802388707926, + "grad_norm": 13.271222462410066, + "learning_rate": 7.023097011000923e-06, + "loss": 0.4124, + "step": 5603 + }, + { + "epoch": 3.042345276872964, + "grad_norm": 12.713200398376557, + "learning_rate": 7.019739800457166e-06, + "loss": 0.5288, + "step": 5604 + }, + { + "epoch": 3.042888165038002, + "grad_norm": 8.938306532181711, + "learning_rate": 7.016382958516741e-06, + "loss": 0.3436, + "step": 5605 + }, + { + "epoch": 3.04343105320304, + "grad_norm": 11.284866978643858, + "learning_rate": 7.0130264855948245e-06, + "loss": 0.3379, + "step": 5606 + }, + { + "epoch": 3.043973941368078, + "grad_norm": 13.192559496466789, + "learning_rate": 7.009670382106558e-06, + "loss": 0.4862, + "step": 5607 + }, + { + "epoch": 3.044516829533116, + "grad_norm": 11.479680729723038, + "learning_rate": 7.006314648467021e-06, + "loss": 0.3483, + "step": 5608 + }, + { + "epoch": 3.045059717698154, + "grad_norm": 12.417526118207578, + "learning_rate": 7.002959285091262e-06, + "loss": 0.4803, + "step": 5609 + }, + { + "epoch": 3.045602605863192, + "grad_norm": 10.73078215472077, + "learning_rate": 6.999604292394271e-06, + "loss": 0.391, + "step": 5610 + }, + { + "epoch": 3.04614549402823, + "grad_norm": 11.713115986580751, + "learning_rate": 6.9962496707910085e-06, + "loss": 0.4674, + "step": 5611 + }, + { + "epoch": 3.046688382193268, + "grad_norm": 11.477819508258428, + "learning_rate": 6.992895420696371e-06, + "loss": 0.4974, + "step": 5612 + }, + { + "epoch": 3.047231270358306, + "grad_norm": 13.625151931379403, + "learning_rate": 6.989541542525218e-06, + "loss": 0.7687, + "step": 5613 + }, + { + "epoch": 3.047774158523344, + "grad_norm": 11.52719950465663, + "learning_rate": 6.986188036692369e-06, + "loss": 0.5544, + "step": 5614 + }, + { + "epoch": 3.048317046688382, + "grad_norm": 13.283507119767767, + "learning_rate": 6.982834903612584e-06, + "loss": 0.6122, + "step": 5615 + }, + { + "epoch": 3.04885993485342, + "grad_norm": 13.11709341600981, + "learning_rate": 6.979482143700591e-06, + "loss": 0.5184, + "step": 5616 + }, + { + "epoch": 3.049402823018458, + "grad_norm": 14.076144429789945, + "learning_rate": 6.976129757371055e-06, + "loss": 0.5594, + "step": 5617 + }, + { + "epoch": 3.049945711183496, + "grad_norm": 13.73222380193425, + "learning_rate": 6.9727777450386124e-06, + "loss": 0.4793, + "step": 5618 + }, + { + "epoch": 3.050488599348534, + "grad_norm": 10.519849404041729, + "learning_rate": 6.969426107117845e-06, + "loss": 0.3503, + "step": 5619 + }, + { + "epoch": 3.0510314875135722, + "grad_norm": 11.598423103761737, + "learning_rate": 6.966074844023284e-06, + "loss": 0.8233, + "step": 5620 + }, + { + "epoch": 3.05157437567861, + "grad_norm": 10.260836656524097, + "learning_rate": 6.9627239561694205e-06, + "loss": 0.4298, + "step": 5621 + }, + { + "epoch": 3.0521172638436482, + "grad_norm": 12.202759177421122, + "learning_rate": 6.9593734439707024e-06, + "loss": 0.5644, + "step": 5622 + }, + { + "epoch": 3.052660152008686, + "grad_norm": 21.73953611396027, + "learning_rate": 6.956023307841524e-06, + "loss": 0.9458, + "step": 5623 + }, + { + "epoch": 3.0532030401737242, + "grad_norm": 12.609994641445128, + "learning_rate": 6.952673548196229e-06, + "loss": 0.5591, + "step": 5624 + }, + { + "epoch": 3.053745928338762, + "grad_norm": 9.019788767418095, + "learning_rate": 6.949324165449129e-06, + "loss": 0.2823, + "step": 5625 + }, + { + "epoch": 3.0542888165038002, + "grad_norm": 9.597041690728307, + "learning_rate": 6.945975160014476e-06, + "loss": 0.339, + "step": 5626 + }, + { + "epoch": 3.054831704668838, + "grad_norm": 9.413655638291356, + "learning_rate": 6.9426265323064865e-06, + "loss": 0.3789, + "step": 5627 + }, + { + "epoch": 3.0553745928338762, + "grad_norm": 11.15027104482265, + "learning_rate": 6.939278282739317e-06, + "loss": 0.5436, + "step": 5628 + }, + { + "epoch": 3.055917480998914, + "grad_norm": 8.752949901728641, + "learning_rate": 6.935930411727091e-06, + "loss": 0.3561, + "step": 5629 + }, + { + "epoch": 3.0564603691639523, + "grad_norm": 13.044879197154156, + "learning_rate": 6.93258291968387e-06, + "loss": 0.5744, + "step": 5630 + }, + { + "epoch": 3.05700325732899, + "grad_norm": 8.940911398612299, + "learning_rate": 6.929235807023689e-06, + "loss": 0.3882, + "step": 5631 + }, + { + "epoch": 3.0575461454940283, + "grad_norm": 11.140820597123383, + "learning_rate": 6.925889074160512e-06, + "loss": 0.4042, + "step": 5632 + }, + { + "epoch": 3.058089033659066, + "grad_norm": 11.477795419029349, + "learning_rate": 6.9225427215082745e-06, + "loss": 0.531, + "step": 5633 + }, + { + "epoch": 3.0586319218241043, + "grad_norm": 11.606939647718864, + "learning_rate": 6.919196749480861e-06, + "loss": 0.3965, + "step": 5634 + }, + { + "epoch": 3.059174809989142, + "grad_norm": 12.112684503158684, + "learning_rate": 6.915851158492099e-06, + "loss": 0.3518, + "step": 5635 + }, + { + "epoch": 3.0597176981541803, + "grad_norm": 11.996768432451738, + "learning_rate": 6.9125059489557855e-06, + "loss": 0.4507, + "step": 5636 + }, + { + "epoch": 3.060260586319218, + "grad_norm": 11.43143134688868, + "learning_rate": 6.909161121285653e-06, + "loss": 0.3375, + "step": 5637 + }, + { + "epoch": 3.0608034744842563, + "grad_norm": 11.944059447932277, + "learning_rate": 6.9058166758954015e-06, + "loss": 0.5155, + "step": 5638 + }, + { + "epoch": 3.061346362649294, + "grad_norm": 13.048822571970177, + "learning_rate": 6.902472613198672e-06, + "loss": 0.4834, + "step": 5639 + }, + { + "epoch": 3.0618892508143323, + "grad_norm": 12.031888695640616, + "learning_rate": 6.899128933609071e-06, + "loss": 0.5558, + "step": 5640 + }, + { + "epoch": 3.06243213897937, + "grad_norm": 12.352798864366527, + "learning_rate": 6.8957856375401466e-06, + "loss": 0.6299, + "step": 5641 + }, + { + "epoch": 3.0629750271444083, + "grad_norm": 8.205311159067556, + "learning_rate": 6.892442725405396e-06, + "loss": 0.2726, + "step": 5642 + }, + { + "epoch": 3.063517915309446, + "grad_norm": 10.36891977803926, + "learning_rate": 6.8891001976182845e-06, + "loss": 0.3787, + "step": 5643 + }, + { + "epoch": 3.0640608034744843, + "grad_norm": 13.053698798672075, + "learning_rate": 6.88575805459222e-06, + "loss": 0.8244, + "step": 5644 + }, + { + "epoch": 3.064603691639522, + "grad_norm": 9.553838049974857, + "learning_rate": 6.88241629674056e-06, + "loss": 0.3611, + "step": 5645 + }, + { + "epoch": 3.0651465798045603, + "grad_norm": 13.221077748060404, + "learning_rate": 6.879074924476621e-06, + "loss": 0.3718, + "step": 5646 + }, + { + "epoch": 3.065689467969598, + "grad_norm": 13.072535493727939, + "learning_rate": 6.8757339382136735e-06, + "loss": 0.4338, + "step": 5647 + }, + { + "epoch": 3.0662323561346363, + "grad_norm": 13.8353344886429, + "learning_rate": 6.872393338364927e-06, + "loss": 0.4158, + "step": 5648 + }, + { + "epoch": 3.066775244299674, + "grad_norm": 12.713981935614036, + "learning_rate": 6.869053125343561e-06, + "loss": 0.4537, + "step": 5649 + }, + { + "epoch": 3.0673181324647123, + "grad_norm": 9.412074844790709, + "learning_rate": 6.865713299562691e-06, + "loss": 0.3842, + "step": 5650 + }, + { + "epoch": 3.06786102062975, + "grad_norm": 12.671904220941121, + "learning_rate": 6.862373861435397e-06, + "loss": 0.5343, + "step": 5651 + }, + { + "epoch": 3.0684039087947883, + "grad_norm": 8.312202499711958, + "learning_rate": 6.859034811374707e-06, + "loss": 0.4265, + "step": 5652 + }, + { + "epoch": 3.068946796959826, + "grad_norm": 8.89496055231579, + "learning_rate": 6.8556961497935925e-06, + "loss": 0.2747, + "step": 5653 + }, + { + "epoch": 3.0694896851248643, + "grad_norm": 12.131106832543901, + "learning_rate": 6.852357877104995e-06, + "loss": 0.4964, + "step": 5654 + }, + { + "epoch": 3.070032573289902, + "grad_norm": 11.860190856186797, + "learning_rate": 6.849019993721786e-06, + "loss": 0.6045, + "step": 5655 + }, + { + "epoch": 3.0705754614549403, + "grad_norm": 9.678033292318382, + "learning_rate": 6.845682500056811e-06, + "loss": 0.331, + "step": 5656 + }, + { + "epoch": 3.071118349619978, + "grad_norm": 15.602537064720222, + "learning_rate": 6.8423453965228485e-06, + "loss": 0.586, + "step": 5657 + }, + { + "epoch": 3.0716612377850163, + "grad_norm": 14.701485694410227, + "learning_rate": 6.839008683532641e-06, + "loss": 0.4876, + "step": 5658 + }, + { + "epoch": 3.072204125950054, + "grad_norm": 10.535230213477728, + "learning_rate": 6.835672361498875e-06, + "loss": 0.6123, + "step": 5659 + }, + { + "epoch": 3.0727470141150923, + "grad_norm": 11.131327768224557, + "learning_rate": 6.832336430834199e-06, + "loss": 0.6257, + "step": 5660 + }, + { + "epoch": 3.07328990228013, + "grad_norm": 8.624075967779987, + "learning_rate": 6.829000891951202e-06, + "loss": 0.3361, + "step": 5661 + }, + { + "epoch": 3.0738327904451683, + "grad_norm": 13.128895205370725, + "learning_rate": 6.825665745262424e-06, + "loss": 0.5965, + "step": 5662 + }, + { + "epoch": 3.074375678610206, + "grad_norm": 12.64008926373084, + "learning_rate": 6.822330991180368e-06, + "loss": 0.62, + "step": 5663 + }, + { + "epoch": 3.0749185667752443, + "grad_norm": 13.416120809361393, + "learning_rate": 6.8189966301174785e-06, + "loss": 0.6674, + "step": 5664 + }, + { + "epoch": 3.075461454940282, + "grad_norm": 9.203438718678973, + "learning_rate": 6.815662662486158e-06, + "loss": 0.3816, + "step": 5665 + }, + { + "epoch": 3.0760043431053203, + "grad_norm": 13.276570487788803, + "learning_rate": 6.81232908869875e-06, + "loss": 0.8179, + "step": 5666 + }, + { + "epoch": 3.076547231270358, + "grad_norm": 12.280157213200626, + "learning_rate": 6.808995909167566e-06, + "loss": 0.5559, + "step": 5667 + }, + { + "epoch": 3.0770901194353963, + "grad_norm": 10.175753421889684, + "learning_rate": 6.805663124304848e-06, + "loss": 0.6191, + "step": 5668 + }, + { + "epoch": 3.077633007600434, + "grad_norm": 12.523871003768551, + "learning_rate": 6.802330734522813e-06, + "loss": 0.4992, + "step": 5669 + }, + { + "epoch": 3.0781758957654723, + "grad_norm": 9.976975930711898, + "learning_rate": 6.798998740233602e-06, + "loss": 0.4104, + "step": 5670 + }, + { + "epoch": 3.07871878393051, + "grad_norm": 11.57486842983445, + "learning_rate": 6.795667141849333e-06, + "loss": 0.5803, + "step": 5671 + }, + { + "epoch": 3.0792616720955484, + "grad_norm": 13.330648489276534, + "learning_rate": 6.79233593978206e-06, + "loss": 0.5116, + "step": 5672 + }, + { + "epoch": 3.079804560260586, + "grad_norm": 10.454358932554449, + "learning_rate": 6.789005134443785e-06, + "loss": 0.4015, + "step": 5673 + }, + { + "epoch": 3.0803474484256244, + "grad_norm": 10.39956810217567, + "learning_rate": 6.785674726246477e-06, + "loss": 0.5832, + "step": 5674 + }, + { + "epoch": 3.080890336590662, + "grad_norm": 10.388603711886958, + "learning_rate": 6.782344715602038e-06, + "loss": 0.2995, + "step": 5675 + }, + { + "epoch": 3.0814332247557004, + "grad_norm": 14.02200163557337, + "learning_rate": 6.779015102922335e-06, + "loss": 0.5132, + "step": 5676 + }, + { + "epoch": 3.081976112920738, + "grad_norm": 9.590771798908678, + "learning_rate": 6.775685888619174e-06, + "loss": 0.2876, + "step": 5677 + }, + { + "epoch": 3.0825190010857764, + "grad_norm": 15.505233514983502, + "learning_rate": 6.7723570731043275e-06, + "loss": 0.8031, + "step": 5678 + }, + { + "epoch": 3.083061889250814, + "grad_norm": 11.180083592225891, + "learning_rate": 6.769028656789497e-06, + "loss": 0.5965, + "step": 5679 + }, + { + "epoch": 3.0836047774158524, + "grad_norm": 12.733678586969875, + "learning_rate": 6.765700640086356e-06, + "loss": 0.4956, + "step": 5680 + }, + { + "epoch": 3.08414766558089, + "grad_norm": 6.888856085950901, + "learning_rate": 6.762373023406515e-06, + "loss": 0.2367, + "step": 5681 + }, + { + "epoch": 3.0846905537459284, + "grad_norm": 12.334377562502302, + "learning_rate": 6.759045807161532e-06, + "loss": 0.5297, + "step": 5682 + }, + { + "epoch": 3.085233441910966, + "grad_norm": 10.637487438782616, + "learning_rate": 6.755718991762934e-06, + "loss": 0.4714, + "step": 5683 + }, + { + "epoch": 3.0857763300760044, + "grad_norm": 10.636835957652801, + "learning_rate": 6.7523925776221775e-06, + "loss": 0.5023, + "step": 5684 + }, + { + "epoch": 3.086319218241042, + "grad_norm": 9.416617539617917, + "learning_rate": 6.749066565150688e-06, + "loss": 0.4012, + "step": 5685 + }, + { + "epoch": 3.0868621064060804, + "grad_norm": 12.384873763197831, + "learning_rate": 6.745740954759823e-06, + "loss": 0.4588, + "step": 5686 + }, + { + "epoch": 3.087404994571118, + "grad_norm": 16.770871731703775, + "learning_rate": 6.7424157468609075e-06, + "loss": 0.8317, + "step": 5687 + }, + { + "epoch": 3.0879478827361564, + "grad_norm": 8.972498742892359, + "learning_rate": 6.7390909418652e-06, + "loss": 0.376, + "step": 5688 + }, + { + "epoch": 3.088490770901194, + "grad_norm": 12.96316859851228, + "learning_rate": 6.735766540183923e-06, + "loss": 0.5285, + "step": 5689 + }, + { + "epoch": 3.0890336590662324, + "grad_norm": 15.891256870281978, + "learning_rate": 6.732442542228245e-06, + "loss": 0.6191, + "step": 5690 + }, + { + "epoch": 3.08957654723127, + "grad_norm": 17.35289550101855, + "learning_rate": 6.729118948409278e-06, + "loss": 0.6878, + "step": 5691 + }, + { + "epoch": 3.0901194353963084, + "grad_norm": 9.022024416599917, + "learning_rate": 6.7257957591380965e-06, + "loss": 0.417, + "step": 5692 + }, + { + "epoch": 3.090662323561346, + "grad_norm": 14.27557655416772, + "learning_rate": 6.722472974825709e-06, + "loss": 0.5535, + "step": 5693 + }, + { + "epoch": 3.0912052117263844, + "grad_norm": 13.610102273496864, + "learning_rate": 6.7191505958830916e-06, + "loss": 0.542, + "step": 5694 + }, + { + "epoch": 3.091748099891422, + "grad_norm": 9.424378370747705, + "learning_rate": 6.715828622721154e-06, + "loss": 0.3447, + "step": 5695 + }, + { + "epoch": 3.0922909880564604, + "grad_norm": 10.815334224243674, + "learning_rate": 6.712507055750768e-06, + "loss": 0.5058, + "step": 5696 + }, + { + "epoch": 3.092833876221498, + "grad_norm": 9.390605474272599, + "learning_rate": 6.709185895382746e-06, + "loss": 0.3778, + "step": 5697 + }, + { + "epoch": 3.0933767643865364, + "grad_norm": 9.621100751873954, + "learning_rate": 6.705865142027863e-06, + "loss": 0.5128, + "step": 5698 + }, + { + "epoch": 3.093919652551574, + "grad_norm": 10.148509981020817, + "learning_rate": 6.7025447960968236e-06, + "loss": 0.3857, + "step": 5699 + }, + { + "epoch": 3.0944625407166124, + "grad_norm": 12.950847437294314, + "learning_rate": 6.699224858000305e-06, + "loss": 0.5364, + "step": 5700 + }, + { + "epoch": 3.09500542888165, + "grad_norm": 11.169096513096612, + "learning_rate": 6.695905328148914e-06, + "loss": 0.7072, + "step": 5701 + }, + { + "epoch": 3.0955483170466884, + "grad_norm": 9.231874354349005, + "learning_rate": 6.692586206953218e-06, + "loss": 0.5096, + "step": 5702 + }, + { + "epoch": 3.096091205211726, + "grad_norm": 13.629340417763906, + "learning_rate": 6.68926749482373e-06, + "loss": 0.5454, + "step": 5703 + }, + { + "epoch": 3.0966340933767644, + "grad_norm": 12.18004127970413, + "learning_rate": 6.685949192170913e-06, + "loss": 0.4463, + "step": 5704 + }, + { + "epoch": 3.097176981541802, + "grad_norm": 10.488473413353695, + "learning_rate": 6.682631299405187e-06, + "loss": 0.4041, + "step": 5705 + }, + { + "epoch": 3.0977198697068404, + "grad_norm": 9.624411116526867, + "learning_rate": 6.679313816936904e-06, + "loss": 0.5316, + "step": 5706 + }, + { + "epoch": 3.098262757871878, + "grad_norm": 13.907233176818153, + "learning_rate": 6.675996745176385e-06, + "loss": 0.6742, + "step": 5707 + }, + { + "epoch": 3.0988056460369164, + "grad_norm": 9.086606489202428, + "learning_rate": 6.672680084533882e-06, + "loss": 0.4101, + "step": 5708 + }, + { + "epoch": 3.099348534201954, + "grad_norm": 9.283349632521654, + "learning_rate": 6.669363835419611e-06, + "loss": 0.318, + "step": 5709 + }, + { + "epoch": 3.0998914223669924, + "grad_norm": 9.668872732852824, + "learning_rate": 6.666047998243732e-06, + "loss": 0.3352, + "step": 5710 + }, + { + "epoch": 3.1004343105320302, + "grad_norm": 9.343099296154486, + "learning_rate": 6.662732573416345e-06, + "loss": 0.4697, + "step": 5711 + }, + { + "epoch": 3.1009771986970684, + "grad_norm": 11.924198282652524, + "learning_rate": 6.659417561347516e-06, + "loss": 0.5743, + "step": 5712 + }, + { + "epoch": 3.1015200868621062, + "grad_norm": 8.560855249443614, + "learning_rate": 6.656102962447244e-06, + "loss": 0.3406, + "step": 5713 + }, + { + "epoch": 3.1020629750271445, + "grad_norm": 14.378777363618713, + "learning_rate": 6.652788777125488e-06, + "loss": 0.6911, + "step": 5714 + }, + { + "epoch": 3.1026058631921822, + "grad_norm": 10.29068899101719, + "learning_rate": 6.649475005792152e-06, + "loss": 0.4449, + "step": 5715 + }, + { + "epoch": 3.1031487513572205, + "grad_norm": 14.02995967568585, + "learning_rate": 6.646161648857085e-06, + "loss": 0.6167, + "step": 5716 + }, + { + "epoch": 3.1036916395222582, + "grad_norm": 9.518238552951749, + "learning_rate": 6.642848706730092e-06, + "loss": 0.3222, + "step": 5717 + }, + { + "epoch": 3.1042345276872965, + "grad_norm": 11.82162879354163, + "learning_rate": 6.639536179820922e-06, + "loss": 0.541, + "step": 5718 + }, + { + "epoch": 3.1047774158523342, + "grad_norm": 10.063084650592682, + "learning_rate": 6.636224068539271e-06, + "loss": 0.3842, + "step": 5719 + }, + { + "epoch": 3.1053203040173725, + "grad_norm": 10.94706260488087, + "learning_rate": 6.632912373294792e-06, + "loss": 0.4513, + "step": 5720 + }, + { + "epoch": 3.1058631921824102, + "grad_norm": 18.39843990173123, + "learning_rate": 6.6296010944970745e-06, + "loss": 0.6059, + "step": 5721 + }, + { + "epoch": 3.1064060803474485, + "grad_norm": 12.609614695754882, + "learning_rate": 6.6262902325556635e-06, + "loss": 0.5504, + "step": 5722 + }, + { + "epoch": 3.1069489685124863, + "grad_norm": 10.109861615464935, + "learning_rate": 6.6229797878800575e-06, + "loss": 0.4609, + "step": 5723 + }, + { + "epoch": 3.1074918566775245, + "grad_norm": 10.309200757998537, + "learning_rate": 6.61966976087969e-06, + "loss": 0.3823, + "step": 5724 + }, + { + "epoch": 3.1080347448425623, + "grad_norm": 13.21038451142413, + "learning_rate": 6.616360151963956e-06, + "loss": 1.0012, + "step": 5725 + }, + { + "epoch": 3.1085776330076005, + "grad_norm": 10.685970514335464, + "learning_rate": 6.613050961542189e-06, + "loss": 0.4587, + "step": 5726 + }, + { + "epoch": 3.1091205211726383, + "grad_norm": 10.001481339897822, + "learning_rate": 6.60974219002368e-06, + "loss": 0.4954, + "step": 5727 + }, + { + "epoch": 3.1096634093376765, + "grad_norm": 10.414474916866299, + "learning_rate": 6.6064338378176585e-06, + "loss": 0.5087, + "step": 5728 + }, + { + "epoch": 3.1102062975027143, + "grad_norm": 12.456658351321629, + "learning_rate": 6.603125905333307e-06, + "loss": 0.5004, + "step": 5729 + }, + { + "epoch": 3.1107491856677525, + "grad_norm": 11.228566931054628, + "learning_rate": 6.599818392979762e-06, + "loss": 0.6843, + "step": 5730 + }, + { + "epoch": 3.1112920738327903, + "grad_norm": 11.262471951903663, + "learning_rate": 6.596511301166093e-06, + "loss": 0.4458, + "step": 5731 + }, + { + "epoch": 3.1118349619978285, + "grad_norm": 14.452306212538934, + "learning_rate": 6.593204630301333e-06, + "loss": 0.4319, + "step": 5732 + }, + { + "epoch": 3.1123778501628663, + "grad_norm": 11.527781393494147, + "learning_rate": 6.589898380794451e-06, + "loss": 0.6469, + "step": 5733 + }, + { + "epoch": 3.1129207383279045, + "grad_norm": 13.390783514445054, + "learning_rate": 6.586592553054374e-06, + "loss": 0.5311, + "step": 5734 + }, + { + "epoch": 3.1134636264929423, + "grad_norm": 9.033138513383328, + "learning_rate": 6.583287147489968e-06, + "loss": 0.4243, + "step": 5735 + }, + { + "epoch": 3.1140065146579805, + "grad_norm": 8.69133142787827, + "learning_rate": 6.579982164510057e-06, + "loss": 0.3874, + "step": 5736 + }, + { + "epoch": 3.1145494028230183, + "grad_norm": 8.998980230350567, + "learning_rate": 6.576677604523399e-06, + "loss": 0.3375, + "step": 5737 + }, + { + "epoch": 3.1150922909880565, + "grad_norm": 13.718371009849239, + "learning_rate": 6.573373467938715e-06, + "loss": 0.8846, + "step": 5738 + }, + { + "epoch": 3.1156351791530943, + "grad_norm": 9.65489128269965, + "learning_rate": 6.5700697551646595e-06, + "loss": 0.5469, + "step": 5739 + }, + { + "epoch": 3.1161780673181325, + "grad_norm": 12.925637209034505, + "learning_rate": 6.566766466609844e-06, + "loss": 0.4983, + "step": 5740 + }, + { + "epoch": 3.1167209554831703, + "grad_norm": 8.77821937172237, + "learning_rate": 6.563463602682823e-06, + "loss": 0.3698, + "step": 5741 + }, + { + "epoch": 3.1172638436482085, + "grad_norm": 9.642465181954359, + "learning_rate": 6.560161163792098e-06, + "loss": 0.3764, + "step": 5742 + }, + { + "epoch": 3.1178067318132463, + "grad_norm": 13.798683831974408, + "learning_rate": 6.5568591503461266e-06, + "loss": 0.6993, + "step": 5743 + }, + { + "epoch": 3.1183496199782845, + "grad_norm": 9.872788283771353, + "learning_rate": 6.553557562753299e-06, + "loss": 0.3386, + "step": 5744 + }, + { + "epoch": 3.1188925081433223, + "grad_norm": 8.862249503192198, + "learning_rate": 6.550256401421969e-06, + "loss": 0.2179, + "step": 5745 + }, + { + "epoch": 3.1194353963083605, + "grad_norm": 12.08451508737181, + "learning_rate": 6.546955666760421e-06, + "loss": 0.6611, + "step": 5746 + }, + { + "epoch": 3.1199782844733983, + "grad_norm": 12.735515016235034, + "learning_rate": 6.543655359176904e-06, + "loss": 0.497, + "step": 5747 + }, + { + "epoch": 3.1205211726384365, + "grad_norm": 11.84865199124047, + "learning_rate": 6.5403554790795974e-06, + "loss": 0.7862, + "step": 5748 + }, + { + "epoch": 3.1210640608034743, + "grad_norm": 10.714296855189131, + "learning_rate": 6.53705602687664e-06, + "loss": 0.4053, + "step": 5749 + }, + { + "epoch": 3.1216069489685125, + "grad_norm": 13.603289352492634, + "learning_rate": 6.533757002976115e-06, + "loss": 0.8514, + "step": 5750 + }, + { + "epoch": 3.1221498371335503, + "grad_norm": 13.12224567132668, + "learning_rate": 6.530458407786045e-06, + "loss": 0.5521, + "step": 5751 + }, + { + "epoch": 3.1226927252985885, + "grad_norm": 12.175088945770037, + "learning_rate": 6.527160241714412e-06, + "loss": 0.5578, + "step": 5752 + }, + { + "epoch": 3.1232356134636263, + "grad_norm": 8.84737172176379, + "learning_rate": 6.52386250516913e-06, + "loss": 0.628, + "step": 5753 + }, + { + "epoch": 3.1237785016286646, + "grad_norm": 9.436585430499584, + "learning_rate": 6.520565198558076e-06, + "loss": 0.3497, + "step": 5754 + }, + { + "epoch": 3.1243213897937023, + "grad_norm": 19.037012868331093, + "learning_rate": 6.517268322289063e-06, + "loss": 0.7797, + "step": 5755 + }, + { + "epoch": 3.1248642779587406, + "grad_norm": 10.13012645392841, + "learning_rate": 6.513971876769857e-06, + "loss": 0.4892, + "step": 5756 + }, + { + "epoch": 3.1254071661237783, + "grad_norm": 12.402167588058873, + "learning_rate": 6.510675862408162e-06, + "loss": 0.4358, + "step": 5757 + }, + { + "epoch": 3.1259500542888166, + "grad_norm": 13.090215605994585, + "learning_rate": 6.5073802796116415e-06, + "loss": 0.4757, + "step": 5758 + }, + { + "epoch": 3.1264929424538543, + "grad_norm": 12.046620866150537, + "learning_rate": 6.504085128787892e-06, + "loss": 0.6104, + "step": 5759 + }, + { + "epoch": 3.1270358306188926, + "grad_norm": 13.860233354070779, + "learning_rate": 6.500790410344463e-06, + "loss": 0.5933, + "step": 5760 + }, + { + "epoch": 3.1275787187839303, + "grad_norm": 10.049188091595793, + "learning_rate": 6.497496124688858e-06, + "loss": 0.294, + "step": 5761 + }, + { + "epoch": 3.1281216069489686, + "grad_norm": 14.96534379979502, + "learning_rate": 6.494202272228509e-06, + "loss": 0.5983, + "step": 5762 + }, + { + "epoch": 3.1286644951140063, + "grad_norm": 12.09845857427764, + "learning_rate": 6.490908853370816e-06, + "loss": 0.6995, + "step": 5763 + }, + { + "epoch": 3.1292073832790446, + "grad_norm": 11.420498115973434, + "learning_rate": 6.487615868523105e-06, + "loss": 0.5189, + "step": 5764 + }, + { + "epoch": 3.1297502714440824, + "grad_norm": 9.433130521020503, + "learning_rate": 6.4843233180926646e-06, + "loss": 0.5145, + "step": 5765 + }, + { + "epoch": 3.1302931596091206, + "grad_norm": 13.10906497158192, + "learning_rate": 6.481031202486714e-06, + "loss": 0.4662, + "step": 5766 + }, + { + "epoch": 3.1308360477741584, + "grad_norm": 11.61717934785526, + "learning_rate": 6.477739522112436e-06, + "loss": 0.4496, + "step": 5767 + }, + { + "epoch": 3.1313789359391966, + "grad_norm": 15.125919718339924, + "learning_rate": 6.474448277376946e-06, + "loss": 0.6205, + "step": 5768 + }, + { + "epoch": 3.1319218241042344, + "grad_norm": 11.779008588612678, + "learning_rate": 6.471157468687315e-06, + "loss": 0.337, + "step": 5769 + }, + { + "epoch": 3.1324647122692726, + "grad_norm": 11.960250814404343, + "learning_rate": 6.467867096450553e-06, + "loss": 0.6284, + "step": 5770 + }, + { + "epoch": 3.1330076004343104, + "grad_norm": 11.356052401682772, + "learning_rate": 6.464577161073613e-06, + "loss": 0.5146, + "step": 5771 + }, + { + "epoch": 3.1335504885993486, + "grad_norm": 12.23671034812476, + "learning_rate": 6.461287662963407e-06, + "loss": 0.4118, + "step": 5772 + }, + { + "epoch": 3.1340933767643864, + "grad_norm": 9.3897356425462, + "learning_rate": 6.457998602526783e-06, + "loss": 0.3312, + "step": 5773 + }, + { + "epoch": 3.1346362649294246, + "grad_norm": 8.285183279593793, + "learning_rate": 6.454709980170537e-06, + "loss": 0.3857, + "step": 5774 + }, + { + "epoch": 3.1351791530944624, + "grad_norm": 10.70178891805582, + "learning_rate": 6.451421796301407e-06, + "loss": 0.4596, + "step": 5775 + }, + { + "epoch": 3.1357220412595006, + "grad_norm": 9.966009510827126, + "learning_rate": 6.448134051326092e-06, + "loss": 0.5223, + "step": 5776 + }, + { + "epoch": 3.1362649294245384, + "grad_norm": 11.165316341245944, + "learning_rate": 6.444846745651212e-06, + "loss": 0.482, + "step": 5777 + }, + { + "epoch": 3.1368078175895766, + "grad_norm": 10.091328730281877, + "learning_rate": 6.441559879683357e-06, + "loss": 0.4232, + "step": 5778 + }, + { + "epoch": 3.1373507057546144, + "grad_norm": 9.885329689068836, + "learning_rate": 6.438273453829047e-06, + "loss": 0.3008, + "step": 5779 + }, + { + "epoch": 3.1378935939196526, + "grad_norm": 11.500516853477965, + "learning_rate": 6.434987468494749e-06, + "loss": 0.5076, + "step": 5780 + }, + { + "epoch": 3.1384364820846904, + "grad_norm": 11.00413836223355, + "learning_rate": 6.431701924086887e-06, + "loss": 0.3688, + "step": 5781 + }, + { + "epoch": 3.1389793702497286, + "grad_norm": 11.998875560827074, + "learning_rate": 6.428416821011814e-06, + "loss": 0.6377, + "step": 5782 + }, + { + "epoch": 3.1395222584147664, + "grad_norm": 13.007472970217544, + "learning_rate": 6.425132159675845e-06, + "loss": 0.5721, + "step": 5783 + }, + { + "epoch": 3.1400651465798046, + "grad_norm": 11.356611803430905, + "learning_rate": 6.4218479404852244e-06, + "loss": 0.3753, + "step": 5784 + }, + { + "epoch": 3.1406080347448424, + "grad_norm": 12.12920282380304, + "learning_rate": 6.418564163846155e-06, + "loss": 0.4842, + "step": 5785 + }, + { + "epoch": 3.1411509229098806, + "grad_norm": 10.041071912027489, + "learning_rate": 6.415280830164779e-06, + "loss": 0.4939, + "step": 5786 + }, + { + "epoch": 3.1416938110749184, + "grad_norm": 12.998407630945819, + "learning_rate": 6.411997939847181e-06, + "loss": 0.636, + "step": 5787 + }, + { + "epoch": 3.1422366992399566, + "grad_norm": 9.655708839558747, + "learning_rate": 6.408715493299398e-06, + "loss": 0.4262, + "step": 5788 + }, + { + "epoch": 3.1427795874049944, + "grad_norm": 18.92187273336628, + "learning_rate": 6.405433490927404e-06, + "loss": 0.6645, + "step": 5789 + }, + { + "epoch": 3.1433224755700326, + "grad_norm": 7.288861789816135, + "learning_rate": 6.402151933137128e-06, + "loss": 0.3247, + "step": 5790 + }, + { + "epoch": 3.1438653637350704, + "grad_norm": 15.171324384191708, + "learning_rate": 6.398870820334431e-06, + "loss": 0.5835, + "step": 5791 + }, + { + "epoch": 3.1444082519001086, + "grad_norm": 19.537893777194515, + "learning_rate": 6.395590152925133e-06, + "loss": 0.4794, + "step": 5792 + }, + { + "epoch": 3.1449511400651464, + "grad_norm": 15.604738367299253, + "learning_rate": 6.3923099313149865e-06, + "loss": 0.8757, + "step": 5793 + }, + { + "epoch": 3.1454940282301846, + "grad_norm": 11.401596673156138, + "learning_rate": 6.389030155909702e-06, + "loss": 0.5278, + "step": 5794 + }, + { + "epoch": 3.1460369163952224, + "grad_norm": 15.026441679185012, + "learning_rate": 6.385750827114919e-06, + "loss": 0.7473, + "step": 5795 + }, + { + "epoch": 3.1465798045602607, + "grad_norm": 10.75641281038809, + "learning_rate": 6.382471945336238e-06, + "loss": 0.7022, + "step": 5796 + }, + { + "epoch": 3.1471226927252984, + "grad_norm": 13.14436770595673, + "learning_rate": 6.379193510979188e-06, + "loss": 0.5036, + "step": 5797 + }, + { + "epoch": 3.1476655808903367, + "grad_norm": 12.687239439350824, + "learning_rate": 6.375915524449259e-06, + "loss": 0.7799, + "step": 5798 + }, + { + "epoch": 3.1482084690553744, + "grad_norm": 12.06538329633808, + "learning_rate": 6.372637986151873e-06, + "loss": 0.6135, + "step": 5799 + }, + { + "epoch": 3.1487513572204127, + "grad_norm": 14.280970019829384, + "learning_rate": 6.369360896492398e-06, + "loss": 0.5631, + "step": 5800 + }, + { + "epoch": 3.1492942453854504, + "grad_norm": 13.49612526729136, + "learning_rate": 6.366084255876159e-06, + "loss": 0.8147, + "step": 5801 + }, + { + "epoch": 3.1498371335504887, + "grad_norm": 16.44633506143303, + "learning_rate": 6.3628080647084074e-06, + "loss": 0.705, + "step": 5802 + }, + { + "epoch": 3.1503800217155264, + "grad_norm": 11.659496492366037, + "learning_rate": 6.359532323394355e-06, + "loss": 0.4872, + "step": 5803 + }, + { + "epoch": 3.1509229098805647, + "grad_norm": 9.20526792646835, + "learning_rate": 6.356257032339141e-06, + "loss": 0.3279, + "step": 5804 + }, + { + "epoch": 3.1514657980456025, + "grad_norm": 15.869195168975136, + "learning_rate": 6.35298219194787e-06, + "loss": 0.6893, + "step": 5805 + }, + { + "epoch": 3.1520086862106407, + "grad_norm": 12.918024481861224, + "learning_rate": 6.349707802625569e-06, + "loss": 0.3561, + "step": 5806 + }, + { + "epoch": 3.1525515743756785, + "grad_norm": 11.948249403281372, + "learning_rate": 6.34643386477723e-06, + "loss": 0.6678, + "step": 5807 + }, + { + "epoch": 3.1530944625407167, + "grad_norm": 10.53882181570608, + "learning_rate": 6.3431603788077736e-06, + "loss": 0.2843, + "step": 5808 + }, + { + "epoch": 3.1536373507057545, + "grad_norm": 14.890522782673512, + "learning_rate": 6.3398873451220644e-06, + "loss": 0.7516, + "step": 5809 + }, + { + "epoch": 3.1541802388707927, + "grad_norm": 13.860961190955267, + "learning_rate": 6.3366147641249266e-06, + "loss": 0.697, + "step": 5810 + }, + { + "epoch": 3.1547231270358305, + "grad_norm": 13.048567845980694, + "learning_rate": 6.3333426362211096e-06, + "loss": 0.5362, + "step": 5811 + }, + { + "epoch": 3.1552660152008687, + "grad_norm": 11.430913085361082, + "learning_rate": 6.33007096181532e-06, + "loss": 0.3698, + "step": 5812 + }, + { + "epoch": 3.1558089033659065, + "grad_norm": 13.799412170578423, + "learning_rate": 6.326799741312202e-06, + "loss": 0.579, + "step": 5813 + }, + { + "epoch": 3.1563517915309447, + "grad_norm": 12.205698077175509, + "learning_rate": 6.323528975116349e-06, + "loss": 0.6956, + "step": 5814 + }, + { + "epoch": 3.1568946796959825, + "grad_norm": 14.923158817152583, + "learning_rate": 6.320258663632288e-06, + "loss": 0.7094, + "step": 5815 + }, + { + "epoch": 3.1574375678610207, + "grad_norm": 9.221494784749213, + "learning_rate": 6.316988807264506e-06, + "loss": 0.3346, + "step": 5816 + }, + { + "epoch": 3.1579804560260585, + "grad_norm": 12.263733681763519, + "learning_rate": 6.3137194064174124e-06, + "loss": 0.6395, + "step": 5817 + }, + { + "epoch": 3.1585233441910967, + "grad_norm": 12.198844326189183, + "learning_rate": 6.31045046149538e-06, + "loss": 0.4894, + "step": 5818 + }, + { + "epoch": 3.1590662323561345, + "grad_norm": 11.16927244530179, + "learning_rate": 6.3071819729027175e-06, + "loss": 0.4653, + "step": 5819 + }, + { + "epoch": 3.1596091205211727, + "grad_norm": 11.182448096771813, + "learning_rate": 6.303913941043669e-06, + "loss": 0.3997, + "step": 5820 + }, + { + "epoch": 3.1601520086862105, + "grad_norm": 9.56488246617197, + "learning_rate": 6.300646366322439e-06, + "loss": 0.6558, + "step": 5821 + }, + { + "epoch": 3.1606948968512487, + "grad_norm": 12.485736110882538, + "learning_rate": 6.297379249143158e-06, + "loss": 0.2747, + "step": 5822 + }, + { + "epoch": 3.1612377850162865, + "grad_norm": 10.29008763858223, + "learning_rate": 6.294112589909919e-06, + "loss": 0.3593, + "step": 5823 + }, + { + "epoch": 3.1617806731813247, + "grad_norm": 10.281965116706075, + "learning_rate": 6.2908463890267345e-06, + "loss": 0.4736, + "step": 5824 + }, + { + "epoch": 3.1623235613463625, + "grad_norm": 9.464092471296539, + "learning_rate": 6.287580646897584e-06, + "loss": 0.4445, + "step": 5825 + }, + { + "epoch": 3.1628664495114007, + "grad_norm": 17.834508294283445, + "learning_rate": 6.284315363926372e-06, + "loss": 0.7063, + "step": 5826 + }, + { + "epoch": 3.1634093376764385, + "grad_norm": 13.155791246426181, + "learning_rate": 6.281050540516963e-06, + "loss": 0.5295, + "step": 5827 + }, + { + "epoch": 3.1639522258414767, + "grad_norm": 11.87732424107669, + "learning_rate": 6.277786177073151e-06, + "loss": 0.6095, + "step": 5828 + }, + { + "epoch": 3.1644951140065145, + "grad_norm": 9.937871872571362, + "learning_rate": 6.274522273998671e-06, + "loss": 0.3788, + "step": 5829 + }, + { + "epoch": 3.1650380021715527, + "grad_norm": 10.26395895690379, + "learning_rate": 6.2712588316972155e-06, + "loss": 0.4352, + "step": 5830 + }, + { + "epoch": 3.1655808903365905, + "grad_norm": 8.627642026627477, + "learning_rate": 6.26799585057241e-06, + "loss": 0.3298, + "step": 5831 + }, + { + "epoch": 3.1661237785016287, + "grad_norm": 12.940542818393178, + "learning_rate": 6.2647333310278295e-06, + "loss": 0.5464, + "step": 5832 + }, + { + "epoch": 3.1666666666666665, + "grad_norm": 11.72079573230472, + "learning_rate": 6.261471273466979e-06, + "loss": 0.3785, + "step": 5833 + }, + { + "epoch": 3.1672095548317047, + "grad_norm": 13.338454285585954, + "learning_rate": 6.258209678293324e-06, + "loss": 0.4455, + "step": 5834 + }, + { + "epoch": 3.1677524429967425, + "grad_norm": 10.755191139797633, + "learning_rate": 6.254948545910256e-06, + "loss": 0.492, + "step": 5835 + }, + { + "epoch": 3.1682953311617807, + "grad_norm": 19.472452158150972, + "learning_rate": 6.251687876721124e-06, + "loss": 0.5031, + "step": 5836 + }, + { + "epoch": 3.1688382193268185, + "grad_norm": 17.265109190709392, + "learning_rate": 6.248427671129207e-06, + "loss": 0.7912, + "step": 5837 + }, + { + "epoch": 3.1693811074918568, + "grad_norm": 15.65566631740727, + "learning_rate": 6.245167929537733e-06, + "loss": 0.5431, + "step": 5838 + }, + { + "epoch": 3.1699239956568945, + "grad_norm": 11.65166282725162, + "learning_rate": 6.241908652349879e-06, + "loss": 0.3528, + "step": 5839 + }, + { + "epoch": 3.1704668838219328, + "grad_norm": 10.29071410809139, + "learning_rate": 6.238649839968746e-06, + "loss": 0.314, + "step": 5840 + }, + { + "epoch": 3.1710097719869705, + "grad_norm": 8.17516522370587, + "learning_rate": 6.2353914927974024e-06, + "loss": 0.3906, + "step": 5841 + }, + { + "epoch": 3.1715526601520088, + "grad_norm": 10.555776499314874, + "learning_rate": 6.232133611238833e-06, + "loss": 0.4222, + "step": 5842 + }, + { + "epoch": 3.1720955483170465, + "grad_norm": 15.841225391604809, + "learning_rate": 6.228876195695986e-06, + "loss": 0.8112, + "step": 5843 + }, + { + "epoch": 3.1726384364820848, + "grad_norm": 12.430872813466289, + "learning_rate": 6.225619246571741e-06, + "loss": 0.5456, + "step": 5844 + }, + { + "epoch": 3.1731813246471225, + "grad_norm": 13.424033486912041, + "learning_rate": 6.222362764268924e-06, + "loss": 0.5942, + "step": 5845 + }, + { + "epoch": 3.1737242128121608, + "grad_norm": 13.261228349891576, + "learning_rate": 6.219106749190298e-06, + "loss": 0.6697, + "step": 5846 + }, + { + "epoch": 3.1742671009771986, + "grad_norm": 12.710362718713364, + "learning_rate": 6.215851201738578e-06, + "loss": 0.5753, + "step": 5847 + }, + { + "epoch": 3.1748099891422368, + "grad_norm": 14.841367294849315, + "learning_rate": 6.212596122316415e-06, + "loss": 0.965, + "step": 5848 + }, + { + "epoch": 3.1753528773072746, + "grad_norm": 13.885344637636523, + "learning_rate": 6.209341511326394e-06, + "loss": 0.5931, + "step": 5849 + }, + { + "epoch": 3.175895765472313, + "grad_norm": 11.460331492201762, + "learning_rate": 6.20608736917106e-06, + "loss": 0.5989, + "step": 5850 + }, + { + "epoch": 3.1764386536373506, + "grad_norm": 11.70057563506978, + "learning_rate": 6.202833696252884e-06, + "loss": 0.4908, + "step": 5851 + }, + { + "epoch": 3.176981541802389, + "grad_norm": 10.645588299206919, + "learning_rate": 6.199580492974291e-06, + "loss": 0.4098, + "step": 5852 + }, + { + "epoch": 3.1775244299674266, + "grad_norm": 12.755952884154276, + "learning_rate": 6.196327759737637e-06, + "loss": 0.6992, + "step": 5853 + }, + { + "epoch": 3.178067318132465, + "grad_norm": 14.709812641950414, + "learning_rate": 6.193075496945231e-06, + "loss": 0.6841, + "step": 5854 + }, + { + "epoch": 3.1786102062975026, + "grad_norm": 13.875280510305055, + "learning_rate": 6.189823704999312e-06, + "loss": 0.7106, + "step": 5855 + }, + { + "epoch": 3.179153094462541, + "grad_norm": 9.281312738229682, + "learning_rate": 6.186572384302072e-06, + "loss": 0.3546, + "step": 5856 + }, + { + "epoch": 3.1796959826275786, + "grad_norm": 12.958826598950415, + "learning_rate": 6.1833215352556375e-06, + "loss": 0.6019, + "step": 5857 + }, + { + "epoch": 3.180238870792617, + "grad_norm": 11.464928690074025, + "learning_rate": 6.180071158262075e-06, + "loss": 0.408, + "step": 5858 + }, + { + "epoch": 3.1807817589576546, + "grad_norm": 9.768529641559107, + "learning_rate": 6.176821253723404e-06, + "loss": 0.4072, + "step": 5859 + }, + { + "epoch": 3.181324647122693, + "grad_norm": 9.678665348787685, + "learning_rate": 6.173571822041568e-06, + "loss": 0.4732, + "step": 5860 + }, + { + "epoch": 3.1818675352877306, + "grad_norm": 14.30027690416035, + "learning_rate": 6.170322863618474e-06, + "loss": 0.7642, + "step": 5861 + }, + { + "epoch": 3.182410423452769, + "grad_norm": 14.961370236846983, + "learning_rate": 6.167074378855946e-06, + "loss": 0.6362, + "step": 5862 + }, + { + "epoch": 3.1829533116178066, + "grad_norm": 9.357640920276147, + "learning_rate": 6.1638263681557695e-06, + "loss": 0.4049, + "step": 5863 + }, + { + "epoch": 3.183496199782845, + "grad_norm": 10.782208269160673, + "learning_rate": 6.160578831919662e-06, + "loss": 0.336, + "step": 5864 + }, + { + "epoch": 3.1840390879478826, + "grad_norm": 14.384452607422604, + "learning_rate": 6.157331770549285e-06, + "loss": 0.5396, + "step": 5865 + }, + { + "epoch": 3.184581976112921, + "grad_norm": 9.250100344453772, + "learning_rate": 6.154085184446237e-06, + "loss": 0.3785, + "step": 5866 + }, + { + "epoch": 3.1851248642779586, + "grad_norm": 15.376158526777193, + "learning_rate": 6.150839074012065e-06, + "loss": 0.5189, + "step": 5867 + }, + { + "epoch": 3.185667752442997, + "grad_norm": 10.177494395248477, + "learning_rate": 6.14759343964825e-06, + "loss": 0.3756, + "step": 5868 + }, + { + "epoch": 3.1862106406080346, + "grad_norm": 17.67659076253852, + "learning_rate": 6.144348281756218e-06, + "loss": 0.7381, + "step": 5869 + }, + { + "epoch": 3.186753528773073, + "grad_norm": 11.56583944116506, + "learning_rate": 6.141103600737337e-06, + "loss": 0.7712, + "step": 5870 + }, + { + "epoch": 3.1872964169381106, + "grad_norm": 15.096190908138054, + "learning_rate": 6.137859396992909e-06, + "loss": 0.6608, + "step": 5871 + }, + { + "epoch": 3.187839305103149, + "grad_norm": 14.055042562247136, + "learning_rate": 6.134615670924191e-06, + "loss": 0.6762, + "step": 5872 + }, + { + "epoch": 3.1883821932681866, + "grad_norm": 12.50506943616142, + "learning_rate": 6.131372422932363e-06, + "loss": 0.4721, + "step": 5873 + }, + { + "epoch": 3.188925081433225, + "grad_norm": 12.679042521649755, + "learning_rate": 6.128129653418562e-06, + "loss": 0.4277, + "step": 5874 + }, + { + "epoch": 3.1894679695982626, + "grad_norm": 13.177050387778413, + "learning_rate": 6.124887362783854e-06, + "loss": 0.7357, + "step": 5875 + }, + { + "epoch": 3.190010857763301, + "grad_norm": 9.589104761627354, + "learning_rate": 6.121645551429255e-06, + "loss": 0.3541, + "step": 5876 + }, + { + "epoch": 3.1905537459283386, + "grad_norm": 15.986950102257723, + "learning_rate": 6.118404219755717e-06, + "loss": 0.5148, + "step": 5877 + }, + { + "epoch": 3.191096634093377, + "grad_norm": 14.855627037048395, + "learning_rate": 6.1151633681641275e-06, + "loss": 0.6874, + "step": 5878 + }, + { + "epoch": 3.1916395222584146, + "grad_norm": 9.68244350868187, + "learning_rate": 6.111922997055328e-06, + "loss": 0.2941, + "step": 5879 + }, + { + "epoch": 3.192182410423453, + "grad_norm": 10.654011696718703, + "learning_rate": 6.108683106830085e-06, + "loss": 0.5009, + "step": 5880 + }, + { + "epoch": 3.1927252985884906, + "grad_norm": 12.855166875044116, + "learning_rate": 6.105443697889121e-06, + "loss": 0.4806, + "step": 5881 + }, + { + "epoch": 3.193268186753529, + "grad_norm": 10.436692542811286, + "learning_rate": 6.102204770633083e-06, + "loss": 0.5723, + "step": 5882 + }, + { + "epoch": 3.1938110749185666, + "grad_norm": 9.990036928653613, + "learning_rate": 6.098966325462574e-06, + "loss": 0.5778, + "step": 5883 + }, + { + "epoch": 3.194353963083605, + "grad_norm": 12.046567955894364, + "learning_rate": 6.095728362778125e-06, + "loss": 0.5015, + "step": 5884 + }, + { + "epoch": 3.1948968512486426, + "grad_norm": 8.725155173492308, + "learning_rate": 6.092490882980219e-06, + "loss": 0.3857, + "step": 5885 + }, + { + "epoch": 3.195439739413681, + "grad_norm": 11.013208621693263, + "learning_rate": 6.089253886469267e-06, + "loss": 0.6389, + "step": 5886 + }, + { + "epoch": 3.1959826275787186, + "grad_norm": 12.794704970363892, + "learning_rate": 6.086017373645624e-06, + "loss": 0.6362, + "step": 5887 + }, + { + "epoch": 3.196525515743757, + "grad_norm": 10.413523492938069, + "learning_rate": 6.082781344909592e-06, + "loss": 0.477, + "step": 5888 + }, + { + "epoch": 3.1970684039087947, + "grad_norm": 19.28502803064406, + "learning_rate": 6.079545800661404e-06, + "loss": 0.6435, + "step": 5889 + }, + { + "epoch": 3.197611292073833, + "grad_norm": 12.8419010375632, + "learning_rate": 6.076310741301244e-06, + "loss": 0.816, + "step": 5890 + }, + { + "epoch": 3.1981541802388707, + "grad_norm": 13.169440134578803, + "learning_rate": 6.0730761672292215e-06, + "loss": 0.4307, + "step": 5891 + }, + { + "epoch": 3.198697068403909, + "grad_norm": 12.487488215101045, + "learning_rate": 6.0698420788454e-06, + "loss": 0.6785, + "step": 5892 + }, + { + "epoch": 3.1992399565689467, + "grad_norm": 13.786560936789252, + "learning_rate": 6.066608476549771e-06, + "loss": 0.866, + "step": 5893 + }, + { + "epoch": 3.199782844733985, + "grad_norm": 14.699571921737691, + "learning_rate": 6.0633753607422785e-06, + "loss": 0.682, + "step": 5894 + }, + { + "epoch": 3.2003257328990227, + "grad_norm": 10.298137592669294, + "learning_rate": 6.060142731822792e-06, + "loss": 0.4765, + "step": 5895 + }, + { + "epoch": 3.200868621064061, + "grad_norm": 11.57501814659113, + "learning_rate": 6.056910590191132e-06, + "loss": 0.4907, + "step": 5896 + }, + { + "epoch": 3.2014115092290987, + "grad_norm": 10.861108561617078, + "learning_rate": 6.0536789362470575e-06, + "loss": 0.4201, + "step": 5897 + }, + { + "epoch": 3.201954397394137, + "grad_norm": 14.726234720260377, + "learning_rate": 6.050447770390258e-06, + "loss": 0.5714, + "step": 5898 + }, + { + "epoch": 3.2024972855591747, + "grad_norm": 10.353294495922004, + "learning_rate": 6.047217093020374e-06, + "loss": 0.3656, + "step": 5899 + }, + { + "epoch": 3.203040173724213, + "grad_norm": 7.811262042752878, + "learning_rate": 6.043986904536979e-06, + "loss": 0.3013, + "step": 5900 + }, + { + "epoch": 3.2035830618892507, + "grad_norm": 9.961324210163603, + "learning_rate": 6.040757205339589e-06, + "loss": 0.3498, + "step": 5901 + }, + { + "epoch": 3.204125950054289, + "grad_norm": 11.897050086158284, + "learning_rate": 6.037527995827659e-06, + "loss": 0.4876, + "step": 5902 + }, + { + "epoch": 3.2046688382193267, + "grad_norm": 10.075300527690798, + "learning_rate": 6.034299276400582e-06, + "loss": 0.3781, + "step": 5903 + }, + { + "epoch": 3.205211726384365, + "grad_norm": 20.471624027064, + "learning_rate": 6.031071047457689e-06, + "loss": 0.7781, + "step": 5904 + }, + { + "epoch": 3.2057546145494027, + "grad_norm": 13.759518261461583, + "learning_rate": 6.027843309398256e-06, + "loss": 0.6616, + "step": 5905 + }, + { + "epoch": 3.206297502714441, + "grad_norm": 11.734095364989608, + "learning_rate": 6.024616062621496e-06, + "loss": 0.4283, + "step": 5906 + }, + { + "epoch": 3.2068403908794787, + "grad_norm": 11.862203608845647, + "learning_rate": 6.02138930752655e-06, + "loss": 0.5662, + "step": 5907 + }, + { + "epoch": 3.207383279044517, + "grad_norm": 15.72669110441235, + "learning_rate": 6.018163044512521e-06, + "loss": 0.5579, + "step": 5908 + }, + { + "epoch": 3.2079261672095547, + "grad_norm": 19.31549988156055, + "learning_rate": 6.014937273978429e-06, + "loss": 0.7692, + "step": 5909 + }, + { + "epoch": 3.208469055374593, + "grad_norm": 10.302000372364747, + "learning_rate": 6.011711996323251e-06, + "loss": 0.5133, + "step": 5910 + }, + { + "epoch": 3.2090119435396307, + "grad_norm": 12.224649649591383, + "learning_rate": 6.008487211945884e-06, + "loss": 0.7677, + "step": 5911 + }, + { + "epoch": 3.209554831704669, + "grad_norm": 17.460170685666615, + "learning_rate": 6.005262921245185e-06, + "loss": 0.6836, + "step": 5912 + }, + { + "epoch": 3.2100977198697067, + "grad_norm": 19.301567461364495, + "learning_rate": 6.00203912461993e-06, + "loss": 0.9095, + "step": 5913 + }, + { + "epoch": 3.210640608034745, + "grad_norm": 10.852675563314923, + "learning_rate": 5.998815822468851e-06, + "loss": 0.5342, + "step": 5914 + }, + { + "epoch": 3.2111834961997827, + "grad_norm": 10.649138913539469, + "learning_rate": 5.995593015190608e-06, + "loss": 0.5665, + "step": 5915 + }, + { + "epoch": 3.211726384364821, + "grad_norm": 10.875890162606636, + "learning_rate": 5.992370703183803e-06, + "loss": 0.3634, + "step": 5916 + }, + { + "epoch": 3.2122692725298587, + "grad_norm": 16.640296114311138, + "learning_rate": 5.9891488868469775e-06, + "loss": 0.7804, + "step": 5917 + }, + { + "epoch": 3.212812160694897, + "grad_norm": 9.73069162630754, + "learning_rate": 5.9859275665786065e-06, + "loss": 0.35, + "step": 5918 + }, + { + "epoch": 3.2133550488599347, + "grad_norm": 10.657768777546748, + "learning_rate": 5.982706742777116e-06, + "loss": 0.529, + "step": 5919 + }, + { + "epoch": 3.213897937024973, + "grad_norm": 12.700699759951497, + "learning_rate": 5.9794864158408535e-06, + "loss": 0.5159, + "step": 5920 + }, + { + "epoch": 3.2144408251900107, + "grad_norm": 8.375283482104031, + "learning_rate": 5.976266586168121e-06, + "loss": 0.2818, + "step": 5921 + }, + { + "epoch": 3.214983713355049, + "grad_norm": 18.061811128959963, + "learning_rate": 5.973047254157148e-06, + "loss": 1.2271, + "step": 5922 + }, + { + "epoch": 3.2155266015200867, + "grad_norm": 10.630424335111323, + "learning_rate": 5.969828420206114e-06, + "loss": 0.3874, + "step": 5923 + }, + { + "epoch": 3.216069489685125, + "grad_norm": 11.324230234064501, + "learning_rate": 5.966610084713118e-06, + "loss": 0.6559, + "step": 5924 + }, + { + "epoch": 3.2166123778501627, + "grad_norm": 12.436990138030758, + "learning_rate": 5.9633922480762205e-06, + "loss": 0.6468, + "step": 5925 + }, + { + "epoch": 3.217155266015201, + "grad_norm": 16.354922341711003, + "learning_rate": 5.960174910693401e-06, + "loss": 0.8733, + "step": 5926 + }, + { + "epoch": 3.2176981541802387, + "grad_norm": 12.756426829873451, + "learning_rate": 5.956958072962586e-06, + "loss": 0.5399, + "step": 5927 + }, + { + "epoch": 3.218241042345277, + "grad_norm": 11.493661345744254, + "learning_rate": 5.953741735281641e-06, + "loss": 0.5487, + "step": 5928 + }, + { + "epoch": 3.2187839305103148, + "grad_norm": 11.431750755108203, + "learning_rate": 5.9505258980483645e-06, + "loss": 0.3997, + "step": 5929 + }, + { + "epoch": 3.219326818675353, + "grad_norm": 12.940367227326554, + "learning_rate": 5.947310561660503e-06, + "loss": 0.8178, + "step": 5930 + }, + { + "epoch": 3.2198697068403908, + "grad_norm": 11.421594805016625, + "learning_rate": 5.9440957265157265e-06, + "loss": 0.5015, + "step": 5931 + }, + { + "epoch": 3.220412595005429, + "grad_norm": 12.215300050799616, + "learning_rate": 5.940881393011659e-06, + "loss": 0.4044, + "step": 5932 + }, + { + "epoch": 3.2209554831704668, + "grad_norm": 11.265138511012573, + "learning_rate": 5.937667561545845e-06, + "loss": 0.6294, + "step": 5933 + }, + { + "epoch": 3.221498371335505, + "grad_norm": 11.883510126446058, + "learning_rate": 5.934454232515785e-06, + "loss": 0.5257, + "step": 5934 + }, + { + "epoch": 3.2220412595005428, + "grad_norm": 14.862260710742142, + "learning_rate": 5.931241406318906e-06, + "loss": 0.7435, + "step": 5935 + }, + { + "epoch": 3.222584147665581, + "grad_norm": 11.010381613248398, + "learning_rate": 5.92802908335257e-06, + "loss": 0.4852, + "step": 5936 + }, + { + "epoch": 3.2231270358306188, + "grad_norm": 10.894562800140209, + "learning_rate": 5.924817264014091e-06, + "loss": 0.4694, + "step": 5937 + }, + { + "epoch": 3.223669923995657, + "grad_norm": 9.91297174934936, + "learning_rate": 5.921605948700704e-06, + "loss": 0.4031, + "step": 5938 + }, + { + "epoch": 3.2242128121606948, + "grad_norm": 10.829816377096837, + "learning_rate": 5.918395137809596e-06, + "loss": 0.5139, + "step": 5939 + }, + { + "epoch": 3.224755700325733, + "grad_norm": 9.940894286293783, + "learning_rate": 5.915184831737881e-06, + "loss": 0.3475, + "step": 5940 + }, + { + "epoch": 3.225298588490771, + "grad_norm": 12.818677443045534, + "learning_rate": 5.911975030882617e-06, + "loss": 0.4293, + "step": 5941 + }, + { + "epoch": 3.225841476655809, + "grad_norm": 8.859395484195392, + "learning_rate": 5.908765735640795e-06, + "loss": 0.4065, + "step": 5942 + }, + { + "epoch": 3.226384364820847, + "grad_norm": 10.145624567540983, + "learning_rate": 5.90555694640935e-06, + "loss": 0.3713, + "step": 5943 + }, + { + "epoch": 3.226927252985885, + "grad_norm": 12.596453073542072, + "learning_rate": 5.902348663585144e-06, + "loss": 0.5128, + "step": 5944 + }, + { + "epoch": 3.227470141150923, + "grad_norm": 13.220304215230376, + "learning_rate": 5.899140887564991e-06, + "loss": 0.7293, + "step": 5945 + }, + { + "epoch": 3.228013029315961, + "grad_norm": 15.436135404867363, + "learning_rate": 5.8959336187456265e-06, + "loss": 0.8522, + "step": 5946 + }, + { + "epoch": 3.228555917480999, + "grad_norm": 11.078819738735637, + "learning_rate": 5.89272685752373e-06, + "loss": 0.5559, + "step": 5947 + }, + { + "epoch": 3.229098805646037, + "grad_norm": 11.732786443102684, + "learning_rate": 5.889520604295926e-06, + "loss": 0.472, + "step": 5948 + }, + { + "epoch": 3.229641693811075, + "grad_norm": 12.744163420827304, + "learning_rate": 5.88631485945876e-06, + "loss": 0.5843, + "step": 5949 + }, + { + "epoch": 3.230184581976113, + "grad_norm": 12.717140075489619, + "learning_rate": 5.8831096234087335e-06, + "loss": 0.4826, + "step": 5950 + }, + { + "epoch": 3.230727470141151, + "grad_norm": 14.460440748409685, + "learning_rate": 5.8799048965422665e-06, + "loss": 0.5057, + "step": 5951 + }, + { + "epoch": 3.231270358306189, + "grad_norm": 13.034475639862867, + "learning_rate": 5.8767006792557316e-06, + "loss": 0.6811, + "step": 5952 + }, + { + "epoch": 3.231813246471227, + "grad_norm": 12.638392990019069, + "learning_rate": 5.873496971945425e-06, + "loss": 0.4532, + "step": 5953 + }, + { + "epoch": 3.232356134636265, + "grad_norm": 11.04321550599593, + "learning_rate": 5.8702937750075914e-06, + "loss": 0.5711, + "step": 5954 + }, + { + "epoch": 3.232899022801303, + "grad_norm": 14.964397050426088, + "learning_rate": 5.867091088838406e-06, + "loss": 0.6479, + "step": 5955 + }, + { + "epoch": 3.233441910966341, + "grad_norm": 10.647120251880176, + "learning_rate": 5.863888913833979e-06, + "loss": 0.5262, + "step": 5956 + }, + { + "epoch": 3.233984799131379, + "grad_norm": 11.288671625393498, + "learning_rate": 5.860687250390367e-06, + "loss": 0.479, + "step": 5957 + }, + { + "epoch": 3.234527687296417, + "grad_norm": 11.49866073465499, + "learning_rate": 5.857486098903549e-06, + "loss": 0.2939, + "step": 5958 + }, + { + "epoch": 3.235070575461455, + "grad_norm": 11.340345898343635, + "learning_rate": 5.854285459769454e-06, + "loss": 0.3908, + "step": 5959 + }, + { + "epoch": 3.235613463626493, + "grad_norm": 10.20477531793921, + "learning_rate": 5.851085333383939e-06, + "loss": 0.4133, + "step": 5960 + }, + { + "epoch": 3.236156351791531, + "grad_norm": 9.818700824916304, + "learning_rate": 5.8478857201428075e-06, + "loss": 0.3324, + "step": 5961 + }, + { + "epoch": 3.236699239956569, + "grad_norm": 9.73255562956133, + "learning_rate": 5.844686620441782e-06, + "loss": 0.7815, + "step": 5962 + }, + { + "epoch": 3.237242128121607, + "grad_norm": 17.504088445923752, + "learning_rate": 5.841488034676542e-06, + "loss": 0.693, + "step": 5963 + }, + { + "epoch": 3.237785016286645, + "grad_norm": 15.010551187805474, + "learning_rate": 5.838289963242684e-06, + "loss": 0.453, + "step": 5964 + }, + { + "epoch": 3.238327904451683, + "grad_norm": 11.512605593104578, + "learning_rate": 5.835092406535763e-06, + "loss": 0.792, + "step": 5965 + }, + { + "epoch": 3.238870792616721, + "grad_norm": 15.034973727458302, + "learning_rate": 5.831895364951247e-06, + "loss": 0.6216, + "step": 5966 + }, + { + "epoch": 3.239413680781759, + "grad_norm": 15.522277540385527, + "learning_rate": 5.828698838884558e-06, + "loss": 0.4771, + "step": 5967 + }, + { + "epoch": 3.239956568946797, + "grad_norm": 11.391715611260866, + "learning_rate": 5.825502828731042e-06, + "loss": 0.4213, + "step": 5968 + }, + { + "epoch": 3.240499457111835, + "grad_norm": 15.534988403627082, + "learning_rate": 5.82230733488599e-06, + "loss": 0.6494, + "step": 5969 + }, + { + "epoch": 3.241042345276873, + "grad_norm": 18.181760899119496, + "learning_rate": 5.819112357744626e-06, + "loss": 0.5825, + "step": 5970 + }, + { + "epoch": 3.241585233441911, + "grad_norm": 10.971869300023029, + "learning_rate": 5.815917897702104e-06, + "loss": 0.7647, + "step": 5971 + }, + { + "epoch": 3.242128121606949, + "grad_norm": 11.760933490969961, + "learning_rate": 5.812723955153533e-06, + "loss": 0.6203, + "step": 5972 + }, + { + "epoch": 3.242671009771987, + "grad_norm": 17.51328901218973, + "learning_rate": 5.809530530493929e-06, + "loss": 0.6598, + "step": 5973 + }, + { + "epoch": 3.243213897937025, + "grad_norm": 12.500120263524371, + "learning_rate": 5.8063376241182745e-06, + "loss": 0.4958, + "step": 5974 + }, + { + "epoch": 3.243756786102063, + "grad_norm": 11.050563890247632, + "learning_rate": 5.803145236421464e-06, + "loss": 0.4204, + "step": 5975 + }, + { + "epoch": 3.244299674267101, + "grad_norm": 6.058184111347752, + "learning_rate": 5.799953367798335e-06, + "loss": 0.2181, + "step": 5976 + }, + { + "epoch": 3.244842562432139, + "grad_norm": 9.648373148728487, + "learning_rate": 5.796762018643675e-06, + "loss": 0.4304, + "step": 5977 + }, + { + "epoch": 3.245385450597177, + "grad_norm": 11.906735912997219, + "learning_rate": 5.793571189352179e-06, + "loss": 0.5365, + "step": 5978 + }, + { + "epoch": 3.245928338762215, + "grad_norm": 12.12325593772168, + "learning_rate": 5.790380880318511e-06, + "loss": 0.3811, + "step": 5979 + }, + { + "epoch": 3.246471226927253, + "grad_norm": 18.625337422182657, + "learning_rate": 5.787191091937236e-06, + "loss": 0.6784, + "step": 5980 + }, + { + "epoch": 3.247014115092291, + "grad_norm": 9.093728620631289, + "learning_rate": 5.784001824602891e-06, + "loss": 0.4545, + "step": 5981 + }, + { + "epoch": 3.247557003257329, + "grad_norm": 12.360404516544222, + "learning_rate": 5.78081307870991e-06, + "loss": 0.5982, + "step": 5982 + }, + { + "epoch": 3.248099891422367, + "grad_norm": 12.023338602328888, + "learning_rate": 5.777624854652696e-06, + "loss": 0.4595, + "step": 5983 + }, + { + "epoch": 3.248642779587405, + "grad_norm": 11.60238178463911, + "learning_rate": 5.774437152825574e-06, + "loss": 0.4674, + "step": 5984 + }, + { + "epoch": 3.249185667752443, + "grad_norm": 13.826357134916073, + "learning_rate": 5.77124997362279e-06, + "loss": 0.4552, + "step": 5985 + }, + { + "epoch": 3.249728555917481, + "grad_norm": 13.531388043207002, + "learning_rate": 5.7680633174385595e-06, + "loss": 0.4926, + "step": 5986 + }, + { + "epoch": 3.250271444082519, + "grad_norm": 11.384053416563145, + "learning_rate": 5.764877184666993e-06, + "loss": 0.5127, + "step": 5987 + }, + { + "epoch": 3.250814332247557, + "grad_norm": 12.390918037366044, + "learning_rate": 5.761691575702168e-06, + "loss": 0.6428, + "step": 5988 + }, + { + "epoch": 3.251357220412595, + "grad_norm": 12.158948584345922, + "learning_rate": 5.758506490938087e-06, + "loss": 0.5374, + "step": 5989 + }, + { + "epoch": 3.251900108577633, + "grad_norm": 12.352114570156113, + "learning_rate": 5.755321930768683e-06, + "loss": 0.4149, + "step": 5990 + }, + { + "epoch": 3.252442996742671, + "grad_norm": 11.451850663530495, + "learning_rate": 5.752137895587826e-06, + "loss": 0.4864, + "step": 5991 + }, + { + "epoch": 3.252985884907709, + "grad_norm": 14.59595398014034, + "learning_rate": 5.748954385789325e-06, + "loss": 0.6791, + "step": 5992 + }, + { + "epoch": 3.253528773072747, + "grad_norm": 10.572257438915425, + "learning_rate": 5.74577140176692e-06, + "loss": 0.3851, + "step": 5993 + }, + { + "epoch": 3.254071661237785, + "grad_norm": 13.496500778871066, + "learning_rate": 5.742588943914289e-06, + "loss": 0.8371, + "step": 5994 + }, + { + "epoch": 3.254614549402823, + "grad_norm": 10.366843508091446, + "learning_rate": 5.739407012625043e-06, + "loss": 0.5289, + "step": 5995 + }, + { + "epoch": 3.255157437567861, + "grad_norm": 13.59778716822848, + "learning_rate": 5.736225608292727e-06, + "loss": 0.6353, + "step": 5996 + }, + { + "epoch": 3.255700325732899, + "grad_norm": 8.13083770907386, + "learning_rate": 5.7330447313108236e-06, + "loss": 0.2591, + "step": 5997 + }, + { + "epoch": 3.256243213897937, + "grad_norm": 11.705959913597166, + "learning_rate": 5.72986438207275e-06, + "loss": 0.4058, + "step": 5998 + }, + { + "epoch": 3.256786102062975, + "grad_norm": 9.25230495085822, + "learning_rate": 5.726684560971852e-06, + "loss": 0.3647, + "step": 5999 + }, + { + "epoch": 3.257328990228013, + "grad_norm": 10.705513260623167, + "learning_rate": 5.72350526840142e-06, + "loss": 0.6986, + "step": 6000 + }, + { + "epoch": 3.257871878393051, + "grad_norm": 11.426736023101553, + "learning_rate": 5.720326504754672e-06, + "loss": 0.548, + "step": 6001 + }, + { + "epoch": 3.258414766558089, + "grad_norm": 11.105330737807535, + "learning_rate": 5.71714827042476e-06, + "loss": 0.5345, + "step": 6002 + }, + { + "epoch": 3.258957654723127, + "grad_norm": 11.734545793783724, + "learning_rate": 5.713970565804782e-06, + "loss": 0.4435, + "step": 6003 + }, + { + "epoch": 3.259500542888165, + "grad_norm": 13.234029513207872, + "learning_rate": 5.710793391287751e-06, + "loss": 0.538, + "step": 6004 + }, + { + "epoch": 3.260043431053203, + "grad_norm": 9.429540505637686, + "learning_rate": 5.707616747266631e-06, + "loss": 0.4042, + "step": 6005 + }, + { + "epoch": 3.260586319218241, + "grad_norm": 10.73830038342973, + "learning_rate": 5.704440634134312e-06, + "loss": 0.4394, + "step": 6006 + }, + { + "epoch": 3.261129207383279, + "grad_norm": 15.154563307319142, + "learning_rate": 5.70126505228362e-06, + "loss": 0.5733, + "step": 6007 + }, + { + "epoch": 3.261672095548317, + "grad_norm": 11.231865527791925, + "learning_rate": 5.6980900021073196e-06, + "loss": 0.3472, + "step": 6008 + }, + { + "epoch": 3.262214983713355, + "grad_norm": 10.315608836480246, + "learning_rate": 5.694915483998099e-06, + "loss": 0.3231, + "step": 6009 + }, + { + "epoch": 3.262757871878393, + "grad_norm": 13.017299228674494, + "learning_rate": 5.691741498348601e-06, + "loss": 0.4652, + "step": 6010 + }, + { + "epoch": 3.263300760043431, + "grad_norm": 13.428357165583693, + "learning_rate": 5.688568045551373e-06, + "loss": 0.51, + "step": 6011 + }, + { + "epoch": 3.263843648208469, + "grad_norm": 12.887391562992057, + "learning_rate": 5.685395125998927e-06, + "loss": 0.7156, + "step": 6012 + }, + { + "epoch": 3.264386536373507, + "grad_norm": 11.112402459616494, + "learning_rate": 5.682222740083683e-06, + "loss": 0.3077, + "step": 6013 + }, + { + "epoch": 3.264929424538545, + "grad_norm": 17.334186329467677, + "learning_rate": 5.6790508881980136e-06, + "loss": 0.487, + "step": 6014 + }, + { + "epoch": 3.265472312703583, + "grad_norm": 11.272673037009941, + "learning_rate": 5.675879570734222e-06, + "loss": 0.5596, + "step": 6015 + }, + { + "epoch": 3.266015200868621, + "grad_norm": 10.57062964165952, + "learning_rate": 5.672708788084528e-06, + "loss": 0.4693, + "step": 6016 + }, + { + "epoch": 3.266558089033659, + "grad_norm": 12.532994743872049, + "learning_rate": 5.669538540641116e-06, + "loss": 0.5709, + "step": 6017 + }, + { + "epoch": 3.267100977198697, + "grad_norm": 12.467434828440602, + "learning_rate": 5.666368828796072e-06, + "loss": 0.5328, + "step": 6018 + }, + { + "epoch": 3.267643865363735, + "grad_norm": 10.349926895360463, + "learning_rate": 5.663199652941444e-06, + "loss": 0.3587, + "step": 6019 + }, + { + "epoch": 3.268186753528773, + "grad_norm": 10.889914966114198, + "learning_rate": 5.660031013469189e-06, + "loss": 0.469, + "step": 6020 + }, + { + "epoch": 3.268729641693811, + "grad_norm": 16.63961667582981, + "learning_rate": 5.656862910771218e-06, + "loss": 0.7146, + "step": 6021 + }, + { + "epoch": 3.269272529858849, + "grad_norm": 10.447294419775371, + "learning_rate": 5.653695345239365e-06, + "loss": 0.3975, + "step": 6022 + }, + { + "epoch": 3.269815418023887, + "grad_norm": 12.85070217834024, + "learning_rate": 5.650528317265398e-06, + "loss": 0.7395, + "step": 6023 + }, + { + "epoch": 3.270358306188925, + "grad_norm": 8.341742485534219, + "learning_rate": 5.647361827241025e-06, + "loss": 0.2861, + "step": 6024 + }, + { + "epoch": 3.270901194353963, + "grad_norm": 9.898309581037385, + "learning_rate": 5.64419587555787e-06, + "loss": 0.6192, + "step": 6025 + }, + { + "epoch": 3.271444082519001, + "grad_norm": 12.335345691658691, + "learning_rate": 5.641030462607514e-06, + "loss": 0.5161, + "step": 6026 + }, + { + "epoch": 3.271986970684039, + "grad_norm": 10.304455881328025, + "learning_rate": 5.637865588781459e-06, + "loss": 0.5773, + "step": 6027 + }, + { + "epoch": 3.272529858849077, + "grad_norm": 11.527834619584397, + "learning_rate": 5.63470125447114e-06, + "loss": 0.4415, + "step": 6028 + }, + { + "epoch": 3.273072747014115, + "grad_norm": 12.835285732552766, + "learning_rate": 5.631537460067926e-06, + "loss": 0.4513, + "step": 6029 + }, + { + "epoch": 3.273615635179153, + "grad_norm": 15.34754753916867, + "learning_rate": 5.628374205963123e-06, + "loss": 0.9626, + "step": 6030 + }, + { + "epoch": 3.274158523344191, + "grad_norm": 13.537553719954104, + "learning_rate": 5.625211492547965e-06, + "loss": 0.6022, + "step": 6031 + }, + { + "epoch": 3.274701411509229, + "grad_norm": 13.03189959063881, + "learning_rate": 5.622049320213622e-06, + "loss": 0.3918, + "step": 6032 + }, + { + "epoch": 3.275244299674267, + "grad_norm": 11.883924111949161, + "learning_rate": 5.618887689351195e-06, + "loss": 0.6166, + "step": 6033 + }, + { + "epoch": 3.2757871878393052, + "grad_norm": 12.574278462374288, + "learning_rate": 5.615726600351723e-06, + "loss": 0.364, + "step": 6034 + }, + { + "epoch": 3.276330076004343, + "grad_norm": 11.646167790458746, + "learning_rate": 5.612566053606172e-06, + "loss": 0.4893, + "step": 6035 + }, + { + "epoch": 3.2768729641693812, + "grad_norm": 15.464095407664189, + "learning_rate": 5.6094060495054435e-06, + "loss": 0.5033, + "step": 6036 + }, + { + "epoch": 3.277415852334419, + "grad_norm": 9.470859172197285, + "learning_rate": 5.606246588440374e-06, + "loss": 0.3201, + "step": 6037 + }, + { + "epoch": 3.2779587404994572, + "grad_norm": 10.186080712958546, + "learning_rate": 5.603087670801728e-06, + "loss": 0.375, + "step": 6038 + }, + { + "epoch": 3.278501628664495, + "grad_norm": 12.069192340552016, + "learning_rate": 5.599929296980205e-06, + "loss": 0.521, + "step": 6039 + }, + { + "epoch": 3.2790445168295332, + "grad_norm": 13.470203218211036, + "learning_rate": 5.596771467366442e-06, + "loss": 0.6999, + "step": 6040 + }, + { + "epoch": 3.279587404994571, + "grad_norm": 15.234328984284964, + "learning_rate": 5.5936141823509995e-06, + "loss": 0.658, + "step": 6041 + }, + { + "epoch": 3.2801302931596092, + "grad_norm": 10.28714486065803, + "learning_rate": 5.590457442324374e-06, + "loss": 0.49, + "step": 6042 + }, + { + "epoch": 3.280673181324647, + "grad_norm": 13.229082345644139, + "learning_rate": 5.587301247677009e-06, + "loss": 0.8617, + "step": 6043 + }, + { + "epoch": 3.2812160694896852, + "grad_norm": 19.161743435780973, + "learning_rate": 5.5841455987992536e-06, + "loss": 0.673, + "step": 6044 + }, + { + "epoch": 3.281758957654723, + "grad_norm": 14.576260981710798, + "learning_rate": 5.580990496081407e-06, + "loss": 0.4843, + "step": 6045 + }, + { + "epoch": 3.2823018458197613, + "grad_norm": 10.509886153036398, + "learning_rate": 5.577835939913701e-06, + "loss": 0.3372, + "step": 6046 + }, + { + "epoch": 3.282844733984799, + "grad_norm": 14.23830213747384, + "learning_rate": 5.574681930686289e-06, + "loss": 0.5041, + "step": 6047 + }, + { + "epoch": 3.2833876221498373, + "grad_norm": 16.736802729512746, + "learning_rate": 5.571528468789276e-06, + "loss": 0.6261, + "step": 6048 + }, + { + "epoch": 3.283930510314875, + "grad_norm": 16.53917203611799, + "learning_rate": 5.5683755546126724e-06, + "loss": 0.6671, + "step": 6049 + }, + { + "epoch": 3.2844733984799133, + "grad_norm": 9.26057962235423, + "learning_rate": 5.565223188546452e-06, + "loss": 0.4135, + "step": 6050 + }, + { + "epoch": 3.285016286644951, + "grad_norm": 13.238225399242737, + "learning_rate": 5.562071370980486e-06, + "loss": 0.558, + "step": 6051 + }, + { + "epoch": 3.2855591748099893, + "grad_norm": 9.005338285525495, + "learning_rate": 5.558920102304615e-06, + "loss": 0.3677, + "step": 6052 + }, + { + "epoch": 3.286102062975027, + "grad_norm": 14.118931865854162, + "learning_rate": 5.55576938290858e-06, + "loss": 0.8303, + "step": 6053 + }, + { + "epoch": 3.2866449511400653, + "grad_norm": 9.596979737331942, + "learning_rate": 5.5526192131820665e-06, + "loss": 0.3322, + "step": 6054 + }, + { + "epoch": 3.287187839305103, + "grad_norm": 11.143732344567033, + "learning_rate": 5.549469593514705e-06, + "loss": 0.4143, + "step": 6055 + }, + { + "epoch": 3.2877307274701413, + "grad_norm": 12.135282131920155, + "learning_rate": 5.546320524296028e-06, + "loss": 0.5302, + "step": 6056 + }, + { + "epoch": 3.288273615635179, + "grad_norm": 12.209025529059957, + "learning_rate": 5.543172005915536e-06, + "loss": 0.5643, + "step": 6057 + }, + { + "epoch": 3.2888165038002173, + "grad_norm": 16.880330158776534, + "learning_rate": 5.540024038762623e-06, + "loss": 0.6743, + "step": 6058 + }, + { + "epoch": 3.289359391965255, + "grad_norm": 13.623361718010658, + "learning_rate": 5.536876623226652e-06, + "loss": 0.6925, + "step": 6059 + }, + { + "epoch": 3.2899022801302933, + "grad_norm": 11.161796182879957, + "learning_rate": 5.533729759696893e-06, + "loss": 0.5875, + "step": 6060 + }, + { + "epoch": 3.290445168295331, + "grad_norm": 14.324802623539963, + "learning_rate": 5.5305834485625545e-06, + "loss": 0.9071, + "step": 6061 + }, + { + "epoch": 3.2909880564603693, + "grad_norm": 12.43112855099258, + "learning_rate": 5.527437690212778e-06, + "loss": 0.3761, + "step": 6062 + }, + { + "epoch": 3.291530944625407, + "grad_norm": 14.231447955496243, + "learning_rate": 5.524292485036638e-06, + "loss": 0.7001, + "step": 6063 + }, + { + "epoch": 3.2920738327904453, + "grad_norm": 11.474598912194304, + "learning_rate": 5.521147833423135e-06, + "loss": 0.419, + "step": 6064 + }, + { + "epoch": 3.292616720955483, + "grad_norm": 13.615753477381617, + "learning_rate": 5.518003735761206e-06, + "loss": 0.7142, + "step": 6065 + }, + { + "epoch": 3.2931596091205213, + "grad_norm": 10.341494655988399, + "learning_rate": 5.514860192439719e-06, + "loss": 0.4286, + "step": 6066 + }, + { + "epoch": 3.293702497285559, + "grad_norm": 11.37742694154389, + "learning_rate": 5.511717203847472e-06, + "loss": 0.4302, + "step": 6067 + }, + { + "epoch": 3.2942453854505973, + "grad_norm": 14.04977404747662, + "learning_rate": 5.508574770373194e-06, + "loss": 0.4538, + "step": 6068 + }, + { + "epoch": 3.294788273615635, + "grad_norm": 13.008904129849688, + "learning_rate": 5.505432892405547e-06, + "loss": 0.419, + "step": 6069 + }, + { + "epoch": 3.2953311617806733, + "grad_norm": 8.086590828855625, + "learning_rate": 5.502291570333122e-06, + "loss": 0.276, + "step": 6070 + }, + { + "epoch": 3.295874049945711, + "grad_norm": 8.20494775069595, + "learning_rate": 5.499150804544445e-06, + "loss": 0.349, + "step": 6071 + }, + { + "epoch": 3.2964169381107493, + "grad_norm": 13.336711424526726, + "learning_rate": 5.496010595427969e-06, + "loss": 0.3958, + "step": 6072 + }, + { + "epoch": 3.296959826275787, + "grad_norm": 21.927197817240263, + "learning_rate": 5.492870943372082e-06, + "loss": 0.5561, + "step": 6073 + }, + { + "epoch": 3.2975027144408253, + "grad_norm": 14.168102563986727, + "learning_rate": 5.489731848765098e-06, + "loss": 0.6231, + "step": 6074 + }, + { + "epoch": 3.298045602605863, + "grad_norm": 9.428786171987674, + "learning_rate": 5.48659331199527e-06, + "loss": 0.3004, + "step": 6075 + }, + { + "epoch": 3.2985884907709013, + "grad_norm": 13.58619098409203, + "learning_rate": 5.483455333450774e-06, + "loss": 0.7655, + "step": 6076 + }, + { + "epoch": 3.299131378935939, + "grad_norm": 10.799157987601124, + "learning_rate": 5.480317913519718e-06, + "loss": 0.4688, + "step": 6077 + }, + { + "epoch": 3.2996742671009773, + "grad_norm": 11.88277850667338, + "learning_rate": 5.477181052590148e-06, + "loss": 0.576, + "step": 6078 + }, + { + "epoch": 3.300217155266015, + "grad_norm": 9.399361473432771, + "learning_rate": 5.474044751050032e-06, + "loss": 0.3612, + "step": 6079 + }, + { + "epoch": 3.3007600434310533, + "grad_norm": 12.83439735807638, + "learning_rate": 5.470909009287273e-06, + "loss": 0.4007, + "step": 6080 + }, + { + "epoch": 3.301302931596091, + "grad_norm": 10.935369972874188, + "learning_rate": 5.467773827689712e-06, + "loss": 0.5349, + "step": 6081 + }, + { + "epoch": 3.3018458197611293, + "grad_norm": 13.981231839826476, + "learning_rate": 5.464639206645104e-06, + "loss": 0.4123, + "step": 6082 + }, + { + "epoch": 3.302388707926167, + "grad_norm": 10.388999075429536, + "learning_rate": 5.461505146541147e-06, + "loss": 0.666, + "step": 6083 + }, + { + "epoch": 3.3029315960912053, + "grad_norm": 10.483727188037403, + "learning_rate": 5.4583716477654665e-06, + "loss": 0.4385, + "step": 6084 + }, + { + "epoch": 3.303474484256243, + "grad_norm": 15.004888765224539, + "learning_rate": 5.455238710705616e-06, + "loss": 0.6089, + "step": 6085 + }, + { + "epoch": 3.3040173724212814, + "grad_norm": 12.510560050118539, + "learning_rate": 5.452106335749092e-06, + "loss": 0.5481, + "step": 6086 + }, + { + "epoch": 3.304560260586319, + "grad_norm": 12.402037551065197, + "learning_rate": 5.448974523283297e-06, + "loss": 0.5909, + "step": 6087 + }, + { + "epoch": 3.3051031487513574, + "grad_norm": 9.456679882341124, + "learning_rate": 5.445843273695595e-06, + "loss": 0.4422, + "step": 6088 + }, + { + "epoch": 3.305646036916395, + "grad_norm": 12.165182627083702, + "learning_rate": 5.4427125873732455e-06, + "loss": 0.8632, + "step": 6089 + }, + { + "epoch": 3.3061889250814334, + "grad_norm": 10.815051939615287, + "learning_rate": 5.439582464703476e-06, + "loss": 0.5143, + "step": 6090 + }, + { + "epoch": 3.306731813246471, + "grad_norm": 8.659965055909717, + "learning_rate": 5.436452906073406e-06, + "loss": 0.3732, + "step": 6091 + }, + { + "epoch": 3.3072747014115094, + "grad_norm": 12.566885836457084, + "learning_rate": 5.433323911870119e-06, + "loss": 0.7708, + "step": 6092 + }, + { + "epoch": 3.307817589576547, + "grad_norm": 12.162201810127142, + "learning_rate": 5.430195482480611e-06, + "loss": 0.5414, + "step": 6093 + }, + { + "epoch": 3.3083604777415854, + "grad_norm": 10.50777462312628, + "learning_rate": 5.427067618291804e-06, + "loss": 0.4301, + "step": 6094 + }, + { + "epoch": 3.308903365906623, + "grad_norm": 10.640857438281406, + "learning_rate": 5.423940319690568e-06, + "loss": 0.4682, + "step": 6095 + }, + { + "epoch": 3.3094462540716614, + "grad_norm": 10.370610442117792, + "learning_rate": 5.420813587063678e-06, + "loss": 0.3056, + "step": 6096 + }, + { + "epoch": 3.309989142236699, + "grad_norm": 10.60821300635486, + "learning_rate": 5.417687420797867e-06, + "loss": 0.3344, + "step": 6097 + }, + { + "epoch": 3.3105320304017374, + "grad_norm": 10.353507117674113, + "learning_rate": 5.414561821279778e-06, + "loss": 0.656, + "step": 6098 + }, + { + "epoch": 3.311074918566775, + "grad_norm": 8.344962502930883, + "learning_rate": 5.411436788895992e-06, + "loss": 0.3413, + "step": 6099 + }, + { + "epoch": 3.3116178067318134, + "grad_norm": 9.483199213032712, + "learning_rate": 5.408312324033016e-06, + "loss": 0.4156, + "step": 6100 + }, + { + "epoch": 3.312160694896851, + "grad_norm": 10.849967748759967, + "learning_rate": 5.40518842707729e-06, + "loss": 0.3849, + "step": 6101 + }, + { + "epoch": 3.3127035830618894, + "grad_norm": 11.385869965644408, + "learning_rate": 5.402065098415188e-06, + "loss": 0.5767, + "step": 6102 + }, + { + "epoch": 3.313246471226927, + "grad_norm": 9.846646631682042, + "learning_rate": 5.398942338432993e-06, + "loss": 0.3517, + "step": 6103 + }, + { + "epoch": 3.3137893593919654, + "grad_norm": 12.208760557718977, + "learning_rate": 5.3958201475169455e-06, + "loss": 0.3989, + "step": 6104 + }, + { + "epoch": 3.314332247557003, + "grad_norm": 14.044852806162847, + "learning_rate": 5.392698526053203e-06, + "loss": 0.6161, + "step": 6105 + }, + { + "epoch": 3.3148751357220414, + "grad_norm": 16.302573867426716, + "learning_rate": 5.389577474427848e-06, + "loss": 0.5668, + "step": 6106 + }, + { + "epoch": 3.315418023887079, + "grad_norm": 15.241837519747031, + "learning_rate": 5.3864569930268986e-06, + "loss": 0.531, + "step": 6107 + }, + { + "epoch": 3.3159609120521174, + "grad_norm": 16.054606664926766, + "learning_rate": 5.383337082236302e-06, + "loss": 0.5549, + "step": 6108 + }, + { + "epoch": 3.316503800217155, + "grad_norm": 7.873315866697446, + "learning_rate": 5.3802177424419335e-06, + "loss": 0.3505, + "step": 6109 + }, + { + "epoch": 3.3170466883821934, + "grad_norm": 7.233865234327245, + "learning_rate": 5.377098974029595e-06, + "loss": 0.3402, + "step": 6110 + }, + { + "epoch": 3.317589576547231, + "grad_norm": 13.489019875241555, + "learning_rate": 5.373980777385026e-06, + "loss": 0.5829, + "step": 6111 + }, + { + "epoch": 3.3181324647122694, + "grad_norm": 8.944352234530756, + "learning_rate": 5.370863152893886e-06, + "loss": 0.2591, + "step": 6112 + }, + { + "epoch": 3.318675352877307, + "grad_norm": 11.421952434092379, + "learning_rate": 5.367746100941769e-06, + "loss": 0.4549, + "step": 6113 + }, + { + "epoch": 3.3192182410423454, + "grad_norm": 9.61822328391865, + "learning_rate": 5.364629621914197e-06, + "loss": 0.3805, + "step": 6114 + }, + { + "epoch": 3.319761129207383, + "grad_norm": 9.466765582290753, + "learning_rate": 5.3615137161966205e-06, + "loss": 0.2804, + "step": 6115 + }, + { + "epoch": 3.3203040173724214, + "grad_norm": 10.228565001712072, + "learning_rate": 5.358398384174421e-06, + "loss": 0.3802, + "step": 6116 + }, + { + "epoch": 3.320846905537459, + "grad_norm": 12.471667235628182, + "learning_rate": 5.355283626232907e-06, + "loss": 0.7176, + "step": 6117 + }, + { + "epoch": 3.3213897937024974, + "grad_norm": 8.698625687388546, + "learning_rate": 5.352169442757312e-06, + "loss": 0.2962, + "step": 6118 + }, + { + "epoch": 3.321932681867535, + "grad_norm": 14.02974515499815, + "learning_rate": 5.349055834132817e-06, + "loss": 0.4804, + "step": 6119 + }, + { + "epoch": 3.3224755700325734, + "grad_norm": 14.91877137041168, + "learning_rate": 5.345942800744499e-06, + "loss": 0.4992, + "step": 6120 + }, + { + "epoch": 3.323018458197611, + "grad_norm": 13.54006240651423, + "learning_rate": 5.342830342977403e-06, + "loss": 0.822, + "step": 6121 + }, + { + "epoch": 3.3235613463626494, + "grad_norm": 11.467004290914351, + "learning_rate": 5.3397184612164676e-06, + "loss": 0.4878, + "step": 6122 + }, + { + "epoch": 3.324104234527687, + "grad_norm": 11.864551909227606, + "learning_rate": 5.33660715584658e-06, + "loss": 0.5312, + "step": 6123 + }, + { + "epoch": 3.3246471226927254, + "grad_norm": 10.288196378126806, + "learning_rate": 5.333496427252551e-06, + "loss": 0.4778, + "step": 6124 + }, + { + "epoch": 3.3251900108577632, + "grad_norm": 11.010968624281599, + "learning_rate": 5.330386275819119e-06, + "loss": 0.4471, + "step": 6125 + }, + { + "epoch": 3.3257328990228014, + "grad_norm": 9.902175725807343, + "learning_rate": 5.327276701930961e-06, + "loss": 0.3581, + "step": 6126 + }, + { + "epoch": 3.3262757871878392, + "grad_norm": 14.462369688525524, + "learning_rate": 5.324167705972661e-06, + "loss": 0.4262, + "step": 6127 + }, + { + "epoch": 3.3268186753528775, + "grad_norm": 12.522646032715151, + "learning_rate": 5.32105928832876e-06, + "loss": 0.3605, + "step": 6128 + }, + { + "epoch": 3.3273615635179152, + "grad_norm": 8.804303263881113, + "learning_rate": 5.317951449383693e-06, + "loss": 0.2589, + "step": 6129 + }, + { + "epoch": 3.3279044516829535, + "grad_norm": 16.231015853377418, + "learning_rate": 5.314844189521859e-06, + "loss": 0.6517, + "step": 6130 + }, + { + "epoch": 3.3284473398479912, + "grad_norm": 15.588313322776349, + "learning_rate": 5.311737509127561e-06, + "loss": 0.6392, + "step": 6131 + }, + { + "epoch": 3.3289902280130295, + "grad_norm": 8.118329570429164, + "learning_rate": 5.308631408585041e-06, + "loss": 0.2433, + "step": 6132 + }, + { + "epoch": 3.3295331161780672, + "grad_norm": 17.927534810088048, + "learning_rate": 5.305525888278469e-06, + "loss": 0.8314, + "step": 6133 + }, + { + "epoch": 3.3300760043431055, + "grad_norm": 12.7719407908283, + "learning_rate": 5.302420948591929e-06, + "loss": 0.5422, + "step": 6134 + }, + { + "epoch": 3.3306188925081432, + "grad_norm": 13.715317083116945, + "learning_rate": 5.29931658990946e-06, + "loss": 0.4756, + "step": 6135 + }, + { + "epoch": 3.3311617806731815, + "grad_norm": 14.232906623385299, + "learning_rate": 5.296212812615001e-06, + "loss": 0.5507, + "step": 6136 + }, + { + "epoch": 3.3317046688382193, + "grad_norm": 11.07024486678308, + "learning_rate": 5.29310961709244e-06, + "loss": 0.4916, + "step": 6137 + }, + { + "epoch": 3.3322475570032575, + "grad_norm": 12.322287003987269, + "learning_rate": 5.290007003725585e-06, + "loss": 0.426, + "step": 6138 + }, + { + "epoch": 3.3327904451682953, + "grad_norm": 11.084747737931762, + "learning_rate": 5.286904972898168e-06, + "loss": 0.2768, + "step": 6139 + }, + { + "epoch": 3.3333333333333335, + "grad_norm": 12.758536777666697, + "learning_rate": 5.283803524993858e-06, + "loss": 0.9266, + "step": 6140 + }, + { + "epoch": 3.3338762214983713, + "grad_norm": 13.098970900751402, + "learning_rate": 5.280702660396243e-06, + "loss": 0.512, + "step": 6141 + }, + { + "epoch": 3.3344191096634095, + "grad_norm": 12.317106256281363, + "learning_rate": 5.277602379488844e-06, + "loss": 0.4155, + "step": 6142 + }, + { + "epoch": 3.3349619978284473, + "grad_norm": 11.38584068826573, + "learning_rate": 5.274502682655112e-06, + "loss": 0.4051, + "step": 6143 + }, + { + "epoch": 3.3355048859934855, + "grad_norm": 14.501004041816849, + "learning_rate": 5.271403570278417e-06, + "loss": 0.5548, + "step": 6144 + }, + { + "epoch": 3.3360477741585233, + "grad_norm": 12.998451360538299, + "learning_rate": 5.268305042742065e-06, + "loss": 0.4021, + "step": 6145 + }, + { + "epoch": 3.3365906623235615, + "grad_norm": 9.038839543671822, + "learning_rate": 5.2652071004292855e-06, + "loss": 0.2625, + "step": 6146 + }, + { + "epoch": 3.3371335504885993, + "grad_norm": 14.728843022444776, + "learning_rate": 5.26210974372324e-06, + "loss": 0.7124, + "step": 6147 + }, + { + "epoch": 3.3376764386536375, + "grad_norm": 12.24691457854214, + "learning_rate": 5.259012973007011e-06, + "loss": 0.3889, + "step": 6148 + }, + { + "epoch": 3.3382193268186753, + "grad_norm": 12.889746719808702, + "learning_rate": 5.255916788663614e-06, + "loss": 0.6392, + "step": 6149 + }, + { + "epoch": 3.3387622149837135, + "grad_norm": 15.16855584494826, + "learning_rate": 5.252821191075989e-06, + "loss": 0.6308, + "step": 6150 + }, + { + "epoch": 3.3393051031487513, + "grad_norm": 15.137292971607204, + "learning_rate": 5.249726180627006e-06, + "loss": 0.4732, + "step": 6151 + }, + { + "epoch": 3.3398479913137895, + "grad_norm": 12.827169669915257, + "learning_rate": 5.246631757699461e-06, + "loss": 0.4235, + "step": 6152 + }, + { + "epoch": 3.3403908794788273, + "grad_norm": 8.000573554523031, + "learning_rate": 5.243537922676074e-06, + "loss": 0.3355, + "step": 6153 + }, + { + "epoch": 3.3409337676438655, + "grad_norm": 11.071696970347325, + "learning_rate": 5.240444675939498e-06, + "loss": 0.3129, + "step": 6154 + }, + { + "epoch": 3.3414766558089033, + "grad_norm": 13.751963016664462, + "learning_rate": 5.237352017872308e-06, + "loss": 0.7001, + "step": 6155 + }, + { + "epoch": 3.3420195439739415, + "grad_norm": 11.736421365008567, + "learning_rate": 5.23425994885701e-06, + "loss": 0.5775, + "step": 6156 + }, + { + "epoch": 3.3425624321389793, + "grad_norm": 13.034138205582762, + "learning_rate": 5.231168469276044e-06, + "loss": 0.4829, + "step": 6157 + }, + { + "epoch": 3.3431053203040175, + "grad_norm": 13.897122730322542, + "learning_rate": 5.228077579511754e-06, + "loss": 0.6586, + "step": 6158 + }, + { + "epoch": 3.3436482084690553, + "grad_norm": 13.765426696215664, + "learning_rate": 5.224987279946441e-06, + "loss": 0.4817, + "step": 6159 + }, + { + "epoch": 3.3441910966340935, + "grad_norm": 13.446494398430909, + "learning_rate": 5.221897570962304e-06, + "loss": 0.5121, + "step": 6160 + }, + { + "epoch": 3.3447339847991313, + "grad_norm": 9.817713520136255, + "learning_rate": 5.2188084529415e-06, + "loss": 0.2662, + "step": 6161 + }, + { + "epoch": 3.3452768729641695, + "grad_norm": 10.025373895120913, + "learning_rate": 5.215719926266082e-06, + "loss": 0.4928, + "step": 6162 + }, + { + "epoch": 3.3458197611292073, + "grad_norm": 13.267267741014804, + "learning_rate": 5.212631991318044e-06, + "loss": 0.6133, + "step": 6163 + }, + { + "epoch": 3.3463626492942455, + "grad_norm": 13.654831265704226, + "learning_rate": 5.209544648479319e-06, + "loss": 0.4202, + "step": 6164 + }, + { + "epoch": 3.3469055374592833, + "grad_norm": 14.003058740046605, + "learning_rate": 5.20645789813174e-06, + "loss": 0.6629, + "step": 6165 + }, + { + "epoch": 3.3474484256243215, + "grad_norm": 8.761766797305249, + "learning_rate": 5.203371740657095e-06, + "loss": 0.3107, + "step": 6166 + }, + { + "epoch": 3.3479913137893593, + "grad_norm": 13.961944426146035, + "learning_rate": 5.2002861764370705e-06, + "loss": 0.4029, + "step": 6167 + }, + { + "epoch": 3.3485342019543975, + "grad_norm": 11.942844125447042, + "learning_rate": 5.1972012058533035e-06, + "loss": 0.2327, + "step": 6168 + }, + { + "epoch": 3.3490770901194353, + "grad_norm": 15.94835762092326, + "learning_rate": 5.194116829287348e-06, + "loss": 0.4844, + "step": 6169 + }, + { + "epoch": 3.3496199782844736, + "grad_norm": 14.492034902597197, + "learning_rate": 5.191033047120682e-06, + "loss": 0.7219, + "step": 6170 + }, + { + "epoch": 3.3501628664495113, + "grad_norm": 16.04200227643707, + "learning_rate": 5.187949859734715e-06, + "loss": 0.4706, + "step": 6171 + }, + { + "epoch": 3.3507057546145496, + "grad_norm": 16.558616346746106, + "learning_rate": 5.184867267510774e-06, + "loss": 0.6355, + "step": 6172 + }, + { + "epoch": 3.3512486427795873, + "grad_norm": 14.202794625614665, + "learning_rate": 5.18178527083013e-06, + "loss": 0.6307, + "step": 6173 + }, + { + "epoch": 3.3517915309446256, + "grad_norm": 13.034510344424431, + "learning_rate": 5.178703870073954e-06, + "loss": 0.614, + "step": 6174 + }, + { + "epoch": 3.3523344191096633, + "grad_norm": 8.768533900570704, + "learning_rate": 5.1756230656233715e-06, + "loss": 0.4036, + "step": 6175 + }, + { + "epoch": 3.3528773072747016, + "grad_norm": 16.580405700504947, + "learning_rate": 5.172542857859418e-06, + "loss": 0.5804, + "step": 6176 + }, + { + "epoch": 3.3534201954397393, + "grad_norm": 11.843457868265668, + "learning_rate": 5.169463247163058e-06, + "loss": 0.5303, + "step": 6177 + }, + { + "epoch": 3.3539630836047776, + "grad_norm": 12.879389378162015, + "learning_rate": 5.166384233915182e-06, + "loss": 0.5253, + "step": 6178 + }, + { + "epoch": 3.3545059717698154, + "grad_norm": 9.5479358372862, + "learning_rate": 5.163305818496607e-06, + "loss": 0.2689, + "step": 6179 + }, + { + "epoch": 3.3550488599348536, + "grad_norm": 11.148812089885375, + "learning_rate": 5.160228001288077e-06, + "loss": 0.5212, + "step": 6180 + }, + { + "epoch": 3.3555917480998914, + "grad_norm": 10.77047326599204, + "learning_rate": 5.157150782670261e-06, + "loss": 0.3674, + "step": 6181 + }, + { + "epoch": 3.3561346362649296, + "grad_norm": 11.804189327484117, + "learning_rate": 5.154074163023756e-06, + "loss": 0.5972, + "step": 6182 + }, + { + "epoch": 3.3566775244299674, + "grad_norm": 11.540083386779209, + "learning_rate": 5.15099814272908e-06, + "loss": 0.703, + "step": 6183 + }, + { + "epoch": 3.3572204125950056, + "grad_norm": 12.780871236150638, + "learning_rate": 5.147922722166683e-06, + "loss": 0.4786, + "step": 6184 + }, + { + "epoch": 3.3577633007600434, + "grad_norm": 11.170640555611453, + "learning_rate": 5.144847901716936e-06, + "loss": 0.6836, + "step": 6185 + }, + { + "epoch": 3.3583061889250816, + "grad_norm": 11.202376321881136, + "learning_rate": 5.1417736817601386e-06, + "loss": 0.327, + "step": 6186 + }, + { + "epoch": 3.3588490770901194, + "grad_norm": 10.62227316496751, + "learning_rate": 5.138700062676516e-06, + "loss": 0.4885, + "step": 6187 + }, + { + "epoch": 3.3593919652551576, + "grad_norm": 11.784247976226597, + "learning_rate": 5.135627044846216e-06, + "loss": 0.4551, + "step": 6188 + }, + { + "epoch": 3.3599348534201954, + "grad_norm": 9.425213682768682, + "learning_rate": 5.132554628649313e-06, + "loss": 0.3285, + "step": 6189 + }, + { + "epoch": 3.3604777415852336, + "grad_norm": 10.622928852300923, + "learning_rate": 5.1294828144658185e-06, + "loss": 0.4265, + "step": 6190 + }, + { + "epoch": 3.3610206297502714, + "grad_norm": 11.666591539214851, + "learning_rate": 5.126411602675649e-06, + "loss": 0.2681, + "step": 6191 + }, + { + "epoch": 3.3615635179153096, + "grad_norm": 19.598513215176627, + "learning_rate": 5.123340993658658e-06, + "loss": 0.7577, + "step": 6192 + }, + { + "epoch": 3.3621064060803474, + "grad_norm": 13.077915822970434, + "learning_rate": 5.120270987794627e-06, + "loss": 0.653, + "step": 6193 + }, + { + "epoch": 3.3626492942453856, + "grad_norm": 15.592283919989452, + "learning_rate": 5.117201585463256e-06, + "loss": 0.4488, + "step": 6194 + }, + { + "epoch": 3.3631921824104234, + "grad_norm": 9.288688824641968, + "learning_rate": 5.114132787044175e-06, + "loss": 0.4185, + "step": 6195 + }, + { + "epoch": 3.3637350705754616, + "grad_norm": 15.306973087542367, + "learning_rate": 5.111064592916935e-06, + "loss": 0.8354, + "step": 6196 + }, + { + "epoch": 3.3642779587404994, + "grad_norm": 11.499674684661942, + "learning_rate": 5.107997003461023e-06, + "loss": 0.5953, + "step": 6197 + }, + { + "epoch": 3.3648208469055376, + "grad_norm": 12.276252733631965, + "learning_rate": 5.104930019055834e-06, + "loss": 0.5574, + "step": 6198 + }, + { + "epoch": 3.3653637350705754, + "grad_norm": 15.0044996135686, + "learning_rate": 5.1018636400807075e-06, + "loss": 0.403, + "step": 6199 + }, + { + "epoch": 3.3659066232356136, + "grad_norm": 8.91478305391761, + "learning_rate": 5.098797866914889e-06, + "loss": 0.3447, + "step": 6200 + }, + { + "epoch": 3.3664495114006514, + "grad_norm": 12.707988441543339, + "learning_rate": 5.095732699937559e-06, + "loss": 0.6121, + "step": 6201 + }, + { + "epoch": 3.3669923995656896, + "grad_norm": 11.893526329599132, + "learning_rate": 5.092668139527831e-06, + "loss": 0.4349, + "step": 6202 + }, + { + "epoch": 3.3675352877307274, + "grad_norm": 9.573100623698757, + "learning_rate": 5.08960418606472e-06, + "loss": 0.3979, + "step": 6203 + }, + { + "epoch": 3.3680781758957656, + "grad_norm": 12.055211170267912, + "learning_rate": 5.0865408399271995e-06, + "loss": 0.7061, + "step": 6204 + }, + { + "epoch": 3.3686210640608034, + "grad_norm": 10.746677950946136, + "learning_rate": 5.08347810149413e-06, + "loss": 0.4687, + "step": 6205 + }, + { + "epoch": 3.3691639522258416, + "grad_norm": 11.592472853876425, + "learning_rate": 5.080415971144332e-06, + "loss": 0.4447, + "step": 6206 + }, + { + "epoch": 3.3697068403908794, + "grad_norm": 14.931359329572437, + "learning_rate": 5.077354449256521e-06, + "loss": 0.7571, + "step": 6207 + }, + { + "epoch": 3.3702497285559176, + "grad_norm": 11.016079444425253, + "learning_rate": 5.07429353620936e-06, + "loss": 0.3964, + "step": 6208 + }, + { + "epoch": 3.3707926167209554, + "grad_norm": 9.567910765955927, + "learning_rate": 5.071233232381425e-06, + "loss": 0.395, + "step": 6209 + }, + { + "epoch": 3.3713355048859937, + "grad_norm": 14.85470239623418, + "learning_rate": 5.0681735381512195e-06, + "loss": 0.557, + "step": 6210 + }, + { + "epoch": 3.3718783930510314, + "grad_norm": 11.383297704214762, + "learning_rate": 5.0651144538971746e-06, + "loss": 0.4522, + "step": 6211 + }, + { + "epoch": 3.3724212812160697, + "grad_norm": 13.445463967563244, + "learning_rate": 5.062055979997631e-06, + "loss": 0.5483, + "step": 6212 + }, + { + "epoch": 3.3729641693811074, + "grad_norm": 12.602441542764687, + "learning_rate": 5.058998116830878e-06, + "loss": 0.6498, + "step": 6213 + }, + { + "epoch": 3.3735070575461457, + "grad_norm": 11.110392326995717, + "learning_rate": 5.055940864775113e-06, + "loss": 0.4893, + "step": 6214 + }, + { + "epoch": 3.3740499457111834, + "grad_norm": 12.47623770525584, + "learning_rate": 5.052884224208461e-06, + "loss": 0.4237, + "step": 6215 + }, + { + "epoch": 3.3745928338762217, + "grad_norm": 9.744420395590724, + "learning_rate": 5.049828195508972e-06, + "loss": 0.2669, + "step": 6216 + }, + { + "epoch": 3.3751357220412594, + "grad_norm": 15.871527506273058, + "learning_rate": 5.046772779054622e-06, + "loss": 0.4374, + "step": 6217 + }, + { + "epoch": 3.3756786102062977, + "grad_norm": 9.442731773451024, + "learning_rate": 5.043717975223308e-06, + "loss": 0.3363, + "step": 6218 + }, + { + "epoch": 3.3762214983713354, + "grad_norm": 9.122598498293966, + "learning_rate": 5.040663784392855e-06, + "loss": 0.366, + "step": 6219 + }, + { + "epoch": 3.3767643865363737, + "grad_norm": 11.713343466114539, + "learning_rate": 5.037610206941009e-06, + "loss": 0.4296, + "step": 6220 + }, + { + "epoch": 3.3773072747014115, + "grad_norm": 9.906918527878904, + "learning_rate": 5.034557243245441e-06, + "loss": 0.3189, + "step": 6221 + }, + { + "epoch": 3.3778501628664497, + "grad_norm": 12.91483922171168, + "learning_rate": 5.031504893683748e-06, + "loss": 0.4095, + "step": 6222 + }, + { + "epoch": 3.3783930510314875, + "grad_norm": 12.288487294242882, + "learning_rate": 5.028453158633448e-06, + "loss": 0.5457, + "step": 6223 + }, + { + "epoch": 3.3789359391965257, + "grad_norm": 8.504130528670414, + "learning_rate": 5.025402038471984e-06, + "loss": 0.3373, + "step": 6224 + }, + { + "epoch": 3.3794788273615635, + "grad_norm": 13.190667855419926, + "learning_rate": 5.022351533576725e-06, + "loss": 0.4901, + "step": 6225 + }, + { + "epoch": 3.3800217155266017, + "grad_norm": 15.552140760061034, + "learning_rate": 5.019301644324961e-06, + "loss": 1.0988, + "step": 6226 + }, + { + "epoch": 3.3805646036916395, + "grad_norm": 9.04502851979091, + "learning_rate": 5.016252371093904e-06, + "loss": 0.3088, + "step": 6227 + }, + { + "epoch": 3.3811074918566777, + "grad_norm": 10.595816125042607, + "learning_rate": 5.0132037142607035e-06, + "loss": 0.4645, + "step": 6228 + }, + { + "epoch": 3.3816503800217155, + "grad_norm": 12.60889622464924, + "learning_rate": 5.010155674202409e-06, + "loss": 0.4484, + "step": 6229 + }, + { + "epoch": 3.3821932681867537, + "grad_norm": 11.424705925472448, + "learning_rate": 5.00710825129602e-06, + "loss": 0.3788, + "step": 6230 + }, + { + "epoch": 3.3827361563517915, + "grad_norm": 12.859629128090067, + "learning_rate": 5.004061445918438e-06, + "loss": 0.6222, + "step": 6231 + }, + { + "epoch": 3.3832790445168297, + "grad_norm": 9.436094386037109, + "learning_rate": 5.001015258446497e-06, + "loss": 0.4005, + "step": 6232 + }, + { + "epoch": 3.3838219326818675, + "grad_norm": 13.270237716590554, + "learning_rate": 4.997969689256957e-06, + "loss": 0.5421, + "step": 6233 + }, + { + "epoch": 3.3843648208469057, + "grad_norm": 8.823605358833275, + "learning_rate": 4.994924738726493e-06, + "loss": 0.3196, + "step": 6234 + }, + { + "epoch": 3.3849077090119435, + "grad_norm": 15.35267771565606, + "learning_rate": 4.991880407231722e-06, + "loss": 0.7645, + "step": 6235 + }, + { + "epoch": 3.3854505971769817, + "grad_norm": 12.948816197321962, + "learning_rate": 4.988836695149156e-06, + "loss": 0.4939, + "step": 6236 + }, + { + "epoch": 3.3859934853420195, + "grad_norm": 11.08333997503354, + "learning_rate": 4.985793602855264e-06, + "loss": 0.386, + "step": 6237 + }, + { + "epoch": 3.3865363735070577, + "grad_norm": 11.954902260666392, + "learning_rate": 4.9827511307264006e-06, + "loss": 0.3809, + "step": 6238 + }, + { + "epoch": 3.3870792616720955, + "grad_norm": 13.826055371275897, + "learning_rate": 4.979709279138879e-06, + "loss": 0.488, + "step": 6239 + }, + { + "epoch": 3.3876221498371337, + "grad_norm": 12.084219110446965, + "learning_rate": 4.976668048468918e-06, + "loss": 0.5239, + "step": 6240 + }, + { + "epoch": 3.3881650380021715, + "grad_norm": 10.947811118815906, + "learning_rate": 4.973627439092651e-06, + "loss": 0.3714, + "step": 6241 + }, + { + "epoch": 3.3887079261672097, + "grad_norm": 12.522059184173699, + "learning_rate": 4.97058745138616e-06, + "loss": 0.689, + "step": 6242 + }, + { + "epoch": 3.3892508143322475, + "grad_norm": 10.527706094398768, + "learning_rate": 4.967548085725423e-06, + "loss": 0.6032, + "step": 6243 + }, + { + "epoch": 3.3897937024972857, + "grad_norm": 13.88984081976927, + "learning_rate": 4.964509342486365e-06, + "loss": 0.7107, + "step": 6244 + }, + { + "epoch": 3.3903365906623235, + "grad_norm": 12.254968564513517, + "learning_rate": 4.961471222044811e-06, + "loss": 0.5961, + "step": 6245 + }, + { + "epoch": 3.3908794788273617, + "grad_norm": 14.460789356302293, + "learning_rate": 4.95843372477653e-06, + "loss": 0.4468, + "step": 6246 + }, + { + "epoch": 3.3914223669923995, + "grad_norm": 11.641301023056258, + "learning_rate": 4.955396851057201e-06, + "loss": 0.4814, + "step": 6247 + }, + { + "epoch": 3.3919652551574377, + "grad_norm": 16.08871509775364, + "learning_rate": 4.9523606012624285e-06, + "loss": 0.6589, + "step": 6248 + }, + { + "epoch": 3.3925081433224755, + "grad_norm": 10.999987937636945, + "learning_rate": 4.9493249757677454e-06, + "loss": 0.4282, + "step": 6249 + }, + { + "epoch": 3.3930510314875137, + "grad_norm": 13.140914544731421, + "learning_rate": 4.946289974948591e-06, + "loss": 0.5447, + "step": 6250 + }, + { + "epoch": 3.3935939196525515, + "grad_norm": 16.948288206914583, + "learning_rate": 4.943255599180352e-06, + "loss": 0.3733, + "step": 6251 + }, + { + "epoch": 3.3941368078175898, + "grad_norm": 13.872557767694476, + "learning_rate": 4.940221848838319e-06, + "loss": 0.407, + "step": 6252 + }, + { + "epoch": 3.3946796959826275, + "grad_norm": 12.01068343710924, + "learning_rate": 4.937188724297713e-06, + "loss": 0.3866, + "step": 6253 + }, + { + "epoch": 3.3952225841476658, + "grad_norm": 12.980406743985375, + "learning_rate": 4.934156225933673e-06, + "loss": 0.6519, + "step": 6254 + }, + { + "epoch": 3.3957654723127035, + "grad_norm": 11.103766528107464, + "learning_rate": 4.931124354121265e-06, + "loss": 0.2982, + "step": 6255 + }, + { + "epoch": 3.3963083604777418, + "grad_norm": 7.2969268723114675, + "learning_rate": 4.928093109235476e-06, + "loss": 0.3553, + "step": 6256 + }, + { + "epoch": 3.3968512486427795, + "grad_norm": 8.742701237748966, + "learning_rate": 4.925062491651213e-06, + "loss": 0.3673, + "step": 6257 + }, + { + "epoch": 3.3973941368078178, + "grad_norm": 11.343078244137898, + "learning_rate": 4.922032501743311e-06, + "loss": 0.5849, + "step": 6258 + }, + { + "epoch": 3.3979370249728555, + "grad_norm": 10.498271159334314, + "learning_rate": 4.919003139886522e-06, + "loss": 0.4092, + "step": 6259 + }, + { + "epoch": 3.3984799131378938, + "grad_norm": 10.97821864845177, + "learning_rate": 4.915974406455522e-06, + "loss": 0.3967, + "step": 6260 + }, + { + "epoch": 3.3990228013029316, + "grad_norm": 13.345358070056497, + "learning_rate": 4.912946301824911e-06, + "loss": 0.61, + "step": 6261 + }, + { + "epoch": 3.3995656894679698, + "grad_norm": 11.235526091701708, + "learning_rate": 4.909918826369209e-06, + "loss": 0.577, + "step": 6262 + }, + { + "epoch": 3.4001085776330076, + "grad_norm": 11.63018111198774, + "learning_rate": 4.9068919804628575e-06, + "loss": 0.5611, + "step": 6263 + }, + { + "epoch": 3.400651465798046, + "grad_norm": 14.484517727976659, + "learning_rate": 4.903865764480224e-06, + "loss": 0.8804, + "step": 6264 + }, + { + "epoch": 3.4011943539630836, + "grad_norm": 13.063076605432173, + "learning_rate": 4.9008401787955964e-06, + "loss": 0.6872, + "step": 6265 + }, + { + "epoch": 3.401737242128122, + "grad_norm": 10.26193448056392, + "learning_rate": 4.89781522378318e-06, + "loss": 0.4086, + "step": 6266 + }, + { + "epoch": 3.4022801302931596, + "grad_norm": 14.7618293483203, + "learning_rate": 4.894790899817106e-06, + "loss": 0.5349, + "step": 6267 + }, + { + "epoch": 3.402823018458198, + "grad_norm": 15.246141002277422, + "learning_rate": 4.8917672072714364e-06, + "loss": 0.4647, + "step": 6268 + }, + { + "epoch": 3.4033659066232356, + "grad_norm": 12.999965847014597, + "learning_rate": 4.888744146520137e-06, + "loss": 0.6288, + "step": 6269 + }, + { + "epoch": 3.403908794788274, + "grad_norm": 12.403108581616499, + "learning_rate": 4.885721717937106e-06, + "loss": 0.4168, + "step": 6270 + }, + { + "epoch": 3.4044516829533116, + "grad_norm": 9.010405275529319, + "learning_rate": 4.882699921896166e-06, + "loss": 0.2848, + "step": 6271 + }, + { + "epoch": 3.40499457111835, + "grad_norm": 9.731197165536528, + "learning_rate": 4.87967875877105e-06, + "loss": 0.4276, + "step": 6272 + }, + { + "epoch": 3.4055374592833876, + "grad_norm": 16.141362422365116, + "learning_rate": 4.876658228935434e-06, + "loss": 0.425, + "step": 6273 + }, + { + "epoch": 3.406080347448426, + "grad_norm": 10.09711287157442, + "learning_rate": 4.873638332762887e-06, + "loss": 0.6085, + "step": 6274 + }, + { + "epoch": 3.4066232356134636, + "grad_norm": 15.877829530437113, + "learning_rate": 4.8706190706269276e-06, + "loss": 0.7737, + "step": 6275 + }, + { + "epoch": 3.407166123778502, + "grad_norm": 16.848740635164727, + "learning_rate": 4.867600442900969e-06, + "loss": 0.9117, + "step": 6276 + }, + { + "epoch": 3.4077090119435396, + "grad_norm": 12.439996859861845, + "learning_rate": 4.8645824499583764e-06, + "loss": 0.6655, + "step": 6277 + }, + { + "epoch": 3.408251900108578, + "grad_norm": 15.742400137527083, + "learning_rate": 4.861565092172402e-06, + "loss": 0.4276, + "step": 6278 + }, + { + "epoch": 3.4087947882736156, + "grad_norm": 10.574321270699965, + "learning_rate": 4.8585483699162505e-06, + "loss": 0.3849, + "step": 6279 + }, + { + "epoch": 3.409337676438654, + "grad_norm": 14.691129894472994, + "learning_rate": 4.8555322835630345e-06, + "loss": 0.8061, + "step": 6280 + }, + { + "epoch": 3.4098805646036916, + "grad_norm": 15.252415938946724, + "learning_rate": 4.852516833485778e-06, + "loss": 0.8019, + "step": 6281 + }, + { + "epoch": 3.41042345276873, + "grad_norm": 15.141237929554537, + "learning_rate": 4.849502020057449e-06, + "loss": 0.8201, + "step": 6282 + }, + { + "epoch": 3.4109663409337676, + "grad_norm": 12.751756702153507, + "learning_rate": 4.846487843650914e-06, + "loss": 0.5871, + "step": 6283 + }, + { + "epoch": 3.411509229098806, + "grad_norm": 10.510847411766605, + "learning_rate": 4.843474304638977e-06, + "loss": 0.5543, + "step": 6284 + }, + { + "epoch": 3.4120521172638436, + "grad_norm": 11.747946589267375, + "learning_rate": 4.8404614033943586e-06, + "loss": 0.5981, + "step": 6285 + }, + { + "epoch": 3.412595005428882, + "grad_norm": 10.03814207660896, + "learning_rate": 4.837449140289696e-06, + "loss": 0.3838, + "step": 6286 + }, + { + "epoch": 3.4131378935939196, + "grad_norm": 9.955014298679115, + "learning_rate": 4.8344375156975525e-06, + "loss": 0.4807, + "step": 6287 + }, + { + "epoch": 3.413680781758958, + "grad_norm": 16.70257123694898, + "learning_rate": 4.8314265299904085e-06, + "loss": 0.5431, + "step": 6288 + }, + { + "epoch": 3.4142236699239956, + "grad_norm": 12.494957483834643, + "learning_rate": 4.828416183540668e-06, + "loss": 0.428, + "step": 6289 + }, + { + "epoch": 3.414766558089034, + "grad_norm": 13.556838718061059, + "learning_rate": 4.825406476720658e-06, + "loss": 0.5487, + "step": 6290 + }, + { + "epoch": 3.4153094462540716, + "grad_norm": 11.716720789445835, + "learning_rate": 4.822397409902622e-06, + "loss": 0.8306, + "step": 6291 + }, + { + "epoch": 3.41585233441911, + "grad_norm": 18.495547243648275, + "learning_rate": 4.819388983458725e-06, + "loss": 1.0386, + "step": 6292 + }, + { + "epoch": 3.4163952225841476, + "grad_norm": 11.600851672889053, + "learning_rate": 4.816381197761055e-06, + "loss": 0.3111, + "step": 6293 + }, + { + "epoch": 3.416938110749186, + "grad_norm": 10.513433670937403, + "learning_rate": 4.813374053181621e-06, + "loss": 0.3881, + "step": 6294 + }, + { + "epoch": 3.4174809989142236, + "grad_norm": 12.120445230198527, + "learning_rate": 4.810367550092349e-06, + "loss": 0.6089, + "step": 6295 + }, + { + "epoch": 3.418023887079262, + "grad_norm": 14.617438317762286, + "learning_rate": 4.807361688865091e-06, + "loss": 0.725, + "step": 6296 + }, + { + "epoch": 3.4185667752442996, + "grad_norm": 13.459119769065767, + "learning_rate": 4.804356469871615e-06, + "loss": 0.678, + "step": 6297 + }, + { + "epoch": 3.419109663409338, + "grad_norm": 11.259674657240234, + "learning_rate": 4.801351893483611e-06, + "loss": 0.4629, + "step": 6298 + }, + { + "epoch": 3.4196525515743756, + "grad_norm": 12.169301874189793, + "learning_rate": 4.7983479600726904e-06, + "loss": 0.6371, + "step": 6299 + }, + { + "epoch": 3.420195439739414, + "grad_norm": 10.211316992248761, + "learning_rate": 4.795344670010385e-06, + "loss": 0.591, + "step": 6300 + }, + { + "epoch": 3.4207383279044516, + "grad_norm": 10.221898080949522, + "learning_rate": 4.792342023668144e-06, + "loss": 0.4058, + "step": 6301 + }, + { + "epoch": 3.42128121606949, + "grad_norm": 11.161796287973816, + "learning_rate": 4.789340021417343e-06, + "loss": 0.6044, + "step": 6302 + }, + { + "epoch": 3.4218241042345277, + "grad_norm": 13.57180649881648, + "learning_rate": 4.7863386636292705e-06, + "loss": 0.5895, + "step": 6303 + }, + { + "epoch": 3.422366992399566, + "grad_norm": 14.249903450132871, + "learning_rate": 4.783337950675143e-06, + "loss": 0.9115, + "step": 6304 + }, + { + "epoch": 3.4229098805646037, + "grad_norm": 8.778670420410771, + "learning_rate": 4.780337882926088e-06, + "loss": 0.368, + "step": 6305 + }, + { + "epoch": 3.423452768729642, + "grad_norm": 14.418437807091719, + "learning_rate": 4.77733846075317e-06, + "loss": 0.7248, + "step": 6306 + }, + { + "epoch": 3.4239956568946797, + "grad_norm": 12.207383551156253, + "learning_rate": 4.774339684527348e-06, + "loss": 0.715, + "step": 6307 + }, + { + "epoch": 3.424538545059718, + "grad_norm": 12.625928846965705, + "learning_rate": 4.7713415546195285e-06, + "loss": 0.4691, + "step": 6308 + }, + { + "epoch": 3.4250814332247557, + "grad_norm": 12.54338617690595, + "learning_rate": 4.768344071400516e-06, + "loss": 0.7746, + "step": 6309 + }, + { + "epoch": 3.425624321389794, + "grad_norm": 12.757091279673624, + "learning_rate": 4.765347235241042e-06, + "loss": 0.5129, + "step": 6310 + }, + { + "epoch": 3.4261672095548317, + "grad_norm": 14.719338319419373, + "learning_rate": 4.762351046511774e-06, + "loss": 0.6281, + "step": 6311 + }, + { + "epoch": 3.42671009771987, + "grad_norm": 14.600430485134494, + "learning_rate": 4.759355505583267e-06, + "loss": 0.6332, + "step": 6312 + }, + { + "epoch": 3.4272529858849077, + "grad_norm": 12.499065575741712, + "learning_rate": 4.756360612826032e-06, + "loss": 0.5188, + "step": 6313 + }, + { + "epoch": 3.427795874049946, + "grad_norm": 13.595282233685417, + "learning_rate": 4.753366368610466e-06, + "loss": 0.6934, + "step": 6314 + }, + { + "epoch": 3.4283387622149837, + "grad_norm": 18.513402859066886, + "learning_rate": 4.750372773306916e-06, + "loss": 0.8336, + "step": 6315 + }, + { + "epoch": 3.428881650380022, + "grad_norm": 12.541674490834746, + "learning_rate": 4.747379827285621e-06, + "loss": 0.4741, + "step": 6316 + }, + { + "epoch": 3.4294245385450597, + "grad_norm": 13.06139830486867, + "learning_rate": 4.744387530916764e-06, + "loss": 0.6489, + "step": 6317 + }, + { + "epoch": 3.429967426710098, + "grad_norm": 9.57665402322532, + "learning_rate": 4.741395884570437e-06, + "loss": 0.4976, + "step": 6318 + }, + { + "epoch": 3.4305103148751357, + "grad_norm": 9.338701533011545, + "learning_rate": 4.738404888616641e-06, + "loss": 0.4947, + "step": 6319 + }, + { + "epoch": 3.431053203040174, + "grad_norm": 13.17347762890872, + "learning_rate": 4.735414543425321e-06, + "loss": 0.5455, + "step": 6320 + }, + { + "epoch": 3.4315960912052117, + "grad_norm": 14.408997044186131, + "learning_rate": 4.732424849366314e-06, + "loss": 0.5761, + "step": 6321 + }, + { + "epoch": 3.43213897937025, + "grad_norm": 10.201121637580389, + "learning_rate": 4.729435806809401e-06, + "loss": 0.4271, + "step": 6322 + }, + { + "epoch": 3.4326818675352877, + "grad_norm": 11.714865692839483, + "learning_rate": 4.726447416124266e-06, + "loss": 0.4182, + "step": 6323 + }, + { + "epoch": 3.433224755700326, + "grad_norm": 21.671744658819392, + "learning_rate": 4.72345967768052e-06, + "loss": 1.1518, + "step": 6324 + }, + { + "epoch": 3.4337676438653637, + "grad_norm": 10.904130560504068, + "learning_rate": 4.72047259184769e-06, + "loss": 0.533, + "step": 6325 + }, + { + "epoch": 3.434310532030402, + "grad_norm": 17.26082351569084, + "learning_rate": 4.717486158995225e-06, + "loss": 0.5579, + "step": 6326 + }, + { + "epoch": 3.4348534201954397, + "grad_norm": 13.562029451071414, + "learning_rate": 4.7145003794924905e-06, + "loss": 0.6275, + "step": 6327 + }, + { + "epoch": 3.435396308360478, + "grad_norm": 10.679128408418185, + "learning_rate": 4.711515253708774e-06, + "loss": 0.3116, + "step": 6328 + }, + { + "epoch": 3.4359391965255157, + "grad_norm": 6.813829973820327, + "learning_rate": 4.708530782013277e-06, + "loss": 0.2313, + "step": 6329 + }, + { + "epoch": 3.436482084690554, + "grad_norm": 10.696300631140947, + "learning_rate": 4.705546964775128e-06, + "loss": 0.4345, + "step": 6330 + }, + { + "epoch": 3.4370249728555917, + "grad_norm": 15.665510743577896, + "learning_rate": 4.702563802363369e-06, + "loss": 0.6155, + "step": 6331 + }, + { + "epoch": 3.4375678610206295, + "grad_norm": 9.437457272265185, + "learning_rate": 4.699581295146961e-06, + "loss": 0.3972, + "step": 6332 + }, + { + "epoch": 3.4381107491856677, + "grad_norm": 11.253233608357364, + "learning_rate": 4.696599443494787e-06, + "loss": 0.4871, + "step": 6333 + }, + { + "epoch": 3.438653637350706, + "grad_norm": 13.323571403470238, + "learning_rate": 4.693618247775645e-06, + "loss": 0.4443, + "step": 6334 + }, + { + "epoch": 3.4391965255157437, + "grad_norm": 10.853976251462788, + "learning_rate": 4.6906377083582556e-06, + "loss": 0.4012, + "step": 6335 + }, + { + "epoch": 3.4397394136807815, + "grad_norm": 11.483324693854579, + "learning_rate": 4.687657825611256e-06, + "loss": 0.4913, + "step": 6336 + }, + { + "epoch": 3.4402823018458197, + "grad_norm": 16.90515830826938, + "learning_rate": 4.684678599903204e-06, + "loss": 0.5872, + "step": 6337 + }, + { + "epoch": 3.440825190010858, + "grad_norm": 11.339161987387357, + "learning_rate": 4.681700031602573e-06, + "loss": 0.484, + "step": 6338 + }, + { + "epoch": 3.4413680781758957, + "grad_norm": 13.116517374892055, + "learning_rate": 4.678722121077759e-06, + "loss": 0.8588, + "step": 6339 + }, + { + "epoch": 3.4419109663409335, + "grad_norm": 10.313990325741717, + "learning_rate": 4.675744868697073e-06, + "loss": 0.4864, + "step": 6340 + }, + { + "epoch": 3.4424538545059717, + "grad_norm": 12.464768763015064, + "learning_rate": 4.672768274828748e-06, + "loss": 0.7793, + "step": 6341 + }, + { + "epoch": 3.44299674267101, + "grad_norm": 7.491565779729114, + "learning_rate": 4.669792339840933e-06, + "loss": 0.2551, + "step": 6342 + }, + { + "epoch": 3.4435396308360477, + "grad_norm": 11.582532266159841, + "learning_rate": 4.666817064101693e-06, + "loss": 0.4687, + "step": 6343 + }, + { + "epoch": 3.4440825190010855, + "grad_norm": 13.223648734250315, + "learning_rate": 4.663842447979026e-06, + "loss": 0.5676, + "step": 6344 + }, + { + "epoch": 3.4446254071661238, + "grad_norm": 9.420907388669441, + "learning_rate": 4.660868491840821e-06, + "loss": 0.4091, + "step": 6345 + }, + { + "epoch": 3.445168295331162, + "grad_norm": 10.937341000869324, + "learning_rate": 4.657895196054919e-06, + "loss": 0.4299, + "step": 6346 + }, + { + "epoch": 3.4457111834961998, + "grad_norm": 10.997320970308403, + "learning_rate": 4.654922560989049e-06, + "loss": 0.4176, + "step": 6347 + }, + { + "epoch": 3.4462540716612375, + "grad_norm": 11.008951815202312, + "learning_rate": 4.651950587010875e-06, + "loss": 0.3343, + "step": 6348 + }, + { + "epoch": 3.4467969598262758, + "grad_norm": 13.714805717613569, + "learning_rate": 4.6489792744879755e-06, + "loss": 0.8672, + "step": 6349 + }, + { + "epoch": 3.447339847991314, + "grad_norm": 12.722366469633661, + "learning_rate": 4.646008623787845e-06, + "loss": 0.601, + "step": 6350 + }, + { + "epoch": 3.4478827361563518, + "grad_norm": 10.3662276435613, + "learning_rate": 4.643038635277908e-06, + "loss": 0.4597, + "step": 6351 + }, + { + "epoch": 3.4484256243213895, + "grad_norm": 11.28259690900768, + "learning_rate": 4.640069309325484e-06, + "loss": 0.4909, + "step": 6352 + }, + { + "epoch": 3.4489685124864278, + "grad_norm": 15.163346859296523, + "learning_rate": 4.6371006462978355e-06, + "loss": 0.8422, + "step": 6353 + }, + { + "epoch": 3.449511400651466, + "grad_norm": 9.625566221898923, + "learning_rate": 4.634132646562119e-06, + "loss": 0.2858, + "step": 6354 + }, + { + "epoch": 3.450054288816504, + "grad_norm": 14.927715657666415, + "learning_rate": 4.631165310485434e-06, + "loss": 0.5556, + "step": 6355 + }, + { + "epoch": 3.4505971769815416, + "grad_norm": 14.575103603282974, + "learning_rate": 4.62819863843478e-06, + "loss": 0.5422, + "step": 6356 + }, + { + "epoch": 3.45114006514658, + "grad_norm": 11.84763157429604, + "learning_rate": 4.625232630777079e-06, + "loss": 0.5382, + "step": 6357 + }, + { + "epoch": 3.451682953311618, + "grad_norm": 14.334352386311854, + "learning_rate": 4.622267287879176e-06, + "loss": 0.6771, + "step": 6358 + }, + { + "epoch": 3.452225841476656, + "grad_norm": 13.03121716154961, + "learning_rate": 4.619302610107819e-06, + "loss": 0.702, + "step": 6359 + }, + { + "epoch": 3.4527687296416936, + "grad_norm": 9.331217131822793, + "learning_rate": 4.616338597829697e-06, + "loss": 0.4664, + "step": 6360 + }, + { + "epoch": 3.453311617806732, + "grad_norm": 16.384420267273573, + "learning_rate": 4.61337525141139e-06, + "loss": 1.1202, + "step": 6361 + }, + { + "epoch": 3.45385450597177, + "grad_norm": 12.656834373054709, + "learning_rate": 4.610412571219421e-06, + "loss": 0.8405, + "step": 6362 + }, + { + "epoch": 3.454397394136808, + "grad_norm": 10.6285388663689, + "learning_rate": 4.607450557620216e-06, + "loss": 0.4027, + "step": 6363 + }, + { + "epoch": 3.4549402823018456, + "grad_norm": 14.149860401799353, + "learning_rate": 4.60448921098012e-06, + "loss": 0.5762, + "step": 6364 + }, + { + "epoch": 3.455483170466884, + "grad_norm": 15.249693808816065, + "learning_rate": 4.601528531665397e-06, + "loss": 1.0724, + "step": 6365 + }, + { + "epoch": 3.456026058631922, + "grad_norm": 11.967042611547605, + "learning_rate": 4.598568520042229e-06, + "loss": 0.572, + "step": 6366 + }, + { + "epoch": 3.45656894679696, + "grad_norm": 13.806580531191415, + "learning_rate": 4.595609176476715e-06, + "loss": 0.7437, + "step": 6367 + }, + { + "epoch": 3.4571118349619976, + "grad_norm": 15.062073427987208, + "learning_rate": 4.592650501334872e-06, + "loss": 0.6357, + "step": 6368 + }, + { + "epoch": 3.457654723127036, + "grad_norm": 8.804207344489894, + "learning_rate": 4.589692494982632e-06, + "loss": 0.3561, + "step": 6369 + }, + { + "epoch": 3.458197611292074, + "grad_norm": 13.940651501182272, + "learning_rate": 4.5867351577858475e-06, + "loss": 0.6055, + "step": 6370 + }, + { + "epoch": 3.458740499457112, + "grad_norm": 12.874778392486414, + "learning_rate": 4.583778490110287e-06, + "loss": 0.466, + "step": 6371 + }, + { + "epoch": 3.4592833876221496, + "grad_norm": 13.362591165530915, + "learning_rate": 4.580822492321634e-06, + "loss": 0.837, + "step": 6372 + }, + { + "epoch": 3.459826275787188, + "grad_norm": 13.501338998690153, + "learning_rate": 4.577867164785492e-06, + "loss": 0.4919, + "step": 6373 + }, + { + "epoch": 3.460369163952226, + "grad_norm": 8.394152440486542, + "learning_rate": 4.574912507867382e-06, + "loss": 0.3555, + "step": 6374 + }, + { + "epoch": 3.460912052117264, + "grad_norm": 11.611965148146934, + "learning_rate": 4.571958521932738e-06, + "loss": 0.4808, + "step": 6375 + }, + { + "epoch": 3.4614549402823016, + "grad_norm": 11.297446435669148, + "learning_rate": 4.569005207346911e-06, + "loss": 0.4093, + "step": 6376 + }, + { + "epoch": 3.46199782844734, + "grad_norm": 13.040685712021983, + "learning_rate": 4.566052564475184e-06, + "loss": 0.5487, + "step": 6377 + }, + { + "epoch": 3.462540716612378, + "grad_norm": 10.160191141450042, + "learning_rate": 4.563100593682732e-06, + "loss": 0.5555, + "step": 6378 + }, + { + "epoch": 3.463083604777416, + "grad_norm": 10.641601715805423, + "learning_rate": 4.560149295334664e-06, + "loss": 0.4319, + "step": 6379 + }, + { + "epoch": 3.4636264929424536, + "grad_norm": 11.742982035764232, + "learning_rate": 4.557198669796001e-06, + "loss": 0.77, + "step": 6380 + }, + { + "epoch": 3.464169381107492, + "grad_norm": 15.297545906946976, + "learning_rate": 4.554248717431678e-06, + "loss": 0.753, + "step": 6381 + }, + { + "epoch": 3.46471226927253, + "grad_norm": 11.061047467202666, + "learning_rate": 4.551299438606559e-06, + "loss": 0.428, + "step": 6382 + }, + { + "epoch": 3.465255157437568, + "grad_norm": 12.238086621939612, + "learning_rate": 4.548350833685402e-06, + "loss": 0.4874, + "step": 6383 + }, + { + "epoch": 3.4657980456026056, + "grad_norm": 13.316349313453275, + "learning_rate": 4.54540290303291e-06, + "loss": 0.7391, + "step": 6384 + }, + { + "epoch": 3.466340933767644, + "grad_norm": 14.524029697923662, + "learning_rate": 4.5424556470136735e-06, + "loss": 0.6782, + "step": 6385 + }, + { + "epoch": 3.466883821932682, + "grad_norm": 11.699705038001243, + "learning_rate": 4.539509065992229e-06, + "loss": 0.6561, + "step": 6386 + }, + { + "epoch": 3.46742671009772, + "grad_norm": 16.5251401693806, + "learning_rate": 4.536563160333001e-06, + "loss": 0.6788, + "step": 6387 + }, + { + "epoch": 3.4679695982627576, + "grad_norm": 8.228129211788293, + "learning_rate": 4.533617930400345e-06, + "loss": 0.4383, + "step": 6388 + }, + { + "epoch": 3.468512486427796, + "grad_norm": 10.016870671639179, + "learning_rate": 4.530673376558543e-06, + "loss": 0.5374, + "step": 6389 + }, + { + "epoch": 3.469055374592834, + "grad_norm": 12.529286864818076, + "learning_rate": 4.527729499171767e-06, + "loss": 0.3919, + "step": 6390 + }, + { + "epoch": 3.469598262757872, + "grad_norm": 13.31372906596201, + "learning_rate": 4.524786298604136e-06, + "loss": 0.7436, + "step": 6391 + }, + { + "epoch": 3.4701411509229096, + "grad_norm": 11.420174160084546, + "learning_rate": 4.521843775219654e-06, + "loss": 0.4285, + "step": 6392 + }, + { + "epoch": 3.470684039087948, + "grad_norm": 11.461560335128281, + "learning_rate": 4.518901929382267e-06, + "loss": 0.5651, + "step": 6393 + }, + { + "epoch": 3.471226927252986, + "grad_norm": 14.298269585226816, + "learning_rate": 4.515960761455827e-06, + "loss": 0.3989, + "step": 6394 + }, + { + "epoch": 3.471769815418024, + "grad_norm": 12.421864505486035, + "learning_rate": 4.5130202718041004e-06, + "loss": 0.8773, + "step": 6395 + }, + { + "epoch": 3.4723127035830617, + "grad_norm": 12.947288721903634, + "learning_rate": 4.510080460790775e-06, + "loss": 0.5396, + "step": 6396 + }, + { + "epoch": 3.4728555917481, + "grad_norm": 10.339717202308544, + "learning_rate": 4.507141328779439e-06, + "loss": 0.5154, + "step": 6397 + }, + { + "epoch": 3.473398479913138, + "grad_norm": 8.810723375318315, + "learning_rate": 4.504202876133627e-06, + "loss": 0.4513, + "step": 6398 + }, + { + "epoch": 3.473941368078176, + "grad_norm": 13.655631619886284, + "learning_rate": 4.501265103216755e-06, + "loss": 0.5043, + "step": 6399 + }, + { + "epoch": 3.4744842562432137, + "grad_norm": 9.947920274498516, + "learning_rate": 4.49832801039218e-06, + "loss": 0.6907, + "step": 6400 + }, + { + "epoch": 3.475027144408252, + "grad_norm": 9.800586651795994, + "learning_rate": 4.495391598023167e-06, + "loss": 0.3988, + "step": 6401 + }, + { + "epoch": 3.47557003257329, + "grad_norm": 14.696382692377027, + "learning_rate": 4.492455866472894e-06, + "loss": 0.5805, + "step": 6402 + }, + { + "epoch": 3.476112920738328, + "grad_norm": 13.06051257277803, + "learning_rate": 4.489520816104455e-06, + "loss": 0.7204, + "step": 6403 + }, + { + "epoch": 3.4766558089033657, + "grad_norm": 12.99399781439784, + "learning_rate": 4.486586447280866e-06, + "loss": 0.7112, + "step": 6404 + }, + { + "epoch": 3.477198697068404, + "grad_norm": 12.422813336616981, + "learning_rate": 4.483652760365052e-06, + "loss": 0.3927, + "step": 6405 + }, + { + "epoch": 3.477741585233442, + "grad_norm": 10.948217532391471, + "learning_rate": 4.480719755719857e-06, + "loss": 0.5587, + "step": 6406 + }, + { + "epoch": 3.47828447339848, + "grad_norm": 11.657205565439646, + "learning_rate": 4.477787433708038e-06, + "loss": 0.5584, + "step": 6407 + }, + { + "epoch": 3.4788273615635177, + "grad_norm": 7.795661051863107, + "learning_rate": 4.474855794692271e-06, + "loss": 0.3232, + "step": 6408 + }, + { + "epoch": 3.479370249728556, + "grad_norm": 7.775932230943303, + "learning_rate": 4.4719248390351446e-06, + "loss": 0.3613, + "step": 6409 + }, + { + "epoch": 3.479913137893594, + "grad_norm": 9.097718161588768, + "learning_rate": 4.468994567099165e-06, + "loss": 0.3896, + "step": 6410 + }, + { + "epoch": 3.480456026058632, + "grad_norm": 7.087231016744389, + "learning_rate": 4.466064979246751e-06, + "loss": 0.2486, + "step": 6411 + }, + { + "epoch": 3.4809989142236697, + "grad_norm": 17.009119363822553, + "learning_rate": 4.463136075840242e-06, + "loss": 0.7777, + "step": 6412 + }, + { + "epoch": 3.481541802388708, + "grad_norm": 11.54915190749092, + "learning_rate": 4.460207857241887e-06, + "loss": 0.4528, + "step": 6413 + }, + { + "epoch": 3.482084690553746, + "grad_norm": 14.237806827589361, + "learning_rate": 4.45728032381385e-06, + "loss": 0.8941, + "step": 6414 + }, + { + "epoch": 3.482627578718784, + "grad_norm": 11.85837340138276, + "learning_rate": 4.454353475918223e-06, + "loss": 0.5289, + "step": 6415 + }, + { + "epoch": 3.4831704668838217, + "grad_norm": 13.987937632959078, + "learning_rate": 4.4514273139169925e-06, + "loss": 0.5152, + "step": 6416 + }, + { + "epoch": 3.48371335504886, + "grad_norm": 10.643372071284285, + "learning_rate": 4.4485018381720755e-06, + "loss": 0.4758, + "step": 6417 + }, + { + "epoch": 3.484256243213898, + "grad_norm": 8.692254926051978, + "learning_rate": 4.445577049045299e-06, + "loss": 0.382, + "step": 6418 + }, + { + "epoch": 3.484799131378936, + "grad_norm": 9.767782914786244, + "learning_rate": 4.4426529468984055e-06, + "loss": 0.4674, + "step": 6419 + }, + { + "epoch": 3.4853420195439737, + "grad_norm": 14.118719943852668, + "learning_rate": 4.4397295320930525e-06, + "loss": 0.5859, + "step": 6420 + }, + { + "epoch": 3.485884907709012, + "grad_norm": 9.120918716734044, + "learning_rate": 4.4368068049908085e-06, + "loss": 0.37, + "step": 6421 + }, + { + "epoch": 3.48642779587405, + "grad_norm": 12.706215416375834, + "learning_rate": 4.4338847659531735e-06, + "loss": 0.8322, + "step": 6422 + }, + { + "epoch": 3.486970684039088, + "grad_norm": 10.926968726715666, + "learning_rate": 4.430963415341533e-06, + "loss": 0.4146, + "step": 6423 + }, + { + "epoch": 3.4875135722041257, + "grad_norm": 14.53230992435977, + "learning_rate": 4.428042753517222e-06, + "loss": 0.5612, + "step": 6424 + }, + { + "epoch": 3.488056460369164, + "grad_norm": 12.614428085329433, + "learning_rate": 4.425122780841456e-06, + "loss": 0.4886, + "step": 6425 + }, + { + "epoch": 3.488599348534202, + "grad_norm": 9.07248478126251, + "learning_rate": 4.422203497675394e-06, + "loss": 0.3342, + "step": 6426 + }, + { + "epoch": 3.48914223669924, + "grad_norm": 7.63850442052429, + "learning_rate": 4.419284904380095e-06, + "loss": 0.2709, + "step": 6427 + }, + { + "epoch": 3.4896851248642777, + "grad_norm": 8.139657158702438, + "learning_rate": 4.416367001316526e-06, + "loss": 0.3276, + "step": 6428 + }, + { + "epoch": 3.490228013029316, + "grad_norm": 12.800427760538255, + "learning_rate": 4.413449788845594e-06, + "loss": 0.8097, + "step": 6429 + }, + { + "epoch": 3.490770901194354, + "grad_norm": 9.88357123496715, + "learning_rate": 4.410533267328087e-06, + "loss": 0.3224, + "step": 6430 + }, + { + "epoch": 3.491313789359392, + "grad_norm": 10.15573231324521, + "learning_rate": 4.407617437124741e-06, + "loss": 0.5535, + "step": 6431 + }, + { + "epoch": 3.4918566775244297, + "grad_norm": 17.972024507407102, + "learning_rate": 4.404702298596177e-06, + "loss": 0.9413, + "step": 6432 + }, + { + "epoch": 3.492399565689468, + "grad_norm": 11.366140193603957, + "learning_rate": 4.401787852102955e-06, + "loss": 0.7224, + "step": 6433 + }, + { + "epoch": 3.492942453854506, + "grad_norm": 11.183293759896978, + "learning_rate": 4.398874098005532e-06, + "loss": 0.6597, + "step": 6434 + }, + { + "epoch": 3.493485342019544, + "grad_norm": 11.409722406543372, + "learning_rate": 4.395961036664288e-06, + "loss": 0.6165, + "step": 6435 + }, + { + "epoch": 3.4940282301845818, + "grad_norm": 8.532518886519954, + "learning_rate": 4.393048668439518e-06, + "loss": 0.3045, + "step": 6436 + }, + { + "epoch": 3.49457111834962, + "grad_norm": 15.499278790978442, + "learning_rate": 4.390136993691417e-06, + "loss": 0.6505, + "step": 6437 + }, + { + "epoch": 3.495114006514658, + "grad_norm": 9.855565799480534, + "learning_rate": 4.387226012780117e-06, + "loss": 0.4787, + "step": 6438 + }, + { + "epoch": 3.495656894679696, + "grad_norm": 13.510584259969823, + "learning_rate": 4.38431572606565e-06, + "loss": 0.5388, + "step": 6439 + }, + { + "epoch": 3.4961997828447338, + "grad_norm": 8.040242399067179, + "learning_rate": 4.381406133907964e-06, + "loss": 0.3596, + "step": 6440 + }, + { + "epoch": 3.496742671009772, + "grad_norm": 9.402081395938351, + "learning_rate": 4.378497236666922e-06, + "loss": 0.3326, + "step": 6441 + }, + { + "epoch": 3.49728555917481, + "grad_norm": 15.188738381211488, + "learning_rate": 4.3755890347023e-06, + "loss": 0.7356, + "step": 6442 + }, + { + "epoch": 3.497828447339848, + "grad_norm": 13.644691532416777, + "learning_rate": 4.372681528373791e-06, + "loss": 0.5727, + "step": 6443 + }, + { + "epoch": 3.4983713355048858, + "grad_norm": 12.036428423892223, + "learning_rate": 4.369774718041e-06, + "loss": 0.4333, + "step": 6444 + }, + { + "epoch": 3.498914223669924, + "grad_norm": 9.814264562422439, + "learning_rate": 4.366868604063444e-06, + "loss": 0.4396, + "step": 6445 + }, + { + "epoch": 3.499457111834962, + "grad_norm": 10.623481464500886, + "learning_rate": 4.363963186800557e-06, + "loss": 0.509, + "step": 6446 + }, + { + "epoch": 3.5, + "grad_norm": 9.931711278540687, + "learning_rate": 4.361058466611686e-06, + "loss": 0.352, + "step": 6447 + }, + { + "epoch": 3.500542888165038, + "grad_norm": 10.709530760490054, + "learning_rate": 4.358154443856091e-06, + "loss": 0.3233, + "step": 6448 + }, + { + "epoch": 3.501085776330076, + "grad_norm": 12.489815680915486, + "learning_rate": 4.355251118892946e-06, + "loss": 0.8317, + "step": 6449 + }, + { + "epoch": 3.5016286644951142, + "grad_norm": 13.400967427781875, + "learning_rate": 4.35234849208134e-06, + "loss": 0.601, + "step": 6450 + }, + { + "epoch": 3.502171552660152, + "grad_norm": 11.710683499628859, + "learning_rate": 4.349446563780272e-06, + "loss": 0.7362, + "step": 6451 + }, + { + "epoch": 3.50271444082519, + "grad_norm": 14.51448957958314, + "learning_rate": 4.346545334348658e-06, + "loss": 0.8007, + "step": 6452 + }, + { + "epoch": 3.503257328990228, + "grad_norm": 10.464776296844045, + "learning_rate": 4.343644804145329e-06, + "loss": 0.3993, + "step": 6453 + }, + { + "epoch": 3.5038002171552662, + "grad_norm": 9.718005412495163, + "learning_rate": 4.340744973529022e-06, + "loss": 0.269, + "step": 6454 + }, + { + "epoch": 3.504343105320304, + "grad_norm": 10.85577962939331, + "learning_rate": 4.337845842858402e-06, + "loss": 0.5342, + "step": 6455 + }, + { + "epoch": 3.504885993485342, + "grad_norm": 12.181308768808934, + "learning_rate": 4.33494741249203e-06, + "loss": 0.4005, + "step": 6456 + }, + { + "epoch": 3.50542888165038, + "grad_norm": 11.779746087085687, + "learning_rate": 4.332049682788391e-06, + "loss": 0.4414, + "step": 6457 + }, + { + "epoch": 3.5059717698154182, + "grad_norm": 11.664912784701475, + "learning_rate": 4.3291526541058795e-06, + "loss": 0.548, + "step": 6458 + }, + { + "epoch": 3.506514657980456, + "grad_norm": 13.373343929291233, + "learning_rate": 4.326256326802802e-06, + "loss": 0.564, + "step": 6459 + }, + { + "epoch": 3.507057546145494, + "grad_norm": 11.223466720082536, + "learning_rate": 4.3233607012373925e-06, + "loss": 0.3611, + "step": 6460 + }, + { + "epoch": 3.507600434310532, + "grad_norm": 13.530570397967848, + "learning_rate": 4.32046577776777e-06, + "loss": 0.6113, + "step": 6461 + }, + { + "epoch": 3.5081433224755703, + "grad_norm": 12.62657182669542, + "learning_rate": 4.3175715567520015e-06, + "loss": 0.4457, + "step": 6462 + }, + { + "epoch": 3.508686210640608, + "grad_norm": 14.69985392835765, + "learning_rate": 4.31467803854803e-06, + "loss": 0.8015, + "step": 6463 + }, + { + "epoch": 3.509229098805646, + "grad_norm": 12.910242277352095, + "learning_rate": 4.311785223513744e-06, + "loss": 0.5957, + "step": 6464 + }, + { + "epoch": 3.509771986970684, + "grad_norm": 14.38667526612287, + "learning_rate": 4.308893112006932e-06, + "loss": 0.6333, + "step": 6465 + }, + { + "epoch": 3.5103148751357223, + "grad_norm": 10.811689404016661, + "learning_rate": 4.306001704385282e-06, + "loss": 0.379, + "step": 6466 + }, + { + "epoch": 3.51085776330076, + "grad_norm": 12.950655194202454, + "learning_rate": 4.303111001006424e-06, + "loss": 0.3791, + "step": 6467 + }, + { + "epoch": 3.511400651465798, + "grad_norm": 10.403413919758718, + "learning_rate": 4.300221002227869e-06, + "loss": 0.5023, + "step": 6468 + }, + { + "epoch": 3.511943539630836, + "grad_norm": 14.86724875684143, + "learning_rate": 4.297331708407072e-06, + "loss": 0.5985, + "step": 6469 + }, + { + "epoch": 3.5124864277958743, + "grad_norm": 9.211711120681478, + "learning_rate": 4.29444311990137e-06, + "loss": 0.2738, + "step": 6470 + }, + { + "epoch": 3.513029315960912, + "grad_norm": 11.398896094597806, + "learning_rate": 4.29155523706804e-06, + "loss": 0.6349, + "step": 6471 + }, + { + "epoch": 3.51357220412595, + "grad_norm": 9.97395797119566, + "learning_rate": 4.288668060264257e-06, + "loss": 0.3945, + "step": 6472 + }, + { + "epoch": 3.514115092290988, + "grad_norm": 10.965919373159068, + "learning_rate": 4.2857815898471114e-06, + "loss": 0.6313, + "step": 6473 + }, + { + "epoch": 3.5146579804560263, + "grad_norm": 14.272003668320222, + "learning_rate": 4.2828958261736045e-06, + "loss": 0.8566, + "step": 6474 + }, + { + "epoch": 3.515200868621064, + "grad_norm": 9.155542594309077, + "learning_rate": 4.280010769600653e-06, + "loss": 0.4155, + "step": 6475 + }, + { + "epoch": 3.515743756786102, + "grad_norm": 12.664316762774577, + "learning_rate": 4.277126420485087e-06, + "loss": 0.4668, + "step": 6476 + }, + { + "epoch": 3.51628664495114, + "grad_norm": 13.218980598800185, + "learning_rate": 4.274242779183646e-06, + "loss": 0.5638, + "step": 6477 + }, + { + "epoch": 3.5168295331161783, + "grad_norm": 10.791302144957823, + "learning_rate": 4.271359846052983e-06, + "loss": 0.5458, + "step": 6478 + }, + { + "epoch": 3.517372421281216, + "grad_norm": 11.294781461737724, + "learning_rate": 4.268477621449666e-06, + "loss": 0.5746, + "step": 6479 + }, + { + "epoch": 3.517915309446254, + "grad_norm": 8.318333652865173, + "learning_rate": 4.2655961057301695e-06, + "loss": 0.3447, + "step": 6480 + }, + { + "epoch": 3.518458197611292, + "grad_norm": 14.570486409127577, + "learning_rate": 4.2627152992508865e-06, + "loss": 0.6757, + "step": 6481 + }, + { + "epoch": 3.5190010857763303, + "grad_norm": 12.036103572116089, + "learning_rate": 4.25983520236812e-06, + "loss": 0.4259, + "step": 6482 + }, + { + "epoch": 3.519543973941368, + "grad_norm": 9.254932202972295, + "learning_rate": 4.256955815438084e-06, + "loss": 0.4689, + "step": 6483 + }, + { + "epoch": 3.520086862106406, + "grad_norm": 10.562719431036207, + "learning_rate": 4.254077138816905e-06, + "loss": 0.3951, + "step": 6484 + }, + { + "epoch": 3.520629750271444, + "grad_norm": 10.475306089406416, + "learning_rate": 4.251199172860624e-06, + "loss": 0.6583, + "step": 6485 + }, + { + "epoch": 3.5211726384364823, + "grad_norm": 11.146639535946015, + "learning_rate": 4.2483219179251915e-06, + "loss": 0.5873, + "step": 6486 + }, + { + "epoch": 3.52171552660152, + "grad_norm": 13.607190728220559, + "learning_rate": 4.245445374366469e-06, + "loss": 0.5111, + "step": 6487 + }, + { + "epoch": 3.522258414766558, + "grad_norm": 13.73052324797901, + "learning_rate": 4.242569542540236e-06, + "loss": 0.6546, + "step": 6488 + }, + { + "epoch": 3.522801302931596, + "grad_norm": 11.846325790998266, + "learning_rate": 4.239694422802177e-06, + "loss": 0.3475, + "step": 6489 + }, + { + "epoch": 3.5233441910966343, + "grad_norm": 9.944462539493102, + "learning_rate": 4.236820015507893e-06, + "loss": 0.3798, + "step": 6490 + }, + { + "epoch": 3.523887079261672, + "grad_norm": 11.901268081862414, + "learning_rate": 4.233946321012895e-06, + "loss": 0.503, + "step": 6491 + }, + { + "epoch": 3.52442996742671, + "grad_norm": 11.373264829742677, + "learning_rate": 4.231073339672601e-06, + "loss": 0.3985, + "step": 6492 + }, + { + "epoch": 3.524972855591748, + "grad_norm": 8.462656388647403, + "learning_rate": 4.228201071842358e-06, + "loss": 0.3526, + "step": 6493 + }, + { + "epoch": 3.5255157437567863, + "grad_norm": 13.319319520255643, + "learning_rate": 4.2253295178774036e-06, + "loss": 0.3119, + "step": 6494 + }, + { + "epoch": 3.526058631921824, + "grad_norm": 11.873988390722962, + "learning_rate": 4.222458678132896e-06, + "loss": 0.3681, + "step": 6495 + }, + { + "epoch": 3.526601520086862, + "grad_norm": 14.615618896842006, + "learning_rate": 4.219588552963907e-06, + "loss": 0.5678, + "step": 6496 + }, + { + "epoch": 3.5271444082519, + "grad_norm": 12.902856396087687, + "learning_rate": 4.216719142725415e-06, + "loss": 0.4386, + "step": 6497 + }, + { + "epoch": 3.5276872964169383, + "grad_norm": 14.124175164997677, + "learning_rate": 4.213850447772324e-06, + "loss": 0.752, + "step": 6498 + }, + { + "epoch": 3.528230184581976, + "grad_norm": 17.881101453026922, + "learning_rate": 4.210982468459425e-06, + "loss": 1.0417, + "step": 6499 + }, + { + "epoch": 3.528773072747014, + "grad_norm": 12.44176561816019, + "learning_rate": 4.2081152051414464e-06, + "loss": 0.5422, + "step": 6500 + }, + { + "epoch": 3.529315960912052, + "grad_norm": 11.595014008440414, + "learning_rate": 4.205248658173005e-06, + "loss": 0.5242, + "step": 6501 + }, + { + "epoch": 3.5298588490770904, + "grad_norm": 13.42551753781462, + "learning_rate": 4.20238282790865e-06, + "loss": 0.4613, + "step": 6502 + }, + { + "epoch": 3.530401737242128, + "grad_norm": 14.699795423214242, + "learning_rate": 4.1995177147028195e-06, + "loss": 0.7931, + "step": 6503 + }, + { + "epoch": 3.530944625407166, + "grad_norm": 13.875599451841557, + "learning_rate": 4.196653318909888e-06, + "loss": 0.7584, + "step": 6504 + }, + { + "epoch": 3.531487513572204, + "grad_norm": 10.802073588313325, + "learning_rate": 4.193789640884126e-06, + "loss": 0.4051, + "step": 6505 + }, + { + "epoch": 3.5320304017372424, + "grad_norm": 12.895721698436635, + "learning_rate": 4.190926680979708e-06, + "loss": 0.5394, + "step": 6506 + }, + { + "epoch": 3.53257328990228, + "grad_norm": 9.06469669507275, + "learning_rate": 4.188064439550743e-06, + "loss": 0.3421, + "step": 6507 + }, + { + "epoch": 3.533116178067318, + "grad_norm": 9.872074130422732, + "learning_rate": 4.185202916951224e-06, + "loss": 0.4934, + "step": 6508 + }, + { + "epoch": 3.533659066232356, + "grad_norm": 14.460060519698754, + "learning_rate": 4.1823421135350796e-06, + "loss": 0.9002, + "step": 6509 + }, + { + "epoch": 3.5342019543973944, + "grad_norm": 10.80679699799082, + "learning_rate": 4.179482029656134e-06, + "loss": 0.5466, + "step": 6510 + }, + { + "epoch": 3.534744842562432, + "grad_norm": 17.39434448772562, + "learning_rate": 4.176622665668127e-06, + "loss": 0.7686, + "step": 6511 + }, + { + "epoch": 3.53528773072747, + "grad_norm": 18.02312862709536, + "learning_rate": 4.173764021924711e-06, + "loss": 0.6285, + "step": 6512 + }, + { + "epoch": 3.535830618892508, + "grad_norm": 16.367851061651017, + "learning_rate": 4.170906098779446e-06, + "loss": 0.7154, + "step": 6513 + }, + { + "epoch": 3.5363735070575464, + "grad_norm": 14.114924740024234, + "learning_rate": 4.168048896585809e-06, + "loss": 0.6001, + "step": 6514 + }, + { + "epoch": 3.536916395222584, + "grad_norm": 11.148204893242914, + "learning_rate": 4.165192415697171e-06, + "loss": 0.4029, + "step": 6515 + }, + { + "epoch": 3.537459283387622, + "grad_norm": 12.586445991725615, + "learning_rate": 4.162336656466839e-06, + "loss": 0.53, + "step": 6516 + }, + { + "epoch": 3.53800217155266, + "grad_norm": 10.601436138316005, + "learning_rate": 4.159481619248012e-06, + "loss": 0.5671, + "step": 6517 + }, + { + "epoch": 3.5385450597176984, + "grad_norm": 21.51381975898863, + "learning_rate": 4.156627304393808e-06, + "loss": 0.8123, + "step": 6518 + }, + { + "epoch": 3.539087947882736, + "grad_norm": 11.786864237322536, + "learning_rate": 4.153773712257251e-06, + "loss": 0.5383, + "step": 6519 + }, + { + "epoch": 3.539630836047774, + "grad_norm": 13.251189336378648, + "learning_rate": 4.1509208431912785e-06, + "loss": 0.4292, + "step": 6520 + }, + { + "epoch": 3.540173724212812, + "grad_norm": 11.978453022863066, + "learning_rate": 4.14806869754874e-06, + "loss": 0.5242, + "step": 6521 + }, + { + "epoch": 3.5407166123778504, + "grad_norm": 14.020243792854204, + "learning_rate": 4.145217275682389e-06, + "loss": 0.4147, + "step": 6522 + }, + { + "epoch": 3.541259500542888, + "grad_norm": 12.871747424704319, + "learning_rate": 4.142366577944897e-06, + "loss": 0.479, + "step": 6523 + }, + { + "epoch": 3.541802388707926, + "grad_norm": 16.745498758328235, + "learning_rate": 4.139516604688843e-06, + "loss": 0.732, + "step": 6524 + }, + { + "epoch": 3.542345276872964, + "grad_norm": 18.803948548990295, + "learning_rate": 4.136667356266716e-06, + "loss": 0.6007, + "step": 6525 + }, + { + "epoch": 3.5428881650380024, + "grad_norm": 12.83744283380288, + "learning_rate": 4.133818833030915e-06, + "loss": 0.556, + "step": 6526 + }, + { + "epoch": 3.54343105320304, + "grad_norm": 10.490416435869522, + "learning_rate": 4.13097103533375e-06, + "loss": 0.3933, + "step": 6527 + }, + { + "epoch": 3.543973941368078, + "grad_norm": 13.443902999757336, + "learning_rate": 4.128123963527441e-06, + "loss": 0.7184, + "step": 6528 + }, + { + "epoch": 3.544516829533116, + "grad_norm": 9.68823576515441, + "learning_rate": 4.125277617964119e-06, + "loss": 0.4799, + "step": 6529 + }, + { + "epoch": 3.5450597176981544, + "grad_norm": 8.977409878928121, + "learning_rate": 4.122431998995819e-06, + "loss": 0.2998, + "step": 6530 + }, + { + "epoch": 3.545602605863192, + "grad_norm": 10.738202771196406, + "learning_rate": 4.119587106974505e-06, + "loss": 0.4038, + "step": 6531 + }, + { + "epoch": 3.54614549402823, + "grad_norm": 13.534073718206239, + "learning_rate": 4.116742942252024e-06, + "loss": 0.6041, + "step": 6532 + }, + { + "epoch": 3.546688382193268, + "grad_norm": 9.554994157525224, + "learning_rate": 4.113899505180157e-06, + "loss": 0.3955, + "step": 6533 + }, + { + "epoch": 3.5472312703583064, + "grad_norm": 14.083745919964446, + "learning_rate": 4.1110567961105775e-06, + "loss": 0.5232, + "step": 6534 + }, + { + "epoch": 3.547774158523344, + "grad_norm": 9.270850995609294, + "learning_rate": 4.108214815394876e-06, + "loss": 0.3932, + "step": 6535 + }, + { + "epoch": 3.548317046688382, + "grad_norm": 8.658156641649512, + "learning_rate": 4.105373563384563e-06, + "loss": 0.3587, + "step": 6536 + }, + { + "epoch": 3.54885993485342, + "grad_norm": 22.179187215860352, + "learning_rate": 4.102533040431034e-06, + "loss": 0.6502, + "step": 6537 + }, + { + "epoch": 3.5494028230184584, + "grad_norm": 12.115851203141874, + "learning_rate": 4.0996932468856265e-06, + "loss": 0.6969, + "step": 6538 + }, + { + "epoch": 3.549945711183496, + "grad_norm": 9.686560788090254, + "learning_rate": 4.096854183099554e-06, + "loss": 0.4376, + "step": 6539 + }, + { + "epoch": 3.550488599348534, + "grad_norm": 14.280429673703154, + "learning_rate": 4.0940158494239725e-06, + "loss": 0.7158, + "step": 6540 + }, + { + "epoch": 3.5510314875135722, + "grad_norm": 12.581936500402087, + "learning_rate": 4.091178246209916e-06, + "loss": 0.5323, + "step": 6541 + }, + { + "epoch": 3.5515743756786105, + "grad_norm": 14.994050222936808, + "learning_rate": 4.088341373808354e-06, + "loss": 0.6231, + "step": 6542 + }, + { + "epoch": 3.5521172638436482, + "grad_norm": 11.419896198480089, + "learning_rate": 4.085505232570157e-06, + "loss": 0.4173, + "step": 6543 + }, + { + "epoch": 3.552660152008686, + "grad_norm": 9.002355774397408, + "learning_rate": 4.082669822846092e-06, + "loss": 0.4046, + "step": 6544 + }, + { + "epoch": 3.5532030401737242, + "grad_norm": 8.59747484019196, + "learning_rate": 4.079835144986861e-06, + "loss": 0.3399, + "step": 6545 + }, + { + "epoch": 3.5537459283387625, + "grad_norm": 15.204451286324845, + "learning_rate": 4.0770011993430465e-06, + "loss": 0.6488, + "step": 6546 + }, + { + "epoch": 3.5542888165038002, + "grad_norm": 12.454918944013878, + "learning_rate": 4.074167986265169e-06, + "loss": 0.7895, + "step": 6547 + }, + { + "epoch": 3.554831704668838, + "grad_norm": 12.640544797370467, + "learning_rate": 4.071335506103639e-06, + "loss": 0.3486, + "step": 6548 + }, + { + "epoch": 3.5553745928338762, + "grad_norm": 14.54014915273436, + "learning_rate": 4.068503759208782e-06, + "loss": 0.8116, + "step": 6549 + }, + { + "epoch": 3.5559174809989145, + "grad_norm": 13.706653138925564, + "learning_rate": 4.065672745930833e-06, + "loss": 0.6351, + "step": 6550 + }, + { + "epoch": 3.5564603691639523, + "grad_norm": 7.990687725437862, + "learning_rate": 4.062842466619937e-06, + "loss": 0.356, + "step": 6551 + }, + { + "epoch": 3.55700325732899, + "grad_norm": 12.892934153832375, + "learning_rate": 4.060012921626146e-06, + "loss": 0.5433, + "step": 6552 + }, + { + "epoch": 3.5575461454940283, + "grad_norm": 11.135918453001386, + "learning_rate": 4.057184111299426e-06, + "loss": 0.5487, + "step": 6553 + }, + { + "epoch": 3.5580890336590665, + "grad_norm": 8.023158669990128, + "learning_rate": 4.054356035989645e-06, + "loss": 0.2137, + "step": 6554 + }, + { + "epoch": 3.5586319218241043, + "grad_norm": 15.187504399840787, + "learning_rate": 4.051528696046586e-06, + "loss": 0.513, + "step": 6555 + }, + { + "epoch": 3.559174809989142, + "grad_norm": 11.432478939914837, + "learning_rate": 4.048702091819938e-06, + "loss": 0.4349, + "step": 6556 + }, + { + "epoch": 3.5597176981541803, + "grad_norm": 8.365367809450516, + "learning_rate": 4.045876223659301e-06, + "loss": 0.3435, + "step": 6557 + }, + { + "epoch": 3.5602605863192185, + "grad_norm": 12.186833110184857, + "learning_rate": 4.0430510919141815e-06, + "loss": 0.6071, + "step": 6558 + }, + { + "epoch": 3.5608034744842563, + "grad_norm": 12.640045326528098, + "learning_rate": 4.040226696933998e-06, + "loss": 0.4771, + "step": 6559 + }, + { + "epoch": 3.561346362649294, + "grad_norm": 12.859046455514006, + "learning_rate": 4.037403039068073e-06, + "loss": 0.6651, + "step": 6560 + }, + { + "epoch": 3.5618892508143323, + "grad_norm": 10.219552402377104, + "learning_rate": 4.034580118665644e-06, + "loss": 0.3973, + "step": 6561 + }, + { + "epoch": 3.5624321389793705, + "grad_norm": 11.468743456972462, + "learning_rate": 4.031757936075854e-06, + "loss": 0.5699, + "step": 6562 + }, + { + "epoch": 3.5629750271444083, + "grad_norm": 14.399787814630601, + "learning_rate": 4.028936491647753e-06, + "loss": 0.7208, + "step": 6563 + }, + { + "epoch": 3.563517915309446, + "grad_norm": 12.303594318406011, + "learning_rate": 4.026115785730305e-06, + "loss": 0.4314, + "step": 6564 + }, + { + "epoch": 3.5640608034744843, + "grad_norm": 9.21496866787658, + "learning_rate": 4.023295818672377e-06, + "loss": 0.4055, + "step": 6565 + }, + { + "epoch": 3.5646036916395225, + "grad_norm": 13.534731784395452, + "learning_rate": 4.0204765908227475e-06, + "loss": 0.4986, + "step": 6566 + }, + { + "epoch": 3.5651465798045603, + "grad_norm": 10.67198457926589, + "learning_rate": 4.017658102530103e-06, + "loss": 0.424, + "step": 6567 + }, + { + "epoch": 3.565689467969598, + "grad_norm": 10.45173885950783, + "learning_rate": 4.014840354143035e-06, + "loss": 0.3562, + "step": 6568 + }, + { + "epoch": 3.5662323561346363, + "grad_norm": 17.950537818667204, + "learning_rate": 4.012023346010059e-06, + "loss": 0.6163, + "step": 6569 + }, + { + "epoch": 3.5667752442996745, + "grad_norm": 12.324798504977831, + "learning_rate": 4.009207078479571e-06, + "loss": 0.4225, + "step": 6570 + }, + { + "epoch": 3.5673181324647123, + "grad_norm": 14.602931671691664, + "learning_rate": 4.006391551899906e-06, + "loss": 0.5519, + "step": 6571 + }, + { + "epoch": 3.56786102062975, + "grad_norm": 11.70892852397016, + "learning_rate": 4.00357676661928e-06, + "loss": 0.483, + "step": 6572 + }, + { + "epoch": 3.5684039087947883, + "grad_norm": 9.372765530009781, + "learning_rate": 4.000762722985844e-06, + "loss": 0.2418, + "step": 6573 + }, + { + "epoch": 3.5689467969598265, + "grad_norm": 9.365246601466223, + "learning_rate": 3.997949421347631e-06, + "loss": 0.3748, + "step": 6574 + }, + { + "epoch": 3.5694896851248643, + "grad_norm": 21.027585535755822, + "learning_rate": 3.995136862052597e-06, + "loss": 0.6116, + "step": 6575 + }, + { + "epoch": 3.570032573289902, + "grad_norm": 9.558242987641426, + "learning_rate": 3.992325045448613e-06, + "loss": 0.3608, + "step": 6576 + }, + { + "epoch": 3.5705754614549403, + "grad_norm": 15.196445736424556, + "learning_rate": 3.989513971883434e-06, + "loss": 0.5847, + "step": 6577 + }, + { + "epoch": 3.5711183496199785, + "grad_norm": 13.049400371079434, + "learning_rate": 3.9867036417047546e-06, + "loss": 0.9054, + "step": 6578 + }, + { + "epoch": 3.5716612377850163, + "grad_norm": 10.359620340512786, + "learning_rate": 3.983894055260146e-06, + "loss": 0.3928, + "step": 6579 + }, + { + "epoch": 3.572204125950054, + "grad_norm": 11.96747351920452, + "learning_rate": 3.981085212897111e-06, + "loss": 0.4581, + "step": 6580 + }, + { + "epoch": 3.5727470141150923, + "grad_norm": 13.611366572987158, + "learning_rate": 3.97827711496305e-06, + "loss": 0.8939, + "step": 6581 + }, + { + "epoch": 3.5732899022801305, + "grad_norm": 12.527292279303031, + "learning_rate": 3.975469761805273e-06, + "loss": 0.3713, + "step": 6582 + }, + { + "epoch": 3.5738327904451683, + "grad_norm": 11.7570760452819, + "learning_rate": 3.9726631537710005e-06, + "loss": 0.442, + "step": 6583 + }, + { + "epoch": 3.574375678610206, + "grad_norm": 10.8606079711249, + "learning_rate": 3.969857291207349e-06, + "loss": 0.3717, + "step": 6584 + }, + { + "epoch": 3.5749185667752443, + "grad_norm": 10.89094058468825, + "learning_rate": 3.9670521744613645e-06, + "loss": 0.5867, + "step": 6585 + }, + { + "epoch": 3.5754614549402826, + "grad_norm": 11.08527189234627, + "learning_rate": 3.964247803879976e-06, + "loss": 0.4183, + "step": 6586 + }, + { + "epoch": 3.5760043431053203, + "grad_norm": 14.584182784882016, + "learning_rate": 3.9614441798100415e-06, + "loss": 0.6388, + "step": 6587 + }, + { + "epoch": 3.576547231270358, + "grad_norm": 12.70805104327354, + "learning_rate": 3.958641302598315e-06, + "loss": 0.4211, + "step": 6588 + }, + { + "epoch": 3.5770901194353963, + "grad_norm": 13.935385901164445, + "learning_rate": 3.95583917259146e-06, + "loss": 0.4955, + "step": 6589 + }, + { + "epoch": 3.5776330076004346, + "grad_norm": 8.25831878298027, + "learning_rate": 3.953037790136051e-06, + "loss": 0.3929, + "step": 6590 + }, + { + "epoch": 3.5781758957654723, + "grad_norm": 13.224695924777794, + "learning_rate": 3.950237155578563e-06, + "loss": 0.9157, + "step": 6591 + }, + { + "epoch": 3.57871878393051, + "grad_norm": 14.476470240562103, + "learning_rate": 3.947437269265387e-06, + "loss": 0.6185, + "step": 6592 + }, + { + "epoch": 3.5792616720955484, + "grad_norm": 14.623277057541618, + "learning_rate": 3.944638131542816e-06, + "loss": 0.415, + "step": 6593 + }, + { + "epoch": 3.5798045602605866, + "grad_norm": 10.928897970897149, + "learning_rate": 3.941839742757052e-06, + "loss": 0.4612, + "step": 6594 + }, + { + "epoch": 3.5803474484256244, + "grad_norm": 11.795057907246525, + "learning_rate": 3.939042103254204e-06, + "loss": 0.449, + "step": 6595 + }, + { + "epoch": 3.580890336590662, + "grad_norm": 13.011176416375783, + "learning_rate": 3.9362452133802866e-06, + "loss": 0.4674, + "step": 6596 + }, + { + "epoch": 3.5814332247557004, + "grad_norm": 12.733097120202023, + "learning_rate": 3.933449073481227e-06, + "loss": 0.5925, + "step": 6597 + }, + { + "epoch": 3.5819761129207386, + "grad_norm": 11.53786371590528, + "learning_rate": 3.930653683902854e-06, + "loss": 0.6089, + "step": 6598 + }, + { + "epoch": 3.5825190010857764, + "grad_norm": 10.840934046340113, + "learning_rate": 3.927859044990907e-06, + "loss": 0.3937, + "step": 6599 + }, + { + "epoch": 3.583061889250814, + "grad_norm": 13.435349756607788, + "learning_rate": 3.92506515709103e-06, + "loss": 0.5642, + "step": 6600 + }, + { + "epoch": 3.5836047774158524, + "grad_norm": 15.725827415410208, + "learning_rate": 3.922272020548773e-06, + "loss": 0.7689, + "step": 6601 + }, + { + "epoch": 3.5841476655808906, + "grad_norm": 10.5862295430844, + "learning_rate": 3.919479635709606e-06, + "loss": 0.4151, + "step": 6602 + }, + { + "epoch": 3.5846905537459284, + "grad_norm": 13.70191819426941, + "learning_rate": 3.9166880029188855e-06, + "loss": 0.5176, + "step": 6603 + }, + { + "epoch": 3.585233441910966, + "grad_norm": 20.653564386817997, + "learning_rate": 3.9138971225218895e-06, + "loss": 0.5782, + "step": 6604 + }, + { + "epoch": 3.5857763300760044, + "grad_norm": 13.964645747661887, + "learning_rate": 3.911106994863794e-06, + "loss": 0.629, + "step": 6605 + }, + { + "epoch": 3.5863192182410426, + "grad_norm": 15.152047460754902, + "learning_rate": 3.908317620289689e-06, + "loss": 0.7885, + "step": 6606 + }, + { + "epoch": 3.5868621064060804, + "grad_norm": 10.596374095925908, + "learning_rate": 3.905528999144576e-06, + "loss": 0.3883, + "step": 6607 + }, + { + "epoch": 3.587404994571118, + "grad_norm": 11.600211796857451, + "learning_rate": 3.902741131773341e-06, + "loss": 0.4589, + "step": 6608 + }, + { + "epoch": 3.5879478827361564, + "grad_norm": 9.294471096012794, + "learning_rate": 3.899954018520808e-06, + "loss": 0.334, + "step": 6609 + }, + { + "epoch": 3.5884907709011946, + "grad_norm": 12.555701694026313, + "learning_rate": 3.8971676597316775e-06, + "loss": 0.3707, + "step": 6610 + }, + { + "epoch": 3.5890336590662324, + "grad_norm": 10.572122929609495, + "learning_rate": 3.894382055750584e-06, + "loss": 0.5635, + "step": 6611 + }, + { + "epoch": 3.58957654723127, + "grad_norm": 9.702545363892131, + "learning_rate": 3.891597206922046e-06, + "loss": 0.3634, + "step": 6612 + }, + { + "epoch": 3.5901194353963084, + "grad_norm": 10.430633527197228, + "learning_rate": 3.888813113590496e-06, + "loss": 0.3777, + "step": 6613 + }, + { + "epoch": 3.5906623235613466, + "grad_norm": 16.05235504376427, + "learning_rate": 3.886029776100287e-06, + "loss": 0.5759, + "step": 6614 + }, + { + "epoch": 3.5912052117263844, + "grad_norm": 9.348911210053458, + "learning_rate": 3.883247194795653e-06, + "loss": 0.2619, + "step": 6615 + }, + { + "epoch": 3.591748099891422, + "grad_norm": 8.462834598999683, + "learning_rate": 3.8804653700207615e-06, + "loss": 0.3012, + "step": 6616 + }, + { + "epoch": 3.5922909880564604, + "grad_norm": 16.64696159175473, + "learning_rate": 3.87768430211966e-06, + "loss": 0.9137, + "step": 6617 + }, + { + "epoch": 3.5928338762214986, + "grad_norm": 12.223148301717917, + "learning_rate": 3.874903991436324e-06, + "loss": 0.4287, + "step": 6618 + }, + { + "epoch": 3.5933767643865364, + "grad_norm": 8.953587319006854, + "learning_rate": 3.872124438314624e-06, + "loss": 0.3932, + "step": 6619 + }, + { + "epoch": 3.593919652551574, + "grad_norm": 13.721613828536574, + "learning_rate": 3.869345643098341e-06, + "loss": 0.5449, + "step": 6620 + }, + { + "epoch": 3.5944625407166124, + "grad_norm": 10.931345128421508, + "learning_rate": 3.866567606131159e-06, + "loss": 0.3144, + "step": 6621 + }, + { + "epoch": 3.5950054288816506, + "grad_norm": 16.911421303952586, + "learning_rate": 3.86379032775667e-06, + "loss": 0.8115, + "step": 6622 + }, + { + "epoch": 3.5955483170466884, + "grad_norm": 13.179299764136056, + "learning_rate": 3.861013808318378e-06, + "loss": 0.557, + "step": 6623 + }, + { + "epoch": 3.596091205211726, + "grad_norm": 14.513913457624946, + "learning_rate": 3.858238048159674e-06, + "loss": 0.7132, + "step": 6624 + }, + { + "epoch": 3.5966340933767644, + "grad_norm": 18.20977238222908, + "learning_rate": 3.8554630476238805e-06, + "loss": 0.9001, + "step": 6625 + }, + { + "epoch": 3.5971769815418027, + "grad_norm": 13.411920870051427, + "learning_rate": 3.852688807054211e-06, + "loss": 0.6691, + "step": 6626 + }, + { + "epoch": 3.5977198697068404, + "grad_norm": 12.94431912351694, + "learning_rate": 3.849915326793786e-06, + "loss": 0.8245, + "step": 6627 + }, + { + "epoch": 3.598262757871878, + "grad_norm": 11.378136214690231, + "learning_rate": 3.847142607185636e-06, + "loss": 0.5369, + "step": 6628 + }, + { + "epoch": 3.5988056460369164, + "grad_norm": 18.107740640174118, + "learning_rate": 3.844370648572694e-06, + "loss": 0.5532, + "step": 6629 + }, + { + "epoch": 3.5993485342019547, + "grad_norm": 13.376518628320737, + "learning_rate": 3.841599451297802e-06, + "loss": 0.6283, + "step": 6630 + }, + { + "epoch": 3.5998914223669924, + "grad_norm": 17.68256418630599, + "learning_rate": 3.8388290157037034e-06, + "loss": 0.5916, + "step": 6631 + }, + { + "epoch": 3.6004343105320302, + "grad_norm": 9.436411945807468, + "learning_rate": 3.836059342133052e-06, + "loss": 0.488, + "step": 6632 + }, + { + "epoch": 3.6009771986970684, + "grad_norm": 12.744056075258795, + "learning_rate": 3.8332904309284045e-06, + "loss": 0.568, + "step": 6633 + }, + { + "epoch": 3.6015200868621067, + "grad_norm": 15.484356986063371, + "learning_rate": 3.830522282432227e-06, + "loss": 0.5619, + "step": 6634 + }, + { + "epoch": 3.6020629750271445, + "grad_norm": 22.466452052149346, + "learning_rate": 3.827754896986884e-06, + "loss": 0.754, + "step": 6635 + }, + { + "epoch": 3.6026058631921822, + "grad_norm": 10.907712543607193, + "learning_rate": 3.824988274934654e-06, + "loss": 0.3684, + "step": 6636 + }, + { + "epoch": 3.6031487513572205, + "grad_norm": 11.436863024139019, + "learning_rate": 3.822222416617714e-06, + "loss": 0.3772, + "step": 6637 + }, + { + "epoch": 3.6036916395222587, + "grad_norm": 8.249398777832742, + "learning_rate": 3.8194573223781515e-06, + "loss": 0.3908, + "step": 6638 + }, + { + "epoch": 3.6042345276872965, + "grad_norm": 10.129204140499496, + "learning_rate": 3.816692992557955e-06, + "loss": 0.5793, + "step": 6639 + }, + { + "epoch": 3.6047774158523342, + "grad_norm": 16.22320006976241, + "learning_rate": 3.8139294274990313e-06, + "loss": 0.7805, + "step": 6640 + }, + { + "epoch": 3.6053203040173725, + "grad_norm": 16.13099309974979, + "learning_rate": 3.8111666275431715e-06, + "loss": 0.5848, + "step": 6641 + }, + { + "epoch": 3.6058631921824107, + "grad_norm": 16.513602515759278, + "learning_rate": 3.8084045930320855e-06, + "loss": 0.5467, + "step": 6642 + }, + { + "epoch": 3.6064060803474485, + "grad_norm": 17.235827177127607, + "learning_rate": 3.8056433243073888e-06, + "loss": 0.7889, + "step": 6643 + }, + { + "epoch": 3.6069489685124863, + "grad_norm": 8.437058780004433, + "learning_rate": 3.8028828217105974e-06, + "loss": 0.4132, + "step": 6644 + }, + { + "epoch": 3.6074918566775245, + "grad_norm": 10.802576839366907, + "learning_rate": 3.800123085583135e-06, + "loss": 0.4123, + "step": 6645 + }, + { + "epoch": 3.6080347448425627, + "grad_norm": 11.08981584907767, + "learning_rate": 3.7973641162663276e-06, + "loss": 0.3947, + "step": 6646 + }, + { + "epoch": 3.6085776330076005, + "grad_norm": 9.95657545789331, + "learning_rate": 3.794605914101419e-06, + "loss": 0.6197, + "step": 6647 + }, + { + "epoch": 3.6091205211726383, + "grad_norm": 8.688663473460801, + "learning_rate": 3.791848479429534e-06, + "loss": 0.2914, + "step": 6648 + }, + { + "epoch": 3.6096634093376765, + "grad_norm": 13.569135821875664, + "learning_rate": 3.78909181259173e-06, + "loss": 0.4108, + "step": 6649 + }, + { + "epoch": 3.6102062975027147, + "grad_norm": 12.845103922950058, + "learning_rate": 3.7863359139289425e-06, + "loss": 0.5261, + "step": 6650 + }, + { + "epoch": 3.6107491856677525, + "grad_norm": 19.751616468555607, + "learning_rate": 3.783580783782035e-06, + "loss": 0.7023, + "step": 6651 + }, + { + "epoch": 3.6112920738327903, + "grad_norm": 13.094345121958364, + "learning_rate": 3.780826422491768e-06, + "loss": 0.5129, + "step": 6652 + }, + { + "epoch": 3.6118349619978285, + "grad_norm": 9.608815012306174, + "learning_rate": 3.7780728303987935e-06, + "loss": 0.3692, + "step": 6653 + }, + { + "epoch": 3.6123778501628667, + "grad_norm": 12.846531429400725, + "learning_rate": 3.7753200078436947e-06, + "loss": 0.5278, + "step": 6654 + }, + { + "epoch": 3.6129207383279045, + "grad_norm": 10.62078519937276, + "learning_rate": 3.7725679551669316e-06, + "loss": 0.5868, + "step": 6655 + }, + { + "epoch": 3.6134636264929423, + "grad_norm": 12.42176232319324, + "learning_rate": 3.769816672708895e-06, + "loss": 0.5746, + "step": 6656 + }, + { + "epoch": 3.6140065146579805, + "grad_norm": 15.37568549128414, + "learning_rate": 3.7670661608098545e-06, + "loss": 0.5671, + "step": 6657 + }, + { + "epoch": 3.6145494028230187, + "grad_norm": 14.651318253713534, + "learning_rate": 3.764316419810009e-06, + "loss": 0.4817, + "step": 6658 + }, + { + "epoch": 3.6150922909880565, + "grad_norm": 17.772625754725265, + "learning_rate": 3.761567450049448e-06, + "loss": 0.6106, + "step": 6659 + }, + { + "epoch": 3.6156351791530943, + "grad_norm": 12.785953562325862, + "learning_rate": 3.7588192518681664e-06, + "loss": 0.458, + "step": 6660 + }, + { + "epoch": 3.6161780673181325, + "grad_norm": 14.872223886784575, + "learning_rate": 3.7560718256060704e-06, + "loss": 0.5833, + "step": 6661 + }, + { + "epoch": 3.6167209554831707, + "grad_norm": 10.482699567034725, + "learning_rate": 3.753325171602955e-06, + "loss": 0.4262, + "step": 6662 + }, + { + "epoch": 3.6172638436482085, + "grad_norm": 12.097306912879557, + "learning_rate": 3.7505792901985417e-06, + "loss": 0.4847, + "step": 6663 + }, + { + "epoch": 3.6178067318132463, + "grad_norm": 11.56080631055532, + "learning_rate": 3.7478341817324414e-06, + "loss": 0.4421, + "step": 6664 + }, + { + "epoch": 3.6183496199782845, + "grad_norm": 17.08246749281531, + "learning_rate": 3.7450898465441744e-06, + "loss": 0.6724, + "step": 6665 + }, + { + "epoch": 3.6188925081433228, + "grad_norm": 14.083035375418559, + "learning_rate": 3.742346284973164e-06, + "loss": 0.7099, + "step": 6666 + }, + { + "epoch": 3.6194353963083605, + "grad_norm": 12.257812595787682, + "learning_rate": 3.7396034973587392e-06, + "loss": 0.5108, + "step": 6667 + }, + { + "epoch": 3.6199782844733983, + "grad_norm": 16.583982679880926, + "learning_rate": 3.7368614840401297e-06, + "loss": 0.5608, + "step": 6668 + }, + { + "epoch": 3.6205211726384365, + "grad_norm": 9.31804613123421, + "learning_rate": 3.734120245356475e-06, + "loss": 0.2673, + "step": 6669 + }, + { + "epoch": 3.6210640608034748, + "grad_norm": 8.73729193728082, + "learning_rate": 3.7313797816468145e-06, + "loss": 0.2867, + "step": 6670 + }, + { + "epoch": 3.6216069489685125, + "grad_norm": 10.574625431794137, + "learning_rate": 3.728640093250093e-06, + "loss": 0.6508, + "step": 6671 + }, + { + "epoch": 3.6221498371335503, + "grad_norm": 13.116215370859619, + "learning_rate": 3.7259011805051605e-06, + "loss": 0.8055, + "step": 6672 + }, + { + "epoch": 3.6226927252985885, + "grad_norm": 12.674214863430533, + "learning_rate": 3.723163043750768e-06, + "loss": 0.4149, + "step": 6673 + }, + { + "epoch": 3.6232356134636268, + "grad_norm": 15.457787982715566, + "learning_rate": 3.720425683325576e-06, + "loss": 0.5407, + "step": 6674 + }, + { + "epoch": 3.6237785016286646, + "grad_norm": 13.509659223593033, + "learning_rate": 3.7176890995681426e-06, + "loss": 0.3499, + "step": 6675 + }, + { + "epoch": 3.6243213897937023, + "grad_norm": 16.313287760559906, + "learning_rate": 3.714953292816934e-06, + "loss": 0.697, + "step": 6676 + }, + { + "epoch": 3.6248642779587406, + "grad_norm": 14.440828134958558, + "learning_rate": 3.712218263410319e-06, + "loss": 0.5698, + "step": 6677 + }, + { + "epoch": 3.6254071661237783, + "grad_norm": 9.354441770636624, + "learning_rate": 3.7094840116865693e-06, + "loss": 0.3452, + "step": 6678 + }, + { + "epoch": 3.6259500542888166, + "grad_norm": 9.584690153683102, + "learning_rate": 3.706750537983861e-06, + "loss": 0.4305, + "step": 6679 + }, + { + "epoch": 3.6264929424538543, + "grad_norm": 18.051709305408423, + "learning_rate": 3.7040178426402817e-06, + "loss": 0.7387, + "step": 6680 + }, + { + "epoch": 3.6270358306188926, + "grad_norm": 10.05950366415881, + "learning_rate": 3.7012859259938074e-06, + "loss": 0.3443, + "step": 6681 + }, + { + "epoch": 3.6275787187839303, + "grad_norm": 12.034931076516903, + "learning_rate": 3.698554788382328e-06, + "loss": 0.542, + "step": 6682 + }, + { + "epoch": 3.6281216069489686, + "grad_norm": 12.814483154092642, + "learning_rate": 3.6958244301436353e-06, + "loss": 0.6022, + "step": 6683 + }, + { + "epoch": 3.6286644951140063, + "grad_norm": 12.602310642914164, + "learning_rate": 3.693094851615421e-06, + "loss": 0.5764, + "step": 6684 + }, + { + "epoch": 3.6292073832790446, + "grad_norm": 13.840451235628782, + "learning_rate": 3.6903660531352947e-06, + "loss": 0.673, + "step": 6685 + }, + { + "epoch": 3.6297502714440824, + "grad_norm": 11.319388737736466, + "learning_rate": 3.6876380350407428e-06, + "loss": 0.5201, + "step": 6686 + }, + { + "epoch": 3.6302931596091206, + "grad_norm": 8.195460173712483, + "learning_rate": 3.6849107976691867e-06, + "loss": 0.2834, + "step": 6687 + }, + { + "epoch": 3.6308360477741584, + "grad_norm": 11.24454405051031, + "learning_rate": 3.6821843413579217e-06, + "loss": 0.4174, + "step": 6688 + }, + { + "epoch": 3.6313789359391966, + "grad_norm": 13.272520776849182, + "learning_rate": 3.679458666444169e-06, + "loss": 0.5333, + "step": 6689 + }, + { + "epoch": 3.6319218241042344, + "grad_norm": 11.044629755858516, + "learning_rate": 3.6767337732650454e-06, + "loss": 0.4734, + "step": 6690 + }, + { + "epoch": 3.6324647122692726, + "grad_norm": 9.202516399757954, + "learning_rate": 3.6740096621575604e-06, + "loss": 0.28, + "step": 6691 + }, + { + "epoch": 3.6330076004343104, + "grad_norm": 11.820877288548346, + "learning_rate": 3.6712863334586478e-06, + "loss": 0.2672, + "step": 6692 + }, + { + "epoch": 3.6335504885993486, + "grad_norm": 11.093852797013197, + "learning_rate": 3.668563787505123e-06, + "loss": 0.3923, + "step": 6693 + }, + { + "epoch": 3.6340933767643864, + "grad_norm": 10.10902067133091, + "learning_rate": 3.6658420246337256e-06, + "loss": 0.3633, + "step": 6694 + }, + { + "epoch": 3.6346362649294246, + "grad_norm": 11.211935387408433, + "learning_rate": 3.663121045181075e-06, + "loss": 0.4325, + "step": 6695 + }, + { + "epoch": 3.6351791530944624, + "grad_norm": 7.899984820961417, + "learning_rate": 3.660400849483716e-06, + "loss": 0.2753, + "step": 6696 + }, + { + "epoch": 3.6357220412595006, + "grad_norm": 15.593004160513892, + "learning_rate": 3.6576814378780844e-06, + "loss": 0.5612, + "step": 6697 + }, + { + "epoch": 3.6362649294245384, + "grad_norm": 14.607948256302272, + "learning_rate": 3.654962810700521e-06, + "loss": 0.5311, + "step": 6698 + }, + { + "epoch": 3.6368078175895766, + "grad_norm": 14.159583660700658, + "learning_rate": 3.6522449682872687e-06, + "loss": 0.4704, + "step": 6699 + }, + { + "epoch": 3.6373507057546144, + "grad_norm": 12.5139018167431, + "learning_rate": 3.6495279109744764e-06, + "loss": 0.5437, + "step": 6700 + }, + { + "epoch": 3.6378935939196526, + "grad_norm": 13.188903276553058, + "learning_rate": 3.6468116390981922e-06, + "loss": 0.5713, + "step": 6701 + }, + { + "epoch": 3.6384364820846904, + "grad_norm": 9.269396910628474, + "learning_rate": 3.644096152994371e-06, + "loss": 0.3362, + "step": 6702 + }, + { + "epoch": 3.6389793702497286, + "grad_norm": 15.76289925225522, + "learning_rate": 3.6413814529988666e-06, + "loss": 0.6549, + "step": 6703 + }, + { + "epoch": 3.6395222584147664, + "grad_norm": 14.218821330155238, + "learning_rate": 3.6386675394474393e-06, + "loss": 0.5017, + "step": 6704 + }, + { + "epoch": 3.6400651465798046, + "grad_norm": 13.035114169534788, + "learning_rate": 3.6359544126757483e-06, + "loss": 0.4331, + "step": 6705 + }, + { + "epoch": 3.6406080347448424, + "grad_norm": 20.82770787825795, + "learning_rate": 3.6332420730193584e-06, + "loss": 0.7242, + "step": 6706 + }, + { + "epoch": 3.6411509229098806, + "grad_norm": 14.084479195896975, + "learning_rate": 3.630530520813735e-06, + "loss": 0.5915, + "step": 6707 + }, + { + "epoch": 3.6416938110749184, + "grad_norm": 12.24391802241245, + "learning_rate": 3.6278197563942497e-06, + "loss": 0.3637, + "step": 6708 + }, + { + "epoch": 3.6422366992399566, + "grad_norm": 19.068246989978615, + "learning_rate": 3.625109780096173e-06, + "loss": 0.5401, + "step": 6709 + }, + { + "epoch": 3.6427795874049944, + "grad_norm": 12.923571390979863, + "learning_rate": 3.6224005922546777e-06, + "loss": 0.4372, + "step": 6710 + }, + { + "epoch": 3.6433224755700326, + "grad_norm": 18.72956431478281, + "learning_rate": 3.6196921932048424e-06, + "loss": 0.5212, + "step": 6711 + }, + { + "epoch": 3.6438653637350704, + "grad_norm": 9.246217525160361, + "learning_rate": 3.6169845832816443e-06, + "loss": 0.2875, + "step": 6712 + }, + { + "epoch": 3.6444082519001086, + "grad_norm": 12.538873429253352, + "learning_rate": 3.6142777628199667e-06, + "loss": 0.7951, + "step": 6713 + }, + { + "epoch": 3.6449511400651464, + "grad_norm": 9.970157389027023, + "learning_rate": 3.611571732154592e-06, + "loss": 0.3658, + "step": 6714 + }, + { + "epoch": 3.6454940282301846, + "grad_norm": 12.203072582682385, + "learning_rate": 3.6088664916202077e-06, + "loss": 0.5721, + "step": 6715 + }, + { + "epoch": 3.6460369163952224, + "grad_norm": 12.279248739912763, + "learning_rate": 3.606162041551402e-06, + "loss": 0.4408, + "step": 6716 + }, + { + "epoch": 3.6465798045602607, + "grad_norm": 12.802459902824975, + "learning_rate": 3.6034583822826607e-06, + "loss": 0.6177, + "step": 6717 + }, + { + "epoch": 3.6471226927252984, + "grad_norm": 14.71635978144862, + "learning_rate": 3.6007555141483897e-06, + "loss": 0.6333, + "step": 6718 + }, + { + "epoch": 3.6476655808903367, + "grad_norm": 15.254287066882977, + "learning_rate": 3.5980534374828667e-06, + "loss": 0.6541, + "step": 6719 + }, + { + "epoch": 3.6482084690553744, + "grad_norm": 17.306678251097242, + "learning_rate": 3.595352152620306e-06, + "loss": 0.5605, + "step": 6720 + }, + { + "epoch": 3.6487513572204127, + "grad_norm": 13.28374885202886, + "learning_rate": 3.5926516598947946e-06, + "loss": 0.553, + "step": 6721 + }, + { + "epoch": 3.6492942453854504, + "grad_norm": 15.674888174232077, + "learning_rate": 3.5899519596403356e-06, + "loss": 0.5473, + "step": 6722 + }, + { + "epoch": 3.6498371335504887, + "grad_norm": 9.387854198306167, + "learning_rate": 3.58725305219084e-06, + "loss": 0.4405, + "step": 6723 + }, + { + "epoch": 3.6503800217155264, + "grad_norm": 9.686200542254724, + "learning_rate": 3.5845549378801005e-06, + "loss": 0.4019, + "step": 6724 + }, + { + "epoch": 3.6509229098805647, + "grad_norm": 15.112849538872657, + "learning_rate": 3.5818576170418373e-06, + "loss": 0.7064, + "step": 6725 + }, + { + "epoch": 3.6514657980456025, + "grad_norm": 11.208536381794502, + "learning_rate": 3.5791610900096465e-06, + "loss": 0.3352, + "step": 6726 + }, + { + "epoch": 3.6520086862106407, + "grad_norm": 7.334988680617685, + "learning_rate": 3.5764653571170527e-06, + "loss": 0.288, + "step": 6727 + }, + { + "epoch": 3.6525515743756785, + "grad_norm": 11.838685637818697, + "learning_rate": 3.573770418697454e-06, + "loss": 0.6447, + "step": 6728 + }, + { + "epoch": 3.6530944625407167, + "grad_norm": 15.976115688824532, + "learning_rate": 3.5710762750841756e-06, + "loss": 0.4407, + "step": 6729 + }, + { + "epoch": 3.6536373507057545, + "grad_norm": 12.25069694802214, + "learning_rate": 3.5683829266104318e-06, + "loss": 0.4395, + "step": 6730 + }, + { + "epoch": 3.6541802388707927, + "grad_norm": 9.192123102498591, + "learning_rate": 3.5656903736093328e-06, + "loss": 0.3363, + "step": 6731 + }, + { + "epoch": 3.6547231270358305, + "grad_norm": 14.524314713398583, + "learning_rate": 3.5629986164139095e-06, + "loss": 0.3778, + "step": 6732 + }, + { + "epoch": 3.6552660152008687, + "grad_norm": 16.927339643078007, + "learning_rate": 3.5603076553570692e-06, + "loss": 0.6248, + "step": 6733 + }, + { + "epoch": 3.6558089033659065, + "grad_norm": 12.599717116787803, + "learning_rate": 3.5576174907716455e-06, + "loss": 0.6944, + "step": 6734 + }, + { + "epoch": 3.6563517915309447, + "grad_norm": 10.646444115088963, + "learning_rate": 3.5549281229903575e-06, + "loss": 0.3772, + "step": 6735 + }, + { + "epoch": 3.6568946796959825, + "grad_norm": 15.498764867333419, + "learning_rate": 3.552239552345831e-06, + "loss": 0.5415, + "step": 6736 + }, + { + "epoch": 3.6574375678610207, + "grad_norm": 15.747483640049493, + "learning_rate": 3.5495517791705937e-06, + "loss": 0.4526, + "step": 6737 + }, + { + "epoch": 3.6579804560260585, + "grad_norm": 8.811218308037992, + "learning_rate": 3.546864803797072e-06, + "loss": 0.2853, + "step": 6738 + }, + { + "epoch": 3.6585233441910967, + "grad_norm": 13.307618617651949, + "learning_rate": 3.5441786265576006e-06, + "loss": 0.5929, + "step": 6739 + }, + { + "epoch": 3.6590662323561345, + "grad_norm": 16.35655181341609, + "learning_rate": 3.541493247784398e-06, + "loss": 0.5969, + "step": 6740 + }, + { + "epoch": 3.6596091205211727, + "grad_norm": 13.468394744914283, + "learning_rate": 3.5388086678096077e-06, + "loss": 0.7131, + "step": 6741 + }, + { + "epoch": 3.6601520086862105, + "grad_norm": 10.610479158018709, + "learning_rate": 3.5361248869652586e-06, + "loss": 0.3491, + "step": 6742 + }, + { + "epoch": 3.6606948968512487, + "grad_norm": 10.63033092512971, + "learning_rate": 3.5334419055832857e-06, + "loss": 0.4148, + "step": 6743 + }, + { + "epoch": 3.6612377850162865, + "grad_norm": 12.066673558862, + "learning_rate": 3.530759723995524e-06, + "loss": 0.5922, + "step": 6744 + }, + { + "epoch": 3.6617806731813247, + "grad_norm": 13.669464818674667, + "learning_rate": 3.5280783425337097e-06, + "loss": 0.4984, + "step": 6745 + }, + { + "epoch": 3.6623235613463625, + "grad_norm": 11.385901052590954, + "learning_rate": 3.525397761529481e-06, + "loss": 0.4451, + "step": 6746 + }, + { + "epoch": 3.6628664495114007, + "grad_norm": 13.127975944644733, + "learning_rate": 3.5227179813143763e-06, + "loss": 0.4786, + "step": 6747 + }, + { + "epoch": 3.6634093376764385, + "grad_norm": 8.286555642541957, + "learning_rate": 3.520039002219835e-06, + "loss": 0.3308, + "step": 6748 + }, + { + "epoch": 3.6639522258414767, + "grad_norm": 13.989369716944896, + "learning_rate": 3.5173608245771964e-06, + "loss": 0.6301, + "step": 6749 + }, + { + "epoch": 3.6644951140065145, + "grad_norm": 11.280241380801893, + "learning_rate": 3.514683448717704e-06, + "loss": 0.4333, + "step": 6750 + }, + { + "epoch": 3.6650380021715527, + "grad_norm": 14.109069012865124, + "learning_rate": 3.5120068749724977e-06, + "loss": 0.4928, + "step": 6751 + }, + { + "epoch": 3.6655808903365905, + "grad_norm": 9.420479780236201, + "learning_rate": 3.509331103672623e-06, + "loss": 0.3409, + "step": 6752 + }, + { + "epoch": 3.6661237785016287, + "grad_norm": 13.634334295467264, + "learning_rate": 3.5066561351490202e-06, + "loss": 0.4104, + "step": 6753 + }, + { + "epoch": 3.6666666666666665, + "grad_norm": 12.151334809320293, + "learning_rate": 3.503981969732536e-06, + "loss": 0.6371, + "step": 6754 + }, + { + "epoch": 3.6672095548317047, + "grad_norm": 15.410529391564157, + "learning_rate": 3.5013086077539125e-06, + "loss": 0.6359, + "step": 6755 + }, + { + "epoch": 3.6677524429967425, + "grad_norm": 14.234740703322796, + "learning_rate": 3.4986360495438033e-06, + "loss": 0.5397, + "step": 6756 + }, + { + "epoch": 3.6682953311617807, + "grad_norm": 12.465036402535198, + "learning_rate": 3.4959642954327435e-06, + "loss": 0.5274, + "step": 6757 + }, + { + "epoch": 3.6688382193268185, + "grad_norm": 12.485874013130712, + "learning_rate": 3.4932933457511907e-06, + "loss": 0.4669, + "step": 6758 + }, + { + "epoch": 3.6693811074918568, + "grad_norm": 13.300754823528772, + "learning_rate": 3.490623200829485e-06, + "loss": 0.5991, + "step": 6759 + }, + { + "epoch": 3.6699239956568945, + "grad_norm": 13.6162114238038, + "learning_rate": 3.4879538609978724e-06, + "loss": 0.5753, + "step": 6760 + }, + { + "epoch": 3.6704668838219328, + "grad_norm": 15.055930735004507, + "learning_rate": 3.4852853265865117e-06, + "loss": 0.5937, + "step": 6761 + }, + { + "epoch": 3.6710097719869705, + "grad_norm": 19.021493329558147, + "learning_rate": 3.4826175979254382e-06, + "loss": 0.7309, + "step": 6762 + }, + { + "epoch": 3.6715526601520088, + "grad_norm": 7.879252217720616, + "learning_rate": 3.4799506753446133e-06, + "loss": 0.2324, + "step": 6763 + }, + { + "epoch": 3.6720955483170465, + "grad_norm": 15.566429874979375, + "learning_rate": 3.477284559173875e-06, + "loss": 0.8685, + "step": 6764 + }, + { + "epoch": 3.6726384364820848, + "grad_norm": 14.932288315207366, + "learning_rate": 3.474619249742983e-06, + "loss": 0.4121, + "step": 6765 + }, + { + "epoch": 3.6731813246471225, + "grad_norm": 12.826517078026129, + "learning_rate": 3.471954747381577e-06, + "loss": 0.4172, + "step": 6766 + }, + { + "epoch": 3.6737242128121608, + "grad_norm": 11.27094341605798, + "learning_rate": 3.4692910524192145e-06, + "loss": 0.6047, + "step": 6767 + }, + { + "epoch": 3.6742671009771986, + "grad_norm": 10.227384928715546, + "learning_rate": 3.4666281651853426e-06, + "loss": 0.4978, + "step": 6768 + }, + { + "epoch": 3.6748099891422368, + "grad_norm": 9.140068130601126, + "learning_rate": 3.463966086009313e-06, + "loss": 0.3314, + "step": 6769 + }, + { + "epoch": 3.6753528773072746, + "grad_norm": 8.57213348310477, + "learning_rate": 3.4613048152203776e-06, + "loss": 0.3135, + "step": 6770 + }, + { + "epoch": 3.675895765472313, + "grad_norm": 18.06864303821088, + "learning_rate": 3.4586443531476777e-06, + "loss": 0.7083, + "step": 6771 + }, + { + "epoch": 3.6764386536373506, + "grad_norm": 12.06612974538307, + "learning_rate": 3.455984700120272e-06, + "loss": 0.4263, + "step": 6772 + }, + { + "epoch": 3.676981541802389, + "grad_norm": 15.51068609579363, + "learning_rate": 3.453325856467108e-06, + "loss": 0.5577, + "step": 6773 + }, + { + "epoch": 3.6775244299674266, + "grad_norm": 10.276565099526616, + "learning_rate": 3.4506678225170363e-06, + "loss": 0.3197, + "step": 6774 + }, + { + "epoch": 3.678067318132465, + "grad_norm": 9.001248775476501, + "learning_rate": 3.4480105985988065e-06, + "loss": 0.376, + "step": 6775 + }, + { + "epoch": 3.6786102062975026, + "grad_norm": 10.679849220418427, + "learning_rate": 3.4453541850410686e-06, + "loss": 0.3383, + "step": 6776 + }, + { + "epoch": 3.679153094462541, + "grad_norm": 10.331306790662499, + "learning_rate": 3.4426985821723714e-06, + "loss": 0.5359, + "step": 6777 + }, + { + "epoch": 3.6796959826275786, + "grad_norm": 12.802407333669652, + "learning_rate": 3.4400437903211635e-06, + "loss": 0.4397, + "step": 6778 + }, + { + "epoch": 3.680238870792617, + "grad_norm": 14.73981252395514, + "learning_rate": 3.4373898098157953e-06, + "loss": 0.5717, + "step": 6779 + }, + { + "epoch": 3.6807817589576546, + "grad_norm": 16.362575863101522, + "learning_rate": 3.434736640984515e-06, + "loss": 0.5125, + "step": 6780 + }, + { + "epoch": 3.681324647122693, + "grad_norm": 11.576699027697012, + "learning_rate": 3.4320842841554702e-06, + "loss": 0.3924, + "step": 6781 + }, + { + "epoch": 3.6818675352877306, + "grad_norm": 12.430829059754513, + "learning_rate": 3.4294327396567085e-06, + "loss": 0.5816, + "step": 6782 + }, + { + "epoch": 3.682410423452769, + "grad_norm": 12.320695413208654, + "learning_rate": 3.4267820078161785e-06, + "loss": 0.5416, + "step": 6783 + }, + { + "epoch": 3.6829533116178066, + "grad_norm": 13.000253896056515, + "learning_rate": 3.4241320889617245e-06, + "loss": 0.536, + "step": 6784 + }, + { + "epoch": 3.683496199782845, + "grad_norm": 10.91584072091396, + "learning_rate": 3.4214829834210948e-06, + "loss": 0.4491, + "step": 6785 + }, + { + "epoch": 3.6840390879478826, + "grad_norm": 13.085054837243629, + "learning_rate": 3.418834691521934e-06, + "loss": 0.378, + "step": 6786 + }, + { + "epoch": 3.684581976112921, + "grad_norm": 11.116151442052336, + "learning_rate": 3.4161872135917873e-06, + "loss": 0.4053, + "step": 6787 + }, + { + "epoch": 3.6851248642779586, + "grad_norm": 14.044458874643864, + "learning_rate": 3.4135405499580998e-06, + "loss": 0.3851, + "step": 6788 + }, + { + "epoch": 3.685667752442997, + "grad_norm": 15.351406534798858, + "learning_rate": 3.410894700948214e-06, + "loss": 0.7129, + "step": 6789 + }, + { + "epoch": 3.6862106406080346, + "grad_norm": 18.553137629768052, + "learning_rate": 3.408249666889373e-06, + "loss": 0.9605, + "step": 6790 + }, + { + "epoch": 3.686753528773073, + "grad_norm": 12.059229166245107, + "learning_rate": 3.4056054481087187e-06, + "loss": 0.6246, + "step": 6791 + }, + { + "epoch": 3.6872964169381106, + "grad_norm": 8.164629976709701, + "learning_rate": 3.4029620449332925e-06, + "loss": 0.2484, + "step": 6792 + }, + { + "epoch": 3.687839305103149, + "grad_norm": 12.956574155251525, + "learning_rate": 3.400319457690031e-06, + "loss": 0.4846, + "step": 6793 + }, + { + "epoch": 3.6883821932681866, + "grad_norm": 13.218562644434163, + "learning_rate": 3.397677686705785e-06, + "loss": 0.4956, + "step": 6794 + }, + { + "epoch": 3.688925081433225, + "grad_norm": 12.52521210536022, + "learning_rate": 3.3950367323072777e-06, + "loss": 0.5609, + "step": 6795 + }, + { + "epoch": 3.6894679695982626, + "grad_norm": 14.858442377207329, + "learning_rate": 3.3923965948211612e-06, + "loss": 0.5361, + "step": 6796 + }, + { + "epoch": 3.690010857763301, + "grad_norm": 10.357147809279667, + "learning_rate": 3.3897572745739584e-06, + "loss": 0.4834, + "step": 6797 + }, + { + "epoch": 3.6905537459283386, + "grad_norm": 10.579748653808505, + "learning_rate": 3.3871187718921173e-06, + "loss": 0.4685, + "step": 6798 + }, + { + "epoch": 3.691096634093377, + "grad_norm": 16.620322523965804, + "learning_rate": 3.3844810871019617e-06, + "loss": 0.6373, + "step": 6799 + }, + { + "epoch": 3.6916395222584146, + "grad_norm": 15.999425270589017, + "learning_rate": 3.3818442205297262e-06, + "loss": 0.895, + "step": 6800 + }, + { + "epoch": 3.692182410423453, + "grad_norm": 10.78206305263471, + "learning_rate": 3.3792081725015525e-06, + "loss": 0.5564, + "step": 6801 + }, + { + "epoch": 3.6927252985884906, + "grad_norm": 11.641189355388065, + "learning_rate": 3.376572943343457e-06, + "loss": 0.5319, + "step": 6802 + }, + { + "epoch": 3.693268186753529, + "grad_norm": 14.695430550731649, + "learning_rate": 3.3739385333813824e-06, + "loss": 0.5868, + "step": 6803 + }, + { + "epoch": 3.6938110749185666, + "grad_norm": 10.078064673313023, + "learning_rate": 3.3713049429411435e-06, + "loss": 0.5002, + "step": 6804 + }, + { + "epoch": 3.694353963083605, + "grad_norm": 13.958804490496405, + "learning_rate": 3.368672172348478e-06, + "loss": 0.5521, + "step": 6805 + }, + { + "epoch": 3.6948968512486426, + "grad_norm": 14.349798638841735, + "learning_rate": 3.366040221929007e-06, + "loss": 0.8658, + "step": 6806 + }, + { + "epoch": 3.695439739413681, + "grad_norm": 19.918583954387678, + "learning_rate": 3.363409092008253e-06, + "loss": 0.7743, + "step": 6807 + }, + { + "epoch": 3.6959826275787186, + "grad_norm": 17.742647882095138, + "learning_rate": 3.3607787829116446e-06, + "loss": 0.4515, + "step": 6808 + }, + { + "epoch": 3.696525515743757, + "grad_norm": 10.287874260684818, + "learning_rate": 3.35814929496449e-06, + "loss": 0.4691, + "step": 6809 + }, + { + "epoch": 3.6970684039087947, + "grad_norm": 13.988040829949796, + "learning_rate": 3.3555206284920226e-06, + "loss": 0.5869, + "step": 6810 + }, + { + "epoch": 3.697611292073833, + "grad_norm": 12.599509307267757, + "learning_rate": 3.352892783819347e-06, + "loss": 0.5899, + "step": 6811 + }, + { + "epoch": 3.6981541802388707, + "grad_norm": 9.433309182729886, + "learning_rate": 3.3502657612714894e-06, + "loss": 0.3442, + "step": 6812 + }, + { + "epoch": 3.698697068403909, + "grad_norm": 16.010015776538683, + "learning_rate": 3.3476395611733616e-06, + "loss": 0.5644, + "step": 6813 + }, + { + "epoch": 3.6992399565689467, + "grad_norm": 12.170086596108076, + "learning_rate": 3.345014183849774e-06, + "loss": 0.5416, + "step": 6814 + }, + { + "epoch": 3.699782844733985, + "grad_norm": 9.466907002380811, + "learning_rate": 3.342389629625439e-06, + "loss": 0.5223, + "step": 6815 + }, + { + "epoch": 3.7003257328990227, + "grad_norm": 10.374676588070637, + "learning_rate": 3.3397658988249647e-06, + "loss": 0.327, + "step": 6816 + }, + { + "epoch": 3.700868621064061, + "grad_norm": 16.837753149163387, + "learning_rate": 3.337142991772859e-06, + "loss": 0.3853, + "step": 6817 + }, + { + "epoch": 3.7014115092290987, + "grad_norm": 19.187866841083977, + "learning_rate": 3.334520908793527e-06, + "loss": 0.7495, + "step": 6818 + }, + { + "epoch": 3.701954397394137, + "grad_norm": 12.434378548251997, + "learning_rate": 3.3318996502112733e-06, + "loss": 0.4615, + "step": 6819 + }, + { + "epoch": 3.7024972855591747, + "grad_norm": 14.472283943179146, + "learning_rate": 3.3292792163502973e-06, + "loss": 0.6001, + "step": 6820 + }, + { + "epoch": 3.703040173724213, + "grad_norm": 10.597662524177974, + "learning_rate": 3.3266596075346993e-06, + "loss": 0.4247, + "step": 6821 + }, + { + "epoch": 3.7035830618892507, + "grad_norm": 10.287091606357066, + "learning_rate": 3.324040824088477e-06, + "loss": 0.3372, + "step": 6822 + }, + { + "epoch": 3.704125950054289, + "grad_norm": 16.36152453578531, + "learning_rate": 3.3214228663355252e-06, + "loss": 0.4507, + "step": 6823 + }, + { + "epoch": 3.7046688382193267, + "grad_norm": 11.166011131196704, + "learning_rate": 3.3188057345996372e-06, + "loss": 0.4297, + "step": 6824 + }, + { + "epoch": 3.705211726384365, + "grad_norm": 13.97058303490548, + "learning_rate": 3.316189429204505e-06, + "loss": 0.5737, + "step": 6825 + }, + { + "epoch": 3.7057546145494027, + "grad_norm": 13.078068910284376, + "learning_rate": 3.313573950473713e-06, + "loss": 0.3998, + "step": 6826 + }, + { + "epoch": 3.706297502714441, + "grad_norm": 17.144621177815008, + "learning_rate": 3.3109592987307583e-06, + "loss": 0.6534, + "step": 6827 + }, + { + "epoch": 3.7068403908794787, + "grad_norm": 12.887234362542056, + "learning_rate": 3.308345474299014e-06, + "loss": 0.6388, + "step": 6828 + }, + { + "epoch": 3.707383279044517, + "grad_norm": 14.117206302860215, + "learning_rate": 3.305732477501766e-06, + "loss": 0.8498, + "step": 6829 + }, + { + "epoch": 3.7079261672095547, + "grad_norm": 16.668445763886474, + "learning_rate": 3.3031203086621953e-06, + "loss": 0.5491, + "step": 6830 + }, + { + "epoch": 3.708469055374593, + "grad_norm": 9.687167039444047, + "learning_rate": 3.300508968103375e-06, + "loss": 0.3493, + "step": 6831 + }, + { + "epoch": 3.7090119435396307, + "grad_norm": 8.884631092991913, + "learning_rate": 3.2978984561482898e-06, + "loss": 0.3227, + "step": 6832 + }, + { + "epoch": 3.709554831704669, + "grad_norm": 10.708989687832828, + "learning_rate": 3.295288773119797e-06, + "loss": 0.5266, + "step": 6833 + }, + { + "epoch": 3.7100977198697067, + "grad_norm": 11.602581058072978, + "learning_rate": 3.292679919340682e-06, + "loss": 0.4903, + "step": 6834 + }, + { + "epoch": 3.710640608034745, + "grad_norm": 11.22107577843576, + "learning_rate": 3.2900718951335973e-06, + "loss": 0.2758, + "step": 6835 + }, + { + "epoch": 3.7111834961997827, + "grad_norm": 11.654780601665442, + "learning_rate": 3.2874647008211223e-06, + "loss": 0.4478, + "step": 6836 + }, + { + "epoch": 3.711726384364821, + "grad_norm": 13.952341454584447, + "learning_rate": 3.2848583367257037e-06, + "loss": 0.6784, + "step": 6837 + }, + { + "epoch": 3.7122692725298587, + "grad_norm": 9.937530151430312, + "learning_rate": 3.2822528031697123e-06, + "loss": 0.3417, + "step": 6838 + }, + { + "epoch": 3.712812160694897, + "grad_norm": 12.66955971380617, + "learning_rate": 3.279648100475403e-06, + "loss": 0.5978, + "step": 6839 + }, + { + "epoch": 3.7133550488599347, + "grad_norm": 11.746942398326137, + "learning_rate": 3.2770442289649218e-06, + "loss": 0.422, + "step": 6840 + }, + { + "epoch": 3.713897937024973, + "grad_norm": 9.742871784049646, + "learning_rate": 3.274441188960331e-06, + "loss": 0.2878, + "step": 6841 + }, + { + "epoch": 3.7144408251900107, + "grad_norm": 16.102195766631716, + "learning_rate": 3.271838980783567e-06, + "loss": 0.6011, + "step": 6842 + }, + { + "epoch": 3.714983713355049, + "grad_norm": 12.78519494193334, + "learning_rate": 3.2692376047564843e-06, + "loss": 0.5416, + "step": 6843 + }, + { + "epoch": 3.7155266015200867, + "grad_norm": 7.51189016165619, + "learning_rate": 3.266637061200821e-06, + "loss": 0.2993, + "step": 6844 + }, + { + "epoch": 3.716069489685125, + "grad_norm": 11.448398733111881, + "learning_rate": 3.264037350438218e-06, + "loss": 0.469, + "step": 6845 + }, + { + "epoch": 3.7166123778501627, + "grad_norm": 11.057137423687962, + "learning_rate": 3.2614384727902106e-06, + "loss": 0.5315, + "step": 6846 + }, + { + "epoch": 3.717155266015201, + "grad_norm": 10.948114702379822, + "learning_rate": 3.2588404285782327e-06, + "loss": 0.5224, + "step": 6847 + }, + { + "epoch": 3.7176981541802387, + "grad_norm": 14.169026887876377, + "learning_rate": 3.2562432181236182e-06, + "loss": 0.4101, + "step": 6848 + }, + { + "epoch": 3.718241042345277, + "grad_norm": 10.211418365179567, + "learning_rate": 3.253646841747583e-06, + "loss": 0.356, + "step": 6849 + }, + { + "epoch": 3.7187839305103148, + "grad_norm": 10.750722744873066, + "learning_rate": 3.251051299771263e-06, + "loss": 0.5381, + "step": 6850 + }, + { + "epoch": 3.719326818675353, + "grad_norm": 10.07832966516926, + "learning_rate": 3.2484565925156732e-06, + "loss": 0.3729, + "step": 6851 + }, + { + "epoch": 3.7198697068403908, + "grad_norm": 12.95382559219359, + "learning_rate": 3.2458627203017335e-06, + "loss": 0.5247, + "step": 6852 + }, + { + "epoch": 3.720412595005429, + "grad_norm": 11.982999422321642, + "learning_rate": 3.243269683450256e-06, + "loss": 0.4333, + "step": 6853 + }, + { + "epoch": 3.7209554831704668, + "grad_norm": 14.445578614379583, + "learning_rate": 3.240677482281953e-06, + "loss": 0.6296, + "step": 6854 + }, + { + "epoch": 3.721498371335505, + "grad_norm": 14.769553336755148, + "learning_rate": 3.2380861171174317e-06, + "loss": 0.48, + "step": 6855 + }, + { + "epoch": 3.7220412595005428, + "grad_norm": 13.97227360884286, + "learning_rate": 3.2354955882771965e-06, + "loss": 0.8019, + "step": 6856 + }, + { + "epoch": 3.722584147665581, + "grad_norm": 10.17994965485879, + "learning_rate": 3.2329058960816483e-06, + "loss": 0.4003, + "step": 6857 + }, + { + "epoch": 3.7231270358306188, + "grad_norm": 11.647057134479718, + "learning_rate": 3.2303170408510832e-06, + "loss": 0.367, + "step": 6858 + }, + { + "epoch": 3.723669923995657, + "grad_norm": 13.75957119148339, + "learning_rate": 3.227729022905697e-06, + "loss": 0.5791, + "step": 6859 + }, + { + "epoch": 3.7242128121606948, + "grad_norm": 9.205512068687561, + "learning_rate": 3.225141842565577e-06, + "loss": 0.3107, + "step": 6860 + }, + { + "epoch": 3.724755700325733, + "grad_norm": 11.81216687228549, + "learning_rate": 3.2225555001507124e-06, + "loss": 0.555, + "step": 6861 + }, + { + "epoch": 3.725298588490771, + "grad_norm": 9.593384384681405, + "learning_rate": 3.219969995980985e-06, + "loss": 0.4445, + "step": 6862 + }, + { + "epoch": 3.725841476655809, + "grad_norm": 6.990210033223102, + "learning_rate": 3.217385330376174e-06, + "loss": 0.1827, + "step": 6863 + }, + { + "epoch": 3.726384364820847, + "grad_norm": 16.17725365334651, + "learning_rate": 3.2148015036559522e-06, + "loss": 0.6442, + "step": 6864 + }, + { + "epoch": 3.726927252985885, + "grad_norm": 9.936484066410749, + "learning_rate": 3.2122185161399012e-06, + "loss": 0.3459, + "step": 6865 + }, + { + "epoch": 3.727470141150923, + "grad_norm": 16.59297040971604, + "learning_rate": 3.209636368147475e-06, + "loss": 0.5169, + "step": 6866 + }, + { + "epoch": 3.728013029315961, + "grad_norm": 14.156312439459766, + "learning_rate": 3.207055059998052e-06, + "loss": 0.6236, + "step": 6867 + }, + { + "epoch": 3.728555917480999, + "grad_norm": 10.18075916418277, + "learning_rate": 3.2044745920108832e-06, + "loss": 0.3108, + "step": 6868 + }, + { + "epoch": 3.729098805646037, + "grad_norm": 11.677366741727328, + "learning_rate": 3.2018949645051257e-06, + "loss": 0.5047, + "step": 6869 + }, + { + "epoch": 3.729641693811075, + "grad_norm": 12.56566346772868, + "learning_rate": 3.199316177799835e-06, + "loss": 0.4698, + "step": 6870 + }, + { + "epoch": 3.730184581976113, + "grad_norm": 14.518568356077754, + "learning_rate": 3.1967382322139538e-06, + "loss": 0.6106, + "step": 6871 + }, + { + "epoch": 3.730727470141151, + "grad_norm": 12.595862204241472, + "learning_rate": 3.194161128066338e-06, + "loss": 0.5617, + "step": 6872 + }, + { + "epoch": 3.731270358306189, + "grad_norm": 14.836027445862813, + "learning_rate": 3.191584865675713e-06, + "loss": 0.5989, + "step": 6873 + }, + { + "epoch": 3.731813246471227, + "grad_norm": 15.308949249609173, + "learning_rate": 3.189009445360731e-06, + "loss": 0.5385, + "step": 6874 + }, + { + "epoch": 3.732356134636265, + "grad_norm": 12.000029094911469, + "learning_rate": 3.1864348674399083e-06, + "loss": 0.4519, + "step": 6875 + }, + { + "epoch": 3.732899022801303, + "grad_norm": 15.618812442704373, + "learning_rate": 3.1838611322316836e-06, + "loss": 0.7854, + "step": 6876 + }, + { + "epoch": 3.733441910966341, + "grad_norm": 11.71029558715352, + "learning_rate": 3.181288240054381e-06, + "loss": 0.4303, + "step": 6877 + }, + { + "epoch": 3.733984799131379, + "grad_norm": 10.712473236039532, + "learning_rate": 3.17871619122621e-06, + "loss": 0.4896, + "step": 6878 + }, + { + "epoch": 3.734527687296417, + "grad_norm": 10.854545722563955, + "learning_rate": 3.176144986065298e-06, + "loss": 0.4795, + "step": 6879 + }, + { + "epoch": 3.735070575461455, + "grad_norm": 10.845848101056692, + "learning_rate": 3.1735746248896436e-06, + "loss": 0.3512, + "step": 6880 + }, + { + "epoch": 3.735613463626493, + "grad_norm": 11.675294486688255, + "learning_rate": 3.1710051080171657e-06, + "loss": 0.3521, + "step": 6881 + }, + { + "epoch": 3.736156351791531, + "grad_norm": 13.325204505254437, + "learning_rate": 3.168436435765654e-06, + "loss": 0.9413, + "step": 6882 + }, + { + "epoch": 3.736699239956569, + "grad_norm": 11.623023017043877, + "learning_rate": 3.1658686084528146e-06, + "loss": 0.491, + "step": 6883 + }, + { + "epoch": 3.737242128121607, + "grad_norm": 7.846579009143442, + "learning_rate": 3.163301626396237e-06, + "loss": 0.2837, + "step": 6884 + }, + { + "epoch": 3.737785016286645, + "grad_norm": 13.308504927624488, + "learning_rate": 3.1607354899134102e-06, + "loss": 0.5548, + "step": 6885 + }, + { + "epoch": 3.738327904451683, + "grad_norm": 8.7685040860308, + "learning_rate": 3.1581701993217197e-06, + "loss": 0.2844, + "step": 6886 + }, + { + "epoch": 3.738870792616721, + "grad_norm": 12.979195325374219, + "learning_rate": 3.1556057549384424e-06, + "loss": 0.6242, + "step": 6887 + }, + { + "epoch": 3.739413680781759, + "grad_norm": 15.955341391994917, + "learning_rate": 3.1530421570807536e-06, + "loss": 0.5053, + "step": 6888 + }, + { + "epoch": 3.739956568946797, + "grad_norm": 15.39892028189493, + "learning_rate": 3.1504794060657228e-06, + "loss": 0.8675, + "step": 6889 + }, + { + "epoch": 3.740499457111835, + "grad_norm": 15.189388217562758, + "learning_rate": 3.147917502210316e-06, + "loss": 0.6521, + "step": 6890 + }, + { + "epoch": 3.741042345276873, + "grad_norm": 15.85941037907273, + "learning_rate": 3.1453564458313923e-06, + "loss": 0.6711, + "step": 6891 + }, + { + "epoch": 3.741585233441911, + "grad_norm": 10.599235827332738, + "learning_rate": 3.142796237245709e-06, + "loss": 0.3489, + "step": 6892 + }, + { + "epoch": 3.742128121606949, + "grad_norm": 11.991913744676399, + "learning_rate": 3.1402368767699152e-06, + "loss": 0.4865, + "step": 6893 + }, + { + "epoch": 3.742671009771987, + "grad_norm": 15.429458030904042, + "learning_rate": 3.137678364720559e-06, + "loss": 0.8425, + "step": 6894 + }, + { + "epoch": 3.743213897937025, + "grad_norm": 13.89952284029879, + "learning_rate": 3.135120701414077e-06, + "loss": 0.5919, + "step": 6895 + }, + { + "epoch": 3.743756786102063, + "grad_norm": 14.355639876320694, + "learning_rate": 3.1325638871668094e-06, + "loss": 0.517, + "step": 6896 + }, + { + "epoch": 3.744299674267101, + "grad_norm": 9.998711427681842, + "learning_rate": 3.1300079222949854e-06, + "loss": 0.4837, + "step": 6897 + }, + { + "epoch": 3.744842562432139, + "grad_norm": 12.948353433621449, + "learning_rate": 3.1274528071147303e-06, + "loss": 0.4702, + "step": 6898 + }, + { + "epoch": 3.745385450597177, + "grad_norm": 11.300075344262957, + "learning_rate": 3.124898541942065e-06, + "loss": 0.4163, + "step": 6899 + }, + { + "epoch": 3.745928338762215, + "grad_norm": 10.014842311292028, + "learning_rate": 3.122345127092905e-06, + "loss": 0.4597, + "step": 6900 + }, + { + "epoch": 3.746471226927253, + "grad_norm": 14.937187078292135, + "learning_rate": 3.1197925628830616e-06, + "loss": 1.0359, + "step": 6901 + }, + { + "epoch": 3.747014115092291, + "grad_norm": 11.821812911772865, + "learning_rate": 3.117240849628239e-06, + "loss": 0.369, + "step": 6902 + }, + { + "epoch": 3.747557003257329, + "grad_norm": 18.601455125565817, + "learning_rate": 3.1146899876440383e-06, + "loss": 0.3967, + "step": 6903 + }, + { + "epoch": 3.748099891422367, + "grad_norm": 26.38167902221625, + "learning_rate": 3.1121399772459495e-06, + "loss": 0.7955, + "step": 6904 + }, + { + "epoch": 3.748642779587405, + "grad_norm": 9.036657138578375, + "learning_rate": 3.109590818749373e-06, + "loss": 0.3234, + "step": 6905 + }, + { + "epoch": 3.749185667752443, + "grad_norm": 12.029611743069063, + "learning_rate": 3.1070425124695825e-06, + "loss": 0.4689, + "step": 6906 + }, + { + "epoch": 3.749728555917481, + "grad_norm": 11.16075799831913, + "learning_rate": 3.1044950587217595e-06, + "loss": 0.455, + "step": 6907 + }, + { + "epoch": 3.750271444082519, + "grad_norm": 12.64123781208839, + "learning_rate": 3.1019484578209768e-06, + "loss": 0.4525, + "step": 6908 + }, + { + "epoch": 3.750814332247557, + "grad_norm": 10.873865130895668, + "learning_rate": 3.0994027100822e-06, + "loss": 0.7957, + "step": 6909 + }, + { + "epoch": 3.751357220412595, + "grad_norm": 15.826997708730714, + "learning_rate": 3.0968578158203e-06, + "loss": 0.4531, + "step": 6910 + }, + { + "epoch": 3.751900108577633, + "grad_norm": 16.864174578861103, + "learning_rate": 3.094313775350022e-06, + "loss": 0.6061, + "step": 6911 + }, + { + "epoch": 3.752442996742671, + "grad_norm": 16.360017183257842, + "learning_rate": 3.0917705889860283e-06, + "loss": 0.849, + "step": 6912 + }, + { + "epoch": 3.752985884907709, + "grad_norm": 11.473341112084933, + "learning_rate": 3.0892282570428513e-06, + "loss": 0.4494, + "step": 6913 + }, + { + "epoch": 3.753528773072747, + "grad_norm": 12.456658860480365, + "learning_rate": 3.0866867798349407e-06, + "loss": 0.4774, + "step": 6914 + }, + { + "epoch": 3.754071661237785, + "grad_norm": 16.476960663849457, + "learning_rate": 3.0841461576766284e-06, + "loss": 0.8323, + "step": 6915 + }, + { + "epoch": 3.754614549402823, + "grad_norm": 13.888913904073622, + "learning_rate": 3.0816063908821403e-06, + "loss": 0.5861, + "step": 6916 + }, + { + "epoch": 3.755157437567861, + "grad_norm": 10.764080500005052, + "learning_rate": 3.079067479765604e-06, + "loss": 0.3865, + "step": 6917 + }, + { + "epoch": 3.755700325732899, + "grad_norm": 13.47504295740381, + "learning_rate": 3.076529424641026e-06, + "loss": 0.628, + "step": 6918 + }, + { + "epoch": 3.756243213897937, + "grad_norm": 14.742498671541298, + "learning_rate": 3.0739922258223286e-06, + "loss": 0.6586, + "step": 6919 + }, + { + "epoch": 3.756786102062975, + "grad_norm": 11.002483024658138, + "learning_rate": 3.0714558836233056e-06, + "loss": 0.539, + "step": 6920 + }, + { + "epoch": 3.757328990228013, + "grad_norm": 13.952905049231472, + "learning_rate": 3.068920398357663e-06, + "loss": 0.8338, + "step": 6921 + }, + { + "epoch": 3.757871878393051, + "grad_norm": 10.43091762446858, + "learning_rate": 3.0663857703389943e-06, + "loss": 0.4779, + "step": 6922 + }, + { + "epoch": 3.758414766558089, + "grad_norm": 16.415629535468835, + "learning_rate": 3.0638519998807826e-06, + "loss": 0.4277, + "step": 6923 + }, + { + "epoch": 3.758957654723127, + "grad_norm": 8.867208839527924, + "learning_rate": 3.0613190872964104e-06, + "loss": 0.3777, + "step": 6924 + }, + { + "epoch": 3.759500542888165, + "grad_norm": 11.642294707216093, + "learning_rate": 3.058787032899152e-06, + "loss": 0.5056, + "step": 6925 + }, + { + "epoch": 3.760043431053203, + "grad_norm": 11.206786718860933, + "learning_rate": 3.0562558370021765e-06, + "loss": 0.446, + "step": 6926 + }, + { + "epoch": 3.760586319218241, + "grad_norm": 10.068491984488928, + "learning_rate": 3.053725499918545e-06, + "loss": 0.5076, + "step": 6927 + }, + { + "epoch": 3.761129207383279, + "grad_norm": 16.07206135268971, + "learning_rate": 3.0511960219612147e-06, + "loss": 0.5143, + "step": 6928 + }, + { + "epoch": 3.761672095548317, + "grad_norm": 11.602660115945492, + "learning_rate": 3.0486674034430352e-06, + "loss": 0.508, + "step": 6929 + }, + { + "epoch": 3.762214983713355, + "grad_norm": 13.3862945751605, + "learning_rate": 3.046139644676751e-06, + "loss": 0.3828, + "step": 6930 + }, + { + "epoch": 3.762757871878393, + "grad_norm": 10.733311225895106, + "learning_rate": 3.0436127459749975e-06, + "loss": 0.37, + "step": 6931 + }, + { + "epoch": 3.763300760043431, + "grad_norm": 11.261346397862807, + "learning_rate": 3.041086707650306e-06, + "loss": 0.4808, + "step": 6932 + }, + { + "epoch": 3.763843648208469, + "grad_norm": 16.501582349438152, + "learning_rate": 3.0385615300151018e-06, + "loss": 0.6239, + "step": 6933 + }, + { + "epoch": 3.764386536373507, + "grad_norm": 11.553010310131304, + "learning_rate": 3.036037213381702e-06, + "loss": 0.4541, + "step": 6934 + }, + { + "epoch": 3.764929424538545, + "grad_norm": 11.723556454514561, + "learning_rate": 3.0335137580623154e-06, + "loss": 0.4217, + "step": 6935 + }, + { + "epoch": 3.765472312703583, + "grad_norm": 15.843208183133134, + "learning_rate": 3.030991164369057e-06, + "loss": 0.628, + "step": 6936 + }, + { + "epoch": 3.766015200868621, + "grad_norm": 8.046500189967162, + "learning_rate": 3.028469432613915e-06, + "loss": 0.288, + "step": 6937 + }, + { + "epoch": 3.766558089033659, + "grad_norm": 10.115796532018264, + "learning_rate": 3.0259485631087846e-06, + "loss": 0.3899, + "step": 6938 + }, + { + "epoch": 3.767100977198697, + "grad_norm": 13.289116360712832, + "learning_rate": 3.023428556165451e-06, + "loss": 0.6068, + "step": 6939 + }, + { + "epoch": 3.767643865363735, + "grad_norm": 11.871048514073903, + "learning_rate": 3.020909412095592e-06, + "loss": 0.6089, + "step": 6940 + }, + { + "epoch": 3.768186753528773, + "grad_norm": 8.084122023965138, + "learning_rate": 3.0183911312107806e-06, + "loss": 0.3247, + "step": 6941 + }, + { + "epoch": 3.768729641693811, + "grad_norm": 10.871234822080087, + "learning_rate": 3.015873713822478e-06, + "loss": 0.4928, + "step": 6942 + }, + { + "epoch": 3.769272529858849, + "grad_norm": 13.956354691016482, + "learning_rate": 3.013357160242051e-06, + "loss": 0.7454, + "step": 6943 + }, + { + "epoch": 3.769815418023887, + "grad_norm": 17.342203574294, + "learning_rate": 3.01084147078074e-06, + "loss": 0.6629, + "step": 6944 + }, + { + "epoch": 3.770358306188925, + "grad_norm": 9.435776085708705, + "learning_rate": 3.008326645749701e-06, + "loss": 0.2969, + "step": 6945 + }, + { + "epoch": 3.770901194353963, + "grad_norm": 14.35523924414514, + "learning_rate": 3.0058126854599624e-06, + "loss": 0.6859, + "step": 6946 + }, + { + "epoch": 3.771444082519001, + "grad_norm": 10.162640497221313, + "learning_rate": 3.003299590222454e-06, + "loss": 0.4144, + "step": 6947 + }, + { + "epoch": 3.771986970684039, + "grad_norm": 10.611516903885335, + "learning_rate": 3.0007873603480107e-06, + "loss": 0.5975, + "step": 6948 + }, + { + "epoch": 3.772529858849077, + "grad_norm": 13.809727664609994, + "learning_rate": 2.998275996147335e-06, + "loss": 0.4714, + "step": 6949 + }, + { + "epoch": 3.773072747014115, + "grad_norm": 11.317211520834723, + "learning_rate": 2.995765497931049e-06, + "loss": 0.4646, + "step": 6950 + }, + { + "epoch": 3.773615635179153, + "grad_norm": 10.86923796914853, + "learning_rate": 2.9932558660096443e-06, + "loss": 0.3228, + "step": 6951 + }, + { + "epoch": 3.774158523344191, + "grad_norm": 7.470997903259584, + "learning_rate": 2.990747100693526e-06, + "loss": 0.2531, + "step": 6952 + }, + { + "epoch": 3.774701411509229, + "grad_norm": 8.627824874780737, + "learning_rate": 2.988239202292972e-06, + "loss": 0.361, + "step": 6953 + }, + { + "epoch": 3.775244299674267, + "grad_norm": 9.14607939221467, + "learning_rate": 2.985732171118172e-06, + "loss": 0.3116, + "step": 6954 + }, + { + "epoch": 3.7757871878393052, + "grad_norm": 10.155112379070209, + "learning_rate": 2.9832260074791984e-06, + "loss": 0.4723, + "step": 6955 + }, + { + "epoch": 3.776330076004343, + "grad_norm": 10.506852840497306, + "learning_rate": 2.9807207116860094e-06, + "loss": 0.4232, + "step": 6956 + }, + { + "epoch": 3.7768729641693812, + "grad_norm": 11.193678853889507, + "learning_rate": 2.9782162840484765e-06, + "loss": 0.4915, + "step": 6957 + }, + { + "epoch": 3.777415852334419, + "grad_norm": 14.323380391376173, + "learning_rate": 2.9757127248763375e-06, + "loss": 0.6207, + "step": 6958 + }, + { + "epoch": 3.7779587404994572, + "grad_norm": 11.773695664037353, + "learning_rate": 2.973210034479247e-06, + "loss": 0.3957, + "step": 6959 + }, + { + "epoch": 3.778501628664495, + "grad_norm": 8.978033406900165, + "learning_rate": 2.97070821316674e-06, + "loss": 0.2935, + "step": 6960 + }, + { + "epoch": 3.7790445168295332, + "grad_norm": 12.4187581608995, + "learning_rate": 2.9682072612482427e-06, + "loss": 0.515, + "step": 6961 + }, + { + "epoch": 3.779587404994571, + "grad_norm": 12.551466191074608, + "learning_rate": 2.965707179033078e-06, + "loss": 0.4014, + "step": 6962 + }, + { + "epoch": 3.7801302931596092, + "grad_norm": 8.193678184944888, + "learning_rate": 2.9632079668304613e-06, + "loss": 0.2947, + "step": 6963 + }, + { + "epoch": 3.780673181324647, + "grad_norm": 12.277305597806611, + "learning_rate": 2.960709624949498e-06, + "loss": 0.4283, + "step": 6964 + }, + { + "epoch": 3.7812160694896852, + "grad_norm": 12.873825056251757, + "learning_rate": 2.958212153699187e-06, + "loss": 0.3719, + "step": 6965 + }, + { + "epoch": 3.781758957654723, + "grad_norm": 11.733322711394022, + "learning_rate": 2.955715553388421e-06, + "loss": 0.4547, + "step": 6966 + }, + { + "epoch": 3.7823018458197613, + "grad_norm": 13.421562379885513, + "learning_rate": 2.95321982432598e-06, + "loss": 0.8095, + "step": 6967 + }, + { + "epoch": 3.782844733984799, + "grad_norm": 10.751529697853956, + "learning_rate": 2.950724966820544e-06, + "loss": 0.3795, + "step": 6968 + }, + { + "epoch": 3.7833876221498373, + "grad_norm": 11.23769589292945, + "learning_rate": 2.9482309811806785e-06, + "loss": 0.4324, + "step": 6969 + }, + { + "epoch": 3.783930510314875, + "grad_norm": 10.760821500115565, + "learning_rate": 2.945737867714844e-06, + "loss": 0.4112, + "step": 6970 + }, + { + "epoch": 3.7844733984799133, + "grad_norm": 10.577088653058517, + "learning_rate": 2.9432456267313924e-06, + "loss": 0.3819, + "step": 6971 + }, + { + "epoch": 3.785016286644951, + "grad_norm": 9.877099189181129, + "learning_rate": 2.9407542585385683e-06, + "loss": 0.2344, + "step": 6972 + }, + { + "epoch": 3.7855591748099893, + "grad_norm": 18.577198876664223, + "learning_rate": 2.9382637634445087e-06, + "loss": 0.3254, + "step": 6973 + }, + { + "epoch": 3.786102062975027, + "grad_norm": 14.663768331220036, + "learning_rate": 2.9357741417572415e-06, + "loss": 0.5764, + "step": 6974 + }, + { + "epoch": 3.7866449511400653, + "grad_norm": 12.695557652262636, + "learning_rate": 2.9332853937846873e-06, + "loss": 0.4016, + "step": 6975 + }, + { + "epoch": 3.787187839305103, + "grad_norm": 13.456679312848303, + "learning_rate": 2.930797519834658e-06, + "loss": 0.438, + "step": 6976 + }, + { + "epoch": 3.7877307274701413, + "grad_norm": 10.441195425871536, + "learning_rate": 2.928310520214859e-06, + "loss": 0.4062, + "step": 6977 + }, + { + "epoch": 3.788273615635179, + "grad_norm": 10.300692057422426, + "learning_rate": 2.9258243952328846e-06, + "loss": 0.368, + "step": 6978 + }, + { + "epoch": 3.7888165038002173, + "grad_norm": 11.532450200794159, + "learning_rate": 2.9233391451962233e-06, + "loss": 0.5835, + "step": 6979 + }, + { + "epoch": 3.789359391965255, + "grad_norm": 13.465329246550354, + "learning_rate": 2.9208547704122525e-06, + "loss": 0.5226, + "step": 6980 + }, + { + "epoch": 3.7899022801302933, + "grad_norm": 14.107590055474905, + "learning_rate": 2.9183712711882526e-06, + "loss": 0.6195, + "step": 6981 + }, + { + "epoch": 3.790445168295331, + "grad_norm": 10.577784567014287, + "learning_rate": 2.915888647831374e-06, + "loss": 0.3655, + "step": 6982 + }, + { + "epoch": 3.7909880564603693, + "grad_norm": 13.136687160957166, + "learning_rate": 2.913406900648683e-06, + "loss": 0.6708, + "step": 6983 + }, + { + "epoch": 3.791530944625407, + "grad_norm": 15.57844160050923, + "learning_rate": 2.9109260299471165e-06, + "loss": 0.7229, + "step": 6984 + }, + { + "epoch": 3.7920738327904453, + "grad_norm": 14.528051815895886, + "learning_rate": 2.908446036033519e-06, + "loss": 0.4592, + "step": 6985 + }, + { + "epoch": 3.792616720955483, + "grad_norm": 20.212297672580796, + "learning_rate": 2.90596691921462e-06, + "loss": 0.759, + "step": 6986 + }, + { + "epoch": 3.7931596091205213, + "grad_norm": 9.84665764740946, + "learning_rate": 2.9034886797970342e-06, + "loss": 0.3681, + "step": 6987 + }, + { + "epoch": 3.793702497285559, + "grad_norm": 8.475948428830446, + "learning_rate": 2.9010113180872847e-06, + "loss": 0.2304, + "step": 6988 + }, + { + "epoch": 3.7942453854505973, + "grad_norm": 11.955148009858453, + "learning_rate": 2.898534834391762e-06, + "loss": 0.5278, + "step": 6989 + }, + { + "epoch": 3.794788273615635, + "grad_norm": 18.8535666021868, + "learning_rate": 2.8960592290167754e-06, + "loss": 1.1773, + "step": 6990 + }, + { + "epoch": 3.7953311617806733, + "grad_norm": 14.603134582974139, + "learning_rate": 2.8935845022685006e-06, + "loss": 0.7039, + "step": 6991 + }, + { + "epoch": 3.795874049945711, + "grad_norm": 17.09419124573872, + "learning_rate": 2.891110654453022e-06, + "loss": 0.6918, + "step": 6992 + }, + { + "epoch": 3.7964169381107493, + "grad_norm": 14.480524368793315, + "learning_rate": 2.888637685876309e-06, + "loss": 0.5987, + "step": 6993 + }, + { + "epoch": 3.796959826275787, + "grad_norm": 13.958888615564357, + "learning_rate": 2.8861655968442192e-06, + "loss": 0.7819, + "step": 6994 + }, + { + "epoch": 3.7975027144408253, + "grad_norm": 13.785514235662655, + "learning_rate": 2.883694387662511e-06, + "loss": 0.6469, + "step": 6995 + }, + { + "epoch": 3.798045602605863, + "grad_norm": 15.113325361621122, + "learning_rate": 2.8812240586368157e-06, + "loss": 0.6732, + "step": 6996 + }, + { + "epoch": 3.7985884907709013, + "grad_norm": 13.447315748094697, + "learning_rate": 2.8787546100726773e-06, + "loss": 0.8885, + "step": 6997 + }, + { + "epoch": 3.799131378935939, + "grad_norm": 15.942374082087648, + "learning_rate": 2.876286042275518e-06, + "loss": 0.5496, + "step": 6998 + }, + { + "epoch": 3.7996742671009773, + "grad_norm": 13.841633437355465, + "learning_rate": 2.873818355550655e-06, + "loss": 0.2924, + "step": 6999 + }, + { + "epoch": 3.800217155266015, + "grad_norm": 12.814203732560337, + "learning_rate": 2.871351550203295e-06, + "loss": 0.5538, + "step": 7000 + }, + { + "epoch": 3.8007600434310533, + "grad_norm": 12.191028501042764, + "learning_rate": 2.8688856265385367e-06, + "loss": 0.4576, + "step": 7001 + }, + { + "epoch": 3.801302931596091, + "grad_norm": 10.517009605166134, + "learning_rate": 2.86642058486137e-06, + "loss": 0.4275, + "step": 7002 + }, + { + "epoch": 3.8018458197611293, + "grad_norm": 13.976454434703868, + "learning_rate": 2.863956425476674e-06, + "loss": 0.5466, + "step": 7003 + }, + { + "epoch": 3.802388707926167, + "grad_norm": 11.5016640245614, + "learning_rate": 2.8614931486892207e-06, + "loss": 0.366, + "step": 7004 + }, + { + "epoch": 3.8029315960912053, + "grad_norm": 11.939582442434286, + "learning_rate": 2.8590307548036712e-06, + "loss": 0.4141, + "step": 7005 + }, + { + "epoch": 3.803474484256243, + "grad_norm": 12.987510367990025, + "learning_rate": 2.8565692441245796e-06, + "loss": 0.5152, + "step": 7006 + }, + { + "epoch": 3.8040173724212814, + "grad_norm": 12.832896642663844, + "learning_rate": 2.8541086169563894e-06, + "loss": 0.4585, + "step": 7007 + }, + { + "epoch": 3.804560260586319, + "grad_norm": 11.863347803098113, + "learning_rate": 2.851648873603433e-06, + "loss": 0.4196, + "step": 7008 + }, + { + "epoch": 3.8051031487513574, + "grad_norm": 13.239707667599935, + "learning_rate": 2.8491900143699367e-06, + "loss": 0.5748, + "step": 7009 + }, + { + "epoch": 3.805646036916395, + "grad_norm": 12.665696921336465, + "learning_rate": 2.846732039560017e-06, + "loss": 0.4688, + "step": 7010 + }, + { + "epoch": 3.8061889250814334, + "grad_norm": 10.268698833296764, + "learning_rate": 2.844274949477679e-06, + "loss": 0.3858, + "step": 7011 + }, + { + "epoch": 3.806731813246471, + "grad_norm": 12.660908637644715, + "learning_rate": 2.8418187444268198e-06, + "loss": 0.445, + "step": 7012 + }, + { + "epoch": 3.8072747014115094, + "grad_norm": 10.571667634245497, + "learning_rate": 2.8393634247112232e-06, + "loss": 0.4066, + "step": 7013 + }, + { + "epoch": 3.807817589576547, + "grad_norm": 19.94546394586301, + "learning_rate": 2.836908990634578e-06, + "loss": 0.5101, + "step": 7014 + }, + { + "epoch": 3.8083604777415854, + "grad_norm": 13.85083509805195, + "learning_rate": 2.834455442500441e-06, + "loss": 0.4073, + "step": 7015 + }, + { + "epoch": 3.808903365906623, + "grad_norm": 10.847008608128625, + "learning_rate": 2.8320027806122753e-06, + "loss": 0.3582, + "step": 7016 + }, + { + "epoch": 3.8094462540716614, + "grad_norm": 13.798257428735926, + "learning_rate": 2.82955100527343e-06, + "loss": 0.7057, + "step": 7017 + }, + { + "epoch": 3.809989142236699, + "grad_norm": 15.062376260700791, + "learning_rate": 2.8271001167871405e-06, + "loss": 0.6019, + "step": 7018 + }, + { + "epoch": 3.8105320304017374, + "grad_norm": 11.109879992701345, + "learning_rate": 2.8246501154565476e-06, + "loss": 0.3265, + "step": 7019 + }, + { + "epoch": 3.811074918566775, + "grad_norm": 13.70633009915418, + "learning_rate": 2.822201001584657e-06, + "loss": 0.4019, + "step": 7020 + }, + { + "epoch": 3.8116178067318134, + "grad_norm": 12.339132223947464, + "learning_rate": 2.8197527754743926e-06, + "loss": 0.3272, + "step": 7021 + }, + { + "epoch": 3.812160694896851, + "grad_norm": 9.817254835111205, + "learning_rate": 2.8173054374285434e-06, + "loss": 0.4118, + "step": 7022 + }, + { + "epoch": 3.8127035830618894, + "grad_norm": 15.18363193003276, + "learning_rate": 2.81485898774981e-06, + "loss": 0.7261, + "step": 7023 + }, + { + "epoch": 3.813246471226927, + "grad_norm": 11.984870936412664, + "learning_rate": 2.8124134267407665e-06, + "loss": 0.3734, + "step": 7024 + }, + { + "epoch": 3.8137893593919654, + "grad_norm": 10.678523359563961, + "learning_rate": 2.8099687547038824e-06, + "loss": 0.5318, + "step": 7025 + }, + { + "epoch": 3.814332247557003, + "grad_norm": 18.489591388355425, + "learning_rate": 2.807524971941528e-06, + "loss": 0.8301, + "step": 7026 + }, + { + "epoch": 3.8148751357220414, + "grad_norm": 10.459115256144122, + "learning_rate": 2.805082078755942e-06, + "loss": 0.3638, + "step": 7027 + }, + { + "epoch": 3.815418023887079, + "grad_norm": 12.778148708574054, + "learning_rate": 2.8026400754492767e-06, + "loss": 0.6202, + "step": 7028 + }, + { + "epoch": 3.8159609120521174, + "grad_norm": 14.76075800024929, + "learning_rate": 2.8001989623235528e-06, + "loss": 0.8516, + "step": 7029 + }, + { + "epoch": 3.816503800217155, + "grad_norm": 15.243047141569159, + "learning_rate": 2.797758739680698e-06, + "loss": 0.7363, + "step": 7030 + }, + { + "epoch": 3.8170466883821934, + "grad_norm": 12.332962959117395, + "learning_rate": 2.7953194078225223e-06, + "loss": 0.5528, + "step": 7031 + }, + { + "epoch": 3.817589576547231, + "grad_norm": 14.15721396911131, + "learning_rate": 2.792880967050724e-06, + "loss": 0.6093, + "step": 7032 + }, + { + "epoch": 3.8181324647122694, + "grad_norm": 16.149812128817175, + "learning_rate": 2.790443417666894e-06, + "loss": 0.7232, + "step": 7033 + }, + { + "epoch": 3.818675352877307, + "grad_norm": 6.884614594609857, + "learning_rate": 2.7880067599725123e-06, + "loss": 0.2341, + "step": 7034 + }, + { + "epoch": 3.8192182410423454, + "grad_norm": 13.604715542482158, + "learning_rate": 2.7855709942689525e-06, + "loss": 0.5522, + "step": 7035 + }, + { + "epoch": 3.819761129207383, + "grad_norm": 12.822857567671976, + "learning_rate": 2.7831361208574636e-06, + "loss": 0.459, + "step": 7036 + }, + { + "epoch": 3.8203040173724214, + "grad_norm": 16.80060104026762, + "learning_rate": 2.780702140039204e-06, + "loss": 0.6408, + "step": 7037 + }, + { + "epoch": 3.820846905537459, + "grad_norm": 9.25304440984111, + "learning_rate": 2.77826905211521e-06, + "loss": 0.3628, + "step": 7038 + }, + { + "epoch": 3.8213897937024974, + "grad_norm": 9.5263200358754, + "learning_rate": 2.775836857386409e-06, + "loss": 0.4398, + "step": 7039 + }, + { + "epoch": 3.821932681867535, + "grad_norm": 9.394535823999632, + "learning_rate": 2.77340555615362e-06, + "loss": 0.4318, + "step": 7040 + }, + { + "epoch": 3.8224755700325734, + "grad_norm": 16.75570479373163, + "learning_rate": 2.7709751487175486e-06, + "loss": 0.621, + "step": 7041 + }, + { + "epoch": 3.823018458197611, + "grad_norm": 11.68838601944592, + "learning_rate": 2.768545635378792e-06, + "loss": 0.4332, + "step": 7042 + }, + { + "epoch": 3.8235613463626494, + "grad_norm": 13.068740742878463, + "learning_rate": 2.7661170164378372e-06, + "loss": 0.6055, + "step": 7043 + }, + { + "epoch": 3.824104234527687, + "grad_norm": 12.808750119841285, + "learning_rate": 2.7636892921950586e-06, + "loss": 0.3966, + "step": 7044 + }, + { + "epoch": 3.8246471226927254, + "grad_norm": 13.368527360017245, + "learning_rate": 2.7612624629507213e-06, + "loss": 0.4222, + "step": 7045 + }, + { + "epoch": 3.8251900108577632, + "grad_norm": 14.624704158928514, + "learning_rate": 2.758836529004979e-06, + "loss": 0.5196, + "step": 7046 + }, + { + "epoch": 3.8257328990228014, + "grad_norm": 11.425494087567786, + "learning_rate": 2.756411490657875e-06, + "loss": 0.5476, + "step": 7047 + }, + { + "epoch": 3.8262757871878392, + "grad_norm": 11.438998129235056, + "learning_rate": 2.753987348209344e-06, + "loss": 0.3902, + "step": 7048 + }, + { + "epoch": 3.8268186753528775, + "grad_norm": 13.662522916551874, + "learning_rate": 2.7515641019592053e-06, + "loss": 0.494, + "step": 7049 + }, + { + "epoch": 3.8273615635179152, + "grad_norm": 10.296671849651366, + "learning_rate": 2.7491417522071706e-06, + "loss": 0.3226, + "step": 7050 + }, + { + "epoch": 3.8279044516829535, + "grad_norm": 10.454941310693076, + "learning_rate": 2.7467202992528376e-06, + "loss": 0.2803, + "step": 7051 + }, + { + "epoch": 3.8284473398479912, + "grad_norm": 11.648384847247753, + "learning_rate": 2.744299743395703e-06, + "loss": 0.5601, + "step": 7052 + }, + { + "epoch": 3.8289902280130295, + "grad_norm": 8.944598429537628, + "learning_rate": 2.741880084935138e-06, + "loss": 0.3296, + "step": 7053 + }, + { + "epoch": 3.8295331161780672, + "grad_norm": 13.218169787059113, + "learning_rate": 2.7394613241704117e-06, + "loss": 0.5443, + "step": 7054 + }, + { + "epoch": 3.8300760043431055, + "grad_norm": 9.804324070437223, + "learning_rate": 2.7370434614006803e-06, + "loss": 0.4269, + "step": 7055 + }, + { + "epoch": 3.8306188925081432, + "grad_norm": 16.684309046147042, + "learning_rate": 2.734626496924986e-06, + "loss": 0.7608, + "step": 7056 + }, + { + "epoch": 3.8311617806731815, + "grad_norm": 11.594370437365152, + "learning_rate": 2.7322104310422713e-06, + "loss": 0.4286, + "step": 7057 + }, + { + "epoch": 3.8317046688382193, + "grad_norm": 9.387708800824917, + "learning_rate": 2.7297952640513483e-06, + "loss": 0.4659, + "step": 7058 + }, + { + "epoch": 3.8322475570032575, + "grad_norm": 15.36709432735338, + "learning_rate": 2.727380996250939e-06, + "loss": 0.4572, + "step": 7059 + }, + { + "epoch": 3.8327904451682953, + "grad_norm": 10.866529562034719, + "learning_rate": 2.724967627939634e-06, + "loss": 0.5208, + "step": 7060 + }, + { + "epoch": 3.8333333333333335, + "grad_norm": 11.215185901261178, + "learning_rate": 2.722555159415934e-06, + "loss": 0.4077, + "step": 7061 + }, + { + "epoch": 3.8338762214983713, + "grad_norm": 13.560818928747905, + "learning_rate": 2.7201435909782027e-06, + "loss": 0.4849, + "step": 7062 + }, + { + "epoch": 3.8344191096634095, + "grad_norm": 12.76281749363346, + "learning_rate": 2.7177329229247186e-06, + "loss": 0.6375, + "step": 7063 + }, + { + "epoch": 3.8349619978284473, + "grad_norm": 11.109626055006531, + "learning_rate": 2.715323155553635e-06, + "loss": 0.6538, + "step": 7064 + }, + { + "epoch": 3.8355048859934855, + "grad_norm": 16.73906444542459, + "learning_rate": 2.712914289162989e-06, + "loss": 0.4548, + "step": 7065 + }, + { + "epoch": 3.8360477741585233, + "grad_norm": 15.832313422557572, + "learning_rate": 2.7105063240507222e-06, + "loss": 0.5417, + "step": 7066 + }, + { + "epoch": 3.8365906623235615, + "grad_norm": 10.615862418603832, + "learning_rate": 2.708099260514645e-06, + "loss": 0.6034, + "step": 7067 + }, + { + "epoch": 3.8371335504885993, + "grad_norm": 16.513578882781225, + "learning_rate": 2.7056930988524763e-06, + "loss": 0.9556, + "step": 7068 + }, + { + "epoch": 3.8376764386536375, + "grad_norm": 17.31989234293219, + "learning_rate": 2.703287839361811e-06, + "loss": 0.744, + "step": 7069 + }, + { + "epoch": 3.8382193268186753, + "grad_norm": 10.336380611592002, + "learning_rate": 2.7008834823401344e-06, + "loss": 0.5354, + "step": 7070 + }, + { + "epoch": 3.8387622149837135, + "grad_norm": 13.577288380592488, + "learning_rate": 2.698480028084821e-06, + "loss": 0.5647, + "step": 7071 + }, + { + "epoch": 3.8393051031487513, + "grad_norm": 15.062012119646445, + "learning_rate": 2.6960774768931365e-06, + "loss": 0.6897, + "step": 7072 + }, + { + "epoch": 3.8398479913137895, + "grad_norm": 9.679061664826015, + "learning_rate": 2.6936758290622324e-06, + "loss": 0.3364, + "step": 7073 + }, + { + "epoch": 3.8403908794788273, + "grad_norm": 12.06717080820273, + "learning_rate": 2.69127508488914e-06, + "loss": 0.473, + "step": 7074 + }, + { + "epoch": 3.8409337676438655, + "grad_norm": 12.515990246581914, + "learning_rate": 2.6888752446707965e-06, + "loss": 0.7188, + "step": 7075 + }, + { + "epoch": 3.8414766558089033, + "grad_norm": 12.151182717520161, + "learning_rate": 2.6864763087040145e-06, + "loss": 0.4494, + "step": 7076 + }, + { + "epoch": 3.8420195439739415, + "grad_norm": 13.03785855389188, + "learning_rate": 2.6840782772855002e-06, + "loss": 0.6012, + "step": 7077 + }, + { + "epoch": 3.8425624321389793, + "grad_norm": 13.049506054659528, + "learning_rate": 2.6816811507118436e-06, + "loss": 0.6763, + "step": 7078 + }, + { + "epoch": 3.8431053203040175, + "grad_norm": 10.797345758374234, + "learning_rate": 2.6792849292795253e-06, + "loss": 0.3712, + "step": 7079 + }, + { + "epoch": 3.8436482084690553, + "grad_norm": 7.303579416108625, + "learning_rate": 2.6768896132849144e-06, + "loss": 0.2622, + "step": 7080 + }, + { + "epoch": 3.8441910966340935, + "grad_norm": 12.992302429769211, + "learning_rate": 2.674495203024269e-06, + "loss": 0.5022, + "step": 7081 + }, + { + "epoch": 3.8447339847991313, + "grad_norm": 11.355323594435934, + "learning_rate": 2.67210169879373e-06, + "loss": 0.3898, + "step": 7082 + }, + { + "epoch": 3.8452768729641695, + "grad_norm": 10.043141357807462, + "learning_rate": 2.669709100889333e-06, + "loss": 0.4849, + "step": 7083 + }, + { + "epoch": 3.8458197611292073, + "grad_norm": 10.154968040172625, + "learning_rate": 2.6673174096069976e-06, + "loss": 0.6587, + "step": 7084 + }, + { + "epoch": 3.8463626492942455, + "grad_norm": 15.332496592868017, + "learning_rate": 2.6649266252425297e-06, + "loss": 0.5297, + "step": 7085 + }, + { + "epoch": 3.8469055374592833, + "grad_norm": 12.176208814956482, + "learning_rate": 2.6625367480916285e-06, + "loss": 0.603, + "step": 7086 + }, + { + "epoch": 3.8474484256243215, + "grad_norm": 12.08600431730549, + "learning_rate": 2.660147778449876e-06, + "loss": 0.6609, + "step": 7087 + }, + { + "epoch": 3.8479913137893593, + "grad_norm": 12.157193141505328, + "learning_rate": 2.6577597166127455e-06, + "loss": 0.4869, + "step": 7088 + }, + { + "epoch": 3.8485342019543975, + "grad_norm": 13.235050132598122, + "learning_rate": 2.655372562875591e-06, + "loss": 0.4279, + "step": 7089 + }, + { + "epoch": 3.8490770901194353, + "grad_norm": 12.171527155216868, + "learning_rate": 2.652986317533669e-06, + "loss": 0.4388, + "step": 7090 + }, + { + "epoch": 3.8496199782844736, + "grad_norm": 10.123066437797494, + "learning_rate": 2.650600980882104e-06, + "loss": 0.3086, + "step": 7091 + }, + { + "epoch": 3.8501628664495113, + "grad_norm": 12.006465921202063, + "learning_rate": 2.6482165532159275e-06, + "loss": 0.4117, + "step": 7092 + }, + { + "epoch": 3.8507057546145496, + "grad_norm": 15.608180907150922, + "learning_rate": 2.645833034830043e-06, + "loss": 0.5819, + "step": 7093 + }, + { + "epoch": 3.8512486427795873, + "grad_norm": 18.214304962348578, + "learning_rate": 2.6434504260192485e-06, + "loss": 0.6414, + "step": 7094 + }, + { + "epoch": 3.8517915309446256, + "grad_norm": 9.263854181848101, + "learning_rate": 2.64106872707823e-06, + "loss": 0.3855, + "step": 7095 + }, + { + "epoch": 3.8523344191096633, + "grad_norm": 11.56441830671296, + "learning_rate": 2.638687938301557e-06, + "loss": 0.3565, + "step": 7096 + }, + { + "epoch": 3.8528773072747016, + "grad_norm": 13.250463401216958, + "learning_rate": 2.636308059983699e-06, + "loss": 0.6425, + "step": 7097 + }, + { + "epoch": 3.8534201954397393, + "grad_norm": 14.438578870413643, + "learning_rate": 2.6339290924189886e-06, + "loss": 0.5401, + "step": 7098 + }, + { + "epoch": 3.8539630836047776, + "grad_norm": 19.454185693944922, + "learning_rate": 2.6315510359016736e-06, + "loss": 0.5113, + "step": 7099 + }, + { + "epoch": 3.8545059717698154, + "grad_norm": 14.281154826114456, + "learning_rate": 2.629173890725866e-06, + "loss": 0.6589, + "step": 7100 + }, + { + "epoch": 3.8550488599348536, + "grad_norm": 10.71424151041571, + "learning_rate": 2.62679765718558e-06, + "loss": 0.5477, + "step": 7101 + }, + { + "epoch": 3.8555917480998914, + "grad_norm": 15.907412596919608, + "learning_rate": 2.6244223355747144e-06, + "loss": 0.5586, + "step": 7102 + }, + { + "epoch": 3.8561346362649296, + "grad_norm": 14.500284217616562, + "learning_rate": 2.6220479261870436e-06, + "loss": 0.4131, + "step": 7103 + }, + { + "epoch": 3.8566775244299674, + "grad_norm": 16.114243210903357, + "learning_rate": 2.6196744293162503e-06, + "loss": 0.8063, + "step": 7104 + }, + { + "epoch": 3.8572204125950056, + "grad_norm": 10.762654220587946, + "learning_rate": 2.6173018452558786e-06, + "loss": 0.3736, + "step": 7105 + }, + { + "epoch": 3.8577633007600434, + "grad_norm": 8.619518240537495, + "learning_rate": 2.614930174299388e-06, + "loss": 0.4294, + "step": 7106 + }, + { + "epoch": 3.8583061889250816, + "grad_norm": 13.731123733938277, + "learning_rate": 2.612559416740098e-06, + "loss": 0.5039, + "step": 7107 + }, + { + "epoch": 3.8588490770901194, + "grad_norm": 9.925845082716284, + "learning_rate": 2.6101895728712355e-06, + "loss": 0.4914, + "step": 7108 + }, + { + "epoch": 3.8593919652551576, + "grad_norm": 9.305166559901256, + "learning_rate": 2.6078206429859044e-06, + "loss": 0.2875, + "step": 7109 + }, + { + "epoch": 3.8599348534201954, + "grad_norm": 11.732054868905763, + "learning_rate": 2.6054526273770975e-06, + "loss": 0.3843, + "step": 7110 + }, + { + "epoch": 3.8604777415852336, + "grad_norm": 11.857351610067585, + "learning_rate": 2.603085526337694e-06, + "loss": 0.431, + "step": 7111 + }, + { + "epoch": 3.8610206297502714, + "grad_norm": 8.758502750249699, + "learning_rate": 2.6007193401604626e-06, + "loss": 0.2799, + "step": 7112 + }, + { + "epoch": 3.8615635179153096, + "grad_norm": 11.732493468231032, + "learning_rate": 2.598354069138056e-06, + "loss": 0.4618, + "step": 7113 + }, + { + "epoch": 3.8621064060803474, + "grad_norm": 12.849925851696895, + "learning_rate": 2.595989713563014e-06, + "loss": 0.6998, + "step": 7114 + }, + { + "epoch": 3.8626492942453856, + "grad_norm": 8.79473787534307, + "learning_rate": 2.593626273727765e-06, + "loss": 0.2626, + "step": 7115 + }, + { + "epoch": 3.8631921824104234, + "grad_norm": 10.941535841138563, + "learning_rate": 2.5912637499246218e-06, + "loss": 0.4595, + "step": 7116 + }, + { + "epoch": 3.8637350705754616, + "grad_norm": 12.201214800033737, + "learning_rate": 2.588902142445786e-06, + "loss": 0.3282, + "step": 7117 + }, + { + "epoch": 3.8642779587404994, + "grad_norm": 12.348184440426978, + "learning_rate": 2.5865414515833455e-06, + "loss": 0.5828, + "step": 7118 + }, + { + "epoch": 3.8648208469055376, + "grad_norm": 13.160180910276585, + "learning_rate": 2.5841816776292727e-06, + "loss": 0.3516, + "step": 7119 + }, + { + "epoch": 3.8653637350705754, + "grad_norm": 19.57808856867751, + "learning_rate": 2.581822820875429e-06, + "loss": 0.8939, + "step": 7120 + }, + { + "epoch": 3.8659066232356136, + "grad_norm": 10.92966233467966, + "learning_rate": 2.5794648816135627e-06, + "loss": 0.3509, + "step": 7121 + }, + { + "epoch": 3.8664495114006514, + "grad_norm": 14.189811177607414, + "learning_rate": 2.577107860135305e-06, + "loss": 0.5024, + "step": 7122 + }, + { + "epoch": 3.8669923995656896, + "grad_norm": 10.963014195472391, + "learning_rate": 2.5747517567321776e-06, + "loss": 0.3253, + "step": 7123 + }, + { + "epoch": 3.8675352877307274, + "grad_norm": 16.06753478220038, + "learning_rate": 2.5723965716955878e-06, + "loss": 0.5281, + "step": 7124 + }, + { + "epoch": 3.8680781758957656, + "grad_norm": 10.96537764415174, + "learning_rate": 2.5700423053168276e-06, + "loss": 0.5059, + "step": 7125 + }, + { + "epoch": 3.8686210640608034, + "grad_norm": 9.185810250309261, + "learning_rate": 2.567688957887077e-06, + "loss": 0.3768, + "step": 7126 + }, + { + "epoch": 3.8691639522258416, + "grad_norm": 13.984895212503027, + "learning_rate": 2.565336529697401e-06, + "loss": 0.5294, + "step": 7127 + }, + { + "epoch": 3.8697068403908794, + "grad_norm": 14.97489290708819, + "learning_rate": 2.562985021038752e-06, + "loss": 0.6177, + "step": 7128 + }, + { + "epoch": 3.8702497285559176, + "grad_norm": 15.68383969722191, + "learning_rate": 2.560634432201966e-06, + "loss": 0.4874, + "step": 7129 + }, + { + "epoch": 3.8707926167209554, + "grad_norm": 13.06746691168529, + "learning_rate": 2.5582847634777753e-06, + "loss": 0.771, + "step": 7130 + }, + { + "epoch": 3.8713355048859937, + "grad_norm": 11.661980178584857, + "learning_rate": 2.55593601515678e-06, + "loss": 0.4957, + "step": 7131 + }, + { + "epoch": 3.8718783930510314, + "grad_norm": 9.687387371930855, + "learning_rate": 2.5535881875294875e-06, + "loss": 0.238, + "step": 7132 + }, + { + "epoch": 3.8724212812160697, + "grad_norm": 12.53249739465081, + "learning_rate": 2.551241280886274e-06, + "loss": 0.3521, + "step": 7133 + }, + { + "epoch": 3.8729641693811074, + "grad_norm": 18.057246108983335, + "learning_rate": 2.548895295517406e-06, + "loss": 0.9307, + "step": 7134 + }, + { + "epoch": 3.8735070575461457, + "grad_norm": 15.513117876289204, + "learning_rate": 2.546550231713051e-06, + "loss": 0.6529, + "step": 7135 + }, + { + "epoch": 3.8740499457111834, + "grad_norm": 11.883743610451969, + "learning_rate": 2.544206089763235e-06, + "loss": 0.7861, + "step": 7136 + }, + { + "epoch": 3.8745928338762217, + "grad_norm": 14.684059381744154, + "learning_rate": 2.5418628699578986e-06, + "loss": 0.412, + "step": 7137 + }, + { + "epoch": 3.8751357220412594, + "grad_norm": 7.135402796193207, + "learning_rate": 2.539520572586843e-06, + "loss": 0.2079, + "step": 7138 + }, + { + "epoch": 3.8756786102062977, + "grad_norm": 10.289771258088079, + "learning_rate": 2.5371791979397766e-06, + "loss": 0.3274, + "step": 7139 + }, + { + "epoch": 3.8762214983713354, + "grad_norm": 9.949109620860526, + "learning_rate": 2.5348387463062808e-06, + "loss": 0.4355, + "step": 7140 + }, + { + "epoch": 3.8767643865363732, + "grad_norm": 10.991037946192776, + "learning_rate": 2.5324992179758268e-06, + "loss": 0.4432, + "step": 7141 + }, + { + "epoch": 3.8773072747014115, + "grad_norm": 15.18957067683597, + "learning_rate": 2.5301606132377733e-06, + "loss": 0.8666, + "step": 7142 + }, + { + "epoch": 3.8778501628664497, + "grad_norm": 14.818504355332566, + "learning_rate": 2.5278229323813553e-06, + "loss": 0.6253, + "step": 7143 + }, + { + "epoch": 3.8783930510314875, + "grad_norm": 13.754502247150116, + "learning_rate": 2.5254861756957115e-06, + "loss": 0.5842, + "step": 7144 + }, + { + "epoch": 3.8789359391965252, + "grad_norm": 19.3020908869945, + "learning_rate": 2.5231503434698435e-06, + "loss": 0.6372, + "step": 7145 + }, + { + "epoch": 3.8794788273615635, + "grad_norm": 8.868163440639846, + "learning_rate": 2.5208154359926606e-06, + "loss": 0.4229, + "step": 7146 + }, + { + "epoch": 3.8800217155266017, + "grad_norm": 10.130535143618188, + "learning_rate": 2.5184814535529457e-06, + "loss": 0.3024, + "step": 7147 + }, + { + "epoch": 3.8805646036916395, + "grad_norm": 10.20257149827293, + "learning_rate": 2.5161483964393676e-06, + "loss": 0.3697, + "step": 7148 + }, + { + "epoch": 3.8811074918566772, + "grad_norm": 17.03033594136655, + "learning_rate": 2.513816264940483e-06, + "loss": 0.6084, + "step": 7149 + }, + { + "epoch": 3.8816503800217155, + "grad_norm": 9.767302667294265, + "learning_rate": 2.5114850593447336e-06, + "loss": 0.2872, + "step": 7150 + }, + { + "epoch": 3.8821932681867537, + "grad_norm": 11.550149537445538, + "learning_rate": 2.509154779940447e-06, + "loss": 0.4796, + "step": 7151 + }, + { + "epoch": 3.8827361563517915, + "grad_norm": 13.607872338694929, + "learning_rate": 2.5068254270158364e-06, + "loss": 0.6471, + "step": 7152 + }, + { + "epoch": 3.8832790445168293, + "grad_norm": 13.786046110369192, + "learning_rate": 2.504497000859e-06, + "loss": 0.6666, + "step": 7153 + }, + { + "epoch": 3.8838219326818675, + "grad_norm": 11.281665232889665, + "learning_rate": 2.5021695017579193e-06, + "loss": 0.5587, + "step": 7154 + }, + { + "epoch": 3.8843648208469057, + "grad_norm": 11.481761454219619, + "learning_rate": 2.4998429300004657e-06, + "loss": 0.3071, + "step": 7155 + }, + { + "epoch": 3.8849077090119435, + "grad_norm": 13.28075154587505, + "learning_rate": 2.4975172858743914e-06, + "loss": 0.3665, + "step": 7156 + }, + { + "epoch": 3.8854505971769813, + "grad_norm": 9.802453085562453, + "learning_rate": 2.4951925696673352e-06, + "loss": 0.3148, + "step": 7157 + }, + { + "epoch": 3.8859934853420195, + "grad_norm": 14.520688569604944, + "learning_rate": 2.492868781666824e-06, + "loss": 0.4942, + "step": 7158 + }, + { + "epoch": 3.8865363735070577, + "grad_norm": 10.711275878066797, + "learning_rate": 2.4905459221602667e-06, + "loss": 0.3353, + "step": 7159 + }, + { + "epoch": 3.8870792616720955, + "grad_norm": 12.549002018795433, + "learning_rate": 2.488223991434955e-06, + "loss": 0.3241, + "step": 7160 + }, + { + "epoch": 3.8876221498371333, + "grad_norm": 13.961472675915795, + "learning_rate": 2.485902989778077e-06, + "loss": 0.5143, + "step": 7161 + }, + { + "epoch": 3.8881650380021715, + "grad_norm": 10.194253119170678, + "learning_rate": 2.483582917476691e-06, + "loss": 0.3732, + "step": 7162 + }, + { + "epoch": 3.8887079261672097, + "grad_norm": 11.408277950635817, + "learning_rate": 2.481263774817748e-06, + "loss": 0.5475, + "step": 7163 + }, + { + "epoch": 3.8892508143322475, + "grad_norm": 11.720624145965107, + "learning_rate": 2.4789455620880855e-06, + "loss": 0.4152, + "step": 7164 + }, + { + "epoch": 3.8897937024972853, + "grad_norm": 17.498990068476704, + "learning_rate": 2.4766282795744225e-06, + "loss": 0.4959, + "step": 7165 + }, + { + "epoch": 3.8903365906623235, + "grad_norm": 10.025081073758155, + "learning_rate": 2.474311927563364e-06, + "loss": 0.2563, + "step": 7166 + }, + { + "epoch": 3.8908794788273617, + "grad_norm": 11.064975553515668, + "learning_rate": 2.4719965063413975e-06, + "loss": 0.4892, + "step": 7167 + }, + { + "epoch": 3.8914223669923995, + "grad_norm": 11.4169177186445, + "learning_rate": 2.4696820161949076e-06, + "loss": 0.4661, + "step": 7168 + }, + { + "epoch": 3.8919652551574373, + "grad_norm": 11.127726171001362, + "learning_rate": 2.4673684574101407e-06, + "loss": 0.5133, + "step": 7169 + }, + { + "epoch": 3.8925081433224755, + "grad_norm": 14.4532269361977, + "learning_rate": 2.4650558302732554e-06, + "loss": 0.6211, + "step": 7170 + }, + { + "epoch": 3.8930510314875137, + "grad_norm": 11.501842216375938, + "learning_rate": 2.4627441350702697e-06, + "loss": 0.43, + "step": 7171 + }, + { + "epoch": 3.8935939196525515, + "grad_norm": 11.827951767183402, + "learning_rate": 2.460433372087099e-06, + "loss": 0.4745, + "step": 7172 + }, + { + "epoch": 3.8941368078175893, + "grad_norm": 16.63470564409548, + "learning_rate": 2.4581235416095516e-06, + "loss": 0.7871, + "step": 7173 + }, + { + "epoch": 3.8946796959826275, + "grad_norm": 13.957180122419292, + "learning_rate": 2.455814643923298e-06, + "loss": 0.435, + "step": 7174 + }, + { + "epoch": 3.8952225841476658, + "grad_norm": 14.844456130947902, + "learning_rate": 2.4535066793139194e-06, + "loss": 0.3498, + "step": 7175 + }, + { + "epoch": 3.8957654723127035, + "grad_norm": 13.063011297960943, + "learning_rate": 2.4511996480668554e-06, + "loss": 0.4807, + "step": 7176 + }, + { + "epoch": 3.8963083604777413, + "grad_norm": 11.343167538907203, + "learning_rate": 2.448893550467456e-06, + "loss": 0.6566, + "step": 7177 + }, + { + "epoch": 3.8968512486427795, + "grad_norm": 21.134566936836904, + "learning_rate": 2.4465883868009323e-06, + "loss": 0.7527, + "step": 7178 + }, + { + "epoch": 3.8973941368078178, + "grad_norm": 12.658548113922626, + "learning_rate": 2.4442841573523967e-06, + "loss": 0.388, + "step": 7179 + }, + { + "epoch": 3.8979370249728555, + "grad_norm": 10.941499945328948, + "learning_rate": 2.4419808624068396e-06, + "loss": 0.585, + "step": 7180 + }, + { + "epoch": 3.8984799131378933, + "grad_norm": 14.25189382311016, + "learning_rate": 2.4396785022491343e-06, + "loss": 0.6702, + "step": 7181 + }, + { + "epoch": 3.8990228013029316, + "grad_norm": 16.90343774367846, + "learning_rate": 2.4373770771640448e-06, + "loss": 0.5498, + "step": 7182 + }, + { + "epoch": 3.8995656894679698, + "grad_norm": 14.478488557861258, + "learning_rate": 2.4350765874362047e-06, + "loss": 0.6703, + "step": 7183 + }, + { + "epoch": 3.9001085776330076, + "grad_norm": 11.280629928495003, + "learning_rate": 2.4327770333501522e-06, + "loss": 0.387, + "step": 7184 + }, + { + "epoch": 3.9006514657980453, + "grad_norm": 10.389018994464337, + "learning_rate": 2.430478415190297e-06, + "loss": 0.4396, + "step": 7185 + }, + { + "epoch": 3.9011943539630836, + "grad_norm": 12.229139584290909, + "learning_rate": 2.4281807332409358e-06, + "loss": 0.639, + "step": 7186 + }, + { + "epoch": 3.901737242128122, + "grad_norm": 18.238591257772253, + "learning_rate": 2.425883987786248e-06, + "loss": 0.6162, + "step": 7187 + }, + { + "epoch": 3.9022801302931596, + "grad_norm": 14.326334376777995, + "learning_rate": 2.423588179110301e-06, + "loss": 0.8227, + "step": 7188 + }, + { + "epoch": 3.9028230184581973, + "grad_norm": 14.43500891639456, + "learning_rate": 2.4212933074970423e-06, + "loss": 0.5057, + "step": 7189 + }, + { + "epoch": 3.9033659066232356, + "grad_norm": 14.30031793020835, + "learning_rate": 2.4189993732303063e-06, + "loss": 0.4514, + "step": 7190 + }, + { + "epoch": 3.903908794788274, + "grad_norm": 9.814097341569619, + "learning_rate": 2.4167063765938103e-06, + "loss": 0.3212, + "step": 7191 + }, + { + "epoch": 3.9044516829533116, + "grad_norm": 13.100322061273497, + "learning_rate": 2.4144143178711555e-06, + "loss": 0.5785, + "step": 7192 + }, + { + "epoch": 3.9049945711183494, + "grad_norm": 15.547319070400764, + "learning_rate": 2.412123197345827e-06, + "loss": 0.6, + "step": 7193 + }, + { + "epoch": 3.9055374592833876, + "grad_norm": 10.568502118830102, + "learning_rate": 2.409833015301195e-06, + "loss": 0.4499, + "step": 7194 + }, + { + "epoch": 3.906080347448426, + "grad_norm": 12.216551577948717, + "learning_rate": 2.4075437720205132e-06, + "loss": 0.3987, + "step": 7195 + }, + { + "epoch": 3.9066232356134636, + "grad_norm": 14.203010493004756, + "learning_rate": 2.4052554677869165e-06, + "loss": 0.7274, + "step": 7196 + }, + { + "epoch": 3.9071661237785014, + "grad_norm": 16.842358223683046, + "learning_rate": 2.4029681028834293e-06, + "loss": 0.6862, + "step": 7197 + }, + { + "epoch": 3.9077090119435396, + "grad_norm": 17.087528448096148, + "learning_rate": 2.4006816775929553e-06, + "loss": 0.6367, + "step": 7198 + }, + { + "epoch": 3.908251900108578, + "grad_norm": 13.559074748741233, + "learning_rate": 2.3983961921982823e-06, + "loss": 0.497, + "step": 7199 + }, + { + "epoch": 3.9087947882736156, + "grad_norm": 11.82349002805914, + "learning_rate": 2.3961116469820834e-06, + "loss": 0.3808, + "step": 7200 + }, + { + "epoch": 3.9093376764386534, + "grad_norm": 9.536111713709516, + "learning_rate": 2.3938280422269143e-06, + "loss": 0.2577, + "step": 7201 + }, + { + "epoch": 3.9098805646036916, + "grad_norm": 13.40690087309448, + "learning_rate": 2.3915453782152166e-06, + "loss": 0.5696, + "step": 7202 + }, + { + "epoch": 3.91042345276873, + "grad_norm": 11.767180310470792, + "learning_rate": 2.3892636552293114e-06, + "loss": 0.447, + "step": 7203 + }, + { + "epoch": 3.9109663409337676, + "grad_norm": 15.579650605609013, + "learning_rate": 2.3869828735514076e-06, + "loss": 0.5241, + "step": 7204 + }, + { + "epoch": 3.9115092290988054, + "grad_norm": 10.113811350963385, + "learning_rate": 2.3847030334635923e-06, + "loss": 0.3542, + "step": 7205 + }, + { + "epoch": 3.9120521172638436, + "grad_norm": 10.589793992165902, + "learning_rate": 2.3824241352478484e-06, + "loss": 0.2795, + "step": 7206 + }, + { + "epoch": 3.912595005428882, + "grad_norm": 7.9421075555522505, + "learning_rate": 2.3801461791860226e-06, + "loss": 0.2936, + "step": 7207 + }, + { + "epoch": 3.9131378935939196, + "grad_norm": 9.535162659306835, + "learning_rate": 2.377869165559867e-06, + "loss": 0.5889, + "step": 7208 + }, + { + "epoch": 3.9136807817589574, + "grad_norm": 11.900400034485365, + "learning_rate": 2.375593094650995e-06, + "loss": 0.5422, + "step": 7209 + }, + { + "epoch": 3.9142236699239956, + "grad_norm": 11.395565814320083, + "learning_rate": 2.3733179667409247e-06, + "loss": 0.4194, + "step": 7210 + }, + { + "epoch": 3.914766558089034, + "grad_norm": 13.14302661684852, + "learning_rate": 2.3710437821110456e-06, + "loss": 0.5611, + "step": 7211 + }, + { + "epoch": 3.9153094462540716, + "grad_norm": 14.854169528681664, + "learning_rate": 2.3687705410426242e-06, + "loss": 0.676, + "step": 7212 + }, + { + "epoch": 3.9158523344191094, + "grad_norm": 14.377106944895534, + "learning_rate": 2.3664982438168305e-06, + "loss": 0.5029, + "step": 7213 + }, + { + "epoch": 3.9163952225841476, + "grad_norm": 16.2949751889776, + "learning_rate": 2.364226890714694e-06, + "loss": 0.4424, + "step": 7214 + }, + { + "epoch": 3.916938110749186, + "grad_norm": 12.453892587793705, + "learning_rate": 2.3619564820171515e-06, + "loss": 0.3498, + "step": 7215 + }, + { + "epoch": 3.9174809989142236, + "grad_norm": 10.510207014160567, + "learning_rate": 2.359687018004998e-06, + "loss": 0.4104, + "step": 7216 + }, + { + "epoch": 3.9180238870792614, + "grad_norm": 10.817797408933902, + "learning_rate": 2.3574184989589345e-06, + "loss": 0.466, + "step": 7217 + }, + { + "epoch": 3.9185667752442996, + "grad_norm": 9.615419411562353, + "learning_rate": 2.3551509251595315e-06, + "loss": 0.3975, + "step": 7218 + }, + { + "epoch": 3.919109663409338, + "grad_norm": 10.629184513933357, + "learning_rate": 2.3528842968872456e-06, + "loss": 0.3305, + "step": 7219 + }, + { + "epoch": 3.9196525515743756, + "grad_norm": 14.741934283339353, + "learning_rate": 2.350618614422421e-06, + "loss": 0.5822, + "step": 7220 + }, + { + "epoch": 3.9201954397394134, + "grad_norm": 10.145231872720867, + "learning_rate": 2.348353878045272e-06, + "loss": 0.3223, + "step": 7221 + }, + { + "epoch": 3.9207383279044516, + "grad_norm": 9.719525157659627, + "learning_rate": 2.346090088035913e-06, + "loss": 0.323, + "step": 7222 + }, + { + "epoch": 3.92128121606949, + "grad_norm": 11.153861466428655, + "learning_rate": 2.3438272446743293e-06, + "loss": 0.4453, + "step": 7223 + }, + { + "epoch": 3.9218241042345277, + "grad_norm": 16.881204557198902, + "learning_rate": 2.3415653482403954e-06, + "loss": 0.8422, + "step": 7224 + }, + { + "epoch": 3.9223669923995654, + "grad_norm": 13.409647968671297, + "learning_rate": 2.339304399013864e-06, + "loss": 0.5316, + "step": 7225 + }, + { + "epoch": 3.9229098805646037, + "grad_norm": 17.78528943034469, + "learning_rate": 2.337044397274375e-06, + "loss": 1.0815, + "step": 7226 + }, + { + "epoch": 3.923452768729642, + "grad_norm": 13.521330366302392, + "learning_rate": 2.3347853433014467e-06, + "loss": 0.6285, + "step": 7227 + }, + { + "epoch": 3.9239956568946797, + "grad_norm": 15.033911656714881, + "learning_rate": 2.3325272373744844e-06, + "loss": 0.4578, + "step": 7228 + }, + { + "epoch": 3.9245385450597174, + "grad_norm": 13.063924995112636, + "learning_rate": 2.330270079772774e-06, + "loss": 0.4246, + "step": 7229 + }, + { + "epoch": 3.9250814332247557, + "grad_norm": 11.872225399295557, + "learning_rate": 2.328013870775483e-06, + "loss": 0.354, + "step": 7230 + }, + { + "epoch": 3.925624321389794, + "grad_norm": 10.802228004866322, + "learning_rate": 2.325758610661664e-06, + "loss": 0.2926, + "step": 7231 + }, + { + "epoch": 3.9261672095548317, + "grad_norm": 10.24423290229698, + "learning_rate": 2.323504299710251e-06, + "loss": 0.3887, + "step": 7232 + }, + { + "epoch": 3.9267100977198695, + "grad_norm": 10.40790875242795, + "learning_rate": 2.32125093820006e-06, + "loss": 0.3774, + "step": 7233 + }, + { + "epoch": 3.9272529858849077, + "grad_norm": 10.819416071470908, + "learning_rate": 2.3189985264097925e-06, + "loss": 0.4433, + "step": 7234 + }, + { + "epoch": 3.927795874049946, + "grad_norm": 14.400478824494995, + "learning_rate": 2.316747064618028e-06, + "loss": 0.6139, + "step": 7235 + }, + { + "epoch": 3.9283387622149837, + "grad_norm": 13.400576026026663, + "learning_rate": 2.3144965531032314e-06, + "loss": 0.6807, + "step": 7236 + }, + { + "epoch": 3.9288816503800215, + "grad_norm": 12.144971157259214, + "learning_rate": 2.3122469921437507e-06, + "loss": 0.406, + "step": 7237 + }, + { + "epoch": 3.9294245385450597, + "grad_norm": 13.566955931406746, + "learning_rate": 2.3099983820178116e-06, + "loss": 0.6583, + "step": 7238 + }, + { + "epoch": 3.929967426710098, + "grad_norm": 13.453764479761338, + "learning_rate": 2.3077507230035345e-06, + "loss": 0.8706, + "step": 7239 + }, + { + "epoch": 3.9305103148751357, + "grad_norm": 15.149717368244767, + "learning_rate": 2.305504015378904e-06, + "loss": 0.6808, + "step": 7240 + }, + { + "epoch": 3.9310532030401735, + "grad_norm": 15.68496135669598, + "learning_rate": 2.303258259421801e-06, + "loss": 0.4163, + "step": 7241 + }, + { + "epoch": 3.9315960912052117, + "grad_norm": 18.73387442818072, + "learning_rate": 2.301013455409983e-06, + "loss": 0.972, + "step": 7242 + }, + { + "epoch": 3.93213897937025, + "grad_norm": 11.659704005587447, + "learning_rate": 2.298769603621088e-06, + "loss": 0.4766, + "step": 7243 + }, + { + "epoch": 3.9326818675352877, + "grad_norm": 12.907778650132892, + "learning_rate": 2.296526704332648e-06, + "loss": 0.5669, + "step": 7244 + }, + { + "epoch": 3.9332247557003255, + "grad_norm": 14.095811490299557, + "learning_rate": 2.294284757822057e-06, + "loss": 0.6699, + "step": 7245 + }, + { + "epoch": 3.9337676438653637, + "grad_norm": 14.347126906125892, + "learning_rate": 2.292043764366615e-06, + "loss": 0.5478, + "step": 7246 + }, + { + "epoch": 3.934310532030402, + "grad_norm": 14.15574104757881, + "learning_rate": 2.289803724243478e-06, + "loss": 0.3451, + "step": 7247 + }, + { + "epoch": 3.9348534201954397, + "grad_norm": 16.273077860218656, + "learning_rate": 2.28756463772971e-06, + "loss": 0.433, + "step": 7248 + }, + { + "epoch": 3.9353963083604775, + "grad_norm": 12.905791138340465, + "learning_rate": 2.2853265051022376e-06, + "loss": 0.3549, + "step": 7249 + }, + { + "epoch": 3.9359391965255157, + "grad_norm": 10.740402790195867, + "learning_rate": 2.283089326637875e-06, + "loss": 0.294, + "step": 7250 + }, + { + "epoch": 3.936482084690554, + "grad_norm": 16.206654981483148, + "learning_rate": 2.2808531026133297e-06, + "loss": 0.4773, + "step": 7251 + }, + { + "epoch": 3.9370249728555917, + "grad_norm": 13.723535114050927, + "learning_rate": 2.278617833305169e-06, + "loss": 0.5248, + "step": 7252 + }, + { + "epoch": 3.9375678610206295, + "grad_norm": 12.638788461264767, + "learning_rate": 2.2763835189898665e-06, + "loss": 0.6087, + "step": 7253 + }, + { + "epoch": 3.9381107491856677, + "grad_norm": 12.333038982803954, + "learning_rate": 2.2741501599437543e-06, + "loss": 0.3829, + "step": 7254 + }, + { + "epoch": 3.938653637350706, + "grad_norm": 11.07699141293442, + "learning_rate": 2.2719177564430662e-06, + "loss": 0.6625, + "step": 7255 + }, + { + "epoch": 3.9391965255157437, + "grad_norm": 14.354975596263804, + "learning_rate": 2.2696863087639063e-06, + "loss": 0.6906, + "step": 7256 + }, + { + "epoch": 3.9397394136807815, + "grad_norm": 10.159046635813173, + "learning_rate": 2.2674558171822646e-06, + "loss": 0.4371, + "step": 7257 + }, + { + "epoch": 3.9402823018458197, + "grad_norm": 18.017466534966523, + "learning_rate": 2.265226281974011e-06, + "loss": 0.5347, + "step": 7258 + }, + { + "epoch": 3.940825190010858, + "grad_norm": 13.962772198264076, + "learning_rate": 2.2629977034148988e-06, + "loss": 0.4712, + "step": 7259 + }, + { + "epoch": 3.9413680781758957, + "grad_norm": 10.058544580261096, + "learning_rate": 2.2607700817805635e-06, + "loss": 0.3168, + "step": 7260 + }, + { + "epoch": 3.9419109663409335, + "grad_norm": 18.61661644453424, + "learning_rate": 2.258543417346514e-06, + "loss": 0.5639, + "step": 7261 + }, + { + "epoch": 3.9424538545059717, + "grad_norm": 12.38740485358688, + "learning_rate": 2.256317710388155e-06, + "loss": 0.4182, + "step": 7262 + }, + { + "epoch": 3.94299674267101, + "grad_norm": 12.633562464203537, + "learning_rate": 2.254092961180764e-06, + "loss": 0.5299, + "step": 7263 + }, + { + "epoch": 3.9435396308360477, + "grad_norm": 15.51128932246205, + "learning_rate": 2.2518691699995e-06, + "loss": 0.6563, + "step": 7264 + }, + { + "epoch": 3.9440825190010855, + "grad_norm": 14.124949663680717, + "learning_rate": 2.2496463371194065e-06, + "loss": 0.7836, + "step": 7265 + }, + { + "epoch": 3.9446254071661238, + "grad_norm": 14.035611568552724, + "learning_rate": 2.247424462815405e-06, + "loss": 0.4113, + "step": 7266 + }, + { + "epoch": 3.945168295331162, + "grad_norm": 13.30563635940546, + "learning_rate": 2.2452035473623022e-06, + "loss": 0.449, + "step": 7267 + }, + { + "epoch": 3.9457111834961998, + "grad_norm": 12.093768184800798, + "learning_rate": 2.242983591034784e-06, + "loss": 0.5745, + "step": 7268 + }, + { + "epoch": 3.9462540716612375, + "grad_norm": 13.19161002685443, + "learning_rate": 2.2407645941074185e-06, + "loss": 0.6073, + "step": 7269 + }, + { + "epoch": 3.9467969598262758, + "grad_norm": 10.454523102469183, + "learning_rate": 2.238546556854655e-06, + "loss": 0.265, + "step": 7270 + }, + { + "epoch": 3.947339847991314, + "grad_norm": 12.97022293790845, + "learning_rate": 2.2363294795508217e-06, + "loss": 0.5445, + "step": 7271 + }, + { + "epoch": 3.9478827361563518, + "grad_norm": 12.83418150037974, + "learning_rate": 2.234113362470133e-06, + "loss": 0.4941, + "step": 7272 + }, + { + "epoch": 3.9484256243213895, + "grad_norm": 12.615689050586889, + "learning_rate": 2.23189820588668e-06, + "loss": 0.8588, + "step": 7273 + }, + { + "epoch": 3.9489685124864278, + "grad_norm": 14.878456274053418, + "learning_rate": 2.2296840100744375e-06, + "loss": 0.4019, + "step": 7274 + }, + { + "epoch": 3.949511400651466, + "grad_norm": 13.017249525734119, + "learning_rate": 2.227470775307261e-06, + "loss": 0.5343, + "step": 7275 + }, + { + "epoch": 3.950054288816504, + "grad_norm": 11.575062166378828, + "learning_rate": 2.2252585018588836e-06, + "loss": 0.3775, + "step": 7276 + }, + { + "epoch": 3.9505971769815416, + "grad_norm": 15.290489125989911, + "learning_rate": 2.2230471900029303e-06, + "loss": 0.4912, + "step": 7277 + }, + { + "epoch": 3.95114006514658, + "grad_norm": 17.764334443185927, + "learning_rate": 2.220836840012891e-06, + "loss": 0.7215, + "step": 7278 + }, + { + "epoch": 3.951682953311618, + "grad_norm": 14.154407180706324, + "learning_rate": 2.218627452162154e-06, + "loss": 0.6472, + "step": 7279 + }, + { + "epoch": 3.952225841476656, + "grad_norm": 11.225658955685093, + "learning_rate": 2.2164190267239737e-06, + "loss": 0.2478, + "step": 7280 + }, + { + "epoch": 3.9527687296416936, + "grad_norm": 12.814332016509574, + "learning_rate": 2.2142115639714935e-06, + "loss": 0.5622, + "step": 7281 + }, + { + "epoch": 3.953311617806732, + "grad_norm": 19.419294319476467, + "learning_rate": 2.2120050641777345e-06, + "loss": 0.7996, + "step": 7282 + }, + { + "epoch": 3.95385450597177, + "grad_norm": 19.022041488664335, + "learning_rate": 2.209799527615599e-06, + "loss": 0.7117, + "step": 7283 + }, + { + "epoch": 3.954397394136808, + "grad_norm": 12.40634141862146, + "learning_rate": 2.20759495455788e-06, + "loss": 0.3793, + "step": 7284 + }, + { + "epoch": 3.9549402823018456, + "grad_norm": 8.873847512885336, + "learning_rate": 2.20539134527723e-06, + "loss": 0.3723, + "step": 7285 + }, + { + "epoch": 3.955483170466884, + "grad_norm": 15.833917455665446, + "learning_rate": 2.203188700046207e-06, + "loss": 0.6743, + "step": 7286 + }, + { + "epoch": 3.956026058631922, + "grad_norm": 12.353488572936646, + "learning_rate": 2.2009870191372263e-06, + "loss": 0.4195, + "step": 7287 + }, + { + "epoch": 3.95656894679696, + "grad_norm": 19.03028178846917, + "learning_rate": 2.198786302822603e-06, + "loss": 0.9055, + "step": 7288 + }, + { + "epoch": 3.9571118349619976, + "grad_norm": 15.81794212526427, + "learning_rate": 2.1965865513745265e-06, + "loss": 0.4088, + "step": 7289 + }, + { + "epoch": 3.957654723127036, + "grad_norm": 14.909566500608973, + "learning_rate": 2.1943877650650556e-06, + "loss": 0.9969, + "step": 7290 + }, + { + "epoch": 3.958197611292074, + "grad_norm": 12.332907917932245, + "learning_rate": 2.192189944166153e-06, + "loss": 0.3762, + "step": 7291 + }, + { + "epoch": 3.958740499457112, + "grad_norm": 13.036980814291068, + "learning_rate": 2.189993088949636e-06, + "loss": 0.4931, + "step": 7292 + }, + { + "epoch": 3.9592833876221496, + "grad_norm": 15.24047006419779, + "learning_rate": 2.187797199687224e-06, + "loss": 0.5143, + "step": 7293 + }, + { + "epoch": 3.959826275787188, + "grad_norm": 12.214872928235861, + "learning_rate": 2.185602276650505e-06, + "loss": 0.4249, + "step": 7294 + }, + { + "epoch": 3.960369163952226, + "grad_norm": 11.734417487796039, + "learning_rate": 2.18340832011095e-06, + "loss": 0.5159, + "step": 7295 + }, + { + "epoch": 3.960912052117264, + "grad_norm": 16.316727205367513, + "learning_rate": 2.181215330339912e-06, + "loss": 0.5141, + "step": 7296 + }, + { + "epoch": 3.9614549402823016, + "grad_norm": 13.896917482997276, + "learning_rate": 2.1790233076086243e-06, + "loss": 1.0396, + "step": 7297 + }, + { + "epoch": 3.96199782844734, + "grad_norm": 17.3480021128049, + "learning_rate": 2.1768322521882003e-06, + "loss": 0.43, + "step": 7298 + }, + { + "epoch": 3.962540716612378, + "grad_norm": 11.942778635499716, + "learning_rate": 2.1746421643496264e-06, + "loss": 0.383, + "step": 7299 + }, + { + "epoch": 3.963083604777416, + "grad_norm": 10.46804392176075, + "learning_rate": 2.1724530443637836e-06, + "loss": 0.6297, + "step": 7300 + }, + { + "epoch": 3.9636264929424536, + "grad_norm": 12.678863612803186, + "learning_rate": 2.1702648925014248e-06, + "loss": 0.4674, + "step": 7301 + }, + { + "epoch": 3.964169381107492, + "grad_norm": 7.353045500973433, + "learning_rate": 2.1680777090331816e-06, + "loss": 0.3103, + "step": 7302 + }, + { + "epoch": 3.96471226927253, + "grad_norm": 12.29878912174572, + "learning_rate": 2.1658914942295706e-06, + "loss": 0.5663, + "step": 7303 + }, + { + "epoch": 3.965255157437568, + "grad_norm": 8.497343343626396, + "learning_rate": 2.163706248360985e-06, + "loss": 0.3524, + "step": 7304 + }, + { + "epoch": 3.9657980456026056, + "grad_norm": 11.764041467674721, + "learning_rate": 2.1615219716977e-06, + "loss": 0.4548, + "step": 7305 + }, + { + "epoch": 3.966340933767644, + "grad_norm": 17.313346132201072, + "learning_rate": 2.1593386645098692e-06, + "loss": 1.0529, + "step": 7306 + }, + { + "epoch": 3.966883821932682, + "grad_norm": 13.90349930971142, + "learning_rate": 2.15715632706753e-06, + "loss": 0.5808, + "step": 7307 + }, + { + "epoch": 3.96742671009772, + "grad_norm": 13.776936265783004, + "learning_rate": 2.1549749596405945e-06, + "loss": 0.6697, + "step": 7308 + }, + { + "epoch": 3.9679695982627576, + "grad_norm": 16.504532446022996, + "learning_rate": 2.15279456249886e-06, + "loss": 0.5508, + "step": 7309 + }, + { + "epoch": 3.968512486427796, + "grad_norm": 9.948445138257954, + "learning_rate": 2.1506151359119997e-06, + "loss": 0.4461, + "step": 7310 + }, + { + "epoch": 3.969055374592834, + "grad_norm": 14.495691133016251, + "learning_rate": 2.1484366801495705e-06, + "loss": 0.7312, + "step": 7311 + }, + { + "epoch": 3.969598262757872, + "grad_norm": 16.82715961877643, + "learning_rate": 2.1462591954810054e-06, + "loss": 0.7524, + "step": 7312 + }, + { + "epoch": 3.9701411509229096, + "grad_norm": 11.225827985724983, + "learning_rate": 2.1440826821756213e-06, + "loss": 0.4354, + "step": 7313 + }, + { + "epoch": 3.970684039087948, + "grad_norm": 9.865289892982581, + "learning_rate": 2.141907140502607e-06, + "loss": 0.2906, + "step": 7314 + }, + { + "epoch": 3.971226927252986, + "grad_norm": 8.988273380587568, + "learning_rate": 2.139732570731048e-06, + "loss": 0.3062, + "step": 7315 + }, + { + "epoch": 3.971769815418024, + "grad_norm": 13.659816709804225, + "learning_rate": 2.1375589731298864e-06, + "loss": 0.382, + "step": 7316 + }, + { + "epoch": 3.9723127035830617, + "grad_norm": 15.628687568254344, + "learning_rate": 2.135386347967967e-06, + "loss": 0.6687, + "step": 7317 + }, + { + "epoch": 3.9728555917481, + "grad_norm": 10.910710203208446, + "learning_rate": 2.1332146955139963e-06, + "loss": 0.3909, + "step": 7318 + }, + { + "epoch": 3.973398479913138, + "grad_norm": 13.270954739017608, + "learning_rate": 2.13104401603657e-06, + "loss": 0.4808, + "step": 7319 + }, + { + "epoch": 3.973941368078176, + "grad_norm": 13.799228933349236, + "learning_rate": 2.12887430980416e-06, + "loss": 0.4717, + "step": 7320 + }, + { + "epoch": 3.9744842562432137, + "grad_norm": 12.928285689167119, + "learning_rate": 2.1267055770851185e-06, + "loss": 0.5013, + "step": 7321 + }, + { + "epoch": 3.975027144408252, + "grad_norm": 11.400185046713785, + "learning_rate": 2.124537818147684e-06, + "loss": 0.304, + "step": 7322 + }, + { + "epoch": 3.97557003257329, + "grad_norm": 10.537157565034327, + "learning_rate": 2.122371033259959e-06, + "loss": 0.3975, + "step": 7323 + }, + { + "epoch": 3.976112920738328, + "grad_norm": 14.40105822826117, + "learning_rate": 2.120205222689944e-06, + "loss": 0.6778, + "step": 7324 + }, + { + "epoch": 3.9766558089033657, + "grad_norm": 13.229824974968945, + "learning_rate": 2.118040386705501e-06, + "loss": 0.6712, + "step": 7325 + }, + { + "epoch": 3.977198697068404, + "grad_norm": 13.793638194835786, + "learning_rate": 2.1158765255743872e-06, + "loss": 0.4601, + "step": 7326 + }, + { + "epoch": 3.977741585233442, + "grad_norm": 15.505430296521286, + "learning_rate": 2.113713639564231e-06, + "loss": 0.8212, + "step": 7327 + }, + { + "epoch": 3.97828447339848, + "grad_norm": 17.09119757360545, + "learning_rate": 2.11155172894254e-06, + "loss": 0.7094, + "step": 7328 + }, + { + "epoch": 3.9788273615635177, + "grad_norm": 13.556531251232467, + "learning_rate": 2.1093907939767065e-06, + "loss": 0.4832, + "step": 7329 + }, + { + "epoch": 3.979370249728556, + "grad_norm": 14.237329918169115, + "learning_rate": 2.10723083493399e-06, + "loss": 0.3663, + "step": 7330 + }, + { + "epoch": 3.979913137893594, + "grad_norm": 9.012447122900646, + "learning_rate": 2.1050718520815485e-06, + "loss": 0.5217, + "step": 7331 + }, + { + "epoch": 3.980456026058632, + "grad_norm": 12.91832304111153, + "learning_rate": 2.1029138456863973e-06, + "loss": 0.7424, + "step": 7332 + }, + { + "epoch": 3.9809989142236697, + "grad_norm": 17.140779656879335, + "learning_rate": 2.1007568160154502e-06, + "loss": 0.5607, + "step": 7333 + }, + { + "epoch": 3.981541802388708, + "grad_norm": 15.535441443258492, + "learning_rate": 2.09860076333549e-06, + "loss": 0.6682, + "step": 7334 + }, + { + "epoch": 3.982084690553746, + "grad_norm": 9.108166871624157, + "learning_rate": 2.09644568791318e-06, + "loss": 0.3471, + "step": 7335 + }, + { + "epoch": 3.982627578718784, + "grad_norm": 13.8585496563886, + "learning_rate": 2.094291590015064e-06, + "loss": 0.4907, + "step": 7336 + }, + { + "epoch": 3.9831704668838217, + "grad_norm": 17.309058810158508, + "learning_rate": 2.0921384699075644e-06, + "loss": 0.7292, + "step": 7337 + }, + { + "epoch": 3.98371335504886, + "grad_norm": 11.888938470455745, + "learning_rate": 2.089986327856981e-06, + "loss": 0.5343, + "step": 7338 + }, + { + "epoch": 3.984256243213898, + "grad_norm": 14.96573909686316, + "learning_rate": 2.087835164129496e-06, + "loss": 0.436, + "step": 7339 + }, + { + "epoch": 3.984799131378936, + "grad_norm": 11.965090702213786, + "learning_rate": 2.085684978991168e-06, + "loss": 0.4307, + "step": 7340 + }, + { + "epoch": 3.9853420195439737, + "grad_norm": 9.089611131661139, + "learning_rate": 2.083535772707935e-06, + "loss": 0.3474, + "step": 7341 + }, + { + "epoch": 3.985884907709012, + "grad_norm": 15.457161419441805, + "learning_rate": 2.0813875455456156e-06, + "loss": 0.6491, + "step": 7342 + }, + { + "epoch": 3.98642779587405, + "grad_norm": 10.250596479107204, + "learning_rate": 2.0792402977699033e-06, + "loss": 0.3589, + "step": 7343 + }, + { + "epoch": 3.986970684039088, + "grad_norm": 14.949877353645148, + "learning_rate": 2.077094029646376e-06, + "loss": 0.8685, + "step": 7344 + }, + { + "epoch": 3.9875135722041257, + "grad_norm": 15.23923375936191, + "learning_rate": 2.074948741440486e-06, + "loss": 0.744, + "step": 7345 + }, + { + "epoch": 3.988056460369164, + "grad_norm": 11.133935275967207, + "learning_rate": 2.0728044334175667e-06, + "loss": 0.7823, + "step": 7346 + }, + { + "epoch": 3.988599348534202, + "grad_norm": 10.744562653679704, + "learning_rate": 2.0706611058428285e-06, + "loss": 0.3669, + "step": 7347 + }, + { + "epoch": 3.98914223669924, + "grad_norm": 12.814655704488615, + "learning_rate": 2.0685187589813625e-06, + "loss": 0.3812, + "step": 7348 + }, + { + "epoch": 3.9896851248642777, + "grad_norm": 10.282758906730834, + "learning_rate": 2.0663773930981367e-06, + "loss": 0.3229, + "step": 7349 + }, + { + "epoch": 3.990228013029316, + "grad_norm": 14.644174681230187, + "learning_rate": 2.064237008458e-06, + "loss": 0.7615, + "step": 7350 + }, + { + "epoch": 3.990770901194354, + "grad_norm": 16.98502298622554, + "learning_rate": 2.062097605325678e-06, + "loss": 0.5847, + "step": 7351 + }, + { + "epoch": 3.991313789359392, + "grad_norm": 12.815769655827422, + "learning_rate": 2.059959183965775e-06, + "loss": 0.3939, + "step": 7352 + }, + { + "epoch": 3.9918566775244297, + "grad_norm": 13.188644656309128, + "learning_rate": 2.057821744642774e-06, + "loss": 0.6187, + "step": 7353 + }, + { + "epoch": 3.992399565689468, + "grad_norm": 12.875041049936685, + "learning_rate": 2.0556852876210354e-06, + "loss": 0.3272, + "step": 7354 + }, + { + "epoch": 3.992942453854506, + "grad_norm": 13.754004827000545, + "learning_rate": 2.053549813164808e-06, + "loss": 0.4637, + "step": 7355 + }, + { + "epoch": 3.993485342019544, + "grad_norm": 10.58203405942224, + "learning_rate": 2.0514153215381983e-06, + "loss": 0.5784, + "step": 7356 + }, + { + "epoch": 3.9940282301845818, + "grad_norm": 10.96146456022653, + "learning_rate": 2.049281813005215e-06, + "loss": 0.4389, + "step": 7357 + }, + { + "epoch": 3.99457111834962, + "grad_norm": 12.017836415062291, + "learning_rate": 2.047149287829726e-06, + "loss": 0.5839, + "step": 7358 + }, + { + "epoch": 3.995114006514658, + "grad_norm": 10.822043166102832, + "learning_rate": 2.045017746275485e-06, + "loss": 0.4467, + "step": 7359 + }, + { + "epoch": 3.995656894679696, + "grad_norm": 13.151255799546034, + "learning_rate": 2.0428871886061343e-06, + "loss": 0.6173, + "step": 7360 + }, + { + "epoch": 3.9961997828447338, + "grad_norm": 11.118960416841484, + "learning_rate": 2.0407576150851705e-06, + "loss": 0.3935, + "step": 7361 + }, + { + "epoch": 3.996742671009772, + "grad_norm": 12.928271003217045, + "learning_rate": 2.0386290259759967e-06, + "loss": 0.5222, + "step": 7362 + }, + { + "epoch": 3.99728555917481, + "grad_norm": 14.237952289938509, + "learning_rate": 2.0365014215418676e-06, + "loss": 0.495, + "step": 7363 + }, + { + "epoch": 3.997828447339848, + "grad_norm": 15.550337043796231, + "learning_rate": 2.0343748020459374e-06, + "loss": 0.4867, + "step": 7364 + }, + { + "epoch": 3.9983713355048858, + "grad_norm": 12.364600226217249, + "learning_rate": 2.032249167751228e-06, + "loss": 0.6296, + "step": 7365 + }, + { + "epoch": 3.998914223669924, + "grad_norm": 13.569416929145026, + "learning_rate": 2.0301245189206385e-06, + "loss": 0.5857, + "step": 7366 + }, + { + "epoch": 3.999457111834962, + "grad_norm": 14.915873617390455, + "learning_rate": 2.028000855816954e-06, + "loss": 0.7152, + "step": 7367 + }, + { + "epoch": 4.0, + "grad_norm": 13.686794191538802, + "learning_rate": 2.025878178702825e-06, + "loss": 0.3451, + "step": 7368 + }, + { + "epoch": 4.000542888165038, + "grad_norm": 14.607866561626649, + "learning_rate": 2.0237564878407956e-06, + "loss": 0.5273, + "step": 7369 + }, + { + "epoch": 4.001085776330076, + "grad_norm": 10.130331812248723, + "learning_rate": 2.021635783493271e-06, + "loss": 0.2723, + "step": 7370 + }, + { + "epoch": 4.001628664495114, + "grad_norm": 10.032934671154406, + "learning_rate": 2.019516065922551e-06, + "loss": 0.3576, + "step": 7371 + }, + { + "epoch": 4.002171552660152, + "grad_norm": 11.657828066715073, + "learning_rate": 2.017397335390803e-06, + "loss": 0.3434, + "step": 7372 + }, + { + "epoch": 4.00271444082519, + "grad_norm": 13.248765523472253, + "learning_rate": 2.0152795921600745e-06, + "loss": 0.5575, + "step": 7373 + }, + { + "epoch": 4.003257328990228, + "grad_norm": 13.375690344309364, + "learning_rate": 2.013162836492293e-06, + "loss": 0.4909, + "step": 7374 + }, + { + "epoch": 4.003800217155266, + "grad_norm": 13.29236800364953, + "learning_rate": 2.01104706864926e-06, + "loss": 0.4056, + "step": 7375 + }, + { + "epoch": 4.004343105320304, + "grad_norm": 11.112608581516877, + "learning_rate": 2.0089322888926577e-06, + "loss": 0.3764, + "step": 7376 + }, + { + "epoch": 4.004885993485342, + "grad_norm": 13.400379725287804, + "learning_rate": 2.0068184974840464e-06, + "loss": 0.5315, + "step": 7377 + }, + { + "epoch": 4.00542888165038, + "grad_norm": 12.399245820402392, + "learning_rate": 2.004705694684863e-06, + "loss": 0.4755, + "step": 7378 + }, + { + "epoch": 4.005971769815418, + "grad_norm": 7.235424189620777, + "learning_rate": 2.0025938807564207e-06, + "loss": 0.2616, + "step": 7379 + }, + { + "epoch": 4.006514657980456, + "grad_norm": 14.18643915790045, + "learning_rate": 2.0004830559599143e-06, + "loss": 0.3706, + "step": 7380 + }, + { + "epoch": 4.007057546145494, + "grad_norm": 10.127056422809925, + "learning_rate": 1.998373220556412e-06, + "loss": 0.3258, + "step": 7381 + }, + { + "epoch": 4.007600434310532, + "grad_norm": 17.273359103095373, + "learning_rate": 1.9962643748068633e-06, + "loss": 0.6007, + "step": 7382 + }, + { + "epoch": 4.00814332247557, + "grad_norm": 10.953175473484771, + "learning_rate": 1.994156518972092e-06, + "loss": 0.4094, + "step": 7383 + }, + { + "epoch": 4.008686210640608, + "grad_norm": 10.927954583193763, + "learning_rate": 1.992049653312802e-06, + "loss": 0.4199, + "step": 7384 + }, + { + "epoch": 4.009229098805646, + "grad_norm": 13.711331521508846, + "learning_rate": 1.9899437780895716e-06, + "loss": 0.6639, + "step": 7385 + }, + { + "epoch": 4.009771986970684, + "grad_norm": 12.537819678821979, + "learning_rate": 1.9878388935628655e-06, + "loss": 0.3551, + "step": 7386 + }, + { + "epoch": 4.010314875135722, + "grad_norm": 11.448288732996893, + "learning_rate": 1.985734999993013e-06, + "loss": 0.4247, + "step": 7387 + }, + { + "epoch": 4.01085776330076, + "grad_norm": 17.00322545288514, + "learning_rate": 1.9836320976402267e-06, + "loss": 0.5907, + "step": 7388 + }, + { + "epoch": 4.011400651465798, + "grad_norm": 10.797125749762163, + "learning_rate": 1.981530186764601e-06, + "loss": 0.4138, + "step": 7389 + }, + { + "epoch": 4.011943539630836, + "grad_norm": 17.95998607345293, + "learning_rate": 1.9794292676260997e-06, + "loss": 0.6222, + "step": 7390 + }, + { + "epoch": 4.012486427795874, + "grad_norm": 14.411778567127596, + "learning_rate": 1.9773293404845696e-06, + "loss": 0.421, + "step": 7391 + }, + { + "epoch": 4.013029315960912, + "grad_norm": 10.2409856593537, + "learning_rate": 1.9752304055997305e-06, + "loss": 0.3125, + "step": 7392 + }, + { + "epoch": 4.01357220412595, + "grad_norm": 11.292522294180968, + "learning_rate": 1.97313246323119e-06, + "loss": 0.2754, + "step": 7393 + }, + { + "epoch": 4.014115092290988, + "grad_norm": 10.928696930486545, + "learning_rate": 1.971035513638414e-06, + "loss": 0.4488, + "step": 7394 + }, + { + "epoch": 4.014657980456026, + "grad_norm": 12.415811489917902, + "learning_rate": 1.968939557080767e-06, + "loss": 0.436, + "step": 7395 + }, + { + "epoch": 4.015200868621064, + "grad_norm": 11.13510857929724, + "learning_rate": 1.9668445938174717e-06, + "loss": 0.3495, + "step": 7396 + }, + { + "epoch": 4.015743756786102, + "grad_norm": 9.229951548837793, + "learning_rate": 1.9647506241076387e-06, + "loss": 0.2861, + "step": 7397 + }, + { + "epoch": 4.01628664495114, + "grad_norm": 16.847132987227848, + "learning_rate": 1.96265764821026e-06, + "loss": 0.5219, + "step": 7398 + }, + { + "epoch": 4.016829533116178, + "grad_norm": 11.84958702376863, + "learning_rate": 1.9605656663841867e-06, + "loss": 0.4128, + "step": 7399 + }, + { + "epoch": 4.017372421281216, + "grad_norm": 14.742001988102418, + "learning_rate": 1.95847467888817e-06, + "loss": 0.4796, + "step": 7400 + }, + { + "epoch": 4.017915309446254, + "grad_norm": 11.28536118564977, + "learning_rate": 1.956384685980818e-06, + "loss": 0.3599, + "step": 7401 + }, + { + "epoch": 4.018458197611292, + "grad_norm": 15.721826059355514, + "learning_rate": 1.954295687920631e-06, + "loss": 0.5619, + "step": 7402 + }, + { + "epoch": 4.01900108577633, + "grad_norm": 15.983663047461183, + "learning_rate": 1.952207684965971e-06, + "loss": 0.6122, + "step": 7403 + }, + { + "epoch": 4.019543973941368, + "grad_norm": 12.128334107658644, + "learning_rate": 1.9501206773750947e-06, + "loss": 0.4605, + "step": 7404 + }, + { + "epoch": 4.020086862106406, + "grad_norm": 11.36806204402688, + "learning_rate": 1.948034665406121e-06, + "loss": 0.3116, + "step": 7405 + }, + { + "epoch": 4.020629750271444, + "grad_norm": 10.68555301238784, + "learning_rate": 1.9459496493170536e-06, + "loss": 0.321, + "step": 7406 + }, + { + "epoch": 4.021172638436482, + "grad_norm": 9.879738893549789, + "learning_rate": 1.9438656293657733e-06, + "loss": 0.2991, + "step": 7407 + }, + { + "epoch": 4.02171552660152, + "grad_norm": 13.176679205393846, + "learning_rate": 1.9417826058100253e-06, + "loss": 0.6426, + "step": 7408 + }, + { + "epoch": 4.022258414766558, + "grad_norm": 9.5138082645493, + "learning_rate": 1.9397005789074497e-06, + "loss": 0.3611, + "step": 7409 + }, + { + "epoch": 4.022801302931596, + "grad_norm": 9.82645446405319, + "learning_rate": 1.9376195489155537e-06, + "loss": 0.2537, + "step": 7410 + }, + { + "epoch": 4.023344191096634, + "grad_norm": 12.161440619011307, + "learning_rate": 1.935539516091721e-06, + "loss": 0.3498, + "step": 7411 + }, + { + "epoch": 4.023887079261672, + "grad_norm": 14.534994929925375, + "learning_rate": 1.9334604806932143e-06, + "loss": 0.5933, + "step": 7412 + }, + { + "epoch": 4.02442996742671, + "grad_norm": 13.507360209479582, + "learning_rate": 1.931382442977171e-06, + "loss": 0.3563, + "step": 7413 + }, + { + "epoch": 4.024972855591748, + "grad_norm": 15.046753479367776, + "learning_rate": 1.9293054032006063e-06, + "loss": 0.4387, + "step": 7414 + }, + { + "epoch": 4.025515743756786, + "grad_norm": 12.258959062712908, + "learning_rate": 1.927229361620412e-06, + "loss": 0.4434, + "step": 7415 + }, + { + "epoch": 4.026058631921824, + "grad_norm": 12.015424398969936, + "learning_rate": 1.9251543184933576e-06, + "loss": 0.4404, + "step": 7416 + }, + { + "epoch": 4.026601520086862, + "grad_norm": 9.477634559865093, + "learning_rate": 1.9230802740760857e-06, + "loss": 0.3345, + "step": 7417 + }, + { + "epoch": 4.0271444082519, + "grad_norm": 14.279619141455536, + "learning_rate": 1.921007228625118e-06, + "loss": 0.3984, + "step": 7418 + }, + { + "epoch": 4.027687296416938, + "grad_norm": 21.075479101285925, + "learning_rate": 1.9189351823968518e-06, + "loss": 0.4166, + "step": 7419 + }, + { + "epoch": 4.028230184581976, + "grad_norm": 9.220949774925229, + "learning_rate": 1.9168641356475625e-06, + "loss": 0.2762, + "step": 7420 + }, + { + "epoch": 4.028773072747014, + "grad_norm": 14.44225601755817, + "learning_rate": 1.9147940886333994e-06, + "loss": 0.564, + "step": 7421 + }, + { + "epoch": 4.029315960912052, + "grad_norm": 9.06166197984856, + "learning_rate": 1.912725041610388e-06, + "loss": 0.3853, + "step": 7422 + }, + { + "epoch": 4.02985884907709, + "grad_norm": 13.498032050668492, + "learning_rate": 1.9106569948344344e-06, + "loss": 0.4768, + "step": 7423 + }, + { + "epoch": 4.030401737242128, + "grad_norm": 13.191503934950113, + "learning_rate": 1.908589948561316e-06, + "loss": 0.3802, + "step": 7424 + }, + { + "epoch": 4.030944625407166, + "grad_norm": 12.094983329355872, + "learning_rate": 1.9065239030466854e-06, + "loss": 0.3355, + "step": 7425 + }, + { + "epoch": 4.031487513572204, + "grad_norm": 10.332292929620861, + "learning_rate": 1.9044588585460821e-06, + "loss": 0.3172, + "step": 7426 + }, + { + "epoch": 4.032030401737242, + "grad_norm": 13.8560757790004, + "learning_rate": 1.9023948153149075e-06, + "loss": 0.3627, + "step": 7427 + }, + { + "epoch": 4.03257328990228, + "grad_norm": 10.601587706717948, + "learning_rate": 1.9003317736084481e-06, + "loss": 0.3288, + "step": 7428 + }, + { + "epoch": 4.033116178067318, + "grad_norm": 11.935383011132533, + "learning_rate": 1.8982697336818633e-06, + "loss": 0.5207, + "step": 7429 + }, + { + "epoch": 4.033659066232356, + "grad_norm": 16.729428248551148, + "learning_rate": 1.8962086957901882e-06, + "loss": 0.495, + "step": 7430 + }, + { + "epoch": 4.034201954397394, + "grad_norm": 9.607193659544246, + "learning_rate": 1.894148660188342e-06, + "loss": 0.326, + "step": 7431 + }, + { + "epoch": 4.034744842562432, + "grad_norm": 12.36629523451521, + "learning_rate": 1.892089627131103e-06, + "loss": 0.5407, + "step": 7432 + }, + { + "epoch": 4.03528773072747, + "grad_norm": 12.984997271109028, + "learning_rate": 1.8900315968731465e-06, + "loss": 0.401, + "step": 7433 + }, + { + "epoch": 4.035830618892508, + "grad_norm": 10.733134132257513, + "learning_rate": 1.8879745696690022e-06, + "loss": 0.2782, + "step": 7434 + }, + { + "epoch": 4.036373507057546, + "grad_norm": 18.013649168300645, + "learning_rate": 1.8859185457730944e-06, + "loss": 0.8034, + "step": 7435 + }, + { + "epoch": 4.036916395222584, + "grad_norm": 13.544413626676313, + "learning_rate": 1.8838635254397154e-06, + "loss": 0.3762, + "step": 7436 + }, + { + "epoch": 4.037459283387622, + "grad_norm": 11.586743250886352, + "learning_rate": 1.8818095089230248e-06, + "loss": 0.5606, + "step": 7437 + }, + { + "epoch": 4.03800217155266, + "grad_norm": 10.323656236449189, + "learning_rate": 1.8797564964770787e-06, + "loss": 0.3772, + "step": 7438 + }, + { + "epoch": 4.038545059717698, + "grad_norm": 11.956263255604744, + "learning_rate": 1.877704488355785e-06, + "loss": 0.3292, + "step": 7439 + }, + { + "epoch": 4.039087947882736, + "grad_norm": 11.706565985563376, + "learning_rate": 1.8756534848129504e-06, + "loss": 0.3045, + "step": 7440 + }, + { + "epoch": 4.039630836047774, + "grad_norm": 17.55780625111388, + "learning_rate": 1.8736034861022368e-06, + "loss": 0.6189, + "step": 7441 + }, + { + "epoch": 4.040173724212812, + "grad_norm": 12.815942770561511, + "learning_rate": 1.8715544924771977e-06, + "loss": 0.3457, + "step": 7442 + }, + { + "epoch": 4.04071661237785, + "grad_norm": 15.230354926789687, + "learning_rate": 1.869506504191253e-06, + "loss": 0.3725, + "step": 7443 + }, + { + "epoch": 4.041259500542888, + "grad_norm": 12.610726529266895, + "learning_rate": 1.8674595214977031e-06, + "loss": 0.4368, + "step": 7444 + }, + { + "epoch": 4.041802388707926, + "grad_norm": 16.959731764835915, + "learning_rate": 1.8654135446497234e-06, + "loss": 0.7666, + "step": 7445 + }, + { + "epoch": 4.042345276872964, + "grad_norm": 13.908609561441018, + "learning_rate": 1.8633685739003548e-06, + "loss": 0.2762, + "step": 7446 + }, + { + "epoch": 4.042888165038002, + "grad_norm": 12.835858815757566, + "learning_rate": 1.8613246095025329e-06, + "loss": 0.4223, + "step": 7447 + }, + { + "epoch": 4.04343105320304, + "grad_norm": 15.055795233826663, + "learning_rate": 1.859281651709053e-06, + "loss": 0.4317, + "step": 7448 + }, + { + "epoch": 4.043973941368078, + "grad_norm": 13.581558077614412, + "learning_rate": 1.857239700772594e-06, + "loss": 0.4312, + "step": 7449 + }, + { + "epoch": 4.044516829533116, + "grad_norm": 13.841266542296328, + "learning_rate": 1.8551987569457053e-06, + "loss": 0.3242, + "step": 7450 + }, + { + "epoch": 4.045059717698154, + "grad_norm": 10.057661085861465, + "learning_rate": 1.853158820480816e-06, + "loss": 0.4803, + "step": 7451 + }, + { + "epoch": 4.045602605863192, + "grad_norm": 11.127274589006348, + "learning_rate": 1.8511198916302274e-06, + "loss": 0.3005, + "step": 7452 + }, + { + "epoch": 4.04614549402823, + "grad_norm": 10.414927056352687, + "learning_rate": 1.8490819706461193e-06, + "loss": 0.4651, + "step": 7453 + }, + { + "epoch": 4.046688382193268, + "grad_norm": 11.548866791691525, + "learning_rate": 1.847045057780542e-06, + "loss": 0.416, + "step": 7454 + }, + { + "epoch": 4.047231270358306, + "grad_norm": 8.995318003205785, + "learning_rate": 1.8450091532854264e-06, + "loss": 0.1938, + "step": 7455 + }, + { + "epoch": 4.047774158523344, + "grad_norm": 12.54663924123166, + "learning_rate": 1.8429742574125765e-06, + "loss": 0.4278, + "step": 7456 + }, + { + "epoch": 4.048317046688382, + "grad_norm": 9.475847861618263, + "learning_rate": 1.8409403704136707e-06, + "loss": 0.2686, + "step": 7457 + }, + { + "epoch": 4.04885993485342, + "grad_norm": 14.596699092222988, + "learning_rate": 1.8389074925402628e-06, + "loss": 0.6494, + "step": 7458 + }, + { + "epoch": 4.049402823018458, + "grad_norm": 8.923910821563455, + "learning_rate": 1.8368756240437836e-06, + "loss": 0.2667, + "step": 7459 + }, + { + "epoch": 4.049945711183496, + "grad_norm": 12.475431384264864, + "learning_rate": 1.8348447651755364e-06, + "loss": 0.2887, + "step": 7460 + }, + { + "epoch": 4.050488599348534, + "grad_norm": 10.730423522471147, + "learning_rate": 1.8328149161867025e-06, + "loss": 0.3498, + "step": 7461 + }, + { + "epoch": 4.051031487513572, + "grad_norm": 16.724674239679032, + "learning_rate": 1.830786077328337e-06, + "loss": 0.7322, + "step": 7462 + }, + { + "epoch": 4.0515743756786105, + "grad_norm": 12.182586543041394, + "learning_rate": 1.8287582488513656e-06, + "loss": 0.4257, + "step": 7463 + }, + { + "epoch": 4.052117263843648, + "grad_norm": 6.932472045654237, + "learning_rate": 1.826731431006603e-06, + "loss": 0.2157, + "step": 7464 + }, + { + "epoch": 4.052660152008686, + "grad_norm": 8.646135191685339, + "learning_rate": 1.8247056240447203e-06, + "loss": 0.2646, + "step": 7465 + }, + { + "epoch": 4.053203040173724, + "grad_norm": 16.802820609812404, + "learning_rate": 1.8226808282162755e-06, + "loss": 0.5392, + "step": 7466 + }, + { + "epoch": 4.0537459283387625, + "grad_norm": 13.75433513899018, + "learning_rate": 1.8206570437716986e-06, + "loss": 0.367, + "step": 7467 + }, + { + "epoch": 4.0542888165038, + "grad_norm": 12.642507240641134, + "learning_rate": 1.8186342709612925e-06, + "loss": 0.4884, + "step": 7468 + }, + { + "epoch": 4.054831704668838, + "grad_norm": 12.446058308802613, + "learning_rate": 1.8166125100352427e-06, + "loss": 0.3996, + "step": 7469 + }, + { + "epoch": 4.055374592833876, + "grad_norm": 9.780058065307577, + "learning_rate": 1.814591761243596e-06, + "loss": 0.2016, + "step": 7470 + }, + { + "epoch": 4.0559174809989145, + "grad_norm": 13.137545091734312, + "learning_rate": 1.812572024836291e-06, + "loss": 0.5646, + "step": 7471 + }, + { + "epoch": 4.056460369163952, + "grad_norm": 13.982451209627083, + "learning_rate": 1.8105533010631215e-06, + "loss": 0.3899, + "step": 7472 + }, + { + "epoch": 4.05700325732899, + "grad_norm": 14.830274354521338, + "learning_rate": 1.8085355901737767e-06, + "loss": 0.6667, + "step": 7473 + }, + { + "epoch": 4.057546145494028, + "grad_norm": 15.746342293850457, + "learning_rate": 1.8065188924178002e-06, + "loss": 0.6769, + "step": 7474 + }, + { + "epoch": 4.0580890336590665, + "grad_norm": 19.880439031191987, + "learning_rate": 1.8045032080446279e-06, + "loss": 0.6128, + "step": 7475 + }, + { + "epoch": 4.058631921824104, + "grad_norm": 12.174293630332373, + "learning_rate": 1.8024885373035628e-06, + "loss": 0.351, + "step": 7476 + }, + { + "epoch": 4.059174809989142, + "grad_norm": 13.340029936420176, + "learning_rate": 1.8004748804437755e-06, + "loss": 0.4044, + "step": 7477 + }, + { + "epoch": 4.05971769815418, + "grad_norm": 10.710849412967622, + "learning_rate": 1.798462237714329e-06, + "loss": 0.2545, + "step": 7478 + }, + { + "epoch": 4.0602605863192185, + "grad_norm": 14.145700312877072, + "learning_rate": 1.7964506093641388e-06, + "loss": 0.442, + "step": 7479 + }, + { + "epoch": 4.060803474484256, + "grad_norm": 11.782368982891633, + "learning_rate": 1.7944399956420133e-06, + "loss": 0.3699, + "step": 7480 + }, + { + "epoch": 4.061346362649294, + "grad_norm": 16.400383314329044, + "learning_rate": 1.7924303967966283e-06, + "loss": 0.4159, + "step": 7481 + }, + { + "epoch": 4.061889250814332, + "grad_norm": 14.158517117689607, + "learning_rate": 1.7904218130765316e-06, + "loss": 0.3477, + "step": 7482 + }, + { + "epoch": 4.0624321389793705, + "grad_norm": 10.538895009930492, + "learning_rate": 1.78841424473015e-06, + "loss": 0.4439, + "step": 7483 + }, + { + "epoch": 4.062975027144408, + "grad_norm": 15.749637206952377, + "learning_rate": 1.7864076920057815e-06, + "loss": 0.4063, + "step": 7484 + }, + { + "epoch": 4.063517915309446, + "grad_norm": 14.701502876295633, + "learning_rate": 1.7844021551516023e-06, + "loss": 0.293, + "step": 7485 + }, + { + "epoch": 4.064060803474484, + "grad_norm": 13.441957282892526, + "learning_rate": 1.7823976344156534e-06, + "loss": 0.3922, + "step": 7486 + }, + { + "epoch": 4.0646036916395225, + "grad_norm": 12.860794488151207, + "learning_rate": 1.780394130045865e-06, + "loss": 0.3637, + "step": 7487 + }, + { + "epoch": 4.06514657980456, + "grad_norm": 10.713125214412898, + "learning_rate": 1.7783916422900304e-06, + "loss": 0.3995, + "step": 7488 + }, + { + "epoch": 4.065689467969598, + "grad_norm": 12.736414280259371, + "learning_rate": 1.776390171395821e-06, + "loss": 0.4328, + "step": 7489 + }, + { + "epoch": 4.066232356134636, + "grad_norm": 13.781401823444826, + "learning_rate": 1.7743897176107817e-06, + "loss": 0.5039, + "step": 7490 + }, + { + "epoch": 4.0667752442996745, + "grad_norm": 12.032700692680026, + "learning_rate": 1.7723902811823312e-06, + "loss": 0.2701, + "step": 7491 + }, + { + "epoch": 4.067318132464712, + "grad_norm": 13.357720909095406, + "learning_rate": 1.7703918623577642e-06, + "loss": 0.472, + "step": 7492 + }, + { + "epoch": 4.06786102062975, + "grad_norm": 13.666181670657824, + "learning_rate": 1.7683944613842474e-06, + "loss": 0.4525, + "step": 7493 + }, + { + "epoch": 4.068403908794788, + "grad_norm": 15.204188285424376, + "learning_rate": 1.766398078508823e-06, + "loss": 0.5151, + "step": 7494 + }, + { + "epoch": 4.0689467969598265, + "grad_norm": 14.986125649693381, + "learning_rate": 1.7644027139784058e-06, + "loss": 0.5068, + "step": 7495 + }, + { + "epoch": 4.069489685124864, + "grad_norm": 17.414959041982844, + "learning_rate": 1.7624083680397874e-06, + "loss": 0.5285, + "step": 7496 + }, + { + "epoch": 4.070032573289902, + "grad_norm": 15.018593246149907, + "learning_rate": 1.7604150409396315e-06, + "loss": 0.4034, + "step": 7497 + }, + { + "epoch": 4.07057546145494, + "grad_norm": 7.740045825599994, + "learning_rate": 1.758422732924474e-06, + "loss": 0.2231, + "step": 7498 + }, + { + "epoch": 4.0711183496199785, + "grad_norm": 15.973649660315148, + "learning_rate": 1.7564314442407282e-06, + "loss": 0.4157, + "step": 7499 + }, + { + "epoch": 4.071661237785016, + "grad_norm": 15.522023726804239, + "learning_rate": 1.7544411751346802e-06, + "loss": 0.5029, + "step": 7500 + }, + { + "epoch": 4.072204125950054, + "grad_norm": 19.621921290245965, + "learning_rate": 1.7524519258524863e-06, + "loss": 0.4998, + "step": 7501 + }, + { + "epoch": 4.072747014115092, + "grad_norm": 13.460020491045686, + "learning_rate": 1.7504636966401877e-06, + "loss": 0.4035, + "step": 7502 + }, + { + "epoch": 4.0732899022801305, + "grad_norm": 12.082530806707663, + "learning_rate": 1.7484764877436822e-06, + "loss": 0.4754, + "step": 7503 + }, + { + "epoch": 4.073832790445168, + "grad_norm": 10.41118329664297, + "learning_rate": 1.7464902994087607e-06, + "loss": 0.331, + "step": 7504 + }, + { + "epoch": 4.074375678610206, + "grad_norm": 13.189232599075739, + "learning_rate": 1.7445051318810702e-06, + "loss": 0.4648, + "step": 7505 + }, + { + "epoch": 4.074918566775244, + "grad_norm": 9.349455348221264, + "learning_rate": 1.742520985406143e-06, + "loss": 0.2363, + "step": 7506 + }, + { + "epoch": 4.075461454940283, + "grad_norm": 11.755562149227849, + "learning_rate": 1.74053786022938e-06, + "loss": 0.2546, + "step": 7507 + }, + { + "epoch": 4.07600434310532, + "grad_norm": 13.819551647694206, + "learning_rate": 1.7385557565960564e-06, + "loss": 0.508, + "step": 7508 + }, + { + "epoch": 4.076547231270358, + "grad_norm": 9.181009774294514, + "learning_rate": 1.7365746747513278e-06, + "loss": 0.2214, + "step": 7509 + }, + { + "epoch": 4.077090119435396, + "grad_norm": 8.97042904079199, + "learning_rate": 1.7345946149402094e-06, + "loss": 0.2924, + "step": 7510 + }, + { + "epoch": 4.077633007600435, + "grad_norm": 12.257238316765228, + "learning_rate": 1.7326155774076058e-06, + "loss": 0.446, + "step": 7511 + }, + { + "epoch": 4.078175895765472, + "grad_norm": 18.213109072198556, + "learning_rate": 1.7306375623982796e-06, + "loss": 0.5576, + "step": 7512 + }, + { + "epoch": 4.07871878393051, + "grad_norm": 15.252433630945886, + "learning_rate": 1.7286605701568815e-06, + "loss": 0.5242, + "step": 7513 + }, + { + "epoch": 4.079261672095548, + "grad_norm": 12.724698295224274, + "learning_rate": 1.7266846009279292e-06, + "loss": 0.4852, + "step": 7514 + }, + { + "epoch": 4.079804560260587, + "grad_norm": 14.292147421433285, + "learning_rate": 1.7247096549558062e-06, + "loss": 0.3125, + "step": 7515 + }, + { + "epoch": 4.080347448425624, + "grad_norm": 11.689927807096288, + "learning_rate": 1.722735732484786e-06, + "loss": 0.3077, + "step": 7516 + }, + { + "epoch": 4.080890336590662, + "grad_norm": 17.211462481334554, + "learning_rate": 1.7207628337589988e-06, + "loss": 0.6293, + "step": 7517 + }, + { + "epoch": 4.0814332247557, + "grad_norm": 11.662256116349747, + "learning_rate": 1.7187909590224604e-06, + "loss": 0.3055, + "step": 7518 + }, + { + "epoch": 4.081976112920739, + "grad_norm": 10.371623662202015, + "learning_rate": 1.7168201085190562e-06, + "loss": 0.3695, + "step": 7519 + }, + { + "epoch": 4.082519001085776, + "grad_norm": 14.563522671402897, + "learning_rate": 1.7148502824925418e-06, + "loss": 0.3588, + "step": 7520 + }, + { + "epoch": 4.083061889250814, + "grad_norm": 12.36778161398001, + "learning_rate": 1.712881481186548e-06, + "loss": 0.3054, + "step": 7521 + }, + { + "epoch": 4.083604777415852, + "grad_norm": 13.99470872596806, + "learning_rate": 1.7109137048445812e-06, + "loss": 0.4663, + "step": 7522 + }, + { + "epoch": 4.084147665580891, + "grad_norm": 11.566073320833054, + "learning_rate": 1.7089469537100178e-06, + "loss": 0.4081, + "step": 7523 + }, + { + "epoch": 4.084690553745928, + "grad_norm": 10.823284667656079, + "learning_rate": 1.70698122802611e-06, + "loss": 0.3027, + "step": 7524 + }, + { + "epoch": 4.085233441910966, + "grad_norm": 14.505207806467645, + "learning_rate": 1.705016528035981e-06, + "loss": 0.7583, + "step": 7525 + }, + { + "epoch": 4.085776330076004, + "grad_norm": 13.977656869068836, + "learning_rate": 1.7030528539826275e-06, + "loss": 0.4479, + "step": 7526 + }, + { + "epoch": 4.086319218241043, + "grad_norm": 14.88627029394164, + "learning_rate": 1.7010902061089197e-06, + "loss": 0.5773, + "step": 7527 + }, + { + "epoch": 4.08686210640608, + "grad_norm": 9.998415444955523, + "learning_rate": 1.6991285846576022e-06, + "loss": 0.2414, + "step": 7528 + }, + { + "epoch": 4.087404994571118, + "grad_norm": 11.634301676106471, + "learning_rate": 1.6971679898712912e-06, + "loss": 0.3949, + "step": 7529 + }, + { + "epoch": 4.087947882736156, + "grad_norm": 12.469794055243895, + "learning_rate": 1.6952084219924757e-06, + "loss": 0.4851, + "step": 7530 + }, + { + "epoch": 4.088490770901195, + "grad_norm": 11.162959342238105, + "learning_rate": 1.6932498812635189e-06, + "loss": 0.4459, + "step": 7531 + }, + { + "epoch": 4.089033659066232, + "grad_norm": 17.59176356543345, + "learning_rate": 1.6912923679266557e-06, + "loss": 0.6036, + "step": 7532 + }, + { + "epoch": 4.08957654723127, + "grad_norm": 10.674032941701148, + "learning_rate": 1.6893358822239947e-06, + "loss": 0.3718, + "step": 7533 + }, + { + "epoch": 4.090119435396308, + "grad_norm": 9.119363490702213, + "learning_rate": 1.6873804243975167e-06, + "loss": 0.2871, + "step": 7534 + }, + { + "epoch": 4.090662323561347, + "grad_norm": 13.955880230291728, + "learning_rate": 1.6854259946890762e-06, + "loss": 0.384, + "step": 7535 + }, + { + "epoch": 4.091205211726384, + "grad_norm": 15.830781520521217, + "learning_rate": 1.6834725933403995e-06, + "loss": 0.59, + "step": 7536 + }, + { + "epoch": 4.091748099891422, + "grad_norm": 11.306287185705354, + "learning_rate": 1.681520220593088e-06, + "loss": 0.3796, + "step": 7537 + }, + { + "epoch": 4.09229098805646, + "grad_norm": 14.796920699824259, + "learning_rate": 1.6795688766886132e-06, + "loss": 0.4181, + "step": 7538 + }, + { + "epoch": 4.092833876221499, + "grad_norm": 15.602066419097026, + "learning_rate": 1.6776185618683171e-06, + "loss": 0.5769, + "step": 7539 + }, + { + "epoch": 4.093376764386536, + "grad_norm": 10.75921703150943, + "learning_rate": 1.6756692763734272e-06, + "loss": 0.3376, + "step": 7540 + }, + { + "epoch": 4.093919652551574, + "grad_norm": 12.134257694162375, + "learning_rate": 1.6737210204450226e-06, + "loss": 0.3719, + "step": 7541 + }, + { + "epoch": 4.094462540716612, + "grad_norm": 14.106939136679763, + "learning_rate": 1.6717737943240774e-06, + "loss": 0.6107, + "step": 7542 + }, + { + "epoch": 4.095005428881651, + "grad_norm": 12.295261891318004, + "learning_rate": 1.6698275982514178e-06, + "loss": 0.35, + "step": 7543 + }, + { + "epoch": 4.095548317046688, + "grad_norm": 11.519476162199268, + "learning_rate": 1.6678824324677623e-06, + "loss": 0.3783, + "step": 7544 + }, + { + "epoch": 4.096091205211726, + "grad_norm": 8.77456108175982, + "learning_rate": 1.6659382972136839e-06, + "loss": 0.2232, + "step": 7545 + }, + { + "epoch": 4.096634093376764, + "grad_norm": 11.479634703602272, + "learning_rate": 1.6639951927296371e-06, + "loss": 0.3152, + "step": 7546 + }, + { + "epoch": 4.097176981541803, + "grad_norm": 11.756787071755726, + "learning_rate": 1.6620531192559552e-06, + "loss": 0.4063, + "step": 7547 + }, + { + "epoch": 4.09771986970684, + "grad_norm": 16.583849337588653, + "learning_rate": 1.6601120770328283e-06, + "loss": 0.5936, + "step": 7548 + }, + { + "epoch": 4.098262757871878, + "grad_norm": 12.762971955755509, + "learning_rate": 1.6581720663003354e-06, + "loss": 0.4476, + "step": 7549 + }, + { + "epoch": 4.098805646036916, + "grad_norm": 18.75880484333442, + "learning_rate": 1.6562330872984122e-06, + "loss": 0.6011, + "step": 7550 + }, + { + "epoch": 4.099348534201955, + "grad_norm": 18.29186988565711, + "learning_rate": 1.6542951402668805e-06, + "loss": 0.5015, + "step": 7551 + }, + { + "epoch": 4.099891422366992, + "grad_norm": 12.057222479862308, + "learning_rate": 1.6523582254454273e-06, + "loss": 0.6185, + "step": 7552 + }, + { + "epoch": 4.10043431053203, + "grad_norm": 15.90807833403242, + "learning_rate": 1.6504223430736122e-06, + "loss": 0.5644, + "step": 7553 + }, + { + "epoch": 4.100977198697068, + "grad_norm": 15.27338292218376, + "learning_rate": 1.6484874933908723e-06, + "loss": 0.4299, + "step": 7554 + }, + { + "epoch": 4.101520086862107, + "grad_norm": 12.827779978027912, + "learning_rate": 1.6465536766365043e-06, + "loss": 0.4768, + "step": 7555 + }, + { + "epoch": 4.1020629750271445, + "grad_norm": 19.22780764109623, + "learning_rate": 1.6446208930496954e-06, + "loss": 0.3838, + "step": 7556 + }, + { + "epoch": 4.102605863192182, + "grad_norm": 14.630988970729627, + "learning_rate": 1.642689142869487e-06, + "loss": 0.3607, + "step": 7557 + }, + { + "epoch": 4.10314875135722, + "grad_norm": 11.07230018602903, + "learning_rate": 1.6407584263348065e-06, + "loss": 0.2996, + "step": 7558 + }, + { + "epoch": 4.103691639522259, + "grad_norm": 12.881805511729306, + "learning_rate": 1.6388287436844474e-06, + "loss": 0.4539, + "step": 7559 + }, + { + "epoch": 4.1042345276872965, + "grad_norm": 12.06776906866114, + "learning_rate": 1.6369000951570746e-06, + "loss": 0.534, + "step": 7560 + }, + { + "epoch": 4.104777415852334, + "grad_norm": 16.25382663633683, + "learning_rate": 1.634972480991226e-06, + "loss": 0.5308, + "step": 7561 + }, + { + "epoch": 4.105320304017372, + "grad_norm": 10.586138014834138, + "learning_rate": 1.6330459014253141e-06, + "loss": 0.3352, + "step": 7562 + }, + { + "epoch": 4.105863192182411, + "grad_norm": 15.890912108947397, + "learning_rate": 1.6311203566976196e-06, + "loss": 0.5464, + "step": 7563 + }, + { + "epoch": 4.1064060803474485, + "grad_norm": 12.756769761348481, + "learning_rate": 1.6291958470462966e-06, + "loss": 0.3651, + "step": 7564 + }, + { + "epoch": 4.106948968512486, + "grad_norm": 10.76450128852776, + "learning_rate": 1.6272723727093732e-06, + "loss": 0.3137, + "step": 7565 + }, + { + "epoch": 4.107491856677524, + "grad_norm": 13.308264213836948, + "learning_rate": 1.6253499339247468e-06, + "loss": 0.5733, + "step": 7566 + }, + { + "epoch": 4.108034744842563, + "grad_norm": 13.76589828286844, + "learning_rate": 1.6234285309301878e-06, + "loss": 0.473, + "step": 7567 + }, + { + "epoch": 4.1085776330076005, + "grad_norm": 14.628341432791526, + "learning_rate": 1.621508163963338e-06, + "loss": 0.2876, + "step": 7568 + }, + { + "epoch": 4.109120521172638, + "grad_norm": 11.72578228984837, + "learning_rate": 1.6195888332617126e-06, + "loss": 0.4284, + "step": 7569 + }, + { + "epoch": 4.109663409337676, + "grad_norm": 20.77607641436691, + "learning_rate": 1.6176705390626956e-06, + "loss": 0.7359, + "step": 7570 + }, + { + "epoch": 4.110206297502715, + "grad_norm": 15.592080343031311, + "learning_rate": 1.615753281603546e-06, + "loss": 0.5176, + "step": 7571 + }, + { + "epoch": 4.1107491856677525, + "grad_norm": 14.765460153242566, + "learning_rate": 1.6138370611213917e-06, + "loss": 0.3779, + "step": 7572 + }, + { + "epoch": 4.11129207383279, + "grad_norm": 11.227048300024913, + "learning_rate": 1.6119218778532386e-06, + "loss": 0.2835, + "step": 7573 + }, + { + "epoch": 4.111834961997828, + "grad_norm": 10.73327552489785, + "learning_rate": 1.6100077320359552e-06, + "loss": 0.3355, + "step": 7574 + }, + { + "epoch": 4.112377850162867, + "grad_norm": 12.220262589971611, + "learning_rate": 1.6080946239062856e-06, + "loss": 0.2647, + "step": 7575 + }, + { + "epoch": 4.1129207383279045, + "grad_norm": 13.578576106400867, + "learning_rate": 1.6061825537008491e-06, + "loss": 0.4504, + "step": 7576 + }, + { + "epoch": 4.113463626492942, + "grad_norm": 12.666206872990253, + "learning_rate": 1.6042715216561312e-06, + "loss": 0.4362, + "step": 7577 + }, + { + "epoch": 4.11400651465798, + "grad_norm": 14.81928570898845, + "learning_rate": 1.6023615280084925e-06, + "loss": 0.5685, + "step": 7578 + }, + { + "epoch": 4.114549402823019, + "grad_norm": 13.058193252197759, + "learning_rate": 1.6004525729941622e-06, + "loss": 0.5339, + "step": 7579 + }, + { + "epoch": 4.1150922909880565, + "grad_norm": 16.238471411046245, + "learning_rate": 1.598544656849248e-06, + "loss": 0.6546, + "step": 7580 + }, + { + "epoch": 4.115635179153094, + "grad_norm": 11.090517158246884, + "learning_rate": 1.5966377798097165e-06, + "loss": 0.4045, + "step": 7581 + }, + { + "epoch": 4.116178067318132, + "grad_norm": 15.25648505046589, + "learning_rate": 1.5947319421114227e-06, + "loss": 0.5787, + "step": 7582 + }, + { + "epoch": 4.116720955483171, + "grad_norm": 11.253131068704839, + "learning_rate": 1.5928271439900755e-06, + "loss": 0.3468, + "step": 7583 + }, + { + "epoch": 4.1172638436482085, + "grad_norm": 22.87822267893679, + "learning_rate": 1.5909233856812634e-06, + "loss": 0.5457, + "step": 7584 + }, + { + "epoch": 4.117806731813246, + "grad_norm": 11.373191676935885, + "learning_rate": 1.5890206674204535e-06, + "loss": 0.4627, + "step": 7585 + }, + { + "epoch": 4.118349619978284, + "grad_norm": 14.408977573167448, + "learning_rate": 1.5871189894429683e-06, + "loss": 0.3613, + "step": 7586 + }, + { + "epoch": 4.118892508143323, + "grad_norm": 13.14025988474886, + "learning_rate": 1.5852183519840203e-06, + "loss": 0.5449, + "step": 7587 + }, + { + "epoch": 4.1194353963083605, + "grad_norm": 13.859357618288538, + "learning_rate": 1.583318755278671e-06, + "loss": 0.5978, + "step": 7588 + }, + { + "epoch": 4.119978284473398, + "grad_norm": 12.173600067172606, + "learning_rate": 1.581420199561876e-06, + "loss": 0.4165, + "step": 7589 + }, + { + "epoch": 4.120521172638436, + "grad_norm": 11.35642341805059, + "learning_rate": 1.5795226850684476e-06, + "loss": 0.4934, + "step": 7590 + }, + { + "epoch": 4.121064060803475, + "grad_norm": 15.267752687607832, + "learning_rate": 1.5776262120330743e-06, + "loss": 0.4089, + "step": 7591 + }, + { + "epoch": 4.1216069489685125, + "grad_norm": 14.238479740595348, + "learning_rate": 1.5757307806903133e-06, + "loss": 0.4299, + "step": 7592 + }, + { + "epoch": 4.12214983713355, + "grad_norm": 11.160357781980832, + "learning_rate": 1.5738363912745959e-06, + "loss": 0.4948, + "step": 7593 + }, + { + "epoch": 4.122692725298588, + "grad_norm": 16.47054898460848, + "learning_rate": 1.5719430440202255e-06, + "loss": 0.5639, + "step": 7594 + }, + { + "epoch": 4.123235613463627, + "grad_norm": 15.099982308232326, + "learning_rate": 1.5700507391613673e-06, + "loss": 0.3724, + "step": 7595 + }, + { + "epoch": 4.1237785016286646, + "grad_norm": 13.904934094160584, + "learning_rate": 1.5681594769320706e-06, + "loss": 0.5423, + "step": 7596 + }, + { + "epoch": 4.124321389793702, + "grad_norm": 10.17978139897218, + "learning_rate": 1.5662692575662486e-06, + "loss": 0.3509, + "step": 7597 + }, + { + "epoch": 4.12486427795874, + "grad_norm": 10.662732400077363, + "learning_rate": 1.5643800812976873e-06, + "loss": 0.306, + "step": 7598 + }, + { + "epoch": 4.125407166123779, + "grad_norm": 10.833314489423646, + "learning_rate": 1.56249194836004e-06, + "loss": 0.3767, + "step": 7599 + }, + { + "epoch": 4.125950054288817, + "grad_norm": 14.593277788934097, + "learning_rate": 1.5606048589868372e-06, + "loss": 0.3886, + "step": 7600 + }, + { + "epoch": 4.126492942453854, + "grad_norm": 19.995368160150967, + "learning_rate": 1.558718813411475e-06, + "loss": 0.3951, + "step": 7601 + }, + { + "epoch": 4.127035830618892, + "grad_norm": 8.750505077849805, + "learning_rate": 1.5568338118672243e-06, + "loss": 0.2172, + "step": 7602 + }, + { + "epoch": 4.127578718783931, + "grad_norm": 13.314482415981553, + "learning_rate": 1.5549498545872233e-06, + "loss": 0.2903, + "step": 7603 + }, + { + "epoch": 4.128121606948969, + "grad_norm": 10.865437650317455, + "learning_rate": 1.5530669418044842e-06, + "loss": 0.3271, + "step": 7604 + }, + { + "epoch": 4.128664495114006, + "grad_norm": 13.939733125814856, + "learning_rate": 1.5511850737518875e-06, + "loss": 0.3687, + "step": 7605 + }, + { + "epoch": 4.129207383279044, + "grad_norm": 13.380151193528324, + "learning_rate": 1.549304250662187e-06, + "loss": 0.3158, + "step": 7606 + }, + { + "epoch": 4.129750271444083, + "grad_norm": 9.920776483442411, + "learning_rate": 1.5474244727680055e-06, + "loss": 0.3943, + "step": 7607 + }, + { + "epoch": 4.130293159609121, + "grad_norm": 11.896762787826454, + "learning_rate": 1.545545740301836e-06, + "loss": 0.5542, + "step": 7608 + }, + { + "epoch": 4.130836047774158, + "grad_norm": 12.91558903974088, + "learning_rate": 1.5436680534960434e-06, + "loss": 0.4152, + "step": 7609 + }, + { + "epoch": 4.131378935939196, + "grad_norm": 10.961854763205224, + "learning_rate": 1.5417914125828616e-06, + "loss": 0.4836, + "step": 7610 + }, + { + "epoch": 4.131921824104235, + "grad_norm": 10.548604561082739, + "learning_rate": 1.5399158177944019e-06, + "loss": 0.4001, + "step": 7611 + }, + { + "epoch": 4.132464712269273, + "grad_norm": 15.534076149917789, + "learning_rate": 1.5380412693626356e-06, + "loss": 0.5564, + "step": 7612 + }, + { + "epoch": 4.13300760043431, + "grad_norm": 12.354229348605834, + "learning_rate": 1.536167767519412e-06, + "loss": 0.5188, + "step": 7613 + }, + { + "epoch": 4.133550488599348, + "grad_norm": 15.550407127657197, + "learning_rate": 1.5342953124964477e-06, + "loss": 0.7004, + "step": 7614 + }, + { + "epoch": 4.134093376764387, + "grad_norm": 13.960316578712156, + "learning_rate": 1.532423904525332e-06, + "loss": 0.3999, + "step": 7615 + }, + { + "epoch": 4.134636264929425, + "grad_norm": 12.1140917102908, + "learning_rate": 1.5305535438375218e-06, + "loss": 0.3274, + "step": 7616 + }, + { + "epoch": 4.135179153094462, + "grad_norm": 12.38487299229727, + "learning_rate": 1.5286842306643468e-06, + "loss": 0.4094, + "step": 7617 + }, + { + "epoch": 4.1357220412595, + "grad_norm": 10.14740725625142, + "learning_rate": 1.5268159652370108e-06, + "loss": 0.4163, + "step": 7618 + }, + { + "epoch": 4.136264929424539, + "grad_norm": 14.207848794460347, + "learning_rate": 1.5249487477865765e-06, + "loss": 0.3613, + "step": 7619 + }, + { + "epoch": 4.136807817589577, + "grad_norm": 17.170203385892172, + "learning_rate": 1.523082578543993e-06, + "loss": 0.4012, + "step": 7620 + }, + { + "epoch": 4.137350705754614, + "grad_norm": 11.31582486367915, + "learning_rate": 1.5212174577400618e-06, + "loss": 0.3245, + "step": 7621 + }, + { + "epoch": 4.137893593919652, + "grad_norm": 15.179364893086714, + "learning_rate": 1.519353385605471e-06, + "loss": 0.418, + "step": 7622 + }, + { + "epoch": 4.138436482084691, + "grad_norm": 16.498670863026067, + "learning_rate": 1.5174903623707716e-06, + "loss": 0.4128, + "step": 7623 + }, + { + "epoch": 4.138979370249729, + "grad_norm": 19.148612028425138, + "learning_rate": 1.515628388266378e-06, + "loss": 0.6773, + "step": 7624 + }, + { + "epoch": 4.139522258414766, + "grad_norm": 21.1656943566629, + "learning_rate": 1.513767463522593e-06, + "loss": 0.6674, + "step": 7625 + }, + { + "epoch": 4.140065146579804, + "grad_norm": 13.460184953663427, + "learning_rate": 1.5119075883695678e-06, + "loss": 0.4615, + "step": 7626 + }, + { + "epoch": 4.140608034744843, + "grad_norm": 13.260423844727587, + "learning_rate": 1.510048763037345e-06, + "loss": 0.3272, + "step": 7627 + }, + { + "epoch": 4.141150922909881, + "grad_norm": 11.331130692905933, + "learning_rate": 1.5081909877558177e-06, + "loss": 0.2903, + "step": 7628 + }, + { + "epoch": 4.141693811074918, + "grad_norm": 15.865149055955152, + "learning_rate": 1.506334262754765e-06, + "loss": 0.4817, + "step": 7629 + }, + { + "epoch": 4.142236699239956, + "grad_norm": 11.349271790017943, + "learning_rate": 1.5044785882638279e-06, + "loss": 0.3778, + "step": 7630 + }, + { + "epoch": 4.142779587404995, + "grad_norm": 12.519516513475779, + "learning_rate": 1.5026239645125196e-06, + "loss": 0.2995, + "step": 7631 + }, + { + "epoch": 4.143322475570033, + "grad_norm": 13.209913219504104, + "learning_rate": 1.500770391730223e-06, + "loss": 0.4709, + "step": 7632 + }, + { + "epoch": 4.14386536373507, + "grad_norm": 9.929305640423868, + "learning_rate": 1.498917870146186e-06, + "loss": 0.269, + "step": 7633 + }, + { + "epoch": 4.144408251900108, + "grad_norm": 17.43652576717896, + "learning_rate": 1.4970663999895384e-06, + "loss": 0.7616, + "step": 7634 + }, + { + "epoch": 4.144951140065147, + "grad_norm": 10.810139421001562, + "learning_rate": 1.4952159814892697e-06, + "loss": 0.3408, + "step": 7635 + }, + { + "epoch": 4.145494028230185, + "grad_norm": 17.00710504331134, + "learning_rate": 1.4933666148742432e-06, + "loss": 0.6018, + "step": 7636 + }, + { + "epoch": 4.146036916395222, + "grad_norm": 8.623042522508, + "learning_rate": 1.4915183003731904e-06, + "loss": 0.2758, + "step": 7637 + }, + { + "epoch": 4.14657980456026, + "grad_norm": 16.04766738068284, + "learning_rate": 1.489671038214715e-06, + "loss": 0.6531, + "step": 7638 + }, + { + "epoch": 4.147122692725299, + "grad_norm": 11.752307004240341, + "learning_rate": 1.48782482862729e-06, + "loss": 0.3995, + "step": 7639 + }, + { + "epoch": 4.147665580890337, + "grad_norm": 12.146925988953807, + "learning_rate": 1.485979671839256e-06, + "loss": 0.4834, + "step": 7640 + }, + { + "epoch": 4.148208469055374, + "grad_norm": 17.18792212776091, + "learning_rate": 1.4841355680788261e-06, + "loss": 0.5534, + "step": 7641 + }, + { + "epoch": 4.148751357220412, + "grad_norm": 12.648802834009173, + "learning_rate": 1.4822925175740798e-06, + "loss": 0.5194, + "step": 7642 + }, + { + "epoch": 4.149294245385451, + "grad_norm": 12.067595847261051, + "learning_rate": 1.4804505205529718e-06, + "loss": 0.3579, + "step": 7643 + }, + { + "epoch": 4.149837133550489, + "grad_norm": 10.279884523990727, + "learning_rate": 1.4786095772433206e-06, + "loss": 0.3087, + "step": 7644 + }, + { + "epoch": 4.1503800217155264, + "grad_norm": 12.341347959843054, + "learning_rate": 1.4767696878728188e-06, + "loss": 0.314, + "step": 7645 + }, + { + "epoch": 4.150922909880564, + "grad_norm": 9.540652635109405, + "learning_rate": 1.4749308526690253e-06, + "loss": 0.3572, + "step": 7646 + }, + { + "epoch": 4.151465798045603, + "grad_norm": 12.941719739828546, + "learning_rate": 1.4730930718593705e-06, + "loss": 0.3057, + "step": 7647 + }, + { + "epoch": 4.152008686210641, + "grad_norm": 10.404139051977397, + "learning_rate": 1.471256345671156e-06, + "loss": 0.3796, + "step": 7648 + }, + { + "epoch": 4.1525515743756785, + "grad_norm": 11.552988822806185, + "learning_rate": 1.4694206743315476e-06, + "loss": 0.3922, + "step": 7649 + }, + { + "epoch": 4.153094462540716, + "grad_norm": 10.698397299218179, + "learning_rate": 1.467586058067585e-06, + "loss": 0.3779, + "step": 7650 + }, + { + "epoch": 4.153637350705755, + "grad_norm": 14.80756632794264, + "learning_rate": 1.465752497106181e-06, + "loss": 0.2783, + "step": 7651 + }, + { + "epoch": 4.154180238870793, + "grad_norm": 14.09754720490473, + "learning_rate": 1.4639199916741087e-06, + "loss": 0.315, + "step": 7652 + }, + { + "epoch": 4.1547231270358305, + "grad_norm": 8.268573012134576, + "learning_rate": 1.4620885419980147e-06, + "loss": 0.321, + "step": 7653 + }, + { + "epoch": 4.155266015200868, + "grad_norm": 18.11398789359469, + "learning_rate": 1.4602581483044188e-06, + "loss": 0.5637, + "step": 7654 + }, + { + "epoch": 4.155808903365907, + "grad_norm": 13.40679373737216, + "learning_rate": 1.458428810819702e-06, + "loss": 0.3844, + "step": 7655 + }, + { + "epoch": 4.156351791530945, + "grad_norm": 19.530436228333638, + "learning_rate": 1.456600529770128e-06, + "loss": 0.7487, + "step": 7656 + }, + { + "epoch": 4.1568946796959825, + "grad_norm": 15.655440408772145, + "learning_rate": 1.4547733053818125e-06, + "loss": 0.5907, + "step": 7657 + }, + { + "epoch": 4.15743756786102, + "grad_norm": 11.05518654892871, + "learning_rate": 1.452947137880757e-06, + "loss": 0.3539, + "step": 7658 + }, + { + "epoch": 4.157980456026059, + "grad_norm": 13.535214104936339, + "learning_rate": 1.4511220274928183e-06, + "loss": 0.4481, + "step": 7659 + }, + { + "epoch": 4.158523344191097, + "grad_norm": 15.227632590619127, + "learning_rate": 1.4492979744437342e-06, + "loss": 0.5095, + "step": 7660 + }, + { + "epoch": 4.1590662323561345, + "grad_norm": 12.117485924978645, + "learning_rate": 1.4474749789591058e-06, + "loss": 0.3035, + "step": 7661 + }, + { + "epoch": 4.159609120521172, + "grad_norm": 15.47621703654388, + "learning_rate": 1.4456530412643988e-06, + "loss": 0.7481, + "step": 7662 + }, + { + "epoch": 4.160152008686211, + "grad_norm": 16.71458160401543, + "learning_rate": 1.4438321615849615e-06, + "loss": 0.7957, + "step": 7663 + }, + { + "epoch": 4.160694896851249, + "grad_norm": 16.66028628014589, + "learning_rate": 1.442012340145995e-06, + "loss": 0.4658, + "step": 7664 + }, + { + "epoch": 4.1612377850162865, + "grad_norm": 14.178694096135883, + "learning_rate": 1.440193577172585e-06, + "loss": 0.5093, + "step": 7665 + }, + { + "epoch": 4.161780673181324, + "grad_norm": 13.385243486849467, + "learning_rate": 1.4383758728896724e-06, + "loss": 0.448, + "step": 7666 + }, + { + "epoch": 4.162323561346363, + "grad_norm": 12.920254235657909, + "learning_rate": 1.4365592275220787e-06, + "loss": 0.3069, + "step": 7667 + }, + { + "epoch": 4.162866449511401, + "grad_norm": 17.879050333979382, + "learning_rate": 1.4347436412944882e-06, + "loss": 0.6627, + "step": 7668 + }, + { + "epoch": 4.1634093376764385, + "grad_norm": 13.04178990878481, + "learning_rate": 1.4329291144314549e-06, + "loss": 0.3993, + "step": 7669 + }, + { + "epoch": 4.163952225841476, + "grad_norm": 13.341317379498156, + "learning_rate": 1.4311156471574027e-06, + "loss": 0.2989, + "step": 7670 + }, + { + "epoch": 4.164495114006515, + "grad_norm": 15.262211702534875, + "learning_rate": 1.4293032396966234e-06, + "loss": 0.6915, + "step": 7671 + }, + { + "epoch": 4.165038002171553, + "grad_norm": 10.866372137845064, + "learning_rate": 1.4274918922732805e-06, + "loss": 0.3446, + "step": 7672 + }, + { + "epoch": 4.1655808903365905, + "grad_norm": 9.213769718587553, + "learning_rate": 1.425681605111402e-06, + "loss": 0.2418, + "step": 7673 + }, + { + "epoch": 4.166123778501628, + "grad_norm": 14.937646538347469, + "learning_rate": 1.4238723784348873e-06, + "loss": 0.4886, + "step": 7674 + }, + { + "epoch": 4.166666666666667, + "grad_norm": 14.913733879373519, + "learning_rate": 1.4220642124675066e-06, + "loss": 0.549, + "step": 7675 + }, + { + "epoch": 4.167209554831705, + "grad_norm": 13.717360959964097, + "learning_rate": 1.4202571074328941e-06, + "loss": 0.5719, + "step": 7676 + }, + { + "epoch": 4.1677524429967425, + "grad_norm": 16.071560326957982, + "learning_rate": 1.4184510635545568e-06, + "loss": 0.6427, + "step": 7677 + }, + { + "epoch": 4.16829533116178, + "grad_norm": 12.694831131159624, + "learning_rate": 1.4166460810558691e-06, + "loss": 0.502, + "step": 7678 + }, + { + "epoch": 4.168838219326819, + "grad_norm": 9.505941995564902, + "learning_rate": 1.4148421601600749e-06, + "loss": 0.357, + "step": 7679 + }, + { + "epoch": 4.169381107491857, + "grad_norm": 10.441603292556694, + "learning_rate": 1.4130393010902831e-06, + "loss": 0.3607, + "step": 7680 + }, + { + "epoch": 4.1699239956568945, + "grad_norm": 11.694087750701021, + "learning_rate": 1.4112375040694781e-06, + "loss": 0.4678, + "step": 7681 + }, + { + "epoch": 4.170466883821932, + "grad_norm": 14.852563594982858, + "learning_rate": 1.4094367693205057e-06, + "loss": 0.5183, + "step": 7682 + }, + { + "epoch": 4.171009771986971, + "grad_norm": 13.451301475700863, + "learning_rate": 1.4076370970660858e-06, + "loss": 0.402, + "step": 7683 + }, + { + "epoch": 4.171552660152009, + "grad_norm": 12.466975166766947, + "learning_rate": 1.4058384875288034e-06, + "loss": 0.5729, + "step": 7684 + }, + { + "epoch": 4.1720955483170465, + "grad_norm": 14.018971166921462, + "learning_rate": 1.404040940931114e-06, + "loss": 0.3046, + "step": 7685 + }, + { + "epoch": 4.172638436482084, + "grad_norm": 9.441888668504816, + "learning_rate": 1.402244457495341e-06, + "loss": 0.2794, + "step": 7686 + }, + { + "epoch": 4.173181324647123, + "grad_norm": 14.181424515746851, + "learning_rate": 1.4004490374436764e-06, + "loss": 0.584, + "step": 7687 + }, + { + "epoch": 4.173724212812161, + "grad_norm": 10.806677133942568, + "learning_rate": 1.3986546809981783e-06, + "loss": 0.4671, + "step": 7688 + }, + { + "epoch": 4.1742671009771986, + "grad_norm": 13.238385393014816, + "learning_rate": 1.3968613883807814e-06, + "loss": 0.8039, + "step": 7689 + }, + { + "epoch": 4.174809989142236, + "grad_norm": 10.501288736320749, + "learning_rate": 1.395069159813276e-06, + "loss": 0.2945, + "step": 7690 + }, + { + "epoch": 4.175352877307275, + "grad_norm": 15.172510573487276, + "learning_rate": 1.3932779955173349e-06, + "loss": 0.4151, + "step": 7691 + }, + { + "epoch": 4.175895765472313, + "grad_norm": 15.13893230910218, + "learning_rate": 1.3914878957144862e-06, + "loss": 0.6144, + "step": 7692 + }, + { + "epoch": 4.176438653637351, + "grad_norm": 9.703023506384875, + "learning_rate": 1.389698860626133e-06, + "loss": 0.2173, + "step": 7693 + }, + { + "epoch": 4.176981541802388, + "grad_norm": 9.6655026395734, + "learning_rate": 1.3879108904735505e-06, + "loss": 0.3047, + "step": 7694 + }, + { + "epoch": 4.177524429967427, + "grad_norm": 13.323140178847439, + "learning_rate": 1.3861239854778719e-06, + "loss": 0.4717, + "step": 7695 + }, + { + "epoch": 4.178067318132465, + "grad_norm": 9.405943562091032, + "learning_rate": 1.38433814586011e-06, + "loss": 0.2174, + "step": 7696 + }, + { + "epoch": 4.178610206297503, + "grad_norm": 13.771494832808791, + "learning_rate": 1.3825533718411343e-06, + "loss": 0.3924, + "step": 7697 + }, + { + "epoch": 4.17915309446254, + "grad_norm": 15.26443665837414, + "learning_rate": 1.3807696636416956e-06, + "loss": 0.4529, + "step": 7698 + }, + { + "epoch": 4.179695982627579, + "grad_norm": 19.71010753607603, + "learning_rate": 1.3789870214823987e-06, + "loss": 1.155, + "step": 7699 + }, + { + "epoch": 4.180238870792617, + "grad_norm": 10.482042942898147, + "learning_rate": 1.377205445583728e-06, + "loss": 0.2945, + "step": 7700 + }, + { + "epoch": 4.180781758957655, + "grad_norm": 23.55094181120616, + "learning_rate": 1.375424936166032e-06, + "loss": 0.5429, + "step": 7701 + }, + { + "epoch": 4.181324647122692, + "grad_norm": 12.840338456448858, + "learning_rate": 1.3736454934495225e-06, + "loss": 0.5054, + "step": 7702 + }, + { + "epoch": 4.181867535287731, + "grad_norm": 15.318914270528728, + "learning_rate": 1.3718671176542897e-06, + "loss": 0.5164, + "step": 7703 + }, + { + "epoch": 4.182410423452769, + "grad_norm": 12.76929564122516, + "learning_rate": 1.3700898090002802e-06, + "loss": 0.5994, + "step": 7704 + }, + { + "epoch": 4.182953311617807, + "grad_norm": 10.099666037602478, + "learning_rate": 1.368313567707318e-06, + "loss": 0.2488, + "step": 7705 + }, + { + "epoch": 4.183496199782844, + "grad_norm": 14.142637199722767, + "learning_rate": 1.3665383939950916e-06, + "loss": 0.5153, + "step": 7706 + }, + { + "epoch": 4.184039087947883, + "grad_norm": 17.926261733912252, + "learning_rate": 1.3647642880831556e-06, + "loss": 0.647, + "step": 7707 + }, + { + "epoch": 4.184581976112921, + "grad_norm": 17.298123683774026, + "learning_rate": 1.3629912501909347e-06, + "loss": 0.8169, + "step": 7708 + }, + { + "epoch": 4.185124864277959, + "grad_norm": 12.451648500428053, + "learning_rate": 1.3612192805377222e-06, + "loss": 0.3808, + "step": 7709 + }, + { + "epoch": 4.185667752442996, + "grad_norm": 9.322957640674206, + "learning_rate": 1.3594483793426794e-06, + "loss": 0.2851, + "step": 7710 + }, + { + "epoch": 4.186210640608035, + "grad_norm": 13.229946902809969, + "learning_rate": 1.3576785468248265e-06, + "loss": 0.4352, + "step": 7711 + }, + { + "epoch": 4.186753528773073, + "grad_norm": 11.991137980793829, + "learning_rate": 1.3559097832030677e-06, + "loss": 0.3591, + "step": 7712 + }, + { + "epoch": 4.187296416938111, + "grad_norm": 13.197661861619876, + "learning_rate": 1.3541420886961631e-06, + "loss": 0.3863, + "step": 7713 + }, + { + "epoch": 4.187839305103148, + "grad_norm": 15.68139113867389, + "learning_rate": 1.352375463522746e-06, + "loss": 0.6056, + "step": 7714 + }, + { + "epoch": 4.188382193268187, + "grad_norm": 12.308133143281433, + "learning_rate": 1.3506099079013124e-06, + "loss": 0.3132, + "step": 7715 + }, + { + "epoch": 4.188925081433225, + "grad_norm": 11.907660119383362, + "learning_rate": 1.34884542205023e-06, + "loss": 0.3939, + "step": 7716 + }, + { + "epoch": 4.189467969598263, + "grad_norm": 19.521959289662206, + "learning_rate": 1.3470820061877344e-06, + "loss": 0.5875, + "step": 7717 + }, + { + "epoch": 4.1900108577633, + "grad_norm": 12.438243625833383, + "learning_rate": 1.3453196605319274e-06, + "loss": 0.307, + "step": 7718 + }, + { + "epoch": 4.190553745928339, + "grad_norm": 12.048829129615415, + "learning_rate": 1.3435583853007783e-06, + "loss": 0.3221, + "step": 7719 + }, + { + "epoch": 4.191096634093377, + "grad_norm": 10.135648901522302, + "learning_rate": 1.3417981807121239e-06, + "loss": 0.2947, + "step": 7720 + }, + { + "epoch": 4.191639522258415, + "grad_norm": 12.238680241464417, + "learning_rate": 1.34003904698367e-06, + "loss": 0.4503, + "step": 7721 + }, + { + "epoch": 4.192182410423452, + "grad_norm": 15.537341199882544, + "learning_rate": 1.338280984332988e-06, + "loss": 0.6186, + "step": 7722 + }, + { + "epoch": 4.192725298588491, + "grad_norm": 14.855361885159425, + "learning_rate": 1.3365239929775186e-06, + "loss": 0.5375, + "step": 7723 + }, + { + "epoch": 4.193268186753529, + "grad_norm": 12.128751807129769, + "learning_rate": 1.3347680731345703e-06, + "loss": 0.3844, + "step": 7724 + }, + { + "epoch": 4.193811074918567, + "grad_norm": 14.352544667740425, + "learning_rate": 1.3330132250213158e-06, + "loss": 0.3476, + "step": 7725 + }, + { + "epoch": 4.194353963083604, + "grad_norm": 11.233393992477202, + "learning_rate": 1.331259448854798e-06, + "loss": 0.3616, + "step": 7726 + }, + { + "epoch": 4.194896851248643, + "grad_norm": 16.82629514482139, + "learning_rate": 1.3295067448519305e-06, + "loss": 0.5184, + "step": 7727 + }, + { + "epoch": 4.195439739413681, + "grad_norm": 16.36127203758974, + "learning_rate": 1.3277551132294842e-06, + "loss": 0.3893, + "step": 7728 + }, + { + "epoch": 4.195982627578719, + "grad_norm": 12.108570761810808, + "learning_rate": 1.32600455420411e-06, + "loss": 0.4403, + "step": 7729 + }, + { + "epoch": 4.196525515743756, + "grad_norm": 9.120521528279125, + "learning_rate": 1.3242550679923162e-06, + "loss": 0.2306, + "step": 7730 + }, + { + "epoch": 4.197068403908795, + "grad_norm": 10.947328867337708, + "learning_rate": 1.3225066548104814e-06, + "loss": 0.4367, + "step": 7731 + }, + { + "epoch": 4.197611292073833, + "grad_norm": 13.012382022226626, + "learning_rate": 1.3207593148748543e-06, + "loss": 0.3955, + "step": 7732 + }, + { + "epoch": 4.198154180238871, + "grad_norm": 12.631500933000643, + "learning_rate": 1.3190130484015452e-06, + "loss": 0.3416, + "step": 7733 + }, + { + "epoch": 4.198697068403908, + "grad_norm": 10.924514493419068, + "learning_rate": 1.3172678556065421e-06, + "loss": 0.2377, + "step": 7734 + }, + { + "epoch": 4.199239956568947, + "grad_norm": 17.37092245490247, + "learning_rate": 1.315523736705684e-06, + "loss": 0.3534, + "step": 7735 + }, + { + "epoch": 4.199782844733985, + "grad_norm": 10.606509484945631, + "learning_rate": 1.3137806919146967e-06, + "loss": 0.367, + "step": 7736 + }, + { + "epoch": 4.200325732899023, + "grad_norm": 12.557115475350406, + "learning_rate": 1.3120387214491514e-06, + "loss": 0.3562, + "step": 7737 + }, + { + "epoch": 4.2008686210640604, + "grad_norm": 12.911761209858065, + "learning_rate": 1.310297825524507e-06, + "loss": 0.4865, + "step": 7738 + }, + { + "epoch": 4.201411509229099, + "grad_norm": 16.757054584318308, + "learning_rate": 1.3085580043560763e-06, + "loss": 0.4604, + "step": 7739 + }, + { + "epoch": 4.201954397394137, + "grad_norm": 11.675532522546053, + "learning_rate": 1.3068192581590433e-06, + "loss": 0.403, + "step": 7740 + }, + { + "epoch": 4.202497285559175, + "grad_norm": 16.159527030741106, + "learning_rate": 1.3050815871484622e-06, + "loss": 0.4909, + "step": 7741 + }, + { + "epoch": 4.2030401737242125, + "grad_norm": 16.38078706465185, + "learning_rate": 1.3033449915392437e-06, + "loss": 0.6048, + "step": 7742 + }, + { + "epoch": 4.203583061889251, + "grad_norm": 12.471147411789133, + "learning_rate": 1.301609471546179e-06, + "loss": 0.5796, + "step": 7743 + }, + { + "epoch": 4.204125950054289, + "grad_norm": 12.706419284404552, + "learning_rate": 1.2998750273839179e-06, + "loss": 0.2934, + "step": 7744 + }, + { + "epoch": 4.204668838219327, + "grad_norm": 11.938920053403065, + "learning_rate": 1.2981416592669794e-06, + "loss": 0.4079, + "step": 7745 + }, + { + "epoch": 4.2052117263843645, + "grad_norm": 14.656307060991027, + "learning_rate": 1.2964093674097488e-06, + "loss": 0.4356, + "step": 7746 + }, + { + "epoch": 4.205754614549403, + "grad_norm": 12.993216203856585, + "learning_rate": 1.2946781520264795e-06, + "loss": 0.3562, + "step": 7747 + }, + { + "epoch": 4.206297502714441, + "grad_norm": 14.182239777867444, + "learning_rate": 1.292948013331291e-06, + "loss": 0.4632, + "step": 7748 + }, + { + "epoch": 4.206840390879479, + "grad_norm": 17.007767789893737, + "learning_rate": 1.291218951538169e-06, + "loss": 0.5598, + "step": 7749 + }, + { + "epoch": 4.2073832790445165, + "grad_norm": 10.545945938754782, + "learning_rate": 1.2894909668609657e-06, + "loss": 0.3098, + "step": 7750 + }, + { + "epoch": 4.207926167209555, + "grad_norm": 14.053283729714321, + "learning_rate": 1.2877640595134033e-06, + "loss": 0.5349, + "step": 7751 + }, + { + "epoch": 4.208469055374593, + "grad_norm": 11.754708732243547, + "learning_rate": 1.2860382297090669e-06, + "loss": 0.2941, + "step": 7752 + }, + { + "epoch": 4.209011943539631, + "grad_norm": 13.75780850163015, + "learning_rate": 1.2843134776614108e-06, + "loss": 0.4709, + "step": 7753 + }, + { + "epoch": 4.2095548317046685, + "grad_norm": 14.679629107183144, + "learning_rate": 1.2825898035837537e-06, + "loss": 0.4888, + "step": 7754 + }, + { + "epoch": 4.210097719869707, + "grad_norm": 11.061342428653278, + "learning_rate": 1.2808672076892837e-06, + "loss": 0.2758, + "step": 7755 + }, + { + "epoch": 4.210640608034745, + "grad_norm": 11.723468756168664, + "learning_rate": 1.2791456901910526e-06, + "loss": 0.4185, + "step": 7756 + }, + { + "epoch": 4.211183496199783, + "grad_norm": 10.0084471239469, + "learning_rate": 1.2774252513019825e-06, + "loss": 0.2693, + "step": 7757 + }, + { + "epoch": 4.2117263843648205, + "grad_norm": 14.164307521757856, + "learning_rate": 1.2757058912348596e-06, + "loss": 0.5625, + "step": 7758 + }, + { + "epoch": 4.212269272529859, + "grad_norm": 15.322157151421756, + "learning_rate": 1.2739876102023352e-06, + "loss": 0.4828, + "step": 7759 + }, + { + "epoch": 4.212812160694897, + "grad_norm": 17.051847694097003, + "learning_rate": 1.2722704084169313e-06, + "loss": 0.5323, + "step": 7760 + }, + { + "epoch": 4.213355048859935, + "grad_norm": 11.059167577716101, + "learning_rate": 1.2705542860910325e-06, + "loss": 0.2758, + "step": 7761 + }, + { + "epoch": 4.2138979370249725, + "grad_norm": 11.89278028580855, + "learning_rate": 1.2688392434368924e-06, + "loss": 0.315, + "step": 7762 + }, + { + "epoch": 4.214440825190011, + "grad_norm": 10.03832961212838, + "learning_rate": 1.2671252806666312e-06, + "loss": 0.316, + "step": 7763 + }, + { + "epoch": 4.214983713355049, + "grad_norm": 14.358273720821714, + "learning_rate": 1.2654123979922294e-06, + "loss": 0.447, + "step": 7764 + }, + { + "epoch": 4.215526601520087, + "grad_norm": 13.267216152626093, + "learning_rate": 1.2637005956255489e-06, + "loss": 0.3846, + "step": 7765 + }, + { + "epoch": 4.2160694896851245, + "grad_norm": 10.109215242139403, + "learning_rate": 1.2619898737782977e-06, + "loss": 0.2809, + "step": 7766 + }, + { + "epoch": 4.216612377850163, + "grad_norm": 16.400315992543526, + "learning_rate": 1.2602802326620688e-06, + "loss": 0.632, + "step": 7767 + }, + { + "epoch": 4.217155266015201, + "grad_norm": 11.109933889769897, + "learning_rate": 1.2585716724883057e-06, + "loss": 0.484, + "step": 7768 + }, + { + "epoch": 4.217698154180239, + "grad_norm": 14.07156919017978, + "learning_rate": 1.2568641934683345e-06, + "loss": 0.7668, + "step": 7769 + }, + { + "epoch": 4.2182410423452765, + "grad_norm": 11.176348337381958, + "learning_rate": 1.255157795813332e-06, + "loss": 0.4213, + "step": 7770 + }, + { + "epoch": 4.218783930510315, + "grad_norm": 16.063587698095667, + "learning_rate": 1.2534524797343494e-06, + "loss": 0.6027, + "step": 7771 + }, + { + "epoch": 4.219326818675353, + "grad_norm": 15.762904336294534, + "learning_rate": 1.2517482454423068e-06, + "loss": 0.6919, + "step": 7772 + }, + { + "epoch": 4.219869706840391, + "grad_norm": 11.914370317943318, + "learning_rate": 1.2500450931479801e-06, + "loss": 0.4781, + "step": 7773 + }, + { + "epoch": 4.2204125950054285, + "grad_norm": 17.015123181348883, + "learning_rate": 1.2483430230620264e-06, + "loss": 0.31, + "step": 7774 + }, + { + "epoch": 4.220955483170467, + "grad_norm": 12.337173743701737, + "learning_rate": 1.2466420353949515e-06, + "loss": 0.3487, + "step": 7775 + }, + { + "epoch": 4.221498371335505, + "grad_norm": 9.196709204828075, + "learning_rate": 1.2449421303571418e-06, + "loss": 0.2396, + "step": 7776 + }, + { + "epoch": 4.222041259500543, + "grad_norm": 14.267880372674293, + "learning_rate": 1.2432433081588424e-06, + "loss": 0.7252, + "step": 7777 + }, + { + "epoch": 4.2225841476655805, + "grad_norm": 10.078600378968735, + "learning_rate": 1.2415455690101685e-06, + "loss": 0.3152, + "step": 7778 + }, + { + "epoch": 4.223127035830619, + "grad_norm": 15.372509732415313, + "learning_rate": 1.2398489131210989e-06, + "loss": 0.5178, + "step": 7779 + }, + { + "epoch": 4.223669923995657, + "grad_norm": 11.547384831594652, + "learning_rate": 1.2381533407014722e-06, + "loss": 0.3999, + "step": 7780 + }, + { + "epoch": 4.224212812160695, + "grad_norm": 11.21628120223001, + "learning_rate": 1.23645885196101e-06, + "loss": 0.2856, + "step": 7781 + }, + { + "epoch": 4.2247557003257326, + "grad_norm": 12.171889903747351, + "learning_rate": 1.2347654471092797e-06, + "loss": 0.3303, + "step": 7782 + }, + { + "epoch": 4.225298588490771, + "grad_norm": 11.210591190805236, + "learning_rate": 1.2330731263557293e-06, + "loss": 0.321, + "step": 7783 + }, + { + "epoch": 4.225841476655809, + "grad_norm": 11.56900032090966, + "learning_rate": 1.2313818899096685e-06, + "loss": 0.3416, + "step": 7784 + }, + { + "epoch": 4.226384364820847, + "grad_norm": 10.580053290763194, + "learning_rate": 1.2296917379802709e-06, + "loss": 0.3201, + "step": 7785 + }, + { + "epoch": 4.226927252985885, + "grad_norm": 18.553406932221385, + "learning_rate": 1.2280026707765758e-06, + "loss": 0.4094, + "step": 7786 + }, + { + "epoch": 4.227470141150923, + "grad_norm": 11.769072597193505, + "learning_rate": 1.2263146885074917e-06, + "loss": 0.3517, + "step": 7787 + }, + { + "epoch": 4.228013029315961, + "grad_norm": 14.383741980090374, + "learning_rate": 1.2246277913817895e-06, + "loss": 0.4665, + "step": 7788 + }, + { + "epoch": 4.228555917480999, + "grad_norm": 13.049438484607592, + "learning_rate": 1.2229419796081077e-06, + "loss": 0.4222, + "step": 7789 + }, + { + "epoch": 4.229098805646037, + "grad_norm": 9.676387763683955, + "learning_rate": 1.2212572533949506e-06, + "loss": 0.3977, + "step": 7790 + }, + { + "epoch": 4.229641693811075, + "grad_norm": 20.33530097040713, + "learning_rate": 1.2195736129506885e-06, + "loss": 0.446, + "step": 7791 + }, + { + "epoch": 4.230184581976113, + "grad_norm": 14.282100526459736, + "learning_rate": 1.2178910584835556e-06, + "loss": 0.3286, + "step": 7792 + }, + { + "epoch": 4.230727470141151, + "grad_norm": 19.14182283529482, + "learning_rate": 1.2162095902016525e-06, + "loss": 0.6531, + "step": 7793 + }, + { + "epoch": 4.231270358306189, + "grad_norm": 9.949859543104305, + "learning_rate": 1.2145292083129467e-06, + "loss": 0.3108, + "step": 7794 + }, + { + "epoch": 4.231813246471227, + "grad_norm": 12.280893090679601, + "learning_rate": 1.2128499130252714e-06, + "loss": 0.3321, + "step": 7795 + }, + { + "epoch": 4.232356134636265, + "grad_norm": 13.441725051837208, + "learning_rate": 1.2111717045463223e-06, + "loss": 0.4196, + "step": 7796 + }, + { + "epoch": 4.232899022801303, + "grad_norm": 15.03461419306544, + "learning_rate": 1.2094945830836625e-06, + "loss": 0.3983, + "step": 7797 + }, + { + "epoch": 4.233441910966341, + "grad_norm": 16.773668359582217, + "learning_rate": 1.207818548844727e-06, + "loss": 0.386, + "step": 7798 + }, + { + "epoch": 4.233984799131379, + "grad_norm": 11.119171257862224, + "learning_rate": 1.2061436020368033e-06, + "loss": 0.4611, + "step": 7799 + }, + { + "epoch": 4.234527687296417, + "grad_norm": 8.229976665110055, + "learning_rate": 1.204469742867055e-06, + "loss": 0.2615, + "step": 7800 + }, + { + "epoch": 4.235070575461455, + "grad_norm": 13.435979628871456, + "learning_rate": 1.2027969715425058e-06, + "loss": 0.4595, + "step": 7801 + }, + { + "epoch": 4.235613463626493, + "grad_norm": 17.70018909852596, + "learning_rate": 1.2011252882700486e-06, + "loss": 0.56, + "step": 7802 + }, + { + "epoch": 4.236156351791531, + "grad_norm": 13.260555683605508, + "learning_rate": 1.1994546932564388e-06, + "loss": 0.384, + "step": 7803 + }, + { + "epoch": 4.236699239956569, + "grad_norm": 10.203419005145452, + "learning_rate": 1.1977851867082968e-06, + "loss": 0.3906, + "step": 7804 + }, + { + "epoch": 4.237242128121607, + "grad_norm": 19.37091442524948, + "learning_rate": 1.1961167688321151e-06, + "loss": 0.5527, + "step": 7805 + }, + { + "epoch": 4.237785016286645, + "grad_norm": 17.208395947554312, + "learning_rate": 1.1944494398342388e-06, + "loss": 0.597, + "step": 7806 + }, + { + "epoch": 4.238327904451683, + "grad_norm": 12.251023694428005, + "learning_rate": 1.1927831999208938e-06, + "loss": 0.4034, + "step": 7807 + }, + { + "epoch": 4.238870792616721, + "grad_norm": 16.49490715109582, + "learning_rate": 1.1911180492981578e-06, + "loss": 0.6092, + "step": 7808 + }, + { + "epoch": 4.239413680781759, + "grad_norm": 14.390923091386458, + "learning_rate": 1.189453988171978e-06, + "loss": 0.5611, + "step": 7809 + }, + { + "epoch": 4.239956568946797, + "grad_norm": 16.171314155557898, + "learning_rate": 1.1877910167481754e-06, + "loss": 0.3232, + "step": 7810 + }, + { + "epoch": 4.240499457111835, + "grad_norm": 15.308750990763075, + "learning_rate": 1.18612913523242e-06, + "loss": 0.3301, + "step": 7811 + }, + { + "epoch": 4.241042345276873, + "grad_norm": 11.75895557202086, + "learning_rate": 1.1844683438302652e-06, + "loss": 0.4324, + "step": 7812 + }, + { + "epoch": 4.241585233441911, + "grad_norm": 11.620344969621105, + "learning_rate": 1.1828086427471107e-06, + "loss": 0.5246, + "step": 7813 + }, + { + "epoch": 4.242128121606949, + "grad_norm": 11.17860914452438, + "learning_rate": 1.1811500321882374e-06, + "loss": 0.4626, + "step": 7814 + }, + { + "epoch": 4.242671009771987, + "grad_norm": 12.479770513980384, + "learning_rate": 1.1794925123587841e-06, + "loss": 0.375, + "step": 7815 + }, + { + "epoch": 4.243213897937025, + "grad_norm": 13.677272569244249, + "learning_rate": 1.1778360834637549e-06, + "loss": 0.3313, + "step": 7816 + }, + { + "epoch": 4.243756786102063, + "grad_norm": 11.135594932532166, + "learning_rate": 1.1761807457080187e-06, + "loss": 0.4987, + "step": 7817 + }, + { + "epoch": 4.244299674267101, + "grad_norm": 16.738993364102154, + "learning_rate": 1.1745264992963112e-06, + "loss": 0.367, + "step": 7818 + }, + { + "epoch": 4.244842562432139, + "grad_norm": 12.305866749951548, + "learning_rate": 1.1728733444332342e-06, + "loss": 0.3499, + "step": 7819 + }, + { + "epoch": 4.245385450597177, + "grad_norm": 13.571484126012804, + "learning_rate": 1.1712212813232448e-06, + "loss": 0.4086, + "step": 7820 + }, + { + "epoch": 4.245928338762215, + "grad_norm": 18.75127331038876, + "learning_rate": 1.1695703101706823e-06, + "loss": 0.6246, + "step": 7821 + }, + { + "epoch": 4.246471226927253, + "grad_norm": 17.48505499787076, + "learning_rate": 1.1679204311797365e-06, + "loss": 0.5487, + "step": 7822 + }, + { + "epoch": 4.247014115092291, + "grad_norm": 9.684180280120152, + "learning_rate": 1.1662716445544676e-06, + "loss": 0.2527, + "step": 7823 + }, + { + "epoch": 4.247557003257329, + "grad_norm": 13.169958031763949, + "learning_rate": 1.164623950498801e-06, + "loss": 0.3898, + "step": 7824 + }, + { + "epoch": 4.248099891422367, + "grad_norm": 11.019724381747643, + "learning_rate": 1.1629773492165264e-06, + "loss": 0.3471, + "step": 7825 + }, + { + "epoch": 4.248642779587405, + "grad_norm": 9.968852095122674, + "learning_rate": 1.1613318409112962e-06, + "loss": 0.3911, + "step": 7826 + }, + { + "epoch": 4.249185667752443, + "grad_norm": 15.553115617722527, + "learning_rate": 1.159687425786632e-06, + "loss": 0.425, + "step": 7827 + }, + { + "epoch": 4.249728555917481, + "grad_norm": 12.82652568103273, + "learning_rate": 1.1580441040459146e-06, + "loss": 0.436, + "step": 7828 + }, + { + "epoch": 4.250271444082519, + "grad_norm": 10.72479783442799, + "learning_rate": 1.1564018758923956e-06, + "loss": 0.3967, + "step": 7829 + }, + { + "epoch": 4.250814332247557, + "grad_norm": 17.35073007540839, + "learning_rate": 1.1547607415291874e-06, + "loss": 0.473, + "step": 7830 + }, + { + "epoch": 4.251357220412595, + "grad_norm": 11.61546854660444, + "learning_rate": 1.153120701159267e-06, + "loss": 0.3997, + "step": 7831 + }, + { + "epoch": 4.251900108577633, + "grad_norm": 14.988521210943322, + "learning_rate": 1.1514817549854784e-06, + "loss": 0.5253, + "step": 7832 + }, + { + "epoch": 4.252442996742671, + "grad_norm": 12.668110351197043, + "learning_rate": 1.149843903210529e-06, + "loss": 0.4776, + "step": 7833 + }, + { + "epoch": 4.252985884907709, + "grad_norm": 15.448405948912091, + "learning_rate": 1.1482071460369903e-06, + "loss": 0.4487, + "step": 7834 + }, + { + "epoch": 4.253528773072747, + "grad_norm": 13.582390046184774, + "learning_rate": 1.146571483667298e-06, + "loss": 0.3464, + "step": 7835 + }, + { + "epoch": 4.254071661237785, + "grad_norm": 16.689486575643958, + "learning_rate": 1.1449369163037582e-06, + "loss": 0.4516, + "step": 7836 + }, + { + "epoch": 4.254614549402823, + "grad_norm": 12.868709622460063, + "learning_rate": 1.1433034441485303e-06, + "loss": 0.2851, + "step": 7837 + }, + { + "epoch": 4.255157437567861, + "grad_norm": 8.159166476090338, + "learning_rate": 1.1416710674036512e-06, + "loss": 0.1686, + "step": 7838 + }, + { + "epoch": 4.255700325732899, + "grad_norm": 10.431332020179449, + "learning_rate": 1.1400397862710112e-06, + "loss": 0.3455, + "step": 7839 + }, + { + "epoch": 4.256243213897937, + "grad_norm": 12.657036094463699, + "learning_rate": 1.1384096009523705e-06, + "loss": 0.3586, + "step": 7840 + }, + { + "epoch": 4.256786102062975, + "grad_norm": 17.536752353552888, + "learning_rate": 1.1367805116493534e-06, + "loss": 0.5641, + "step": 7841 + }, + { + "epoch": 4.257328990228013, + "grad_norm": 12.422330381694959, + "learning_rate": 1.135152518563447e-06, + "loss": 0.5808, + "step": 7842 + }, + { + "epoch": 4.257871878393051, + "grad_norm": 14.406429659207616, + "learning_rate": 1.13352562189601e-06, + "loss": 0.4333, + "step": 7843 + }, + { + "epoch": 4.258414766558089, + "grad_norm": 13.093016383546724, + "learning_rate": 1.1318998218482501e-06, + "loss": 0.4737, + "step": 7844 + }, + { + "epoch": 4.258957654723127, + "grad_norm": 12.327912324997381, + "learning_rate": 1.130275118621259e-06, + "loss": 0.3371, + "step": 7845 + }, + { + "epoch": 4.259500542888165, + "grad_norm": 12.784200365328555, + "learning_rate": 1.1286515124159726e-06, + "loss": 0.4967, + "step": 7846 + }, + { + "epoch": 4.260043431053203, + "grad_norm": 14.609662199743113, + "learning_rate": 1.1270290034332076e-06, + "loss": 0.5266, + "step": 7847 + }, + { + "epoch": 4.260586319218241, + "grad_norm": 16.132913961447098, + "learning_rate": 1.1254075918736395e-06, + "loss": 0.7818, + "step": 7848 + }, + { + "epoch": 4.261129207383279, + "grad_norm": 12.362451065508255, + "learning_rate": 1.1237872779378013e-06, + "loss": 0.3534, + "step": 7849 + }, + { + "epoch": 4.261672095548317, + "grad_norm": 11.787990488474275, + "learning_rate": 1.1221680618261022e-06, + "loss": 0.3316, + "step": 7850 + }, + { + "epoch": 4.262214983713355, + "grad_norm": 16.520849175369758, + "learning_rate": 1.1205499437388024e-06, + "loss": 0.6686, + "step": 7851 + }, + { + "epoch": 4.262757871878393, + "grad_norm": 15.087215559431693, + "learning_rate": 1.1189329238760416e-06, + "loss": 0.4386, + "step": 7852 + }, + { + "epoch": 4.263300760043431, + "grad_norm": 14.92562475156193, + "learning_rate": 1.1173170024378067e-06, + "loss": 0.6797, + "step": 7853 + }, + { + "epoch": 4.263843648208469, + "grad_norm": 10.695763423794212, + "learning_rate": 1.1157021796239631e-06, + "loss": 0.2609, + "step": 7854 + }, + { + "epoch": 4.264386536373507, + "grad_norm": 13.337827871285958, + "learning_rate": 1.1140884556342346e-06, + "loss": 0.6186, + "step": 7855 + }, + { + "epoch": 4.264929424538545, + "grad_norm": 13.122182163353113, + "learning_rate": 1.1124758306682081e-06, + "loss": 0.518, + "step": 7856 + }, + { + "epoch": 4.265472312703583, + "grad_norm": 8.714454332149657, + "learning_rate": 1.1108643049253364e-06, + "loss": 0.3084, + "step": 7857 + }, + { + "epoch": 4.266015200868621, + "grad_norm": 12.760955138975337, + "learning_rate": 1.1092538786049311e-06, + "loss": 0.5716, + "step": 7858 + }, + { + "epoch": 4.266558089033659, + "grad_norm": 17.235261430649693, + "learning_rate": 1.1076445519061784e-06, + "loss": 0.4657, + "step": 7859 + }, + { + "epoch": 4.267100977198697, + "grad_norm": 14.044151036119151, + "learning_rate": 1.1060363250281192e-06, + "loss": 0.6461, + "step": 7860 + }, + { + "epoch": 4.267643865363735, + "grad_norm": 12.741420750608738, + "learning_rate": 1.104429198169662e-06, + "loss": 0.4156, + "step": 7861 + }, + { + "epoch": 4.268186753528773, + "grad_norm": 16.38667997198235, + "learning_rate": 1.10282317152958e-06, + "loss": 0.4167, + "step": 7862 + }, + { + "epoch": 4.268729641693811, + "grad_norm": 12.512768079843084, + "learning_rate": 1.1012182453065069e-06, + "loss": 0.4918, + "step": 7863 + }, + { + "epoch": 4.269272529858849, + "grad_norm": 13.720175584312965, + "learning_rate": 1.0996144196989444e-06, + "loss": 0.3484, + "step": 7864 + }, + { + "epoch": 4.269815418023887, + "grad_norm": 12.963348508780628, + "learning_rate": 1.0980116949052554e-06, + "loss": 0.4639, + "step": 7865 + }, + { + "epoch": 4.270358306188925, + "grad_norm": 16.8876044604694, + "learning_rate": 1.096410071123668e-06, + "loss": 0.4278, + "step": 7866 + }, + { + "epoch": 4.270901194353963, + "grad_norm": 14.644807785059461, + "learning_rate": 1.0948095485522725e-06, + "loss": 0.3409, + "step": 7867 + }, + { + "epoch": 4.271444082519001, + "grad_norm": 13.719922365450167, + "learning_rate": 1.093210127389025e-06, + "loss": 0.5096, + "step": 7868 + }, + { + "epoch": 4.271986970684039, + "grad_norm": 15.219211444329652, + "learning_rate": 1.091611807831745e-06, + "loss": 0.4472, + "step": 7869 + }, + { + "epoch": 4.272529858849077, + "grad_norm": 12.392349389977586, + "learning_rate": 1.0900145900781133e-06, + "loss": 0.2404, + "step": 7870 + }, + { + "epoch": 4.273072747014115, + "grad_norm": 15.937024579734995, + "learning_rate": 1.0884184743256788e-06, + "loss": 0.5215, + "step": 7871 + }, + { + "epoch": 4.273615635179153, + "grad_norm": 11.853999121791691, + "learning_rate": 1.0868234607718498e-06, + "loss": 0.4883, + "step": 7872 + }, + { + "epoch": 4.274158523344191, + "grad_norm": 13.690567615009499, + "learning_rate": 1.0852295496139008e-06, + "loss": 0.5029, + "step": 7873 + }, + { + "epoch": 4.274701411509229, + "grad_norm": 9.345078157701355, + "learning_rate": 1.083636741048969e-06, + "loss": 0.34, + "step": 7874 + }, + { + "epoch": 4.2752442996742674, + "grad_norm": 10.355626815319667, + "learning_rate": 1.082045035274053e-06, + "loss": 0.2761, + "step": 7875 + }, + { + "epoch": 4.275787187839305, + "grad_norm": 9.570407579073933, + "learning_rate": 1.0804544324860245e-06, + "loss": 0.2851, + "step": 7876 + }, + { + "epoch": 4.276330076004343, + "grad_norm": 12.52101649250042, + "learning_rate": 1.0788649328816059e-06, + "loss": 0.4681, + "step": 7877 + }, + { + "epoch": 4.276872964169381, + "grad_norm": 10.848327417784757, + "learning_rate": 1.0772765366573902e-06, + "loss": 0.4557, + "step": 7878 + }, + { + "epoch": 4.2774158523344195, + "grad_norm": 14.851238098236974, + "learning_rate": 1.0756892440098333e-06, + "loss": 0.3483, + "step": 7879 + }, + { + "epoch": 4.277958740499457, + "grad_norm": 10.516600101157945, + "learning_rate": 1.0741030551352517e-06, + "loss": 0.2466, + "step": 7880 + }, + { + "epoch": 4.278501628664495, + "grad_norm": 13.065799904684205, + "learning_rate": 1.0725179702298338e-06, + "loss": 0.3965, + "step": 7881 + }, + { + "epoch": 4.279044516829533, + "grad_norm": 11.826945310503726, + "learning_rate": 1.0709339894896186e-06, + "loss": 0.3297, + "step": 7882 + }, + { + "epoch": 4.2795874049945715, + "grad_norm": 12.014578152352707, + "learning_rate": 1.0693511131105216e-06, + "loss": 0.6114, + "step": 7883 + }, + { + "epoch": 4.280130293159609, + "grad_norm": 13.074176249077697, + "learning_rate": 1.0677693412883082e-06, + "loss": 0.4467, + "step": 7884 + }, + { + "epoch": 4.280673181324647, + "grad_norm": 16.7050290896531, + "learning_rate": 1.0661886742186212e-06, + "loss": 0.7551, + "step": 7885 + }, + { + "epoch": 4.281216069489685, + "grad_norm": 9.762649259962595, + "learning_rate": 1.064609112096956e-06, + "loss": 0.3769, + "step": 7886 + }, + { + "epoch": 4.2817589576547235, + "grad_norm": 13.322662274779637, + "learning_rate": 1.0630306551186775e-06, + "loss": 0.6132, + "step": 7887 + }, + { + "epoch": 4.282301845819761, + "grad_norm": 14.512051944801728, + "learning_rate": 1.061453303479013e-06, + "loss": 0.2763, + "step": 7888 + }, + { + "epoch": 4.282844733984799, + "grad_norm": 11.38093344044884, + "learning_rate": 1.0598770573730465e-06, + "loss": 0.4843, + "step": 7889 + }, + { + "epoch": 4.283387622149837, + "grad_norm": 12.749261916645077, + "learning_rate": 1.0583019169957377e-06, + "loss": 0.3451, + "step": 7890 + }, + { + "epoch": 4.2839305103148755, + "grad_norm": 11.929936221189454, + "learning_rate": 1.0567278825418937e-06, + "loss": 0.7333, + "step": 7891 + }, + { + "epoch": 4.284473398479913, + "grad_norm": 7.644818532912615, + "learning_rate": 1.0551549542062023e-06, + "loss": 0.2397, + "step": 7892 + }, + { + "epoch": 4.285016286644951, + "grad_norm": 13.22419891221089, + "learning_rate": 1.0535831321832003e-06, + "loss": 0.348, + "step": 7893 + }, + { + "epoch": 4.285559174809989, + "grad_norm": 16.960704066599227, + "learning_rate": 1.0520124166672964e-06, + "loss": 0.5281, + "step": 7894 + }, + { + "epoch": 4.2861020629750275, + "grad_norm": 13.957960998474602, + "learning_rate": 1.0504428078527573e-06, + "loss": 0.4479, + "step": 7895 + }, + { + "epoch": 4.286644951140065, + "grad_norm": 10.380635061146368, + "learning_rate": 1.0488743059337148e-06, + "loss": 0.2593, + "step": 7896 + }, + { + "epoch": 4.287187839305103, + "grad_norm": 14.044042926670437, + "learning_rate": 1.0473069111041644e-06, + "loss": 0.3819, + "step": 7897 + }, + { + "epoch": 4.287730727470141, + "grad_norm": 11.351044141908234, + "learning_rate": 1.0457406235579638e-06, + "loss": 0.3519, + "step": 7898 + }, + { + "epoch": 4.2882736156351795, + "grad_norm": 15.23295931267919, + "learning_rate": 1.044175443488834e-06, + "loss": 0.4457, + "step": 7899 + }, + { + "epoch": 4.288816503800217, + "grad_norm": 15.871156671188771, + "learning_rate": 1.0426113710903586e-06, + "loss": 0.4451, + "step": 7900 + }, + { + "epoch": 4.289359391965255, + "grad_norm": 15.223931858136798, + "learning_rate": 1.0410484065559845e-06, + "loss": 0.7813, + "step": 7901 + }, + { + "epoch": 4.289902280130293, + "grad_norm": 17.47270174804404, + "learning_rate": 1.039486550079023e-06, + "loss": 0.6555, + "step": 7902 + }, + { + "epoch": 4.2904451682953315, + "grad_norm": 14.66163351293163, + "learning_rate": 1.0379258018526451e-06, + "loss": 0.4451, + "step": 7903 + }, + { + "epoch": 4.290988056460369, + "grad_norm": 14.775375275089438, + "learning_rate": 1.0363661620698872e-06, + "loss": 0.5579, + "step": 7904 + }, + { + "epoch": 4.291530944625407, + "grad_norm": 17.26244179697533, + "learning_rate": 1.0348076309236487e-06, + "loss": 0.436, + "step": 7905 + }, + { + "epoch": 4.292073832790445, + "grad_norm": 11.967809939669856, + "learning_rate": 1.0332502086066909e-06, + "loss": 0.251, + "step": 7906 + }, + { + "epoch": 4.2926167209554835, + "grad_norm": 18.886544353437575, + "learning_rate": 1.0316938953116374e-06, + "loss": 0.7729, + "step": 7907 + }, + { + "epoch": 4.293159609120521, + "grad_norm": 16.082653194833412, + "learning_rate": 1.0301386912309763e-06, + "loss": 0.6706, + "step": 7908 + }, + { + "epoch": 4.293702497285559, + "grad_norm": 10.115336308458334, + "learning_rate": 1.0285845965570584e-06, + "loss": 0.238, + "step": 7909 + }, + { + "epoch": 4.294245385450597, + "grad_norm": 10.909597072320835, + "learning_rate": 1.0270316114820954e-06, + "loss": 0.2976, + "step": 7910 + }, + { + "epoch": 4.2947882736156355, + "grad_norm": 13.390757276403917, + "learning_rate": 1.025479736198164e-06, + "loss": 0.8118, + "step": 7911 + }, + { + "epoch": 4.295331161780673, + "grad_norm": 10.149705690991324, + "learning_rate": 1.0239289708972022e-06, + "loss": 0.3365, + "step": 7912 + }, + { + "epoch": 4.295874049945711, + "grad_norm": 14.439616764678872, + "learning_rate": 1.0223793157710083e-06, + "loss": 0.4143, + "step": 7913 + }, + { + "epoch": 4.296416938110749, + "grad_norm": 12.606369995049853, + "learning_rate": 1.020830771011253e-06, + "loss": 0.4796, + "step": 7914 + }, + { + "epoch": 4.2969598262757875, + "grad_norm": 12.593118847476834, + "learning_rate": 1.0192833368094556e-06, + "loss": 0.4738, + "step": 7915 + }, + { + "epoch": 4.297502714440825, + "grad_norm": 16.435528985746867, + "learning_rate": 1.0177370133570108e-06, + "loss": 0.4431, + "step": 7916 + }, + { + "epoch": 4.298045602605863, + "grad_norm": 10.261492535138068, + "learning_rate": 1.0161918008451665e-06, + "loss": 0.3084, + "step": 7917 + }, + { + "epoch": 4.298588490770901, + "grad_norm": 19.67761715432872, + "learning_rate": 1.0146476994650367e-06, + "loss": 0.4107, + "step": 7918 + }, + { + "epoch": 4.2991313789359396, + "grad_norm": 12.628487019222309, + "learning_rate": 1.013104709407603e-06, + "loss": 0.3636, + "step": 7919 + }, + { + "epoch": 4.299674267100977, + "grad_norm": 8.48683461551969, + "learning_rate": 1.0115628308636981e-06, + "loss": 0.2398, + "step": 7920 + }, + { + "epoch": 4.300217155266015, + "grad_norm": 10.946283608256392, + "learning_rate": 1.010022064024032e-06, + "loss": 0.3367, + "step": 7921 + }, + { + "epoch": 4.300760043431053, + "grad_norm": 11.690515621064748, + "learning_rate": 1.0084824090791612e-06, + "loss": 0.3374, + "step": 7922 + }, + { + "epoch": 4.301302931596092, + "grad_norm": 13.693996659420456, + "learning_rate": 1.0069438662195197e-06, + "loss": 0.5506, + "step": 7923 + }, + { + "epoch": 4.301845819761129, + "grad_norm": 10.325683049665914, + "learning_rate": 1.0054064356353898e-06, + "loss": 0.3755, + "step": 7924 + }, + { + "epoch": 4.302388707926167, + "grad_norm": 10.906414174966567, + "learning_rate": 1.0038701175169295e-06, + "loss": 0.3714, + "step": 7925 + }, + { + "epoch": 4.302931596091205, + "grad_norm": 11.218531762718778, + "learning_rate": 1.0023349120541526e-06, + "loss": 0.256, + "step": 7926 + }, + { + "epoch": 4.303474484256244, + "grad_norm": 11.942829819879181, + "learning_rate": 1.00080081943693e-06, + "loss": 0.4458, + "step": 7927 + }, + { + "epoch": 4.304017372421281, + "grad_norm": 10.47979684406596, + "learning_rate": 9.992678398550082e-07, + "loss": 0.2921, + "step": 7928 + }, + { + "epoch": 4.304560260586319, + "grad_norm": 13.649113498748063, + "learning_rate": 9.977359734979818e-07, + "loss": 0.3592, + "step": 7929 + }, + { + "epoch": 4.305103148751357, + "grad_norm": 15.946994134404349, + "learning_rate": 9.962052205553196e-07, + "loss": 0.5153, + "step": 7930 + }, + { + "epoch": 4.305646036916396, + "grad_norm": 11.373437657544107, + "learning_rate": 9.946755812163445e-07, + "loss": 0.34, + "step": 7931 + }, + { + "epoch": 4.306188925081433, + "grad_norm": 13.701873624030634, + "learning_rate": 9.931470556702472e-07, + "loss": 0.4921, + "step": 7932 + }, + { + "epoch": 4.306731813246471, + "grad_norm": 15.895989124158662, + "learning_rate": 9.916196441060765e-07, + "loss": 0.5287, + "step": 7933 + }, + { + "epoch": 4.307274701411509, + "grad_norm": 12.624151740529374, + "learning_rate": 9.900933467127448e-07, + "loss": 0.4054, + "step": 7934 + }, + { + "epoch": 4.307817589576548, + "grad_norm": 10.664486127172733, + "learning_rate": 9.88568163679028e-07, + "loss": 0.2721, + "step": 7935 + }, + { + "epoch": 4.308360477741585, + "grad_norm": 11.459203733504268, + "learning_rate": 9.870440951935633e-07, + "loss": 0.4345, + "step": 7936 + }, + { + "epoch": 4.308903365906623, + "grad_norm": 9.387573986721975, + "learning_rate": 9.855211414448484e-07, + "loss": 0.2966, + "step": 7937 + }, + { + "epoch": 4.309446254071661, + "grad_norm": 11.914282073062253, + "learning_rate": 9.839993026212458e-07, + "loss": 0.4385, + "step": 7938 + }, + { + "epoch": 4.3099891422367, + "grad_norm": 13.814254907233607, + "learning_rate": 9.824785789109792e-07, + "loss": 0.4864, + "step": 7939 + }, + { + "epoch": 4.310532030401737, + "grad_norm": 9.848682371430808, + "learning_rate": 9.809589705021339e-07, + "loss": 0.3087, + "step": 7940 + }, + { + "epoch": 4.311074918566775, + "grad_norm": 12.717733643585042, + "learning_rate": 9.79440477582657e-07, + "loss": 0.3316, + "step": 7941 + }, + { + "epoch": 4.311617806731813, + "grad_norm": 8.913158382091432, + "learning_rate": 9.77923100340359e-07, + "loss": 0.2116, + "step": 7942 + }, + { + "epoch": 4.312160694896852, + "grad_norm": 15.897903840435498, + "learning_rate": 9.764068389629112e-07, + "loss": 0.4687, + "step": 7943 + }, + { + "epoch": 4.312703583061889, + "grad_norm": 16.56909882802291, + "learning_rate": 9.748916936378461e-07, + "loss": 0.4037, + "step": 7944 + }, + { + "epoch": 4.313246471226927, + "grad_norm": 11.499341380465028, + "learning_rate": 9.733776645525606e-07, + "loss": 0.4861, + "step": 7945 + }, + { + "epoch": 4.313789359391965, + "grad_norm": 14.229871658466864, + "learning_rate": 9.718647518943115e-07, + "loss": 0.4858, + "step": 7946 + }, + { + "epoch": 4.314332247557004, + "grad_norm": 11.12286071870711, + "learning_rate": 9.70352955850219e-07, + "loss": 0.3882, + "step": 7947 + }, + { + "epoch": 4.314875135722041, + "grad_norm": 19.942030070834193, + "learning_rate": 9.688422766072648e-07, + "loss": 0.7211, + "step": 7948 + }, + { + "epoch": 4.315418023887079, + "grad_norm": 12.959497734741237, + "learning_rate": 9.673327143522904e-07, + "loss": 0.5216, + "step": 7949 + }, + { + "epoch": 4.315960912052117, + "grad_norm": 13.592821886751958, + "learning_rate": 9.65824269272001e-07, + "loss": 0.586, + "step": 7950 + }, + { + "epoch": 4.316503800217156, + "grad_norm": 16.14045501283915, + "learning_rate": 9.64316941552964e-07, + "loss": 0.448, + "step": 7951 + }, + { + "epoch": 4.317046688382193, + "grad_norm": 14.78354837719301, + "learning_rate": 9.628107313816116e-07, + "loss": 0.5419, + "step": 7952 + }, + { + "epoch": 4.317589576547231, + "grad_norm": 11.53798450893877, + "learning_rate": 9.613056389442265e-07, + "loss": 0.3079, + "step": 7953 + }, + { + "epoch": 4.318132464712269, + "grad_norm": 19.061498107117767, + "learning_rate": 9.598016644269703e-07, + "loss": 0.6412, + "step": 7954 + }, + { + "epoch": 4.318675352877308, + "grad_norm": 14.807149115023895, + "learning_rate": 9.582988080158506e-07, + "loss": 0.4823, + "step": 7955 + }, + { + "epoch": 4.319218241042345, + "grad_norm": 15.900590171517315, + "learning_rate": 9.567970698967434e-07, + "loss": 0.4381, + "step": 7956 + }, + { + "epoch": 4.319761129207383, + "grad_norm": 14.580684073605896, + "learning_rate": 9.552964502553885e-07, + "loss": 0.7024, + "step": 7957 + }, + { + "epoch": 4.320304017372421, + "grad_norm": 13.182118947543977, + "learning_rate": 9.537969492773814e-07, + "loss": 0.4845, + "step": 7958 + }, + { + "epoch": 4.32084690553746, + "grad_norm": 16.398747723242373, + "learning_rate": 9.522985671481888e-07, + "loss": 0.4973, + "step": 7959 + }, + { + "epoch": 4.321389793702497, + "grad_norm": 12.433647949014988, + "learning_rate": 9.50801304053125e-07, + "loss": 0.2912, + "step": 7960 + }, + { + "epoch": 4.321932681867535, + "grad_norm": 18.17349554177303, + "learning_rate": 9.493051601773828e-07, + "loss": 0.3915, + "step": 7961 + }, + { + "epoch": 4.322475570032573, + "grad_norm": 8.718427912132642, + "learning_rate": 9.478101357059988e-07, + "loss": 0.2589, + "step": 7962 + }, + { + "epoch": 4.323018458197612, + "grad_norm": 15.540646935471532, + "learning_rate": 9.46316230823886e-07, + "loss": 0.5203, + "step": 7963 + }, + { + "epoch": 4.323561346362649, + "grad_norm": 10.368742305387377, + "learning_rate": 9.448234457158112e-07, + "loss": 0.337, + "step": 7964 + }, + { + "epoch": 4.324104234527687, + "grad_norm": 15.68035062917153, + "learning_rate": 9.433317805664055e-07, + "loss": 0.5318, + "step": 7965 + }, + { + "epoch": 4.324647122692725, + "grad_norm": 11.340254631487928, + "learning_rate": 9.418412355601603e-07, + "loss": 0.4241, + "step": 7966 + }, + { + "epoch": 4.325190010857764, + "grad_norm": 13.451210184334698, + "learning_rate": 9.403518108814236e-07, + "loss": 0.5082, + "step": 7967 + }, + { + "epoch": 4.3257328990228014, + "grad_norm": 11.17927635762652, + "learning_rate": 9.388635067144169e-07, + "loss": 0.3793, + "step": 7968 + }, + { + "epoch": 4.326275787187839, + "grad_norm": 13.902709479850872, + "learning_rate": 9.373763232432121e-07, + "loss": 0.5547, + "step": 7969 + }, + { + "epoch": 4.326818675352877, + "grad_norm": 18.295405200485035, + "learning_rate": 9.358902606517472e-07, + "loss": 0.3459, + "step": 7970 + }, + { + "epoch": 4.327361563517916, + "grad_norm": 11.600290394966255, + "learning_rate": 9.344053191238211e-07, + "loss": 0.3111, + "step": 7971 + }, + { + "epoch": 4.3279044516829535, + "grad_norm": 14.034291419295219, + "learning_rate": 9.329214988430934e-07, + "loss": 0.4003, + "step": 7972 + }, + { + "epoch": 4.328447339847991, + "grad_norm": 11.650796163329682, + "learning_rate": 9.314387999930863e-07, + "loss": 0.5023, + "step": 7973 + }, + { + "epoch": 4.328990228013029, + "grad_norm": 12.709065524457731, + "learning_rate": 9.299572227571807e-07, + "loss": 0.5242, + "step": 7974 + }, + { + "epoch": 4.329533116178068, + "grad_norm": 14.580030445612671, + "learning_rate": 9.284767673186213e-07, + "loss": 0.4893, + "step": 7975 + }, + { + "epoch": 4.3300760043431055, + "grad_norm": 17.03339708205366, + "learning_rate": 9.269974338605137e-07, + "loss": 0.3985, + "step": 7976 + }, + { + "epoch": 4.330618892508143, + "grad_norm": 15.730665291068982, + "learning_rate": 9.255192225658239e-07, + "loss": 0.335, + "step": 7977 + }, + { + "epoch": 4.331161780673181, + "grad_norm": 10.567417866924043, + "learning_rate": 9.240421336173788e-07, + "loss": 0.2627, + "step": 7978 + }, + { + "epoch": 4.33170466883822, + "grad_norm": 10.543118083895102, + "learning_rate": 9.22566167197867e-07, + "loss": 0.5185, + "step": 7979 + }, + { + "epoch": 4.3322475570032575, + "grad_norm": 12.182864373567915, + "learning_rate": 9.210913234898389e-07, + "loss": 0.3888, + "step": 7980 + }, + { + "epoch": 4.332790445168295, + "grad_norm": 11.069697536316085, + "learning_rate": 9.196176026757064e-07, + "loss": 0.4811, + "step": 7981 + }, + { + "epoch": 4.333333333333333, + "grad_norm": 16.843361756349292, + "learning_rate": 9.181450049377405e-07, + "loss": 0.5628, + "step": 7982 + }, + { + "epoch": 4.333876221498372, + "grad_norm": 16.78954259852446, + "learning_rate": 9.166735304580743e-07, + "loss": 0.5249, + "step": 7983 + }, + { + "epoch": 4.3344191096634095, + "grad_norm": 10.463116306289768, + "learning_rate": 9.15203179418701e-07, + "loss": 0.3087, + "step": 7984 + }, + { + "epoch": 4.334961997828447, + "grad_norm": 10.619811686151522, + "learning_rate": 9.13733952001481e-07, + "loss": 0.4424, + "step": 7985 + }, + { + "epoch": 4.335504885993485, + "grad_norm": 9.230465655882801, + "learning_rate": 9.122658483881264e-07, + "loss": 0.307, + "step": 7986 + }, + { + "epoch": 4.336047774158524, + "grad_norm": 9.602884171912605, + "learning_rate": 9.107988687602142e-07, + "loss": 0.4543, + "step": 7987 + }, + { + "epoch": 4.3365906623235615, + "grad_norm": 11.802560272642094, + "learning_rate": 9.09333013299184e-07, + "loss": 0.293, + "step": 7988 + }, + { + "epoch": 4.337133550488599, + "grad_norm": 19.98627599580886, + "learning_rate": 9.078682821863339e-07, + "loss": 0.5988, + "step": 7989 + }, + { + "epoch": 4.337676438653637, + "grad_norm": 14.01016302793737, + "learning_rate": 9.064046756028278e-07, + "loss": 0.3797, + "step": 7990 + }, + { + "epoch": 4.338219326818676, + "grad_norm": 10.54886354090915, + "learning_rate": 9.049421937296821e-07, + "loss": 0.2179, + "step": 7991 + }, + { + "epoch": 4.3387622149837135, + "grad_norm": 14.088344911170685, + "learning_rate": 9.034808367477844e-07, + "loss": 0.5632, + "step": 7992 + }, + { + "epoch": 4.339305103148751, + "grad_norm": 18.19756991978728, + "learning_rate": 9.0202060483787e-07, + "loss": 0.6184, + "step": 7993 + }, + { + "epoch": 4.339847991313789, + "grad_norm": 13.111955449734706, + "learning_rate": 9.005614981805522e-07, + "loss": 0.2811, + "step": 7994 + }, + { + "epoch": 4.340390879478828, + "grad_norm": 16.45443112800284, + "learning_rate": 8.991035169562879e-07, + "loss": 0.7305, + "step": 7995 + }, + { + "epoch": 4.3409337676438655, + "grad_norm": 15.613823352223783, + "learning_rate": 8.976466613454038e-07, + "loss": 0.6787, + "step": 7996 + }, + { + "epoch": 4.341476655808903, + "grad_norm": 12.421367112935704, + "learning_rate": 8.961909315280914e-07, + "loss": 0.3399, + "step": 7997 + }, + { + "epoch": 4.342019543973941, + "grad_norm": 12.609970065901182, + "learning_rate": 8.947363276843901e-07, + "loss": 0.3253, + "step": 7998 + }, + { + "epoch": 4.34256243213898, + "grad_norm": 10.696313488480198, + "learning_rate": 8.932828499942159e-07, + "loss": 0.3197, + "step": 7999 + }, + { + "epoch": 4.3431053203040175, + "grad_norm": 13.140351352844807, + "learning_rate": 8.918304986373283e-07, + "loss": 0.3766, + "step": 8000 + }, + { + "epoch": 4.343648208469055, + "grad_norm": 9.439332843485358, + "learning_rate": 8.903792737933626e-07, + "loss": 0.3762, + "step": 8001 + }, + { + "epoch": 4.344191096634093, + "grad_norm": 17.7660438004401, + "learning_rate": 8.889291756418073e-07, + "loss": 0.4339, + "step": 8002 + }, + { + "epoch": 4.344733984799132, + "grad_norm": 12.278395472072871, + "learning_rate": 8.874802043620112e-07, + "loss": 0.4535, + "step": 8003 + }, + { + "epoch": 4.3452768729641695, + "grad_norm": 16.55900970801957, + "learning_rate": 8.860323601331899e-07, + "loss": 0.511, + "step": 8004 + }, + { + "epoch": 4.345819761129207, + "grad_norm": 11.14928157377559, + "learning_rate": 8.84585643134408e-07, + "loss": 0.417, + "step": 8005 + }, + { + "epoch": 4.346362649294245, + "grad_norm": 21.55707350645272, + "learning_rate": 8.831400535446044e-07, + "loss": 0.5588, + "step": 8006 + }, + { + "epoch": 4.346905537459284, + "grad_norm": 14.787630844212787, + "learning_rate": 8.816955915425663e-07, + "loss": 0.4701, + "step": 8007 + }, + { + "epoch": 4.3474484256243215, + "grad_norm": 12.957973486859906, + "learning_rate": 8.802522573069505e-07, + "loss": 0.3898, + "step": 8008 + }, + { + "epoch": 4.347991313789359, + "grad_norm": 13.858932325098396, + "learning_rate": 8.788100510162712e-07, + "loss": 0.4319, + "step": 8009 + }, + { + "epoch": 4.348534201954397, + "grad_norm": 11.44820285977616, + "learning_rate": 8.773689728489021e-07, + "loss": 0.346, + "step": 8010 + }, + { + "epoch": 4.349077090119436, + "grad_norm": 13.32788926130813, + "learning_rate": 8.759290229830774e-07, + "loss": 0.5118, + "step": 8011 + }, + { + "epoch": 4.3496199782844736, + "grad_norm": 13.65313572883658, + "learning_rate": 8.744902015968926e-07, + "loss": 0.6354, + "step": 8012 + }, + { + "epoch": 4.350162866449511, + "grad_norm": 15.440111761472027, + "learning_rate": 8.730525088683028e-07, + "loss": 0.3655, + "step": 8013 + }, + { + "epoch": 4.350705754614549, + "grad_norm": 13.392699739564616, + "learning_rate": 8.71615944975126e-07, + "loss": 0.5795, + "step": 8014 + }, + { + "epoch": 4.351248642779588, + "grad_norm": 12.572118935109678, + "learning_rate": 8.701805100950378e-07, + "loss": 0.4298, + "step": 8015 + }, + { + "epoch": 4.351791530944626, + "grad_norm": 13.635858966392217, + "learning_rate": 8.687462044055739e-07, + "loss": 0.3448, + "step": 8016 + }, + { + "epoch": 4.352334419109663, + "grad_norm": 13.828043330695678, + "learning_rate": 8.673130280841335e-07, + "loss": 0.645, + "step": 8017 + }, + { + "epoch": 4.352877307274701, + "grad_norm": 15.111038785972921, + "learning_rate": 8.658809813079716e-07, + "loss": 0.451, + "step": 8018 + }, + { + "epoch": 4.35342019543974, + "grad_norm": 11.79984417087872, + "learning_rate": 8.644500642542086e-07, + "loss": 0.3503, + "step": 8019 + }, + { + "epoch": 4.353963083604778, + "grad_norm": 11.058205432853121, + "learning_rate": 8.63020277099822e-07, + "loss": 0.3616, + "step": 8020 + }, + { + "epoch": 4.354505971769815, + "grad_norm": 17.658740606783095, + "learning_rate": 8.61591620021649e-07, + "loss": 0.4659, + "step": 8021 + }, + { + "epoch": 4.355048859934853, + "grad_norm": 15.417869712043583, + "learning_rate": 8.601640931963873e-07, + "loss": 0.5888, + "step": 8022 + }, + { + "epoch": 4.355591748099892, + "grad_norm": 17.637210382444906, + "learning_rate": 8.587376968006011e-07, + "loss": 0.6608, + "step": 8023 + }, + { + "epoch": 4.35613463626493, + "grad_norm": 11.883245195971504, + "learning_rate": 8.573124310107028e-07, + "loss": 0.3882, + "step": 8024 + }, + { + "epoch": 4.356677524429967, + "grad_norm": 13.09142985069161, + "learning_rate": 8.558882960029758e-07, + "loss": 0.4295, + "step": 8025 + }, + { + "epoch": 4.357220412595005, + "grad_norm": 13.42588355451286, + "learning_rate": 8.544652919535568e-07, + "loss": 0.5258, + "step": 8026 + }, + { + "epoch": 4.357763300760044, + "grad_norm": 11.055862131666451, + "learning_rate": 8.530434190384463e-07, + "loss": 0.3189, + "step": 8027 + }, + { + "epoch": 4.358306188925082, + "grad_norm": 14.270581711188854, + "learning_rate": 8.516226774335046e-07, + "loss": 0.4594, + "step": 8028 + }, + { + "epoch": 4.358849077090119, + "grad_norm": 13.14559769122887, + "learning_rate": 8.502030673144479e-07, + "loss": 0.6237, + "step": 8029 + }, + { + "epoch": 4.359391965255157, + "grad_norm": 16.47710832181487, + "learning_rate": 8.487845888568625e-07, + "loss": 0.4926, + "step": 8030 + }, + { + "epoch": 4.359934853420196, + "grad_norm": 9.957560938239181, + "learning_rate": 8.473672422361801e-07, + "loss": 0.3599, + "step": 8031 + }, + { + "epoch": 4.360477741585234, + "grad_norm": 13.852815438372643, + "learning_rate": 8.459510276277083e-07, + "loss": 0.3684, + "step": 8032 + }, + { + "epoch": 4.361020629750271, + "grad_norm": 15.597466583644815, + "learning_rate": 8.445359452065982e-07, + "loss": 0.6425, + "step": 8033 + }, + { + "epoch": 4.361563517915309, + "grad_norm": 10.260784495869965, + "learning_rate": 8.431219951478775e-07, + "loss": 0.3258, + "step": 8034 + }, + { + "epoch": 4.362106406080348, + "grad_norm": 12.005771351192994, + "learning_rate": 8.41709177626423e-07, + "loss": 0.3723, + "step": 8035 + }, + { + "epoch": 4.362649294245386, + "grad_norm": 12.407275509701295, + "learning_rate": 8.402974928169705e-07, + "loss": 0.4967, + "step": 8036 + }, + { + "epoch": 4.363192182410423, + "grad_norm": 14.727533661174913, + "learning_rate": 8.388869408941247e-07, + "loss": 0.4133, + "step": 8037 + }, + { + "epoch": 4.363735070575461, + "grad_norm": 11.403317480094385, + "learning_rate": 8.374775220323406e-07, + "loss": 0.3032, + "step": 8038 + }, + { + "epoch": 4.3642779587405, + "grad_norm": 12.660090294532273, + "learning_rate": 8.360692364059409e-07, + "loss": 0.3985, + "step": 8039 + }, + { + "epoch": 4.364820846905538, + "grad_norm": 9.17954467960599, + "learning_rate": 8.346620841891029e-07, + "loss": 0.287, + "step": 8040 + }, + { + "epoch": 4.365363735070575, + "grad_norm": 16.649840062856516, + "learning_rate": 8.332560655558663e-07, + "loss": 0.5538, + "step": 8041 + }, + { + "epoch": 4.365906623235613, + "grad_norm": 12.11674130758851, + "learning_rate": 8.318511806801288e-07, + "loss": 0.6752, + "step": 8042 + }, + { + "epoch": 4.366449511400652, + "grad_norm": 15.055685550752939, + "learning_rate": 8.304474297356479e-07, + "loss": 0.4629, + "step": 8043 + }, + { + "epoch": 4.36699239956569, + "grad_norm": 13.344971903872295, + "learning_rate": 8.29044812896046e-07, + "loss": 0.3308, + "step": 8044 + }, + { + "epoch": 4.367535287730727, + "grad_norm": 10.304984524921624, + "learning_rate": 8.276433303347953e-07, + "loss": 0.7107, + "step": 8045 + }, + { + "epoch": 4.368078175895765, + "grad_norm": 11.722406609137764, + "learning_rate": 8.26242982225236e-07, + "loss": 0.5312, + "step": 8046 + }, + { + "epoch": 4.368621064060804, + "grad_norm": 9.83110831313041, + "learning_rate": 8.248437687405664e-07, + "loss": 0.2615, + "step": 8047 + }, + { + "epoch": 4.369163952225842, + "grad_norm": 12.152410682750114, + "learning_rate": 8.234456900538424e-07, + "loss": 0.3105, + "step": 8048 + }, + { + "epoch": 4.369706840390879, + "grad_norm": 13.31087133977307, + "learning_rate": 8.220487463379811e-07, + "loss": 0.4315, + "step": 8049 + }, + { + "epoch": 4.370249728555917, + "grad_norm": 11.94526177677569, + "learning_rate": 8.206529377657579e-07, + "loss": 0.3136, + "step": 8050 + }, + { + "epoch": 4.370792616720956, + "grad_norm": 16.337088129910562, + "learning_rate": 8.192582645098091e-07, + "loss": 0.6422, + "step": 8051 + }, + { + "epoch": 4.371335504885994, + "grad_norm": 15.612171822665877, + "learning_rate": 8.178647267426299e-07, + "loss": 0.4182, + "step": 8052 + }, + { + "epoch": 4.371878393051031, + "grad_norm": 15.057359339910816, + "learning_rate": 8.164723246365758e-07, + "loss": 0.5622, + "step": 8053 + }, + { + "epoch": 4.372421281216069, + "grad_norm": 14.162937228652032, + "learning_rate": 8.150810583638602e-07, + "loss": 0.4681, + "step": 8054 + }, + { + "epoch": 4.372964169381108, + "grad_norm": 15.27678092027012, + "learning_rate": 8.136909280965589e-07, + "loss": 0.5663, + "step": 8055 + }, + { + "epoch": 4.373507057546146, + "grad_norm": 9.416661735214964, + "learning_rate": 8.123019340066029e-07, + "loss": 0.4337, + "step": 8056 + }, + { + "epoch": 4.374049945711183, + "grad_norm": 8.853547844230004, + "learning_rate": 8.109140762657885e-07, + "loss": 0.2355, + "step": 8057 + }, + { + "epoch": 4.374592833876221, + "grad_norm": 13.636809573527607, + "learning_rate": 8.095273550457649e-07, + "loss": 0.3061, + "step": 8058 + }, + { + "epoch": 4.37513572204126, + "grad_norm": 10.074225429783628, + "learning_rate": 8.081417705180461e-07, + "loss": 0.5573, + "step": 8059 + }, + { + "epoch": 4.375678610206298, + "grad_norm": 16.626739722709576, + "learning_rate": 8.067573228540015e-07, + "loss": 0.8822, + "step": 8060 + }, + { + "epoch": 4.3762214983713354, + "grad_norm": 13.021514377648314, + "learning_rate": 8.053740122248665e-07, + "loss": 0.5488, + "step": 8061 + }, + { + "epoch": 4.376764386536373, + "grad_norm": 22.411620281238108, + "learning_rate": 8.039918388017231e-07, + "loss": 1.0347, + "step": 8062 + }, + { + "epoch": 4.377307274701412, + "grad_norm": 9.303857597914824, + "learning_rate": 8.026108027555302e-07, + "loss": 0.4273, + "step": 8063 + }, + { + "epoch": 4.37785016286645, + "grad_norm": 13.977505292581387, + "learning_rate": 8.012309042570887e-07, + "loss": 0.4826, + "step": 8064 + }, + { + "epoch": 4.3783930510314875, + "grad_norm": 16.681445518035993, + "learning_rate": 7.9985214347707e-07, + "loss": 0.6674, + "step": 8065 + }, + { + "epoch": 4.378935939196525, + "grad_norm": 15.765471997364271, + "learning_rate": 7.984745205860022e-07, + "loss": 0.4246, + "step": 8066 + }, + { + "epoch": 4.379478827361564, + "grad_norm": 12.65619134059525, + "learning_rate": 7.970980357542679e-07, + "loss": 0.4719, + "step": 8067 + }, + { + "epoch": 4.380021715526602, + "grad_norm": 14.558989497065397, + "learning_rate": 7.957226891521185e-07, + "loss": 0.3499, + "step": 8068 + }, + { + "epoch": 4.3805646036916395, + "grad_norm": 10.308079110981643, + "learning_rate": 7.943484809496549e-07, + "loss": 0.2406, + "step": 8069 + }, + { + "epoch": 4.381107491856677, + "grad_norm": 10.563122177925374, + "learning_rate": 7.929754113168442e-07, + "loss": 0.4711, + "step": 8070 + }, + { + "epoch": 4.381650380021716, + "grad_norm": 16.713706157595226, + "learning_rate": 7.916034804235062e-07, + "loss": 0.4097, + "step": 8071 + }, + { + "epoch": 4.382193268186754, + "grad_norm": 13.661917790603933, + "learning_rate": 7.902326884393275e-07, + "loss": 0.421, + "step": 8072 + }, + { + "epoch": 4.3827361563517915, + "grad_norm": 11.192296315079558, + "learning_rate": 7.888630355338489e-07, + "loss": 0.2375, + "step": 8073 + }, + { + "epoch": 4.383279044516829, + "grad_norm": 10.108216567015306, + "learning_rate": 7.874945218764685e-07, + "loss": 0.2637, + "step": 8074 + }, + { + "epoch": 4.383821932681868, + "grad_norm": 16.67098266690521, + "learning_rate": 7.861271476364496e-07, + "loss": 0.7568, + "step": 8075 + }, + { + "epoch": 4.384364820846906, + "grad_norm": 14.31792498162585, + "learning_rate": 7.847609129829081e-07, + "loss": 0.5722, + "step": 8076 + }, + { + "epoch": 4.3849077090119435, + "grad_norm": 11.783645413114316, + "learning_rate": 7.833958180848267e-07, + "loss": 0.4316, + "step": 8077 + }, + { + "epoch": 4.385450597176981, + "grad_norm": 12.328500975614492, + "learning_rate": 7.820318631110357e-07, + "loss": 0.3031, + "step": 8078 + }, + { + "epoch": 4.38599348534202, + "grad_norm": 15.932687060138331, + "learning_rate": 7.806690482302371e-07, + "loss": 0.5638, + "step": 8079 + }, + { + "epoch": 4.386536373507058, + "grad_norm": 11.335548440871465, + "learning_rate": 7.793073736109846e-07, + "loss": 0.2642, + "step": 8080 + }, + { + "epoch": 4.3870792616720955, + "grad_norm": 14.219082798674243, + "learning_rate": 7.779468394216905e-07, + "loss": 0.5914, + "step": 8081 + }, + { + "epoch": 4.387622149837133, + "grad_norm": 11.171865534559577, + "learning_rate": 7.765874458306299e-07, + "loss": 0.3573, + "step": 8082 + }, + { + "epoch": 4.388165038002172, + "grad_norm": 14.949429459159015, + "learning_rate": 7.752291930059341e-07, + "loss": 0.491, + "step": 8083 + }, + { + "epoch": 4.38870792616721, + "grad_norm": 11.682058767447176, + "learning_rate": 7.738720811155931e-07, + "loss": 0.4278, + "step": 8084 + }, + { + "epoch": 4.3892508143322475, + "grad_norm": 9.720345707653586, + "learning_rate": 7.725161103274581e-07, + "loss": 0.3931, + "step": 8085 + }, + { + "epoch": 4.389793702497285, + "grad_norm": 13.696977621677252, + "learning_rate": 7.711612808092372e-07, + "loss": 0.4812, + "step": 8086 + }, + { + "epoch": 4.390336590662324, + "grad_norm": 13.326736087170834, + "learning_rate": 7.698075927284964e-07, + "loss": 0.5377, + "step": 8087 + }, + { + "epoch": 4.390879478827362, + "grad_norm": 13.222977582929962, + "learning_rate": 7.684550462526641e-07, + "loss": 0.4528, + "step": 8088 + }, + { + "epoch": 4.3914223669923995, + "grad_norm": 15.11572868941985, + "learning_rate": 7.671036415490252e-07, + "loss": 0.6123, + "step": 8089 + }, + { + "epoch": 4.391965255157437, + "grad_norm": 13.903887430270185, + "learning_rate": 7.657533787847215e-07, + "loss": 0.493, + "step": 8090 + }, + { + "epoch": 4.392508143322476, + "grad_norm": 13.728638124426194, + "learning_rate": 7.644042581267585e-07, + "loss": 0.5071, + "step": 8091 + }, + { + "epoch": 4.393051031487514, + "grad_norm": 16.05244491657159, + "learning_rate": 7.630562797419949e-07, + "loss": 0.4212, + "step": 8092 + }, + { + "epoch": 4.3935939196525515, + "grad_norm": 11.275459672619295, + "learning_rate": 7.617094437971529e-07, + "loss": 0.2391, + "step": 8093 + }, + { + "epoch": 4.394136807817589, + "grad_norm": 15.521537268341413, + "learning_rate": 7.603637504588112e-07, + "loss": 0.4127, + "step": 8094 + }, + { + "epoch": 4.394679695982628, + "grad_norm": 11.264052644917689, + "learning_rate": 7.59019199893406e-07, + "loss": 0.3287, + "step": 8095 + }, + { + "epoch": 4.395222584147666, + "grad_norm": 11.628986278608336, + "learning_rate": 7.576757922672339e-07, + "loss": 0.3109, + "step": 8096 + }, + { + "epoch": 4.3957654723127035, + "grad_norm": 12.05990932087969, + "learning_rate": 7.56333527746449e-07, + "loss": 0.4798, + "step": 8097 + }, + { + "epoch": 4.396308360477741, + "grad_norm": 11.758858029678734, + "learning_rate": 7.54992406497067e-07, + "loss": 0.3736, + "step": 8098 + }, + { + "epoch": 4.39685124864278, + "grad_norm": 16.559994897045023, + "learning_rate": 7.536524286849578e-07, + "loss": 0.5436, + "step": 8099 + }, + { + "epoch": 4.397394136807818, + "grad_norm": 12.81236009916768, + "learning_rate": 7.523135944758519e-07, + "loss": 0.5033, + "step": 8100 + }, + { + "epoch": 4.3979370249728555, + "grad_norm": 11.076059660720759, + "learning_rate": 7.509759040353415e-07, + "loss": 0.337, + "step": 8101 + }, + { + "epoch": 4.398479913137893, + "grad_norm": 9.275703031226358, + "learning_rate": 7.496393575288708e-07, + "loss": 0.2984, + "step": 8102 + }, + { + "epoch": 4.399022801302932, + "grad_norm": 15.86969949487026, + "learning_rate": 7.483039551217475e-07, + "loss": 0.5653, + "step": 8103 + }, + { + "epoch": 4.39956568946797, + "grad_norm": 12.679331406935141, + "learning_rate": 7.469696969791351e-07, + "loss": 0.3381, + "step": 8104 + }, + { + "epoch": 4.400108577633008, + "grad_norm": 14.791912312085929, + "learning_rate": 7.456365832660572e-07, + "loss": 0.3494, + "step": 8105 + }, + { + "epoch": 4.400651465798045, + "grad_norm": 10.790537111566087, + "learning_rate": 7.443046141473986e-07, + "loss": 0.3024, + "step": 8106 + }, + { + "epoch": 4.401194353963084, + "grad_norm": 16.74012144354708, + "learning_rate": 7.429737897878942e-07, + "loss": 0.7146, + "step": 8107 + }, + { + "epoch": 4.401737242128122, + "grad_norm": 18.112779551139106, + "learning_rate": 7.416441103521476e-07, + "loss": 0.787, + "step": 8108 + }, + { + "epoch": 4.40228013029316, + "grad_norm": 18.192800558391657, + "learning_rate": 7.403155760046099e-07, + "loss": 0.6249, + "step": 8109 + }, + { + "epoch": 4.402823018458197, + "grad_norm": 12.394483609056916, + "learning_rate": 7.389881869096027e-07, + "loss": 0.3235, + "step": 8110 + }, + { + "epoch": 4.403365906623236, + "grad_norm": 12.486317574320127, + "learning_rate": 7.376619432312926e-07, + "loss": 0.4458, + "step": 8111 + }, + { + "epoch": 4.403908794788274, + "grad_norm": 11.546942923476625, + "learning_rate": 7.363368451337183e-07, + "loss": 0.2904, + "step": 8112 + }, + { + "epoch": 4.404451682953312, + "grad_norm": 11.661763155608599, + "learning_rate": 7.350128927807676e-07, + "loss": 0.6103, + "step": 8113 + }, + { + "epoch": 4.404994571118349, + "grad_norm": 12.401624870044014, + "learning_rate": 7.336900863361851e-07, + "loss": 0.5543, + "step": 8114 + }, + { + "epoch": 4.405537459283388, + "grad_norm": 14.577278093491197, + "learning_rate": 7.323684259635855e-07, + "loss": 0.5578, + "step": 8115 + }, + { + "epoch": 4.406080347448426, + "grad_norm": 12.374504432300087, + "learning_rate": 7.310479118264247e-07, + "loss": 0.4786, + "step": 8116 + }, + { + "epoch": 4.406623235613464, + "grad_norm": 13.53250221886951, + "learning_rate": 7.297285440880331e-07, + "loss": 0.5691, + "step": 8117 + }, + { + "epoch": 4.407166123778501, + "grad_norm": 15.930406260794717, + "learning_rate": 7.284103229115891e-07, + "loss": 0.4541, + "step": 8118 + }, + { + "epoch": 4.40770901194354, + "grad_norm": 12.90844460573453, + "learning_rate": 7.270932484601345e-07, + "loss": 0.43, + "step": 8119 + }, + { + "epoch": 4.408251900108578, + "grad_norm": 11.272038151188799, + "learning_rate": 7.257773208965646e-07, + "loss": 0.5185, + "step": 8120 + }, + { + "epoch": 4.408794788273616, + "grad_norm": 17.139510748072922, + "learning_rate": 7.24462540383637e-07, + "loss": 0.6829, + "step": 8121 + }, + { + "epoch": 4.409337676438653, + "grad_norm": 12.62321799336504, + "learning_rate": 7.231489070839648e-07, + "loss": 0.3177, + "step": 8122 + }, + { + "epoch": 4.409880564603692, + "grad_norm": 11.177325186499452, + "learning_rate": 7.218364211600215e-07, + "loss": 0.3812, + "step": 8123 + }, + { + "epoch": 4.41042345276873, + "grad_norm": 16.25406531253722, + "learning_rate": 7.205250827741361e-07, + "loss": 0.4482, + "step": 8124 + }, + { + "epoch": 4.410966340933768, + "grad_norm": 16.76577388765637, + "learning_rate": 7.192148920884989e-07, + "loss": 0.4272, + "step": 8125 + }, + { + "epoch": 4.411509229098805, + "grad_norm": 14.687321011203691, + "learning_rate": 7.179058492651547e-07, + "loss": 0.5207, + "step": 8126 + }, + { + "epoch": 4.412052117263844, + "grad_norm": 11.956370304080279, + "learning_rate": 7.165979544660085e-07, + "loss": 0.3902, + "step": 8127 + }, + { + "epoch": 4.412595005428882, + "grad_norm": 16.74860718331433, + "learning_rate": 7.152912078528229e-07, + "loss": 0.4425, + "step": 8128 + }, + { + "epoch": 4.41313789359392, + "grad_norm": 12.718796959860708, + "learning_rate": 7.139856095872177e-07, + "loss": 0.4485, + "step": 8129 + }, + { + "epoch": 4.413680781758957, + "grad_norm": 11.351426129816947, + "learning_rate": 7.126811598306726e-07, + "loss": 0.4284, + "step": 8130 + }, + { + "epoch": 4.414223669923996, + "grad_norm": 15.528891800944992, + "learning_rate": 7.11377858744522e-07, + "loss": 0.4391, + "step": 8131 + }, + { + "epoch": 4.414766558089034, + "grad_norm": 15.062549425859391, + "learning_rate": 7.100757064899644e-07, + "loss": 0.4085, + "step": 8132 + }, + { + "epoch": 4.415309446254072, + "grad_norm": 13.147141392832742, + "learning_rate": 7.087747032280479e-07, + "loss": 0.5011, + "step": 8133 + }, + { + "epoch": 4.415852334419109, + "grad_norm": 13.452451139567685, + "learning_rate": 7.074748491196837e-07, + "loss": 0.5419, + "step": 8134 + }, + { + "epoch": 4.416395222584148, + "grad_norm": 15.154494684843028, + "learning_rate": 7.061761443256409e-07, + "loss": 0.4214, + "step": 8135 + }, + { + "epoch": 4.416938110749186, + "grad_norm": 14.435292009300346, + "learning_rate": 7.048785890065446e-07, + "loss": 0.3093, + "step": 8136 + }, + { + "epoch": 4.417480998914224, + "grad_norm": 14.372369930518246, + "learning_rate": 7.035821833228785e-07, + "loss": 0.4875, + "step": 8137 + }, + { + "epoch": 4.418023887079261, + "grad_norm": 13.382996066133376, + "learning_rate": 7.02286927434983e-07, + "loss": 0.5495, + "step": 8138 + }, + { + "epoch": 4.4185667752443, + "grad_norm": 18.053750531135215, + "learning_rate": 7.009928215030615e-07, + "loss": 0.4188, + "step": 8139 + }, + { + "epoch": 4.419109663409338, + "grad_norm": 16.22341652670191, + "learning_rate": 6.996998656871646e-07, + "loss": 0.6233, + "step": 8140 + }, + { + "epoch": 4.419652551574376, + "grad_norm": 12.414389082202195, + "learning_rate": 6.984080601472143e-07, + "loss": 0.441, + "step": 8141 + }, + { + "epoch": 4.420195439739413, + "grad_norm": 15.971472074811146, + "learning_rate": 6.971174050429786e-07, + "loss": 0.4502, + "step": 8142 + }, + { + "epoch": 4.420738327904452, + "grad_norm": 8.57673948818228, + "learning_rate": 6.958279005340874e-07, + "loss": 0.3876, + "step": 8143 + }, + { + "epoch": 4.42128121606949, + "grad_norm": 14.907293994220137, + "learning_rate": 6.94539546780032e-07, + "loss": 0.4499, + "step": 8144 + }, + { + "epoch": 4.421824104234528, + "grad_norm": 11.784072399397415, + "learning_rate": 6.93252343940154e-07, + "loss": 0.4506, + "step": 8145 + }, + { + "epoch": 4.422366992399565, + "grad_norm": 13.773309280609004, + "learning_rate": 6.919662921736614e-07, + "loss": 0.6268, + "step": 8146 + }, + { + "epoch": 4.422909880564604, + "grad_norm": 14.034219630419418, + "learning_rate": 6.906813916396104e-07, + "loss": 0.557, + "step": 8147 + }, + { + "epoch": 4.423452768729642, + "grad_norm": 11.54689354028201, + "learning_rate": 6.89397642496924e-07, + "loss": 0.293, + "step": 8148 + }, + { + "epoch": 4.42399565689468, + "grad_norm": 12.196466085866508, + "learning_rate": 6.881150449043727e-07, + "loss": 0.4063, + "step": 8149 + }, + { + "epoch": 4.424538545059717, + "grad_norm": 13.943110002700577, + "learning_rate": 6.868335990205965e-07, + "loss": 0.4629, + "step": 8150 + }, + { + "epoch": 4.425081433224756, + "grad_norm": 11.136718980358994, + "learning_rate": 6.855533050040841e-07, + "loss": 0.353, + "step": 8151 + }, + { + "epoch": 4.425624321389794, + "grad_norm": 12.815867518018083, + "learning_rate": 6.84274163013181e-07, + "loss": 0.3692, + "step": 8152 + }, + { + "epoch": 4.426167209554832, + "grad_norm": 15.673409819920156, + "learning_rate": 6.829961732060997e-07, + "loss": 0.7404, + "step": 8153 + }, + { + "epoch": 4.4267100977198695, + "grad_norm": 15.185723434348903, + "learning_rate": 6.81719335740898e-07, + "loss": 0.6198, + "step": 8154 + }, + { + "epoch": 4.427252985884908, + "grad_norm": 13.985897143479551, + "learning_rate": 6.804436507755008e-07, + "loss": 0.4197, + "step": 8155 + }, + { + "epoch": 4.427795874049946, + "grad_norm": 15.277569295336207, + "learning_rate": 6.791691184676863e-07, + "loss": 0.5533, + "step": 8156 + }, + { + "epoch": 4.428338762214984, + "grad_norm": 16.725284925242832, + "learning_rate": 6.778957389750907e-07, + "loss": 0.6159, + "step": 8157 + }, + { + "epoch": 4.4288816503800215, + "grad_norm": 14.027576538249749, + "learning_rate": 6.766235124552079e-07, + "loss": 0.3277, + "step": 8158 + }, + { + "epoch": 4.42942453854506, + "grad_norm": 9.599302156075451, + "learning_rate": 6.753524390653876e-07, + "loss": 0.2496, + "step": 8159 + }, + { + "epoch": 4.429967426710098, + "grad_norm": 15.48284468466842, + "learning_rate": 6.740825189628386e-07, + "loss": 0.3937, + "step": 8160 + }, + { + "epoch": 4.430510314875136, + "grad_norm": 13.417678670406026, + "learning_rate": 6.728137523046274e-07, + "loss": 0.4335, + "step": 8161 + }, + { + "epoch": 4.4310532030401735, + "grad_norm": 16.002263273785378, + "learning_rate": 6.715461392476763e-07, + "loss": 0.7239, + "step": 8162 + }, + { + "epoch": 4.431596091205212, + "grad_norm": 13.513757533100362, + "learning_rate": 6.702796799487665e-07, + "loss": 0.5011, + "step": 8163 + }, + { + "epoch": 4.43213897937025, + "grad_norm": 17.068661392730355, + "learning_rate": 6.69014374564535e-07, + "loss": 0.5128, + "step": 8164 + }, + { + "epoch": 4.432681867535288, + "grad_norm": 16.016384575594365, + "learning_rate": 6.677502232514777e-07, + "loss": 0.2884, + "step": 8165 + }, + { + "epoch": 4.4332247557003255, + "grad_norm": 14.947992178628231, + "learning_rate": 6.66487226165945e-07, + "loss": 0.603, + "step": 8166 + }, + { + "epoch": 4.433767643865364, + "grad_norm": 14.940395600883463, + "learning_rate": 6.652253834641487e-07, + "loss": 0.4589, + "step": 8167 + }, + { + "epoch": 4.434310532030402, + "grad_norm": 10.553864429327028, + "learning_rate": 6.639646953021539e-07, + "loss": 0.2977, + "step": 8168 + }, + { + "epoch": 4.43485342019544, + "grad_norm": 13.081302650262185, + "learning_rate": 6.627051618358848e-07, + "loss": 0.5568, + "step": 8169 + }, + { + "epoch": 4.4353963083604775, + "grad_norm": 11.615908937272211, + "learning_rate": 6.614467832211224e-07, + "loss": 0.3839, + "step": 8170 + }, + { + "epoch": 4.435939196525516, + "grad_norm": 13.81874812862262, + "learning_rate": 6.601895596135055e-07, + "loss": 0.4983, + "step": 8171 + }, + { + "epoch": 4.436482084690554, + "grad_norm": 10.994119413078606, + "learning_rate": 6.589334911685297e-07, + "loss": 0.2608, + "step": 8172 + }, + { + "epoch": 4.437024972855592, + "grad_norm": 13.512853171310855, + "learning_rate": 6.576785780415474e-07, + "loss": 0.3852, + "step": 8173 + }, + { + "epoch": 4.4375678610206295, + "grad_norm": 13.099445891838716, + "learning_rate": 6.564248203877677e-07, + "loss": 0.3673, + "step": 8174 + }, + { + "epoch": 4.438110749185668, + "grad_norm": 13.344678881473703, + "learning_rate": 6.551722183622578e-07, + "loss": 0.3684, + "step": 8175 + }, + { + "epoch": 4.438653637350706, + "grad_norm": 11.89837504148065, + "learning_rate": 6.539207721199392e-07, + "loss": 0.3663, + "step": 8176 + }, + { + "epoch": 4.439196525515744, + "grad_norm": 16.57103010045736, + "learning_rate": 6.526704818155983e-07, + "loss": 0.3226, + "step": 8177 + }, + { + "epoch": 4.4397394136807815, + "grad_norm": 13.302285956970346, + "learning_rate": 6.514213476038667e-07, + "loss": 0.5893, + "step": 8178 + }, + { + "epoch": 4.440282301845819, + "grad_norm": 16.633046645169372, + "learning_rate": 6.501733696392454e-07, + "loss": 0.3461, + "step": 8179 + }, + { + "epoch": 4.440825190010858, + "grad_norm": 11.411100884112152, + "learning_rate": 6.489265480760798e-07, + "loss": 0.3867, + "step": 8180 + }, + { + "epoch": 4.441368078175896, + "grad_norm": 11.251827734607014, + "learning_rate": 6.476808830685855e-07, + "loss": 0.4055, + "step": 8181 + }, + { + "epoch": 4.4419109663409335, + "grad_norm": 12.272578715881838, + "learning_rate": 6.464363747708236e-07, + "loss": 0.489, + "step": 8182 + }, + { + "epoch": 4.442453854505972, + "grad_norm": 10.150536439617065, + "learning_rate": 6.451930233367154e-07, + "loss": 0.2645, + "step": 8183 + }, + { + "epoch": 4.44299674267101, + "grad_norm": 16.617783744787634, + "learning_rate": 6.43950828920048e-07, + "loss": 0.452, + "step": 8184 + }, + { + "epoch": 4.443539630836048, + "grad_norm": 13.318664745403998, + "learning_rate": 6.427097916744496e-07, + "loss": 0.2832, + "step": 8185 + }, + { + "epoch": 4.4440825190010855, + "grad_norm": 14.768021540885634, + "learning_rate": 6.414699117534207e-07, + "loss": 0.6652, + "step": 8186 + }, + { + "epoch": 4.444625407166123, + "grad_norm": 12.248982189139824, + "learning_rate": 6.402311893103052e-07, + "loss": 0.3722, + "step": 8187 + }, + { + "epoch": 4.445168295331162, + "grad_norm": 17.38625490353128, + "learning_rate": 6.389936244983153e-07, + "loss": 0.4537, + "step": 8188 + }, + { + "epoch": 4.4457111834962, + "grad_norm": 11.805512735858951, + "learning_rate": 6.377572174705127e-07, + "loss": 0.3552, + "step": 8189 + }, + { + "epoch": 4.4462540716612375, + "grad_norm": 11.017672732553946, + "learning_rate": 6.365219683798197e-07, + "loss": 0.3821, + "step": 8190 + }, + { + "epoch": 4.446796959826276, + "grad_norm": 13.850511847705805, + "learning_rate": 6.352878773790128e-07, + "loss": 0.3766, + "step": 8191 + }, + { + "epoch": 4.447339847991314, + "grad_norm": 11.980286969929338, + "learning_rate": 6.340549446207245e-07, + "loss": 0.3077, + "step": 8192 + }, + { + "epoch": 4.447882736156352, + "grad_norm": 12.593566275833842, + "learning_rate": 6.328231702574483e-07, + "loss": 0.3866, + "step": 8193 + }, + { + "epoch": 4.4484256243213895, + "grad_norm": 16.906397392736164, + "learning_rate": 6.315925544415324e-07, + "loss": 0.6914, + "step": 8194 + }, + { + "epoch": 4.448968512486427, + "grad_norm": 13.626782408196018, + "learning_rate": 6.303630973251795e-07, + "loss": 0.3687, + "step": 8195 + }, + { + "epoch": 4.449511400651466, + "grad_norm": 12.864518300164796, + "learning_rate": 6.291347990604524e-07, + "loss": 0.3668, + "step": 8196 + }, + { + "epoch": 4.450054288816504, + "grad_norm": 14.217617041842738, + "learning_rate": 6.279076597992683e-07, + "loss": 0.4862, + "step": 8197 + }, + { + "epoch": 4.450597176981542, + "grad_norm": 10.62667862211288, + "learning_rate": 6.266816796934016e-07, + "loss": 0.5427, + "step": 8198 + }, + { + "epoch": 4.45114006514658, + "grad_norm": 13.714935436924003, + "learning_rate": 6.254568588944832e-07, + "loss": 0.5349, + "step": 8199 + }, + { + "epoch": 4.451682953311618, + "grad_norm": 10.890652969571276, + "learning_rate": 6.242331975540017e-07, + "loss": 0.3663, + "step": 8200 + }, + { + "epoch": 4.452225841476656, + "grad_norm": 15.543319095934837, + "learning_rate": 6.230106958233006e-07, + "loss": 0.4374, + "step": 8201 + }, + { + "epoch": 4.452768729641694, + "grad_norm": 13.007089382527154, + "learning_rate": 6.217893538535813e-07, + "loss": 0.5068, + "step": 8202 + }, + { + "epoch": 4.453311617806731, + "grad_norm": 10.246126043767955, + "learning_rate": 6.205691717959017e-07, + "loss": 0.2651, + "step": 8203 + }, + { + "epoch": 4.45385450597177, + "grad_norm": 14.357913078114876, + "learning_rate": 6.193501498011756e-07, + "loss": 0.7913, + "step": 8204 + }, + { + "epoch": 4.454397394136808, + "grad_norm": 9.017306171160634, + "learning_rate": 6.181322880201734e-07, + "loss": 0.3367, + "step": 8205 + }, + { + "epoch": 4.454940282301846, + "grad_norm": 9.76917208852768, + "learning_rate": 6.169155866035226e-07, + "loss": 0.3176, + "step": 8206 + }, + { + "epoch": 4.455483170466884, + "grad_norm": 12.851456706342148, + "learning_rate": 6.15700045701706e-07, + "loss": 0.363, + "step": 8207 + }, + { + "epoch": 4.456026058631922, + "grad_norm": 12.21030677086537, + "learning_rate": 6.144856654650644e-07, + "loss": 0.3758, + "step": 8208 + }, + { + "epoch": 4.45656894679696, + "grad_norm": 12.105703346136071, + "learning_rate": 6.132724460437923e-07, + "loss": 0.4042, + "step": 8209 + }, + { + "epoch": 4.457111834961998, + "grad_norm": 9.134328872162778, + "learning_rate": 6.120603875879472e-07, + "loss": 0.2771, + "step": 8210 + }, + { + "epoch": 4.457654723127035, + "grad_norm": 11.516060563585224, + "learning_rate": 6.108494902474349e-07, + "loss": 0.2969, + "step": 8211 + }, + { + "epoch": 4.458197611292074, + "grad_norm": 12.99985841750232, + "learning_rate": 6.096397541720201e-07, + "loss": 0.4357, + "step": 8212 + }, + { + "epoch": 4.458740499457112, + "grad_norm": 17.46011030710575, + "learning_rate": 6.084311795113273e-07, + "loss": 0.3445, + "step": 8213 + }, + { + "epoch": 4.45928338762215, + "grad_norm": 13.1182184160893, + "learning_rate": 6.072237664148328e-07, + "loss": 0.4747, + "step": 8214 + }, + { + "epoch": 4.459826275787188, + "grad_norm": 12.980936251031281, + "learning_rate": 6.060175150318759e-07, + "loss": 0.3913, + "step": 8215 + }, + { + "epoch": 4.460369163952226, + "grad_norm": 17.440326310172946, + "learning_rate": 6.048124255116417e-07, + "loss": 0.6855, + "step": 8216 + }, + { + "epoch": 4.460912052117264, + "grad_norm": 9.688408464276451, + "learning_rate": 6.036084980031831e-07, + "loss": 0.258, + "step": 8217 + }, + { + "epoch": 4.461454940282302, + "grad_norm": 12.394258791386804, + "learning_rate": 6.024057326553978e-07, + "loss": 0.2525, + "step": 8218 + }, + { + "epoch": 4.461997828447339, + "grad_norm": 22.30304317136707, + "learning_rate": 6.012041296170523e-07, + "loss": 0.9278, + "step": 8219 + }, + { + "epoch": 4.462540716612378, + "grad_norm": 15.580394413629943, + "learning_rate": 6.000036890367577e-07, + "loss": 0.5253, + "step": 8220 + }, + { + "epoch": 4.463083604777416, + "grad_norm": 12.8104075419022, + "learning_rate": 5.988044110629864e-07, + "loss": 0.4224, + "step": 8221 + }, + { + "epoch": 4.463626492942454, + "grad_norm": 15.277018990574911, + "learning_rate": 5.97606295844072e-07, + "loss": 0.4786, + "step": 8222 + }, + { + "epoch": 4.464169381107492, + "grad_norm": 10.424466200619575, + "learning_rate": 5.964093435281937e-07, + "loss": 0.4286, + "step": 8223 + }, + { + "epoch": 4.46471226927253, + "grad_norm": 10.219964992820138, + "learning_rate": 5.952135542633975e-07, + "loss": 0.4709, + "step": 8224 + }, + { + "epoch": 4.465255157437568, + "grad_norm": 15.141978803020056, + "learning_rate": 5.94018928197575e-07, + "loss": 0.5745, + "step": 8225 + }, + { + "epoch": 4.465798045602606, + "grad_norm": 12.814624758233451, + "learning_rate": 5.928254654784837e-07, + "loss": 0.451, + "step": 8226 + }, + { + "epoch": 4.466340933767643, + "grad_norm": 11.052255193283386, + "learning_rate": 5.916331662537322e-07, + "loss": 0.353, + "step": 8227 + }, + { + "epoch": 4.466883821932682, + "grad_norm": 10.286265305406184, + "learning_rate": 5.904420306707848e-07, + "loss": 0.2147, + "step": 8228 + }, + { + "epoch": 4.46742671009772, + "grad_norm": 8.65857853802515, + "learning_rate": 5.892520588769646e-07, + "loss": 0.2347, + "step": 8229 + }, + { + "epoch": 4.467969598262758, + "grad_norm": 12.354192118862283, + "learning_rate": 5.880632510194473e-07, + "loss": 0.4059, + "step": 8230 + }, + { + "epoch": 4.468512486427796, + "grad_norm": 11.57115618341546, + "learning_rate": 5.868756072452686e-07, + "loss": 0.3671, + "step": 8231 + }, + { + "epoch": 4.469055374592834, + "grad_norm": 12.575647382464236, + "learning_rate": 5.856891277013154e-07, + "loss": 0.557, + "step": 8232 + }, + { + "epoch": 4.469598262757872, + "grad_norm": 15.538783933288979, + "learning_rate": 5.84503812534335e-07, + "loss": 0.3137, + "step": 8233 + }, + { + "epoch": 4.47014115092291, + "grad_norm": 10.745613891988105, + "learning_rate": 5.833196618909309e-07, + "loss": 0.243, + "step": 8234 + }, + { + "epoch": 4.470684039087947, + "grad_norm": 14.21692386565347, + "learning_rate": 5.821366759175573e-07, + "loss": 0.3973, + "step": 8235 + }, + { + "epoch": 4.471226927252986, + "grad_norm": 14.050911061509822, + "learning_rate": 5.809548547605304e-07, + "loss": 0.4404, + "step": 8236 + }, + { + "epoch": 4.471769815418024, + "grad_norm": 13.715645322282858, + "learning_rate": 5.797741985660188e-07, + "loss": 0.4384, + "step": 8237 + }, + { + "epoch": 4.472312703583062, + "grad_norm": 14.288594139401848, + "learning_rate": 5.78594707480048e-07, + "loss": 0.5333, + "step": 8238 + }, + { + "epoch": 4.4728555917481, + "grad_norm": 8.76616154408981, + "learning_rate": 5.77416381648499e-07, + "loss": 0.3037, + "step": 8239 + }, + { + "epoch": 4.473398479913138, + "grad_norm": 19.474660708412426, + "learning_rate": 5.762392212171086e-07, + "loss": 0.6935, + "step": 8240 + }, + { + "epoch": 4.473941368078176, + "grad_norm": 10.918792962612473, + "learning_rate": 5.750632263314715e-07, + "loss": 0.5132, + "step": 8241 + }, + { + "epoch": 4.474484256243214, + "grad_norm": 12.225987090529173, + "learning_rate": 5.738883971370357e-07, + "loss": 0.3604, + "step": 8242 + }, + { + "epoch": 4.4750271444082514, + "grad_norm": 15.19011773226198, + "learning_rate": 5.72714733779105e-07, + "loss": 0.5368, + "step": 8243 + }, + { + "epoch": 4.47557003257329, + "grad_norm": 15.315623852151898, + "learning_rate": 5.715422364028423e-07, + "loss": 0.6617, + "step": 8244 + }, + { + "epoch": 4.476112920738328, + "grad_norm": 14.565005688916782, + "learning_rate": 5.703709051532613e-07, + "loss": 0.3753, + "step": 8245 + }, + { + "epoch": 4.476655808903366, + "grad_norm": 13.306086406222777, + "learning_rate": 5.692007401752353e-07, + "loss": 0.297, + "step": 8246 + }, + { + "epoch": 4.477198697068404, + "grad_norm": 11.685314077983634, + "learning_rate": 5.680317416134917e-07, + "loss": 0.3943, + "step": 8247 + }, + { + "epoch": 4.477741585233442, + "grad_norm": 15.78610523944711, + "learning_rate": 5.668639096126172e-07, + "loss": 0.568, + "step": 8248 + }, + { + "epoch": 4.47828447339848, + "grad_norm": 17.678808024002272, + "learning_rate": 5.65697244317045e-07, + "loss": 0.5074, + "step": 8249 + }, + { + "epoch": 4.478827361563518, + "grad_norm": 13.613139148009028, + "learning_rate": 5.645317458710775e-07, + "loss": 0.4749, + "step": 8250 + }, + { + "epoch": 4.4793702497285555, + "grad_norm": 11.810323226840032, + "learning_rate": 5.633674144188594e-07, + "loss": 0.3615, + "step": 8251 + }, + { + "epoch": 4.479913137893594, + "grad_norm": 12.479823316808458, + "learning_rate": 5.62204250104399e-07, + "loss": 0.3417, + "step": 8252 + }, + { + "epoch": 4.480456026058632, + "grad_norm": 10.507991166659478, + "learning_rate": 5.610422530715597e-07, + "loss": 0.2787, + "step": 8253 + }, + { + "epoch": 4.48099891422367, + "grad_norm": 11.553633131438758, + "learning_rate": 5.59881423464056e-07, + "loss": 0.304, + "step": 8254 + }, + { + "epoch": 4.481541802388708, + "grad_norm": 12.545712767514072, + "learning_rate": 5.587217614254658e-07, + "loss": 0.5601, + "step": 8255 + }, + { + "epoch": 4.482084690553746, + "grad_norm": 14.917437145721351, + "learning_rate": 5.575632670992126e-07, + "loss": 0.6768, + "step": 8256 + }, + { + "epoch": 4.482627578718784, + "grad_norm": 20.337480802460437, + "learning_rate": 5.56405940628586e-07, + "loss": 0.602, + "step": 8257 + }, + { + "epoch": 4.483170466883822, + "grad_norm": 11.359050219678316, + "learning_rate": 5.552497821567216e-07, + "loss": 0.3441, + "step": 8258 + }, + { + "epoch": 4.4837133550488595, + "grad_norm": 14.281903579588091, + "learning_rate": 5.540947918266171e-07, + "loss": 0.3607, + "step": 8259 + }, + { + "epoch": 4.484256243213898, + "grad_norm": 10.394969225634863, + "learning_rate": 5.52940969781125e-07, + "loss": 0.2491, + "step": 8260 + }, + { + "epoch": 4.484799131378936, + "grad_norm": 15.233116056219048, + "learning_rate": 5.517883161629478e-07, + "loss": 0.4715, + "step": 8261 + }, + { + "epoch": 4.485342019543974, + "grad_norm": 13.370733413078632, + "learning_rate": 5.506368311146526e-07, + "loss": 0.5931, + "step": 8262 + }, + { + "epoch": 4.485884907709012, + "grad_norm": 11.613070671019594, + "learning_rate": 5.49486514778651e-07, + "loss": 0.3663, + "step": 8263 + }, + { + "epoch": 4.48642779587405, + "grad_norm": 12.305931901351924, + "learning_rate": 5.483373672972215e-07, + "loss": 0.3761, + "step": 8264 + }, + { + "epoch": 4.486970684039088, + "grad_norm": 11.00494181177833, + "learning_rate": 5.471893888124891e-07, + "loss": 0.3874, + "step": 8265 + }, + { + "epoch": 4.487513572204126, + "grad_norm": 12.920317758044833, + "learning_rate": 5.460425794664381e-07, + "loss": 0.5323, + "step": 8266 + }, + { + "epoch": 4.4880564603691635, + "grad_norm": 16.420345201613245, + "learning_rate": 5.448969394009096e-07, + "loss": 0.4206, + "step": 8267 + }, + { + "epoch": 4.488599348534202, + "grad_norm": 14.573525619555683, + "learning_rate": 5.437524687575957e-07, + "loss": 0.4514, + "step": 8268 + }, + { + "epoch": 4.48914223669924, + "grad_norm": 22.781040116815415, + "learning_rate": 5.426091676780499e-07, + "loss": 0.7505, + "step": 8269 + }, + { + "epoch": 4.489685124864278, + "grad_norm": 8.882692706900785, + "learning_rate": 5.414670363036722e-07, + "loss": 0.2729, + "step": 8270 + }, + { + "epoch": 4.490228013029316, + "grad_norm": 14.363794979804172, + "learning_rate": 5.403260747757266e-07, + "loss": 0.453, + "step": 8271 + }, + { + "epoch": 4.490770901194354, + "grad_norm": 10.245019537386296, + "learning_rate": 5.391862832353289e-07, + "loss": 0.3372, + "step": 8272 + }, + { + "epoch": 4.491313789359392, + "grad_norm": 12.50265023247342, + "learning_rate": 5.380476618234498e-07, + "loss": 0.4035, + "step": 8273 + }, + { + "epoch": 4.49185667752443, + "grad_norm": 11.646457698141697, + "learning_rate": 5.369102106809165e-07, + "loss": 0.3997, + "step": 8274 + }, + { + "epoch": 4.4923995656894675, + "grad_norm": 9.464982035604422, + "learning_rate": 5.3577392994841e-07, + "loss": 0.2514, + "step": 8275 + }, + { + "epoch": 4.492942453854506, + "grad_norm": 17.529725496816823, + "learning_rate": 5.346388197664675e-07, + "loss": 0.4352, + "step": 8276 + }, + { + "epoch": 4.493485342019544, + "grad_norm": 17.57064929965146, + "learning_rate": 5.335048802754817e-07, + "loss": 0.4559, + "step": 8277 + }, + { + "epoch": 4.494028230184582, + "grad_norm": 13.71860363844268, + "learning_rate": 5.323721116156999e-07, + "loss": 0.3552, + "step": 8278 + }, + { + "epoch": 4.49457111834962, + "grad_norm": 10.880133558033618, + "learning_rate": 5.312405139272237e-07, + "loss": 0.3963, + "step": 8279 + }, + { + "epoch": 4.495114006514658, + "grad_norm": 20.247587867745704, + "learning_rate": 5.30110087350012e-07, + "loss": 0.5438, + "step": 8280 + }, + { + "epoch": 4.495656894679696, + "grad_norm": 9.454005269809635, + "learning_rate": 5.289808320238776e-07, + "loss": 0.2168, + "step": 8281 + }, + { + "epoch": 4.496199782844734, + "grad_norm": 13.194855975031016, + "learning_rate": 5.278527480884888e-07, + "loss": 0.3221, + "step": 8282 + }, + { + "epoch": 4.4967426710097715, + "grad_norm": 16.741747646625797, + "learning_rate": 5.267258356833671e-07, + "loss": 0.5343, + "step": 8283 + }, + { + "epoch": 4.49728555917481, + "grad_norm": 17.72898677538945, + "learning_rate": 5.256000949478934e-07, + "loss": 0.5213, + "step": 8284 + }, + { + "epoch": 4.497828447339848, + "grad_norm": 23.94089901495696, + "learning_rate": 5.244755260212986e-07, + "loss": 1.1702, + "step": 8285 + }, + { + "epoch": 4.498371335504886, + "grad_norm": 15.217551359960183, + "learning_rate": 5.233521290426746e-07, + "loss": 0.667, + "step": 8286 + }, + { + "epoch": 4.498914223669924, + "grad_norm": 15.41827808885883, + "learning_rate": 5.222299041509604e-07, + "loss": 0.3961, + "step": 8287 + }, + { + "epoch": 4.499457111834962, + "grad_norm": 10.397823177247577, + "learning_rate": 5.211088514849594e-07, + "loss": 0.3334, + "step": 8288 + }, + { + "epoch": 4.5, + "grad_norm": 10.913567291837683, + "learning_rate": 5.199889711833217e-07, + "loss": 0.3493, + "step": 8289 + }, + { + "epoch": 4.500542888165038, + "grad_norm": 12.37886318983014, + "learning_rate": 5.188702633845566e-07, + "loss": 0.3399, + "step": 8290 + }, + { + "epoch": 4.501085776330076, + "grad_norm": 11.119221436517403, + "learning_rate": 5.177527282270278e-07, + "loss": 0.2982, + "step": 8291 + }, + { + "epoch": 4.501628664495114, + "grad_norm": 11.838482844225242, + "learning_rate": 5.166363658489537e-07, + "loss": 0.4248, + "step": 8292 + }, + { + "epoch": 4.502171552660152, + "grad_norm": 13.111861380868095, + "learning_rate": 5.155211763884094e-07, + "loss": 0.521, + "step": 8293 + }, + { + "epoch": 4.50271444082519, + "grad_norm": 12.88810562354075, + "learning_rate": 5.1440715998332e-07, + "loss": 0.3804, + "step": 8294 + }, + { + "epoch": 4.5032573289902285, + "grad_norm": 10.514953303860349, + "learning_rate": 5.132943167714744e-07, + "loss": 0.3674, + "step": 8295 + }, + { + "epoch": 4.503800217155266, + "grad_norm": 10.023365825501816, + "learning_rate": 5.121826468905033e-07, + "loss": 0.5211, + "step": 8296 + }, + { + "epoch": 4.504343105320304, + "grad_norm": 11.606294239792984, + "learning_rate": 5.110721504779059e-07, + "loss": 0.7045, + "step": 8297 + }, + { + "epoch": 4.504885993485342, + "grad_norm": 15.029276069995117, + "learning_rate": 5.099628276710278e-07, + "loss": 0.602, + "step": 8298 + }, + { + "epoch": 4.50542888165038, + "grad_norm": 14.866091084726918, + "learning_rate": 5.088546786070714e-07, + "loss": 0.4688, + "step": 8299 + }, + { + "epoch": 4.505971769815418, + "grad_norm": 12.486305532545822, + "learning_rate": 5.07747703423097e-07, + "loss": 0.6218, + "step": 8300 + }, + { + "epoch": 4.506514657980456, + "grad_norm": 13.755190146974774, + "learning_rate": 5.066419022560121e-07, + "loss": 0.5051, + "step": 8301 + }, + { + "epoch": 4.507057546145494, + "grad_norm": 17.222866237868953, + "learning_rate": 5.055372752425902e-07, + "loss": 0.5394, + "step": 8302 + }, + { + "epoch": 4.5076004343105325, + "grad_norm": 13.245212020095746, + "learning_rate": 5.044338225194467e-07, + "loss": 0.2677, + "step": 8303 + }, + { + "epoch": 4.50814332247557, + "grad_norm": 17.027782293780568, + "learning_rate": 5.033315442230636e-07, + "loss": 0.6007, + "step": 8304 + }, + { + "epoch": 4.508686210640608, + "grad_norm": 13.708233013149469, + "learning_rate": 5.022304404897693e-07, + "loss": 0.623, + "step": 8305 + }, + { + "epoch": 4.509229098805646, + "grad_norm": 12.927772993543751, + "learning_rate": 5.011305114557519e-07, + "loss": 0.3889, + "step": 8306 + }, + { + "epoch": 4.509771986970684, + "grad_norm": 14.732891345640356, + "learning_rate": 5.000317572570523e-07, + "loss": 0.3646, + "step": 8307 + }, + { + "epoch": 4.510314875135722, + "grad_norm": 12.052951834491875, + "learning_rate": 4.989341780295654e-07, + "loss": 0.3066, + "step": 8308 + }, + { + "epoch": 4.51085776330076, + "grad_norm": 12.16750268825224, + "learning_rate": 4.978377739090424e-07, + "loss": 0.317, + "step": 8309 + }, + { + "epoch": 4.511400651465798, + "grad_norm": 15.533245706529485, + "learning_rate": 4.96742545031087e-07, + "loss": 0.2653, + "step": 8310 + }, + { + "epoch": 4.5119435396308365, + "grad_norm": 15.668556903572544, + "learning_rate": 4.956484915311599e-07, + "loss": 0.3729, + "step": 8311 + }, + { + "epoch": 4.512486427795874, + "grad_norm": 17.819603638078444, + "learning_rate": 4.945556135445739e-07, + "loss": 0.8422, + "step": 8312 + }, + { + "epoch": 4.513029315960912, + "grad_norm": 11.03718224282426, + "learning_rate": 4.934639112064998e-07, + "loss": 0.3522, + "step": 8313 + }, + { + "epoch": 4.51357220412595, + "grad_norm": 17.833077129344446, + "learning_rate": 4.923733846519607e-07, + "loss": 0.3477, + "step": 8314 + }, + { + "epoch": 4.514115092290988, + "grad_norm": 13.75370085535341, + "learning_rate": 4.912840340158343e-07, + "loss": 0.4303, + "step": 8315 + }, + { + "epoch": 4.514657980456026, + "grad_norm": 13.416980517253936, + "learning_rate": 4.901958594328527e-07, + "loss": 0.5717, + "step": 8316 + }, + { + "epoch": 4.515200868621064, + "grad_norm": 13.697716950123397, + "learning_rate": 4.891088610376038e-07, + "loss": 0.4465, + "step": 8317 + }, + { + "epoch": 4.515743756786102, + "grad_norm": 11.647982736650992, + "learning_rate": 4.880230389645291e-07, + "loss": 0.3537, + "step": 8318 + }, + { + "epoch": 4.5162866449511405, + "grad_norm": 17.71700440336686, + "learning_rate": 4.869383933479254e-07, + "loss": 0.5677, + "step": 8319 + }, + { + "epoch": 4.516829533116178, + "grad_norm": 8.473597132737073, + "learning_rate": 4.858549243219423e-07, + "loss": 0.2425, + "step": 8320 + }, + { + "epoch": 4.517372421281216, + "grad_norm": 9.186745419175947, + "learning_rate": 4.847726320205847e-07, + "loss": 0.2741, + "step": 8321 + }, + { + "epoch": 4.517915309446254, + "grad_norm": 10.521404049591816, + "learning_rate": 4.836915165777134e-07, + "loss": 0.3232, + "step": 8322 + }, + { + "epoch": 4.518458197611292, + "grad_norm": 18.349893458827204, + "learning_rate": 4.826115781270424e-07, + "loss": 0.5938, + "step": 8323 + }, + { + "epoch": 4.51900108577633, + "grad_norm": 10.506129387254754, + "learning_rate": 4.815328168021405e-07, + "loss": 0.2614, + "step": 8324 + }, + { + "epoch": 4.519543973941368, + "grad_norm": 13.11135730425014, + "learning_rate": 4.804552327364276e-07, + "loss": 0.3003, + "step": 8325 + }, + { + "epoch": 4.520086862106406, + "grad_norm": 9.56019008291403, + "learning_rate": 4.79378826063186e-07, + "loss": 0.2494, + "step": 8326 + }, + { + "epoch": 4.5206297502714445, + "grad_norm": 15.77492779940971, + "learning_rate": 4.783035969155425e-07, + "loss": 0.6032, + "step": 8327 + }, + { + "epoch": 4.521172638436482, + "grad_norm": 12.024645355321516, + "learning_rate": 4.772295454264886e-07, + "loss": 0.3562, + "step": 8328 + }, + { + "epoch": 4.52171552660152, + "grad_norm": 17.154707742654768, + "learning_rate": 4.7615667172885903e-07, + "loss": 0.4941, + "step": 8329 + }, + { + "epoch": 4.522258414766558, + "grad_norm": 11.691277186483006, + "learning_rate": 4.750849759553511e-07, + "loss": 0.3484, + "step": 8330 + }, + { + "epoch": 4.522801302931596, + "grad_norm": 11.471731467185506, + "learning_rate": 4.7401445823851534e-07, + "loss": 0.4616, + "step": 8331 + }, + { + "epoch": 4.523344191096634, + "grad_norm": 14.556670225131105, + "learning_rate": 4.729451187107514e-07, + "loss": 0.3747, + "step": 8332 + }, + { + "epoch": 4.523887079261672, + "grad_norm": 10.060443485377643, + "learning_rate": 4.718769575043214e-07, + "loss": 0.2432, + "step": 8333 + }, + { + "epoch": 4.52442996742671, + "grad_norm": 17.265452310056222, + "learning_rate": 4.708099747513328e-07, + "loss": 0.5352, + "step": 8334 + }, + { + "epoch": 4.5249728555917486, + "grad_norm": 14.537157483573262, + "learning_rate": 4.6974417058375574e-07, + "loss": 0.3997, + "step": 8335 + }, + { + "epoch": 4.525515743756786, + "grad_norm": 12.731997994806079, + "learning_rate": 4.68679545133407e-07, + "loss": 0.4316, + "step": 8336 + }, + { + "epoch": 4.526058631921824, + "grad_norm": 16.64941354972263, + "learning_rate": 4.676160985319633e-07, + "loss": 0.5068, + "step": 8337 + }, + { + "epoch": 4.526601520086862, + "grad_norm": 14.967442903073584, + "learning_rate": 4.6655383091095405e-07, + "loss": 0.5022, + "step": 8338 + }, + { + "epoch": 4.5271444082519, + "grad_norm": 14.760527489157035, + "learning_rate": 4.654927424017586e-07, + "loss": 0.5753, + "step": 8339 + }, + { + "epoch": 4.527687296416938, + "grad_norm": 18.12528620784058, + "learning_rate": 4.644328331356196e-07, + "loss": 0.5, + "step": 8340 + }, + { + "epoch": 4.528230184581976, + "grad_norm": 11.591072114682094, + "learning_rate": 4.633741032436223e-07, + "loss": 0.458, + "step": 8341 + }, + { + "epoch": 4.528773072747014, + "grad_norm": 14.15798620351717, + "learning_rate": 4.623165528567164e-07, + "loss": 0.3498, + "step": 8342 + }, + { + "epoch": 4.529315960912053, + "grad_norm": 23.12722823259664, + "learning_rate": 4.6126018210569946e-07, + "loss": 0.5397, + "step": 8343 + }, + { + "epoch": 4.52985884907709, + "grad_norm": 14.338174694861847, + "learning_rate": 4.602049911212259e-07, + "loss": 0.5037, + "step": 8344 + }, + { + "epoch": 4.530401737242128, + "grad_norm": 13.947445388030966, + "learning_rate": 4.5915098003380343e-07, + "loss": 0.7601, + "step": 8345 + }, + { + "epoch": 4.530944625407166, + "grad_norm": 15.423631300358158, + "learning_rate": 4.5809814897379345e-07, + "loss": 0.3918, + "step": 8346 + }, + { + "epoch": 4.531487513572204, + "grad_norm": 11.526636704758273, + "learning_rate": 4.570464980714129e-07, + "loss": 0.3192, + "step": 8347 + }, + { + "epoch": 4.532030401737242, + "grad_norm": 10.597564388470815, + "learning_rate": 4.559960274567299e-07, + "loss": 0.2513, + "step": 8348 + }, + { + "epoch": 4.53257328990228, + "grad_norm": 17.637678470826174, + "learning_rate": 4.5494673725966833e-07, + "loss": 0.3689, + "step": 8349 + }, + { + "epoch": 4.533116178067318, + "grad_norm": 15.332131840534869, + "learning_rate": 4.5389862761000767e-07, + "loss": 0.5437, + "step": 8350 + }, + { + "epoch": 4.533659066232357, + "grad_norm": 18.544251971087917, + "learning_rate": 4.5285169863737874e-07, + "loss": 0.9338, + "step": 8351 + }, + { + "epoch": 4.534201954397394, + "grad_norm": 14.845994288337273, + "learning_rate": 4.5180595047126795e-07, + "loss": 0.4528, + "step": 8352 + }, + { + "epoch": 4.534744842562432, + "grad_norm": 10.966600100926046, + "learning_rate": 4.5076138324101516e-07, + "loss": 0.3213, + "step": 8353 + }, + { + "epoch": 4.53528773072747, + "grad_norm": 13.979375013420652, + "learning_rate": 4.4971799707581254e-07, + "loss": 0.4328, + "step": 8354 + }, + { + "epoch": 4.535830618892508, + "grad_norm": 13.199579774530013, + "learning_rate": 4.4867579210470915e-07, + "loss": 0.4421, + "step": 8355 + }, + { + "epoch": 4.536373507057546, + "grad_norm": 15.679271986245535, + "learning_rate": 4.4763476845660627e-07, + "loss": 0.686, + "step": 8356 + }, + { + "epoch": 4.536916395222584, + "grad_norm": 10.32839293787231, + "learning_rate": 4.465949262602609e-07, + "loss": 0.3593, + "step": 8357 + }, + { + "epoch": 4.537459283387622, + "grad_norm": 19.97455202826578, + "learning_rate": 4.4555626564428024e-07, + "loss": 0.6335, + "step": 8358 + }, + { + "epoch": 4.538002171552661, + "grad_norm": 17.035674620541695, + "learning_rate": 4.44518786737127e-07, + "loss": 0.5835, + "step": 8359 + }, + { + "epoch": 4.538545059717698, + "grad_norm": 15.009806208164479, + "learning_rate": 4.434824896671208e-07, + "loss": 0.4077, + "step": 8360 + }, + { + "epoch": 4.539087947882736, + "grad_norm": 13.778069593523517, + "learning_rate": 4.4244737456243025e-07, + "loss": 0.3833, + "step": 8361 + }, + { + "epoch": 4.539630836047774, + "grad_norm": 12.494743874325179, + "learning_rate": 4.4141344155108066e-07, + "loss": 0.2996, + "step": 8362 + }, + { + "epoch": 4.540173724212812, + "grad_norm": 13.814732420411174, + "learning_rate": 4.4038069076094983e-07, + "loss": 0.319, + "step": 8363 + }, + { + "epoch": 4.54071661237785, + "grad_norm": 18.350946091244264, + "learning_rate": 4.3934912231977433e-07, + "loss": 0.7825, + "step": 8364 + }, + { + "epoch": 4.541259500542888, + "grad_norm": 11.565882281033971, + "learning_rate": 4.383187363551333e-07, + "loss": 0.3073, + "step": 8365 + }, + { + "epoch": 4.541802388707926, + "grad_norm": 16.013246124631713, + "learning_rate": 4.3728953299447265e-07, + "loss": 0.6382, + "step": 8366 + }, + { + "epoch": 4.542345276872965, + "grad_norm": 13.181941989917167, + "learning_rate": 4.362615123650815e-07, + "loss": 0.3552, + "step": 8367 + }, + { + "epoch": 4.542888165038002, + "grad_norm": 11.189597225295508, + "learning_rate": 4.352346745941083e-07, + "loss": 0.3848, + "step": 8368 + }, + { + "epoch": 4.54343105320304, + "grad_norm": 14.67310600028813, + "learning_rate": 4.342090198085569e-07, + "loss": 0.3936, + "step": 8369 + }, + { + "epoch": 4.543973941368078, + "grad_norm": 13.25839722725421, + "learning_rate": 4.33184548135277e-07, + "loss": 0.499, + "step": 8370 + }, + { + "epoch": 4.544516829533116, + "grad_norm": 9.9899345545398, + "learning_rate": 4.3216125970098164e-07, + "loss": 0.366, + "step": 8371 + }, + { + "epoch": 4.545059717698154, + "grad_norm": 8.836049199804664, + "learning_rate": 4.3113915463222855e-07, + "loss": 0.2981, + "step": 8372 + }, + { + "epoch": 4.545602605863192, + "grad_norm": 10.295970227608981, + "learning_rate": 4.3011823305543766e-07, + "loss": 0.3629, + "step": 8373 + }, + { + "epoch": 4.54614549402823, + "grad_norm": 12.00515544737705, + "learning_rate": 4.290984950968724e-07, + "loss": 0.3902, + "step": 8374 + }, + { + "epoch": 4.546688382193269, + "grad_norm": 15.20456661725038, + "learning_rate": 4.280799408826619e-07, + "loss": 0.3767, + "step": 8375 + }, + { + "epoch": 4.547231270358306, + "grad_norm": 16.982779080955037, + "learning_rate": 4.270625705387788e-07, + "loss": 0.3967, + "step": 8376 + }, + { + "epoch": 4.547774158523344, + "grad_norm": 10.63460587566383, + "learning_rate": 4.2604638419105336e-07, + "loss": 0.5597, + "step": 8377 + }, + { + "epoch": 4.548317046688382, + "grad_norm": 14.423686882901746, + "learning_rate": 4.250313819651719e-07, + "loss": 0.4968, + "step": 8378 + }, + { + "epoch": 4.54885993485342, + "grad_norm": 14.591468749508062, + "learning_rate": 4.240175639866662e-07, + "loss": 0.4191, + "step": 8379 + }, + { + "epoch": 4.549402823018458, + "grad_norm": 13.309197670984693, + "learning_rate": 4.230049303809325e-07, + "loss": 0.5255, + "step": 8380 + }, + { + "epoch": 4.549945711183496, + "grad_norm": 17.125364753150805, + "learning_rate": 4.219934812732107e-07, + "loss": 0.8478, + "step": 8381 + }, + { + "epoch": 4.550488599348534, + "grad_norm": 14.128057894087553, + "learning_rate": 4.209832167886019e-07, + "loss": 0.3886, + "step": 8382 + }, + { + "epoch": 4.551031487513573, + "grad_norm": 15.17581204817665, + "learning_rate": 4.199741370520538e-07, + "loss": 0.5868, + "step": 8383 + }, + { + "epoch": 4.5515743756786105, + "grad_norm": 13.255434170354311, + "learning_rate": 4.1896624218837444e-07, + "loss": 0.2633, + "step": 8384 + }, + { + "epoch": 4.552117263843648, + "grad_norm": 10.395013531026416, + "learning_rate": 4.1795953232221966e-07, + "loss": 0.2439, + "step": 8385 + }, + { + "epoch": 4.552660152008686, + "grad_norm": 16.020764006465836, + "learning_rate": 4.169540075781009e-07, + "loss": 0.4021, + "step": 8386 + }, + { + "epoch": 4.553203040173724, + "grad_norm": 16.237245900097843, + "learning_rate": 4.159496680803832e-07, + "loss": 0.4336, + "step": 8387 + }, + { + "epoch": 4.5537459283387625, + "grad_norm": 13.197573051421786, + "learning_rate": 4.14946513953286e-07, + "loss": 0.361, + "step": 8388 + }, + { + "epoch": 4.5542888165038, + "grad_norm": 14.796764385747446, + "learning_rate": 4.1394454532088015e-07, + "loss": 0.3395, + "step": 8389 + }, + { + "epoch": 4.554831704668838, + "grad_norm": 14.559402156578185, + "learning_rate": 4.129437623070909e-07, + "loss": 0.3893, + "step": 8390 + }, + { + "epoch": 4.555374592833877, + "grad_norm": 17.202126229361323, + "learning_rate": 4.1194416503569703e-07, + "loss": 0.7156, + "step": 8391 + }, + { + "epoch": 4.5559174809989145, + "grad_norm": 9.358596584111847, + "learning_rate": 4.109457536303285e-07, + "loss": 0.4116, + "step": 8392 + }, + { + "epoch": 4.556460369163952, + "grad_norm": 12.636931486569619, + "learning_rate": 4.099485282144733e-07, + "loss": 0.4355, + "step": 8393 + }, + { + "epoch": 4.55700325732899, + "grad_norm": 13.073013470376315, + "learning_rate": 4.089524889114671e-07, + "loss": 0.2467, + "step": 8394 + }, + { + "epoch": 4.557546145494028, + "grad_norm": 13.801266347348164, + "learning_rate": 4.0795763584450366e-07, + "loss": 0.7289, + "step": 8395 + }, + { + "epoch": 4.5580890336590665, + "grad_norm": 9.171162460594205, + "learning_rate": 4.069639691366267e-07, + "loss": 0.2739, + "step": 8396 + }, + { + "epoch": 4.558631921824104, + "grad_norm": 12.729495446250427, + "learning_rate": 4.059714889107369e-07, + "loss": 0.4215, + "step": 8397 + }, + { + "epoch": 4.559174809989142, + "grad_norm": 18.76182078049615, + "learning_rate": 4.0498019528958265e-07, + "loss": 0.4525, + "step": 8398 + }, + { + "epoch": 4.559717698154181, + "grad_norm": 10.419041741316235, + "learning_rate": 4.0399008839576927e-07, + "loss": 0.2068, + "step": 8399 + }, + { + "epoch": 4.5602605863192185, + "grad_norm": 16.500852833343252, + "learning_rate": 4.0300116835175653e-07, + "loss": 0.5465, + "step": 8400 + }, + { + "epoch": 4.560803474484256, + "grad_norm": 13.29028900926885, + "learning_rate": 4.020134352798533e-07, + "loss": 0.4574, + "step": 8401 + }, + { + "epoch": 4.561346362649294, + "grad_norm": 14.113439209552388, + "learning_rate": 4.010268893022273e-07, + "loss": 0.3419, + "step": 8402 + }, + { + "epoch": 4.561889250814332, + "grad_norm": 11.962313774958856, + "learning_rate": 4.0004153054089223e-07, + "loss": 0.2653, + "step": 8403 + }, + { + "epoch": 4.5624321389793705, + "grad_norm": 15.681458008075632, + "learning_rate": 3.990573591177227e-07, + "loss": 0.9879, + "step": 8404 + }, + { + "epoch": 4.562975027144408, + "grad_norm": 7.896889702285753, + "learning_rate": 3.9807437515443915e-07, + "loss": 0.266, + "step": 8405 + }, + { + "epoch": 4.563517915309446, + "grad_norm": 14.317793413141796, + "learning_rate": 3.97092578772621e-07, + "loss": 0.5589, + "step": 8406 + }, + { + "epoch": 4.564060803474485, + "grad_norm": 14.701295872646838, + "learning_rate": 3.9611197009369774e-07, + "loss": 0.4754, + "step": 8407 + }, + { + "epoch": 4.5646036916395225, + "grad_norm": 9.751091449679476, + "learning_rate": 3.9513254923895017e-07, + "loss": 0.3138, + "step": 8408 + }, + { + "epoch": 4.56514657980456, + "grad_norm": 13.368140777634292, + "learning_rate": 3.9415431632951917e-07, + "loss": 0.4099, + "step": 8409 + }, + { + "epoch": 4.565689467969598, + "grad_norm": 15.380332314106305, + "learning_rate": 3.9317727148638906e-07, + "loss": 0.528, + "step": 8410 + }, + { + "epoch": 4.566232356134636, + "grad_norm": 11.966998113136334, + "learning_rate": 3.9220141483040873e-07, + "loss": 0.3129, + "step": 8411 + }, + { + "epoch": 4.5667752442996745, + "grad_norm": 10.655786798583739, + "learning_rate": 3.912267464822661e-07, + "loss": 0.3956, + "step": 8412 + }, + { + "epoch": 4.567318132464712, + "grad_norm": 11.348329273545943, + "learning_rate": 3.9025326656251583e-07, + "loss": 0.2485, + "step": 8413 + }, + { + "epoch": 4.56786102062975, + "grad_norm": 10.93357120051505, + "learning_rate": 3.892809751915572e-07, + "loss": 0.1911, + "step": 8414 + }, + { + "epoch": 4.568403908794789, + "grad_norm": 16.65942853342247, + "learning_rate": 3.8830987248964394e-07, + "loss": 0.5467, + "step": 8415 + }, + { + "epoch": 4.5689467969598265, + "grad_norm": 14.213503704674388, + "learning_rate": 3.8733995857688664e-07, + "loss": 0.5076, + "step": 8416 + }, + { + "epoch": 4.569489685124864, + "grad_norm": 9.07435692786808, + "learning_rate": 3.8637123357324057e-07, + "loss": 0.314, + "step": 8417 + }, + { + "epoch": 4.570032573289902, + "grad_norm": 13.204052175322088, + "learning_rate": 3.8540369759852313e-07, + "loss": 0.4221, + "step": 8418 + }, + { + "epoch": 4.57057546145494, + "grad_norm": 13.165097473842113, + "learning_rate": 3.8443735077239975e-07, + "loss": 0.5271, + "step": 8419 + }, + { + "epoch": 4.5711183496199785, + "grad_norm": 14.175386650762487, + "learning_rate": 3.8347219321439033e-07, + "loss": 0.8023, + "step": 8420 + }, + { + "epoch": 4.571661237785016, + "grad_norm": 14.868459939871421, + "learning_rate": 3.825082250438661e-07, + "loss": 0.4034, + "step": 8421 + }, + { + "epoch": 4.572204125950054, + "grad_norm": 10.271953811670828, + "learning_rate": 3.8154544638005275e-07, + "loss": 0.2737, + "step": 8422 + }, + { + "epoch": 4.572747014115093, + "grad_norm": 12.599854900654838, + "learning_rate": 3.805838573420273e-07, + "loss": 0.3811, + "step": 8423 + }, + { + "epoch": 4.5732899022801305, + "grad_norm": 12.34104633408041, + "learning_rate": 3.7962345804872235e-07, + "loss": 0.4701, + "step": 8424 + }, + { + "epoch": 4.573832790445168, + "grad_norm": 16.049446226050716, + "learning_rate": 3.786642486189207e-07, + "loss": 0.6162, + "step": 8425 + }, + { + "epoch": 4.574375678610206, + "grad_norm": 12.429280513300323, + "learning_rate": 3.7770622917125857e-07, + "loss": 0.3345, + "step": 8426 + }, + { + "epoch": 4.574918566775244, + "grad_norm": 13.767089346919395, + "learning_rate": 3.7674939982422555e-07, + "loss": 0.4666, + "step": 8427 + }, + { + "epoch": 4.575461454940283, + "grad_norm": 11.650625569284712, + "learning_rate": 3.757937606961648e-07, + "loss": 0.446, + "step": 8428 + }, + { + "epoch": 4.57600434310532, + "grad_norm": 14.190570945854875, + "learning_rate": 3.7483931190526956e-07, + "loss": 0.5539, + "step": 8429 + }, + { + "epoch": 4.576547231270358, + "grad_norm": 14.141845515233648, + "learning_rate": 3.7388605356958873e-07, + "loss": 0.5402, + "step": 8430 + }, + { + "epoch": 4.577090119435397, + "grad_norm": 9.796163273061495, + "learning_rate": 3.7293398580702244e-07, + "loss": 0.1934, + "step": 8431 + }, + { + "epoch": 4.577633007600435, + "grad_norm": 14.965355025993395, + "learning_rate": 3.719831087353243e-07, + "loss": 0.6887, + "step": 8432 + }, + { + "epoch": 4.578175895765472, + "grad_norm": 10.131802869294864, + "learning_rate": 3.7103342247210015e-07, + "loss": 0.305, + "step": 8433 + }, + { + "epoch": 4.57871878393051, + "grad_norm": 13.343860286662204, + "learning_rate": 3.700849271348073e-07, + "loss": 0.573, + "step": 8434 + }, + { + "epoch": 4.579261672095548, + "grad_norm": 15.966222976200056, + "learning_rate": 3.691376228407606e-07, + "loss": 0.57, + "step": 8435 + }, + { + "epoch": 4.579804560260587, + "grad_norm": 11.11642134835898, + "learning_rate": 3.6819150970712093e-07, + "loss": 0.4264, + "step": 8436 + }, + { + "epoch": 4.580347448425624, + "grad_norm": 12.39535666720846, + "learning_rate": 3.672465878509057e-07, + "loss": 0.2841, + "step": 8437 + }, + { + "epoch": 4.580890336590662, + "grad_norm": 19.74473442455948, + "learning_rate": 3.66302857388986e-07, + "loss": 0.6739, + "step": 8438 + }, + { + "epoch": 4.581433224755701, + "grad_norm": 19.226999685559647, + "learning_rate": 3.653603184380805e-07, + "loss": 0.5112, + "step": 8439 + }, + { + "epoch": 4.581976112920739, + "grad_norm": 14.270097098199749, + "learning_rate": 3.644189711147683e-07, + "loss": 0.3839, + "step": 8440 + }, + { + "epoch": 4.582519001085776, + "grad_norm": 22.049747355061914, + "learning_rate": 3.634788155354729e-07, + "loss": 1.0213, + "step": 8441 + }, + { + "epoch": 4.583061889250814, + "grad_norm": 13.290670656397504, + "learning_rate": 3.62539851816478e-07, + "loss": 0.3051, + "step": 8442 + }, + { + "epoch": 4.583604777415852, + "grad_norm": 12.069309467099986, + "learning_rate": 3.616020800739117e-07, + "loss": 0.3633, + "step": 8443 + }, + { + "epoch": 4.584147665580891, + "grad_norm": 12.59014501860219, + "learning_rate": 3.606655004237647e-07, + "loss": 0.4773, + "step": 8444 + }, + { + "epoch": 4.584690553745928, + "grad_norm": 19.27546265345427, + "learning_rate": 3.5973011298186756e-07, + "loss": 0.6292, + "step": 8445 + }, + { + "epoch": 4.585233441910966, + "grad_norm": 12.759660559474787, + "learning_rate": 3.5879591786391667e-07, + "loss": 0.4249, + "step": 8446 + }, + { + "epoch": 4.585776330076005, + "grad_norm": 9.774391903683185, + "learning_rate": 3.5786291518545293e-07, + "loss": 0.2522, + "step": 8447 + }, + { + "epoch": 4.586319218241043, + "grad_norm": 17.765613416632792, + "learning_rate": 3.5693110506186956e-07, + "loss": 0.3391, + "step": 8448 + }, + { + "epoch": 4.58686210640608, + "grad_norm": 8.844583546256278, + "learning_rate": 3.5600048760841776e-07, + "loss": 0.29, + "step": 8449 + }, + { + "epoch": 4.587404994571118, + "grad_norm": 15.035031432995032, + "learning_rate": 3.5507106294019323e-07, + "loss": 0.517, + "step": 8450 + }, + { + "epoch": 4.587947882736156, + "grad_norm": 11.610533233131106, + "learning_rate": 3.5414283117215285e-07, + "loss": 0.3211, + "step": 8451 + }, + { + "epoch": 4.588490770901195, + "grad_norm": 16.920965677227937, + "learning_rate": 3.5321579241910043e-07, + "loss": 0.5429, + "step": 8452 + }, + { + "epoch": 4.589033659066232, + "grad_norm": 16.08329007012344, + "learning_rate": 3.5228994679569307e-07, + "loss": 0.535, + "step": 8453 + }, + { + "epoch": 4.58957654723127, + "grad_norm": 12.092722534315392, + "learning_rate": 3.513652944164414e-07, + "loss": 0.3239, + "step": 8454 + }, + { + "epoch": 4.590119435396309, + "grad_norm": 15.030040393058352, + "learning_rate": 3.504418353957073e-07, + "loss": 0.7184, + "step": 8455 + }, + { + "epoch": 4.590662323561347, + "grad_norm": 13.30266812560096, + "learning_rate": 3.495195698477083e-07, + "loss": 0.5521, + "step": 8456 + }, + { + "epoch": 4.591205211726384, + "grad_norm": 8.94936555843634, + "learning_rate": 3.4859849788650647e-07, + "loss": 0.2933, + "step": 8457 + }, + { + "epoch": 4.591748099891422, + "grad_norm": 12.005638724180985, + "learning_rate": 3.4767861962602624e-07, + "loss": 0.4297, + "step": 8458 + }, + { + "epoch": 4.59229098805646, + "grad_norm": 16.7349439675274, + "learning_rate": 3.467599351800366e-07, + "loss": 0.4244, + "step": 8459 + }, + { + "epoch": 4.592833876221499, + "grad_norm": 11.893802858804712, + "learning_rate": 3.458424446621644e-07, + "loss": 0.398, + "step": 8460 + }, + { + "epoch": 4.593376764386536, + "grad_norm": 14.139039290089297, + "learning_rate": 3.4492614818588457e-07, + "loss": 0.4137, + "step": 8461 + }, + { + "epoch": 4.593919652551574, + "grad_norm": 11.191416980837655, + "learning_rate": 3.440110458645263e-07, + "loss": 0.4034, + "step": 8462 + }, + { + "epoch": 4.594462540716613, + "grad_norm": 11.78684784372542, + "learning_rate": 3.430971378112724e-07, + "loss": 0.4019, + "step": 8463 + }, + { + "epoch": 4.595005428881651, + "grad_norm": 18.939664291756173, + "learning_rate": 3.4218442413915477e-07, + "loss": 0.7571, + "step": 8464 + }, + { + "epoch": 4.595548317046688, + "grad_norm": 15.556000892744493, + "learning_rate": 3.412729049610586e-07, + "loss": 0.6587, + "step": 8465 + }, + { + "epoch": 4.596091205211726, + "grad_norm": 11.440255717066862, + "learning_rate": 3.403625803897248e-07, + "loss": 0.5029, + "step": 8466 + }, + { + "epoch": 4.596634093376764, + "grad_norm": 13.988507935932384, + "learning_rate": 3.3945345053774115e-07, + "loss": 0.5368, + "step": 8467 + }, + { + "epoch": 4.597176981541803, + "grad_norm": 12.212309724820258, + "learning_rate": 3.385455155175521e-07, + "loss": 0.5259, + "step": 8468 + }, + { + "epoch": 4.59771986970684, + "grad_norm": 15.416156056006928, + "learning_rate": 3.376387754414523e-07, + "loss": 0.5056, + "step": 8469 + }, + { + "epoch": 4.598262757871878, + "grad_norm": 13.475505300424985, + "learning_rate": 3.3673323042158645e-07, + "loss": 0.4761, + "step": 8470 + }, + { + "epoch": 4.598805646036917, + "grad_norm": 14.065679411115752, + "learning_rate": 3.358288805699572e-07, + "loss": 0.42, + "step": 8471 + }, + { + "epoch": 4.599348534201955, + "grad_norm": 13.604312597253596, + "learning_rate": 3.349257259984129e-07, + "loss": 0.3117, + "step": 8472 + }, + { + "epoch": 4.599891422366992, + "grad_norm": 12.84454204935273, + "learning_rate": 3.3402376681866076e-07, + "loss": 0.3665, + "step": 8473 + }, + { + "epoch": 4.60043431053203, + "grad_norm": 11.499709529011515, + "learning_rate": 3.3312300314225166e-07, + "loss": 0.4646, + "step": 8474 + }, + { + "epoch": 4.600977198697068, + "grad_norm": 16.60616452066744, + "learning_rate": 3.3222343508059975e-07, + "loss": 0.6893, + "step": 8475 + }, + { + "epoch": 4.601520086862107, + "grad_norm": 11.281168428156999, + "learning_rate": 3.3132506274495936e-07, + "loss": 0.2993, + "step": 8476 + }, + { + "epoch": 4.6020629750271445, + "grad_norm": 9.791961525364469, + "learning_rate": 3.3042788624644496e-07, + "loss": 0.366, + "step": 8477 + }, + { + "epoch": 4.602605863192182, + "grad_norm": 13.850560788317166, + "learning_rate": 3.2953190569602e-07, + "loss": 0.657, + "step": 8478 + }, + { + "epoch": 4.603148751357221, + "grad_norm": 15.50838377377847, + "learning_rate": 3.2863712120450144e-07, + "loss": 0.4996, + "step": 8479 + }, + { + "epoch": 4.603691639522259, + "grad_norm": 14.002029740570508, + "learning_rate": 3.2774353288255957e-07, + "loss": 0.2919, + "step": 8480 + }, + { + "epoch": 4.6042345276872965, + "grad_norm": 11.327925279919503, + "learning_rate": 3.268511408407105e-07, + "loss": 0.3677, + "step": 8481 + }, + { + "epoch": 4.604777415852334, + "grad_norm": 12.329600797590148, + "learning_rate": 3.259599451893303e-07, + "loss": 0.5164, + "step": 8482 + }, + { + "epoch": 4.605320304017372, + "grad_norm": 12.05441506895834, + "learning_rate": 3.2506994603864085e-07, + "loss": 0.373, + "step": 8483 + }, + { + "epoch": 4.605863192182411, + "grad_norm": 14.077644499086377, + "learning_rate": 3.241811434987219e-07, + "loss": 0.5312, + "step": 8484 + }, + { + "epoch": 4.6064060803474485, + "grad_norm": 13.177790677829083, + "learning_rate": 3.232935376794999e-07, + "loss": 0.4356, + "step": 8485 + }, + { + "epoch": 4.606948968512486, + "grad_norm": 13.304561923438573, + "learning_rate": 3.224071286907537e-07, + "loss": 0.5409, + "step": 8486 + }, + { + "epoch": 4.607491856677525, + "grad_norm": 12.635364751343143, + "learning_rate": 3.2152191664212016e-07, + "loss": 0.364, + "step": 8487 + }, + { + "epoch": 4.608034744842563, + "grad_norm": 13.255155227139575, + "learning_rate": 3.2063790164307827e-07, + "loss": 0.5079, + "step": 8488 + }, + { + "epoch": 4.6085776330076005, + "grad_norm": 10.295823117219982, + "learning_rate": 3.197550838029684e-07, + "loss": 0.4099, + "step": 8489 + }, + { + "epoch": 4.609120521172638, + "grad_norm": 11.793953833291377, + "learning_rate": 3.188734632309787e-07, + "loss": 0.3837, + "step": 8490 + }, + { + "epoch": 4.609663409337676, + "grad_norm": 9.860632156842737, + "learning_rate": 3.179930400361475e-07, + "loss": 0.3377, + "step": 8491 + }, + { + "epoch": 4.610206297502715, + "grad_norm": 14.1155595223432, + "learning_rate": 3.171138143273689e-07, + "loss": 0.555, + "step": 8492 + }, + { + "epoch": 4.6107491856677525, + "grad_norm": 12.713917889288181, + "learning_rate": 3.162357862133858e-07, + "loss": 0.7882, + "step": 8493 + }, + { + "epoch": 4.61129207383279, + "grad_norm": 10.439021701558202, + "learning_rate": 3.1535895580279364e-07, + "loss": 0.2997, + "step": 8494 + }, + { + "epoch": 4.611834961997829, + "grad_norm": 20.571810234682463, + "learning_rate": 3.1448332320404116e-07, + "loss": 0.8567, + "step": 8495 + }, + { + "epoch": 4.612377850162867, + "grad_norm": 15.758885478695694, + "learning_rate": 3.1360888852542735e-07, + "loss": 0.6149, + "step": 8496 + }, + { + "epoch": 4.6129207383279045, + "grad_norm": 11.258643577667833, + "learning_rate": 3.1273565187510455e-07, + "loss": 0.3024, + "step": 8497 + }, + { + "epoch": 4.613463626492942, + "grad_norm": 8.375690337720608, + "learning_rate": 3.118636133610753e-07, + "loss": 0.2935, + "step": 8498 + }, + { + "epoch": 4.61400651465798, + "grad_norm": 14.247335073986578, + "learning_rate": 3.1099277309119544e-07, + "loss": 0.478, + "step": 8499 + }, + { + "epoch": 4.614549402823019, + "grad_norm": 13.654635451453982, + "learning_rate": 3.1012313117317007e-07, + "loss": 0.4477, + "step": 8500 + }, + { + "epoch": 4.6150922909880565, + "grad_norm": 18.13015910990287, + "learning_rate": 3.092546877145608e-07, + "loss": 0.4855, + "step": 8501 + }, + { + "epoch": 4.615635179153094, + "grad_norm": 12.036553317442285, + "learning_rate": 3.083874428227751e-07, + "loss": 0.3039, + "step": 8502 + }, + { + "epoch": 4.616178067318133, + "grad_norm": 11.335798067486618, + "learning_rate": 3.0752139660507716e-07, + "loss": 0.4776, + "step": 8503 + }, + { + "epoch": 4.616720955483171, + "grad_norm": 8.618445905749738, + "learning_rate": 3.0665654916858136e-07, + "loss": 0.2891, + "step": 8504 + }, + { + "epoch": 4.6172638436482085, + "grad_norm": 15.36813056497114, + "learning_rate": 3.057929006202509e-07, + "loss": 0.7977, + "step": 8505 + }, + { + "epoch": 4.617806731813246, + "grad_norm": 14.372663174892352, + "learning_rate": 3.04930451066906e-07, + "loss": 0.4023, + "step": 8506 + }, + { + "epoch": 4.618349619978284, + "grad_norm": 11.295565033762697, + "learning_rate": 3.040692006152135e-07, + "loss": 0.3676, + "step": 8507 + }, + { + "epoch": 4.618892508143323, + "grad_norm": 16.469747315227195, + "learning_rate": 3.03209149371696e-07, + "loss": 0.513, + "step": 8508 + }, + { + "epoch": 4.6194353963083605, + "grad_norm": 12.642088556027979, + "learning_rate": 3.0235029744272503e-07, + "loss": 0.2876, + "step": 8509 + }, + { + "epoch": 4.619978284473398, + "grad_norm": 11.969568380769612, + "learning_rate": 3.0149264493452345e-07, + "loss": 0.3763, + "step": 8510 + }, + { + "epoch": 4.620521172638437, + "grad_norm": 10.582102519076841, + "learning_rate": 3.0063619195317197e-07, + "loss": 0.3355, + "step": 8511 + }, + { + "epoch": 4.621064060803475, + "grad_norm": 16.902071395253756, + "learning_rate": 2.9978093860459133e-07, + "loss": 0.4931, + "step": 8512 + }, + { + "epoch": 4.6216069489685125, + "grad_norm": 12.087668761128803, + "learning_rate": 2.98926884994567e-07, + "loss": 0.5744, + "step": 8513 + }, + { + "epoch": 4.62214983713355, + "grad_norm": 12.845222810452263, + "learning_rate": 2.9807403122872556e-07, + "loss": 0.5582, + "step": 8514 + }, + { + "epoch": 4.622692725298588, + "grad_norm": 14.946886726818494, + "learning_rate": 2.972223774125504e-07, + "loss": 0.255, + "step": 8515 + }, + { + "epoch": 4.623235613463627, + "grad_norm": 11.713196783156775, + "learning_rate": 2.963719236513751e-07, + "loss": 0.3617, + "step": 8516 + }, + { + "epoch": 4.6237785016286646, + "grad_norm": 13.896782178184436, + "learning_rate": 2.955226700503855e-07, + "loss": 0.4964, + "step": 8517 + }, + { + "epoch": 4.624321389793702, + "grad_norm": 17.65687683800585, + "learning_rate": 2.946746167146197e-07, + "loss": 0.5491, + "step": 8518 + }, + { + "epoch": 4.624864277958741, + "grad_norm": 17.240713061010005, + "learning_rate": 2.938277637489639e-07, + "loss": 0.6051, + "step": 8519 + }, + { + "epoch": 4.625407166123779, + "grad_norm": 12.679756933018908, + "learning_rate": 2.929821112581621e-07, + "loss": 0.3541, + "step": 8520 + }, + { + "epoch": 4.625950054288817, + "grad_norm": 13.390576698021126, + "learning_rate": 2.921376593468006e-07, + "loss": 0.4184, + "step": 8521 + }, + { + "epoch": 4.626492942453854, + "grad_norm": 13.360144383030772, + "learning_rate": 2.9129440811932694e-07, + "loss": 0.4945, + "step": 8522 + }, + { + "epoch": 4.627035830618892, + "grad_norm": 11.822031606359241, + "learning_rate": 2.9045235768003334e-07, + "loss": 0.407, + "step": 8523 + }, + { + "epoch": 4.627578718783931, + "grad_norm": 15.383117428348534, + "learning_rate": 2.896115081330675e-07, + "loss": 0.3646, + "step": 8524 + }, + { + "epoch": 4.628121606948969, + "grad_norm": 9.313479496169867, + "learning_rate": 2.8877185958242846e-07, + "loss": 0.4177, + "step": 8525 + }, + { + "epoch": 4.628664495114006, + "grad_norm": 10.269974892795535, + "learning_rate": 2.879334121319599e-07, + "loss": 0.2665, + "step": 8526 + }, + { + "epoch": 4.629207383279044, + "grad_norm": 17.13009555473781, + "learning_rate": 2.870961658853677e-07, + "loss": 0.7038, + "step": 8527 + }, + { + "epoch": 4.629750271444083, + "grad_norm": 13.86007402381537, + "learning_rate": 2.8626012094620016e-07, + "loss": 0.4005, + "step": 8528 + }, + { + "epoch": 4.630293159609121, + "grad_norm": 12.672456984260394, + "learning_rate": 2.8542527741786343e-07, + "loss": 0.3733, + "step": 8529 + }, + { + "epoch": 4.630836047774158, + "grad_norm": 14.341898459464153, + "learning_rate": 2.8459163540361044e-07, + "loss": 0.44, + "step": 8530 + }, + { + "epoch": 4.631378935939196, + "grad_norm": 13.870748930087846, + "learning_rate": 2.837591950065477e-07, + "loss": 0.384, + "step": 8531 + }, + { + "epoch": 4.631921824104235, + "grad_norm": 13.917731753801313, + "learning_rate": 2.829279563296339e-07, + "loss": 0.4294, + "step": 8532 + }, + { + "epoch": 4.632464712269273, + "grad_norm": 11.929659844878131, + "learning_rate": 2.8209791947567786e-07, + "loss": 0.41, + "step": 8533 + }, + { + "epoch": 4.63300760043431, + "grad_norm": 13.166803611206449, + "learning_rate": 2.812690845473376e-07, + "loss": 0.5053, + "step": 8534 + }, + { + "epoch": 4.633550488599348, + "grad_norm": 14.961346446295355, + "learning_rate": 2.804414516471277e-07, + "loss": 0.5023, + "step": 8535 + }, + { + "epoch": 4.634093376764387, + "grad_norm": 12.220085895146651, + "learning_rate": 2.7961502087740845e-07, + "loss": 0.2651, + "step": 8536 + }, + { + "epoch": 4.634636264929425, + "grad_norm": 12.031722727757703, + "learning_rate": 2.787897923403959e-07, + "loss": 0.3387, + "step": 8537 + }, + { + "epoch": 4.635179153094462, + "grad_norm": 8.385703592908602, + "learning_rate": 2.779657661381552e-07, + "loss": 0.1904, + "step": 8538 + }, + { + "epoch": 4.6357220412595, + "grad_norm": 10.906284934193355, + "learning_rate": 2.7714294237260353e-07, + "loss": 0.2436, + "step": 8539 + }, + { + "epoch": 4.636264929424539, + "grad_norm": 10.369806572631566, + "learning_rate": 2.7632132114550846e-07, + "loss": 0.3501, + "step": 8540 + }, + { + "epoch": 4.636807817589577, + "grad_norm": 9.022715700382092, + "learning_rate": 2.7550090255848984e-07, + "loss": 0.2257, + "step": 8541 + }, + { + "epoch": 4.637350705754614, + "grad_norm": 12.181969040940151, + "learning_rate": 2.7468168671301756e-07, + "loss": 0.2761, + "step": 8542 + }, + { + "epoch": 4.637893593919652, + "grad_norm": 14.91305612612671, + "learning_rate": 2.7386367371041387e-07, + "loss": 0.5462, + "step": 8543 + }, + { + "epoch": 4.638436482084691, + "grad_norm": 14.378563278437591, + "learning_rate": 2.730468636518535e-07, + "loss": 0.465, + "step": 8544 + }, + { + "epoch": 4.638979370249729, + "grad_norm": 16.775535206614833, + "learning_rate": 2.722312566383589e-07, + "loss": 0.5409, + "step": 8545 + }, + { + "epoch": 4.639522258414766, + "grad_norm": 12.973657108911796, + "learning_rate": 2.714168527708061e-07, + "loss": 0.588, + "step": 8546 + }, + { + "epoch": 4.640065146579804, + "grad_norm": 11.59497961581488, + "learning_rate": 2.7060365214992114e-07, + "loss": 0.2145, + "step": 8547 + }, + { + "epoch": 4.640608034744843, + "grad_norm": 11.458027506031687, + "learning_rate": 2.6979165487628354e-07, + "loss": 0.4019, + "step": 8548 + }, + { + "epoch": 4.641150922909881, + "grad_norm": 10.405492855343923, + "learning_rate": 2.6898086105032193e-07, + "loss": 0.3135, + "step": 8549 + }, + { + "epoch": 4.641693811074918, + "grad_norm": 18.94716197151622, + "learning_rate": 2.6817127077231495e-07, + "loss": 0.681, + "step": 8550 + }, + { + "epoch": 4.642236699239956, + "grad_norm": 10.670452690665341, + "learning_rate": 2.6736288414239806e-07, + "loss": 0.3912, + "step": 8551 + }, + { + "epoch": 4.642779587404995, + "grad_norm": 12.348673232962904, + "learning_rate": 2.665557012605491e-07, + "loss": 0.388, + "step": 8552 + }, + { + "epoch": 4.643322475570033, + "grad_norm": 21.51174818589849, + "learning_rate": 2.6574972222660477e-07, + "loss": 0.5009, + "step": 8553 + }, + { + "epoch": 4.64386536373507, + "grad_norm": 12.513732617765974, + "learning_rate": 2.649449471402488e-07, + "loss": 0.5977, + "step": 8554 + }, + { + "epoch": 4.644408251900108, + "grad_norm": 15.098381270766813, + "learning_rate": 2.641413761010159e-07, + "loss": 0.661, + "step": 8555 + }, + { + "epoch": 4.644951140065147, + "grad_norm": 12.750147316660394, + "learning_rate": 2.633390092082966e-07, + "loss": 0.4127, + "step": 8556 + }, + { + "epoch": 4.645494028230185, + "grad_norm": 14.376855822812393, + "learning_rate": 2.6253784656132375e-07, + "loss": 0.4477, + "step": 8557 + }, + { + "epoch": 4.646036916395222, + "grad_norm": 9.209626776279498, + "learning_rate": 2.6173788825919253e-07, + "loss": 0.2738, + "step": 8558 + }, + { + "epoch": 4.64657980456026, + "grad_norm": 14.646741912366064, + "learning_rate": 2.6093913440083715e-07, + "loss": 0.5399, + "step": 8559 + }, + { + "epoch": 4.647122692725299, + "grad_norm": 13.357674411747405, + "learning_rate": 2.6014158508505304e-07, + "loss": 0.4451, + "step": 8560 + }, + { + "epoch": 4.647665580890337, + "grad_norm": 12.343655571585558, + "learning_rate": 2.5934524041047904e-07, + "loss": 0.25, + "step": 8561 + }, + { + "epoch": 4.648208469055374, + "grad_norm": 13.60791433197796, + "learning_rate": 2.585501004756108e-07, + "loss": 0.3931, + "step": 8562 + }, + { + "epoch": 4.648751357220412, + "grad_norm": 13.34834410008458, + "learning_rate": 2.577561653787919e-07, + "loss": 0.4463, + "step": 8563 + }, + { + "epoch": 4.649294245385451, + "grad_norm": 14.512470015045476, + "learning_rate": 2.569634352182171e-07, + "loss": 0.4353, + "step": 8564 + }, + { + "epoch": 4.649837133550489, + "grad_norm": 14.941554464032452, + "learning_rate": 2.5617191009193356e-07, + "loss": 0.5627, + "step": 8565 + }, + { + "epoch": 4.6503800217155264, + "grad_norm": 19.483475442382513, + "learning_rate": 2.5538159009783516e-07, + "loss": 0.4743, + "step": 8566 + }, + { + "epoch": 4.650922909880564, + "grad_norm": 14.823894906192926, + "learning_rate": 2.545924753336737e-07, + "loss": 0.7184, + "step": 8567 + }, + { + "epoch": 4.651465798045603, + "grad_norm": 10.450395109397446, + "learning_rate": 2.538045658970478e-07, + "loss": 0.2966, + "step": 8568 + }, + { + "epoch": 4.652008686210641, + "grad_norm": 15.67717478106929, + "learning_rate": 2.530178618854051e-07, + "loss": 0.7021, + "step": 8569 + }, + { + "epoch": 4.6525515743756785, + "grad_norm": 11.80163747115902, + "learning_rate": 2.5223236339604775e-07, + "loss": 0.8058, + "step": 8570 + }, + { + "epoch": 4.653094462540716, + "grad_norm": 13.651780146568179, + "learning_rate": 2.51448070526128e-07, + "loss": 0.6028, + "step": 8571 + }, + { + "epoch": 4.653637350705755, + "grad_norm": 11.185785067606833, + "learning_rate": 2.506649833726471e-07, + "loss": 0.3177, + "step": 8572 + }, + { + "epoch": 4.654180238870793, + "grad_norm": 12.995594956299897, + "learning_rate": 2.498831020324588e-07, + "loss": 0.4614, + "step": 8573 + }, + { + "epoch": 4.6547231270358305, + "grad_norm": 9.97604838386349, + "learning_rate": 2.49102426602269e-07, + "loss": 0.3827, + "step": 8574 + }, + { + "epoch": 4.655266015200868, + "grad_norm": 12.661162432576356, + "learning_rate": 2.4832295717863054e-07, + "loss": 0.4313, + "step": 8575 + }, + { + "epoch": 4.655808903365907, + "grad_norm": 17.062394951641533, + "learning_rate": 2.4754469385795177e-07, + "loss": 0.5858, + "step": 8576 + }, + { + "epoch": 4.656351791530945, + "grad_norm": 12.7320209210787, + "learning_rate": 2.4676763673648687e-07, + "loss": 0.3646, + "step": 8577 + }, + { + "epoch": 4.6568946796959825, + "grad_norm": 12.069345026621932, + "learning_rate": 2.4599178591034666e-07, + "loss": 0.4396, + "step": 8578 + }, + { + "epoch": 4.65743756786102, + "grad_norm": 11.365880065156547, + "learning_rate": 2.452171414754867e-07, + "loss": 0.27, + "step": 8579 + }, + { + "epoch": 4.657980456026059, + "grad_norm": 16.52465140027476, + "learning_rate": 2.4444370352771807e-07, + "loss": 0.3874, + "step": 8580 + }, + { + "epoch": 4.658523344191097, + "grad_norm": 9.574363596986716, + "learning_rate": 2.4367147216269873e-07, + "loss": 0.2478, + "step": 8581 + }, + { + "epoch": 4.6590662323561345, + "grad_norm": 15.080081475667454, + "learning_rate": 2.4290044747594333e-07, + "loss": 0.4246, + "step": 8582 + }, + { + "epoch": 4.659609120521172, + "grad_norm": 17.58550881221689, + "learning_rate": 2.421306295628101e-07, + "loss": 0.3626, + "step": 8583 + }, + { + "epoch": 4.660152008686211, + "grad_norm": 11.302404165947946, + "learning_rate": 2.413620185185128e-07, + "loss": 0.3163, + "step": 8584 + }, + { + "epoch": 4.660694896851249, + "grad_norm": 12.604954301784339, + "learning_rate": 2.4059461443811325e-07, + "loss": 0.4633, + "step": 8585 + }, + { + "epoch": 4.6612377850162865, + "grad_norm": 16.576550436899478, + "learning_rate": 2.398284174165255e-07, + "loss": 0.5843, + "step": 8586 + }, + { + "epoch": 4.661780673181324, + "grad_norm": 18.037796417236624, + "learning_rate": 2.3906342754851595e-07, + "loss": 0.8494, + "step": 8587 + }, + { + "epoch": 4.662323561346363, + "grad_norm": 10.736278337401655, + "learning_rate": 2.3829964492869673e-07, + "loss": 0.4509, + "step": 8588 + }, + { + "epoch": 4.662866449511401, + "grad_norm": 13.335894542183691, + "learning_rate": 2.3753706965153666e-07, + "loss": 0.6105, + "step": 8589 + }, + { + "epoch": 4.6634093376764385, + "grad_norm": 11.486510470019578, + "learning_rate": 2.3677570181134923e-07, + "loss": 0.3567, + "step": 8590 + }, + { + "epoch": 4.663952225841476, + "grad_norm": 19.53299668429002, + "learning_rate": 2.3601554150230577e-07, + "loss": 0.6259, + "step": 8591 + }, + { + "epoch": 4.664495114006515, + "grad_norm": 12.957043525497816, + "learning_rate": 2.3525658881841884e-07, + "loss": 0.4851, + "step": 8592 + }, + { + "epoch": 4.665038002171553, + "grad_norm": 11.145868206087584, + "learning_rate": 2.3449884385356114e-07, + "loss": 0.2943, + "step": 8593 + }, + { + "epoch": 4.6655808903365905, + "grad_norm": 17.411940658097325, + "learning_rate": 2.3374230670145104e-07, + "loss": 0.6137, + "step": 8594 + }, + { + "epoch": 4.666123778501628, + "grad_norm": 13.68690986112585, + "learning_rate": 2.329869774556559e-07, + "loss": 0.374, + "step": 8595 + }, + { + "epoch": 4.666666666666667, + "grad_norm": 15.428074760731318, + "learning_rate": 2.322328562095988e-07, + "loss": 0.3807, + "step": 8596 + }, + { + "epoch": 4.667209554831705, + "grad_norm": 14.483267133862492, + "learning_rate": 2.314799430565484e-07, + "loss": 0.4133, + "step": 8597 + }, + { + "epoch": 4.6677524429967425, + "grad_norm": 12.727450078217835, + "learning_rate": 2.3072823808962918e-07, + "loss": 0.424, + "step": 8598 + }, + { + "epoch": 4.66829533116178, + "grad_norm": 10.546825509905323, + "learning_rate": 2.2997774140181007e-07, + "loss": 0.3447, + "step": 8599 + }, + { + "epoch": 4.668838219326819, + "grad_norm": 12.26173003486407, + "learning_rate": 2.2922845308591457e-07, + "loss": 0.3995, + "step": 8600 + }, + { + "epoch": 4.669381107491857, + "grad_norm": 11.862665121581074, + "learning_rate": 2.2848037323461748e-07, + "loss": 0.396, + "step": 8601 + }, + { + "epoch": 4.6699239956568945, + "grad_norm": 11.53214186874608, + "learning_rate": 2.277335019404403e-07, + "loss": 0.3117, + "step": 8602 + }, + { + "epoch": 4.670466883821932, + "grad_norm": 9.677126503825807, + "learning_rate": 2.2698783929576029e-07, + "loss": 0.3299, + "step": 8603 + }, + { + "epoch": 4.671009771986971, + "grad_norm": 12.489313816621161, + "learning_rate": 2.26243385392797e-07, + "loss": 0.5942, + "step": 8604 + }, + { + "epoch": 4.671552660152009, + "grad_norm": 13.63702388723821, + "learning_rate": 2.255001403236312e-07, + "loss": 0.5161, + "step": 8605 + }, + { + "epoch": 4.6720955483170465, + "grad_norm": 12.047997675365789, + "learning_rate": 2.2475810418018496e-07, + "loss": 0.3852, + "step": 8606 + }, + { + "epoch": 4.672638436482084, + "grad_norm": 14.787350942388676, + "learning_rate": 2.2401727705423704e-07, + "loss": 0.4452, + "step": 8607 + }, + { + "epoch": 4.673181324647123, + "grad_norm": 15.453200600201, + "learning_rate": 2.2327765903741194e-07, + "loss": 0.4722, + "step": 8608 + }, + { + "epoch": 4.673724212812161, + "grad_norm": 14.118602165974508, + "learning_rate": 2.2253925022118872e-07, + "loss": 0.464, + "step": 8609 + }, + { + "epoch": 4.6742671009771986, + "grad_norm": 14.255945897987628, + "learning_rate": 2.218020506968932e-07, + "loss": 0.6369, + "step": 8610 + }, + { + "epoch": 4.674809989142236, + "grad_norm": 15.496989818750507, + "learning_rate": 2.2106606055570468e-07, + "loss": 0.2463, + "step": 8611 + }, + { + "epoch": 4.675352877307275, + "grad_norm": 11.272611567068735, + "learning_rate": 2.2033127988865032e-07, + "loss": 0.3691, + "step": 8612 + }, + { + "epoch": 4.675895765472313, + "grad_norm": 11.496100548853523, + "learning_rate": 2.195977087866097e-07, + "loss": 0.2911, + "step": 8613 + }, + { + "epoch": 4.676438653637351, + "grad_norm": 8.624224394988135, + "learning_rate": 2.1886534734031238e-07, + "loss": 0.2357, + "step": 8614 + }, + { + "epoch": 4.676981541802388, + "grad_norm": 15.681027668557938, + "learning_rate": 2.1813419564033823e-07, + "loss": 0.5629, + "step": 8615 + }, + { + "epoch": 4.677524429967427, + "grad_norm": 15.960892979111017, + "learning_rate": 2.1740425377711704e-07, + "loss": 0.5147, + "step": 8616 + }, + { + "epoch": 4.678067318132465, + "grad_norm": 14.11243172747272, + "learning_rate": 2.1667552184092778e-07, + "loss": 0.4392, + "step": 8617 + }, + { + "epoch": 4.678610206297503, + "grad_norm": 10.877174468176161, + "learning_rate": 2.1594799992190274e-07, + "loss": 0.3015, + "step": 8618 + }, + { + "epoch": 4.67915309446254, + "grad_norm": 15.805676119173855, + "learning_rate": 2.152216881100222e-07, + "loss": 0.6413, + "step": 8619 + }, + { + "epoch": 4.679695982627579, + "grad_norm": 10.914077977600847, + "learning_rate": 2.1449658649511873e-07, + "loss": 0.3409, + "step": 8620 + }, + { + "epoch": 4.680238870792617, + "grad_norm": 18.998010438264245, + "learning_rate": 2.137726951668717e-07, + "loss": 0.7325, + "step": 8621 + }, + { + "epoch": 4.680781758957655, + "grad_norm": 11.034223861305692, + "learning_rate": 2.130500142148162e-07, + "loss": 0.3421, + "step": 8622 + }, + { + "epoch": 4.681324647122692, + "grad_norm": 13.223093235714936, + "learning_rate": 2.123285437283329e-07, + "loss": 0.3714, + "step": 8623 + }, + { + "epoch": 4.681867535287731, + "grad_norm": 15.922027216206333, + "learning_rate": 2.1160828379665378e-07, + "loss": 0.6786, + "step": 8624 + }, + { + "epoch": 4.682410423452769, + "grad_norm": 12.103763081737, + "learning_rate": 2.1088923450886312e-07, + "loss": 0.2379, + "step": 8625 + }, + { + "epoch": 4.682953311617807, + "grad_norm": 8.921346684769288, + "learning_rate": 2.1017139595389202e-07, + "loss": 0.2147, + "step": 8626 + }, + { + "epoch": 4.683496199782844, + "grad_norm": 17.606964595238484, + "learning_rate": 2.0945476822052725e-07, + "loss": 0.5509, + "step": 8627 + }, + { + "epoch": 4.684039087947883, + "grad_norm": 16.30507726038034, + "learning_rate": 2.0873935139739898e-07, + "loss": 0.9206, + "step": 8628 + }, + { + "epoch": 4.684581976112921, + "grad_norm": 10.642327565305825, + "learning_rate": 2.0802514557299424e-07, + "loss": 0.4332, + "step": 8629 + }, + { + "epoch": 4.685124864277959, + "grad_norm": 11.145256035049321, + "learning_rate": 2.073121508356446e-07, + "loss": 0.3007, + "step": 8630 + }, + { + "epoch": 4.685667752442996, + "grad_norm": 12.907940052414924, + "learning_rate": 2.0660036727353615e-07, + "loss": 0.4258, + "step": 8631 + }, + { + "epoch": 4.686210640608035, + "grad_norm": 12.333985136599837, + "learning_rate": 2.0588979497470185e-07, + "loss": 0.3495, + "step": 8632 + }, + { + "epoch": 4.686753528773073, + "grad_norm": 15.062071959789007, + "learning_rate": 2.0518043402702692e-07, + "loss": 0.3947, + "step": 8633 + }, + { + "epoch": 4.687296416938111, + "grad_norm": 17.461595809708303, + "learning_rate": 2.0447228451824785e-07, + "loss": 0.6401, + "step": 8634 + }, + { + "epoch": 4.687839305103148, + "grad_norm": 15.753893232071881, + "learning_rate": 2.037653465359457e-07, + "loss": 0.5768, + "step": 8635 + }, + { + "epoch": 4.688382193268187, + "grad_norm": 7.552176112941645, + "learning_rate": 2.0305962016756164e-07, + "loss": 0.2491, + "step": 8636 + }, + { + "epoch": 4.688925081433225, + "grad_norm": 14.000517906223289, + "learning_rate": 2.0235510550037584e-07, + "loss": 0.4483, + "step": 8637 + }, + { + "epoch": 4.689467969598263, + "grad_norm": 11.530560060275077, + "learning_rate": 2.0165180262152528e-07, + "loss": 0.3759, + "step": 8638 + }, + { + "epoch": 4.6900108577633, + "grad_norm": 12.821236187937325, + "learning_rate": 2.0094971161799703e-07, + "loss": 0.3813, + "step": 8639 + }, + { + "epoch": 4.690553745928339, + "grad_norm": 10.067727608425038, + "learning_rate": 2.0024883257662608e-07, + "loss": 0.3769, + "step": 8640 + }, + { + "epoch": 4.691096634093377, + "grad_norm": 12.525413966092579, + "learning_rate": 1.9954916558409753e-07, + "loss": 0.5518, + "step": 8641 + }, + { + "epoch": 4.691639522258415, + "grad_norm": 10.599027738171559, + "learning_rate": 1.9885071072694773e-07, + "loss": 0.2527, + "step": 8642 + }, + { + "epoch": 4.692182410423452, + "grad_norm": 18.15731060573008, + "learning_rate": 1.9815346809156311e-07, + "loss": 0.4417, + "step": 8643 + }, + { + "epoch": 4.692725298588491, + "grad_norm": 10.69232848876475, + "learning_rate": 1.9745743776417914e-07, + "loss": 0.3218, + "step": 8644 + }, + { + "epoch": 4.693268186753529, + "grad_norm": 12.33953438764356, + "learning_rate": 1.967626198308814e-07, + "loss": 0.3422, + "step": 8645 + }, + { + "epoch": 4.693811074918567, + "grad_norm": 14.649644582793519, + "learning_rate": 1.9606901437760785e-07, + "loss": 0.5192, + "step": 8646 + }, + { + "epoch": 4.694353963083604, + "grad_norm": 13.239419010285507, + "learning_rate": 1.9537662149014312e-07, + "loss": 0.3623, + "step": 8647 + }, + { + "epoch": 4.694896851248643, + "grad_norm": 13.905675199978676, + "learning_rate": 1.9468544125412549e-07, + "loss": 0.5053, + "step": 8648 + }, + { + "epoch": 4.695439739413681, + "grad_norm": 12.990136427182705, + "learning_rate": 1.939954737550387e-07, + "loss": 0.3557, + "step": 8649 + }, + { + "epoch": 4.695982627578719, + "grad_norm": 17.66732769429285, + "learning_rate": 1.9330671907822006e-07, + "loss": 0.5037, + "step": 8650 + }, + { + "epoch": 4.696525515743756, + "grad_norm": 10.181756534303782, + "learning_rate": 1.9261917730885592e-07, + "loss": 0.3003, + "step": 8651 + }, + { + "epoch": 4.697068403908795, + "grad_norm": 13.590242162946018, + "learning_rate": 1.9193284853198268e-07, + "loss": 0.5647, + "step": 8652 + }, + { + "epoch": 4.697611292073833, + "grad_norm": 12.17175015122372, + "learning_rate": 1.9124773283248688e-07, + "loss": 0.4198, + "step": 8653 + }, + { + "epoch": 4.698154180238871, + "grad_norm": 11.105538961089028, + "learning_rate": 1.9056383029510406e-07, + "loss": 0.3301, + "step": 8654 + }, + { + "epoch": 4.698697068403908, + "grad_norm": 16.019376029880355, + "learning_rate": 1.898811410044199e-07, + "loss": 0.5822, + "step": 8655 + }, + { + "epoch": 4.699239956568947, + "grad_norm": 15.67141763731205, + "learning_rate": 1.8919966504487242e-07, + "loss": 0.5289, + "step": 8656 + }, + { + "epoch": 4.699782844733985, + "grad_norm": 12.06247311043435, + "learning_rate": 1.885194025007464e-07, + "loss": 0.305, + "step": 8657 + }, + { + "epoch": 4.700325732899023, + "grad_norm": 9.89364921552198, + "learning_rate": 1.878403534561768e-07, + "loss": 0.3231, + "step": 8658 + }, + { + "epoch": 4.7008686210640604, + "grad_norm": 14.67715265064961, + "learning_rate": 1.8716251799515083e-07, + "loss": 0.4446, + "step": 8659 + }, + { + "epoch": 4.701411509229099, + "grad_norm": 11.673006836898997, + "learning_rate": 1.864858962015048e-07, + "loss": 0.556, + "step": 8660 + }, + { + "epoch": 4.701954397394137, + "grad_norm": 10.608708966780606, + "learning_rate": 1.8581048815892288e-07, + "loss": 0.2792, + "step": 8661 + }, + { + "epoch": 4.702497285559175, + "grad_norm": 13.134592713431639, + "learning_rate": 1.8513629395094158e-07, + "loss": 0.3235, + "step": 8662 + }, + { + "epoch": 4.7030401737242125, + "grad_norm": 10.263003236237122, + "learning_rate": 1.844633136609453e-07, + "loss": 0.2699, + "step": 8663 + }, + { + "epoch": 4.703583061889251, + "grad_norm": 15.219462479114338, + "learning_rate": 1.8379154737216963e-07, + "loss": 0.3003, + "step": 8664 + }, + { + "epoch": 4.704125950054289, + "grad_norm": 13.049475534310673, + "learning_rate": 1.8312099516770265e-07, + "loss": 0.4857, + "step": 8665 + }, + { + "epoch": 4.704668838219327, + "grad_norm": 10.788197456357453, + "learning_rate": 1.824516571304735e-07, + "loss": 0.3011, + "step": 8666 + }, + { + "epoch": 4.7052117263843645, + "grad_norm": 9.043271241783842, + "learning_rate": 1.817835333432727e-07, + "loss": 0.2357, + "step": 8667 + }, + { + "epoch": 4.705754614549403, + "grad_norm": 9.656532376361252, + "learning_rate": 1.8111662388873074e-07, + "loss": 0.248, + "step": 8668 + }, + { + "epoch": 4.706297502714441, + "grad_norm": 16.70271141414241, + "learning_rate": 1.80450928849335e-07, + "loss": 0.4306, + "step": 8669 + }, + { + "epoch": 4.706840390879479, + "grad_norm": 13.146831118988962, + "learning_rate": 1.7978644830741743e-07, + "loss": 0.3665, + "step": 8670 + }, + { + "epoch": 4.7073832790445165, + "grad_norm": 15.26133993523942, + "learning_rate": 1.7912318234516334e-07, + "loss": 0.6296, + "step": 8671 + }, + { + "epoch": 4.707926167209555, + "grad_norm": 12.19166523931649, + "learning_rate": 1.7846113104460716e-07, + "loss": 0.3262, + "step": 8672 + }, + { + "epoch": 4.708469055374593, + "grad_norm": 14.50300563374095, + "learning_rate": 1.7780029448763004e-07, + "loss": 0.6108, + "step": 8673 + }, + { + "epoch": 4.709011943539631, + "grad_norm": 12.157023060827033, + "learning_rate": 1.771406727559688e-07, + "loss": 0.3338, + "step": 8674 + }, + { + "epoch": 4.7095548317046685, + "grad_norm": 16.358634798576084, + "learning_rate": 1.7648226593120154e-07, + "loss": 0.5102, + "step": 8675 + }, + { + "epoch": 4.710097719869707, + "grad_norm": 15.026104365171472, + "learning_rate": 1.7582507409476534e-07, + "loss": 0.5946, + "step": 8676 + }, + { + "epoch": 4.710640608034745, + "grad_norm": 10.191961599550515, + "learning_rate": 1.7516909732794186e-07, + "loss": 0.2927, + "step": 8677 + }, + { + "epoch": 4.711183496199783, + "grad_norm": 13.319008024488364, + "learning_rate": 1.7451433571186282e-07, + "loss": 0.3696, + "step": 8678 + }, + { + "epoch": 4.7117263843648205, + "grad_norm": 9.989147946007154, + "learning_rate": 1.7386078932751016e-07, + "loss": 0.2435, + "step": 8679 + }, + { + "epoch": 4.712269272529859, + "grad_norm": 14.90238420603907, + "learning_rate": 1.7320845825571474e-07, + "loss": 0.4526, + "step": 8680 + }, + { + "epoch": 4.712812160694897, + "grad_norm": 19.24089773059805, + "learning_rate": 1.7255734257716094e-07, + "loss": 0.6766, + "step": 8681 + }, + { + "epoch": 4.713355048859935, + "grad_norm": 14.187464466284949, + "learning_rate": 1.7190744237237545e-07, + "loss": 0.7976, + "step": 8682 + }, + { + "epoch": 4.7138979370249725, + "grad_norm": 14.841750204586576, + "learning_rate": 1.7125875772174172e-07, + "loss": 0.6627, + "step": 8683 + }, + { + "epoch": 4.714440825190011, + "grad_norm": 13.266714679527853, + "learning_rate": 1.7061128870548893e-07, + "loss": 0.3231, + "step": 8684 + }, + { + "epoch": 4.714983713355049, + "grad_norm": 10.735930720776885, + "learning_rate": 1.6996503540369857e-07, + "loss": 0.3564, + "step": 8685 + }, + { + "epoch": 4.715526601520087, + "grad_norm": 13.418944607227882, + "learning_rate": 1.6931999789629895e-07, + "loss": 0.4069, + "step": 8686 + }, + { + "epoch": 4.7160694896851245, + "grad_norm": 7.840135585116192, + "learning_rate": 1.6867617626307064e-07, + "loss": 0.2062, + "step": 8687 + }, + { + "epoch": 4.716612377850163, + "grad_norm": 16.197182822841835, + "learning_rate": 1.6803357058364113e-07, + "loss": 0.6624, + "step": 8688 + }, + { + "epoch": 4.717155266015201, + "grad_norm": 13.971201461249894, + "learning_rate": 1.6739218093748898e-07, + "loss": 0.6094, + "step": 8689 + }, + { + "epoch": 4.717698154180239, + "grad_norm": 13.176526767197032, + "learning_rate": 1.6675200740394303e-07, + "loss": 0.5601, + "step": 8690 + }, + { + "epoch": 4.7182410423452765, + "grad_norm": 11.043215518250438, + "learning_rate": 1.6611305006218103e-07, + "loss": 0.3768, + "step": 8691 + }, + { + "epoch": 4.718783930510315, + "grad_norm": 11.048074306636915, + "learning_rate": 1.6547530899122866e-07, + "loss": 0.3214, + "step": 8692 + }, + { + "epoch": 4.719326818675353, + "grad_norm": 12.953674925372624, + "learning_rate": 1.64838784269965e-07, + "loss": 0.3367, + "step": 8693 + }, + { + "epoch": 4.719869706840391, + "grad_norm": 19.600125509779307, + "learning_rate": 1.6420347597711607e-07, + "loss": 0.5111, + "step": 8694 + }, + { + "epoch": 4.7204125950054285, + "grad_norm": 12.73836642608448, + "learning_rate": 1.6356938419125556e-07, + "loss": 0.4027, + "step": 8695 + }, + { + "epoch": 4.720955483170467, + "grad_norm": 13.775692913408399, + "learning_rate": 1.6293650899081193e-07, + "loss": 0.6892, + "step": 8696 + }, + { + "epoch": 4.721498371335505, + "grad_norm": 13.475886438070386, + "learning_rate": 1.6230485045405697e-07, + "loss": 0.4, + "step": 8697 + }, + { + "epoch": 4.722041259500543, + "grad_norm": 14.97051098396697, + "learning_rate": 1.6167440865911932e-07, + "loss": 0.3445, + "step": 8698 + }, + { + "epoch": 4.7225841476655805, + "grad_norm": 16.47361547249056, + "learning_rate": 1.610451836839688e-07, + "loss": 0.3719, + "step": 8699 + }, + { + "epoch": 4.723127035830619, + "grad_norm": 11.078494155572205, + "learning_rate": 1.604171756064321e-07, + "loss": 0.3196, + "step": 8700 + }, + { + "epoch": 4.723669923995657, + "grad_norm": 17.74776123788687, + "learning_rate": 1.597903845041815e-07, + "loss": 0.686, + "step": 8701 + }, + { + "epoch": 4.724212812160695, + "grad_norm": 10.382852901941714, + "learning_rate": 1.591648104547383e-07, + "loss": 0.3855, + "step": 8702 + }, + { + "epoch": 4.7247557003257326, + "grad_norm": 13.409173794328078, + "learning_rate": 1.5854045353547508e-07, + "loss": 0.4286, + "step": 8703 + }, + { + "epoch": 4.725298588490771, + "grad_norm": 13.843911786331836, + "learning_rate": 1.579173138236134e-07, + "loss": 0.2781, + "step": 8704 + }, + { + "epoch": 4.725841476655809, + "grad_norm": 15.962352266628493, + "learning_rate": 1.5729539139622496e-07, + "loss": 0.573, + "step": 8705 + }, + { + "epoch": 4.726384364820847, + "grad_norm": 13.180235377705445, + "learning_rate": 1.566746863302293e-07, + "loss": 0.6014, + "step": 8706 + }, + { + "epoch": 4.726927252985885, + "grad_norm": 13.52518276203702, + "learning_rate": 1.5605519870239615e-07, + "loss": 0.4553, + "step": 8707 + }, + { + "epoch": 4.727470141150923, + "grad_norm": 16.51381961667992, + "learning_rate": 1.5543692858934535e-07, + "loss": 0.435, + "step": 8708 + }, + { + "epoch": 4.728013029315961, + "grad_norm": 7.676924114723452, + "learning_rate": 1.5481987606754456e-07, + "loss": 0.2235, + "step": 8709 + }, + { + "epoch": 4.728555917480999, + "grad_norm": 10.953711460471121, + "learning_rate": 1.542040412133128e-07, + "loss": 0.4336, + "step": 8710 + }, + { + "epoch": 4.729098805646037, + "grad_norm": 14.934986133675514, + "learning_rate": 1.5358942410281685e-07, + "loss": 0.461, + "step": 8711 + }, + { + "epoch": 4.729641693811075, + "grad_norm": 15.50829782966211, + "learning_rate": 1.5297602481207485e-07, + "loss": 0.4241, + "step": 8712 + }, + { + "epoch": 4.730184581976113, + "grad_norm": 12.740867628061949, + "learning_rate": 1.5236384341694944e-07, + "loss": 0.3727, + "step": 8713 + }, + { + "epoch": 4.730727470141151, + "grad_norm": 17.63779367168277, + "learning_rate": 1.5175287999316e-07, + "loss": 0.522, + "step": 8714 + }, + { + "epoch": 4.731270358306189, + "grad_norm": 12.751132476411902, + "learning_rate": 1.511431346162706e-07, + "loss": 0.3198, + "step": 8715 + }, + { + "epoch": 4.731813246471227, + "grad_norm": 13.490836748308888, + "learning_rate": 1.5053460736169535e-07, + "loss": 0.6525, + "step": 8716 + }, + { + "epoch": 4.732356134636265, + "grad_norm": 14.36545355486523, + "learning_rate": 1.499272983046962e-07, + "loss": 0.3776, + "step": 8717 + }, + { + "epoch": 4.732899022801303, + "grad_norm": 12.319339869403628, + "learning_rate": 1.493212075203887e-07, + "loss": 0.3082, + "step": 8718 + }, + { + "epoch": 4.733441910966341, + "grad_norm": 14.700540476410424, + "learning_rate": 1.4871633508373285e-07, + "loss": 0.5825, + "step": 8719 + }, + { + "epoch": 4.733984799131379, + "grad_norm": 12.437335028743151, + "learning_rate": 1.4811268106954212e-07, + "loss": 0.5425, + "step": 8720 + }, + { + "epoch": 4.734527687296417, + "grad_norm": 13.731965452187058, + "learning_rate": 1.4751024555247683e-07, + "loss": 0.3165, + "step": 8721 + }, + { + "epoch": 4.735070575461455, + "grad_norm": 12.201944872037965, + "learning_rate": 1.4690902860704626e-07, + "loss": 0.3219, + "step": 8722 + }, + { + "epoch": 4.735613463626493, + "grad_norm": 10.718132799597125, + "learning_rate": 1.463090303076109e-07, + "loss": 0.3031, + "step": 8723 + }, + { + "epoch": 4.736156351791531, + "grad_norm": 10.91023746799804, + "learning_rate": 1.457102507283781e-07, + "loss": 0.2936, + "step": 8724 + }, + { + "epoch": 4.736699239956569, + "grad_norm": 18.09494215671278, + "learning_rate": 1.4511268994340742e-07, + "loss": 0.481, + "step": 8725 + }, + { + "epoch": 4.737242128121607, + "grad_norm": 11.40586943270243, + "learning_rate": 1.4451634802660654e-07, + "loss": 0.335, + "step": 8726 + }, + { + "epoch": 4.737785016286645, + "grad_norm": 11.869445402678647, + "learning_rate": 1.4392122505172968e-07, + "loss": 0.4306, + "step": 8727 + }, + { + "epoch": 4.738327904451683, + "grad_norm": 15.51945714681897, + "learning_rate": 1.4332732109238466e-07, + "loss": 0.4721, + "step": 8728 + }, + { + "epoch": 4.738870792616721, + "grad_norm": 16.334459977316087, + "learning_rate": 1.427346362220261e-07, + "loss": 0.3783, + "step": 8729 + }, + { + "epoch": 4.739413680781759, + "grad_norm": 16.14244485338596, + "learning_rate": 1.4214317051395753e-07, + "loss": 0.5549, + "step": 8730 + }, + { + "epoch": 4.739956568946797, + "grad_norm": 10.731389089987651, + "learning_rate": 1.415529240413327e-07, + "loss": 0.2656, + "step": 8731 + }, + { + "epoch": 4.740499457111835, + "grad_norm": 11.843454339533057, + "learning_rate": 1.4096389687715427e-07, + "loss": 0.4719, + "step": 8732 + }, + { + "epoch": 4.741042345276873, + "grad_norm": 18.272514152254576, + "learning_rate": 1.4037608909427402e-07, + "loss": 0.6369, + "step": 8733 + }, + { + "epoch": 4.741585233441911, + "grad_norm": 12.608713904711752, + "learning_rate": 1.3978950076539265e-07, + "loss": 0.3475, + "step": 8734 + }, + { + "epoch": 4.742128121606949, + "grad_norm": 13.64628443407127, + "learning_rate": 1.3920413196305992e-07, + "loss": 0.578, + "step": 8735 + }, + { + "epoch": 4.742671009771987, + "grad_norm": 16.32438818863956, + "learning_rate": 1.3861998275967793e-07, + "loss": 0.6912, + "step": 8736 + }, + { + "epoch": 4.743213897937025, + "grad_norm": 11.713933611270436, + "learning_rate": 1.3803705322749107e-07, + "loss": 0.2049, + "step": 8737 + }, + { + "epoch": 4.743756786102063, + "grad_norm": 12.260983477641858, + "learning_rate": 1.3745534343860166e-07, + "loss": 0.4776, + "step": 8738 + }, + { + "epoch": 4.744299674267101, + "grad_norm": 12.840258489630633, + "learning_rate": 1.368748534649511e-07, + "loss": 0.3826, + "step": 8739 + }, + { + "epoch": 4.744842562432139, + "grad_norm": 16.03466717753298, + "learning_rate": 1.362955833783408e-07, + "loss": 0.6061, + "step": 8740 + }, + { + "epoch": 4.745385450597177, + "grad_norm": 9.356241312014683, + "learning_rate": 1.3571753325041126e-07, + "loss": 0.2633, + "step": 8741 + }, + { + "epoch": 4.745928338762215, + "grad_norm": 16.828204569000025, + "learning_rate": 1.351407031526586e-07, + "loss": 0.4258, + "step": 8742 + }, + { + "epoch": 4.746471226927253, + "grad_norm": 15.686474603102551, + "learning_rate": 1.3456509315642685e-07, + "loss": 0.3753, + "step": 8743 + }, + { + "epoch": 4.747014115092291, + "grad_norm": 9.198083120731805, + "learning_rate": 1.3399070333290688e-07, + "loss": 0.2984, + "step": 8744 + }, + { + "epoch": 4.747557003257329, + "grad_norm": 10.265356555589412, + "learning_rate": 1.3341753375314182e-07, + "loss": 0.3292, + "step": 8745 + }, + { + "epoch": 4.748099891422367, + "grad_norm": 10.896246230002951, + "learning_rate": 1.3284558448802054e-07, + "loss": 0.3922, + "step": 8746 + }, + { + "epoch": 4.748642779587405, + "grad_norm": 14.67283008436615, + "learning_rate": 1.3227485560828312e-07, + "loss": 0.5232, + "step": 8747 + }, + { + "epoch": 4.749185667752443, + "grad_norm": 9.52295940041419, + "learning_rate": 1.3170534718451754e-07, + "loss": 0.3544, + "step": 8748 + }, + { + "epoch": 4.749728555917481, + "grad_norm": 18.509918462924368, + "learning_rate": 1.3113705928716415e-07, + "loss": 0.4525, + "step": 8749 + }, + { + "epoch": 4.750271444082519, + "grad_norm": 12.780045325033498, + "learning_rate": 1.3056999198650778e-07, + "loss": 0.4545, + "step": 8750 + }, + { + "epoch": 4.750814332247557, + "grad_norm": 14.885728814999055, + "learning_rate": 1.3000414535268236e-07, + "loss": 0.2981, + "step": 8751 + }, + { + "epoch": 4.751357220412595, + "grad_norm": 13.266942664774517, + "learning_rate": 1.2943951945567633e-07, + "loss": 0.492, + "step": 8752 + }, + { + "epoch": 4.751900108577633, + "grad_norm": 16.094462901734268, + "learning_rate": 1.2887611436532167e-07, + "loss": 0.444, + "step": 8753 + }, + { + "epoch": 4.752442996742671, + "grad_norm": 14.798576298672852, + "learning_rate": 1.2831393015130034e-07, + "loss": 0.6268, + "step": 8754 + }, + { + "epoch": 4.752985884907709, + "grad_norm": 14.276704947243562, + "learning_rate": 1.2775296688314675e-07, + "loss": 0.4696, + "step": 8755 + }, + { + "epoch": 4.753528773072747, + "grad_norm": 11.206999668115289, + "learning_rate": 1.2719322463023877e-07, + "loss": 0.3501, + "step": 8756 + }, + { + "epoch": 4.754071661237785, + "grad_norm": 9.692669879532469, + "learning_rate": 1.2663470346180872e-07, + "loss": 0.317, + "step": 8757 + }, + { + "epoch": 4.754614549402823, + "grad_norm": 11.794412367069619, + "learning_rate": 1.2607740344693474e-07, + "loss": 0.5, + "step": 8758 + }, + { + "epoch": 4.755157437567861, + "grad_norm": 10.997145578080557, + "learning_rate": 1.2552132465454282e-07, + "loss": 0.2869, + "step": 8759 + }, + { + "epoch": 4.755700325732899, + "grad_norm": 11.55907046766932, + "learning_rate": 1.2496646715341231e-07, + "loss": 0.2764, + "step": 8760 + }, + { + "epoch": 4.756243213897937, + "grad_norm": 24.08335246288593, + "learning_rate": 1.2441283101216727e-07, + "loss": 1.1823, + "step": 8761 + }, + { + "epoch": 4.756786102062975, + "grad_norm": 14.279299755644365, + "learning_rate": 1.238604162992818e-07, + "loss": 0.7198, + "step": 8762 + }, + { + "epoch": 4.757328990228013, + "grad_norm": 13.336435628960723, + "learning_rate": 1.2330922308308125e-07, + "loss": 0.341, + "step": 8763 + }, + { + "epoch": 4.757871878393051, + "grad_norm": 13.97355889589316, + "learning_rate": 1.2275925143173662e-07, + "loss": 0.4514, + "step": 8764 + }, + { + "epoch": 4.758414766558089, + "grad_norm": 10.201594507019205, + "learning_rate": 1.2221050141327018e-07, + "loss": 0.4758, + "step": 8765 + }, + { + "epoch": 4.758957654723127, + "grad_norm": 8.477171079397465, + "learning_rate": 1.2166297309555098e-07, + "loss": 0.2141, + "step": 8766 + }, + { + "epoch": 4.759500542888165, + "grad_norm": 11.363514355058907, + "learning_rate": 1.211166665462993e-07, + "loss": 0.2775, + "step": 8767 + }, + { + "epoch": 4.760043431053203, + "grad_norm": 12.151898918162795, + "learning_rate": 1.2057158183308215e-07, + "loss": 0.3867, + "step": 8768 + }, + { + "epoch": 4.760586319218241, + "grad_norm": 12.369021167087942, + "learning_rate": 1.20027719023319e-07, + "loss": 0.3602, + "step": 8769 + }, + { + "epoch": 4.761129207383279, + "grad_norm": 16.318528391362594, + "learning_rate": 1.1948507818427269e-07, + "loss": 0.5529, + "step": 8770 + }, + { + "epoch": 4.761672095548317, + "grad_norm": 17.758768019553752, + "learning_rate": 1.1894365938305952e-07, + "loss": 0.7753, + "step": 8771 + }, + { + "epoch": 4.762214983713355, + "grad_norm": 11.920196045244229, + "learning_rate": 1.184034626866426e-07, + "loss": 0.36, + "step": 8772 + }, + { + "epoch": 4.762757871878393, + "grad_norm": 13.291930947753695, + "learning_rate": 1.1786448816183516e-07, + "loss": 0.5906, + "step": 8773 + }, + { + "epoch": 4.763300760043431, + "grad_norm": 14.292097175382699, + "learning_rate": 1.1732673587529609e-07, + "loss": 0.3195, + "step": 8774 + }, + { + "epoch": 4.763843648208469, + "grad_norm": 9.609372443037184, + "learning_rate": 1.1679020589353662e-07, + "loss": 0.2597, + "step": 8775 + }, + { + "epoch": 4.764386536373507, + "grad_norm": 13.447235691488418, + "learning_rate": 1.1625489828291814e-07, + "loss": 0.2508, + "step": 8776 + }, + { + "epoch": 4.764929424538545, + "grad_norm": 16.230999706702107, + "learning_rate": 1.1572081310964434e-07, + "loss": 0.7529, + "step": 8777 + }, + { + "epoch": 4.765472312703583, + "grad_norm": 12.727526228421432, + "learning_rate": 1.151879504397746e-07, + "loss": 0.3311, + "step": 8778 + }, + { + "epoch": 4.766015200868621, + "grad_norm": 10.438531330742874, + "learning_rate": 1.1465631033921288e-07, + "loss": 0.2289, + "step": 8779 + }, + { + "epoch": 4.766558089033659, + "grad_norm": 14.295286154658784, + "learning_rate": 1.1412589287371323e-07, + "loss": 0.3953, + "step": 8780 + }, + { + "epoch": 4.767100977198697, + "grad_norm": 9.693437885130846, + "learning_rate": 1.1359669810887985e-07, + "loss": 0.2993, + "step": 8781 + }, + { + "epoch": 4.767643865363735, + "grad_norm": 15.32504375625243, + "learning_rate": 1.1306872611016151e-07, + "loss": 0.4768, + "step": 8782 + }, + { + "epoch": 4.768186753528773, + "grad_norm": 11.211203320225184, + "learning_rate": 1.1254197694286262e-07, + "loss": 0.2661, + "step": 8783 + }, + { + "epoch": 4.768729641693811, + "grad_norm": 13.67914440990922, + "learning_rate": 1.1201645067212886e-07, + "loss": 0.5022, + "step": 8784 + }, + { + "epoch": 4.769272529858849, + "grad_norm": 11.569900535556993, + "learning_rate": 1.1149214736296044e-07, + "loss": 0.3133, + "step": 8785 + }, + { + "epoch": 4.769815418023887, + "grad_norm": 17.63101629527424, + "learning_rate": 1.1096906708020216e-07, + "loss": 0.5855, + "step": 8786 + }, + { + "epoch": 4.770358306188925, + "grad_norm": 13.27760517619257, + "learning_rate": 1.1044720988855118e-07, + "loss": 0.3421, + "step": 8787 + }, + { + "epoch": 4.770901194353963, + "grad_norm": 15.992691867357234, + "learning_rate": 1.099265758525514e-07, + "loss": 0.4832, + "step": 8788 + }, + { + "epoch": 4.771444082519001, + "grad_norm": 13.15390855364886, + "learning_rate": 1.0940716503659465e-07, + "loss": 0.3082, + "step": 8789 + }, + { + "epoch": 4.771986970684039, + "grad_norm": 18.263335547323074, + "learning_rate": 1.0888897750492289e-07, + "loss": 0.556, + "step": 8790 + }, + { + "epoch": 4.772529858849077, + "grad_norm": 10.441020903025926, + "learning_rate": 1.0837201332162595e-07, + "loss": 0.2007, + "step": 8791 + }, + { + "epoch": 4.773072747014115, + "grad_norm": 15.037462451331834, + "learning_rate": 1.0785627255064379e-07, + "loss": 0.6108, + "step": 8792 + }, + { + "epoch": 4.773615635179153, + "grad_norm": 14.518203414614236, + "learning_rate": 1.0734175525576429e-07, + "loss": 0.444, + "step": 8793 + }, + { + "epoch": 4.774158523344191, + "grad_norm": 11.643159629017385, + "learning_rate": 1.068284615006221e-07, + "loss": 0.4596, + "step": 8794 + }, + { + "epoch": 4.774701411509229, + "grad_norm": 15.528158500984132, + "learning_rate": 1.0631639134870308e-07, + "loss": 0.6026, + "step": 8795 + }, + { + "epoch": 4.7752442996742674, + "grad_norm": 19.44862903645615, + "learning_rate": 1.0580554486334216e-07, + "loss": 0.5733, + "step": 8796 + }, + { + "epoch": 4.775787187839305, + "grad_norm": 13.48756952731863, + "learning_rate": 1.0529592210771988e-07, + "loss": 0.3872, + "step": 8797 + }, + { + "epoch": 4.776330076004343, + "grad_norm": 13.820366819889347, + "learning_rate": 1.0478752314486695e-07, + "loss": 0.3807, + "step": 8798 + }, + { + "epoch": 4.776872964169381, + "grad_norm": 14.931351167549256, + "learning_rate": 1.0428034803766418e-07, + "loss": 0.4707, + "step": 8799 + }, + { + "epoch": 4.7774158523344195, + "grad_norm": 9.35754799853792, + "learning_rate": 1.0377439684884027e-07, + "loss": 0.2955, + "step": 8800 + }, + { + "epoch": 4.777958740499457, + "grad_norm": 15.469307360429545, + "learning_rate": 1.0326966964096963e-07, + "loss": 0.5782, + "step": 8801 + }, + { + "epoch": 4.778501628664495, + "grad_norm": 13.056079685648413, + "learning_rate": 1.0276616647648008e-07, + "loss": 0.5096, + "step": 8802 + }, + { + "epoch": 4.779044516829533, + "grad_norm": 13.862999489833173, + "learning_rate": 1.0226388741764514e-07, + "loss": 0.2736, + "step": 8803 + }, + { + "epoch": 4.7795874049945715, + "grad_norm": 10.890688599470478, + "learning_rate": 1.0176283252658737e-07, + "loss": 0.3334, + "step": 8804 + }, + { + "epoch": 4.780130293159609, + "grad_norm": 14.415866724042699, + "learning_rate": 1.0126300186527715e-07, + "loss": 0.5933, + "step": 8805 + }, + { + "epoch": 4.780673181324647, + "grad_norm": 13.555552920820668, + "learning_rate": 1.0076439549553507e-07, + "loss": 0.3958, + "step": 8806 + }, + { + "epoch": 4.781216069489685, + "grad_norm": 11.893874625329541, + "learning_rate": 1.0026701347903067e-07, + "loss": 0.3873, + "step": 8807 + }, + { + "epoch": 4.7817589576547235, + "grad_norm": 15.261946481554697, + "learning_rate": 9.977085587727919e-08, + "loss": 0.404, + "step": 8808 + }, + { + "epoch": 4.782301845819761, + "grad_norm": 11.516422308885032, + "learning_rate": 9.92759227516471e-08, + "loss": 0.318, + "step": 8809 + }, + { + "epoch": 4.782844733984799, + "grad_norm": 13.915845024156635, + "learning_rate": 9.878221416334877e-08, + "loss": 0.4227, + "step": 8810 + }, + { + "epoch": 4.783387622149837, + "grad_norm": 12.05067037271945, + "learning_rate": 9.828973017344534e-08, + "loss": 0.3802, + "step": 8811 + }, + { + "epoch": 4.7839305103148755, + "grad_norm": 12.366225498487221, + "learning_rate": 9.779847084284921e-08, + "loss": 0.5068, + "step": 8812 + }, + { + "epoch": 4.784473398479913, + "grad_norm": 11.020281825268318, + "learning_rate": 9.730843623231956e-08, + "loss": 0.4372, + "step": 8813 + }, + { + "epoch": 4.785016286644951, + "grad_norm": 11.115605935865265, + "learning_rate": 9.681962640246679e-08, + "loss": 0.3556, + "step": 8814 + }, + { + "epoch": 4.785559174809989, + "grad_norm": 15.746022339127645, + "learning_rate": 9.633204141374475e-08, + "loss": 0.4654, + "step": 8815 + }, + { + "epoch": 4.7861020629750275, + "grad_norm": 11.675199453622405, + "learning_rate": 9.584568132645966e-08, + "loss": 0.3, + "step": 8816 + }, + { + "epoch": 4.786644951140065, + "grad_norm": 15.071011265966053, + "learning_rate": 9.536054620076563e-08, + "loss": 0.3563, + "step": 8817 + }, + { + "epoch": 4.787187839305103, + "grad_norm": 14.392285757682815, + "learning_rate": 9.487663609666576e-08, + "loss": 0.3403, + "step": 8818 + }, + { + "epoch": 4.787730727470141, + "grad_norm": 14.478548975904134, + "learning_rate": 9.439395107400995e-08, + "loss": 0.4303, + "step": 8819 + }, + { + "epoch": 4.7882736156351795, + "grad_norm": 16.873014104043584, + "learning_rate": 9.391249119249601e-08, + "loss": 0.4669, + "step": 8820 + }, + { + "epoch": 4.788816503800217, + "grad_norm": 12.850529236209315, + "learning_rate": 9.34322565116752e-08, + "loss": 0.36, + "step": 8821 + }, + { + "epoch": 4.789359391965255, + "grad_norm": 12.049173435805207, + "learning_rate": 9.29532470909389e-08, + "loss": 0.4809, + "step": 8822 + }, + { + "epoch": 4.789902280130293, + "grad_norm": 14.445567418948913, + "learning_rate": 9.247546298953747e-08, + "loss": 0.375, + "step": 8823 + }, + { + "epoch": 4.7904451682953315, + "grad_norm": 16.331300178692935, + "learning_rate": 9.19989042665581e-08, + "loss": 0.5994, + "step": 8824 + }, + { + "epoch": 4.790988056460369, + "grad_norm": 16.563817844672233, + "learning_rate": 9.1523570980947e-08, + "loss": 0.4818, + "step": 8825 + }, + { + "epoch": 4.791530944625407, + "grad_norm": 12.9314417840519, + "learning_rate": 9.104946319149266e-08, + "loss": 0.5215, + "step": 8826 + }, + { + "epoch": 4.792073832790445, + "grad_norm": 11.600758934710594, + "learning_rate": 9.057658095683264e-08, + "loss": 0.4369, + "step": 8827 + }, + { + "epoch": 4.7926167209554835, + "grad_norm": 11.593366700305912, + "learning_rate": 9.010492433545459e-08, + "loss": 0.3912, + "step": 8828 + }, + { + "epoch": 4.793159609120521, + "grad_norm": 14.357902988841598, + "learning_rate": 8.963449338569297e-08, + "loss": 0.5229, + "step": 8829 + }, + { + "epoch": 4.793702497285559, + "grad_norm": 10.079996265707912, + "learning_rate": 8.916528816573234e-08, + "loss": 0.4725, + "step": 8830 + }, + { + "epoch": 4.794245385450597, + "grad_norm": 13.722106385687223, + "learning_rate": 8.869730873360405e-08, + "loss": 0.4548, + "step": 8831 + }, + { + "epoch": 4.7947882736156355, + "grad_norm": 14.066258092224395, + "learning_rate": 8.82305551471896e-08, + "loss": 0.5001, + "step": 8832 + }, + { + "epoch": 4.795331161780673, + "grad_norm": 12.082732968978428, + "learning_rate": 8.776502746421611e-08, + "loss": 0.3134, + "step": 8833 + }, + { + "epoch": 4.795874049945711, + "grad_norm": 14.022649629603004, + "learning_rate": 8.730072574226311e-08, + "loss": 0.3467, + "step": 8834 + }, + { + "epoch": 4.796416938110749, + "grad_norm": 15.002589519279939, + "learning_rate": 8.683765003875356e-08, + "loss": 0.4259, + "step": 8835 + }, + { + "epoch": 4.7969598262757875, + "grad_norm": 14.91562383479778, + "learning_rate": 8.637580041096382e-08, + "loss": 0.4919, + "step": 8836 + }, + { + "epoch": 4.797502714440825, + "grad_norm": 12.718843473649267, + "learning_rate": 8.59151769160138e-08, + "loss": 0.6194, + "step": 8837 + }, + { + "epoch": 4.798045602605863, + "grad_norm": 12.934801394638406, + "learning_rate": 8.545577961087681e-08, + "loss": 0.4032, + "step": 8838 + }, + { + "epoch": 4.798588490770901, + "grad_norm": 15.44257297198036, + "learning_rate": 8.49976085523696e-08, + "loss": 0.4059, + "step": 8839 + }, + { + "epoch": 4.7991313789359396, + "grad_norm": 11.274515134729002, + "learning_rate": 8.454066379716019e-08, + "loss": 0.2459, + "step": 8840 + }, + { + "epoch": 4.799674267100977, + "grad_norm": 13.639296201505058, + "learning_rate": 8.408494540176448e-08, + "loss": 0.4588, + "step": 8841 + }, + { + "epoch": 4.800217155266015, + "grad_norm": 10.677817137466716, + "learning_rate": 8.363045342254628e-08, + "loss": 0.2931, + "step": 8842 + }, + { + "epoch": 4.800760043431053, + "grad_norm": 14.87764794308729, + "learning_rate": 8.317718791571838e-08, + "loss": 0.5248, + "step": 8843 + }, + { + "epoch": 4.801302931596092, + "grad_norm": 11.064095389068989, + "learning_rate": 8.27251489373404e-08, + "loss": 0.2687, + "step": 8844 + }, + { + "epoch": 4.801845819761129, + "grad_norm": 22.55978148923217, + "learning_rate": 8.227433654332206e-08, + "loss": 0.862, + "step": 8845 + }, + { + "epoch": 4.802388707926167, + "grad_norm": 12.685033016571245, + "learning_rate": 8.182475078941987e-08, + "loss": 0.3498, + "step": 8846 + }, + { + "epoch": 4.802931596091205, + "grad_norm": 14.666620778959023, + "learning_rate": 8.137639173124046e-08, + "loss": 0.4637, + "step": 8847 + }, + { + "epoch": 4.803474484256244, + "grad_norm": 12.025170948479563, + "learning_rate": 8.092925942423613e-08, + "loss": 0.4287, + "step": 8848 + }, + { + "epoch": 4.804017372421281, + "grad_norm": 15.47756406381011, + "learning_rate": 8.048335392371042e-08, + "loss": 0.4397, + "step": 8849 + }, + { + "epoch": 4.804560260586319, + "grad_norm": 15.134699918755489, + "learning_rate": 8.003867528481257e-08, + "loss": 0.7081, + "step": 8850 + }, + { + "epoch": 4.805103148751357, + "grad_norm": 12.320316293618925, + "learning_rate": 7.959522356254079e-08, + "loss": 0.3828, + "step": 8851 + }, + { + "epoch": 4.805646036916396, + "grad_norm": 11.222062063383326, + "learning_rate": 7.915299881174344e-08, + "loss": 0.4006, + "step": 8852 + }, + { + "epoch": 4.806188925081433, + "grad_norm": 12.422853158493984, + "learning_rate": 7.871200108711341e-08, + "loss": 0.422, + "step": 8853 + }, + { + "epoch": 4.806731813246471, + "grad_norm": 9.781525068337004, + "learning_rate": 7.82722304431971e-08, + "loss": 0.2884, + "step": 8854 + }, + { + "epoch": 4.807274701411509, + "grad_norm": 11.760204421271188, + "learning_rate": 7.783368693438209e-08, + "loss": 0.4436, + "step": 8855 + }, + { + "epoch": 4.807817589576548, + "grad_norm": 10.811544328919764, + "learning_rate": 7.739637061491279e-08, + "loss": 0.2036, + "step": 8856 + }, + { + "epoch": 4.808360477741585, + "grad_norm": 16.51916863908083, + "learning_rate": 7.696028153887259e-08, + "loss": 0.7597, + "step": 8857 + }, + { + "epoch": 4.808903365906623, + "grad_norm": 18.728919815805536, + "learning_rate": 7.652541976019945e-08, + "loss": 0.4963, + "step": 8858 + }, + { + "epoch": 4.809446254071661, + "grad_norm": 12.95657669087096, + "learning_rate": 7.609178533268036e-08, + "loss": 0.4693, + "step": 8859 + }, + { + "epoch": 4.8099891422367, + "grad_norm": 11.87170444365128, + "learning_rate": 7.565937830994353e-08, + "loss": 0.4667, + "step": 8860 + }, + { + "epoch": 4.810532030401737, + "grad_norm": 10.885904559128528, + "learning_rate": 7.522819874547283e-08, + "loss": 0.2819, + "step": 8861 + }, + { + "epoch": 4.811074918566775, + "grad_norm": 11.241238204039677, + "learning_rate": 7.479824669259561e-08, + "loss": 0.2828, + "step": 8862 + }, + { + "epoch": 4.811617806731813, + "grad_norm": 12.844422586502446, + "learning_rate": 7.436952220449045e-08, + "loss": 0.4597, + "step": 8863 + }, + { + "epoch": 4.812160694896852, + "grad_norm": 10.410880488204011, + "learning_rate": 7.394202533418049e-08, + "loss": 0.3583, + "step": 8864 + }, + { + "epoch": 4.812703583061889, + "grad_norm": 10.695248824341306, + "learning_rate": 7.351575613454231e-08, + "loss": 0.3529, + "step": 8865 + }, + { + "epoch": 4.813246471226927, + "grad_norm": 15.957870026482412, + "learning_rate": 7.309071465829487e-08, + "loss": 0.4339, + "step": 8866 + }, + { + "epoch": 4.813789359391965, + "grad_norm": 10.337571437282099, + "learning_rate": 7.266690095800944e-08, + "loss": 0.2697, + "step": 8867 + }, + { + "epoch": 4.814332247557004, + "grad_norm": 11.474450789929184, + "learning_rate": 7.224431508610296e-08, + "loss": 0.2208, + "step": 8868 + }, + { + "epoch": 4.814875135722041, + "grad_norm": 9.448799109853498, + "learning_rate": 7.182295709484255e-08, + "loss": 0.24, + "step": 8869 + }, + { + "epoch": 4.815418023887079, + "grad_norm": 10.70255325665073, + "learning_rate": 7.140282703634204e-08, + "loss": 0.4463, + "step": 8870 + }, + { + "epoch": 4.815960912052117, + "grad_norm": 15.464670090867934, + "learning_rate": 7.098392496256324e-08, + "loss": 0.505, + "step": 8871 + }, + { + "epoch": 4.816503800217156, + "grad_norm": 12.515924687979576, + "learning_rate": 7.056625092531688e-08, + "loss": 0.3723, + "step": 8872 + }, + { + "epoch": 4.817046688382193, + "grad_norm": 12.22224030809159, + "learning_rate": 7.014980497626279e-08, + "loss": 0.3378, + "step": 8873 + }, + { + "epoch": 4.817589576547231, + "grad_norm": 18.212411386580676, + "learning_rate": 6.973458716690639e-08, + "loss": 0.5945, + "step": 8874 + }, + { + "epoch": 4.818132464712269, + "grad_norm": 12.548255509888588, + "learning_rate": 6.93205975486022e-08, + "loss": 0.3672, + "step": 8875 + }, + { + "epoch": 4.818675352877308, + "grad_norm": 10.738878454120973, + "learning_rate": 6.89078361725537e-08, + "loss": 0.3582, + "step": 8876 + }, + { + "epoch": 4.819218241042345, + "grad_norm": 13.763430071394062, + "learning_rate": 6.849630308981226e-08, + "loss": 0.4466, + "step": 8877 + }, + { + "epoch": 4.819761129207383, + "grad_norm": 14.698245992825742, + "learning_rate": 6.808599835127605e-08, + "loss": 0.4725, + "step": 8878 + }, + { + "epoch": 4.820304017372421, + "grad_norm": 17.5106844905948, + "learning_rate": 6.767692200769226e-08, + "loss": 0.582, + "step": 8879 + }, + { + "epoch": 4.82084690553746, + "grad_norm": 11.772458890524595, + "learning_rate": 6.726907410965822e-08, + "loss": 0.3311, + "step": 8880 + }, + { + "epoch": 4.821389793702497, + "grad_norm": 10.79164174819329, + "learning_rate": 6.686245470761355e-08, + "loss": 0.299, + "step": 8881 + }, + { + "epoch": 4.821932681867535, + "grad_norm": 12.844928905031605, + "learning_rate": 6.645706385185246e-08, + "loss": 0.3807, + "step": 8882 + }, + { + "epoch": 4.822475570032573, + "grad_norm": 21.17889161763075, + "learning_rate": 6.605290159251376e-08, + "loss": 0.6281, + "step": 8883 + }, + { + "epoch": 4.823018458197612, + "grad_norm": 10.727221853449628, + "learning_rate": 6.564996797958412e-08, + "loss": 0.3694, + "step": 8884 + }, + { + "epoch": 4.823561346362649, + "grad_norm": 16.28670128227416, + "learning_rate": 6.524826306289921e-08, + "loss": 0.4659, + "step": 8885 + }, + { + "epoch": 4.824104234527687, + "grad_norm": 17.1662871993217, + "learning_rate": 6.484778689214266e-08, + "loss": 0.5159, + "step": 8886 + }, + { + "epoch": 4.824647122692725, + "grad_norm": 12.292228244187442, + "learning_rate": 6.444853951684704e-08, + "loss": 0.2194, + "step": 8887 + }, + { + "epoch": 4.825190010857764, + "grad_norm": 13.565563908072575, + "learning_rate": 6.405052098639065e-08, + "loss": 0.4117, + "step": 8888 + }, + { + "epoch": 4.8257328990228014, + "grad_norm": 13.37247191878539, + "learning_rate": 6.365373135000075e-08, + "loss": 0.3119, + "step": 8889 + }, + { + "epoch": 4.826275787187839, + "grad_norm": 14.204941355106659, + "learning_rate": 6.325817065675366e-08, + "loss": 0.3838, + "step": 8890 + }, + { + "epoch": 4.826818675352877, + "grad_norm": 11.742712728879413, + "learning_rate": 6.286383895557357e-08, + "loss": 0.5932, + "step": 8891 + }, + { + "epoch": 4.827361563517916, + "grad_norm": 10.887835991446542, + "learning_rate": 6.247073629523037e-08, + "loss": 0.3227, + "step": 8892 + }, + { + "epoch": 4.8279044516829535, + "grad_norm": 12.144000261688534, + "learning_rate": 6.207886272434516e-08, + "loss": 0.3384, + "step": 8893 + }, + { + "epoch": 4.828447339847991, + "grad_norm": 12.385730983063644, + "learning_rate": 6.168821829138472e-08, + "loss": 0.2895, + "step": 8894 + }, + { + "epoch": 4.828990228013029, + "grad_norm": 15.682138391036652, + "learning_rate": 6.129880304466484e-08, + "loss": 0.5902, + "step": 8895 + }, + { + "epoch": 4.829533116178068, + "grad_norm": 16.02759596723009, + "learning_rate": 6.091061703234924e-08, + "loss": 0.5285, + "step": 8896 + }, + { + "epoch": 4.8300760043431055, + "grad_norm": 14.27361095848103, + "learning_rate": 6.052366030244838e-08, + "loss": 0.4956, + "step": 8897 + }, + { + "epoch": 4.830618892508143, + "grad_norm": 11.386206832016665, + "learning_rate": 6.013793290282289e-08, + "loss": 0.3047, + "step": 8898 + }, + { + "epoch": 4.831161780673181, + "grad_norm": 10.195950730327096, + "learning_rate": 5.975343488118013e-08, + "loss": 0.3495, + "step": 8899 + }, + { + "epoch": 4.83170466883822, + "grad_norm": 12.21542669488425, + "learning_rate": 5.9370166285073195e-08, + "loss": 0.5064, + "step": 8900 + }, + { + "epoch": 4.8322475570032575, + "grad_norm": 14.215817918389403, + "learning_rate": 5.8988127161908603e-08, + "loss": 0.478, + "step": 8901 + }, + { + "epoch": 4.832790445168295, + "grad_norm": 12.971102099976296, + "learning_rate": 5.8607317558935226e-08, + "loss": 0.374, + "step": 8902 + }, + { + "epoch": 4.833333333333333, + "grad_norm": 7.241451153701665, + "learning_rate": 5.822773752325317e-08, + "loss": 0.3883, + "step": 8903 + }, + { + "epoch": 4.833876221498372, + "grad_norm": 12.147250550465268, + "learning_rate": 5.784938710181043e-08, + "loss": 0.3978, + "step": 8904 + }, + { + "epoch": 4.8344191096634095, + "grad_norm": 18.246476649890155, + "learning_rate": 5.747226634139957e-08, + "loss": 0.3516, + "step": 8905 + }, + { + "epoch": 4.834961997828447, + "grad_norm": 10.69413118790195, + "learning_rate": 5.709637528866552e-08, + "loss": 0.4941, + "step": 8906 + }, + { + "epoch": 4.835504885993485, + "grad_norm": 16.53209092388979, + "learning_rate": 5.672171399009774e-08, + "loss": 0.4765, + "step": 8907 + }, + { + "epoch": 4.836047774158524, + "grad_norm": 13.199658324083233, + "learning_rate": 5.6348282492036946e-08, + "loss": 0.3345, + "step": 8908 + }, + { + "epoch": 4.8365906623235615, + "grad_norm": 15.65690883916084, + "learning_rate": 5.59760808406673e-08, + "loss": 0.4327, + "step": 8909 + }, + { + "epoch": 4.837133550488599, + "grad_norm": 13.187099385246231, + "learning_rate": 5.5605109082024215e-08, + "loss": 0.6508, + "step": 8910 + }, + { + "epoch": 4.837676438653637, + "grad_norm": 15.637262642220758, + "learning_rate": 5.523536726199097e-08, + "loss": 0.3265, + "step": 8911 + }, + { + "epoch": 4.838219326818676, + "grad_norm": 16.12059897202224, + "learning_rate": 5.486685542629655e-08, + "loss": 0.5249, + "step": 8912 + }, + { + "epoch": 4.8387622149837135, + "grad_norm": 13.559084799401953, + "learning_rate": 5.449957362052005e-08, + "loss": 0.5622, + "step": 8913 + }, + { + "epoch": 4.839305103148751, + "grad_norm": 10.42459560738085, + "learning_rate": 5.413352189008736e-08, + "loss": 0.2996, + "step": 8914 + }, + { + "epoch": 4.839847991313789, + "grad_norm": 16.46805590997443, + "learning_rate": 5.3768700280271146e-08, + "loss": 0.5508, + "step": 8915 + }, + { + "epoch": 4.840390879478828, + "grad_norm": 15.003668947871548, + "learning_rate": 5.340510883619421e-08, + "loss": 0.4085, + "step": 8916 + }, + { + "epoch": 4.8409337676438655, + "grad_norm": 8.64870666065282, + "learning_rate": 5.3042747602826125e-08, + "loss": 0.2792, + "step": 8917 + }, + { + "epoch": 4.841476655808903, + "grad_norm": 13.21055869887479, + "learning_rate": 5.268161662498439e-08, + "loss": 0.3822, + "step": 8918 + }, + { + "epoch": 4.842019543973941, + "grad_norm": 15.48767742576848, + "learning_rate": 5.2321715947333264e-08, + "loss": 0.6979, + "step": 8919 + }, + { + "epoch": 4.84256243213898, + "grad_norm": 13.032634397464806, + "learning_rate": 5.196304561438714e-08, + "loss": 0.4676, + "step": 8920 + }, + { + "epoch": 4.8431053203040175, + "grad_norm": 12.108816113672754, + "learning_rate": 5.1605605670506095e-08, + "loss": 0.4098, + "step": 8921 + }, + { + "epoch": 4.843648208469055, + "grad_norm": 9.883044917135045, + "learning_rate": 5.124939615989921e-08, + "loss": 0.3021, + "step": 8922 + }, + { + "epoch": 4.844191096634093, + "grad_norm": 15.02353598376969, + "learning_rate": 5.089441712662346e-08, + "loss": 0.5224, + "step": 8923 + }, + { + "epoch": 4.844733984799132, + "grad_norm": 14.500832790037261, + "learning_rate": 5.054066861458151e-08, + "loss": 0.3961, + "step": 8924 + }, + { + "epoch": 4.8452768729641695, + "grad_norm": 14.045099169119101, + "learning_rate": 5.0188150667528356e-08, + "loss": 0.2919, + "step": 8925 + }, + { + "epoch": 4.845819761129207, + "grad_norm": 11.608813846717336, + "learning_rate": 4.9836863329061346e-08, + "loss": 0.303, + "step": 8926 + }, + { + "epoch": 4.846362649294245, + "grad_norm": 13.766577335723845, + "learning_rate": 4.948680664262906e-08, + "loss": 0.4839, + "step": 8927 + }, + { + "epoch": 4.846905537459284, + "grad_norm": 16.843435011316796, + "learning_rate": 4.913798065152797e-08, + "loss": 0.5213, + "step": 8928 + }, + { + "epoch": 4.8474484256243215, + "grad_norm": 10.74962929723452, + "learning_rate": 4.879038539890024e-08, + "loss": 0.3224, + "step": 8929 + }, + { + "epoch": 4.847991313789359, + "grad_norm": 14.297515517628703, + "learning_rate": 4.844402092773815e-08, + "loss": 0.6642, + "step": 8930 + }, + { + "epoch": 4.848534201954397, + "grad_norm": 10.40231274048184, + "learning_rate": 4.809888728087853e-08, + "loss": 0.2788, + "step": 8931 + }, + { + "epoch": 4.849077090119436, + "grad_norm": 14.099191061940111, + "learning_rate": 4.775498450101057e-08, + "loss": 0.455, + "step": 8932 + }, + { + "epoch": 4.8496199782844736, + "grad_norm": 14.245133009805686, + "learning_rate": 4.7412312630666925e-08, + "loss": 0.4375, + "step": 8933 + }, + { + "epoch": 4.850162866449511, + "grad_norm": 13.88233182279541, + "learning_rate": 4.707087171223146e-08, + "loss": 0.367, + "step": 8934 + }, + { + "epoch": 4.850705754614549, + "grad_norm": 16.896162529240044, + "learning_rate": 4.673066178793262e-08, + "loss": 0.7301, + "step": 8935 + }, + { + "epoch": 4.851248642779588, + "grad_norm": 15.42411861883886, + "learning_rate": 4.6391682899848966e-08, + "loss": 0.5421, + "step": 8936 + }, + { + "epoch": 4.851791530944626, + "grad_norm": 13.430931121094932, + "learning_rate": 4.6053935089905856e-08, + "loss": 0.5728, + "step": 8937 + }, + { + "epoch": 4.852334419109663, + "grad_norm": 11.78549017995068, + "learning_rate": 4.5717418399875426e-08, + "loss": 0.3377, + "step": 8938 + }, + { + "epoch": 4.852877307274701, + "grad_norm": 9.979103448244798, + "learning_rate": 4.5382132871381046e-08, + "loss": 0.2352, + "step": 8939 + }, + { + "epoch": 4.85342019543974, + "grad_norm": 14.36690978126345, + "learning_rate": 4.504807854588844e-08, + "loss": 0.4036, + "step": 8940 + }, + { + "epoch": 4.853963083604778, + "grad_norm": 15.465505131974233, + "learning_rate": 4.4715255464715665e-08, + "loss": 0.3305, + "step": 8941 + }, + { + "epoch": 4.854505971769815, + "grad_norm": 11.075613879055394, + "learning_rate": 4.438366366902647e-08, + "loss": 0.3939, + "step": 8942 + }, + { + "epoch": 4.855048859934853, + "grad_norm": 9.494149902633026, + "learning_rate": 4.4053303199832474e-08, + "loss": 0.455, + "step": 8943 + }, + { + "epoch": 4.855591748099892, + "grad_norm": 13.435606504793506, + "learning_rate": 4.372417409799323e-08, + "loss": 0.7939, + "step": 8944 + }, + { + "epoch": 4.85613463626493, + "grad_norm": 15.032529072722495, + "learning_rate": 4.3396276404216175e-08, + "loss": 0.4496, + "step": 8945 + }, + { + "epoch": 4.856677524429967, + "grad_norm": 15.656375813559288, + "learning_rate": 4.3069610159056644e-08, + "loss": 0.4964, + "step": 8946 + }, + { + "epoch": 4.857220412595005, + "grad_norm": 10.906491414551686, + "learning_rate": 4.274417540291564e-08, + "loss": 0.3617, + "step": 8947 + }, + { + "epoch": 4.857763300760044, + "grad_norm": 13.569426587375315, + "learning_rate": 4.241997217604543e-08, + "loss": 0.5283, + "step": 8948 + }, + { + "epoch": 4.858306188925082, + "grad_norm": 10.691629929006671, + "learning_rate": 4.209700051854282e-08, + "loss": 0.1982, + "step": 8949 + }, + { + "epoch": 4.858849077090119, + "grad_norm": 10.71016148764362, + "learning_rate": 4.177526047035363e-08, + "loss": 0.4083, + "step": 8950 + }, + { + "epoch": 4.859391965255157, + "grad_norm": 14.30919702980143, + "learning_rate": 4.1454752071271586e-08, + "loss": 0.4441, + "step": 8951 + }, + { + "epoch": 4.859934853420196, + "grad_norm": 15.73961142029379, + "learning_rate": 4.11354753609372e-08, + "loss": 0.3984, + "step": 8952 + }, + { + "epoch": 4.860477741585234, + "grad_norm": 17.417654306288096, + "learning_rate": 4.0817430378840004e-08, + "loss": 0.4194, + "step": 8953 + }, + { + "epoch": 4.861020629750271, + "grad_norm": 13.52226655944115, + "learning_rate": 4.050061716431408e-08, + "loss": 0.3684, + "step": 8954 + }, + { + "epoch": 4.861563517915309, + "grad_norm": 9.814382598914019, + "learning_rate": 4.018503575654587e-08, + "loss": 0.7053, + "step": 8955 + }, + { + "epoch": 4.862106406080348, + "grad_norm": 11.24338506908338, + "learning_rate": 3.987068619456635e-08, + "loss": 0.3247, + "step": 8956 + }, + { + "epoch": 4.862649294245386, + "grad_norm": 12.546220960893276, + "learning_rate": 3.955756851725334e-08, + "loss": 0.2883, + "step": 8957 + }, + { + "epoch": 4.863192182410423, + "grad_norm": 9.883047461274098, + "learning_rate": 3.9245682763335846e-08, + "loss": 0.2081, + "step": 8958 + }, + { + "epoch": 4.863735070575461, + "grad_norm": 11.723178509404748, + "learning_rate": 3.8935028971386344e-08, + "loss": 0.2868, + "step": 8959 + }, + { + "epoch": 4.8642779587405, + "grad_norm": 14.517146742383616, + "learning_rate": 3.862560717982855e-08, + "loss": 0.4062, + "step": 8960 + }, + { + "epoch": 4.864820846905538, + "grad_norm": 14.948031344029658, + "learning_rate": 3.8317417426931845e-08, + "loss": 0.5253, + "step": 8961 + }, + { + "epoch": 4.865363735070575, + "grad_norm": 12.386658970702634, + "learning_rate": 3.801045975081352e-08, + "loss": 0.4839, + "step": 8962 + }, + { + "epoch": 4.865906623235613, + "grad_norm": 15.354605902256184, + "learning_rate": 3.770473418943876e-08, + "loss": 0.4133, + "step": 8963 + }, + { + "epoch": 4.866449511400652, + "grad_norm": 10.875285258256444, + "learning_rate": 3.7400240780619543e-08, + "loss": 0.3914, + "step": 8964 + }, + { + "epoch": 4.86699239956569, + "grad_norm": 12.106013147842344, + "learning_rate": 3.709697956201686e-08, + "loss": 0.3433, + "step": 8965 + }, + { + "epoch": 4.867535287730727, + "grad_norm": 13.309078271025745, + "learning_rate": 3.6794950571138464e-08, + "loss": 0.331, + "step": 8966 + }, + { + "epoch": 4.868078175895765, + "grad_norm": 10.487793011045266, + "learning_rate": 3.649415384533894e-08, + "loss": 0.2756, + "step": 8967 + }, + { + "epoch": 4.868621064060804, + "grad_norm": 15.14197511158265, + "learning_rate": 3.619458942182297e-08, + "loss": 0.854, + "step": 8968 + }, + { + "epoch": 4.869163952225842, + "grad_norm": 11.373840117881688, + "learning_rate": 3.589625733763869e-08, + "loss": 0.3237, + "step": 8969 + }, + { + "epoch": 4.869706840390879, + "grad_norm": 14.405255774825118, + "learning_rate": 3.5599157629686576e-08, + "loss": 0.5447, + "step": 8970 + }, + { + "epoch": 4.870249728555917, + "grad_norm": 9.844386093357125, + "learning_rate": 3.5303290334711695e-08, + "loss": 0.4378, + "step": 8971 + }, + { + "epoch": 4.870792616720956, + "grad_norm": 13.588627269277811, + "learning_rate": 3.5008655489306984e-08, + "loss": 0.3705, + "step": 8972 + }, + { + "epoch": 4.871335504885994, + "grad_norm": 10.470214013749253, + "learning_rate": 3.471525312991331e-08, + "loss": 0.3385, + "step": 8973 + }, + { + "epoch": 4.871878393051031, + "grad_norm": 22.377934873090744, + "learning_rate": 3.44230832928194e-08, + "loss": 0.6696, + "step": 8974 + }, + { + "epoch": 4.872421281216069, + "grad_norm": 10.012570917492981, + "learning_rate": 3.4132146014161925e-08, + "loss": 0.3796, + "step": 8975 + }, + { + "epoch": 4.872964169381108, + "grad_norm": 10.734857093886871, + "learning_rate": 3.38424413299232e-08, + "loss": 0.3492, + "step": 8976 + }, + { + "epoch": 4.873507057546146, + "grad_norm": 18.0435053595691, + "learning_rate": 3.3553969275935684e-08, + "loss": 0.5868, + "step": 8977 + }, + { + "epoch": 4.874049945711183, + "grad_norm": 10.021666813244016, + "learning_rate": 3.326672988787638e-08, + "loss": 0.2792, + "step": 8978 + }, + { + "epoch": 4.874592833876221, + "grad_norm": 11.363776145547698, + "learning_rate": 3.298072320127355e-08, + "loss": 0.2911, + "step": 8979 + }, + { + "epoch": 4.875135722041259, + "grad_norm": 16.234875062594668, + "learning_rate": 3.2695949251499994e-08, + "loss": 0.4444, + "step": 8980 + }, + { + "epoch": 4.875678610206298, + "grad_norm": 18.354132401016077, + "learning_rate": 3.2412408073776434e-08, + "loss": 0.8921, + "step": 8981 + }, + { + "epoch": 4.8762214983713354, + "grad_norm": 13.549633551421508, + "learning_rate": 3.2130099703171494e-08, + "loss": 0.4195, + "step": 8982 + }, + { + "epoch": 4.876764386536373, + "grad_norm": 9.25197339319036, + "learning_rate": 3.1849024174603895e-08, + "loss": 0.363, + "step": 8983 + }, + { + "epoch": 4.877307274701412, + "grad_norm": 14.423397162730842, + "learning_rate": 3.1569181522834724e-08, + "loss": 0.5159, + "step": 8984 + }, + { + "epoch": 4.87785016286645, + "grad_norm": 15.097447164063551, + "learning_rate": 3.1290571782476296e-08, + "loss": 0.5308, + "step": 8985 + }, + { + "epoch": 4.8783930510314875, + "grad_norm": 14.75727860602389, + "learning_rate": 3.101319498798883e-08, + "loss": 0.5018, + "step": 8986 + }, + { + "epoch": 4.878935939196525, + "grad_norm": 11.612303819226995, + "learning_rate": 3.073705117367709e-08, + "loss": 0.3928, + "step": 8987 + }, + { + "epoch": 4.879478827361563, + "grad_norm": 14.550764874493284, + "learning_rate": 3.046214037369488e-08, + "loss": 0.4143, + "step": 8988 + }, + { + "epoch": 4.880021715526602, + "grad_norm": 14.242337431144017, + "learning_rate": 3.0188462622045e-08, + "loss": 0.4503, + "step": 8989 + }, + { + "epoch": 4.8805646036916395, + "grad_norm": 14.734943195801437, + "learning_rate": 2.991601795257482e-08, + "loss": 0.3427, + "step": 8990 + }, + { + "epoch": 4.881107491856677, + "grad_norm": 14.592718045998431, + "learning_rate": 2.9644806398982928e-08, + "loss": 0.4573, + "step": 8991 + }, + { + "epoch": 4.881650380021716, + "grad_norm": 12.610312778876645, + "learning_rate": 2.9374827994810283e-08, + "loss": 0.4903, + "step": 8992 + }, + { + "epoch": 4.882193268186754, + "grad_norm": 12.128783807352118, + "learning_rate": 2.9106082773450172e-08, + "loss": 0.4459, + "step": 8993 + }, + { + "epoch": 4.8827361563517915, + "grad_norm": 17.48739197019621, + "learning_rate": 2.8838570768141562e-08, + "loss": 0.7127, + "step": 8994 + }, + { + "epoch": 4.883279044516829, + "grad_norm": 14.579178040970136, + "learning_rate": 2.8572292011969095e-08, + "loss": 0.3715, + "step": 8995 + }, + { + "epoch": 4.883821932681867, + "grad_norm": 11.025778405045106, + "learning_rate": 2.830724653786865e-08, + "loss": 0.2711, + "step": 8996 + }, + { + "epoch": 4.884364820846906, + "grad_norm": 20.20923605960106, + "learning_rate": 2.804343437861956e-08, + "loss": 0.6597, + "step": 8997 + }, + { + "epoch": 4.8849077090119435, + "grad_norm": 9.4089323273495, + "learning_rate": 2.7780855566852394e-08, + "loss": 0.2086, + "step": 8998 + }, + { + "epoch": 4.885450597176981, + "grad_norm": 10.97470388984354, + "learning_rate": 2.7519510135041173e-08, + "loss": 0.297, + "step": 8999 + }, + { + "epoch": 4.88599348534202, + "grad_norm": 12.570657915339005, + "learning_rate": 2.7259398115511148e-08, + "loss": 0.5147, + "step": 9000 + }, + { + "epoch": 4.886536373507058, + "grad_norm": 15.54569627943996, + "learning_rate": 2.7000519540434367e-08, + "loss": 0.4506, + "step": 9001 + }, + { + "epoch": 4.8870792616720955, + "grad_norm": 10.42963878683934, + "learning_rate": 2.6742874441826328e-08, + "loss": 0.3591, + "step": 9002 + }, + { + "epoch": 4.887622149837133, + "grad_norm": 12.875386611254111, + "learning_rate": 2.6486462851554874e-08, + "loss": 0.3134, + "step": 9003 + }, + { + "epoch": 4.888165038002171, + "grad_norm": 9.34456052093577, + "learning_rate": 2.6231284801333524e-08, + "loss": 0.2513, + "step": 9004 + }, + { + "epoch": 4.88870792616721, + "grad_norm": 12.286889941053696, + "learning_rate": 2.597734032272148e-08, + "loss": 0.258, + "step": 9005 + }, + { + "epoch": 4.8892508143322475, + "grad_norm": 18.670865914831435, + "learning_rate": 2.5724629447130278e-08, + "loss": 0.7677, + "step": 9006 + }, + { + "epoch": 4.889793702497285, + "grad_norm": 13.505630369786697, + "learning_rate": 2.547315220581159e-08, + "loss": 0.4204, + "step": 9007 + }, + { + "epoch": 4.890336590662324, + "grad_norm": 13.5847940651091, + "learning_rate": 2.5222908629871647e-08, + "loss": 0.3804, + "step": 9008 + }, + { + "epoch": 4.890879478827362, + "grad_norm": 18.30352602432072, + "learning_rate": 2.497389875025902e-08, + "loss": 0.7192, + "step": 9009 + }, + { + "epoch": 4.8914223669923995, + "grad_norm": 17.37999494699269, + "learning_rate": 2.4726122597773516e-08, + "loss": 0.6299, + "step": 9010 + }, + { + "epoch": 4.891965255157437, + "grad_norm": 10.851195927316295, + "learning_rate": 2.4479580203058408e-08, + "loss": 0.5514, + "step": 9011 + }, + { + "epoch": 4.892508143322475, + "grad_norm": 10.563754398632097, + "learning_rate": 2.423427159660707e-08, + "loss": 0.2451, + "step": 9012 + }, + { + "epoch": 4.893051031487514, + "grad_norm": 11.26826600207582, + "learning_rate": 2.39901968087608e-08, + "loss": 0.3286, + "step": 9013 + }, + { + "epoch": 4.8935939196525515, + "grad_norm": 17.973747561161268, + "learning_rate": 2.3747355869706555e-08, + "loss": 0.5601, + "step": 9014 + }, + { + "epoch": 4.894136807817589, + "grad_norm": 9.432719575335916, + "learning_rate": 2.35057488094792e-08, + "loss": 0.1905, + "step": 9015 + }, + { + "epoch": 4.894679695982628, + "grad_norm": 10.718766904219184, + "learning_rate": 2.3265375657960388e-08, + "loss": 0.4343, + "step": 9016 + }, + { + "epoch": 4.895222584147666, + "grad_norm": 13.174705545759618, + "learning_rate": 2.3026236444880778e-08, + "loss": 0.7121, + "step": 9017 + }, + { + "epoch": 4.8957654723127035, + "grad_norm": 14.581493576657595, + "learning_rate": 2.2788331199816715e-08, + "loss": 0.6438, + "step": 9018 + }, + { + "epoch": 4.896308360477741, + "grad_norm": 13.239127120829922, + "learning_rate": 2.255165995219355e-08, + "loss": 0.3997, + "step": 9019 + }, + { + "epoch": 4.896851248642779, + "grad_norm": 19.46295020201305, + "learning_rate": 2.2316222731282312e-08, + "loss": 0.4649, + "step": 9020 + }, + { + "epoch": 4.897394136807818, + "grad_norm": 10.875751867136424, + "learning_rate": 2.208201956620304e-08, + "loss": 0.3195, + "step": 9021 + }, + { + "epoch": 4.8979370249728555, + "grad_norm": 20.350514144081636, + "learning_rate": 2.184905048592256e-08, + "loss": 0.399, + "step": 9022 + }, + { + "epoch": 4.898479913137893, + "grad_norm": 14.514112929300751, + "learning_rate": 2.161731551925339e-08, + "loss": 0.3543, + "step": 9023 + }, + { + "epoch": 4.899022801302932, + "grad_norm": 14.478901017090632, + "learning_rate": 2.1386814694859258e-08, + "loss": 0.341, + "step": 9024 + }, + { + "epoch": 4.89956568946797, + "grad_norm": 13.744752043693394, + "learning_rate": 2.1157548041246258e-08, + "loss": 0.4873, + "step": 9025 + }, + { + "epoch": 4.900108577633008, + "grad_norm": 9.057494433239784, + "learning_rate": 2.09295155867717e-08, + "loss": 0.217, + "step": 9026 + }, + { + "epoch": 4.900651465798045, + "grad_norm": 15.4247048251968, + "learning_rate": 2.070271735963858e-08, + "loss": 0.5049, + "step": 9027 + }, + { + "epoch": 4.901194353963083, + "grad_norm": 11.973584500311773, + "learning_rate": 2.0477153387898908e-08, + "loss": 0.4135, + "step": 9028 + }, + { + "epoch": 4.901737242128122, + "grad_norm": 15.813841494616677, + "learning_rate": 2.0252823699449254e-08, + "loss": 0.5418, + "step": 9029 + }, + { + "epoch": 4.90228013029316, + "grad_norm": 12.85782607647325, + "learning_rate": 2.0029728322036312e-08, + "loss": 0.5075, + "step": 9030 + }, + { + "epoch": 4.902823018458197, + "grad_norm": 14.609268597455578, + "learning_rate": 1.9807867283251346e-08, + "loss": 0.633, + "step": 9031 + }, + { + "epoch": 4.903365906623236, + "grad_norm": 18.076004594690488, + "learning_rate": 1.958724061053574e-08, + "loss": 0.8642, + "step": 9032 + }, + { + "epoch": 4.903908794788274, + "grad_norm": 14.58349613179072, + "learning_rate": 1.9367848331176554e-08, + "loss": 0.4699, + "step": 9033 + }, + { + "epoch": 4.904451682953312, + "grad_norm": 8.68624975589409, + "learning_rate": 1.9149690472309856e-08, + "loss": 0.2521, + "step": 9034 + }, + { + "epoch": 4.904994571118349, + "grad_norm": 13.439742340445157, + "learning_rate": 1.8932767060915182e-08, + "loss": 0.5178, + "step": 9035 + }, + { + "epoch": 4.905537459283387, + "grad_norm": 10.992659888320869, + "learning_rate": 1.8717078123823286e-08, + "loss": 0.344, + "step": 9036 + }, + { + "epoch": 4.906080347448426, + "grad_norm": 13.160181116268273, + "learning_rate": 1.8502623687711717e-08, + "loss": 0.641, + "step": 9037 + }, + { + "epoch": 4.906623235613464, + "grad_norm": 12.762552089795363, + "learning_rate": 1.828940377910371e-08, + "loss": 0.3898, + "step": 9038 + }, + { + "epoch": 4.907166123778501, + "grad_norm": 9.15219473750769, + "learning_rate": 1.8077418424370387e-08, + "loss": 0.3201, + "step": 9039 + }, + { + "epoch": 4.90770901194354, + "grad_norm": 16.992302430875814, + "learning_rate": 1.7866667649730774e-08, + "loss": 0.5003, + "step": 9040 + }, + { + "epoch": 4.908251900108578, + "grad_norm": 13.106449507444296, + "learning_rate": 1.765715148125069e-08, + "loss": 0.4396, + "step": 9041 + }, + { + "epoch": 4.908794788273616, + "grad_norm": 17.602962269162354, + "learning_rate": 1.744886994484385e-08, + "loss": 0.5396, + "step": 9042 + }, + { + "epoch": 4.909337676438653, + "grad_norm": 12.698432190182416, + "learning_rate": 1.724182306627076e-08, + "loss": 0.3839, + "step": 9043 + }, + { + "epoch": 4.909880564603691, + "grad_norm": 18.435733381982295, + "learning_rate": 1.703601087113871e-08, + "loss": 0.523, + "step": 9044 + }, + { + "epoch": 4.91042345276873, + "grad_norm": 15.182522992411029, + "learning_rate": 1.6831433384904006e-08, + "loss": 0.4973, + "step": 9045 + }, + { + "epoch": 4.910966340933768, + "grad_norm": 16.90243953066863, + "learning_rate": 1.662809063286863e-08, + "loss": 0.419, + "step": 9046 + }, + { + "epoch": 4.911509229098805, + "grad_norm": 15.579122908193515, + "learning_rate": 1.642598264018025e-08, + "loss": 0.4699, + "step": 9047 + }, + { + "epoch": 4.912052117263844, + "grad_norm": 11.883257189390664, + "learning_rate": 1.6225109431839968e-08, + "loss": 0.387, + "step": 9048 + }, + { + "epoch": 4.912595005428882, + "grad_norm": 12.57956767087257, + "learning_rate": 1.6025471032689034e-08, + "loss": 0.3389, + "step": 9049 + }, + { + "epoch": 4.91313789359392, + "grad_norm": 10.023724036259201, + "learning_rate": 1.582706746741991e-08, + "loss": 0.181, + "step": 9050 + }, + { + "epoch": 4.913680781758957, + "grad_norm": 10.904566382599436, + "learning_rate": 1.562989876057075e-08, + "loss": 0.2939, + "step": 9051 + }, + { + "epoch": 4.914223669923995, + "grad_norm": 15.131794978062635, + "learning_rate": 1.5433964936529823e-08, + "loss": 0.5318, + "step": 9052 + }, + { + "epoch": 4.914766558089034, + "grad_norm": 11.260253598807191, + "learning_rate": 1.5239266019527743e-08, + "loss": 0.2825, + "step": 9053 + }, + { + "epoch": 4.915309446254072, + "grad_norm": 8.783484863701455, + "learning_rate": 1.504580203364747e-08, + "loss": 0.2619, + "step": 9054 + }, + { + "epoch": 4.915852334419109, + "grad_norm": 17.289518463767752, + "learning_rate": 1.4853573002815425e-08, + "loss": 0.6613, + "step": 9055 + }, + { + "epoch": 4.916395222584148, + "grad_norm": 10.577132965088307, + "learning_rate": 1.4662578950808137e-08, + "loss": 0.34, + "step": 9056 + }, + { + "epoch": 4.916938110749186, + "grad_norm": 13.481640817817254, + "learning_rate": 1.4472819901246715e-08, + "loss": 0.3756, + "step": 9057 + }, + { + "epoch": 4.917480998914224, + "grad_norm": 14.814161562886865, + "learning_rate": 1.4284295877602383e-08, + "loss": 0.5217, + "step": 9058 + }, + { + "epoch": 4.918023887079261, + "grad_norm": 13.911904587945742, + "learning_rate": 1.4097006903190935e-08, + "loss": 0.4312, + "step": 9059 + }, + { + "epoch": 4.918566775244299, + "grad_norm": 13.559974799811506, + "learning_rate": 1.3910953001176063e-08, + "loss": 0.5598, + "step": 9060 + }, + { + "epoch": 4.919109663409338, + "grad_norm": 12.999506118739305, + "learning_rate": 1.3726134194570472e-08, + "loss": 0.4866, + "step": 9061 + }, + { + "epoch": 4.919652551574376, + "grad_norm": 12.16516581528143, + "learning_rate": 1.3542550506232543e-08, + "loss": 0.4469, + "step": 9062 + }, + { + "epoch": 4.920195439739413, + "grad_norm": 13.76187292005182, + "learning_rate": 1.3360201958867447e-08, + "loss": 0.5088, + "step": 9063 + }, + { + "epoch": 4.920738327904452, + "grad_norm": 11.597669183780656, + "learning_rate": 1.3179088575029365e-08, + "loss": 0.3892, + "step": 9064 + }, + { + "epoch": 4.92128121606949, + "grad_norm": 13.03473175291288, + "learning_rate": 1.2999210377118155e-08, + "loss": 0.3606, + "step": 9065 + }, + { + "epoch": 4.921824104234528, + "grad_norm": 11.618399948981745, + "learning_rate": 1.2820567387381577e-08, + "loss": 0.431, + "step": 9066 + }, + { + "epoch": 4.922366992399565, + "grad_norm": 14.444651618891138, + "learning_rate": 1.2643159627914182e-08, + "loss": 0.4671, + "step": 9067 + }, + { + "epoch": 4.922909880564603, + "grad_norm": 19.413144717406254, + "learning_rate": 1.2466987120658413e-08, + "loss": 0.4479, + "step": 9068 + }, + { + "epoch": 4.923452768729642, + "grad_norm": 15.188364084941032, + "learning_rate": 1.2292049887403513e-08, + "loss": 0.518, + "step": 9069 + }, + { + "epoch": 4.92399565689468, + "grad_norm": 17.595339652436596, + "learning_rate": 1.2118347949785503e-08, + "loss": 0.5491, + "step": 9070 + }, + { + "epoch": 4.924538545059717, + "grad_norm": 13.811004803619682, + "learning_rate": 1.1945881329289422e-08, + "loss": 0.521, + "step": 9071 + }, + { + "epoch": 4.925081433224756, + "grad_norm": 17.210661605650973, + "learning_rate": 1.1774650047244873e-08, + "loss": 0.4881, + "step": 9072 + }, + { + "epoch": 4.925624321389794, + "grad_norm": 11.74172359244011, + "learning_rate": 1.1604654124830472e-08, + "loss": 0.2734, + "step": 9073 + }, + { + "epoch": 4.926167209554832, + "grad_norm": 12.999583685539983, + "learning_rate": 1.1435893583071623e-08, + "loss": 0.4139, + "step": 9074 + }, + { + "epoch": 4.9267100977198695, + "grad_norm": 18.891052821295602, + "learning_rate": 1.1268368442840516e-08, + "loss": 0.6411, + "step": 9075 + }, + { + "epoch": 4.927252985884907, + "grad_norm": 9.732481550986785, + "learning_rate": 1.1102078724857247e-08, + "loss": 0.2768, + "step": 9076 + }, + { + "epoch": 4.927795874049946, + "grad_norm": 13.879976567786102, + "learning_rate": 1.0937024449688693e-08, + "loss": 0.3334, + "step": 9077 + }, + { + "epoch": 4.928338762214984, + "grad_norm": 14.2590437939657, + "learning_rate": 1.0773205637749639e-08, + "loss": 0.5803, + "step": 9078 + }, + { + "epoch": 4.9288816503800215, + "grad_norm": 12.223213148169753, + "learning_rate": 1.0610622309300544e-08, + "loss": 0.3341, + "step": 9079 + }, + { + "epoch": 4.92942453854506, + "grad_norm": 9.799661483908018, + "learning_rate": 1.0449274484450877e-08, + "loss": 0.2622, + "step": 9080 + }, + { + "epoch": 4.929967426710098, + "grad_norm": 12.782439661304597, + "learning_rate": 1.0289162183155788e-08, + "loss": 0.4998, + "step": 9081 + }, + { + "epoch": 4.930510314875136, + "grad_norm": 17.86302762326864, + "learning_rate": 1.0130285425218322e-08, + "loss": 0.6778, + "step": 9082 + }, + { + "epoch": 4.9310532030401735, + "grad_norm": 7.999286008192372, + "learning_rate": 9.97264423028832e-09, + "loss": 0.2896, + "step": 9083 + }, + { + "epoch": 4.931596091205211, + "grad_norm": 18.706243688121948, + "learning_rate": 9.816238617864626e-09, + "loss": 0.5758, + "step": 9084 + }, + { + "epoch": 4.93213897937025, + "grad_norm": 12.391173965997686, + "learning_rate": 9.661068607288437e-09, + "loss": 0.2557, + "step": 9085 + }, + { + "epoch": 4.932681867535288, + "grad_norm": 13.206506481368905, + "learning_rate": 9.50713421775551e-09, + "loss": 0.4463, + "step": 9086 + }, + { + "epoch": 4.9332247557003255, + "grad_norm": 12.900701770071388, + "learning_rate": 9.354435468301726e-09, + "loss": 0.459, + "step": 9087 + }, + { + "epoch": 4.933767643865364, + "grad_norm": 12.068288986888435, + "learning_rate": 9.202972377814201e-09, + "loss": 0.3485, + "step": 9088 + }, + { + "epoch": 4.934310532030402, + "grad_norm": 12.75825787671443, + "learning_rate": 9.052744965025728e-09, + "loss": 0.4755, + "step": 9089 + }, + { + "epoch": 4.93485342019544, + "grad_norm": 11.99221513373725, + "learning_rate": 8.903753248516999e-09, + "loss": 0.3729, + "step": 9090 + }, + { + "epoch": 4.9353963083604775, + "grad_norm": 12.257794064544841, + "learning_rate": 8.755997246715498e-09, + "loss": 0.3006, + "step": 9091 + }, + { + "epoch": 4.935939196525515, + "grad_norm": 11.554301500218191, + "learning_rate": 8.609476977896602e-09, + "loss": 0.4357, + "step": 9092 + }, + { + "epoch": 4.936482084690554, + "grad_norm": 12.78413585140587, + "learning_rate": 8.464192460181376e-09, + "loss": 0.3639, + "step": 9093 + }, + { + "epoch": 4.937024972855592, + "grad_norm": 12.250703901070336, + "learning_rate": 8.320143711538775e-09, + "loss": 0.3514, + "step": 9094 + }, + { + "epoch": 4.9375678610206295, + "grad_norm": 14.649810047012926, + "learning_rate": 8.17733074978455e-09, + "loss": 0.3402, + "step": 9095 + }, + { + "epoch": 4.938110749185668, + "grad_norm": 17.43504427698499, + "learning_rate": 8.035753592583461e-09, + "loss": 0.5377, + "step": 9096 + }, + { + "epoch": 4.938653637350706, + "grad_norm": 14.034004629213161, + "learning_rate": 7.895412257444834e-09, + "loss": 0.397, + "step": 9097 + }, + { + "epoch": 4.939196525515744, + "grad_norm": 11.963493138309827, + "learning_rate": 7.756306761725895e-09, + "loss": 0.2919, + "step": 9098 + }, + { + "epoch": 4.9397394136807815, + "grad_norm": 15.061307897752748, + "learning_rate": 7.618437122632882e-09, + "loss": 0.6577, + "step": 9099 + }, + { + "epoch": 4.940282301845819, + "grad_norm": 14.852941299011642, + "learning_rate": 7.48180335721771e-09, + "loss": 0.5519, + "step": 9100 + }, + { + "epoch": 4.940825190010858, + "grad_norm": 12.814384068391547, + "learning_rate": 7.346405482377972e-09, + "loss": 0.5925, + "step": 9101 + }, + { + "epoch": 4.941368078175896, + "grad_norm": 9.563813812350007, + "learning_rate": 7.2122435148613835e-09, + "loss": 0.2204, + "step": 9102 + }, + { + "epoch": 4.9419109663409335, + "grad_norm": 9.564014825276837, + "learning_rate": 7.079317471260228e-09, + "loss": 0.2511, + "step": 9103 + }, + { + "epoch": 4.942453854505972, + "grad_norm": 17.38602510454561, + "learning_rate": 6.947627368015797e-09, + "loss": 0.5656, + "step": 9104 + }, + { + "epoch": 4.94299674267101, + "grad_norm": 9.577450195791936, + "learning_rate": 6.817173221415063e-09, + "loss": 0.2404, + "step": 9105 + }, + { + "epoch": 4.943539630836048, + "grad_norm": 11.92192298760108, + "learning_rate": 6.687955047592898e-09, + "loss": 0.3116, + "step": 9106 + }, + { + "epoch": 4.9440825190010855, + "grad_norm": 14.111860080166249, + "learning_rate": 6.559972862530961e-09, + "loss": 0.4142, + "step": 9107 + }, + { + "epoch": 4.944625407166123, + "grad_norm": 16.85584479160711, + "learning_rate": 6.433226682059923e-09, + "loss": 0.4438, + "step": 9108 + }, + { + "epoch": 4.945168295331162, + "grad_norm": 12.407679682981549, + "learning_rate": 6.307716521853912e-09, + "loss": 0.7122, + "step": 9109 + }, + { + "epoch": 4.9457111834962, + "grad_norm": 18.57218994539643, + "learning_rate": 6.183442397437178e-09, + "loss": 0.4333, + "step": 9110 + }, + { + "epoch": 4.9462540716612375, + "grad_norm": 10.179606005226988, + "learning_rate": 6.0604043241807574e-09, + "loss": 0.2016, + "step": 9111 + }, + { + "epoch": 4.946796959826276, + "grad_norm": 11.47232407870111, + "learning_rate": 5.938602317301368e-09, + "loss": 0.4277, + "step": 9112 + }, + { + "epoch": 4.947339847991314, + "grad_norm": 11.739066706652215, + "learning_rate": 5.8180363918625135e-09, + "loss": 0.5738, + "step": 9113 + }, + { + "epoch": 4.947882736156352, + "grad_norm": 10.786798778444865, + "learning_rate": 5.698706562778933e-09, + "loss": 0.3646, + "step": 9114 + }, + { + "epoch": 4.9484256243213895, + "grad_norm": 14.9560917250246, + "learning_rate": 5.580612844806599e-09, + "loss": 0.4187, + "step": 9115 + }, + { + "epoch": 4.948968512486427, + "grad_norm": 8.906425035542007, + "learning_rate": 5.4637552525538266e-09, + "loss": 0.2081, + "step": 9116 + }, + { + "epoch": 4.949511400651466, + "grad_norm": 15.606788834694017, + "learning_rate": 5.3481338004723886e-09, + "loss": 0.533, + "step": 9117 + }, + { + "epoch": 4.950054288816504, + "grad_norm": 12.28419381741561, + "learning_rate": 5.233748502861957e-09, + "loss": 0.5334, + "step": 9118 + }, + { + "epoch": 4.950597176981542, + "grad_norm": 12.227392912720493, + "learning_rate": 5.120599373872326e-09, + "loss": 0.2228, + "step": 9119 + }, + { + "epoch": 4.95114006514658, + "grad_norm": 16.11553058072441, + "learning_rate": 5.008686427495635e-09, + "loss": 0.4372, + "step": 9120 + }, + { + "epoch": 4.951682953311618, + "grad_norm": 13.761040611076448, + "learning_rate": 4.898009677574145e-09, + "loss": 0.4973, + "step": 9121 + }, + { + "epoch": 4.952225841476656, + "grad_norm": 10.73617260071943, + "learning_rate": 4.788569137796905e-09, + "loss": 0.4157, + "step": 9122 + }, + { + "epoch": 4.952768729641694, + "grad_norm": 20.53036734807641, + "learning_rate": 4.6803648216986465e-09, + "loss": 0.3275, + "step": 9123 + }, + { + "epoch": 4.953311617806731, + "grad_norm": 10.707480347228744, + "learning_rate": 4.5733967426642155e-09, + "loss": 0.3827, + "step": 9124 + }, + { + "epoch": 4.95385450597177, + "grad_norm": 17.496584887499267, + "learning_rate": 4.467664913921921e-09, + "loss": 0.7894, + "step": 9125 + }, + { + "epoch": 4.954397394136808, + "grad_norm": 13.01334205119194, + "learning_rate": 4.363169348549079e-09, + "loss": 0.2762, + "step": 9126 + }, + { + "epoch": 4.954940282301846, + "grad_norm": 19.033412027447728, + "learning_rate": 4.259910059470906e-09, + "loss": 0.4458, + "step": 9127 + }, + { + "epoch": 4.955483170466884, + "grad_norm": 12.446627215988766, + "learning_rate": 4.1578870594571885e-09, + "loss": 0.3732, + "step": 9128 + }, + { + "epoch": 4.956026058631922, + "grad_norm": 11.854175020272, + "learning_rate": 4.05710036112672e-09, + "loss": 0.412, + "step": 9129 + }, + { + "epoch": 4.95656894679696, + "grad_norm": 13.145274907827265, + "learning_rate": 3.957549976946195e-09, + "loss": 0.376, + "step": 9130 + }, + { + "epoch": 4.957111834961998, + "grad_norm": 14.159905703597753, + "learning_rate": 3.859235919226878e-09, + "loss": 0.4783, + "step": 9131 + }, + { + "epoch": 4.957654723127035, + "grad_norm": 10.76209944745829, + "learning_rate": 3.76215820012793e-09, + "loss": 0.3364, + "step": 9132 + }, + { + "epoch": 4.958197611292074, + "grad_norm": 18.14425836712623, + "learning_rate": 3.666316831657524e-09, + "loss": 0.3815, + "step": 9133 + }, + { + "epoch": 4.958740499457112, + "grad_norm": 9.071986096860721, + "learning_rate": 3.5717118256684004e-09, + "loss": 0.3018, + "step": 9134 + }, + { + "epoch": 4.95928338762215, + "grad_norm": 11.250761445029084, + "learning_rate": 3.4783431938623103e-09, + "loss": 0.3972, + "step": 9135 + }, + { + "epoch": 4.959826275787188, + "grad_norm": 14.397121807180921, + "learning_rate": 3.3862109477877935e-09, + "loss": 0.6723, + "step": 9136 + }, + { + "epoch": 4.960369163952226, + "grad_norm": 13.645836729945032, + "learning_rate": 3.295315098837959e-09, + "loss": 0.3039, + "step": 9137 + }, + { + "epoch": 4.960912052117264, + "grad_norm": 13.683432725518449, + "learning_rate": 3.2056556582549247e-09, + "loss": 0.4265, + "step": 9138 + }, + { + "epoch": 4.961454940282302, + "grad_norm": 13.894633750793059, + "learning_rate": 3.117232637130929e-09, + "loss": 0.3269, + "step": 9139 + }, + { + "epoch": 4.961997828447339, + "grad_norm": 12.71603084534765, + "learning_rate": 3.0300460463983384e-09, + "loss": 0.4258, + "step": 9140 + }, + { + "epoch": 4.962540716612378, + "grad_norm": 12.764000039058175, + "learning_rate": 2.944095896844079e-09, + "loss": 0.2995, + "step": 9141 + }, + { + "epoch": 4.963083604777416, + "grad_norm": 9.50398515456868, + "learning_rate": 2.8593821990963167e-09, + "loss": 0.3252, + "step": 9142 + }, + { + "epoch": 4.963626492942454, + "grad_norm": 10.26064116609969, + "learning_rate": 2.7759049636322254e-09, + "loss": 0.4138, + "step": 9143 + }, + { + "epoch": 4.964169381107492, + "grad_norm": 14.706063028722339, + "learning_rate": 2.6936642007791004e-09, + "loss": 0.3732, + "step": 9144 + }, + { + "epoch": 4.96471226927253, + "grad_norm": 16.446836987033343, + "learning_rate": 2.612659920705474e-09, + "loss": 0.6265, + "step": 9145 + }, + { + "epoch": 4.965255157437568, + "grad_norm": 13.694193607083827, + "learning_rate": 2.5328921334322186e-09, + "loss": 0.3205, + "step": 9146 + }, + { + "epoch": 4.965798045602606, + "grad_norm": 21.42453240685977, + "learning_rate": 2.4543608488247774e-09, + "loss": 0.5323, + "step": 9147 + }, + { + "epoch": 4.966340933767643, + "grad_norm": 17.278512554654093, + "learning_rate": 2.3770660765953803e-09, + "loss": 0.4915, + "step": 9148 + }, + { + "epoch": 4.966883821932682, + "grad_norm": 11.488248878514298, + "learning_rate": 2.301007826304158e-09, + "loss": 0.4663, + "step": 9149 + }, + { + "epoch": 4.96742671009772, + "grad_norm": 12.059564726501527, + "learning_rate": 2.2261861073580306e-09, + "loss": 0.354, + "step": 9150 + }, + { + "epoch": 4.967969598262758, + "grad_norm": 12.062263452100884, + "learning_rate": 2.1526009290118165e-09, + "loss": 0.4141, + "step": 9151 + }, + { + "epoch": 4.968512486427796, + "grad_norm": 13.560790820046249, + "learning_rate": 2.0802523003660146e-09, + "loss": 0.4742, + "step": 9152 + }, + { + "epoch": 4.969055374592834, + "grad_norm": 12.550658712203877, + "learning_rate": 2.009140230369022e-09, + "loss": 0.5253, + "step": 9153 + }, + { + "epoch": 4.969598262757872, + "grad_norm": 12.71176738765954, + "learning_rate": 1.939264727816026e-09, + "loss": 0.4024, + "step": 9154 + }, + { + "epoch": 4.97014115092291, + "grad_norm": 11.734799781581923, + "learning_rate": 1.870625801349002e-09, + "loss": 0.2679, + "step": 9155 + }, + { + "epoch": 4.970684039087947, + "grad_norm": 13.891313537034907, + "learning_rate": 1.8032234594578257e-09, + "loss": 0.2977, + "step": 9156 + }, + { + "epoch": 4.971226927252986, + "grad_norm": 13.309663904879017, + "learning_rate": 1.7370577104780518e-09, + "loss": 0.6025, + "step": 9157 + }, + { + "epoch": 4.971769815418024, + "grad_norm": 20.714422472932196, + "learning_rate": 1.672128562594244e-09, + "loss": 0.7059, + "step": 9158 + }, + { + "epoch": 4.972312703583062, + "grad_norm": 16.889615399656492, + "learning_rate": 1.6084360238366458e-09, + "loss": 0.5564, + "step": 9159 + }, + { + "epoch": 4.9728555917481, + "grad_norm": 16.399562625707034, + "learning_rate": 1.5459801020833999e-09, + "loss": 0.5455, + "step": 9160 + }, + { + "epoch": 4.973398479913138, + "grad_norm": 13.826143902617954, + "learning_rate": 1.484760805057217e-09, + "loss": 0.4126, + "step": 9161 + }, + { + "epoch": 4.973941368078176, + "grad_norm": 11.172428599335714, + "learning_rate": 1.4247781403320393e-09, + "loss": 0.4096, + "step": 9162 + }, + { + "epoch": 4.974484256243214, + "grad_norm": 16.468489977898997, + "learning_rate": 1.3660321153252665e-09, + "loss": 0.4705, + "step": 9163 + }, + { + "epoch": 4.9750271444082514, + "grad_norm": 14.439166503302797, + "learning_rate": 1.3085227373033082e-09, + "loss": 0.742, + "step": 9164 + }, + { + "epoch": 4.97557003257329, + "grad_norm": 13.576277807739485, + "learning_rate": 1.2522500133782533e-09, + "loss": 0.3955, + "step": 9165 + }, + { + "epoch": 4.976112920738328, + "grad_norm": 10.042503188679653, + "learning_rate": 1.1972139505112002e-09, + "loss": 0.2315, + "step": 9166 + }, + { + "epoch": 4.976655808903366, + "grad_norm": 15.56377868085397, + "learning_rate": 1.1434145555078157e-09, + "loss": 0.6752, + "step": 9167 + }, + { + "epoch": 4.977198697068404, + "grad_norm": 14.579485607298796, + "learning_rate": 1.0908518350238873e-09, + "loss": 0.4866, + "step": 9168 + }, + { + "epoch": 4.977741585233442, + "grad_norm": 16.774796454586042, + "learning_rate": 1.0395257955586602e-09, + "loss": 0.5845, + "step": 9169 + }, + { + "epoch": 4.97828447339848, + "grad_norm": 11.815529585109633, + "learning_rate": 9.894364434603899e-10, + "loss": 0.4916, + "step": 9170 + }, + { + "epoch": 4.978827361563518, + "grad_norm": 14.632698898536018, + "learning_rate": 9.405837849252308e-10, + "loss": 0.4883, + "step": 9171 + }, + { + "epoch": 4.9793702497285555, + "grad_norm": 14.57532070039393, + "learning_rate": 8.929678259939067e-10, + "loss": 0.3583, + "step": 9172 + }, + { + "epoch": 4.979913137893594, + "grad_norm": 18.102220051162018, + "learning_rate": 8.465885725572609e-10, + "loss": 0.5808, + "step": 9173 + }, + { + "epoch": 4.980456026058632, + "grad_norm": 15.979995931593416, + "learning_rate": 8.014460303507054e-10, + "loss": 0.2909, + "step": 9174 + }, + { + "epoch": 4.98099891422367, + "grad_norm": 12.010762984710725, + "learning_rate": 7.57540204957552e-10, + "loss": 0.4328, + "step": 9175 + }, + { + "epoch": 4.981541802388708, + "grad_norm": 9.665802442722407, + "learning_rate": 7.148711018090115e-10, + "loss": 0.3006, + "step": 9176 + }, + { + "epoch": 4.982084690553746, + "grad_norm": 12.841279252555506, + "learning_rate": 6.734387261808639e-10, + "loss": 0.4085, + "step": 9177 + }, + { + "epoch": 4.982627578718784, + "grad_norm": 17.41329959247634, + "learning_rate": 6.332430831978986e-10, + "loss": 0.4756, + "step": 9178 + }, + { + "epoch": 4.983170466883822, + "grad_norm": 15.463364790372648, + "learning_rate": 5.942841778328045e-10, + "loss": 0.4864, + "step": 9179 + }, + { + "epoch": 4.9837133550488595, + "grad_norm": 18.266485807419585, + "learning_rate": 5.565620149039497e-10, + "loss": 0.5961, + "step": 9180 + }, + { + "epoch": 4.984256243213898, + "grad_norm": 16.902373526165416, + "learning_rate": 5.200765990753809e-10, + "loss": 0.5701, + "step": 9181 + }, + { + "epoch": 4.984799131378936, + "grad_norm": 19.631248888899865, + "learning_rate": 4.848279348601547e-10, + "loss": 0.6822, + "step": 9182 + }, + { + "epoch": 4.985342019543974, + "grad_norm": 12.12267583241551, + "learning_rate": 4.5081602661922743e-10, + "loss": 0.4016, + "step": 9183 + }, + { + "epoch": 4.985884907709012, + "grad_norm": 13.057551391655595, + "learning_rate": 4.180408785570134e-10, + "loss": 0.415, + "step": 9184 + }, + { + "epoch": 4.98642779587405, + "grad_norm": 17.345548037710916, + "learning_rate": 3.8650249472915735e-10, + "loss": 0.6871, + "step": 9185 + }, + { + "epoch": 4.986970684039088, + "grad_norm": 13.256176178212948, + "learning_rate": 3.562008790358729e-10, + "loss": 0.5344, + "step": 9186 + }, + { + "epoch": 4.987513572204126, + "grad_norm": 14.72682874967264, + "learning_rate": 3.271360352252728e-10, + "loss": 0.4598, + "step": 9187 + }, + { + "epoch": 4.9880564603691635, + "grad_norm": 11.832103539399576, + "learning_rate": 2.993079668911492e-10, + "loss": 0.314, + "step": 9188 + }, + { + "epoch": 4.988599348534202, + "grad_norm": 16.48759487025403, + "learning_rate": 2.7271667747519327e-10, + "loss": 0.4666, + "step": 9189 + }, + { + "epoch": 4.98914223669924, + "grad_norm": 11.438831531197014, + "learning_rate": 2.473621702669959e-10, + "loss": 0.3144, + "step": 9190 + }, + { + "epoch": 4.989685124864278, + "grad_norm": 16.2603783146197, + "learning_rate": 2.2324444840293723e-10, + "loss": 0.4698, + "step": 9191 + }, + { + "epoch": 4.990228013029316, + "grad_norm": 13.65337346906678, + "learning_rate": 2.003635148650762e-10, + "loss": 0.296, + "step": 9192 + }, + { + "epoch": 4.990770901194354, + "grad_norm": 13.609445308581654, + "learning_rate": 1.7871937248337134e-10, + "loss": 0.411, + "step": 9193 + }, + { + "epoch": 4.991313789359392, + "grad_norm": 16.227775980791364, + "learning_rate": 1.5831202393457035e-10, + "loss": 0.3289, + "step": 9194 + }, + { + "epoch": 4.99185667752443, + "grad_norm": 18.790945613812905, + "learning_rate": 1.3914147174332037e-10, + "loss": 0.6441, + "step": 9195 + }, + { + "epoch": 4.9923995656894675, + "grad_norm": 11.25550609886848, + "learning_rate": 1.2120771828105783e-10, + "loss": 0.3092, + "step": 9196 + }, + { + "epoch": 4.992942453854506, + "grad_norm": 11.398167541292397, + "learning_rate": 1.0451076576489805e-10, + "loss": 0.2813, + "step": 9197 + }, + { + "epoch": 4.993485342019544, + "grad_norm": 11.282564692027671, + "learning_rate": 8.905061625985589e-11, + "loss": 0.3401, + "step": 9198 + }, + { + "epoch": 4.994028230184582, + "grad_norm": 20.099380241860953, + "learning_rate": 7.482727167884563e-11, + "loss": 0.5747, + "step": 9199 + }, + { + "epoch": 4.99457111834962, + "grad_norm": 11.476742325586816, + "learning_rate": 6.184073378046051e-11, + "loss": 0.2744, + "step": 9200 + }, + { + "epoch": 4.995114006514658, + "grad_norm": 9.30054079035995, + "learning_rate": 5.009100417119328e-11, + "loss": 0.1959, + "step": 9201 + }, + { + "epoch": 4.995656894679696, + "grad_norm": 17.192495762416968, + "learning_rate": 3.957808430432586e-11, + "loss": 0.4294, + "step": 9202 + }, + { + "epoch": 4.996199782844734, + "grad_norm": 15.84133950781789, + "learning_rate": 3.030197548103964e-11, + "loss": 0.4281, + "step": 9203 + }, + { + "epoch": 4.9967426710097715, + "grad_norm": 11.498444122453641, + "learning_rate": 2.226267884597455e-11, + "loss": 0.2538, + "step": 9204 + }, + { + "epoch": 4.99728555917481, + "grad_norm": 10.709506205461455, + "learning_rate": 1.5460195396110878e-11, + "loss": 0.3601, + "step": 9205 + }, + { + "epoch": 4.997828447339848, + "grad_norm": 11.943394957532275, + "learning_rate": 9.894525970777224e-12, + "loss": 0.2582, + "step": 9206 + }, + { + "epoch": 4.998371335504886, + "grad_norm": 18.015533305967693, + "learning_rate": 5.5656712605323126e-12, + "loss": 0.5477, + "step": 9207 + }, + { + "epoch": 4.998914223669924, + "grad_norm": 13.226875446527242, + "learning_rate": 2.4736317982831936e-12, + "loss": 0.4569, + "step": 9208 + }, + { + "epoch": 4.999457111834962, + "grad_norm": 13.424018021065068, + "learning_rate": 6.184079692772571e-13, + "loss": 0.3316, + "step": 9209 + }, + { + "epoch": 5.0, + "grad_norm": 15.798499583731147, + "learning_rate": 0.0, + "loss": 0.5066, + "step": 9210 + } + ], + "logging_steps": 1.0, + "max_steps": 9210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.098902245736448e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/training_args.bin b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b16de63f342ea53d59c4d97aa8a3bd463a79531f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9801fdd09c6ba798e12b77f77221db3b5747b6e1296785789d44cf2bc492ff9e +size 7160 diff --git a/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/zero_to_fp32.py b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/checkpoint-9210/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/llava-next-video-7b_lora-True_qlora-False/trainer_state.json b/llava-next-video-7b_lora-True_qlora-False/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..38f2368aa348c1fbfdeb30d5a7a3d7b8d9b04452 --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/trainer_state.json @@ -0,0 +1,64512 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 9210, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005428881650380022, + "grad_norm": 2.1320506281889067, + "learning_rate": 7.220216606498195e-08, + "loss": 5.1936, + "step": 1 + }, + { + "epoch": 0.0010857763300760044, + "grad_norm": 1.8633488766761366, + "learning_rate": 1.444043321299639e-07, + "loss": 4.6995, + "step": 2 + }, + { + "epoch": 0.0016286644951140066, + "grad_norm": 2.063286098879776, + "learning_rate": 2.1660649819494586e-07, + "loss": 4.4902, + "step": 3 + }, + { + "epoch": 0.002171552660152009, + "grad_norm": 1.7123050230907495, + "learning_rate": 2.888086642599278e-07, + "loss": 4.9125, + "step": 4 + }, + { + "epoch": 0.0027144408251900108, + "grad_norm": 1.7499657276799891, + "learning_rate": 3.610108303249098e-07, + "loss": 3.7648, + "step": 5 + }, + { + "epoch": 0.003257328990228013, + "grad_norm": 2.0419045636066886, + "learning_rate": 4.332129963898917e-07, + "loss": 4.7842, + "step": 6 + }, + { + "epoch": 0.003800217155266015, + "grad_norm": 1.4960282529105982, + "learning_rate": 5.054151624548736e-07, + "loss": 4.9333, + "step": 7 + }, + { + "epoch": 0.004343105320304018, + "grad_norm": 1.640954967476452, + "learning_rate": 5.776173285198556e-07, + "loss": 3.9576, + "step": 8 + }, + { + "epoch": 0.004885993485342019, + "grad_norm": 2.0590500459610244, + "learning_rate": 6.498194945848375e-07, + "loss": 4.3833, + "step": 9 + }, + { + "epoch": 0.0054288816503800215, + "grad_norm": 1.07709353257816, + "learning_rate": 7.220216606498196e-07, + "loss": 3.7669, + "step": 10 + }, + { + "epoch": 0.005971769815418024, + "grad_norm": 1.9506235655782926, + "learning_rate": 7.942238267148016e-07, + "loss": 4.5975, + "step": 11 + }, + { + "epoch": 0.006514657980456026, + "grad_norm": 1.411983096954528, + "learning_rate": 8.664259927797834e-07, + "loss": 3.9706, + "step": 12 + }, + { + "epoch": 0.007057546145494028, + "grad_norm": 1.7110053131090028, + "learning_rate": 9.386281588447654e-07, + "loss": 4.9081, + "step": 13 + }, + { + "epoch": 0.00760043431053203, + "grad_norm": 1.3689315757199692, + "learning_rate": 1.0108303249097473e-06, + "loss": 4.2426, + "step": 14 + }, + { + "epoch": 0.008143322475570033, + "grad_norm": 1.5612091671030481, + "learning_rate": 1.0830324909747293e-06, + "loss": 4.5144, + "step": 15 + }, + { + "epoch": 0.008686210640608035, + "grad_norm": 1.3725168953865754, + "learning_rate": 1.1552346570397112e-06, + "loss": 4.9378, + "step": 16 + }, + { + "epoch": 0.009229098805646038, + "grad_norm": 2.027359913808926, + "learning_rate": 1.2274368231046932e-06, + "loss": 3.7477, + "step": 17 + }, + { + "epoch": 0.009771986970684038, + "grad_norm": 2.1695456962160224, + "learning_rate": 1.299638989169675e-06, + "loss": 5.7636, + "step": 18 + }, + { + "epoch": 0.01031487513572204, + "grad_norm": 3.2637002287680428, + "learning_rate": 1.3718411552346572e-06, + "loss": 6.1883, + "step": 19 + }, + { + "epoch": 0.010857763300760043, + "grad_norm": 2.5723616120643857, + "learning_rate": 1.4440433212996392e-06, + "loss": 4.0267, + "step": 20 + }, + { + "epoch": 0.011400651465798045, + "grad_norm": 1.4473754577877487, + "learning_rate": 1.516245487364621e-06, + "loss": 4.526, + "step": 21 + }, + { + "epoch": 0.011943539630836048, + "grad_norm": 1.6098208731761, + "learning_rate": 1.5884476534296031e-06, + "loss": 4.5163, + "step": 22 + }, + { + "epoch": 0.01248642779587405, + "grad_norm": 2.899624682573202, + "learning_rate": 1.6606498194945849e-06, + "loss": 4.8188, + "step": 23 + }, + { + "epoch": 0.013029315960912053, + "grad_norm": 2.273571202553131, + "learning_rate": 1.7328519855595669e-06, + "loss": 5.0074, + "step": 24 + }, + { + "epoch": 0.013572204125950055, + "grad_norm": 1.1877323869171315, + "learning_rate": 1.805054151624549e-06, + "loss": 4.4331, + "step": 25 + }, + { + "epoch": 0.014115092290988056, + "grad_norm": 1.9365501019385494, + "learning_rate": 1.8772563176895308e-06, + "loss": 5.0525, + "step": 26 + }, + { + "epoch": 0.014657980456026058, + "grad_norm": 1.7710485953105986, + "learning_rate": 1.949458483754513e-06, + "loss": 4.1346, + "step": 27 + }, + { + "epoch": 0.01520086862106406, + "grad_norm": 1.231848881292397, + "learning_rate": 2.0216606498194946e-06, + "loss": 4.2431, + "step": 28 + }, + { + "epoch": 0.015743756786102063, + "grad_norm": 1.6372838462475754, + "learning_rate": 2.0938628158844768e-06, + "loss": 5.4317, + "step": 29 + }, + { + "epoch": 0.016286644951140065, + "grad_norm": 1.7326568199523718, + "learning_rate": 2.1660649819494585e-06, + "loss": 4.4774, + "step": 30 + }, + { + "epoch": 0.016829533116178068, + "grad_norm": 2.1315641946923387, + "learning_rate": 2.2382671480144407e-06, + "loss": 4.1418, + "step": 31 + }, + { + "epoch": 0.01737242128121607, + "grad_norm": 1.916538366799816, + "learning_rate": 2.3104693140794225e-06, + "loss": 5.7926, + "step": 32 + }, + { + "epoch": 0.017915309446254073, + "grad_norm": 2.057531274363787, + "learning_rate": 2.3826714801444047e-06, + "loss": 4.2862, + "step": 33 + }, + { + "epoch": 0.018458197611292075, + "grad_norm": 2.6727877745045023, + "learning_rate": 2.4548736462093864e-06, + "loss": 5.1323, + "step": 34 + }, + { + "epoch": 0.019001085776330078, + "grad_norm": 1.5043525736330898, + "learning_rate": 2.527075812274368e-06, + "loss": 4.0634, + "step": 35 + }, + { + "epoch": 0.019543973941368076, + "grad_norm": 2.6324248758076862, + "learning_rate": 2.59927797833935e-06, + "loss": 5.4211, + "step": 36 + }, + { + "epoch": 0.02008686210640608, + "grad_norm": 1.6581905646911415, + "learning_rate": 2.6714801444043326e-06, + "loss": 4.9825, + "step": 37 + }, + { + "epoch": 0.02062975027144408, + "grad_norm": 1.6667970880898455, + "learning_rate": 2.7436823104693144e-06, + "loss": 4.3219, + "step": 38 + }, + { + "epoch": 0.021172638436482084, + "grad_norm": 1.7448150724822071, + "learning_rate": 2.815884476534296e-06, + "loss": 4.3912, + "step": 39 + }, + { + "epoch": 0.021715526601520086, + "grad_norm": 3.720503777726443, + "learning_rate": 2.8880866425992783e-06, + "loss": 6.0592, + "step": 40 + }, + { + "epoch": 0.02225841476655809, + "grad_norm": 1.6853001074010774, + "learning_rate": 2.96028880866426e-06, + "loss": 3.8395, + "step": 41 + }, + { + "epoch": 0.02280130293159609, + "grad_norm": 1.797029067943667, + "learning_rate": 3.032490974729242e-06, + "loss": 4.4651, + "step": 42 + }, + { + "epoch": 0.023344191096634093, + "grad_norm": 1.6298824221707349, + "learning_rate": 3.1046931407942245e-06, + "loss": 4.4507, + "step": 43 + }, + { + "epoch": 0.023887079261672096, + "grad_norm": 1.4793189209059368, + "learning_rate": 3.1768953068592062e-06, + "loss": 4.7849, + "step": 44 + }, + { + "epoch": 0.024429967426710098, + "grad_norm": 1.1741346570472608, + "learning_rate": 3.249097472924188e-06, + "loss": 3.5897, + "step": 45 + }, + { + "epoch": 0.0249728555917481, + "grad_norm": 2.1468867306169064, + "learning_rate": 3.3212996389891698e-06, + "loss": 3.8969, + "step": 46 + }, + { + "epoch": 0.025515743756786103, + "grad_norm": 1.657494000697021, + "learning_rate": 3.393501805054152e-06, + "loss": 4.3412, + "step": 47 + }, + { + "epoch": 0.026058631921824105, + "grad_norm": 1.7125864055710684, + "learning_rate": 3.4657039711191337e-06, + "loss": 3.8059, + "step": 48 + }, + { + "epoch": 0.026601520086862108, + "grad_norm": 2.9515347000925622, + "learning_rate": 3.5379061371841155e-06, + "loss": 4.1069, + "step": 49 + }, + { + "epoch": 0.02714440825190011, + "grad_norm": 3.0185188918098262, + "learning_rate": 3.610108303249098e-06, + "loss": 5.2646, + "step": 50 + }, + { + "epoch": 0.02768729641693811, + "grad_norm": 1.5546972252207274, + "learning_rate": 3.68231046931408e-06, + "loss": 3.7212, + "step": 51 + }, + { + "epoch": 0.02823018458197611, + "grad_norm": 2.6376587426282474, + "learning_rate": 3.7545126353790616e-06, + "loss": 4.2862, + "step": 52 + }, + { + "epoch": 0.028773072747014114, + "grad_norm": 2.053905428064986, + "learning_rate": 3.826714801444043e-06, + "loss": 3.8304, + "step": 53 + }, + { + "epoch": 0.029315960912052116, + "grad_norm": 2.9083183656849725, + "learning_rate": 3.898916967509026e-06, + "loss": 4.8223, + "step": 54 + }, + { + "epoch": 0.02985884907709012, + "grad_norm": 1.9919239162425852, + "learning_rate": 3.971119133574007e-06, + "loss": 3.9054, + "step": 55 + }, + { + "epoch": 0.03040173724212812, + "grad_norm": 2.4076122411223135, + "learning_rate": 4.043321299638989e-06, + "loss": 4.8568, + "step": 56 + }, + { + "epoch": 0.030944625407166124, + "grad_norm": 3.854434602830178, + "learning_rate": 4.115523465703971e-06, + "loss": 5.7771, + "step": 57 + }, + { + "epoch": 0.031487513572204126, + "grad_norm": 2.510303564095041, + "learning_rate": 4.1877256317689535e-06, + "loss": 5.5851, + "step": 58 + }, + { + "epoch": 0.03203040173724213, + "grad_norm": 2.8884448366576305, + "learning_rate": 4.259927797833936e-06, + "loss": 4.1774, + "step": 59 + }, + { + "epoch": 0.03257328990228013, + "grad_norm": 2.0893778739721, + "learning_rate": 4.332129963898917e-06, + "loss": 4.4679, + "step": 60 + }, + { + "epoch": 0.03311617806731813, + "grad_norm": 1.5816439039579229, + "learning_rate": 4.404332129963899e-06, + "loss": 4.1999, + "step": 61 + }, + { + "epoch": 0.033659066232356136, + "grad_norm": 1.9591574137365761, + "learning_rate": 4.4765342960288814e-06, + "loss": 4.1539, + "step": 62 + }, + { + "epoch": 0.03420195439739414, + "grad_norm": 3.1533909413995924, + "learning_rate": 4.548736462093864e-06, + "loss": 4.5842, + "step": 63 + }, + { + "epoch": 0.03474484256243214, + "grad_norm": 2.543715437360663, + "learning_rate": 4.620938628158845e-06, + "loss": 4.1473, + "step": 64 + }, + { + "epoch": 0.03528773072747014, + "grad_norm": 2.7608093188531955, + "learning_rate": 4.693140794223827e-06, + "loss": 4.9354, + "step": 65 + }, + { + "epoch": 0.035830618892508145, + "grad_norm": 4.192163162864289, + "learning_rate": 4.765342960288809e-06, + "loss": 5.4263, + "step": 66 + }, + { + "epoch": 0.03637350705754615, + "grad_norm": 2.809206875027472, + "learning_rate": 4.837545126353791e-06, + "loss": 4.4259, + "step": 67 + }, + { + "epoch": 0.03691639522258415, + "grad_norm": 3.9318073942863534, + "learning_rate": 4.909747292418773e-06, + "loss": 4.1483, + "step": 68 + }, + { + "epoch": 0.03745928338762215, + "grad_norm": 2.935065385226121, + "learning_rate": 4.981949458483755e-06, + "loss": 4.6854, + "step": 69 + }, + { + "epoch": 0.038002171552660155, + "grad_norm": 2.3687140876301305, + "learning_rate": 5.054151624548736e-06, + "loss": 4.5875, + "step": 70 + }, + { + "epoch": 0.03854505971769816, + "grad_norm": 2.987709377109457, + "learning_rate": 5.126353790613719e-06, + "loss": 5.5784, + "step": 71 + }, + { + "epoch": 0.03908794788273615, + "grad_norm": 2.618514747422557, + "learning_rate": 5.1985559566787e-06, + "loss": 3.9901, + "step": 72 + }, + { + "epoch": 0.039630836047774155, + "grad_norm": 3.729239515680902, + "learning_rate": 5.270758122743683e-06, + "loss": 4.6192, + "step": 73 + }, + { + "epoch": 0.04017372421281216, + "grad_norm": 2.488658936956188, + "learning_rate": 5.342960288808665e-06, + "loss": 3.9889, + "step": 74 + }, + { + "epoch": 0.04071661237785016, + "grad_norm": 2.6055640979236223, + "learning_rate": 5.415162454873647e-06, + "loss": 4.3166, + "step": 75 + }, + { + "epoch": 0.04125950054288816, + "grad_norm": 2.091905540809132, + "learning_rate": 5.487364620938629e-06, + "loss": 3.5848, + "step": 76 + }, + { + "epoch": 0.041802388707926165, + "grad_norm": 4.153140315460812, + "learning_rate": 5.559566787003611e-06, + "loss": 4.4889, + "step": 77 + }, + { + "epoch": 0.04234527687296417, + "grad_norm": 4.828033123535034, + "learning_rate": 5.631768953068592e-06, + "loss": 4.886, + "step": 78 + }, + { + "epoch": 0.04288816503800217, + "grad_norm": 3.106146715385438, + "learning_rate": 5.7039711191335744e-06, + "loss": 4.3698, + "step": 79 + }, + { + "epoch": 0.04343105320304017, + "grad_norm": 4.458698539182903, + "learning_rate": 5.776173285198557e-06, + "loss": 4.7018, + "step": 80 + }, + { + "epoch": 0.043973941368078175, + "grad_norm": 4.619902873416999, + "learning_rate": 5.848375451263538e-06, + "loss": 4.3449, + "step": 81 + }, + { + "epoch": 0.04451682953311618, + "grad_norm": 2.411841615517987, + "learning_rate": 5.92057761732852e-06, + "loss": 3.9952, + "step": 82 + }, + { + "epoch": 0.04505971769815418, + "grad_norm": 3.6594837742651705, + "learning_rate": 5.992779783393502e-06, + "loss": 4.1308, + "step": 83 + }, + { + "epoch": 0.04560260586319218, + "grad_norm": 2.3760888273966074, + "learning_rate": 6.064981949458484e-06, + "loss": 4.0315, + "step": 84 + }, + { + "epoch": 0.046145494028230184, + "grad_norm": 4.579655616004927, + "learning_rate": 6.137184115523466e-06, + "loss": 4.1051, + "step": 85 + }, + { + "epoch": 0.04668838219326819, + "grad_norm": 3.5909339610827846, + "learning_rate": 6.209386281588449e-06, + "loss": 4.3126, + "step": 86 + }, + { + "epoch": 0.04723127035830619, + "grad_norm": 3.1030422711991625, + "learning_rate": 6.28158844765343e-06, + "loss": 4.3038, + "step": 87 + }, + { + "epoch": 0.04777415852334419, + "grad_norm": 4.134211324790428, + "learning_rate": 6.3537906137184125e-06, + "loss": 4.906, + "step": 88 + }, + { + "epoch": 0.048317046688382194, + "grad_norm": 3.538946913288816, + "learning_rate": 6.425992779783395e-06, + "loss": 4.2182, + "step": 89 + }, + { + "epoch": 0.048859934853420196, + "grad_norm": 1.9828152543749473, + "learning_rate": 6.498194945848376e-06, + "loss": 3.8487, + "step": 90 + }, + { + "epoch": 0.0494028230184582, + "grad_norm": 2.127969936938475, + "learning_rate": 6.570397111913358e-06, + "loss": 3.9221, + "step": 91 + }, + { + "epoch": 0.0499457111834962, + "grad_norm": 2.740653026382279, + "learning_rate": 6.6425992779783395e-06, + "loss": 4.4661, + "step": 92 + }, + { + "epoch": 0.050488599348534204, + "grad_norm": 3.278496788216742, + "learning_rate": 6.714801444043322e-06, + "loss": 4.1995, + "step": 93 + }, + { + "epoch": 0.051031487513572206, + "grad_norm": 1.5455749978381657, + "learning_rate": 6.787003610108304e-06, + "loss": 3.3941, + "step": 94 + }, + { + "epoch": 0.05157437567861021, + "grad_norm": 2.6938438745410975, + "learning_rate": 6.859205776173285e-06, + "loss": 3.9863, + "step": 95 + }, + { + "epoch": 0.05211726384364821, + "grad_norm": 2.754193515919841, + "learning_rate": 6.9314079422382674e-06, + "loss": 4.1621, + "step": 96 + }, + { + "epoch": 0.05266015200868621, + "grad_norm": 1.787757508585929, + "learning_rate": 7.00361010830325e-06, + "loss": 3.728, + "step": 97 + }, + { + "epoch": 0.053203040173724216, + "grad_norm": 1.6756343581979376, + "learning_rate": 7.075812274368231e-06, + "loss": 3.4483, + "step": 98 + }, + { + "epoch": 0.05374592833876222, + "grad_norm": 3.0418610323064317, + "learning_rate": 7.148014440433214e-06, + "loss": 4.1131, + "step": 99 + }, + { + "epoch": 0.05428881650380022, + "grad_norm": 2.710846119482301, + "learning_rate": 7.220216606498196e-06, + "loss": 5.0449, + "step": 100 + }, + { + "epoch": 0.054831704668838216, + "grad_norm": 1.906382949957388, + "learning_rate": 7.2924187725631776e-06, + "loss": 3.3181, + "step": 101 + }, + { + "epoch": 0.05537459283387622, + "grad_norm": 2.4669878697837, + "learning_rate": 7.36462093862816e-06, + "loss": 3.7996, + "step": 102 + }, + { + "epoch": 0.05591748099891422, + "grad_norm": 2.3755842712374586, + "learning_rate": 7.436823104693142e-06, + "loss": 4.0321, + "step": 103 + }, + { + "epoch": 0.05646036916395222, + "grad_norm": 2.032793864149989, + "learning_rate": 7.509025270758123e-06, + "loss": 3.4121, + "step": 104 + }, + { + "epoch": 0.057003257328990226, + "grad_norm": 2.1561554352810686, + "learning_rate": 7.5812274368231055e-06, + "loss": 4.2585, + "step": 105 + }, + { + "epoch": 0.05754614549402823, + "grad_norm": 1.9627108654626784, + "learning_rate": 7.653429602888087e-06, + "loss": 4.3085, + "step": 106 + }, + { + "epoch": 0.05808903365906623, + "grad_norm": 1.8622753095278954, + "learning_rate": 7.72563176895307e-06, + "loss": 3.8459, + "step": 107 + }, + { + "epoch": 0.05863192182410423, + "grad_norm": 1.4438765470100654, + "learning_rate": 7.797833935018051e-06, + "loss": 3.4914, + "step": 108 + }, + { + "epoch": 0.059174809989142235, + "grad_norm": 1.6164085905580508, + "learning_rate": 7.870036101083033e-06, + "loss": 3.4958, + "step": 109 + }, + { + "epoch": 0.05971769815418024, + "grad_norm": 1.6075014519578008, + "learning_rate": 7.942238267148014e-06, + "loss": 3.202, + "step": 110 + }, + { + "epoch": 0.06026058631921824, + "grad_norm": 1.7630994896021268, + "learning_rate": 8.014440433212997e-06, + "loss": 3.0916, + "step": 111 + }, + { + "epoch": 0.06080347448425624, + "grad_norm": 1.7600666006428431, + "learning_rate": 8.086642599277978e-06, + "loss": 3.5791, + "step": 112 + }, + { + "epoch": 0.061346362649294245, + "grad_norm": 2.324243905399748, + "learning_rate": 8.158844765342961e-06, + "loss": 3.6247, + "step": 113 + }, + { + "epoch": 0.06188925081433225, + "grad_norm": 1.2715449173786617, + "learning_rate": 8.231046931407943e-06, + "loss": 3.2381, + "step": 114 + }, + { + "epoch": 0.06243213897937025, + "grad_norm": 1.9029014476619814, + "learning_rate": 8.303249097472926e-06, + "loss": 3.5269, + "step": 115 + }, + { + "epoch": 0.06297502714440825, + "grad_norm": 1.8782574049467182, + "learning_rate": 8.375451263537907e-06, + "loss": 2.8698, + "step": 116 + }, + { + "epoch": 0.06351791530944625, + "grad_norm": 1.8522383705853351, + "learning_rate": 8.447653429602888e-06, + "loss": 3.5003, + "step": 117 + }, + { + "epoch": 0.06406080347448426, + "grad_norm": 1.6480852413427443, + "learning_rate": 8.519855595667871e-06, + "loss": 3.5665, + "step": 118 + }, + { + "epoch": 0.06460369163952226, + "grad_norm": 2.1628448626122685, + "learning_rate": 8.592057761732853e-06, + "loss": 3.343, + "step": 119 + }, + { + "epoch": 0.06514657980456026, + "grad_norm": 1.568456113423694, + "learning_rate": 8.664259927797834e-06, + "loss": 2.474, + "step": 120 + }, + { + "epoch": 0.06568946796959826, + "grad_norm": 1.8054357065473743, + "learning_rate": 8.736462093862817e-06, + "loss": 2.9228, + "step": 121 + }, + { + "epoch": 0.06623235613463627, + "grad_norm": 1.9418450135728782, + "learning_rate": 8.808664259927798e-06, + "loss": 3.1995, + "step": 122 + }, + { + "epoch": 0.06677524429967427, + "grad_norm": 1.321851521267606, + "learning_rate": 8.88086642599278e-06, + "loss": 3.5515, + "step": 123 + }, + { + "epoch": 0.06731813246471227, + "grad_norm": 1.390198419578882, + "learning_rate": 8.953068592057763e-06, + "loss": 2.7741, + "step": 124 + }, + { + "epoch": 0.06786102062975027, + "grad_norm": 2.119370139523294, + "learning_rate": 9.025270758122744e-06, + "loss": 3.3427, + "step": 125 + }, + { + "epoch": 0.06840390879478828, + "grad_norm": 2.042189734142807, + "learning_rate": 9.097472924187727e-06, + "loss": 3.1414, + "step": 126 + }, + { + "epoch": 0.06894679695982628, + "grad_norm": 2.0399787373405123, + "learning_rate": 9.169675090252709e-06, + "loss": 3.0862, + "step": 127 + }, + { + "epoch": 0.06948968512486428, + "grad_norm": 2.0523727295223915, + "learning_rate": 9.24187725631769e-06, + "loss": 2.6461, + "step": 128 + }, + { + "epoch": 0.07003257328990228, + "grad_norm": 1.459190908283032, + "learning_rate": 9.314079422382673e-06, + "loss": 2.9907, + "step": 129 + }, + { + "epoch": 0.07057546145494029, + "grad_norm": 1.7403521896723462, + "learning_rate": 9.386281588447654e-06, + "loss": 3.0681, + "step": 130 + }, + { + "epoch": 0.07111834961997829, + "grad_norm": 2.2283604183714383, + "learning_rate": 9.458483754512636e-06, + "loss": 3.4619, + "step": 131 + }, + { + "epoch": 0.07166123778501629, + "grad_norm": 1.3291461832292721, + "learning_rate": 9.530685920577619e-06, + "loss": 3.096, + "step": 132 + }, + { + "epoch": 0.0722041259500543, + "grad_norm": 1.7839357457835638, + "learning_rate": 9.6028880866426e-06, + "loss": 2.7512, + "step": 133 + }, + { + "epoch": 0.0727470141150923, + "grad_norm": 1.7836465130823984, + "learning_rate": 9.675090252707581e-06, + "loss": 3.3118, + "step": 134 + }, + { + "epoch": 0.0732899022801303, + "grad_norm": 2.0407172747925135, + "learning_rate": 9.747292418772564e-06, + "loss": 2.1032, + "step": 135 + }, + { + "epoch": 0.0738327904451683, + "grad_norm": 2.335226908691354, + "learning_rate": 9.819494584837546e-06, + "loss": 2.9039, + "step": 136 + }, + { + "epoch": 0.0743756786102063, + "grad_norm": 2.567081845339778, + "learning_rate": 9.891696750902527e-06, + "loss": 3.1071, + "step": 137 + }, + { + "epoch": 0.0749185667752443, + "grad_norm": 1.9549184920601244, + "learning_rate": 9.96389891696751e-06, + "loss": 2.8253, + "step": 138 + }, + { + "epoch": 0.07546145494028231, + "grad_norm": 1.6389815727698025, + "learning_rate": 1.0036101083032491e-05, + "loss": 2.5701, + "step": 139 + }, + { + "epoch": 0.07600434310532031, + "grad_norm": 1.6558924195158018, + "learning_rate": 1.0108303249097473e-05, + "loss": 2.6428, + "step": 140 + }, + { + "epoch": 0.07654723127035831, + "grad_norm": 2.0257343427634087, + "learning_rate": 1.0180505415162456e-05, + "loss": 2.2828, + "step": 141 + }, + { + "epoch": 0.07709011943539631, + "grad_norm": 1.879991343961764, + "learning_rate": 1.0252707581227437e-05, + "loss": 3.1436, + "step": 142 + }, + { + "epoch": 0.07763300760043432, + "grad_norm": 1.8718236498893788, + "learning_rate": 1.032490974729242e-05, + "loss": 3.2434, + "step": 143 + }, + { + "epoch": 0.0781758957654723, + "grad_norm": 1.8252190287923398, + "learning_rate": 1.03971119133574e-05, + "loss": 2.4851, + "step": 144 + }, + { + "epoch": 0.07871878393051031, + "grad_norm": 2.6564413093838213, + "learning_rate": 1.0469314079422383e-05, + "loss": 2.7315, + "step": 145 + }, + { + "epoch": 0.07926167209554831, + "grad_norm": 2.054556112475997, + "learning_rate": 1.0541516245487366e-05, + "loss": 2.647, + "step": 146 + }, + { + "epoch": 0.07980456026058631, + "grad_norm": 1.072343898976271, + "learning_rate": 1.0613718411552347e-05, + "loss": 2.5868, + "step": 147 + }, + { + "epoch": 0.08034744842562432, + "grad_norm": 1.1443585178571194, + "learning_rate": 1.068592057761733e-05, + "loss": 2.3247, + "step": 148 + }, + { + "epoch": 0.08089033659066232, + "grad_norm": 2.0424670761496646, + "learning_rate": 1.0758122743682312e-05, + "loss": 1.989, + "step": 149 + }, + { + "epoch": 0.08143322475570032, + "grad_norm": 1.3715615162918033, + "learning_rate": 1.0830324909747295e-05, + "loss": 2.3989, + "step": 150 + }, + { + "epoch": 0.08197611292073832, + "grad_norm": 1.3740116291532023, + "learning_rate": 1.0902527075812274e-05, + "loss": 1.8726, + "step": 151 + }, + { + "epoch": 0.08251900108577633, + "grad_norm": 2.5824759155435655, + "learning_rate": 1.0974729241877257e-05, + "loss": 2.3019, + "step": 152 + }, + { + "epoch": 0.08306188925081433, + "grad_norm": 1.5584514490971877, + "learning_rate": 1.1046931407942239e-05, + "loss": 2.4209, + "step": 153 + }, + { + "epoch": 0.08360477741585233, + "grad_norm": 1.6184456176727229, + "learning_rate": 1.1119133574007222e-05, + "loss": 2.2377, + "step": 154 + }, + { + "epoch": 0.08414766558089033, + "grad_norm": 2.6665151568462586, + "learning_rate": 1.1191335740072201e-05, + "loss": 2.6381, + "step": 155 + }, + { + "epoch": 0.08469055374592833, + "grad_norm": 1.442929928712006, + "learning_rate": 1.1263537906137184e-05, + "loss": 2.3405, + "step": 156 + }, + { + "epoch": 0.08523344191096634, + "grad_norm": 1.906547769868643, + "learning_rate": 1.1335740072202166e-05, + "loss": 2.4391, + "step": 157 + }, + { + "epoch": 0.08577633007600434, + "grad_norm": 1.2153727696829317, + "learning_rate": 1.1407942238267149e-05, + "loss": 1.8401, + "step": 158 + }, + { + "epoch": 0.08631921824104234, + "grad_norm": 1.4810729195972283, + "learning_rate": 1.1480144404332132e-05, + "loss": 2.7248, + "step": 159 + }, + { + "epoch": 0.08686210640608034, + "grad_norm": 2.9401118999625058, + "learning_rate": 1.1552346570397113e-05, + "loss": 2.6582, + "step": 160 + }, + { + "epoch": 0.08740499457111835, + "grad_norm": 1.145106458573816, + "learning_rate": 1.1624548736462096e-05, + "loss": 2.6648, + "step": 161 + }, + { + "epoch": 0.08794788273615635, + "grad_norm": 1.9228128025198137, + "learning_rate": 1.1696750902527076e-05, + "loss": 2.3702, + "step": 162 + }, + { + "epoch": 0.08849077090119435, + "grad_norm": 1.318978856783607, + "learning_rate": 1.1768953068592059e-05, + "loss": 1.9388, + "step": 163 + }, + { + "epoch": 0.08903365906623235, + "grad_norm": 1.58104775083856, + "learning_rate": 1.184115523465704e-05, + "loss": 2.0774, + "step": 164 + }, + { + "epoch": 0.08957654723127036, + "grad_norm": 1.8206646323260607, + "learning_rate": 1.1913357400722023e-05, + "loss": 2.5221, + "step": 165 + }, + { + "epoch": 0.09011943539630836, + "grad_norm": 1.0815960351491627, + "learning_rate": 1.1985559566787005e-05, + "loss": 2.128, + "step": 166 + }, + { + "epoch": 0.09066232356134636, + "grad_norm": 1.3341486432977188, + "learning_rate": 1.2057761732851988e-05, + "loss": 1.9418, + "step": 167 + }, + { + "epoch": 0.09120521172638436, + "grad_norm": 1.3772486036170464, + "learning_rate": 1.2129963898916967e-05, + "loss": 1.84, + "step": 168 + }, + { + "epoch": 0.09174809989142237, + "grad_norm": 1.42874152479176, + "learning_rate": 1.220216606498195e-05, + "loss": 2.2172, + "step": 169 + }, + { + "epoch": 0.09229098805646037, + "grad_norm": 1.9691394708584433, + "learning_rate": 1.2274368231046932e-05, + "loss": 2.9247, + "step": 170 + }, + { + "epoch": 0.09283387622149837, + "grad_norm": 1.4773974635910447, + "learning_rate": 1.2346570397111915e-05, + "loss": 2.1028, + "step": 171 + }, + { + "epoch": 0.09337676438653637, + "grad_norm": 1.5937344901476087, + "learning_rate": 1.2418772563176898e-05, + "loss": 2.4073, + "step": 172 + }, + { + "epoch": 0.09391965255157438, + "grad_norm": 1.630783959399718, + "learning_rate": 1.2490974729241878e-05, + "loss": 2.6978, + "step": 173 + }, + { + "epoch": 0.09446254071661238, + "grad_norm": 1.44736447594164, + "learning_rate": 1.256317689530686e-05, + "loss": 1.8851, + "step": 174 + }, + { + "epoch": 0.09500542888165038, + "grad_norm": 1.3909459856510267, + "learning_rate": 1.2635379061371842e-05, + "loss": 1.576, + "step": 175 + }, + { + "epoch": 0.09554831704668838, + "grad_norm": 1.7838393677575515, + "learning_rate": 1.2707581227436825e-05, + "loss": 2.3895, + "step": 176 + }, + { + "epoch": 0.09609120521172639, + "grad_norm": 1.1294127385799255, + "learning_rate": 1.2779783393501806e-05, + "loss": 2.4855, + "step": 177 + }, + { + "epoch": 0.09663409337676439, + "grad_norm": 1.259662421149648, + "learning_rate": 1.285198555956679e-05, + "loss": 2.5469, + "step": 178 + }, + { + "epoch": 0.09717698154180239, + "grad_norm": 1.4621367635995284, + "learning_rate": 1.2924187725631769e-05, + "loss": 1.9015, + "step": 179 + }, + { + "epoch": 0.09771986970684039, + "grad_norm": 1.6960126954903645, + "learning_rate": 1.2996389891696752e-05, + "loss": 2.9545, + "step": 180 + }, + { + "epoch": 0.0982627578718784, + "grad_norm": 1.5419022503429414, + "learning_rate": 1.3068592057761733e-05, + "loss": 1.9912, + "step": 181 + }, + { + "epoch": 0.0988056460369164, + "grad_norm": 1.4073422496120207, + "learning_rate": 1.3140794223826716e-05, + "loss": 2.3353, + "step": 182 + }, + { + "epoch": 0.0993485342019544, + "grad_norm": 1.2881659254471862, + "learning_rate": 1.3212996389891696e-05, + "loss": 2.7784, + "step": 183 + }, + { + "epoch": 0.0998914223669924, + "grad_norm": 1.361969663712106, + "learning_rate": 1.3285198555956679e-05, + "loss": 2.523, + "step": 184 + }, + { + "epoch": 0.1004343105320304, + "grad_norm": 2.5841731119141316, + "learning_rate": 1.3357400722021662e-05, + "loss": 2.43, + "step": 185 + }, + { + "epoch": 0.10097719869706841, + "grad_norm": 1.3203537847943392, + "learning_rate": 1.3429602888086643e-05, + "loss": 1.9404, + "step": 186 + }, + { + "epoch": 0.10152008686210641, + "grad_norm": 1.393097901493087, + "learning_rate": 1.3501805054151626e-05, + "loss": 2.6202, + "step": 187 + }, + { + "epoch": 0.10206297502714441, + "grad_norm": 0.9414377380652101, + "learning_rate": 1.3574007220216608e-05, + "loss": 2.0879, + "step": 188 + }, + { + "epoch": 0.10260586319218241, + "grad_norm": 1.6547581806701488, + "learning_rate": 1.3646209386281591e-05, + "loss": 2.8912, + "step": 189 + }, + { + "epoch": 0.10314875135722042, + "grad_norm": 1.1236261864015666, + "learning_rate": 1.371841155234657e-05, + "loss": 2.1802, + "step": 190 + }, + { + "epoch": 0.10369163952225842, + "grad_norm": 1.1766384878462062, + "learning_rate": 1.3790613718411554e-05, + "loss": 1.9127, + "step": 191 + }, + { + "epoch": 0.10423452768729642, + "grad_norm": 2.0453724866488656, + "learning_rate": 1.3862815884476535e-05, + "loss": 1.8522, + "step": 192 + }, + { + "epoch": 0.10477741585233442, + "grad_norm": 1.550964370734908, + "learning_rate": 1.3935018050541518e-05, + "loss": 2.5852, + "step": 193 + }, + { + "epoch": 0.10532030401737243, + "grad_norm": 1.303227180334529, + "learning_rate": 1.40072202166065e-05, + "loss": 2.407, + "step": 194 + }, + { + "epoch": 0.10586319218241043, + "grad_norm": 1.2009193115283474, + "learning_rate": 1.4079422382671482e-05, + "loss": 1.8978, + "step": 195 + }, + { + "epoch": 0.10640608034744843, + "grad_norm": 1.353648963417667, + "learning_rate": 1.4151624548736462e-05, + "loss": 2.283, + "step": 196 + }, + { + "epoch": 0.10694896851248643, + "grad_norm": 1.266154745219424, + "learning_rate": 1.4223826714801445e-05, + "loss": 2.0679, + "step": 197 + }, + { + "epoch": 0.10749185667752444, + "grad_norm": 4.186447043886807, + "learning_rate": 1.4296028880866428e-05, + "loss": 1.9175, + "step": 198 + }, + { + "epoch": 0.10803474484256244, + "grad_norm": 1.147065238667786, + "learning_rate": 1.436823104693141e-05, + "loss": 2.2748, + "step": 199 + }, + { + "epoch": 0.10857763300760044, + "grad_norm": 1.2760699256137895, + "learning_rate": 1.4440433212996392e-05, + "loss": 2.3031, + "step": 200 + }, + { + "epoch": 0.10912052117263844, + "grad_norm": 0.9452152631580967, + "learning_rate": 1.4512635379061372e-05, + "loss": 1.4833, + "step": 201 + }, + { + "epoch": 0.10966340933767643, + "grad_norm": 1.3980490034032074, + "learning_rate": 1.4584837545126355e-05, + "loss": 1.5865, + "step": 202 + }, + { + "epoch": 0.11020629750271443, + "grad_norm": 1.6440838011007093, + "learning_rate": 1.4657039711191336e-05, + "loss": 1.632, + "step": 203 + }, + { + "epoch": 0.11074918566775244, + "grad_norm": 1.3969764794458068, + "learning_rate": 1.472924187725632e-05, + "loss": 1.4301, + "step": 204 + }, + { + "epoch": 0.11129207383279044, + "grad_norm": 1.223663090997016, + "learning_rate": 1.48014440433213e-05, + "loss": 2.3993, + "step": 205 + }, + { + "epoch": 0.11183496199782844, + "grad_norm": 1.183881027275754, + "learning_rate": 1.4873646209386284e-05, + "loss": 1.4987, + "step": 206 + }, + { + "epoch": 0.11237785016286644, + "grad_norm": 1.3788483357578956, + "learning_rate": 1.4945848375451264e-05, + "loss": 2.7742, + "step": 207 + }, + { + "epoch": 0.11292073832790445, + "grad_norm": 1.1260622559595288, + "learning_rate": 1.5018050541516247e-05, + "loss": 1.755, + "step": 208 + }, + { + "epoch": 0.11346362649294245, + "grad_norm": 1.8439396062381286, + "learning_rate": 1.5090252707581228e-05, + "loss": 2.9875, + "step": 209 + }, + { + "epoch": 0.11400651465798045, + "grad_norm": 1.2789617727573501, + "learning_rate": 1.5162454873646211e-05, + "loss": 2.0119, + "step": 210 + }, + { + "epoch": 0.11454940282301845, + "grad_norm": 1.3385913561767113, + "learning_rate": 1.5234657039711192e-05, + "loss": 1.7693, + "step": 211 + }, + { + "epoch": 0.11509229098805646, + "grad_norm": 1.049162702021749, + "learning_rate": 1.5306859205776174e-05, + "loss": 1.7541, + "step": 212 + }, + { + "epoch": 0.11563517915309446, + "grad_norm": 1.0797407129441892, + "learning_rate": 1.537906137184116e-05, + "loss": 2.0427, + "step": 213 + }, + { + "epoch": 0.11617806731813246, + "grad_norm": 1.3661735952693994, + "learning_rate": 1.545126353790614e-05, + "loss": 2.722, + "step": 214 + }, + { + "epoch": 0.11672095548317046, + "grad_norm": 1.7084933140661187, + "learning_rate": 1.552346570397112e-05, + "loss": 2.2125, + "step": 215 + }, + { + "epoch": 0.11726384364820847, + "grad_norm": 1.3814428141700028, + "learning_rate": 1.5595667870036102e-05, + "loss": 2.448, + "step": 216 + }, + { + "epoch": 0.11780673181324647, + "grad_norm": 1.298309196615407, + "learning_rate": 1.5667870036101084e-05, + "loss": 2.2706, + "step": 217 + }, + { + "epoch": 0.11834961997828447, + "grad_norm": 1.4553148411944858, + "learning_rate": 1.5740072202166065e-05, + "loss": 2.0087, + "step": 218 + }, + { + "epoch": 0.11889250814332247, + "grad_norm": 1.5083115470870359, + "learning_rate": 1.581227436823105e-05, + "loss": 1.7179, + "step": 219 + }, + { + "epoch": 0.11943539630836048, + "grad_norm": 1.2335179878839475, + "learning_rate": 1.5884476534296028e-05, + "loss": 2.5337, + "step": 220 + }, + { + "epoch": 0.11997828447339848, + "grad_norm": 1.1729282902419653, + "learning_rate": 1.5956678700361013e-05, + "loss": 1.9604, + "step": 221 + }, + { + "epoch": 0.12052117263843648, + "grad_norm": 1.5700347334422509, + "learning_rate": 1.6028880866425994e-05, + "loss": 2.3655, + "step": 222 + }, + { + "epoch": 0.12106406080347448, + "grad_norm": 1.435764746128967, + "learning_rate": 1.6101083032490975e-05, + "loss": 2.109, + "step": 223 + }, + { + "epoch": 0.12160694896851248, + "grad_norm": 1.0866216936301165, + "learning_rate": 1.6173285198555957e-05, + "loss": 2.3334, + "step": 224 + }, + { + "epoch": 0.12214983713355049, + "grad_norm": 1.8060789361277199, + "learning_rate": 1.624548736462094e-05, + "loss": 1.6605, + "step": 225 + }, + { + "epoch": 0.12269272529858849, + "grad_norm": 1.539101200174871, + "learning_rate": 1.6317689530685923e-05, + "loss": 2.2597, + "step": 226 + }, + { + "epoch": 0.12323561346362649, + "grad_norm": 1.7392984350243024, + "learning_rate": 1.6389891696750904e-05, + "loss": 2.5979, + "step": 227 + }, + { + "epoch": 0.1237785016286645, + "grad_norm": 1.6792982939949113, + "learning_rate": 1.6462093862815885e-05, + "loss": 2.0626, + "step": 228 + }, + { + "epoch": 0.1243213897937025, + "grad_norm": 2.067856070276011, + "learning_rate": 1.6534296028880867e-05, + "loss": 2.056, + "step": 229 + }, + { + "epoch": 0.1248642779587405, + "grad_norm": 1.2773224600362139, + "learning_rate": 1.660649819494585e-05, + "loss": 1.5852, + "step": 230 + }, + { + "epoch": 0.1254071661237785, + "grad_norm": 1.381245750855955, + "learning_rate": 1.6678700361010833e-05, + "loss": 2.0675, + "step": 231 + }, + { + "epoch": 0.1259500542888165, + "grad_norm": 1.4215755121774605, + "learning_rate": 1.6750902527075814e-05, + "loss": 2.1002, + "step": 232 + }, + { + "epoch": 0.1264929424538545, + "grad_norm": 1.4581673093175722, + "learning_rate": 1.6823104693140795e-05, + "loss": 1.9279, + "step": 233 + }, + { + "epoch": 0.1270358306188925, + "grad_norm": 1.2656102558648035, + "learning_rate": 1.6895306859205777e-05, + "loss": 1.4913, + "step": 234 + }, + { + "epoch": 0.1275787187839305, + "grad_norm": 1.3985246778048182, + "learning_rate": 1.6967509025270758e-05, + "loss": 2.1629, + "step": 235 + }, + { + "epoch": 0.1281216069489685, + "grad_norm": 1.757194392177075, + "learning_rate": 1.7039711191335743e-05, + "loss": 2.5688, + "step": 236 + }, + { + "epoch": 0.12866449511400652, + "grad_norm": 1.3766112824147358, + "learning_rate": 1.711191335740072e-05, + "loss": 2.3344, + "step": 237 + }, + { + "epoch": 0.12920738327904452, + "grad_norm": 1.2985043105700416, + "learning_rate": 1.7184115523465706e-05, + "loss": 2.0626, + "step": 238 + }, + { + "epoch": 0.12975027144408252, + "grad_norm": 1.715796674825951, + "learning_rate": 1.7256317689530687e-05, + "loss": 2.1559, + "step": 239 + }, + { + "epoch": 0.13029315960912052, + "grad_norm": 1.6374209306607368, + "learning_rate": 1.7328519855595668e-05, + "loss": 2.0801, + "step": 240 + }, + { + "epoch": 0.13083604777415853, + "grad_norm": 1.244682506367769, + "learning_rate": 1.7400722021660653e-05, + "loss": 1.6125, + "step": 241 + }, + { + "epoch": 0.13137893593919653, + "grad_norm": 1.3001838781366861, + "learning_rate": 1.7472924187725634e-05, + "loss": 2.0934, + "step": 242 + }, + { + "epoch": 0.13192182410423453, + "grad_norm": 1.1600879910417528, + "learning_rate": 1.7545126353790616e-05, + "loss": 1.7862, + "step": 243 + }, + { + "epoch": 0.13246471226927253, + "grad_norm": 1.3503312968411385, + "learning_rate": 1.7617328519855597e-05, + "loss": 2.1636, + "step": 244 + }, + { + "epoch": 0.13300760043431054, + "grad_norm": 2.1299761628528167, + "learning_rate": 1.768953068592058e-05, + "loss": 2.4826, + "step": 245 + }, + { + "epoch": 0.13355048859934854, + "grad_norm": 1.3232052544377066, + "learning_rate": 1.776173285198556e-05, + "loss": 2.0497, + "step": 246 + }, + { + "epoch": 0.13409337676438654, + "grad_norm": 1.5079892467642568, + "learning_rate": 1.7833935018050544e-05, + "loss": 2.4552, + "step": 247 + }, + { + "epoch": 0.13463626492942454, + "grad_norm": 1.716355318914646, + "learning_rate": 1.7906137184115526e-05, + "loss": 1.5309, + "step": 248 + }, + { + "epoch": 0.13517915309446255, + "grad_norm": 2.1523878955969193, + "learning_rate": 1.7978339350180507e-05, + "loss": 2.7403, + "step": 249 + }, + { + "epoch": 0.13572204125950055, + "grad_norm": 1.8169301927997485, + "learning_rate": 1.805054151624549e-05, + "loss": 2.3324, + "step": 250 + }, + { + "epoch": 0.13626492942453855, + "grad_norm": 1.4464924096148426, + "learning_rate": 1.812274368231047e-05, + "loss": 1.8399, + "step": 251 + }, + { + "epoch": 0.13680781758957655, + "grad_norm": 1.9818298103877654, + "learning_rate": 1.8194945848375454e-05, + "loss": 2.4237, + "step": 252 + }, + { + "epoch": 0.13735070575461455, + "grad_norm": 1.6371003739671923, + "learning_rate": 1.8267148014440436e-05, + "loss": 2.0477, + "step": 253 + }, + { + "epoch": 0.13789359391965256, + "grad_norm": 1.2944832148952543, + "learning_rate": 1.8339350180505417e-05, + "loss": 2.2269, + "step": 254 + }, + { + "epoch": 0.13843648208469056, + "grad_norm": 1.4284289887933213, + "learning_rate": 1.84115523465704e-05, + "loss": 2.2465, + "step": 255 + }, + { + "epoch": 0.13897937024972856, + "grad_norm": 1.4633226693151502, + "learning_rate": 1.848375451263538e-05, + "loss": 2.1582, + "step": 256 + }, + { + "epoch": 0.13952225841476656, + "grad_norm": 1.6169546473154692, + "learning_rate": 1.855595667870036e-05, + "loss": 1.9526, + "step": 257 + }, + { + "epoch": 0.14006514657980457, + "grad_norm": 1.47082651679944, + "learning_rate": 1.8628158844765346e-05, + "loss": 2.0314, + "step": 258 + }, + { + "epoch": 0.14060803474484257, + "grad_norm": 1.2279964838302115, + "learning_rate": 1.8700361010830327e-05, + "loss": 1.4345, + "step": 259 + }, + { + "epoch": 0.14115092290988057, + "grad_norm": 1.5727648286246085, + "learning_rate": 1.877256317689531e-05, + "loss": 2.7371, + "step": 260 + }, + { + "epoch": 0.14169381107491857, + "grad_norm": 1.5536143039045338, + "learning_rate": 1.884476534296029e-05, + "loss": 1.6691, + "step": 261 + }, + { + "epoch": 0.14223669923995658, + "grad_norm": 1.504953118413033, + "learning_rate": 1.891696750902527e-05, + "loss": 1.8422, + "step": 262 + }, + { + "epoch": 0.14277958740499458, + "grad_norm": 1.6092868309857633, + "learning_rate": 1.8989169675090253e-05, + "loss": 1.6446, + "step": 263 + }, + { + "epoch": 0.14332247557003258, + "grad_norm": 1.434484546738067, + "learning_rate": 1.9061371841155237e-05, + "loss": 1.5945, + "step": 264 + }, + { + "epoch": 0.14386536373507058, + "grad_norm": 1.3472204852366787, + "learning_rate": 1.913357400722022e-05, + "loss": 1.4511, + "step": 265 + }, + { + "epoch": 0.1444082519001086, + "grad_norm": 2.493842366649246, + "learning_rate": 1.92057761732852e-05, + "loss": 2.1144, + "step": 266 + }, + { + "epoch": 0.1449511400651466, + "grad_norm": 1.3138428233046642, + "learning_rate": 1.927797833935018e-05, + "loss": 1.6592, + "step": 267 + }, + { + "epoch": 0.1454940282301846, + "grad_norm": 1.7201458829291465, + "learning_rate": 1.9350180505415163e-05, + "loss": 1.6895, + "step": 268 + }, + { + "epoch": 0.1460369163952226, + "grad_norm": 1.9777013455620547, + "learning_rate": 1.9422382671480147e-05, + "loss": 2.2628, + "step": 269 + }, + { + "epoch": 0.1465798045602606, + "grad_norm": 1.8559862870685497, + "learning_rate": 1.949458483754513e-05, + "loss": 1.3374, + "step": 270 + }, + { + "epoch": 0.1471226927252986, + "grad_norm": 1.919404449298866, + "learning_rate": 1.956678700361011e-05, + "loss": 2.3805, + "step": 271 + }, + { + "epoch": 0.1476655808903366, + "grad_norm": 1.4760010543412088, + "learning_rate": 1.963898916967509e-05, + "loss": 1.9791, + "step": 272 + }, + { + "epoch": 0.1482084690553746, + "grad_norm": 1.4491842831658432, + "learning_rate": 1.9711191335740073e-05, + "loss": 1.8119, + "step": 273 + }, + { + "epoch": 0.1487513572204126, + "grad_norm": 1.4098796806726177, + "learning_rate": 1.9783393501805054e-05, + "loss": 1.7398, + "step": 274 + }, + { + "epoch": 0.1492942453854506, + "grad_norm": 1.5009369942229867, + "learning_rate": 1.985559566787004e-05, + "loss": 1.5194, + "step": 275 + }, + { + "epoch": 0.1498371335504886, + "grad_norm": 1.5073462671342261, + "learning_rate": 1.992779783393502e-05, + "loss": 1.8125, + "step": 276 + }, + { + "epoch": 0.1503800217155266, + "grad_norm": 1.9315604890805302, + "learning_rate": 2e-05, + "loss": 1.5429, + "step": 277 + }, + { + "epoch": 0.15092290988056462, + "grad_norm": 1.5059106777628657, + "learning_rate": 1.999999938159203e-05, + "loss": 1.6402, + "step": 278 + }, + { + "epoch": 0.15146579804560262, + "grad_norm": 1.7255630320192266, + "learning_rate": 1.9999997526368205e-05, + "loss": 2.2806, + "step": 279 + }, + { + "epoch": 0.15200868621064062, + "grad_norm": 1.7985407961594548, + "learning_rate": 1.999999443432874e-05, + "loss": 2.5213, + "step": 280 + }, + { + "epoch": 0.15255157437567862, + "grad_norm": 1.801471230172198, + "learning_rate": 1.999999010547403e-05, + "loss": 1.8362, + "step": 281 + }, + { + "epoch": 0.15309446254071662, + "grad_norm": 3.236881896883975, + "learning_rate": 1.999998453980461e-05, + "loss": 2.7148, + "step": 282 + }, + { + "epoch": 0.15363735070575463, + "grad_norm": 1.7343786362856708, + "learning_rate": 1.9999977737321156e-05, + "loss": 1.8195, + "step": 283 + }, + { + "epoch": 0.15418023887079263, + "grad_norm": 1.5006942215232695, + "learning_rate": 1.999996969802452e-05, + "loss": 1.8744, + "step": 284 + }, + { + "epoch": 0.15472312703583063, + "grad_norm": 4.171967924853298, + "learning_rate": 1.99999604219157e-05, + "loss": 2.4329, + "step": 285 + }, + { + "epoch": 0.15526601520086863, + "grad_norm": 1.7546486556973808, + "learning_rate": 1.9999949908995832e-05, + "loss": 1.3931, + "step": 286 + }, + { + "epoch": 0.15580890336590664, + "grad_norm": 2.0223702721102375, + "learning_rate": 1.999993815926622e-05, + "loss": 1.6704, + "step": 287 + }, + { + "epoch": 0.1563517915309446, + "grad_norm": 1.8736762651082781, + "learning_rate": 1.9999925172728324e-05, + "loss": 1.9895, + "step": 288 + }, + { + "epoch": 0.15689467969598261, + "grad_norm": 1.8878827291463698, + "learning_rate": 1.9999910949383742e-05, + "loss": 2.4939, + "step": 289 + }, + { + "epoch": 0.15743756786102062, + "grad_norm": 1.7277080078786893, + "learning_rate": 1.9999895489234234e-05, + "loss": 1.9265, + "step": 290 + }, + { + "epoch": 0.15798045602605862, + "grad_norm": 1.7402172969699101, + "learning_rate": 1.999987879228172e-05, + "loss": 1.8101, + "step": 291 + }, + { + "epoch": 0.15852334419109662, + "grad_norm": 2.077278713661093, + "learning_rate": 1.9999860858528257e-05, + "loss": 2.1763, + "step": 292 + }, + { + "epoch": 0.15906623235613462, + "grad_norm": 2.247624069618552, + "learning_rate": 1.9999841687976067e-05, + "loss": 1.9727, + "step": 293 + }, + { + "epoch": 0.15960912052117263, + "grad_norm": 1.6712153667666279, + "learning_rate": 1.999982128062752e-05, + "loss": 1.9921, + "step": 294 + }, + { + "epoch": 0.16015200868621063, + "grad_norm": 1.7126188149003763, + "learning_rate": 1.999979963648514e-05, + "loss": 1.9375, + "step": 295 + }, + { + "epoch": 0.16069489685124863, + "grad_norm": 1.949228534239786, + "learning_rate": 1.99997767555516e-05, + "loss": 2.1628, + "step": 296 + }, + { + "epoch": 0.16123778501628663, + "grad_norm": 1.8253305155157105, + "learning_rate": 1.9999752637829734e-05, + "loss": 1.8293, + "step": 297 + }, + { + "epoch": 0.16178067318132464, + "grad_norm": 1.5881557560095252, + "learning_rate": 1.9999727283322524e-05, + "loss": 1.3688, + "step": 298 + }, + { + "epoch": 0.16232356134636264, + "grad_norm": 2.4008202914116503, + "learning_rate": 1.9999700692033112e-05, + "loss": 1.5919, + "step": 299 + }, + { + "epoch": 0.16286644951140064, + "grad_norm": 1.610839420604411, + "learning_rate": 1.9999672863964778e-05, + "loss": 1.8724, + "step": 300 + }, + { + "epoch": 0.16340933767643864, + "grad_norm": 1.4663439746230646, + "learning_rate": 1.9999643799120964e-05, + "loss": 1.4628, + "step": 301 + }, + { + "epoch": 0.16395222584147665, + "grad_norm": 1.7051101810518976, + "learning_rate": 1.9999613497505272e-05, + "loss": 1.2916, + "step": 302 + }, + { + "epoch": 0.16449511400651465, + "grad_norm": 1.8967399031374104, + "learning_rate": 1.9999581959121443e-05, + "loss": 1.7559, + "step": 303 + }, + { + "epoch": 0.16503800217155265, + "grad_norm": 1.6206520761555416, + "learning_rate": 1.9999549183973382e-05, + "loss": 1.9196, + "step": 304 + }, + { + "epoch": 0.16558089033659065, + "grad_norm": 2.3545652344963544, + "learning_rate": 1.999951517206514e-05, + "loss": 1.2731, + "step": 305 + }, + { + "epoch": 0.16612377850162866, + "grad_norm": 1.7952736257095967, + "learning_rate": 1.9999479923400926e-05, + "loss": 2.0502, + "step": 306 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 1.7199420678903319, + "learning_rate": 1.99994434379851e-05, + "loss": 1.7059, + "step": 307 + }, + { + "epoch": 0.16720955483170466, + "grad_norm": 2.3318553047346, + "learning_rate": 1.9999405715822167e-05, + "loss": 1.4667, + "step": 308 + }, + { + "epoch": 0.16775244299674266, + "grad_norm": 1.618961362672766, + "learning_rate": 1.9999366756916804e-05, + "loss": 1.1866, + "step": 309 + }, + { + "epoch": 0.16829533116178066, + "grad_norm": 2.0805239071290114, + "learning_rate": 1.999932656127382e-05, + "loss": 1.7601, + "step": 310 + }, + { + "epoch": 0.16883821932681867, + "grad_norm": 1.782022796323131, + "learning_rate": 1.9999285128898193e-05, + "loss": 1.6207, + "step": 311 + }, + { + "epoch": 0.16938110749185667, + "grad_norm": 2.2438870289045028, + "learning_rate": 1.9999242459795045e-05, + "loss": 2.1329, + "step": 312 + }, + { + "epoch": 0.16992399565689467, + "grad_norm": 2.2550840823235587, + "learning_rate": 1.9999198553969652e-05, + "loss": 1.6055, + "step": 313 + }, + { + "epoch": 0.17046688382193267, + "grad_norm": 1.8012459587153988, + "learning_rate": 1.9999153411427445e-05, + "loss": 2.018, + "step": 314 + }, + { + "epoch": 0.17100977198697068, + "grad_norm": 1.8498840013582993, + "learning_rate": 1.9999107032174007e-05, + "loss": 1.8113, + "step": 315 + }, + { + "epoch": 0.17155266015200868, + "grad_norm": 2.2799557163953965, + "learning_rate": 1.9999059416215078e-05, + "loss": 1.9494, + "step": 316 + }, + { + "epoch": 0.17209554831704668, + "grad_norm": 2.383891228288174, + "learning_rate": 1.999901056355654e-05, + "loss": 1.8323, + "step": 317 + }, + { + "epoch": 0.17263843648208468, + "grad_norm": 1.9081077971876759, + "learning_rate": 1.9998960474204443e-05, + "loss": 1.5476, + "step": 318 + }, + { + "epoch": 0.1731813246471227, + "grad_norm": 1.9010783437994137, + "learning_rate": 1.999890914816498e-05, + "loss": 1.4602, + "step": 319 + }, + { + "epoch": 0.1737242128121607, + "grad_norm": 2.295573334252978, + "learning_rate": 1.9998856585444493e-05, + "loss": 1.0965, + "step": 320 + }, + { + "epoch": 0.1742671009771987, + "grad_norm": 2.1014332124066954, + "learning_rate": 1.999880278604949e-05, + "loss": 1.3726, + "step": 321 + }, + { + "epoch": 0.1748099891422367, + "grad_norm": 2.0418722317428206, + "learning_rate": 1.9998747749986625e-05, + "loss": 1.8875, + "step": 322 + }, + { + "epoch": 0.1753528773072747, + "grad_norm": 1.768897707524193, + "learning_rate": 1.99986914772627e-05, + "loss": 1.8841, + "step": 323 + }, + { + "epoch": 0.1758957654723127, + "grad_norm": 2.603335158781435, + "learning_rate": 1.9998633967884676e-05, + "loss": 1.4262, + "step": 324 + }, + { + "epoch": 0.1764386536373507, + "grad_norm": 2.1485447383197607, + "learning_rate": 1.999857522185967e-05, + "loss": 1.9283, + "step": 325 + }, + { + "epoch": 0.1769815418023887, + "grad_norm": 1.8398472624384716, + "learning_rate": 1.9998515239194945e-05, + "loss": 1.9089, + "step": 326 + }, + { + "epoch": 0.1775244299674267, + "grad_norm": 2.6988014220525374, + "learning_rate": 1.9998454019897918e-05, + "loss": 1.3636, + "step": 327 + }, + { + "epoch": 0.1780673181324647, + "grad_norm": 2.182794188412918, + "learning_rate": 1.9998391563976166e-05, + "loss": 1.7041, + "step": 328 + }, + { + "epoch": 0.1786102062975027, + "grad_norm": 1.970858895313084, + "learning_rate": 1.9998327871437405e-05, + "loss": 1.4917, + "step": 329 + }, + { + "epoch": 0.1791530944625407, + "grad_norm": 1.829655959511745, + "learning_rate": 1.9998262942289524e-05, + "loss": 1.9694, + "step": 330 + }, + { + "epoch": 0.17969598262757872, + "grad_norm": 1.9990017706312653, + "learning_rate": 1.9998196776540545e-05, + "loss": 2.067, + "step": 331 + }, + { + "epoch": 0.18023887079261672, + "grad_norm": 2.2829395740777234, + "learning_rate": 1.9998129374198655e-05, + "loss": 1.6707, + "step": 332 + }, + { + "epoch": 0.18078175895765472, + "grad_norm": 2.174263401724357, + "learning_rate": 1.9998060735272186e-05, + "loss": 1.6897, + "step": 333 + }, + { + "epoch": 0.18132464712269272, + "grad_norm": 1.8438064757899841, + "learning_rate": 1.9997990859769633e-05, + "loss": 1.4446, + "step": 334 + }, + { + "epoch": 0.18186753528773072, + "grad_norm": 2.0988396037795507, + "learning_rate": 1.9997919747699638e-05, + "loss": 1.8837, + "step": 335 + }, + { + "epoch": 0.18241042345276873, + "grad_norm": 2.582616969632433, + "learning_rate": 1.999784739907099e-05, + "loss": 2.459, + "step": 336 + }, + { + "epoch": 0.18295331161780673, + "grad_norm": 2.5198677585207707, + "learning_rate": 1.9997773813892644e-05, + "loss": 1.5117, + "step": 337 + }, + { + "epoch": 0.18349619978284473, + "grad_norm": 2.1361317951692635, + "learning_rate": 1.9997698992173697e-05, + "loss": 1.5892, + "step": 338 + }, + { + "epoch": 0.18403908794788273, + "grad_norm": 1.7284990063627486, + "learning_rate": 1.9997622933923406e-05, + "loss": 1.5151, + "step": 339 + }, + { + "epoch": 0.18458197611292074, + "grad_norm": 2.076752898704788, + "learning_rate": 1.9997545639151176e-05, + "loss": 2.3424, + "step": 340 + }, + { + "epoch": 0.18512486427795874, + "grad_norm": 2.479369276757794, + "learning_rate": 1.999746710786657e-05, + "loss": 1.5923, + "step": 341 + }, + { + "epoch": 0.18566775244299674, + "grad_norm": 1.9193986648366674, + "learning_rate": 1.9997387340079294e-05, + "loss": 1.6222, + "step": 342 + }, + { + "epoch": 0.18621064060803474, + "grad_norm": 1.8622986153150223, + "learning_rate": 1.999730633579922e-05, + "loss": 1.8402, + "step": 343 + }, + { + "epoch": 0.18675352877307275, + "grad_norm": 2.4012194427686238, + "learning_rate": 1.999722409503637e-05, + "loss": 1.9406, + "step": 344 + }, + { + "epoch": 0.18729641693811075, + "grad_norm": 2.3641518463127538, + "learning_rate": 1.9997140617800907e-05, + "loss": 2.0108, + "step": 345 + }, + { + "epoch": 0.18783930510314875, + "grad_norm": 2.406749148558192, + "learning_rate": 1.9997055904103156e-05, + "loss": 1.1323, + "step": 346 + }, + { + "epoch": 0.18838219326818675, + "grad_norm": 2.002785642517938, + "learning_rate": 1.99969699539536e-05, + "loss": 1.8745, + "step": 347 + }, + { + "epoch": 0.18892508143322476, + "grad_norm": 2.0002882431712816, + "learning_rate": 1.9996882767362874e-05, + "loss": 1.6311, + "step": 348 + }, + { + "epoch": 0.18946796959826276, + "grad_norm": 2.14418043520414, + "learning_rate": 1.9996794344341744e-05, + "loss": 1.7789, + "step": 349 + }, + { + "epoch": 0.19001085776330076, + "grad_norm": 2.0976983604919144, + "learning_rate": 1.9996704684901163e-05, + "loss": 1.9542, + "step": 350 + }, + { + "epoch": 0.19055374592833876, + "grad_norm": 2.017704706993719, + "learning_rate": 1.9996613789052214e-05, + "loss": 1.7974, + "step": 351 + }, + { + "epoch": 0.19109663409337677, + "grad_norm": 1.9907769296520095, + "learning_rate": 1.999652165680614e-05, + "loss": 1.9079, + "step": 352 + }, + { + "epoch": 0.19163952225841477, + "grad_norm": 2.3031451756435746, + "learning_rate": 1.999642828817433e-05, + "loss": 1.7022, + "step": 353 + }, + { + "epoch": 0.19218241042345277, + "grad_norm": 1.6722822353766713, + "learning_rate": 1.9996333683168342e-05, + "loss": 1.6655, + "step": 354 + }, + { + "epoch": 0.19272529858849077, + "grad_norm": 2.2974087830498027, + "learning_rate": 1.9996237841799874e-05, + "loss": 1.761, + "step": 355 + }, + { + "epoch": 0.19326818675352878, + "grad_norm": 1.9181254747298617, + "learning_rate": 1.9996140764080777e-05, + "loss": 1.8259, + "step": 356 + }, + { + "epoch": 0.19381107491856678, + "grad_norm": 2.07933834321742, + "learning_rate": 1.9996042450023053e-05, + "loss": 1.3783, + "step": 357 + }, + { + "epoch": 0.19435396308360478, + "grad_norm": 2.8332787295887436, + "learning_rate": 1.9995942899638875e-05, + "loss": 1.9796, + "step": 358 + }, + { + "epoch": 0.19489685124864278, + "grad_norm": 1.7924418183793758, + "learning_rate": 1.9995842112940545e-05, + "loss": 1.5063, + "step": 359 + }, + { + "epoch": 0.19543973941368079, + "grad_norm": 2.78627522830663, + "learning_rate": 1.9995740089940532e-05, + "loss": 1.891, + "step": 360 + }, + { + "epoch": 0.1959826275787188, + "grad_norm": 2.067311381839208, + "learning_rate": 1.9995636830651453e-05, + "loss": 1.9235, + "step": 361 + }, + { + "epoch": 0.1965255157437568, + "grad_norm": 2.539112412358468, + "learning_rate": 1.9995532335086078e-05, + "loss": 2.3062, + "step": 362 + }, + { + "epoch": 0.1970684039087948, + "grad_norm": 1.9040915545393904, + "learning_rate": 1.999542660325734e-05, + "loss": 1.9263, + "step": 363 + }, + { + "epoch": 0.1976112920738328, + "grad_norm": 1.7664296640025938, + "learning_rate": 1.9995319635178305e-05, + "loss": 1.5562, + "step": 364 + }, + { + "epoch": 0.1981541802388708, + "grad_norm": 1.687746217581272, + "learning_rate": 1.9995211430862206e-05, + "loss": 1.738, + "step": 365 + }, + { + "epoch": 0.1986970684039088, + "grad_norm": 1.9292930460598687, + "learning_rate": 1.9995101990322428e-05, + "loss": 1.5446, + "step": 366 + }, + { + "epoch": 0.1992399565689468, + "grad_norm": 1.9890390069401638, + "learning_rate": 1.9994991313572508e-05, + "loss": 1.9086, + "step": 367 + }, + { + "epoch": 0.1997828447339848, + "grad_norm": 1.8632430045841502, + "learning_rate": 1.999487940062613e-05, + "loss": 1.9747, + "step": 368 + }, + { + "epoch": 0.2003257328990228, + "grad_norm": 2.0677818303742237, + "learning_rate": 1.999476625149714e-05, + "loss": 1.4377, + "step": 369 + }, + { + "epoch": 0.2008686210640608, + "grad_norm": 1.7412205305358945, + "learning_rate": 1.9994651866199527e-05, + "loss": 1.9352, + "step": 370 + }, + { + "epoch": 0.2014115092290988, + "grad_norm": 1.8562175126383418, + "learning_rate": 1.9994536244747448e-05, + "loss": 1.5469, + "step": 371 + }, + { + "epoch": 0.20195439739413681, + "grad_norm": 1.8688345857586197, + "learning_rate": 1.9994419387155194e-05, + "loss": 1.8327, + "step": 372 + }, + { + "epoch": 0.20249728555917482, + "grad_norm": 2.1538407513899647, + "learning_rate": 1.9994301293437223e-05, + "loss": 1.9621, + "step": 373 + }, + { + "epoch": 0.20304017372421282, + "grad_norm": 2.4849664096044752, + "learning_rate": 1.999418196360814e-05, + "loss": 1.8179, + "step": 374 + }, + { + "epoch": 0.20358306188925082, + "grad_norm": 2.098949652063271, + "learning_rate": 1.99940613976827e-05, + "loss": 1.8719, + "step": 375 + }, + { + "epoch": 0.20412595005428882, + "grad_norm": 1.8802547433747465, + "learning_rate": 1.999393959567582e-05, + "loss": 1.5697, + "step": 376 + }, + { + "epoch": 0.20466883821932683, + "grad_norm": 2.090861017905227, + "learning_rate": 1.9993816557602567e-05, + "loss": 1.8249, + "step": 377 + }, + { + "epoch": 0.20521172638436483, + "grad_norm": 1.775481861127165, + "learning_rate": 1.999369228347815e-05, + "loss": 1.3095, + "step": 378 + }, + { + "epoch": 0.20575461454940283, + "grad_norm": 2.260330600420479, + "learning_rate": 1.999356677331794e-05, + "loss": 1.9153, + "step": 379 + }, + { + "epoch": 0.20629750271444083, + "grad_norm": 2.0769016143551484, + "learning_rate": 1.999344002713747e-05, + "loss": 1.958, + "step": 380 + }, + { + "epoch": 0.20684039087947884, + "grad_norm": 2.063584775140761, + "learning_rate": 1.9993312044952408e-05, + "loss": 1.7887, + "step": 381 + }, + { + "epoch": 0.20738327904451684, + "grad_norm": 1.9638736053910157, + "learning_rate": 1.9993182826778588e-05, + "loss": 1.3178, + "step": 382 + }, + { + "epoch": 0.20792616720955484, + "grad_norm": 1.7752329715548703, + "learning_rate": 1.9993052372631988e-05, + "loss": 1.4473, + "step": 383 + }, + { + "epoch": 0.20846905537459284, + "grad_norm": 2.1002641405528952, + "learning_rate": 1.999292068252874e-05, + "loss": 1.9698, + "step": 384 + }, + { + "epoch": 0.20901194353963085, + "grad_norm": 2.757803188261632, + "learning_rate": 1.999278775648514e-05, + "loss": 1.1517, + "step": 385 + }, + { + "epoch": 0.20955483170466885, + "grad_norm": 3.0341284449704378, + "learning_rate": 1.9992653594517624e-05, + "loss": 1.987, + "step": 386 + }, + { + "epoch": 0.21009771986970685, + "grad_norm": 2.445830333821562, + "learning_rate": 1.9992518196642786e-05, + "loss": 1.7656, + "step": 387 + }, + { + "epoch": 0.21064060803474485, + "grad_norm": 2.737783962568244, + "learning_rate": 1.9992381562877368e-05, + "loss": 2.3012, + "step": 388 + }, + { + "epoch": 0.21118349619978286, + "grad_norm": 2.0108432149096793, + "learning_rate": 1.9992243693238275e-05, + "loss": 2.1096, + "step": 389 + }, + { + "epoch": 0.21172638436482086, + "grad_norm": 2.1448465445226397, + "learning_rate": 1.9992104587742558e-05, + "loss": 1.6912, + "step": 390 + }, + { + "epoch": 0.21226927252985886, + "grad_norm": 2.842739011518123, + "learning_rate": 1.999196424640742e-05, + "loss": 1.9987, + "step": 391 + }, + { + "epoch": 0.21281216069489686, + "grad_norm": 2.026900017823922, + "learning_rate": 1.9991822669250216e-05, + "loss": 1.6251, + "step": 392 + }, + { + "epoch": 0.21335504885993486, + "grad_norm": 2.139818571931489, + "learning_rate": 1.9991679856288462e-05, + "loss": 1.4181, + "step": 393 + }, + { + "epoch": 0.21389793702497287, + "grad_norm": 2.061327407258535, + "learning_rate": 1.999153580753982e-05, + "loss": 2.0976, + "step": 394 + }, + { + "epoch": 0.21444082519001087, + "grad_norm": 1.9510646261617408, + "learning_rate": 1.9991390523022105e-05, + "loss": 1.3091, + "step": 395 + }, + { + "epoch": 0.21498371335504887, + "grad_norm": 2.0892427860245664, + "learning_rate": 1.9991244002753287e-05, + "loss": 1.3693, + "step": 396 + }, + { + "epoch": 0.21552660152008687, + "grad_norm": 1.9528708238712815, + "learning_rate": 1.9991096246751483e-05, + "loss": 1.2807, + "step": 397 + }, + { + "epoch": 0.21606948968512488, + "grad_norm": 2.0084632909809983, + "learning_rate": 1.9990947255034977e-05, + "loss": 1.7429, + "step": 398 + }, + { + "epoch": 0.21661237785016288, + "grad_norm": 2.19414527352725, + "learning_rate": 1.999079702762219e-05, + "loss": 1.5962, + "step": 399 + }, + { + "epoch": 0.21715526601520088, + "grad_norm": 2.184484203373308, + "learning_rate": 1.9990645564531702e-05, + "loss": 1.6887, + "step": 400 + }, + { + "epoch": 0.21769815418023888, + "grad_norm": 2.214473402714986, + "learning_rate": 1.9990492865782248e-05, + "loss": 1.5226, + "step": 401 + }, + { + "epoch": 0.2182410423452769, + "grad_norm": 2.1467748369867414, + "learning_rate": 1.9990338931392714e-05, + "loss": 1.5634, + "step": 402 + }, + { + "epoch": 0.21878393051031486, + "grad_norm": 2.1969353910959684, + "learning_rate": 1.999018376138214e-05, + "loss": 1.8652, + "step": 403 + }, + { + "epoch": 0.21932681867535286, + "grad_norm": 2.245031544698939, + "learning_rate": 1.9990027355769715e-05, + "loss": 2.1811, + "step": 404 + }, + { + "epoch": 0.21986970684039087, + "grad_norm": 2.381275606702572, + "learning_rate": 1.9989869714574784e-05, + "loss": 1.6356, + "step": 405 + }, + { + "epoch": 0.22041259500542887, + "grad_norm": 2.8847531540936053, + "learning_rate": 1.9989710837816846e-05, + "loss": 1.9953, + "step": 406 + }, + { + "epoch": 0.22095548317046687, + "grad_norm": 2.383074182943421, + "learning_rate": 1.9989550725515553e-05, + "loss": 1.5589, + "step": 407 + }, + { + "epoch": 0.22149837133550487, + "grad_norm": 2.3867744628533987, + "learning_rate": 1.99893893776907e-05, + "loss": 1.5351, + "step": 408 + }, + { + "epoch": 0.22204125950054288, + "grad_norm": 2.4139767232435463, + "learning_rate": 1.998922679436225e-05, + "loss": 1.5949, + "step": 409 + }, + { + "epoch": 0.22258414766558088, + "grad_norm": 3.034840719531937, + "learning_rate": 1.9989062975550313e-05, + "loss": 1.7994, + "step": 410 + }, + { + "epoch": 0.22312703583061888, + "grad_norm": 2.6245673280102544, + "learning_rate": 1.9988897921275144e-05, + "loss": 2.2593, + "step": 411 + }, + { + "epoch": 0.22366992399565688, + "grad_norm": 2.403934048820897, + "learning_rate": 1.998873163155716e-05, + "loss": 1.3179, + "step": 412 + }, + { + "epoch": 0.22421281216069489, + "grad_norm": 2.0174357446052524, + "learning_rate": 1.998856410641693e-05, + "loss": 1.8698, + "step": 413 + }, + { + "epoch": 0.2247557003257329, + "grad_norm": 2.25869172419699, + "learning_rate": 1.998839534587517e-05, + "loss": 1.596, + "step": 414 + }, + { + "epoch": 0.2252985884907709, + "grad_norm": 3.384596068816443, + "learning_rate": 1.9988225349952758e-05, + "loss": 1.5232, + "step": 415 + }, + { + "epoch": 0.2258414766558089, + "grad_norm": 2.2109577343653246, + "learning_rate": 1.9988054118670712e-05, + "loss": 1.2403, + "step": 416 + }, + { + "epoch": 0.2263843648208469, + "grad_norm": 2.1825505717325053, + "learning_rate": 1.9987881652050215e-05, + "loss": 1.7205, + "step": 417 + }, + { + "epoch": 0.2269272529858849, + "grad_norm": 2.7992033856118583, + "learning_rate": 1.99877079501126e-05, + "loss": 1.7486, + "step": 418 + }, + { + "epoch": 0.2274701411509229, + "grad_norm": 2.6351781282432793, + "learning_rate": 1.9987533012879344e-05, + "loss": 1.6693, + "step": 419 + }, + { + "epoch": 0.2280130293159609, + "grad_norm": 2.2385207734696384, + "learning_rate": 1.9987356840372088e-05, + "loss": 1.1296, + "step": 420 + }, + { + "epoch": 0.2285559174809989, + "grad_norm": 2.4921881643709107, + "learning_rate": 1.998717943261262e-05, + "loss": 1.1379, + "step": 421 + }, + { + "epoch": 0.2290988056460369, + "grad_norm": 2.842685080646979, + "learning_rate": 1.9987000789622884e-05, + "loss": 1.7963, + "step": 422 + }, + { + "epoch": 0.2296416938110749, + "grad_norm": 2.4776027353414145, + "learning_rate": 1.9986820911424972e-05, + "loss": 1.6939, + "step": 423 + }, + { + "epoch": 0.2301845819761129, + "grad_norm": 2.688785422525701, + "learning_rate": 1.9986639798041134e-05, + "loss": 1.2282, + "step": 424 + }, + { + "epoch": 0.23072747014115091, + "grad_norm": 3.4758865457335495, + "learning_rate": 1.998645744949377e-05, + "loss": 1.7664, + "step": 425 + }, + { + "epoch": 0.23127035830618892, + "grad_norm": 2.5586740557561605, + "learning_rate": 1.9986273865805432e-05, + "loss": 1.5637, + "step": 426 + }, + { + "epoch": 0.23181324647122692, + "grad_norm": 2.645306116903878, + "learning_rate": 1.9986089046998827e-05, + "loss": 1.5331, + "step": 427 + }, + { + "epoch": 0.23235613463626492, + "grad_norm": 2.748145883999443, + "learning_rate": 1.998590299309681e-05, + "loss": 1.7316, + "step": 428 + }, + { + "epoch": 0.23289902280130292, + "grad_norm": 2.4357923833746438, + "learning_rate": 1.99857157041224e-05, + "loss": 1.1479, + "step": 429 + }, + { + "epoch": 0.23344191096634093, + "grad_norm": 2.7322633978331377, + "learning_rate": 1.9985527180098755e-05, + "loss": 1.6089, + "step": 430 + }, + { + "epoch": 0.23398479913137893, + "grad_norm": 2.625531215380387, + "learning_rate": 1.9985337421049193e-05, + "loss": 1.8973, + "step": 431 + }, + { + "epoch": 0.23452768729641693, + "grad_norm": 2.1586487666504754, + "learning_rate": 1.9985146426997185e-05, + "loss": 1.664, + "step": 432 + }, + { + "epoch": 0.23507057546145493, + "grad_norm": 2.748704307086674, + "learning_rate": 1.9984954197966355e-05, + "loss": 1.677, + "step": 433 + }, + { + "epoch": 0.23561346362649294, + "grad_norm": 2.345532526959197, + "learning_rate": 1.9984760733980476e-05, + "loss": 1.5133, + "step": 434 + }, + { + "epoch": 0.23615635179153094, + "grad_norm": 2.1655585346308848, + "learning_rate": 1.9984566035063473e-05, + "loss": 1.4206, + "step": 435 + }, + { + "epoch": 0.23669923995656894, + "grad_norm": 2.697584333049271, + "learning_rate": 1.9984370101239434e-05, + "loss": 1.8131, + "step": 436 + }, + { + "epoch": 0.23724212812160694, + "grad_norm": 2.235069158786981, + "learning_rate": 1.9984172932532583e-05, + "loss": 1.7839, + "step": 437 + }, + { + "epoch": 0.23778501628664495, + "grad_norm": 2.548562657856099, + "learning_rate": 1.998397452896731e-05, + "loss": 1.4876, + "step": 438 + }, + { + "epoch": 0.23832790445168295, + "grad_norm": 2.1698874138883673, + "learning_rate": 1.9983774890568163e-05, + "loss": 1.2916, + "step": 439 + }, + { + "epoch": 0.23887079261672095, + "grad_norm": 2.3237954527681084, + "learning_rate": 1.998357401735982e-05, + "loss": 1.398, + "step": 440 + }, + { + "epoch": 0.23941368078175895, + "grad_norm": 2.448364938314344, + "learning_rate": 1.9983371909367135e-05, + "loss": 1.2663, + "step": 441 + }, + { + "epoch": 0.23995656894679696, + "grad_norm": 1.935899726785714, + "learning_rate": 1.99831685666151e-05, + "loss": 1.2448, + "step": 442 + }, + { + "epoch": 0.24049945711183496, + "grad_norm": 2.14357016947003, + "learning_rate": 1.9982963989128864e-05, + "loss": 1.4786, + "step": 443 + }, + { + "epoch": 0.24104234527687296, + "grad_norm": 2.5830135196717148, + "learning_rate": 1.998275817693373e-05, + "loss": 1.8324, + "step": 444 + }, + { + "epoch": 0.24158523344191096, + "grad_norm": 3.159289299734819, + "learning_rate": 1.9982551130055157e-05, + "loss": 1.9091, + "step": 445 + }, + { + "epoch": 0.24212812160694897, + "grad_norm": 3.2712239674501755, + "learning_rate": 1.9982342848518753e-05, + "loss": 1.7869, + "step": 446 + }, + { + "epoch": 0.24267100977198697, + "grad_norm": 3.35834841128443, + "learning_rate": 1.998213333235027e-05, + "loss": 1.2772, + "step": 447 + }, + { + "epoch": 0.24321389793702497, + "grad_norm": 2.9379666136057354, + "learning_rate": 1.998192258157563e-05, + "loss": 1.3497, + "step": 448 + }, + { + "epoch": 0.24375678610206297, + "grad_norm": 2.334119075845609, + "learning_rate": 1.9981710596220897e-05, + "loss": 1.5541, + "step": 449 + }, + { + "epoch": 0.24429967426710097, + "grad_norm": 2.781469856763384, + "learning_rate": 1.998149737631229e-05, + "loss": 1.9413, + "step": 450 + }, + { + "epoch": 0.24484256243213898, + "grad_norm": 3.0196781534130452, + "learning_rate": 1.9981282921876177e-05, + "loss": 1.3238, + "step": 451 + }, + { + "epoch": 0.24538545059717698, + "grad_norm": 3.547169321727429, + "learning_rate": 1.9981067232939086e-05, + "loss": 1.9952, + "step": 452 + }, + { + "epoch": 0.24592833876221498, + "grad_norm": 3.2411499530913535, + "learning_rate": 1.9980850309527693e-05, + "loss": 1.8244, + "step": 453 + }, + { + "epoch": 0.24647122692725298, + "grad_norm": 2.670313260104859, + "learning_rate": 1.9980632151668822e-05, + "loss": 1.607, + "step": 454 + }, + { + "epoch": 0.247014115092291, + "grad_norm": 2.8822100638306143, + "learning_rate": 1.9980412759389468e-05, + "loss": 1.8868, + "step": 455 + }, + { + "epoch": 0.247557003257329, + "grad_norm": 2.4776152417583317, + "learning_rate": 1.9980192132716748e-05, + "loss": 1.778, + "step": 456 + }, + { + "epoch": 0.248099891422367, + "grad_norm": 2.2001923672712076, + "learning_rate": 1.9979970271677967e-05, + "loss": 1.3544, + "step": 457 + }, + { + "epoch": 0.248642779587405, + "grad_norm": 2.7694932683911837, + "learning_rate": 1.9979747176300553e-05, + "loss": 1.6521, + "step": 458 + }, + { + "epoch": 0.249185667752443, + "grad_norm": 2.8464573075472845, + "learning_rate": 1.99795228466121e-05, + "loss": 1.8803, + "step": 459 + }, + { + "epoch": 0.249728555917481, + "grad_norm": 2.577989994947286, + "learning_rate": 1.9979297282640365e-05, + "loss": 1.8838, + "step": 460 + }, + { + "epoch": 0.250271444082519, + "grad_norm": 2.9859834231033164, + "learning_rate": 1.997907048441323e-05, + "loss": 1.5122, + "step": 461 + }, + { + "epoch": 0.250814332247557, + "grad_norm": 3.0857580735568098, + "learning_rate": 1.9978842451958757e-05, + "loss": 1.7789, + "step": 462 + }, + { + "epoch": 0.251357220412595, + "grad_norm": 2.523783848348806, + "learning_rate": 1.9978613185305145e-05, + "loss": 1.5815, + "step": 463 + }, + { + "epoch": 0.251900108577633, + "grad_norm": 2.699094362013039, + "learning_rate": 1.9978382684480747e-05, + "loss": 1.7448, + "step": 464 + }, + { + "epoch": 0.252442996742671, + "grad_norm": 2.736268550628698, + "learning_rate": 1.997815094951408e-05, + "loss": 1.7052, + "step": 465 + }, + { + "epoch": 0.252985884907709, + "grad_norm": 3.1150828428090014, + "learning_rate": 1.99779179804338e-05, + "loss": 1.7743, + "step": 466 + }, + { + "epoch": 0.253528773072747, + "grad_norm": 3.3513164619888482, + "learning_rate": 1.997768377726872e-05, + "loss": 2.3905, + "step": 467 + }, + { + "epoch": 0.254071661237785, + "grad_norm": 2.400886982379507, + "learning_rate": 1.9977448340047808e-05, + "loss": 1.6096, + "step": 468 + }, + { + "epoch": 0.254614549402823, + "grad_norm": 2.633533044966171, + "learning_rate": 1.9977211668800186e-05, + "loss": 1.7796, + "step": 469 + }, + { + "epoch": 0.255157437567861, + "grad_norm": 3.9297139007235042, + "learning_rate": 1.997697376355512e-05, + "loss": 1.4602, + "step": 470 + }, + { + "epoch": 0.255700325732899, + "grad_norm": 5.588260619709643, + "learning_rate": 1.9976734624342044e-05, + "loss": 2.0389, + "step": 471 + }, + { + "epoch": 0.256243213897937, + "grad_norm": 2.600723171476426, + "learning_rate": 1.9976494251190522e-05, + "loss": 1.3676, + "step": 472 + }, + { + "epoch": 0.25678610206297503, + "grad_norm": 2.61945002649116, + "learning_rate": 1.9976252644130297e-05, + "loss": 1.7902, + "step": 473 + }, + { + "epoch": 0.25732899022801303, + "grad_norm": 3.7513561790803838, + "learning_rate": 1.997600980319124e-05, + "loss": 1.5997, + "step": 474 + }, + { + "epoch": 0.25787187839305103, + "grad_norm": 2.8832359552778737, + "learning_rate": 1.9975765728403395e-05, + "loss": 1.9636, + "step": 475 + }, + { + "epoch": 0.25841476655808904, + "grad_norm": 3.0975018752600243, + "learning_rate": 1.9975520419796942e-05, + "loss": 1.1165, + "step": 476 + }, + { + "epoch": 0.25895765472312704, + "grad_norm": 2.121708502818221, + "learning_rate": 1.9975273877402227e-05, + "loss": 1.5108, + "step": 477 + }, + { + "epoch": 0.25950054288816504, + "grad_norm": 2.4073592870530116, + "learning_rate": 1.997502610124974e-05, + "loss": 1.7828, + "step": 478 + }, + { + "epoch": 0.26004343105320304, + "grad_norm": 2.6858679986632974, + "learning_rate": 1.997477709137013e-05, + "loss": 1.8483, + "step": 479 + }, + { + "epoch": 0.26058631921824105, + "grad_norm": 3.283752190131325, + "learning_rate": 1.997452684779419e-05, + "loss": 1.6105, + "step": 480 + }, + { + "epoch": 0.26112920738327905, + "grad_norm": 2.500181185675909, + "learning_rate": 1.997427537055287e-05, + "loss": 1.5475, + "step": 481 + }, + { + "epoch": 0.26167209554831705, + "grad_norm": 2.3992324550953885, + "learning_rate": 1.9974022659677278e-05, + "loss": 1.6062, + "step": 482 + }, + { + "epoch": 0.26221498371335505, + "grad_norm": 2.7692293822867837, + "learning_rate": 1.9973768715198667e-05, + "loss": 1.4995, + "step": 483 + }, + { + "epoch": 0.26275787187839306, + "grad_norm": 2.971423364277874, + "learning_rate": 1.9973513537148447e-05, + "loss": 1.7904, + "step": 484 + }, + { + "epoch": 0.26330076004343106, + "grad_norm": 2.2769736321644105, + "learning_rate": 1.9973257125558177e-05, + "loss": 1.3121, + "step": 485 + }, + { + "epoch": 0.26384364820846906, + "grad_norm": 2.5949046051899254, + "learning_rate": 1.997299948045957e-05, + "loss": 1.4555, + "step": 486 + }, + { + "epoch": 0.26438653637350706, + "grad_norm": 3.0210593683445204, + "learning_rate": 1.997274060188449e-05, + "loss": 1.8211, + "step": 487 + }, + { + "epoch": 0.26492942453854507, + "grad_norm": 3.076626224616319, + "learning_rate": 1.9972480489864962e-05, + "loss": 1.6366, + "step": 488 + }, + { + "epoch": 0.26547231270358307, + "grad_norm": 3.1369728365663536, + "learning_rate": 1.9972219144433148e-05, + "loss": 1.5027, + "step": 489 + }, + { + "epoch": 0.26601520086862107, + "grad_norm": 2.2870450349164635, + "learning_rate": 1.9971956565621383e-05, + "loss": 1.2784, + "step": 490 + }, + { + "epoch": 0.2665580890336591, + "grad_norm": 3.0130036065633776, + "learning_rate": 1.9971692753462134e-05, + "loss": 1.2083, + "step": 491 + }, + { + "epoch": 0.2671009771986971, + "grad_norm": 2.6063157323029733, + "learning_rate": 1.9971427707988034e-05, + "loss": 1.0083, + "step": 492 + }, + { + "epoch": 0.2676438653637351, + "grad_norm": 2.579371053895234, + "learning_rate": 1.997116142923186e-05, + "loss": 1.1937, + "step": 493 + }, + { + "epoch": 0.2681867535287731, + "grad_norm": 2.532537971800688, + "learning_rate": 1.9970893917226554e-05, + "loss": 1.4735, + "step": 494 + }, + { + "epoch": 0.2687296416938111, + "grad_norm": 2.3483150144294105, + "learning_rate": 1.997062517200519e-05, + "loss": 1.7269, + "step": 495 + }, + { + "epoch": 0.2692725298588491, + "grad_norm": 2.594809867192747, + "learning_rate": 1.997035519360102e-05, + "loss": 1.8283, + "step": 496 + }, + { + "epoch": 0.2698154180238871, + "grad_norm": 2.613813750609998, + "learning_rate": 1.9970083982047428e-05, + "loss": 1.2302, + "step": 497 + }, + { + "epoch": 0.2703583061889251, + "grad_norm": 2.648279162964909, + "learning_rate": 1.9969811537377956e-05, + "loss": 1.6225, + "step": 498 + }, + { + "epoch": 0.2709011943539631, + "grad_norm": 2.3790190706794325, + "learning_rate": 1.9969537859626308e-05, + "loss": 1.5172, + "step": 499 + }, + { + "epoch": 0.2714440825190011, + "grad_norm": 2.7054998578606364, + "learning_rate": 1.9969262948826326e-05, + "loss": 1.4525, + "step": 500 + }, + { + "epoch": 0.2719869706840391, + "grad_norm": 2.412151508264948, + "learning_rate": 1.9968986805012012e-05, + "loss": 1.3299, + "step": 501 + }, + { + "epoch": 0.2725298588490771, + "grad_norm": 2.4836460319285414, + "learning_rate": 1.9968709428217525e-05, + "loss": 1.5217, + "step": 502 + }, + { + "epoch": 0.2730727470141151, + "grad_norm": 3.3112944949184606, + "learning_rate": 1.9968430818477168e-05, + "loss": 2.0643, + "step": 503 + }, + { + "epoch": 0.2736156351791531, + "grad_norm": 3.266395629954733, + "learning_rate": 1.9968150975825397e-05, + "loss": 1.1719, + "step": 504 + }, + { + "epoch": 0.2741585233441911, + "grad_norm": 2.3487351103507073, + "learning_rate": 1.996786990029683e-05, + "loss": 1.6876, + "step": 505 + }, + { + "epoch": 0.2747014115092291, + "grad_norm": 3.0672442719402673, + "learning_rate": 1.9967587591926227e-05, + "loss": 1.3946, + "step": 506 + }, + { + "epoch": 0.2752442996742671, + "grad_norm": 2.84320621483769, + "learning_rate": 1.99673040507485e-05, + "loss": 1.3259, + "step": 507 + }, + { + "epoch": 0.2757871878393051, + "grad_norm": 2.5078449617010707, + "learning_rate": 1.9967019276798728e-05, + "loss": 0.8478, + "step": 508 + }, + { + "epoch": 0.2763300760043431, + "grad_norm": 3.137784699454796, + "learning_rate": 1.9966733270112126e-05, + "loss": 1.1688, + "step": 509 + }, + { + "epoch": 0.2768729641693811, + "grad_norm": 3.7277826491955017, + "learning_rate": 1.996644603072407e-05, + "loss": 1.1091, + "step": 510 + }, + { + "epoch": 0.2774158523344191, + "grad_norm": 2.5276781604415635, + "learning_rate": 1.996615755867008e-05, + "loss": 1.1299, + "step": 511 + }, + { + "epoch": 0.2779587404994571, + "grad_norm": 3.1012700661738744, + "learning_rate": 1.996586785398584e-05, + "loss": 1.7218, + "step": 512 + }, + { + "epoch": 0.2785016286644951, + "grad_norm": 2.285166712515903, + "learning_rate": 1.9965576916707182e-05, + "loss": 1.2868, + "step": 513 + }, + { + "epoch": 0.27904451682953313, + "grad_norm": 2.33097906349044, + "learning_rate": 1.9965284746870088e-05, + "loss": 0.9887, + "step": 514 + }, + { + "epoch": 0.27958740499457113, + "grad_norm": 2.6473787082237927, + "learning_rate": 1.9964991344510697e-05, + "loss": 1.8543, + "step": 515 + }, + { + "epoch": 0.28013029315960913, + "grad_norm": 2.4628160599533366, + "learning_rate": 1.996469670966529e-05, + "loss": 1.2263, + "step": 516 + }, + { + "epoch": 0.28067318132464714, + "grad_norm": 3.2897847068350905, + "learning_rate": 1.9964400842370314e-05, + "loss": 1.6338, + "step": 517 + }, + { + "epoch": 0.28121606948968514, + "grad_norm": 2.4439319341540324, + "learning_rate": 1.9964103742662363e-05, + "loss": 1.0836, + "step": 518 + }, + { + "epoch": 0.28175895765472314, + "grad_norm": 2.3221991020412003, + "learning_rate": 1.996380541057818e-05, + "loss": 1.2331, + "step": 519 + }, + { + "epoch": 0.28230184581976114, + "grad_norm": 2.9571040634251564, + "learning_rate": 1.9963505846154662e-05, + "loss": 1.3066, + "step": 520 + }, + { + "epoch": 0.28284473398479915, + "grad_norm": 3.7512706020225624, + "learning_rate": 1.996320504942886e-05, + "loss": 1.7482, + "step": 521 + }, + { + "epoch": 0.28338762214983715, + "grad_norm": 2.4620109793388267, + "learning_rate": 1.9962903020437983e-05, + "loss": 1.5334, + "step": 522 + }, + { + "epoch": 0.28393051031487515, + "grad_norm": 3.030374272795485, + "learning_rate": 1.9962599759219383e-05, + "loss": 1.8957, + "step": 523 + }, + { + "epoch": 0.28447339847991315, + "grad_norm": 2.452389821491403, + "learning_rate": 1.9962295265810563e-05, + "loss": 1.5438, + "step": 524 + }, + { + "epoch": 0.28501628664495116, + "grad_norm": 2.713028369466205, + "learning_rate": 1.996198954024919e-05, + "loss": 1.4272, + "step": 525 + }, + { + "epoch": 0.28555917480998916, + "grad_norm": 3.810321275175567, + "learning_rate": 1.996168258257307e-05, + "loss": 1.9028, + "step": 526 + }, + { + "epoch": 0.28610206297502716, + "grad_norm": 2.7774100977441236, + "learning_rate": 1.9961374392820173e-05, + "loss": 1.5644, + "step": 527 + }, + { + "epoch": 0.28664495114006516, + "grad_norm": 2.6798712089104186, + "learning_rate": 1.9961064971028616e-05, + "loss": 1.188, + "step": 528 + }, + { + "epoch": 0.28718783930510317, + "grad_norm": 2.889465990486677, + "learning_rate": 1.9960754317236666e-05, + "loss": 1.5393, + "step": 529 + }, + { + "epoch": 0.28773072747014117, + "grad_norm": 3.465251366831076, + "learning_rate": 1.996044243148275e-05, + "loss": 2.1899, + "step": 530 + }, + { + "epoch": 0.28827361563517917, + "grad_norm": 2.645941940974219, + "learning_rate": 1.9960129313805437e-05, + "loss": 1.3691, + "step": 531 + }, + { + "epoch": 0.2888165038002172, + "grad_norm": 2.3914199977194293, + "learning_rate": 1.9959814964243455e-05, + "loss": 1.5219, + "step": 532 + }, + { + "epoch": 0.2893593919652552, + "grad_norm": 2.3023810529281343, + "learning_rate": 1.995949938283569e-05, + "loss": 1.5147, + "step": 533 + }, + { + "epoch": 0.2899022801302932, + "grad_norm": 2.7362205671791155, + "learning_rate": 1.9959182569621164e-05, + "loss": 1.7571, + "step": 534 + }, + { + "epoch": 0.2904451682953312, + "grad_norm": 3.971162331076012, + "learning_rate": 1.9958864524639066e-05, + "loss": 1.3425, + "step": 535 + }, + { + "epoch": 0.2909880564603692, + "grad_norm": 3.4144928239616514, + "learning_rate": 1.9958545247928727e-05, + "loss": 1.6962, + "step": 536 + }, + { + "epoch": 0.2915309446254072, + "grad_norm": 3.5063126675319043, + "learning_rate": 1.9958224739529647e-05, + "loss": 1.6406, + "step": 537 + }, + { + "epoch": 0.2920738327904452, + "grad_norm": 2.9013783116047547, + "learning_rate": 1.995790299948146e-05, + "loss": 1.6376, + "step": 538 + }, + { + "epoch": 0.2926167209554832, + "grad_norm": 2.4827691033904693, + "learning_rate": 1.9957580027823957e-05, + "loss": 1.8672, + "step": 539 + }, + { + "epoch": 0.2931596091205212, + "grad_norm": 3.1245563354940242, + "learning_rate": 1.9957255824597087e-05, + "loss": 1.885, + "step": 540 + }, + { + "epoch": 0.2937024972855592, + "grad_norm": 2.7941860466759896, + "learning_rate": 1.9956930389840945e-05, + "loss": 1.0903, + "step": 541 + }, + { + "epoch": 0.2942453854505972, + "grad_norm": 3.1348104146717772, + "learning_rate": 1.9956603723595784e-05, + "loss": 2.1446, + "step": 542 + }, + { + "epoch": 0.2947882736156352, + "grad_norm": 2.4248222663066747, + "learning_rate": 1.995627582590201e-05, + "loss": 1.3503, + "step": 543 + }, + { + "epoch": 0.2953311617806732, + "grad_norm": 3.1538245864476337, + "learning_rate": 1.995594669680017e-05, + "loss": 1.3275, + "step": 544 + }, + { + "epoch": 0.2958740499457112, + "grad_norm": 3.436905240480997, + "learning_rate": 1.9955616336330976e-05, + "loss": 1.7249, + "step": 545 + }, + { + "epoch": 0.2964169381107492, + "grad_norm": 3.9145352507266393, + "learning_rate": 1.9955284744535287e-05, + "loss": 2.0089, + "step": 546 + }, + { + "epoch": 0.2969598262757872, + "grad_norm": 2.8600509325831185, + "learning_rate": 1.9954951921454113e-05, + "loss": 1.4527, + "step": 547 + }, + { + "epoch": 0.2975027144408252, + "grad_norm": 2.865760632888347, + "learning_rate": 1.995461786712862e-05, + "loss": 1.4553, + "step": 548 + }, + { + "epoch": 0.2980456026058632, + "grad_norm": 3.3559882259900706, + "learning_rate": 1.9954282581600127e-05, + "loss": 2.0456, + "step": 549 + }, + { + "epoch": 0.2985884907709012, + "grad_norm": 3.3424656149019008, + "learning_rate": 1.9953946064910098e-05, + "loss": 1.8253, + "step": 550 + }, + { + "epoch": 0.2991313789359392, + "grad_norm": 2.569119318410074, + "learning_rate": 1.9953608317100153e-05, + "loss": 1.3623, + "step": 551 + }, + { + "epoch": 0.2996742671009772, + "grad_norm": 2.8918953822102424, + "learning_rate": 1.995326933821207e-05, + "loss": 1.7521, + "step": 552 + }, + { + "epoch": 0.3002171552660152, + "grad_norm": 2.7627900710192246, + "learning_rate": 1.995292912828777e-05, + "loss": 1.775, + "step": 553 + }, + { + "epoch": 0.3007600434310532, + "grad_norm": 2.8053609300694804, + "learning_rate": 1.9952587687369334e-05, + "loss": 1.7536, + "step": 554 + }, + { + "epoch": 0.30130293159609123, + "grad_norm": 2.9775715301146803, + "learning_rate": 1.995224501549899e-05, + "loss": 1.6715, + "step": 555 + }, + { + "epoch": 0.30184581976112923, + "grad_norm": 3.146995410263436, + "learning_rate": 1.9951901112719123e-05, + "loss": 1.1032, + "step": 556 + }, + { + "epoch": 0.30238870792616723, + "grad_norm": 3.219197817112143, + "learning_rate": 1.9951555979072266e-05, + "loss": 1.6326, + "step": 557 + }, + { + "epoch": 0.30293159609120524, + "grad_norm": 2.676508071644292, + "learning_rate": 1.99512096146011e-05, + "loss": 1.4836, + "step": 558 + }, + { + "epoch": 0.30347448425624324, + "grad_norm": 3.8806900857620374, + "learning_rate": 1.9950862019348474e-05, + "loss": 1.7794, + "step": 559 + }, + { + "epoch": 0.30401737242128124, + "grad_norm": 2.547409032322543, + "learning_rate": 1.995051319335737e-05, + "loss": 1.3263, + "step": 560 + }, + { + "epoch": 0.30456026058631924, + "grad_norm": 3.025189850713409, + "learning_rate": 1.995016313667094e-05, + "loss": 1.2409, + "step": 561 + }, + { + "epoch": 0.30510314875135724, + "grad_norm": 3.3644665856402614, + "learning_rate": 1.9949811849332476e-05, + "loss": 1.2988, + "step": 562 + }, + { + "epoch": 0.30564603691639525, + "grad_norm": 3.4126932100522755, + "learning_rate": 1.9949459331385422e-05, + "loss": 1.6126, + "step": 563 + }, + { + "epoch": 0.30618892508143325, + "grad_norm": 3.3751971677759416, + "learning_rate": 1.994910558287338e-05, + "loss": 1.6243, + "step": 564 + }, + { + "epoch": 0.30673181324647125, + "grad_norm": 3.0948952637866105, + "learning_rate": 1.9948750603840102e-05, + "loss": 1.6553, + "step": 565 + }, + { + "epoch": 0.30727470141150925, + "grad_norm": 3.1116929004314224, + "learning_rate": 1.9948394394329494e-05, + "loss": 1.0466, + "step": 566 + }, + { + "epoch": 0.30781758957654726, + "grad_norm": 2.649244034262683, + "learning_rate": 1.9948036954385613e-05, + "loss": 1.2914, + "step": 567 + }, + { + "epoch": 0.30836047774158526, + "grad_norm": 3.9338023936000965, + "learning_rate": 1.9947678284052667e-05, + "loss": 1.7532, + "step": 568 + }, + { + "epoch": 0.30890336590662326, + "grad_norm": 2.887133065759567, + "learning_rate": 1.9947318383375017e-05, + "loss": 1.8001, + "step": 569 + }, + { + "epoch": 0.30944625407166126, + "grad_norm": 3.6996119456889915, + "learning_rate": 1.9946957252397173e-05, + "loss": 2.4852, + "step": 570 + }, + { + "epoch": 0.30998914223669927, + "grad_norm": 2.829855772438557, + "learning_rate": 1.9946594891163808e-05, + "loss": 1.6048, + "step": 571 + }, + { + "epoch": 0.31053203040173727, + "grad_norm": 4.288772356209683, + "learning_rate": 1.9946231299719732e-05, + "loss": 1.4841, + "step": 572 + }, + { + "epoch": 0.31107491856677527, + "grad_norm": 2.37958986026152, + "learning_rate": 1.9945866478109914e-05, + "loss": 1.1797, + "step": 573 + }, + { + "epoch": 0.3116178067318133, + "grad_norm": 2.9977970906442932, + "learning_rate": 1.9945500426379483e-05, + "loss": 1.224, + "step": 574 + }, + { + "epoch": 0.3121606948968513, + "grad_norm": 3.406596416686285, + "learning_rate": 1.9945133144573705e-05, + "loss": 1.4793, + "step": 575 + }, + { + "epoch": 0.3127035830618892, + "grad_norm": 3.3772239188722244, + "learning_rate": 1.994476463273801e-05, + "loss": 1.3696, + "step": 576 + }, + { + "epoch": 0.3132464712269272, + "grad_norm": 2.729299768057245, + "learning_rate": 1.9944394890917977e-05, + "loss": 1.448, + "step": 577 + }, + { + "epoch": 0.31378935939196523, + "grad_norm": 2.6526406591248297, + "learning_rate": 1.9944023919159335e-05, + "loss": 1.6905, + "step": 578 + }, + { + "epoch": 0.31433224755700323, + "grad_norm": 2.9512501195242944, + "learning_rate": 1.9943651717507965e-05, + "loss": 1.6277, + "step": 579 + }, + { + "epoch": 0.31487513572204123, + "grad_norm": 2.8292191327941723, + "learning_rate": 1.9943278286009903e-05, + "loss": 1.0532, + "step": 580 + }, + { + "epoch": 0.31541802388707924, + "grad_norm": 3.2105870294745436, + "learning_rate": 1.9942903624711335e-05, + "loss": 1.1823, + "step": 581 + }, + { + "epoch": 0.31596091205211724, + "grad_norm": 3.224137115744835, + "learning_rate": 1.9942527733658602e-05, + "loss": 1.5409, + "step": 582 + }, + { + "epoch": 0.31650380021715524, + "grad_norm": 3.126134123334164, + "learning_rate": 1.9942150612898194e-05, + "loss": 1.2423, + "step": 583 + }, + { + "epoch": 0.31704668838219324, + "grad_norm": 2.709954974374804, + "learning_rate": 1.994177226247675e-05, + "loss": 1.3419, + "step": 584 + }, + { + "epoch": 0.31758957654723124, + "grad_norm": 2.9030367366777927, + "learning_rate": 1.9941392682441066e-05, + "loss": 1.427, + "step": 585 + }, + { + "epoch": 0.31813246471226925, + "grad_norm": 3.272875981776567, + "learning_rate": 1.9941011872838092e-05, + "loss": 2.0196, + "step": 586 + }, + { + "epoch": 0.31867535287730725, + "grad_norm": 3.3480185179867, + "learning_rate": 1.994062983371493e-05, + "loss": 1.6038, + "step": 587 + }, + { + "epoch": 0.31921824104234525, + "grad_norm": 2.801738772846361, + "learning_rate": 1.9940246565118822e-05, + "loss": 1.7505, + "step": 588 + }, + { + "epoch": 0.31976112920738325, + "grad_norm": 2.743550967049156, + "learning_rate": 1.993986206709718e-05, + "loss": 1.3914, + "step": 589 + }, + { + "epoch": 0.32030401737242126, + "grad_norm": 2.481845489278486, + "learning_rate": 1.9939476339697555e-05, + "loss": 1.6927, + "step": 590 + }, + { + "epoch": 0.32084690553745926, + "grad_norm": 3.145019330878407, + "learning_rate": 1.993908938296765e-05, + "loss": 1.42, + "step": 591 + }, + { + "epoch": 0.32138979370249726, + "grad_norm": 3.326427662456394, + "learning_rate": 1.9938701196955335e-05, + "loss": 1.2516, + "step": 592 + }, + { + "epoch": 0.32193268186753526, + "grad_norm": 3.195815443331326, + "learning_rate": 1.9938311781708616e-05, + "loss": 2.1428, + "step": 593 + }, + { + "epoch": 0.32247557003257327, + "grad_norm": 2.461395361556941, + "learning_rate": 1.9937921137275657e-05, + "loss": 1.4448, + "step": 594 + }, + { + "epoch": 0.32301845819761127, + "grad_norm": 3.1793351214549794, + "learning_rate": 1.993752926370477e-05, + "loss": 1.4609, + "step": 595 + }, + { + "epoch": 0.32356134636264927, + "grad_norm": 2.783909288864463, + "learning_rate": 1.9937136161044427e-05, + "loss": 1.3355, + "step": 596 + }, + { + "epoch": 0.3241042345276873, + "grad_norm": 2.8156929696256734, + "learning_rate": 1.9936741829343247e-05, + "loss": 2.101, + "step": 597 + }, + { + "epoch": 0.3246471226927253, + "grad_norm": 3.793352093788154, + "learning_rate": 1.993634626865e-05, + "loss": 2.19, + "step": 598 + }, + { + "epoch": 0.3251900108577633, + "grad_norm": 2.6680863350639545, + "learning_rate": 1.993594947901361e-05, + "loss": 1.5199, + "step": 599 + }, + { + "epoch": 0.3257328990228013, + "grad_norm": 2.6498586813134297, + "learning_rate": 1.9935551460483155e-05, + "loss": 0.9282, + "step": 600 + }, + { + "epoch": 0.3262757871878393, + "grad_norm": 3.443063689484479, + "learning_rate": 1.993515221310786e-05, + "loss": 1.9107, + "step": 601 + }, + { + "epoch": 0.3268186753528773, + "grad_norm": 2.45686855619251, + "learning_rate": 1.9934751736937103e-05, + "loss": 1.2929, + "step": 602 + }, + { + "epoch": 0.3273615635179153, + "grad_norm": 3.3974155688994077, + "learning_rate": 1.9934350032020417e-05, + "loss": 1.546, + "step": 603 + }, + { + "epoch": 0.3279044516829533, + "grad_norm": 2.3692191456624783, + "learning_rate": 1.993394709840749e-05, + "loss": 1.2239, + "step": 604 + }, + { + "epoch": 0.3284473398479913, + "grad_norm": 4.044352925058249, + "learning_rate": 1.993354293614815e-05, + "loss": 1.801, + "step": 605 + }, + { + "epoch": 0.3289902280130293, + "grad_norm": 4.051192057617293, + "learning_rate": 1.993313754529239e-05, + "loss": 1.8043, + "step": 606 + }, + { + "epoch": 0.3295331161780673, + "grad_norm": 3.3680281079594634, + "learning_rate": 1.9932730925890344e-05, + "loss": 1.4915, + "step": 607 + }, + { + "epoch": 0.3300760043431053, + "grad_norm": 2.993677842102555, + "learning_rate": 1.9932323077992312e-05, + "loss": 1.3457, + "step": 608 + }, + { + "epoch": 0.3306188925081433, + "grad_norm": 3.325666927669253, + "learning_rate": 1.9931914001648726e-05, + "loss": 1.4221, + "step": 609 + }, + { + "epoch": 0.3311617806731813, + "grad_norm": 2.459382097232459, + "learning_rate": 1.993150369691019e-05, + "loss": 1.2551, + "step": 610 + }, + { + "epoch": 0.3317046688382193, + "grad_norm": 5.281638549560053, + "learning_rate": 1.993109216382745e-05, + "loss": 1.142, + "step": 611 + }, + { + "epoch": 0.3322475570032573, + "grad_norm": 2.6931033883715374, + "learning_rate": 1.99306794024514e-05, + "loss": 1.2573, + "step": 612 + }, + { + "epoch": 0.3327904451682953, + "grad_norm": 2.8066393604642714, + "learning_rate": 1.9930265412833097e-05, + "loss": 0.9847, + "step": 613 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 3.682172765988149, + "learning_rate": 1.992985019502374e-05, + "loss": 1.4482, + "step": 614 + }, + { + "epoch": 0.3338762214983713, + "grad_norm": 3.7190248655227873, + "learning_rate": 1.9929433749074684e-05, + "loss": 1.3682, + "step": 615 + }, + { + "epoch": 0.3344191096634093, + "grad_norm": 2.8594798198555527, + "learning_rate": 1.9929016075037438e-05, + "loss": 1.2396, + "step": 616 + }, + { + "epoch": 0.3349619978284473, + "grad_norm": 3.5439348644341035, + "learning_rate": 1.992859717296366e-05, + "loss": 1.5843, + "step": 617 + }, + { + "epoch": 0.3355048859934853, + "grad_norm": 2.3447575030783283, + "learning_rate": 1.992817704290516e-05, + "loss": 1.2025, + "step": 618 + }, + { + "epoch": 0.3360477741585233, + "grad_norm": 3.027039186876229, + "learning_rate": 1.99277556849139e-05, + "loss": 1.7133, + "step": 619 + }, + { + "epoch": 0.33659066232356133, + "grad_norm": 2.9868315770124343, + "learning_rate": 1.9927333099041992e-05, + "loss": 1.2309, + "step": 620 + }, + { + "epoch": 0.33713355048859933, + "grad_norm": 3.215889472723888, + "learning_rate": 1.9926909285341705e-05, + "loss": 1.2217, + "step": 621 + }, + { + "epoch": 0.33767643865363733, + "grad_norm": 3.293119071439456, + "learning_rate": 1.992648424386546e-05, + "loss": 1.7828, + "step": 622 + }, + { + "epoch": 0.33821932681867534, + "grad_norm": 2.7023629491691623, + "learning_rate": 1.992605797466582e-05, + "loss": 1.3745, + "step": 623 + }, + { + "epoch": 0.33876221498371334, + "grad_norm": 4.234534944070885, + "learning_rate": 1.9925630477795514e-05, + "loss": 1.3415, + "step": 624 + }, + { + "epoch": 0.33930510314875134, + "grad_norm": 3.6994706259321815, + "learning_rate": 1.9925201753307406e-05, + "loss": 1.4234, + "step": 625 + }, + { + "epoch": 0.33984799131378934, + "grad_norm": 3.156992028368943, + "learning_rate": 1.992477180125453e-05, + "loss": 1.4845, + "step": 626 + }, + { + "epoch": 0.34039087947882735, + "grad_norm": 3.1286393814611397, + "learning_rate": 1.9924340621690058e-05, + "loss": 0.9685, + "step": 627 + }, + { + "epoch": 0.34093376764386535, + "grad_norm": 3.336970766050665, + "learning_rate": 1.9923908214667323e-05, + "loss": 1.3268, + "step": 628 + }, + { + "epoch": 0.34147665580890335, + "grad_norm": 4.290496076646803, + "learning_rate": 1.99234745802398e-05, + "loss": 1.2785, + "step": 629 + }, + { + "epoch": 0.34201954397394135, + "grad_norm": 2.7491777959300023, + "learning_rate": 1.9923039718461127e-05, + "loss": 1.3798, + "step": 630 + }, + { + "epoch": 0.34256243213897936, + "grad_norm": 5.312529540414799, + "learning_rate": 1.992260362938509e-05, + "loss": 2.2512, + "step": 631 + }, + { + "epoch": 0.34310532030401736, + "grad_norm": 3.9557409230221676, + "learning_rate": 1.9922166313065618e-05, + "loss": 1.849, + "step": 632 + }, + { + "epoch": 0.34364820846905536, + "grad_norm": 3.7327574664165786, + "learning_rate": 1.9921727769556805e-05, + "loss": 1.4078, + "step": 633 + }, + { + "epoch": 0.34419109663409336, + "grad_norm": 3.7644055867642594, + "learning_rate": 1.992128799891289e-05, + "loss": 1.714, + "step": 634 + }, + { + "epoch": 0.34473398479913137, + "grad_norm": 3.923952524278675, + "learning_rate": 1.9920847001188258e-05, + "loss": 1.7613, + "step": 635 + }, + { + "epoch": 0.34527687296416937, + "grad_norm": 3.1523646374054155, + "learning_rate": 1.992040477643746e-05, + "loss": 1.2563, + "step": 636 + }, + { + "epoch": 0.34581976112920737, + "grad_norm": 3.6157151039083444, + "learning_rate": 1.991996132471519e-05, + "loss": 1.1795, + "step": 637 + }, + { + "epoch": 0.3463626492942454, + "grad_norm": 4.222007576307266, + "learning_rate": 1.991951664607629e-05, + "loss": 1.6605, + "step": 638 + }, + { + "epoch": 0.3469055374592834, + "grad_norm": 3.1555785318795544, + "learning_rate": 1.9919070740575764e-05, + "loss": 1.526, + "step": 639 + }, + { + "epoch": 0.3474484256243214, + "grad_norm": 3.2645269107818677, + "learning_rate": 1.991862360826876e-05, + "loss": 1.3182, + "step": 640 + }, + { + "epoch": 0.3479913137893594, + "grad_norm": 3.53044955128402, + "learning_rate": 1.991817524921058e-05, + "loss": 1.5583, + "step": 641 + }, + { + "epoch": 0.3485342019543974, + "grad_norm": 3.24710260787145, + "learning_rate": 1.9917725663456678e-05, + "loss": 0.9878, + "step": 642 + }, + { + "epoch": 0.3490770901194354, + "grad_norm": 3.9154742975791654, + "learning_rate": 1.991727485106266e-05, + "loss": 1.6034, + "step": 643 + }, + { + "epoch": 0.3496199782844734, + "grad_norm": 3.8199965805178318, + "learning_rate": 1.9916822812084282e-05, + "loss": 1.3768, + "step": 644 + }, + { + "epoch": 0.3501628664495114, + "grad_norm": 3.198793670499035, + "learning_rate": 1.9916369546577455e-05, + "loss": 1.4308, + "step": 645 + }, + { + "epoch": 0.3507057546145494, + "grad_norm": 2.908177540585374, + "learning_rate": 1.9915915054598237e-05, + "loss": 1.3964, + "step": 646 + }, + { + "epoch": 0.3512486427795874, + "grad_norm": 2.920889003436479, + "learning_rate": 1.9915459336202844e-05, + "loss": 1.056, + "step": 647 + }, + { + "epoch": 0.3517915309446254, + "grad_norm": 3.7046400158340864, + "learning_rate": 1.991500239144763e-05, + "loss": 1.9052, + "step": 648 + }, + { + "epoch": 0.3523344191096634, + "grad_norm": 4.412988121206581, + "learning_rate": 1.9914544220389124e-05, + "loss": 1.944, + "step": 649 + }, + { + "epoch": 0.3528773072747014, + "grad_norm": 3.87124697251994, + "learning_rate": 1.9914084823083988e-05, + "loss": 1.4951, + "step": 650 + }, + { + "epoch": 0.3534201954397394, + "grad_norm": 3.8734663848108584, + "learning_rate": 1.9913624199589037e-05, + "loss": 1.5462, + "step": 651 + }, + { + "epoch": 0.3539630836047774, + "grad_norm": 2.7283359997443126, + "learning_rate": 1.9913162349961248e-05, + "loss": 0.9188, + "step": 652 + }, + { + "epoch": 0.3545059717698154, + "grad_norm": 2.885333878853398, + "learning_rate": 1.991269927425774e-05, + "loss": 1.2292, + "step": 653 + }, + { + "epoch": 0.3550488599348534, + "grad_norm": 3.2425772460053257, + "learning_rate": 1.9912234972535788e-05, + "loss": 1.2863, + "step": 654 + }, + { + "epoch": 0.3555917480998914, + "grad_norm": 3.060966260943164, + "learning_rate": 1.991176944485281e-05, + "loss": 1.5056, + "step": 655 + }, + { + "epoch": 0.3561346362649294, + "grad_norm": 4.533553635387312, + "learning_rate": 1.99113026912664e-05, + "loss": 1.8229, + "step": 656 + }, + { + "epoch": 0.3566775244299674, + "grad_norm": 2.8582150809054045, + "learning_rate": 1.9910834711834267e-05, + "loss": 1.3233, + "step": 657 + }, + { + "epoch": 0.3572204125950054, + "grad_norm": 4.930680768753159, + "learning_rate": 1.9910365506614308e-05, + "loss": 1.5997, + "step": 658 + }, + { + "epoch": 0.3577633007600434, + "grad_norm": 3.0733335483268083, + "learning_rate": 1.9909895075664545e-05, + "loss": 0.9206, + "step": 659 + }, + { + "epoch": 0.3583061889250814, + "grad_norm": 3.4659704922712686, + "learning_rate": 1.990942341904317e-05, + "loss": 1.2752, + "step": 660 + }, + { + "epoch": 0.35884907709011943, + "grad_norm": 3.493850558129149, + "learning_rate": 1.9908950536808508e-05, + "loss": 1.588, + "step": 661 + }, + { + "epoch": 0.35939196525515743, + "grad_norm": 3.61256229272583, + "learning_rate": 1.9908476429019056e-05, + "loss": 1.469, + "step": 662 + }, + { + "epoch": 0.35993485342019543, + "grad_norm": 3.3430037514089803, + "learning_rate": 1.9908001095733445e-05, + "loss": 1.3038, + "step": 663 + }, + { + "epoch": 0.36047774158523344, + "grad_norm": 3.8918915484374557, + "learning_rate": 1.9907524537010467e-05, + "loss": 1.3683, + "step": 664 + }, + { + "epoch": 0.36102062975027144, + "grad_norm": 3.6032272045245053, + "learning_rate": 1.9907046752909064e-05, + "loss": 1.8694, + "step": 665 + }, + { + "epoch": 0.36156351791530944, + "grad_norm": 4.546268377661869, + "learning_rate": 1.9906567743488326e-05, + "loss": 1.2871, + "step": 666 + }, + { + "epoch": 0.36210640608034744, + "grad_norm": 3.5283169402860777, + "learning_rate": 1.9906087508807504e-05, + "loss": 1.5334, + "step": 667 + }, + { + "epoch": 0.36264929424538545, + "grad_norm": 2.7984191262784504, + "learning_rate": 1.9905606048925993e-05, + "loss": 1.1924, + "step": 668 + }, + { + "epoch": 0.36319218241042345, + "grad_norm": 3.54703835950394, + "learning_rate": 1.9905123363903335e-05, + "loss": 1.5972, + "step": 669 + }, + { + "epoch": 0.36373507057546145, + "grad_norm": 3.0126938064772855, + "learning_rate": 1.9904639453799236e-05, + "loss": 1.2294, + "step": 670 + }, + { + "epoch": 0.36427795874049945, + "grad_norm": 3.16358553989926, + "learning_rate": 1.990415431867354e-05, + "loss": 1.3564, + "step": 671 + }, + { + "epoch": 0.36482084690553745, + "grad_norm": 4.118910539945259, + "learning_rate": 1.990366795858626e-05, + "loss": 1.6838, + "step": 672 + }, + { + "epoch": 0.36536373507057546, + "grad_norm": 3.8466533902335502, + "learning_rate": 1.9903180373597534e-05, + "loss": 1.7986, + "step": 673 + }, + { + "epoch": 0.36590662323561346, + "grad_norm": 3.6195384409794684, + "learning_rate": 1.990269156376768e-05, + "loss": 1.6113, + "step": 674 + }, + { + "epoch": 0.36644951140065146, + "grad_norm": 3.509724379422402, + "learning_rate": 1.9902201529157152e-05, + "loss": 1.5496, + "step": 675 + }, + { + "epoch": 0.36699239956568946, + "grad_norm": 3.579737463592409, + "learning_rate": 1.9901710269826554e-05, + "loss": 1.4856, + "step": 676 + }, + { + "epoch": 0.36753528773072747, + "grad_norm": 3.780431786449365, + "learning_rate": 1.9901217785836655e-05, + "loss": 1.8519, + "step": 677 + }, + { + "epoch": 0.36807817589576547, + "grad_norm": 4.0659352627131735, + "learning_rate": 1.9900724077248354e-05, + "loss": 1.4382, + "step": 678 + }, + { + "epoch": 0.36862106406080347, + "grad_norm": 2.9524656376446274, + "learning_rate": 1.9900229144122723e-05, + "loss": 1.4541, + "step": 679 + }, + { + "epoch": 0.3691639522258415, + "grad_norm": 2.388269642575342, + "learning_rate": 1.989973298652097e-05, + "loss": 0.9567, + "step": 680 + }, + { + "epoch": 0.3697068403908795, + "grad_norm": 2.8322419431529453, + "learning_rate": 1.9899235604504467e-05, + "loss": 1.1187, + "step": 681 + }, + { + "epoch": 0.3702497285559175, + "grad_norm": 3.0469435898757613, + "learning_rate": 1.9898736998134726e-05, + "loss": 1.5571, + "step": 682 + }, + { + "epoch": 0.3707926167209555, + "grad_norm": 3.623982705749655, + "learning_rate": 1.9898237167473416e-05, + "loss": 1.2047, + "step": 683 + }, + { + "epoch": 0.3713355048859935, + "grad_norm": 3.669362267695381, + "learning_rate": 1.9897736112582357e-05, + "loss": 1.7747, + "step": 684 + }, + { + "epoch": 0.3718783930510315, + "grad_norm": 2.9109325710576353, + "learning_rate": 1.989723383352352e-05, + "loss": 1.0044, + "step": 685 + }, + { + "epoch": 0.3724212812160695, + "grad_norm": 3.4478434843760146, + "learning_rate": 1.9896730330359032e-05, + "loss": 1.1245, + "step": 686 + }, + { + "epoch": 0.3729641693811075, + "grad_norm": 3.2957514959337275, + "learning_rate": 1.989622560315116e-05, + "loss": 1.3243, + "step": 687 + }, + { + "epoch": 0.3735070575461455, + "grad_norm": 3.5500423086355988, + "learning_rate": 1.989571965196234e-05, + "loss": 2.0478, + "step": 688 + }, + { + "epoch": 0.3740499457111835, + "grad_norm": 2.621987663839103, + "learning_rate": 1.9895212476855136e-05, + "loss": 1.3135, + "step": 689 + }, + { + "epoch": 0.3745928338762215, + "grad_norm": 4.055390555970487, + "learning_rate": 1.989470407789228e-05, + "loss": 1.5165, + "step": 690 + }, + { + "epoch": 0.3751357220412595, + "grad_norm": 2.94204110816024, + "learning_rate": 1.989419445513666e-05, + "loss": 1.3458, + "step": 691 + }, + { + "epoch": 0.3756786102062975, + "grad_norm": 3.5714132672245977, + "learning_rate": 1.98936836086513e-05, + "loss": 1.3034, + "step": 692 + }, + { + "epoch": 0.3762214983713355, + "grad_norm": 3.659137782783938, + "learning_rate": 1.9893171538499382e-05, + "loss": 1.6203, + "step": 693 + }, + { + "epoch": 0.3767643865363735, + "grad_norm": 5.116097300755018, + "learning_rate": 1.9892658244744236e-05, + "loss": 2.1071, + "step": 694 + }, + { + "epoch": 0.3773072747014115, + "grad_norm": 2.8449662018307005, + "learning_rate": 1.9892143727449357e-05, + "loss": 1.0477, + "step": 695 + }, + { + "epoch": 0.3778501628664495, + "grad_norm": 4.021030963638319, + "learning_rate": 1.989162798667838e-05, + "loss": 1.9528, + "step": 696 + }, + { + "epoch": 0.3783930510314875, + "grad_norm": 3.377368666687089, + "learning_rate": 1.989111102249508e-05, + "loss": 1.2481, + "step": 697 + }, + { + "epoch": 0.3789359391965255, + "grad_norm": 3.403268816169458, + "learning_rate": 1.9890592834963406e-05, + "loss": 1.1864, + "step": 698 + }, + { + "epoch": 0.3794788273615635, + "grad_norm": 3.082879601892987, + "learning_rate": 1.9890073424147453e-05, + "loss": 1.7365, + "step": 699 + }, + { + "epoch": 0.3800217155266015, + "grad_norm": 4.122610250215441, + "learning_rate": 1.988955279011145e-05, + "loss": 1.4139, + "step": 700 + }, + { + "epoch": 0.3805646036916395, + "grad_norm": 3.298087120576001, + "learning_rate": 1.98890309329198e-05, + "loss": 1.5904, + "step": 701 + }, + { + "epoch": 0.3811074918566775, + "grad_norm": 3.1183850173888703, + "learning_rate": 1.9888507852637043e-05, + "loss": 1.2146, + "step": 702 + }, + { + "epoch": 0.38165038002171553, + "grad_norm": 3.128042971411966, + "learning_rate": 1.9887983549327873e-05, + "loss": 1.2751, + "step": 703 + }, + { + "epoch": 0.38219326818675353, + "grad_norm": 3.362555156234201, + "learning_rate": 1.988745802305714e-05, + "loss": 1.1363, + "step": 704 + }, + { + "epoch": 0.38273615635179153, + "grad_norm": 3.29574344438116, + "learning_rate": 1.988693127388984e-05, + "loss": 1.227, + "step": 705 + }, + { + "epoch": 0.38327904451682954, + "grad_norm": 4.2418853227437205, + "learning_rate": 1.9886403301891123e-05, + "loss": 1.7091, + "step": 706 + }, + { + "epoch": 0.38382193268186754, + "grad_norm": 4.165071133964158, + "learning_rate": 1.9885874107126287e-05, + "loss": 1.9403, + "step": 707 + }, + { + "epoch": 0.38436482084690554, + "grad_norm": 4.639609386348944, + "learning_rate": 1.9885343689660787e-05, + "loss": 1.221, + "step": 708 + }, + { + "epoch": 0.38490770901194354, + "grad_norm": 3.0746775894794034, + "learning_rate": 1.9884812049560226e-05, + "loss": 1.4958, + "step": 709 + }, + { + "epoch": 0.38545059717698155, + "grad_norm": 4.822804859325608, + "learning_rate": 1.9884279186890357e-05, + "loss": 1.6928, + "step": 710 + }, + { + "epoch": 0.38599348534201955, + "grad_norm": 4.299811491105524, + "learning_rate": 1.9883745101717084e-05, + "loss": 1.3075, + "step": 711 + }, + { + "epoch": 0.38653637350705755, + "grad_norm": 3.882170994492157, + "learning_rate": 1.9883209794106464e-05, + "loss": 1.9712, + "step": 712 + }, + { + "epoch": 0.38707926167209555, + "grad_norm": 4.164019029255246, + "learning_rate": 1.9882673264124705e-05, + "loss": 1.6109, + "step": 713 + }, + { + "epoch": 0.38762214983713356, + "grad_norm": 3.691226317297577, + "learning_rate": 1.9882135511838167e-05, + "loss": 1.3225, + "step": 714 + }, + { + "epoch": 0.38816503800217156, + "grad_norm": 5.777282505586183, + "learning_rate": 1.988159653731336e-05, + "loss": 1.7945, + "step": 715 + }, + { + "epoch": 0.38870792616720956, + "grad_norm": 3.3926512133724165, + "learning_rate": 1.9881056340616944e-05, + "loss": 1.5797, + "step": 716 + }, + { + "epoch": 0.38925081433224756, + "grad_norm": 3.7756965718668467, + "learning_rate": 1.988051492181573e-05, + "loss": 1.5515, + "step": 717 + }, + { + "epoch": 0.38979370249728557, + "grad_norm": 3.5509658648264613, + "learning_rate": 1.987997228097668e-05, + "loss": 1.2004, + "step": 718 + }, + { + "epoch": 0.39033659066232357, + "grad_norm": 4.682493032945398, + "learning_rate": 1.987942841816692e-05, + "loss": 1.2957, + "step": 719 + }, + { + "epoch": 0.39087947882736157, + "grad_norm": 2.923319971884201, + "learning_rate": 1.9878883333453704e-05, + "loss": 0.753, + "step": 720 + }, + { + "epoch": 0.3914223669923996, + "grad_norm": 2.8706465146507845, + "learning_rate": 1.987833702690445e-05, + "loss": 1.1057, + "step": 721 + }, + { + "epoch": 0.3919652551574376, + "grad_norm": 3.4516078574834057, + "learning_rate": 1.987778949858673e-05, + "loss": 1.1275, + "step": 722 + }, + { + "epoch": 0.3925081433224756, + "grad_norm": 3.3400553957950567, + "learning_rate": 1.9877240748568263e-05, + "loss": 1.1538, + "step": 723 + }, + { + "epoch": 0.3930510314875136, + "grad_norm": 3.155540544638446, + "learning_rate": 1.987669077691692e-05, + "loss": 1.2486, + "step": 724 + }, + { + "epoch": 0.3935939196525516, + "grad_norm": 3.1450276486107054, + "learning_rate": 1.987613958370072e-05, + "loss": 1.265, + "step": 725 + }, + { + "epoch": 0.3941368078175896, + "grad_norm": 3.0493327058864406, + "learning_rate": 1.9875587168987834e-05, + "loss": 1.2097, + "step": 726 + }, + { + "epoch": 0.3946796959826276, + "grad_norm": 3.716259137065701, + "learning_rate": 1.987503353284659e-05, + "loss": 1.5386, + "step": 727 + }, + { + "epoch": 0.3952225841476656, + "grad_norm": 3.441725300410296, + "learning_rate": 1.9874478675345458e-05, + "loss": 1.3936, + "step": 728 + }, + { + "epoch": 0.3957654723127036, + "grad_norm": 3.773353353497139, + "learning_rate": 1.9873922596553067e-05, + "loss": 1.731, + "step": 729 + }, + { + "epoch": 0.3963083604777416, + "grad_norm": 3.527645597828566, + "learning_rate": 1.987336529653819e-05, + "loss": 1.708, + "step": 730 + }, + { + "epoch": 0.3968512486427796, + "grad_norm": 3.8440953147603643, + "learning_rate": 1.9872806775369762e-05, + "loss": 1.6102, + "step": 731 + }, + { + "epoch": 0.3973941368078176, + "grad_norm": 3.4036622029265966, + "learning_rate": 1.9872247033116855e-05, + "loss": 0.9298, + "step": 732 + }, + { + "epoch": 0.3979370249728556, + "grad_norm": 4.034027632823603, + "learning_rate": 1.98716860698487e-05, + "loss": 1.78, + "step": 733 + }, + { + "epoch": 0.3984799131378936, + "grad_norm": 4.10524690369694, + "learning_rate": 1.987112388563468e-05, + "loss": 1.2354, + "step": 734 + }, + { + "epoch": 0.3990228013029316, + "grad_norm": 4.059289261823474, + "learning_rate": 1.9870560480544325e-05, + "loss": 1.804, + "step": 735 + }, + { + "epoch": 0.3995656894679696, + "grad_norm": 3.2544839104292667, + "learning_rate": 1.986999585464732e-05, + "loss": 1.1228, + "step": 736 + }, + { + "epoch": 0.4001085776330076, + "grad_norm": 3.741117309135591, + "learning_rate": 1.9869430008013496e-05, + "loss": 1.329, + "step": 737 + }, + { + "epoch": 0.4006514657980456, + "grad_norm": 4.446330579981585, + "learning_rate": 1.9868862940712838e-05, + "loss": 1.6506, + "step": 738 + }, + { + "epoch": 0.4011943539630836, + "grad_norm": 3.139296317304318, + "learning_rate": 1.9868294652815483e-05, + "loss": 1.144, + "step": 739 + }, + { + "epoch": 0.4017372421281216, + "grad_norm": 4.025270331554587, + "learning_rate": 1.986772514439172e-05, + "loss": 1.336, + "step": 740 + }, + { + "epoch": 0.4022801302931596, + "grad_norm": 3.4545518889756557, + "learning_rate": 1.986715441551198e-05, + "loss": 1.6006, + "step": 741 + }, + { + "epoch": 0.4028230184581976, + "grad_norm": 3.748598610621174, + "learning_rate": 1.986658246624686e-05, + "loss": 1.4812, + "step": 742 + }, + { + "epoch": 0.4033659066232356, + "grad_norm": 5.126058311631635, + "learning_rate": 1.9866009296667093e-05, + "loss": 2.0129, + "step": 743 + }, + { + "epoch": 0.40390879478827363, + "grad_norm": 8.145692299678405, + "learning_rate": 1.9865434906843574e-05, + "loss": 1.8321, + "step": 744 + }, + { + "epoch": 0.40445168295331163, + "grad_norm": 4.23918362592579, + "learning_rate": 1.9864859296847343e-05, + "loss": 1.5765, + "step": 745 + }, + { + "epoch": 0.40499457111834963, + "grad_norm": 4.85145039296263, + "learning_rate": 1.986428246674959e-05, + "loss": 1.6486, + "step": 746 + }, + { + "epoch": 0.40553745928338764, + "grad_norm": 3.7681072947070415, + "learning_rate": 1.986370441662166e-05, + "loss": 1.4466, + "step": 747 + }, + { + "epoch": 0.40608034744842564, + "grad_norm": 4.206136090938125, + "learning_rate": 1.986312514653505e-05, + "loss": 1.6717, + "step": 748 + }, + { + "epoch": 0.40662323561346364, + "grad_norm": 3.7336223866799965, + "learning_rate": 1.9862544656561403e-05, + "loss": 0.9599, + "step": 749 + }, + { + "epoch": 0.40716612377850164, + "grad_norm": 4.037759008136602, + "learning_rate": 1.986196294677251e-05, + "loss": 1.9387, + "step": 750 + }, + { + "epoch": 0.40770901194353965, + "grad_norm": 3.8832525632606156, + "learning_rate": 1.9861380017240324e-05, + "loss": 2.0184, + "step": 751 + }, + { + "epoch": 0.40825190010857765, + "grad_norm": 3.9521376069923546, + "learning_rate": 1.986079586803694e-05, + "loss": 1.0926, + "step": 752 + }, + { + "epoch": 0.40879478827361565, + "grad_norm": 5.128973903987411, + "learning_rate": 1.986021049923461e-05, + "loss": 1.5943, + "step": 753 + }, + { + "epoch": 0.40933767643865365, + "grad_norm": 4.5372892013581865, + "learning_rate": 1.9859623910905728e-05, + "loss": 1.643, + "step": 754 + }, + { + "epoch": 0.40988056460369166, + "grad_norm": 4.2137978897196415, + "learning_rate": 1.985903610312285e-05, + "loss": 1.7237, + "step": 755 + }, + { + "epoch": 0.41042345276872966, + "grad_norm": 2.713265124864733, + "learning_rate": 1.985844707595867e-05, + "loss": 1.0521, + "step": 756 + }, + { + "epoch": 0.41096634093376766, + "grad_norm": 3.362872283003281, + "learning_rate": 1.9857856829486045e-05, + "loss": 1.2879, + "step": 757 + }, + { + "epoch": 0.41150922909880566, + "grad_norm": 4.022235417044952, + "learning_rate": 1.9857265363777975e-05, + "loss": 1.398, + "step": 758 + }, + { + "epoch": 0.41205211726384366, + "grad_norm": 4.083280210123792, + "learning_rate": 1.9856672678907616e-05, + "loss": 1.6652, + "step": 759 + }, + { + "epoch": 0.41259500542888167, + "grad_norm": 4.469771898345589, + "learning_rate": 1.985607877494827e-05, + "loss": 1.6242, + "step": 760 + }, + { + "epoch": 0.41313789359391967, + "grad_norm": 3.0490808076449016, + "learning_rate": 1.9855483651973396e-05, + "loss": 1.0783, + "step": 761 + }, + { + "epoch": 0.41368078175895767, + "grad_norm": 3.488230005104396, + "learning_rate": 1.9854887310056593e-05, + "loss": 1.3953, + "step": 762 + }, + { + "epoch": 0.4142236699239957, + "grad_norm": 3.5610333355465653, + "learning_rate": 1.9854289749271624e-05, + "loss": 1.3692, + "step": 763 + }, + { + "epoch": 0.4147665580890337, + "grad_norm": 3.5572639096784533, + "learning_rate": 1.9853690969692393e-05, + "loss": 1.646, + "step": 764 + }, + { + "epoch": 0.4153094462540717, + "grad_norm": 3.811741925703646, + "learning_rate": 1.9853090971392953e-05, + "loss": 1.6637, + "step": 765 + }, + { + "epoch": 0.4158523344191097, + "grad_norm": 2.912609699248709, + "learning_rate": 1.9852489754447526e-05, + "loss": 0.9086, + "step": 766 + }, + { + "epoch": 0.4163952225841477, + "grad_norm": 3.7608762982374646, + "learning_rate": 1.985188731893046e-05, + "loss": 1.5131, + "step": 767 + }, + { + "epoch": 0.4169381107491857, + "grad_norm": 3.776000457813264, + "learning_rate": 1.985128366491627e-05, + "loss": 1.5929, + "step": 768 + }, + { + "epoch": 0.4174809989142237, + "grad_norm": 4.129873624081247, + "learning_rate": 1.9850678792479613e-05, + "loss": 1.5461, + "step": 769 + }, + { + "epoch": 0.4180238870792617, + "grad_norm": 3.2233425858963645, + "learning_rate": 1.9850072701695306e-05, + "loss": 0.9614, + "step": 770 + }, + { + "epoch": 0.4185667752442997, + "grad_norm": 3.8215876727885054, + "learning_rate": 1.984946539263831e-05, + "loss": 1.2549, + "step": 771 + }, + { + "epoch": 0.4191096634093377, + "grad_norm": 3.0062072101032125, + "learning_rate": 1.9848856865383732e-05, + "loss": 0.9796, + "step": 772 + }, + { + "epoch": 0.4196525515743757, + "grad_norm": 2.937513473620578, + "learning_rate": 1.984824712000684e-05, + "loss": 1.0204, + "step": 773 + }, + { + "epoch": 0.4201954397394137, + "grad_norm": 2.949279428268976, + "learning_rate": 1.984763615658305e-05, + "loss": 1.332, + "step": 774 + }, + { + "epoch": 0.4207383279044517, + "grad_norm": 3.455520053658832, + "learning_rate": 1.9847023975187925e-05, + "loss": 1.3786, + "step": 775 + }, + { + "epoch": 0.4212812160694897, + "grad_norm": 3.8352233275650733, + "learning_rate": 1.9846410575897183e-05, + "loss": 1.4897, + "step": 776 + }, + { + "epoch": 0.4218241042345277, + "grad_norm": 3.262791090623995, + "learning_rate": 1.984579595878669e-05, + "loss": 1.5097, + "step": 777 + }, + { + "epoch": 0.4223669923995657, + "grad_norm": 3.603809745585289, + "learning_rate": 1.9845180123932456e-05, + "loss": 1.8419, + "step": 778 + }, + { + "epoch": 0.4229098805646037, + "grad_norm": 3.9648489260279196, + "learning_rate": 1.9844563071410656e-05, + "loss": 1.5159, + "step": 779 + }, + { + "epoch": 0.4234527687296417, + "grad_norm": 4.576987136894564, + "learning_rate": 1.9843944801297605e-05, + "loss": 1.6866, + "step": 780 + }, + { + "epoch": 0.4239956568946797, + "grad_norm": 3.4164331578089406, + "learning_rate": 1.9843325313669774e-05, + "loss": 1.5533, + "step": 781 + }, + { + "epoch": 0.4245385450597177, + "grad_norm": 3.2121441901059367, + "learning_rate": 1.9842704608603774e-05, + "loss": 0.8879, + "step": 782 + }, + { + "epoch": 0.4250814332247557, + "grad_norm": 3.6213877563860537, + "learning_rate": 1.9842082686176388e-05, + "loss": 1.3098, + "step": 783 + }, + { + "epoch": 0.4256243213897937, + "grad_norm": 3.719993086277978, + "learning_rate": 1.9841459546464527e-05, + "loss": 1.331, + "step": 784 + }, + { + "epoch": 0.4261672095548317, + "grad_norm": 2.7952357248876516, + "learning_rate": 1.9840835189545266e-05, + "loss": 1.1298, + "step": 785 + }, + { + "epoch": 0.42671009771986973, + "grad_norm": 3.450597586563723, + "learning_rate": 1.9840209615495822e-05, + "loss": 1.0023, + "step": 786 + }, + { + "epoch": 0.42725298588490773, + "grad_norm": 4.338480682869939, + "learning_rate": 1.983958282439357e-05, + "loss": 1.7044, + "step": 787 + }, + { + "epoch": 0.42779587404994573, + "grad_norm": 3.4831324579282765, + "learning_rate": 1.983895481631603e-05, + "loss": 1.4948, + "step": 788 + }, + { + "epoch": 0.42833876221498374, + "grad_norm": 6.382810905837672, + "learning_rate": 1.9838325591340885e-05, + "loss": 2.3008, + "step": 789 + }, + { + "epoch": 0.42888165038002174, + "grad_norm": 3.063848408601439, + "learning_rate": 1.9837695149545945e-05, + "loss": 1.2045, + "step": 790 + }, + { + "epoch": 0.42942453854505974, + "grad_norm": 4.80361465821882, + "learning_rate": 1.9837063491009193e-05, + "loss": 1.2552, + "step": 791 + }, + { + "epoch": 0.42996742671009774, + "grad_norm": 4.16884813133452, + "learning_rate": 1.9836430615808745e-05, + "loss": 1.6618, + "step": 792 + }, + { + "epoch": 0.43051031487513575, + "grad_norm": 4.215747683493222, + "learning_rate": 1.9835796524022886e-05, + "loss": 1.2571, + "step": 793 + }, + { + "epoch": 0.43105320304017375, + "grad_norm": 4.894853560761769, + "learning_rate": 1.9835161215730038e-05, + "loss": 0.793, + "step": 794 + }, + { + "epoch": 0.43159609120521175, + "grad_norm": 3.264854272979086, + "learning_rate": 1.983452469100877e-05, + "loss": 1.0721, + "step": 795 + }, + { + "epoch": 0.43213897937024975, + "grad_norm": 3.2346237132359335, + "learning_rate": 1.9833886949937823e-05, + "loss": 1.2396, + "step": 796 + }, + { + "epoch": 0.43268186753528776, + "grad_norm": 3.7265266370894023, + "learning_rate": 1.9833247992596058e-05, + "loss": 1.8845, + "step": 797 + }, + { + "epoch": 0.43322475570032576, + "grad_norm": 3.0655010730082126, + "learning_rate": 1.9832607819062513e-05, + "loss": 1.556, + "step": 798 + }, + { + "epoch": 0.43376764386536376, + "grad_norm": 3.209417077142042, + "learning_rate": 1.983196642941636e-05, + "loss": 1.2183, + "step": 799 + }, + { + "epoch": 0.43431053203040176, + "grad_norm": 4.384697511143271, + "learning_rate": 1.9831323823736933e-05, + "loss": 2.0889, + "step": 800 + }, + { + "epoch": 0.43485342019543977, + "grad_norm": 3.2773759507081084, + "learning_rate": 1.9830680002103703e-05, + "loss": 0.919, + "step": 801 + }, + { + "epoch": 0.43539630836047777, + "grad_norm": 4.1436406822638805, + "learning_rate": 1.9830034964596304e-05, + "loss": 1.7759, + "step": 802 + }, + { + "epoch": 0.43593919652551577, + "grad_norm": 5.715436881791243, + "learning_rate": 1.9829388711294512e-05, + "loss": 1.4277, + "step": 803 + }, + { + "epoch": 0.4364820846905538, + "grad_norm": 3.166985124043999, + "learning_rate": 1.982874124227826e-05, + "loss": 1.56, + "step": 804 + }, + { + "epoch": 0.4370249728555918, + "grad_norm": 4.237249076660815, + "learning_rate": 1.9828092557627626e-05, + "loss": 1.2366, + "step": 805 + }, + { + "epoch": 0.4375678610206297, + "grad_norm": 3.073398749866342, + "learning_rate": 1.982744265742284e-05, + "loss": 1.1475, + "step": 806 + }, + { + "epoch": 0.4381107491856677, + "grad_norm": 3.356777818996544, + "learning_rate": 1.9826791541744285e-05, + "loss": 1.2666, + "step": 807 + }, + { + "epoch": 0.4386536373507057, + "grad_norm": 3.171383453112648, + "learning_rate": 1.982613921067249e-05, + "loss": 1.1407, + "step": 808 + }, + { + "epoch": 0.43919652551574373, + "grad_norm": 3.334493176451811, + "learning_rate": 1.9825485664288138e-05, + "loss": 1.3309, + "step": 809 + }, + { + "epoch": 0.43973941368078173, + "grad_norm": 3.199045603289457, + "learning_rate": 1.982483090267206e-05, + "loss": 1.1634, + "step": 810 + }, + { + "epoch": 0.44028230184581973, + "grad_norm": 2.9970299057571856, + "learning_rate": 1.9824174925905235e-05, + "loss": 0.9642, + "step": 811 + }, + { + "epoch": 0.44082519001085774, + "grad_norm": 3.9851247078132994, + "learning_rate": 1.98235177340688e-05, + "loss": 1.4926, + "step": 812 + }, + { + "epoch": 0.44136807817589574, + "grad_norm": 3.586604601363014, + "learning_rate": 1.9822859327244034e-05, + "loss": 1.3105, + "step": 813 + }, + { + "epoch": 0.44191096634093374, + "grad_norm": 4.198634753759357, + "learning_rate": 1.9822199705512372e-05, + "loss": 1.4437, + "step": 814 + }, + { + "epoch": 0.44245385450597174, + "grad_norm": 3.6499375531563096, + "learning_rate": 1.9821538868955394e-05, + "loss": 1.2063, + "step": 815 + }, + { + "epoch": 0.44299674267100975, + "grad_norm": 4.3973905860052716, + "learning_rate": 1.9820876817654836e-05, + "loss": 1.5002, + "step": 816 + }, + { + "epoch": 0.44353963083604775, + "grad_norm": 2.7871960857121287, + "learning_rate": 1.9820213551692585e-05, + "loss": 0.6589, + "step": 817 + }, + { + "epoch": 0.44408251900108575, + "grad_norm": 3.4126665543063184, + "learning_rate": 1.981954907115067e-05, + "loss": 1.1651, + "step": 818 + }, + { + "epoch": 0.44462540716612375, + "grad_norm": 4.229651941034759, + "learning_rate": 1.981888337611127e-05, + "loss": 1.6444, + "step": 819 + }, + { + "epoch": 0.44516829533116176, + "grad_norm": 3.4135993440364056, + "learning_rate": 1.981821646665673e-05, + "loss": 1.2614, + "step": 820 + }, + { + "epoch": 0.44571118349619976, + "grad_norm": 4.274078711060124, + "learning_rate": 1.9817548342869527e-05, + "loss": 1.66, + "step": 821 + }, + { + "epoch": 0.44625407166123776, + "grad_norm": 2.9406863812447903, + "learning_rate": 1.98168790048323e-05, + "loss": 1.0412, + "step": 822 + }, + { + "epoch": 0.44679695982627576, + "grad_norm": 4.300279758003301, + "learning_rate": 1.981620845262783e-05, + "loss": 1.3689, + "step": 823 + }, + { + "epoch": 0.44733984799131377, + "grad_norm": 3.130525530503507, + "learning_rate": 1.9815536686339056e-05, + "loss": 1.1771, + "step": 824 + }, + { + "epoch": 0.44788273615635177, + "grad_norm": 3.6606030046920597, + "learning_rate": 1.981486370604906e-05, + "loss": 1.0323, + "step": 825 + }, + { + "epoch": 0.44842562432138977, + "grad_norm": 3.495865966877518, + "learning_rate": 1.981418951184108e-05, + "loss": 1.4493, + "step": 826 + }, + { + "epoch": 0.4489685124864278, + "grad_norm": 3.621229661878418, + "learning_rate": 1.9813514103798498e-05, + "loss": 1.5392, + "step": 827 + }, + { + "epoch": 0.4495114006514658, + "grad_norm": 3.6798995423063485, + "learning_rate": 1.9812837482004853e-05, + "loss": 1.0593, + "step": 828 + }, + { + "epoch": 0.4500542888165038, + "grad_norm": 3.4514040085362954, + "learning_rate": 1.9812159646543824e-05, + "loss": 0.9262, + "step": 829 + }, + { + "epoch": 0.4505971769815418, + "grad_norm": 3.195454430712678, + "learning_rate": 1.9811480597499257e-05, + "loss": 1.1469, + "step": 830 + }, + { + "epoch": 0.4511400651465798, + "grad_norm": 4.590874916307442, + "learning_rate": 1.981080033495513e-05, + "loss": 1.6534, + "step": 831 + }, + { + "epoch": 0.4516829533116178, + "grad_norm": 3.668702667937113, + "learning_rate": 1.981011885899558e-05, + "loss": 1.0597, + "step": 832 + }, + { + "epoch": 0.4522258414766558, + "grad_norm": 2.7669639542025046, + "learning_rate": 1.98094361697049e-05, + "loss": 1.0387, + "step": 833 + }, + { + "epoch": 0.4527687296416938, + "grad_norm": 4.734510249253564, + "learning_rate": 1.9808752267167515e-05, + "loss": 1.6757, + "step": 834 + }, + { + "epoch": 0.4533116178067318, + "grad_norm": 9.006078348758516, + "learning_rate": 1.9808067151468018e-05, + "loss": 2.1559, + "step": 835 + }, + { + "epoch": 0.4538545059717698, + "grad_norm": 4.169764802563779, + "learning_rate": 1.9807380822691146e-05, + "loss": 1.2557, + "step": 836 + }, + { + "epoch": 0.4543973941368078, + "grad_norm": 3.970541817152741, + "learning_rate": 1.9806693280921783e-05, + "loss": 1.5586, + "step": 837 + }, + { + "epoch": 0.4549402823018458, + "grad_norm": 5.042448582457387, + "learning_rate": 1.9806004526244966e-05, + "loss": 1.3191, + "step": 838 + }, + { + "epoch": 0.4554831704668838, + "grad_norm": 3.7467475303782236, + "learning_rate": 1.9805314558745876e-05, + "loss": 1.3318, + "step": 839 + }, + { + "epoch": 0.4560260586319218, + "grad_norm": 4.214820131429537, + "learning_rate": 1.980462337850986e-05, + "loss": 1.3958, + "step": 840 + }, + { + "epoch": 0.4565689467969598, + "grad_norm": 3.9655409992603965, + "learning_rate": 1.980393098562239e-05, + "loss": 0.9978, + "step": 841 + }, + { + "epoch": 0.4571118349619978, + "grad_norm": 5.522638140748038, + "learning_rate": 1.980323738016912e-05, + "loss": 1.4151, + "step": 842 + }, + { + "epoch": 0.4576547231270358, + "grad_norm": 3.84359618797871, + "learning_rate": 1.9802542562235822e-05, + "loss": 1.3567, + "step": 843 + }, + { + "epoch": 0.4581976112920738, + "grad_norm": 5.115138943098397, + "learning_rate": 1.9801846531908437e-05, + "loss": 1.6437, + "step": 844 + }, + { + "epoch": 0.4587404994571118, + "grad_norm": 3.811847477591368, + "learning_rate": 1.9801149289273054e-05, + "loss": 1.2335, + "step": 845 + }, + { + "epoch": 0.4592833876221498, + "grad_norm": 4.4009848681180115, + "learning_rate": 1.98004508344159e-05, + "loss": 1.2147, + "step": 846 + }, + { + "epoch": 0.4598262757871878, + "grad_norm": 3.850991850744113, + "learning_rate": 1.9799751167423376e-05, + "loss": 1.4865, + "step": 847 + }, + { + "epoch": 0.4603691639522258, + "grad_norm": 3.1303042065802917, + "learning_rate": 1.9799050288382004e-05, + "loss": 1.4926, + "step": 848 + }, + { + "epoch": 0.4609120521172638, + "grad_norm": 4.225513214699549, + "learning_rate": 1.9798348197378475e-05, + "loss": 1.2044, + "step": 849 + }, + { + "epoch": 0.46145494028230183, + "grad_norm": 4.139979880818486, + "learning_rate": 1.9797644894499625e-05, + "loss": 1.5682, + "step": 850 + }, + { + "epoch": 0.46199782844733983, + "grad_norm": 4.414209735978709, + "learning_rate": 1.979694037983244e-05, + "loss": 1.9033, + "step": 851 + }, + { + "epoch": 0.46254071661237783, + "grad_norm": 5.079031167256292, + "learning_rate": 1.9796234653464057e-05, + "loss": 1.8745, + "step": 852 + }, + { + "epoch": 0.46308360477741584, + "grad_norm": 6.59971413476838, + "learning_rate": 1.9795527715481755e-05, + "loss": 1.2449, + "step": 853 + }, + { + "epoch": 0.46362649294245384, + "grad_norm": 3.863740127637264, + "learning_rate": 1.9794819565972973e-05, + "loss": 1.3912, + "step": 854 + }, + { + "epoch": 0.46416938110749184, + "grad_norm": 3.8578186204961327, + "learning_rate": 1.9794110205025302e-05, + "loss": 1.3487, + "step": 855 + }, + { + "epoch": 0.46471226927252984, + "grad_norm": 4.19751015099396, + "learning_rate": 1.9793399632726466e-05, + "loss": 1.129, + "step": 856 + }, + { + "epoch": 0.46525515743756785, + "grad_norm": 4.261483190159906, + "learning_rate": 1.979268784916436e-05, + "loss": 1.7298, + "step": 857 + }, + { + "epoch": 0.46579804560260585, + "grad_norm": 4.1774094914047435, + "learning_rate": 1.9791974854427008e-05, + "loss": 0.8716, + "step": 858 + }, + { + "epoch": 0.46634093376764385, + "grad_norm": 3.845723101011648, + "learning_rate": 1.9791260648602603e-05, + "loss": 0.9086, + "step": 859 + }, + { + "epoch": 0.46688382193268185, + "grad_norm": 3.4922208632687393, + "learning_rate": 1.9790545231779476e-05, + "loss": 1.3688, + "step": 860 + }, + { + "epoch": 0.46742671009771986, + "grad_norm": 2.976761133296452, + "learning_rate": 1.978982860404611e-05, + "loss": 0.7865, + "step": 861 + }, + { + "epoch": 0.46796959826275786, + "grad_norm": 3.934765402497862, + "learning_rate": 1.978911076549114e-05, + "loss": 1.2993, + "step": 862 + }, + { + "epoch": 0.46851248642779586, + "grad_norm": 3.029198772099951, + "learning_rate": 1.978839171620335e-05, + "loss": 1.1737, + "step": 863 + }, + { + "epoch": 0.46905537459283386, + "grad_norm": 3.5637545602445946, + "learning_rate": 1.978767145627167e-05, + "loss": 1.1932, + "step": 864 + }, + { + "epoch": 0.46959826275787186, + "grad_norm": 3.131289093388952, + "learning_rate": 1.9786949985785187e-05, + "loss": 0.69, + "step": 865 + }, + { + "epoch": 0.47014115092290987, + "grad_norm": 3.522468470152659, + "learning_rate": 1.978622730483313e-05, + "loss": 1.0277, + "step": 866 + }, + { + "epoch": 0.47068403908794787, + "grad_norm": 4.046439598363532, + "learning_rate": 1.9785503413504883e-05, + "loss": 1.2917, + "step": 867 + }, + { + "epoch": 0.47122692725298587, + "grad_norm": 5.55259287782346, + "learning_rate": 1.978477831188998e-05, + "loss": 1.866, + "step": 868 + }, + { + "epoch": 0.4717698154180239, + "grad_norm": 5.874766516228883, + "learning_rate": 1.97840520000781e-05, + "loss": 1.6221, + "step": 869 + }, + { + "epoch": 0.4723127035830619, + "grad_norm": 5.530080877073305, + "learning_rate": 1.9783324478159074e-05, + "loss": 1.7617, + "step": 870 + }, + { + "epoch": 0.4728555917480999, + "grad_norm": 4.996268624326209, + "learning_rate": 1.9782595746222886e-05, + "loss": 1.5794, + "step": 871 + }, + { + "epoch": 0.4733984799131379, + "grad_norm": 3.934049315300115, + "learning_rate": 1.9781865804359663e-05, + "loss": 1.2603, + "step": 872 + }, + { + "epoch": 0.4739413680781759, + "grad_norm": 4.534039775579426, + "learning_rate": 1.978113465265969e-05, + "loss": 1.2766, + "step": 873 + }, + { + "epoch": 0.4744842562432139, + "grad_norm": 4.281791787646871, + "learning_rate": 1.9780402291213393e-05, + "loss": 1.6424, + "step": 874 + }, + { + "epoch": 0.4750271444082519, + "grad_norm": 4.443571747878234, + "learning_rate": 1.977966872011135e-05, + "loss": 1.514, + "step": 875 + }, + { + "epoch": 0.4755700325732899, + "grad_norm": 4.432653652742011, + "learning_rate": 1.9778933939444298e-05, + "loss": 1.2405, + "step": 876 + }, + { + "epoch": 0.4761129207383279, + "grad_norm": 3.408851551186198, + "learning_rate": 1.9778197949303107e-05, + "loss": 0.9597, + "step": 877 + }, + { + "epoch": 0.4766558089033659, + "grad_norm": 3.13258933410851, + "learning_rate": 1.9777460749778812e-05, + "loss": 0.8015, + "step": 878 + }, + { + "epoch": 0.4771986970684039, + "grad_norm": 3.535707436018844, + "learning_rate": 1.977672234096259e-05, + "loss": 1.3375, + "step": 879 + }, + { + "epoch": 0.4777415852334419, + "grad_norm": 5.210333639552661, + "learning_rate": 1.9775982722945764e-05, + "loss": 1.5039, + "step": 880 + }, + { + "epoch": 0.4782844733984799, + "grad_norm": 4.14115837125392, + "learning_rate": 1.9775241895819818e-05, + "loss": 1.0267, + "step": 881 + }, + { + "epoch": 0.4788273615635179, + "grad_norm": 4.390843049257208, + "learning_rate": 1.977449985967637e-05, + "loss": 1.6639, + "step": 882 + }, + { + "epoch": 0.4793702497285559, + "grad_norm": 3.5906251649150405, + "learning_rate": 1.9773756614607205e-05, + "loss": 1.2382, + "step": 883 + }, + { + "epoch": 0.4799131378935939, + "grad_norm": 4.78880370244169, + "learning_rate": 1.9773012160704242e-05, + "loss": 1.9436, + "step": 884 + }, + { + "epoch": 0.4804560260586319, + "grad_norm": 4.024679917575935, + "learning_rate": 1.977226649805956e-05, + "loss": 1.7126, + "step": 885 + }, + { + "epoch": 0.4809989142236699, + "grad_norm": 4.789782875875197, + "learning_rate": 1.9771519626765384e-05, + "loss": 1.4639, + "step": 886 + }, + { + "epoch": 0.4815418023887079, + "grad_norm": 4.188325588019175, + "learning_rate": 1.9770771546914088e-05, + "loss": 0.9698, + "step": 887 + }, + { + "epoch": 0.4820846905537459, + "grad_norm": 4.491436069698424, + "learning_rate": 1.9770022258598192e-05, + "loss": 1.5504, + "step": 888 + }, + { + "epoch": 0.4826275787187839, + "grad_norm": 4.889270563723842, + "learning_rate": 1.9769271761910373e-05, + "loss": 1.6052, + "step": 889 + }, + { + "epoch": 0.4831704668838219, + "grad_norm": 3.7615089990618493, + "learning_rate": 1.9768520056943454e-05, + "loss": 1.2445, + "step": 890 + }, + { + "epoch": 0.4837133550488599, + "grad_norm": 4.153654659347469, + "learning_rate": 1.9767767143790403e-05, + "loss": 1.4304, + "step": 891 + }, + { + "epoch": 0.48425624321389793, + "grad_norm": 4.040032964648755, + "learning_rate": 1.9767013022544346e-05, + "loss": 1.2906, + "step": 892 + }, + { + "epoch": 0.48479913137893593, + "grad_norm": 4.6559235863058275, + "learning_rate": 1.976625769329855e-05, + "loss": 1.5214, + "step": 893 + }, + { + "epoch": 0.48534201954397393, + "grad_norm": 4.224494142526306, + "learning_rate": 1.976550115614644e-05, + "loss": 1.2299, + "step": 894 + }, + { + "epoch": 0.48588490770901194, + "grad_norm": 5.225621473809677, + "learning_rate": 1.9764743411181585e-05, + "loss": 1.3489, + "step": 895 + }, + { + "epoch": 0.48642779587404994, + "grad_norm": 3.298440710545404, + "learning_rate": 1.97639844584977e-05, + "loss": 1.2107, + "step": 896 + }, + { + "epoch": 0.48697068403908794, + "grad_norm": 4.8373671778516485, + "learning_rate": 1.9763224298188652e-05, + "loss": 1.7748, + "step": 897 + }, + { + "epoch": 0.48751357220412594, + "grad_norm": 3.9332967347674224, + "learning_rate": 1.9762462930348465e-05, + "loss": 1.1036, + "step": 898 + }, + { + "epoch": 0.48805646036916395, + "grad_norm": 3.555513906222738, + "learning_rate": 1.9761700355071306e-05, + "loss": 1.5988, + "step": 899 + }, + { + "epoch": 0.48859934853420195, + "grad_norm": 4.115001448738943, + "learning_rate": 1.9760936572451487e-05, + "loss": 0.9789, + "step": 900 + }, + { + "epoch": 0.48914223669923995, + "grad_norm": 3.955269044937872, + "learning_rate": 1.9760171582583476e-05, + "loss": 1.3833, + "step": 901 + }, + { + "epoch": 0.48968512486427795, + "grad_norm": 2.718909991804504, + "learning_rate": 1.975940538556189e-05, + "loss": 1.1708, + "step": 902 + }, + { + "epoch": 0.49022801302931596, + "grad_norm": 4.234679090051759, + "learning_rate": 1.975863798148149e-05, + "loss": 1.3002, + "step": 903 + }, + { + "epoch": 0.49077090119435396, + "grad_norm": 4.162638017444828, + "learning_rate": 1.975786937043719e-05, + "loss": 1.246, + "step": 904 + }, + { + "epoch": 0.49131378935939196, + "grad_norm": 4.466027820294199, + "learning_rate": 1.975709955252406e-05, + "loss": 1.7345, + "step": 905 + }, + { + "epoch": 0.49185667752442996, + "grad_norm": 4.120456111509664, + "learning_rate": 1.97563285278373e-05, + "loss": 1.1339, + "step": 906 + }, + { + "epoch": 0.49239956568946797, + "grad_norm": 3.7932999801551346, + "learning_rate": 1.9755556296472286e-05, + "loss": 1.241, + "step": 907 + }, + { + "epoch": 0.49294245385450597, + "grad_norm": 3.3311865780622107, + "learning_rate": 1.9754782858524515e-05, + "loss": 1.2329, + "step": 908 + }, + { + "epoch": 0.49348534201954397, + "grad_norm": 4.21189840914184, + "learning_rate": 1.9754008214089654e-05, + "loss": 1.089, + "step": 909 + }, + { + "epoch": 0.494028230184582, + "grad_norm": 4.882048591403942, + "learning_rate": 1.9753232363263513e-05, + "loss": 1.3579, + "step": 910 + }, + { + "epoch": 0.49457111834962, + "grad_norm": 3.8495412115497833, + "learning_rate": 1.9752455306142052e-05, + "loss": 0.876, + "step": 911 + }, + { + "epoch": 0.495114006514658, + "grad_norm": 5.35378911928701, + "learning_rate": 1.975167704282137e-05, + "loss": 1.5025, + "step": 912 + }, + { + "epoch": 0.495656894679696, + "grad_norm": 3.4776107222093176, + "learning_rate": 1.9750897573397733e-05, + "loss": 1.1576, + "step": 913 + }, + { + "epoch": 0.496199782844734, + "grad_norm": 4.3680755226003525, + "learning_rate": 1.9750116897967545e-05, + "loss": 1.2519, + "step": 914 + }, + { + "epoch": 0.496742671009772, + "grad_norm": 4.154395835162648, + "learning_rate": 1.9749335016627355e-05, + "loss": 1.3397, + "step": 915 + }, + { + "epoch": 0.49728555917481, + "grad_norm": 5.465210053229008, + "learning_rate": 1.9748551929473873e-05, + "loss": 1.7555, + "step": 916 + }, + { + "epoch": 0.497828447339848, + "grad_norm": 4.391089608848125, + "learning_rate": 1.9747767636603953e-05, + "loss": 0.9685, + "step": 917 + }, + { + "epoch": 0.498371335504886, + "grad_norm": 4.215127655197256, + "learning_rate": 1.9746982138114597e-05, + "loss": 1.3889, + "step": 918 + }, + { + "epoch": 0.498914223669924, + "grad_norm": 4.559291789845839, + "learning_rate": 1.9746195434102956e-05, + "loss": 1.2411, + "step": 919 + }, + { + "epoch": 0.499457111834962, + "grad_norm": 4.226267506705079, + "learning_rate": 1.9745407524666327e-05, + "loss": 0.9517, + "step": 920 + }, + { + "epoch": 0.5, + "grad_norm": 4.289151336000736, + "learning_rate": 1.9744618409902164e-05, + "loss": 0.9123, + "step": 921 + }, + { + "epoch": 0.500542888165038, + "grad_norm": 5.356043714827742, + "learning_rate": 1.9743828089908067e-05, + "loss": 1.4039, + "step": 922 + }, + { + "epoch": 0.501085776330076, + "grad_norm": 4.867388642606108, + "learning_rate": 1.9743036564781785e-05, + "loss": 1.877, + "step": 923 + }, + { + "epoch": 0.501628664495114, + "grad_norm": 4.034273796102648, + "learning_rate": 1.9742243834621207e-05, + "loss": 1.6151, + "step": 924 + }, + { + "epoch": 0.502171552660152, + "grad_norm": 4.67131234305618, + "learning_rate": 1.9741449899524393e-05, + "loss": 1.6415, + "step": 925 + }, + { + "epoch": 0.50271444082519, + "grad_norm": 4.609211834325983, + "learning_rate": 1.9740654759589524e-05, + "loss": 1.2702, + "step": 926 + }, + { + "epoch": 0.503257328990228, + "grad_norm": 3.8693712873725064, + "learning_rate": 1.973985841491495e-05, + "loss": 1.0518, + "step": 927 + }, + { + "epoch": 0.503800217155266, + "grad_norm": 4.224951197105749, + "learning_rate": 1.9739060865599163e-05, + "loss": 1.1367, + "step": 928 + }, + { + "epoch": 0.504343105320304, + "grad_norm": 4.851656117293343, + "learning_rate": 1.973826211174081e-05, + "loss": 1.2753, + "step": 929 + }, + { + "epoch": 0.504885993485342, + "grad_norm": 4.255261482914816, + "learning_rate": 1.9737462153438675e-05, + "loss": 1.1723, + "step": 930 + }, + { + "epoch": 0.50542888165038, + "grad_norm": 3.8038614228711616, + "learning_rate": 1.9736660990791705e-05, + "loss": 1.1919, + "step": 931 + }, + { + "epoch": 0.505971769815418, + "grad_norm": 3.934227455395002, + "learning_rate": 1.9735858623898984e-05, + "loss": 1.1311, + "step": 932 + }, + { + "epoch": 0.506514657980456, + "grad_norm": 6.087081162565673, + "learning_rate": 1.9735055052859752e-05, + "loss": 1.4226, + "step": 933 + }, + { + "epoch": 0.507057546145494, + "grad_norm": 3.901680463400142, + "learning_rate": 1.9734250277773398e-05, + "loss": 0.8336, + "step": 934 + }, + { + "epoch": 0.507600434310532, + "grad_norm": 4.29804205692996, + "learning_rate": 1.9733444298739454e-05, + "loss": 1.5767, + "step": 935 + }, + { + "epoch": 0.50814332247557, + "grad_norm": 4.114842645062852, + "learning_rate": 1.9732637115857606e-05, + "loss": 1.3353, + "step": 936 + }, + { + "epoch": 0.508686210640608, + "grad_norm": 3.8525228530533706, + "learning_rate": 1.9731828729227686e-05, + "loss": 0.8208, + "step": 937 + }, + { + "epoch": 0.509229098805646, + "grad_norm": 4.592070280997817, + "learning_rate": 1.973101913894968e-05, + "loss": 1.3908, + "step": 938 + }, + { + "epoch": 0.509771986970684, + "grad_norm": 3.5842386277465628, + "learning_rate": 1.9730208345123718e-05, + "loss": 1.1996, + "step": 939 + }, + { + "epoch": 0.510314875135722, + "grad_norm": 6.053647281932304, + "learning_rate": 1.9729396347850082e-05, + "loss": 1.565, + "step": 940 + }, + { + "epoch": 0.51085776330076, + "grad_norm": 4.515664489508703, + "learning_rate": 1.9728583147229196e-05, + "loss": 1.4823, + "step": 941 + }, + { + "epoch": 0.511400651465798, + "grad_norm": 4.7752750595730475, + "learning_rate": 1.9727768743361644e-05, + "loss": 1.3867, + "step": 942 + }, + { + "epoch": 0.511943539630836, + "grad_norm": 3.983262894008015, + "learning_rate": 1.972695313634815e-05, + "loss": 1.1327, + "step": 943 + }, + { + "epoch": 0.512486427795874, + "grad_norm": 3.8510372454489326, + "learning_rate": 1.9726136326289586e-05, + "loss": 1.0233, + "step": 944 + }, + { + "epoch": 0.5130293159609121, + "grad_norm": 3.7881258264140634, + "learning_rate": 1.9725318313286985e-05, + "loss": 1.0239, + "step": 945 + }, + { + "epoch": 0.5135722041259501, + "grad_norm": 5.532602386318144, + "learning_rate": 1.9724499097441513e-05, + "loss": 1.8325, + "step": 946 + }, + { + "epoch": 0.5141150922909881, + "grad_norm": 5.978249809716352, + "learning_rate": 1.9723678678854493e-05, + "loss": 1.4899, + "step": 947 + }, + { + "epoch": 0.5146579804560261, + "grad_norm": 4.164217442125778, + "learning_rate": 1.9722857057627398e-05, + "loss": 1.5428, + "step": 948 + }, + { + "epoch": 0.5152008686210641, + "grad_norm": 3.315839086509451, + "learning_rate": 1.9722034233861848e-05, + "loss": 1.161, + "step": 949 + }, + { + "epoch": 0.5157437567861021, + "grad_norm": 4.362448119896709, + "learning_rate": 1.9721210207659608e-05, + "loss": 1.3402, + "step": 950 + }, + { + "epoch": 0.5162866449511401, + "grad_norm": 4.196652705893848, + "learning_rate": 1.9720384979122594e-05, + "loss": 1.4782, + "step": 951 + }, + { + "epoch": 0.5168295331161781, + "grad_norm": 4.093486911730782, + "learning_rate": 1.9719558548352876e-05, + "loss": 0.8508, + "step": 952 + }, + { + "epoch": 0.5173724212812161, + "grad_norm": 3.623412868803193, + "learning_rate": 1.9718730915452664e-05, + "loss": 1.4979, + "step": 953 + }, + { + "epoch": 0.5179153094462541, + "grad_norm": 3.4964911024452845, + "learning_rate": 1.9717902080524324e-05, + "loss": 1.0495, + "step": 954 + }, + { + "epoch": 0.5184581976112921, + "grad_norm": 4.021867363378223, + "learning_rate": 1.9717072043670367e-05, + "loss": 1.5284, + "step": 955 + }, + { + "epoch": 0.5190010857763301, + "grad_norm": 4.383918034859983, + "learning_rate": 1.9716240804993454e-05, + "loss": 1.3092, + "step": 956 + }, + { + "epoch": 0.5195439739413681, + "grad_norm": 4.61200677033352, + "learning_rate": 1.971540836459639e-05, + "loss": 1.8809, + "step": 957 + }, + { + "epoch": 0.5200868621064061, + "grad_norm": 4.490685725956694, + "learning_rate": 1.9714574722582142e-05, + "loss": 1.4004, + "step": 958 + }, + { + "epoch": 0.5206297502714441, + "grad_norm": 3.4814367857792736, + "learning_rate": 1.9713739879053802e-05, + "loss": 0.7918, + "step": 959 + }, + { + "epoch": 0.5211726384364821, + "grad_norm": 4.751597566480749, + "learning_rate": 1.9712903834114635e-05, + "loss": 1.6595, + "step": 960 + }, + { + "epoch": 0.5217155266015201, + "grad_norm": 4.535333205317335, + "learning_rate": 1.9712066587868042e-05, + "loss": 1.5676, + "step": 961 + }, + { + "epoch": 0.5222584147665581, + "grad_norm": 4.285001993552699, + "learning_rate": 1.9711228140417577e-05, + "loss": 1.0156, + "step": 962 + }, + { + "epoch": 0.5228013029315961, + "grad_norm": 4.61361421766516, + "learning_rate": 1.9710388491866934e-05, + "loss": 0.9552, + "step": 963 + }, + { + "epoch": 0.5233441910966341, + "grad_norm": 5.760482951437034, + "learning_rate": 1.9709547642319968e-05, + "loss": 2.2573, + "step": 964 + }, + { + "epoch": 0.5238870792616721, + "grad_norm": 4.38974180509626, + "learning_rate": 1.9708705591880674e-05, + "loss": 0.9939, + "step": 965 + }, + { + "epoch": 0.5244299674267101, + "grad_norm": 4.856793718321527, + "learning_rate": 1.97078623406532e-05, + "loss": 1.6161, + "step": 966 + }, + { + "epoch": 0.5249728555917481, + "grad_norm": 4.164767885626303, + "learning_rate": 1.9707017888741838e-05, + "loss": 1.4327, + "step": 967 + }, + { + "epoch": 0.5255157437567861, + "grad_norm": 5.163022804287371, + "learning_rate": 1.970617223625104e-05, + "loss": 1.2063, + "step": 968 + }, + { + "epoch": 0.5260586319218241, + "grad_norm": 4.826390195584734, + "learning_rate": 1.9705325383285384e-05, + "loss": 1.609, + "step": 969 + }, + { + "epoch": 0.5266015200868621, + "grad_norm": 6.3570394668292405, + "learning_rate": 1.9704477329949617e-05, + "loss": 1.5845, + "step": 970 + }, + { + "epoch": 0.5271444082519001, + "grad_norm": 5.176208926963624, + "learning_rate": 1.9703628076348628e-05, + "loss": 1.5792, + "step": 971 + }, + { + "epoch": 0.5276872964169381, + "grad_norm": 4.968725075311641, + "learning_rate": 1.9702777622587452e-05, + "loss": 1.774, + "step": 972 + }, + { + "epoch": 0.5282301845819761, + "grad_norm": 3.733580423042607, + "learning_rate": 1.9701925968771277e-05, + "loss": 0.8424, + "step": 973 + }, + { + "epoch": 0.5287730727470141, + "grad_norm": 4.330993358163065, + "learning_rate": 1.9701073115005437e-05, + "loss": 1.6198, + "step": 974 + }, + { + "epoch": 0.5293159609120521, + "grad_norm": 4.64032700965104, + "learning_rate": 1.9700219061395408e-05, + "loss": 1.2958, + "step": 975 + }, + { + "epoch": 0.5298588490770901, + "grad_norm": 4.08538215893539, + "learning_rate": 1.969936380804683e-05, + "loss": 1.5371, + "step": 976 + }, + { + "epoch": 0.5304017372421281, + "grad_norm": 3.7672102191405012, + "learning_rate": 1.9698507355065478e-05, + "loss": 1.4831, + "step": 977 + }, + { + "epoch": 0.5309446254071661, + "grad_norm": 4.237599031471779, + "learning_rate": 1.969764970255728e-05, + "loss": 1.4561, + "step": 978 + }, + { + "epoch": 0.5314875135722041, + "grad_norm": 4.296489590459447, + "learning_rate": 1.9696790850628308e-05, + "loss": 1.3813, + "step": 979 + }, + { + "epoch": 0.5320304017372421, + "grad_norm": 4.037479773042692, + "learning_rate": 1.9695930799384787e-05, + "loss": 1.103, + "step": 980 + }, + { + "epoch": 0.5325732899022801, + "grad_norm": 3.388491089813824, + "learning_rate": 1.9695069548933097e-05, + "loss": 1.0416, + "step": 981 + }, + { + "epoch": 0.5331161780673181, + "grad_norm": 3.6808593992848873, + "learning_rate": 1.969420709937975e-05, + "loss": 1.3158, + "step": 982 + }, + { + "epoch": 0.5336590662323561, + "grad_norm": 3.574434479685402, + "learning_rate": 1.969334345083142e-05, + "loss": 0.9542, + "step": 983 + }, + { + "epoch": 0.5342019543973942, + "grad_norm": 4.205886917864195, + "learning_rate": 1.9692478603394926e-05, + "loss": 1.1542, + "step": 984 + }, + { + "epoch": 0.5347448425624322, + "grad_norm": 3.89660120144171, + "learning_rate": 1.9691612557177225e-05, + "loss": 1.2374, + "step": 985 + }, + { + "epoch": 0.5352877307274702, + "grad_norm": 5.508260629607227, + "learning_rate": 1.9690745312285443e-05, + "loss": 2.0836, + "step": 986 + }, + { + "epoch": 0.5358306188925082, + "grad_norm": 5.482641917586033, + "learning_rate": 1.968987686882683e-05, + "loss": 1.5608, + "step": 987 + }, + { + "epoch": 0.5363735070575462, + "grad_norm": 5.4640893579957845, + "learning_rate": 1.9689007226908807e-05, + "loss": 1.3632, + "step": 988 + }, + { + "epoch": 0.5369163952225842, + "grad_norm": 4.163793092253212, + "learning_rate": 1.9688136386638926e-05, + "loss": 1.0509, + "step": 989 + }, + { + "epoch": 0.5374592833876222, + "grad_norm": 7.431185891169344, + "learning_rate": 1.96872643481249e-05, + "loss": 2.2147, + "step": 990 + }, + { + "epoch": 0.5380021715526602, + "grad_norm": 4.243102861037138, + "learning_rate": 1.9686391111474574e-05, + "loss": 0.9978, + "step": 991 + }, + { + "epoch": 0.5385450597176982, + "grad_norm": 4.250454886057412, + "learning_rate": 1.968551667679596e-05, + "loss": 1.5265, + "step": 992 + }, + { + "epoch": 0.5390879478827362, + "grad_norm": 5.989466882905416, + "learning_rate": 1.9684641044197207e-05, + "loss": 1.3344, + "step": 993 + }, + { + "epoch": 0.5396308360477742, + "grad_norm": 6.003524822087355, + "learning_rate": 1.9683764213786617e-05, + "loss": 2.0785, + "step": 994 + }, + { + "epoch": 0.5401737242128122, + "grad_norm": 4.535858561115482, + "learning_rate": 1.9682886185672633e-05, + "loss": 1.7654, + "step": 995 + }, + { + "epoch": 0.5407166123778502, + "grad_norm": 3.304287633610362, + "learning_rate": 1.9682006959963854e-05, + "loss": 0.9151, + "step": 996 + }, + { + "epoch": 0.5412595005428882, + "grad_norm": 3.6154194334954797, + "learning_rate": 1.9681126536769022e-05, + "loss": 1.1871, + "step": 997 + }, + { + "epoch": 0.5418023887079262, + "grad_norm": 5.596808777173772, + "learning_rate": 1.968024491619703e-05, + "loss": 1.1773, + "step": 998 + }, + { + "epoch": 0.5423452768729642, + "grad_norm": 4.486728043355676, + "learning_rate": 1.9679362098356923e-05, + "loss": 1.6495, + "step": 999 + }, + { + "epoch": 0.5428881650380022, + "grad_norm": 5.254525660355539, + "learning_rate": 1.9678478083357882e-05, + "loss": 1.7632, + "step": 1000 + }, + { + "epoch": 0.5434310532030402, + "grad_norm": 4.084222439189464, + "learning_rate": 1.9677592871309248e-05, + "loss": 1.4108, + "step": 1001 + }, + { + "epoch": 0.5439739413680782, + "grad_norm": 3.7647794378542248, + "learning_rate": 1.9676706462320504e-05, + "loss": 1.0893, + "step": 1002 + }, + { + "epoch": 0.5445168295331162, + "grad_norm": 3.658128786211824, + "learning_rate": 1.967581885650128e-05, + "loss": 1.0626, + "step": 1003 + }, + { + "epoch": 0.5450597176981542, + "grad_norm": 4.111525087755858, + "learning_rate": 1.967493005396136e-05, + "loss": 1.1915, + "step": 1004 + }, + { + "epoch": 0.5456026058631922, + "grad_norm": 4.130761699023899, + "learning_rate": 1.967404005481067e-05, + "loss": 1.1007, + "step": 1005 + }, + { + "epoch": 0.5461454940282302, + "grad_norm": 4.610164776763781, + "learning_rate": 1.9673148859159292e-05, + "loss": 1.3626, + "step": 1006 + }, + { + "epoch": 0.5466883821932682, + "grad_norm": 3.3516729554956686, + "learning_rate": 1.9672256467117445e-05, + "loss": 1.1138, + "step": 1007 + }, + { + "epoch": 0.5472312703583062, + "grad_norm": 5.185808250719794, + "learning_rate": 1.9671362878795502e-05, + "loss": 1.5639, + "step": 1008 + }, + { + "epoch": 0.5477741585233442, + "grad_norm": 3.9549787550673323, + "learning_rate": 1.9670468094303983e-05, + "loss": 1.0094, + "step": 1009 + }, + { + "epoch": 0.5483170466883822, + "grad_norm": 5.025417088271204, + "learning_rate": 1.966957211375356e-05, + "loss": 1.9741, + "step": 1010 + }, + { + "epoch": 0.5488599348534202, + "grad_norm": 5.1560673391175, + "learning_rate": 1.9668674937255044e-05, + "loss": 0.9653, + "step": 1011 + }, + { + "epoch": 0.5494028230184582, + "grad_norm": 4.843540796107876, + "learning_rate": 1.9667776564919404e-05, + "loss": 1.4745, + "step": 1012 + }, + { + "epoch": 0.5499457111834962, + "grad_norm": 3.8379573247832193, + "learning_rate": 1.966687699685775e-05, + "loss": 1.537, + "step": 1013 + }, + { + "epoch": 0.5504885993485342, + "grad_norm": 4.291804816024517, + "learning_rate": 1.9665976233181342e-05, + "loss": 1.6859, + "step": 1014 + }, + { + "epoch": 0.5510314875135722, + "grad_norm": 4.827630960232388, + "learning_rate": 1.966507427400159e-05, + "loss": 1.3205, + "step": 1015 + }, + { + "epoch": 0.5515743756786102, + "grad_norm": 4.251540690047392, + "learning_rate": 1.9664171119430044e-05, + "loss": 0.8213, + "step": 1016 + }, + { + "epoch": 0.5521172638436482, + "grad_norm": 3.7693342275615476, + "learning_rate": 1.9663266769578414e-05, + "loss": 1.0349, + "step": 1017 + }, + { + "epoch": 0.5526601520086862, + "grad_norm": 4.075472936638236, + "learning_rate": 1.966236122455855e-05, + "loss": 1.0853, + "step": 1018 + }, + { + "epoch": 0.5532030401737242, + "grad_norm": 4.771167333111926, + "learning_rate": 1.9661454484482448e-05, + "loss": 1.2135, + "step": 1019 + }, + { + "epoch": 0.5537459283387622, + "grad_norm": 3.834854303591261, + "learning_rate": 1.9660546549462262e-05, + "loss": 1.0972, + "step": 1020 + }, + { + "epoch": 0.5542888165038002, + "grad_norm": 4.712560042125338, + "learning_rate": 1.9659637419610278e-05, + "loss": 0.8779, + "step": 1021 + }, + { + "epoch": 0.5548317046688382, + "grad_norm": 3.8174143787104566, + "learning_rate": 1.9658727095038942e-05, + "loss": 0.7395, + "step": 1022 + }, + { + "epoch": 0.5553745928338762, + "grad_norm": 5.970646625648977, + "learning_rate": 1.965781557586085e-05, + "loss": 1.4233, + "step": 1023 + }, + { + "epoch": 0.5559174809989142, + "grad_norm": 3.872798187838392, + "learning_rate": 1.9656902862188732e-05, + "loss": 0.8147, + "step": 1024 + }, + { + "epoch": 0.5564603691639523, + "grad_norm": 5.632609234869673, + "learning_rate": 1.9655988954135473e-05, + "loss": 1.1581, + "step": 1025 + }, + { + "epoch": 0.5570032573289903, + "grad_norm": 4.141188191911578, + "learning_rate": 1.9655073851814117e-05, + "loss": 1.1707, + "step": 1026 + }, + { + "epoch": 0.5575461454940283, + "grad_norm": 5.7835323867879564, + "learning_rate": 1.9654157555337837e-05, + "loss": 1.2383, + "step": 1027 + }, + { + "epoch": 0.5580890336590663, + "grad_norm": 6.410603262158535, + "learning_rate": 1.9653240064819965e-05, + "loss": 2.3461, + "step": 1028 + }, + { + "epoch": 0.5586319218241043, + "grad_norm": 4.661944777700264, + "learning_rate": 1.9652321380373974e-05, + "loss": 1.1718, + "step": 1029 + }, + { + "epoch": 0.5591748099891423, + "grad_norm": 4.224334614804915, + "learning_rate": 1.9651401502113497e-05, + "loss": 1.0937, + "step": 1030 + }, + { + "epoch": 0.5597176981541803, + "grad_norm": 4.292561821167698, + "learning_rate": 1.9650480430152295e-05, + "loss": 0.8306, + "step": 1031 + }, + { + "epoch": 0.5602605863192183, + "grad_norm": 4.515070808084738, + "learning_rate": 1.9649558164604293e-05, + "loss": 1.4042, + "step": 1032 + }, + { + "epoch": 0.5608034744842563, + "grad_norm": 4.906478484057087, + "learning_rate": 1.964863470558356e-05, + "loss": 1.3396, + "step": 1033 + }, + { + "epoch": 0.5613463626492943, + "grad_norm": 3.9530548146664306, + "learning_rate": 1.9647710053204307e-05, + "loss": 0.917, + "step": 1034 + }, + { + "epoch": 0.5618892508143323, + "grad_norm": 4.640160859146004, + "learning_rate": 1.96467842075809e-05, + "loss": 1.6625, + "step": 1035 + }, + { + "epoch": 0.5624321389793703, + "grad_norm": 4.921887921875984, + "learning_rate": 1.964585716882785e-05, + "loss": 1.5133, + "step": 1036 + }, + { + "epoch": 0.5629750271444083, + "grad_norm": 4.667679101301949, + "learning_rate": 1.964492893705981e-05, + "loss": 0.9713, + "step": 1037 + }, + { + "epoch": 0.5635179153094463, + "grad_norm": 6.0457527064758825, + "learning_rate": 1.9643999512391586e-05, + "loss": 1.1101, + "step": 1038 + }, + { + "epoch": 0.5640608034744843, + "grad_norm": 5.188403972218186, + "learning_rate": 1.964306889493813e-05, + "loss": 1.8276, + "step": 1039 + }, + { + "epoch": 0.5646036916395223, + "grad_norm": 4.449370010668479, + "learning_rate": 1.9642137084814548e-05, + "loss": 1.0797, + "step": 1040 + }, + { + "epoch": 0.5651465798045603, + "grad_norm": 5.97660491495462, + "learning_rate": 1.9641204082136085e-05, + "loss": 1.913, + "step": 1041 + }, + { + "epoch": 0.5656894679695983, + "grad_norm": 5.611282387277076, + "learning_rate": 1.9640269887018135e-05, + "loss": 2.4593, + "step": 1042 + }, + { + "epoch": 0.5662323561346363, + "grad_norm": 3.6226559029227996, + "learning_rate": 1.9639334499576237e-05, + "loss": 0.7939, + "step": 1043 + }, + { + "epoch": 0.5667752442996743, + "grad_norm": 4.701569628633213, + "learning_rate": 1.963839791992609e-05, + "loss": 1.4092, + "step": 1044 + }, + { + "epoch": 0.5673181324647123, + "grad_norm": 5.185222096624812, + "learning_rate": 1.9637460148183525e-05, + "loss": 1.3851, + "step": 1045 + }, + { + "epoch": 0.5678610206297503, + "grad_norm": 4.094772946058271, + "learning_rate": 1.963652118446453e-05, + "loss": 1.1872, + "step": 1046 + }, + { + "epoch": 0.5684039087947883, + "grad_norm": 4.653748340189311, + "learning_rate": 1.9635581028885233e-05, + "loss": 1.1546, + "step": 1047 + }, + { + "epoch": 0.5689467969598263, + "grad_norm": 3.9426831069951844, + "learning_rate": 1.9634639681561924e-05, + "loss": 0.9324, + "step": 1048 + }, + { + "epoch": 0.5694896851248643, + "grad_norm": 4.949304407204856, + "learning_rate": 1.9633697142611017e-05, + "loss": 1.2372, + "step": 1049 + }, + { + "epoch": 0.5700325732899023, + "grad_norm": 5.364689744345368, + "learning_rate": 1.9632753412149096e-05, + "loss": 1.5104, + "step": 1050 + }, + { + "epoch": 0.5705754614549403, + "grad_norm": 5.307969223857254, + "learning_rate": 1.9631808490292884e-05, + "loss": 1.212, + "step": 1051 + }, + { + "epoch": 0.5711183496199783, + "grad_norm": 5.197310781163775, + "learning_rate": 1.963086237715924e-05, + "loss": 1.0175, + "step": 1052 + }, + { + "epoch": 0.5716612377850163, + "grad_norm": 4.095170540464123, + "learning_rate": 1.9629915072865194e-05, + "loss": 1.2711, + "step": 1053 + }, + { + "epoch": 0.5722041259500543, + "grad_norm": 4.816474093418903, + "learning_rate": 1.9628966577527902e-05, + "loss": 0.7274, + "step": 1054 + }, + { + "epoch": 0.5727470141150923, + "grad_norm": 4.000604469041167, + "learning_rate": 1.962801689126468e-05, + "loss": 1.2913, + "step": 1055 + }, + { + "epoch": 0.5732899022801303, + "grad_norm": 4.31668064977018, + "learning_rate": 1.962706601419298e-05, + "loss": 1.2598, + "step": 1056 + }, + { + "epoch": 0.5738327904451683, + "grad_norm": 5.033983318277113, + "learning_rate": 1.9626113946430414e-05, + "loss": 1.5149, + "step": 1057 + }, + { + "epoch": 0.5743756786102063, + "grad_norm": 4.51738649586986, + "learning_rate": 1.9625160688094733e-05, + "loss": 1.2389, + "step": 1058 + }, + { + "epoch": 0.5749185667752443, + "grad_norm": 3.6899120907028853, + "learning_rate": 1.9624206239303837e-05, + "loss": 1.1632, + "step": 1059 + }, + { + "epoch": 0.5754614549402823, + "grad_norm": 4.423539955787714, + "learning_rate": 1.9623250600175775e-05, + "loss": 0.9526, + "step": 1060 + }, + { + "epoch": 0.5760043431053203, + "grad_norm": 5.062986942342047, + "learning_rate": 1.962229377082874e-05, + "loss": 1.6768, + "step": 1061 + }, + { + "epoch": 0.5765472312703583, + "grad_norm": 5.044415439938055, + "learning_rate": 1.962133575138108e-05, + "loss": 1.3808, + "step": 1062 + }, + { + "epoch": 0.5770901194353963, + "grad_norm": 3.1405283495786587, + "learning_rate": 1.962037654195128e-05, + "loss": 0.7588, + "step": 1063 + }, + { + "epoch": 0.5776330076004343, + "grad_norm": 5.449873922495929, + "learning_rate": 1.9619416142657974e-05, + "loss": 1.1544, + "step": 1064 + }, + { + "epoch": 0.5781758957654723, + "grad_norm": 4.268527750000251, + "learning_rate": 1.961845455361995e-05, + "loss": 1.1941, + "step": 1065 + }, + { + "epoch": 0.5787187839305103, + "grad_norm": 4.6641878464575095, + "learning_rate": 1.9617491774956137e-05, + "loss": 1.7162, + "step": 1066 + }, + { + "epoch": 0.5792616720955484, + "grad_norm": 4.583539478633983, + "learning_rate": 1.9616527806785612e-05, + "loss": 1.1415, + "step": 1067 + }, + { + "epoch": 0.5798045602605864, + "grad_norm": 4.325093393868848, + "learning_rate": 1.96155626492276e-05, + "loss": 1.3284, + "step": 1068 + }, + { + "epoch": 0.5803474484256244, + "grad_norm": 4.180109104666535, + "learning_rate": 1.9614596302401478e-05, + "loss": 1.136, + "step": 1069 + }, + { + "epoch": 0.5808903365906624, + "grad_norm": 4.108700840520527, + "learning_rate": 1.9613628766426762e-05, + "loss": 1.6142, + "step": 1070 + }, + { + "epoch": 0.5814332247557004, + "grad_norm": 5.1931558422083555, + "learning_rate": 1.9612660041423116e-05, + "loss": 0.8479, + "step": 1071 + }, + { + "epoch": 0.5819761129207384, + "grad_norm": 6.008601497593343, + "learning_rate": 1.9611690127510358e-05, + "loss": 1.4743, + "step": 1072 + }, + { + "epoch": 0.5825190010857764, + "grad_norm": 4.531622764164067, + "learning_rate": 1.9610719024808444e-05, + "loss": 1.5755, + "step": 1073 + }, + { + "epoch": 0.5830618892508144, + "grad_norm": 6.18647874774261, + "learning_rate": 1.9609746733437486e-05, + "loss": 1.7369, + "step": 1074 + }, + { + "epoch": 0.5836047774158524, + "grad_norm": 5.5880249590216025, + "learning_rate": 1.9608773253517738e-05, + "loss": 1.4402, + "step": 1075 + }, + { + "epoch": 0.5841476655808904, + "grad_norm": 4.798692544278573, + "learning_rate": 1.9607798585169595e-05, + "loss": 1.2257, + "step": 1076 + }, + { + "epoch": 0.5846905537459284, + "grad_norm": 5.635786862796798, + "learning_rate": 1.960682272851361e-05, + "loss": 1.183, + "step": 1077 + }, + { + "epoch": 0.5852334419109664, + "grad_norm": 3.3979344506068037, + "learning_rate": 1.9605845683670484e-05, + "loss": 0.6463, + "step": 1078 + }, + { + "epoch": 0.5857763300760044, + "grad_norm": 4.5152680576801405, + "learning_rate": 1.960486745076105e-05, + "loss": 0.8476, + "step": 1079 + }, + { + "epoch": 0.5863192182410424, + "grad_norm": 4.4606140059612835, + "learning_rate": 1.9603888029906305e-05, + "loss": 1.185, + "step": 1080 + }, + { + "epoch": 0.5868621064060804, + "grad_norm": 4.229219175605004, + "learning_rate": 1.960290742122738e-05, + "loss": 1.3672, + "step": 1081 + }, + { + "epoch": 0.5874049945711184, + "grad_norm": 4.964172517040214, + "learning_rate": 1.960192562484556e-05, + "loss": 1.0561, + "step": 1082 + }, + { + "epoch": 0.5879478827361564, + "grad_norm": 4.8348336861653785, + "learning_rate": 1.960094264088228e-05, + "loss": 1.1005, + "step": 1083 + }, + { + "epoch": 0.5884907709011944, + "grad_norm": 4.3153540955607514, + "learning_rate": 1.959995846945911e-05, + "loss": 1.6513, + "step": 1084 + }, + { + "epoch": 0.5890336590662324, + "grad_norm": 4.733352525701168, + "learning_rate": 1.9598973110697773e-05, + "loss": 1.6162, + "step": 1085 + }, + { + "epoch": 0.5895765472312704, + "grad_norm": 5.402815147148315, + "learning_rate": 1.959798656472015e-05, + "loss": 1.4178, + "step": 1086 + }, + { + "epoch": 0.5901194353963084, + "grad_norm": 3.8276911935691147, + "learning_rate": 1.9596998831648247e-05, + "loss": 1.1554, + "step": 1087 + }, + { + "epoch": 0.5906623235613464, + "grad_norm": 4.499779496087119, + "learning_rate": 1.9596009911604232e-05, + "loss": 0.9786, + "step": 1088 + }, + { + "epoch": 0.5912052117263844, + "grad_norm": 4.3705352636049595, + "learning_rate": 1.959501980471042e-05, + "loss": 1.0827, + "step": 1089 + }, + { + "epoch": 0.5917480998914224, + "grad_norm": 4.9770508472387025, + "learning_rate": 1.9594028511089264e-05, + "loss": 1.2512, + "step": 1090 + }, + { + "epoch": 0.5922909880564604, + "grad_norm": 4.23068807802663, + "learning_rate": 1.9593036030863376e-05, + "loss": 0.7528, + "step": 1091 + }, + { + "epoch": 0.5928338762214984, + "grad_norm": 4.321323716192747, + "learning_rate": 1.9592042364155496e-05, + "loss": 1.0773, + "step": 1092 + }, + { + "epoch": 0.5933767643865364, + "grad_norm": 3.380036112638087, + "learning_rate": 1.9591047511088535e-05, + "loss": 0.7238, + "step": 1093 + }, + { + "epoch": 0.5939196525515744, + "grad_norm": 3.9722637658815447, + "learning_rate": 1.959005147178553e-05, + "loss": 0.8132, + "step": 1094 + }, + { + "epoch": 0.5944625407166124, + "grad_norm": 4.354488865716506, + "learning_rate": 1.9589054246369673e-05, + "loss": 1.1214, + "step": 1095 + }, + { + "epoch": 0.5950054288816504, + "grad_norm": 4.337355429292092, + "learning_rate": 1.9588055834964307e-05, + "loss": 0.9039, + "step": 1096 + }, + { + "epoch": 0.5955483170466884, + "grad_norm": 5.864047590425705, + "learning_rate": 1.9587056237692912e-05, + "loss": 1.6648, + "step": 1097 + }, + { + "epoch": 0.5960912052117264, + "grad_norm": 4.904173131706714, + "learning_rate": 1.9586055454679123e-05, + "loss": 1.1465, + "step": 1098 + }, + { + "epoch": 0.5966340933767644, + "grad_norm": 4.411877930208604, + "learning_rate": 1.9585053486046717e-05, + "loss": 1.4579, + "step": 1099 + }, + { + "epoch": 0.5971769815418024, + "grad_norm": 4.985171448881133, + "learning_rate": 1.9584050331919616e-05, + "loss": 1.5497, + "step": 1100 + }, + { + "epoch": 0.5977198697068404, + "grad_norm": 4.950532301400581, + "learning_rate": 1.9583045992421902e-05, + "loss": 1.3811, + "step": 1101 + }, + { + "epoch": 0.5982627578718784, + "grad_norm": 4.154119544907876, + "learning_rate": 1.9582040467677782e-05, + "loss": 1.186, + "step": 1102 + }, + { + "epoch": 0.5988056460369164, + "grad_norm": 4.069018604381756, + "learning_rate": 1.9581033757811628e-05, + "loss": 0.9312, + "step": 1103 + }, + { + "epoch": 0.5993485342019544, + "grad_norm": 5.33070083129983, + "learning_rate": 1.9580025862947948e-05, + "loss": 1.4274, + "step": 1104 + }, + { + "epoch": 0.5998914223669924, + "grad_norm": 4.855606632567906, + "learning_rate": 1.95790167832114e-05, + "loss": 1.0875, + "step": 1105 + }, + { + "epoch": 0.6004343105320304, + "grad_norm": 5.30182267546013, + "learning_rate": 1.957800651872679e-05, + "loss": 1.6501, + "step": 1106 + }, + { + "epoch": 0.6009771986970684, + "grad_norm": 4.433903391255917, + "learning_rate": 1.957699506961907e-05, + "loss": 1.3897, + "step": 1107 + }, + { + "epoch": 0.6015200868621065, + "grad_norm": 4.687653782723915, + "learning_rate": 1.9575982436013335e-05, + "loss": 0.8484, + "step": 1108 + }, + { + "epoch": 0.6020629750271445, + "grad_norm": 5.650426140105992, + "learning_rate": 1.957496861803483e-05, + "loss": 1.923, + "step": 1109 + }, + { + "epoch": 0.6026058631921825, + "grad_norm": 4.024650772512973, + "learning_rate": 1.957395361580895e-05, + "loss": 1.2713, + "step": 1110 + }, + { + "epoch": 0.6031487513572205, + "grad_norm": 4.343921133786845, + "learning_rate": 1.9572937429461223e-05, + "loss": 1.3152, + "step": 1111 + }, + { + "epoch": 0.6036916395222585, + "grad_norm": 4.1467176317930265, + "learning_rate": 1.957192005911734e-05, + "loss": 1.3325, + "step": 1112 + }, + { + "epoch": 0.6042345276872965, + "grad_norm": 4.305465221711973, + "learning_rate": 1.9570901504903128e-05, + "loss": 0.9237, + "step": 1113 + }, + { + "epoch": 0.6047774158523345, + "grad_norm": 4.101045495900469, + "learning_rate": 1.9569881766944564e-05, + "loss": 1.3932, + "step": 1114 + }, + { + "epoch": 0.6053203040173725, + "grad_norm": 4.0920748344339835, + "learning_rate": 1.956886084536777e-05, + "loss": 1.0067, + "step": 1115 + }, + { + "epoch": 0.6058631921824105, + "grad_norm": 4.844431281019285, + "learning_rate": 1.956783874029902e-05, + "loss": 1.2726, + "step": 1116 + }, + { + "epoch": 0.6064060803474485, + "grad_norm": 4.329792546128024, + "learning_rate": 1.9566815451864723e-05, + "loss": 1.1144, + "step": 1117 + }, + { + "epoch": 0.6069489685124865, + "grad_norm": 5.430322578642178, + "learning_rate": 1.9565790980191447e-05, + "loss": 1.3307, + "step": 1118 + }, + { + "epoch": 0.6074918566775245, + "grad_norm": 3.672369600344093, + "learning_rate": 1.9564765325405895e-05, + "loss": 0.8514, + "step": 1119 + }, + { + "epoch": 0.6080347448425625, + "grad_norm": 3.6433282223345334, + "learning_rate": 1.9563738487634924e-05, + "loss": 1.1972, + "step": 1120 + }, + { + "epoch": 0.6085776330076005, + "grad_norm": 4.23249445053715, + "learning_rate": 1.956271046700553e-05, + "loss": 1.0638, + "step": 1121 + }, + { + "epoch": 0.6091205211726385, + "grad_norm": 3.9162807435580134, + "learning_rate": 1.956168126364487e-05, + "loss": 1.121, + "step": 1122 + }, + { + "epoch": 0.6096634093376765, + "grad_norm": 3.3651337591577812, + "learning_rate": 1.956065087768023e-05, + "loss": 0.9604, + "step": 1123 + }, + { + "epoch": 0.6102062975027145, + "grad_norm": 5.5905307606778685, + "learning_rate": 1.955961930923905e-05, + "loss": 1.4509, + "step": 1124 + }, + { + "epoch": 0.6107491856677525, + "grad_norm": 3.6120233107886404, + "learning_rate": 1.955858655844892e-05, + "loss": 1.0106, + "step": 1125 + }, + { + "epoch": 0.6112920738327905, + "grad_norm": 3.6242967727445468, + "learning_rate": 1.9557552625437574e-05, + "loss": 1.0768, + "step": 1126 + }, + { + "epoch": 0.6118349619978285, + "grad_norm": 4.856763858492221, + "learning_rate": 1.9556517510332883e-05, + "loss": 1.6372, + "step": 1127 + }, + { + "epoch": 0.6123778501628665, + "grad_norm": 4.648427809898363, + "learning_rate": 1.9555481213262873e-05, + "loss": 1.1622, + "step": 1128 + }, + { + "epoch": 0.6129207383279045, + "grad_norm": 4.228089396781667, + "learning_rate": 1.9554443734355723e-05, + "loss": 1.067, + "step": 1129 + }, + { + "epoch": 0.6134636264929425, + "grad_norm": 4.069786144676624, + "learning_rate": 1.9553405073739743e-05, + "loss": 1.1427, + "step": 1130 + }, + { + "epoch": 0.6140065146579805, + "grad_norm": 3.112182391585594, + "learning_rate": 1.9552365231543395e-05, + "loss": 0.7862, + "step": 1131 + }, + { + "epoch": 0.6145494028230185, + "grad_norm": 4.783518734299315, + "learning_rate": 1.955132420789529e-05, + "loss": 1.0366, + "step": 1132 + }, + { + "epoch": 0.6150922909880565, + "grad_norm": 4.905128048667836, + "learning_rate": 1.9550282002924187e-05, + "loss": 1.2566, + "step": 1133 + }, + { + "epoch": 0.6156351791530945, + "grad_norm": 4.643678450407587, + "learning_rate": 1.9549238616758987e-05, + "loss": 1.2749, + "step": 1134 + }, + { + "epoch": 0.6161780673181325, + "grad_norm": 4.18363664236292, + "learning_rate": 1.954819404952873e-05, + "loss": 1.1523, + "step": 1135 + }, + { + "epoch": 0.6167209554831705, + "grad_norm": 3.9776065520060335, + "learning_rate": 1.9547148301362623e-05, + "loss": 0.8672, + "step": 1136 + }, + { + "epoch": 0.6172638436482085, + "grad_norm": 5.417170242762362, + "learning_rate": 1.9546101372389994e-05, + "loss": 1.6434, + "step": 1137 + }, + { + "epoch": 0.6178067318132465, + "grad_norm": 3.3325133642660068, + "learning_rate": 1.9545053262740335e-05, + "loss": 1.1447, + "step": 1138 + }, + { + "epoch": 0.6183496199782845, + "grad_norm": 5.4727757064949545, + "learning_rate": 1.9544003972543273e-05, + "loss": 1.5013, + "step": 1139 + }, + { + "epoch": 0.6188925081433225, + "grad_norm": 4.619714236951128, + "learning_rate": 1.954295350192859e-05, + "loss": 0.9758, + "step": 1140 + }, + { + "epoch": 0.6194353963083605, + "grad_norm": 5.189787937143953, + "learning_rate": 1.954190185102621e-05, + "loss": 1.2771, + "step": 1141 + }, + { + "epoch": 0.6199782844733985, + "grad_norm": 6.571114312185869, + "learning_rate": 1.9540849019966198e-05, + "loss": 1.2805, + "step": 1142 + }, + { + "epoch": 0.6205211726384365, + "grad_norm": 5.6343515388196455, + "learning_rate": 1.9539795008878774e-05, + "loss": 1.4061, + "step": 1143 + }, + { + "epoch": 0.6210640608034745, + "grad_norm": 5.582702517092888, + "learning_rate": 1.9538739817894302e-05, + "loss": 1.613, + "step": 1144 + }, + { + "epoch": 0.6216069489685125, + "grad_norm": 6.194616996988272, + "learning_rate": 1.9537683447143287e-05, + "loss": 1.114, + "step": 1145 + }, + { + "epoch": 0.6221498371335505, + "grad_norm": 6.0395907363668515, + "learning_rate": 1.9536625896756377e-05, + "loss": 1.5233, + "step": 1146 + }, + { + "epoch": 0.6226927252985885, + "grad_norm": 4.672511820840566, + "learning_rate": 1.9535567166864382e-05, + "loss": 1.1491, + "step": 1147 + }, + { + "epoch": 0.6232356134636265, + "grad_norm": 4.954133576744497, + "learning_rate": 1.9534507257598244e-05, + "loss": 1.1974, + "step": 1148 + }, + { + "epoch": 0.6237785016286646, + "grad_norm": 4.447205012745297, + "learning_rate": 1.953344616908905e-05, + "loss": 1.019, + "step": 1149 + }, + { + "epoch": 0.6243213897937026, + "grad_norm": 4.284087771325503, + "learning_rate": 1.9532383901468038e-05, + "loss": 1.0814, + "step": 1150 + }, + { + "epoch": 0.6248642779587406, + "grad_norm": 3.576067836847051, + "learning_rate": 1.9531320454866595e-05, + "loss": 1.1025, + "step": 1151 + }, + { + "epoch": 0.6254071661237784, + "grad_norm": 3.6831292979487777, + "learning_rate": 1.9530255829416246e-05, + "loss": 0.967, + "step": 1152 + }, + { + "epoch": 0.6259500542888164, + "grad_norm": 3.842496728015195, + "learning_rate": 1.952919002524867e-05, + "loss": 1.0903, + "step": 1153 + }, + { + "epoch": 0.6264929424538545, + "grad_norm": 4.077633480381933, + "learning_rate": 1.952812304249568e-05, + "loss": 0.9227, + "step": 1154 + }, + { + "epoch": 0.6270358306188925, + "grad_norm": 4.157004955121598, + "learning_rate": 1.952705488128925e-05, + "loss": 0.9427, + "step": 1155 + }, + { + "epoch": 0.6275787187839305, + "grad_norm": 5.267614104282471, + "learning_rate": 1.952598554176149e-05, + "loss": 1.9875, + "step": 1156 + }, + { + "epoch": 0.6281216069489685, + "grad_norm": 3.8938350314376735, + "learning_rate": 1.952491502404465e-05, + "loss": 1.1662, + "step": 1157 + }, + { + "epoch": 0.6286644951140065, + "grad_norm": 4.851409026807682, + "learning_rate": 1.9523843328271144e-05, + "loss": 1.9559, + "step": 1158 + }, + { + "epoch": 0.6292073832790445, + "grad_norm": 4.4104613077358845, + "learning_rate": 1.9522770454573513e-05, + "loss": 0.8011, + "step": 1159 + }, + { + "epoch": 0.6297502714440825, + "grad_norm": 3.6883335641947856, + "learning_rate": 1.952169640308446e-05, + "loss": 1.1838, + "step": 1160 + }, + { + "epoch": 0.6302931596091205, + "grad_norm": 3.728203039459734, + "learning_rate": 1.9520621173936818e-05, + "loss": 1.1423, + "step": 1161 + }, + { + "epoch": 0.6308360477741585, + "grad_norm": 3.6272612276564296, + "learning_rate": 1.9519544767263574e-05, + "loss": 0.8376, + "step": 1162 + }, + { + "epoch": 0.6313789359391965, + "grad_norm": 3.928276333701112, + "learning_rate": 1.951846718319786e-05, + "loss": 1.0749, + "step": 1163 + }, + { + "epoch": 0.6319218241042345, + "grad_norm": 5.941949883397345, + "learning_rate": 1.951738842187296e-05, + "loss": 1.8535, + "step": 1164 + }, + { + "epoch": 0.6324647122692725, + "grad_norm": 3.4741200080360644, + "learning_rate": 1.951630848342229e-05, + "loss": 0.7537, + "step": 1165 + }, + { + "epoch": 0.6330076004343105, + "grad_norm": 5.392480994365815, + "learning_rate": 1.9515227367979416e-05, + "loss": 0.8776, + "step": 1166 + }, + { + "epoch": 0.6335504885993485, + "grad_norm": 5.203169892005258, + "learning_rate": 1.951414507567806e-05, + "loss": 1.2678, + "step": 1167 + }, + { + "epoch": 0.6340933767643865, + "grad_norm": 7.499226723318992, + "learning_rate": 1.9513061606652076e-05, + "loss": 1.3295, + "step": 1168 + }, + { + "epoch": 0.6346362649294245, + "grad_norm": 4.355172848230307, + "learning_rate": 1.9511976961035474e-05, + "loss": 0.9188, + "step": 1169 + }, + { + "epoch": 0.6351791530944625, + "grad_norm": 4.983220943925928, + "learning_rate": 1.9510891138962398e-05, + "loss": 1.1951, + "step": 1170 + }, + { + "epoch": 0.6357220412595005, + "grad_norm": 5.3098934312394395, + "learning_rate": 1.950980414056715e-05, + "loss": 0.855, + "step": 1171 + }, + { + "epoch": 0.6362649294245385, + "grad_norm": 5.125683116898323, + "learning_rate": 1.950871596598417e-05, + "loss": 1.1371, + "step": 1172 + }, + { + "epoch": 0.6368078175895765, + "grad_norm": 4.3010814264850215, + "learning_rate": 1.950762661534804e-05, + "loss": 1.7401, + "step": 1173 + }, + { + "epoch": 0.6373507057546145, + "grad_norm": 5.860625964437982, + "learning_rate": 1.95065360887935e-05, + "loss": 1.3557, + "step": 1174 + }, + { + "epoch": 0.6378935939196525, + "grad_norm": 6.3356570754666155, + "learning_rate": 1.9505444386455426e-05, + "loss": 1.2281, + "step": 1175 + }, + { + "epoch": 0.6384364820846905, + "grad_norm": 4.788366086551355, + "learning_rate": 1.9504351508468842e-05, + "loss": 1.3458, + "step": 1176 + }, + { + "epoch": 0.6389793702497285, + "grad_norm": 4.5553889359557065, + "learning_rate": 1.9503257454968914e-05, + "loss": 0.803, + "step": 1177 + }, + { + "epoch": 0.6395222584147665, + "grad_norm": 4.9063187966995345, + "learning_rate": 1.950216222609096e-05, + "loss": 0.8306, + "step": 1178 + }, + { + "epoch": 0.6400651465798045, + "grad_norm": 4.775138114897886, + "learning_rate": 1.9501065821970435e-05, + "loss": 0.9783, + "step": 1179 + }, + { + "epoch": 0.6406080347448425, + "grad_norm": 4.53672284240203, + "learning_rate": 1.9499968242742948e-05, + "loss": 1.3076, + "step": 1180 + }, + { + "epoch": 0.6411509229098805, + "grad_norm": 4.539860050199487, + "learning_rate": 1.949886948854425e-05, + "loss": 1.4212, + "step": 1181 + }, + { + "epoch": 0.6416938110749185, + "grad_norm": 4.251462432005476, + "learning_rate": 1.9497769559510232e-05, + "loss": 1.0274, + "step": 1182 + }, + { + "epoch": 0.6422366992399565, + "grad_norm": 4.49791289564495, + "learning_rate": 1.9496668455776938e-05, + "loss": 1.7531, + "step": 1183 + }, + { + "epoch": 0.6427795874049945, + "grad_norm": 4.996740483321485, + "learning_rate": 1.9495566177480555e-05, + "loss": 1.201, + "step": 1184 + }, + { + "epoch": 0.6433224755700325, + "grad_norm": 5.347352236818569, + "learning_rate": 1.9494462724757413e-05, + "loss": 1.5723, + "step": 1185 + }, + { + "epoch": 0.6438653637350705, + "grad_norm": 4.2242558589374015, + "learning_rate": 1.9493358097743988e-05, + "loss": 1.8014, + "step": 1186 + }, + { + "epoch": 0.6444082519001085, + "grad_norm": 5.042556201755935, + "learning_rate": 1.9492252296576906e-05, + "loss": 1.5466, + "step": 1187 + }, + { + "epoch": 0.6449511400651465, + "grad_norm": 4.284314238035697, + "learning_rate": 1.949114532139293e-05, + "loss": 0.9515, + "step": 1188 + }, + { + "epoch": 0.6454940282301845, + "grad_norm": 4.699826119048575, + "learning_rate": 1.9490037172328974e-05, + "loss": 1.1677, + "step": 1189 + }, + { + "epoch": 0.6460369163952225, + "grad_norm": 6.035247135554916, + "learning_rate": 1.9488927849522095e-05, + "loss": 1.957, + "step": 1190 + }, + { + "epoch": 0.6465798045602605, + "grad_norm": 4.042004235065528, + "learning_rate": 1.94878173531095e-05, + "loss": 1.2182, + "step": 1191 + }, + { + "epoch": 0.6471226927252985, + "grad_norm": 4.114859597350168, + "learning_rate": 1.948670568322853e-05, + "loss": 1.1787, + "step": 1192 + }, + { + "epoch": 0.6476655808903365, + "grad_norm": 4.380177461001504, + "learning_rate": 1.9485592840016682e-05, + "loss": 1.3912, + "step": 1193 + }, + { + "epoch": 0.6482084690553745, + "grad_norm": 5.160710982722015, + "learning_rate": 1.948447882361159e-05, + "loss": 1.2642, + "step": 1194 + }, + { + "epoch": 0.6487513572204126, + "grad_norm": 3.950160510485583, + "learning_rate": 1.9483363634151046e-05, + "loss": 1.1246, + "step": 1195 + }, + { + "epoch": 0.6492942453854506, + "grad_norm": 4.587449618376974, + "learning_rate": 1.9482247271772974e-05, + "loss": 1.3275, + "step": 1196 + }, + { + "epoch": 0.6498371335504886, + "grad_norm": 4.919831513668322, + "learning_rate": 1.9481129736615445e-05, + "loss": 1.4007, + "step": 1197 + }, + { + "epoch": 0.6503800217155266, + "grad_norm": 4.298954423192263, + "learning_rate": 1.948001102881668e-05, + "loss": 0.8076, + "step": 1198 + }, + { + "epoch": 0.6509229098805646, + "grad_norm": 5.525772014179547, + "learning_rate": 1.9478891148515043e-05, + "loss": 1.4558, + "step": 1199 + }, + { + "epoch": 0.6514657980456026, + "grad_norm": 5.473549669240299, + "learning_rate": 1.947777009584904e-05, + "loss": 1.0976, + "step": 1200 + }, + { + "epoch": 0.6520086862106406, + "grad_norm": 5.360010284326786, + "learning_rate": 1.947664787095733e-05, + "loss": 1.0462, + "step": 1201 + }, + { + "epoch": 0.6525515743756786, + "grad_norm": 4.696691081444868, + "learning_rate": 1.9475524473978705e-05, + "loss": 0.933, + "step": 1202 + }, + { + "epoch": 0.6530944625407166, + "grad_norm": 6.0503072087093965, + "learning_rate": 1.947439990505211e-05, + "loss": 0.9628, + "step": 1203 + }, + { + "epoch": 0.6536373507057546, + "grad_norm": 5.476272854511399, + "learning_rate": 1.9473274164316637e-05, + "loss": 0.9716, + "step": 1204 + }, + { + "epoch": 0.6541802388707926, + "grad_norm": 3.910448715873912, + "learning_rate": 1.9472147251911517e-05, + "loss": 1.0732, + "step": 1205 + }, + { + "epoch": 0.6547231270358306, + "grad_norm": 4.596114022138067, + "learning_rate": 1.9471019167976126e-05, + "loss": 0.9399, + "step": 1206 + }, + { + "epoch": 0.6552660152008686, + "grad_norm": 6.764840749864346, + "learning_rate": 1.946988991264999e-05, + "loss": 2.0529, + "step": 1207 + }, + { + "epoch": 0.6558089033659066, + "grad_norm": 5.217328230116913, + "learning_rate": 1.9468759486072778e-05, + "loss": 1.0477, + "step": 1208 + }, + { + "epoch": 0.6563517915309446, + "grad_norm": 4.03429219354135, + "learning_rate": 1.9467627888384303e-05, + "loss": 1.1103, + "step": 1209 + }, + { + "epoch": 0.6568946796959826, + "grad_norm": 4.301388894222403, + "learning_rate": 1.946649511972452e-05, + "loss": 0.8515, + "step": 1210 + }, + { + "epoch": 0.6574375678610206, + "grad_norm": 5.285446092583105, + "learning_rate": 1.9465361180233536e-05, + "loss": 1.2881, + "step": 1211 + }, + { + "epoch": 0.6579804560260586, + "grad_norm": 5.879897026479874, + "learning_rate": 1.9464226070051593e-05, + "loss": 0.9362, + "step": 1212 + }, + { + "epoch": 0.6585233441910966, + "grad_norm": 5.335257600571856, + "learning_rate": 1.9463089789319083e-05, + "loss": 1.4372, + "step": 1213 + }, + { + "epoch": 0.6590662323561346, + "grad_norm": 5.272070125431448, + "learning_rate": 1.9461952338176552e-05, + "loss": 1.2063, + "step": 1214 + }, + { + "epoch": 0.6596091205211726, + "grad_norm": 5.188765715525815, + "learning_rate": 1.946081371676467e-05, + "loss": 0.8912, + "step": 1215 + }, + { + "epoch": 0.6601520086862106, + "grad_norm": 4.747539421286818, + "learning_rate": 1.9459673925224275e-05, + "loss": 1.0381, + "step": 1216 + }, + { + "epoch": 0.6606948968512486, + "grad_norm": 4.486254331013965, + "learning_rate": 1.945853296369633e-05, + "loss": 1.2584, + "step": 1217 + }, + { + "epoch": 0.6612377850162866, + "grad_norm": 4.730012415000151, + "learning_rate": 1.945739083232195e-05, + "loss": 1.1061, + "step": 1218 + }, + { + "epoch": 0.6617806731813246, + "grad_norm": 5.713566521874704, + "learning_rate": 1.9456247531242405e-05, + "loss": 1.4628, + "step": 1219 + }, + { + "epoch": 0.6623235613463626, + "grad_norm": 6.08134803588137, + "learning_rate": 1.9455103060599093e-05, + "loss": 1.6743, + "step": 1220 + }, + { + "epoch": 0.6628664495114006, + "grad_norm": 4.778151760566848, + "learning_rate": 1.9453957420533562e-05, + "loss": 0.8454, + "step": 1221 + }, + { + "epoch": 0.6634093376764386, + "grad_norm": 5.007171333863363, + "learning_rate": 1.945281061118751e-05, + "loss": 1.079, + "step": 1222 + }, + { + "epoch": 0.6639522258414766, + "grad_norm": 4.034543591270571, + "learning_rate": 1.945166263270278e-05, + "loss": 0.8378, + "step": 1223 + }, + { + "epoch": 0.6644951140065146, + "grad_norm": 4.191899722379318, + "learning_rate": 1.9450513485221352e-05, + "loss": 1.3602, + "step": 1224 + }, + { + "epoch": 0.6650380021715526, + "grad_norm": 6.349672663190246, + "learning_rate": 1.944936316888535e-05, + "loss": 1.861, + "step": 1225 + }, + { + "epoch": 0.6655808903365906, + "grad_norm": 5.991894404952759, + "learning_rate": 1.9448211683837055e-05, + "loss": 1.0931, + "step": 1226 + }, + { + "epoch": 0.6661237785016286, + "grad_norm": 4.405247018509644, + "learning_rate": 1.9447059030218876e-05, + "loss": 1.2239, + "step": 1227 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 4.810361623786047, + "learning_rate": 1.9445905208173387e-05, + "loss": 1.1468, + "step": 1228 + }, + { + "epoch": 0.6672095548317046, + "grad_norm": 4.6967843893557015, + "learning_rate": 1.944475021784328e-05, + "loss": 1.5304, + "step": 1229 + }, + { + "epoch": 0.6677524429967426, + "grad_norm": 5.325911382516086, + "learning_rate": 1.9443594059371417e-05, + "loss": 1.2218, + "step": 1230 + }, + { + "epoch": 0.6682953311617806, + "grad_norm": 4.416437363913103, + "learning_rate": 1.9442436732900788e-05, + "loss": 1.223, + "step": 1231 + }, + { + "epoch": 0.6688382193268186, + "grad_norm": 3.720560087272861, + "learning_rate": 1.9441278238574537e-05, + "loss": 0.8155, + "step": 1232 + }, + { + "epoch": 0.6693811074918566, + "grad_norm": 5.273159297494141, + "learning_rate": 1.9440118576535947e-05, + "loss": 1.1958, + "step": 1233 + }, + { + "epoch": 0.6699239956568946, + "grad_norm": 5.684867015184436, + "learning_rate": 1.9438957746928443e-05, + "loss": 1.2384, + "step": 1234 + }, + { + "epoch": 0.6704668838219326, + "grad_norm": 4.952541482610055, + "learning_rate": 1.9437795749895604e-05, + "loss": 1.337, + "step": 1235 + }, + { + "epoch": 0.6710097719869706, + "grad_norm": 4.788863413585636, + "learning_rate": 1.9436632585581145e-05, + "loss": 0.5182, + "step": 1236 + }, + { + "epoch": 0.6715526601520087, + "grad_norm": 5.723878805619219, + "learning_rate": 1.9435468254128925e-05, + "loss": 1.1444, + "step": 1237 + }, + { + "epoch": 0.6720955483170467, + "grad_norm": 4.3068839361839, + "learning_rate": 1.9434302755682958e-05, + "loss": 1.1286, + "step": 1238 + }, + { + "epoch": 0.6726384364820847, + "grad_norm": 3.7588872381668477, + "learning_rate": 1.9433136090387384e-05, + "loss": 0.9718, + "step": 1239 + }, + { + "epoch": 0.6731813246471227, + "grad_norm": 6.802207677110681, + "learning_rate": 1.9431968258386508e-05, + "loss": 1.2389, + "step": 1240 + }, + { + "epoch": 0.6737242128121607, + "grad_norm": 3.6180158875124735, + "learning_rate": 1.9430799259824766e-05, + "loss": 0.644, + "step": 1241 + }, + { + "epoch": 0.6742671009771987, + "grad_norm": 5.559822163300458, + "learning_rate": 1.9429629094846742e-05, + "loss": 1.1157, + "step": 1242 + }, + { + "epoch": 0.6748099891422367, + "grad_norm": 4.084355750224745, + "learning_rate": 1.942845776359716e-05, + "loss": 0.91, + "step": 1243 + }, + { + "epoch": 0.6753528773072747, + "grad_norm": 3.9402358983627583, + "learning_rate": 1.9427285266220895e-05, + "loss": 0.9409, + "step": 1244 + }, + { + "epoch": 0.6758957654723127, + "grad_norm": 5.093172430302549, + "learning_rate": 1.9426111602862968e-05, + "loss": 1.5905, + "step": 1245 + }, + { + "epoch": 0.6764386536373507, + "grad_norm": 4.02536040473524, + "learning_rate": 1.942493677366853e-05, + "loss": 0.9277, + "step": 1246 + }, + { + "epoch": 0.6769815418023887, + "grad_norm": 4.741533426757492, + "learning_rate": 1.942376077878289e-05, + "loss": 1.1027, + "step": 1247 + }, + { + "epoch": 0.6775244299674267, + "grad_norm": 5.365058237804207, + "learning_rate": 1.9422583618351503e-05, + "loss": 1.072, + "step": 1248 + }, + { + "epoch": 0.6780673181324647, + "grad_norm": 4.479339123691231, + "learning_rate": 1.9421405292519956e-05, + "loss": 1.474, + "step": 1249 + }, + { + "epoch": 0.6786102062975027, + "grad_norm": 7.79086989645888, + "learning_rate": 1.942022580143398e-05, + "loss": 1.6973, + "step": 1250 + }, + { + "epoch": 0.6791530944625407, + "grad_norm": 5.621467578044424, + "learning_rate": 1.9419045145239474e-05, + "loss": 1.5432, + "step": 1251 + }, + { + "epoch": 0.6796959826275787, + "grad_norm": 5.041038447258229, + "learning_rate": 1.9417863324082444e-05, + "loss": 1.2977, + "step": 1252 + }, + { + "epoch": 0.6802388707926167, + "grad_norm": 4.717189339301973, + "learning_rate": 1.941668033810907e-05, + "loss": 0.924, + "step": 1253 + }, + { + "epoch": 0.6807817589576547, + "grad_norm": 5.086833595497465, + "learning_rate": 1.9415496187465667e-05, + "loss": 1.5545, + "step": 1254 + }, + { + "epoch": 0.6813246471226927, + "grad_norm": 4.790403842984273, + "learning_rate": 1.941431087229869e-05, + "loss": 1.2545, + "step": 1255 + }, + { + "epoch": 0.6818675352877307, + "grad_norm": 4.470785518257313, + "learning_rate": 1.9413124392754735e-05, + "loss": 1.2036, + "step": 1256 + }, + { + "epoch": 0.6824104234527687, + "grad_norm": 3.820710537112263, + "learning_rate": 1.9411936748980557e-05, + "loss": 1.0856, + "step": 1257 + }, + { + "epoch": 0.6829533116178067, + "grad_norm": 4.453720544384727, + "learning_rate": 1.9410747941123036e-05, + "loss": 1.1645, + "step": 1258 + }, + { + "epoch": 0.6834961997828447, + "grad_norm": 4.692503519890958, + "learning_rate": 1.9409557969329218e-05, + "loss": 1.1174, + "step": 1259 + }, + { + "epoch": 0.6840390879478827, + "grad_norm": 5.666971506629844, + "learning_rate": 1.940836683374627e-05, + "loss": 1.4904, + "step": 1260 + }, + { + "epoch": 0.6845819761129207, + "grad_norm": 5.468580615919992, + "learning_rate": 1.940717453452152e-05, + "loss": 1.2834, + "step": 1261 + }, + { + "epoch": 0.6851248642779587, + "grad_norm": 6.42319990350808, + "learning_rate": 1.9405981071802425e-05, + "loss": 1.1144, + "step": 1262 + }, + { + "epoch": 0.6856677524429967, + "grad_norm": 6.078674980286373, + "learning_rate": 1.9404786445736607e-05, + "loss": 1.4214, + "step": 1263 + }, + { + "epoch": 0.6862106406080347, + "grad_norm": 4.465631588283012, + "learning_rate": 1.9403590656471806e-05, + "loss": 0.7813, + "step": 1264 + }, + { + "epoch": 0.6867535287730727, + "grad_norm": 5.220582351132811, + "learning_rate": 1.940239370415593e-05, + "loss": 1.2112, + "step": 1265 + }, + { + "epoch": 0.6872964169381107, + "grad_norm": 5.1374760487841895, + "learning_rate": 1.9401195588937014e-05, + "loss": 1.2402, + "step": 1266 + }, + { + "epoch": 0.6878393051031487, + "grad_norm": 5.106543779494583, + "learning_rate": 1.9399996310963243e-05, + "loss": 1.1846, + "step": 1267 + }, + { + "epoch": 0.6883821932681867, + "grad_norm": 3.9521299273320363, + "learning_rate": 1.939879587038295e-05, + "loss": 0.8975, + "step": 1268 + }, + { + "epoch": 0.6889250814332247, + "grad_norm": 5.398700470714175, + "learning_rate": 1.9397594267344604e-05, + "loss": 1.4686, + "step": 1269 + }, + { + "epoch": 0.6894679695982627, + "grad_norm": 5.8287973073589265, + "learning_rate": 1.939639150199682e-05, + "loss": 1.011, + "step": 1270 + }, + { + "epoch": 0.6900108577633007, + "grad_norm": 4.812101240482426, + "learning_rate": 1.9395187574488358e-05, + "loss": 0.948, + "step": 1271 + }, + { + "epoch": 0.6905537459283387, + "grad_norm": 5.828731649028842, + "learning_rate": 1.939398248496813e-05, + "loss": 1.2233, + "step": 1272 + }, + { + "epoch": 0.6910966340933767, + "grad_norm": 4.890080173319348, + "learning_rate": 1.9392776233585167e-05, + "loss": 0.8859, + "step": 1273 + }, + { + "epoch": 0.6916395222584147, + "grad_norm": 4.920508250004042, + "learning_rate": 1.9391568820488674e-05, + "loss": 0.73, + "step": 1274 + }, + { + "epoch": 0.6921824104234527, + "grad_norm": 5.635401553567158, + "learning_rate": 1.9390360245827983e-05, + "loss": 1.1432, + "step": 1275 + }, + { + "epoch": 0.6927252985884907, + "grad_norm": 4.588336191443314, + "learning_rate": 1.9389150509752566e-05, + "loss": 1.3326, + "step": 1276 + }, + { + "epoch": 0.6932681867535287, + "grad_norm": 4.631694372660094, + "learning_rate": 1.9387939612412056e-05, + "loss": 1.3762, + "step": 1277 + }, + { + "epoch": 0.6938110749185668, + "grad_norm": 5.585985863722577, + "learning_rate": 1.938672755395621e-05, + "loss": 1.3606, + "step": 1278 + }, + { + "epoch": 0.6943539630836048, + "grad_norm": 4.764783952643876, + "learning_rate": 1.938551433453494e-05, + "loss": 0.9858, + "step": 1279 + }, + { + "epoch": 0.6948968512486428, + "grad_norm": 4.234252918062867, + "learning_rate": 1.9384299954298297e-05, + "loss": 0.7702, + "step": 1280 + }, + { + "epoch": 0.6954397394136808, + "grad_norm": 4.36117548276201, + "learning_rate": 1.938308441339648e-05, + "loss": 1.1846, + "step": 1281 + }, + { + "epoch": 0.6959826275787188, + "grad_norm": 4.311957696864392, + "learning_rate": 1.938186771197983e-05, + "loss": 0.9934, + "step": 1282 + }, + { + "epoch": 0.6965255157437568, + "grad_norm": 5.117807004985933, + "learning_rate": 1.9380649850198824e-05, + "loss": 1.4474, + "step": 1283 + }, + { + "epoch": 0.6970684039087948, + "grad_norm": 5.695329527120276, + "learning_rate": 1.93794308282041e-05, + "loss": 1.3001, + "step": 1284 + }, + { + "epoch": 0.6976112920738328, + "grad_norm": 4.305244126285458, + "learning_rate": 1.937821064614642e-05, + "loss": 1.1421, + "step": 1285 + }, + { + "epoch": 0.6981541802388708, + "grad_norm": 4.574797635236639, + "learning_rate": 1.93769893041767e-05, + "loss": 1.1104, + "step": 1286 + }, + { + "epoch": 0.6986970684039088, + "grad_norm": 5.051193103711606, + "learning_rate": 1.9375766802446002e-05, + "loss": 1.2858, + "step": 1287 + }, + { + "epoch": 0.6992399565689468, + "grad_norm": 5.002235181102662, + "learning_rate": 1.9374543141105518e-05, + "loss": 1.2931, + "step": 1288 + }, + { + "epoch": 0.6997828447339848, + "grad_norm": 5.259386594746194, + "learning_rate": 1.93733183203066e-05, + "loss": 1.0432, + "step": 1289 + }, + { + "epoch": 0.7003257328990228, + "grad_norm": 3.7218449672815117, + "learning_rate": 1.9372092340200736e-05, + "loss": 0.9913, + "step": 1290 + }, + { + "epoch": 0.7008686210640608, + "grad_norm": 4.45109753373239, + "learning_rate": 1.937086520093955e-05, + "loss": 1.2012, + "step": 1291 + }, + { + "epoch": 0.7014115092290988, + "grad_norm": 4.73529586440727, + "learning_rate": 1.9369636902674823e-05, + "loss": 1.3995, + "step": 1292 + }, + { + "epoch": 0.7019543973941368, + "grad_norm": 3.845443933095641, + "learning_rate": 1.936840744555847e-05, + "loss": 1.0971, + "step": 1293 + }, + { + "epoch": 0.7024972855591748, + "grad_norm": 6.67046211091574, + "learning_rate": 1.9367176829742553e-05, + "loss": 1.5273, + "step": 1294 + }, + { + "epoch": 0.7030401737242128, + "grad_norm": 4.454761961182613, + "learning_rate": 1.9365945055379275e-05, + "loss": 1.1282, + "step": 1295 + }, + { + "epoch": 0.7035830618892508, + "grad_norm": 6.552953195473038, + "learning_rate": 1.936471212262099e-05, + "loss": 2.171, + "step": 1296 + }, + { + "epoch": 0.7041259500542888, + "grad_norm": 4.894073730847946, + "learning_rate": 1.9363478031620182e-05, + "loss": 1.2303, + "step": 1297 + }, + { + "epoch": 0.7046688382193268, + "grad_norm": 3.863202427039748, + "learning_rate": 1.936224278252949e-05, + "loss": 0.8515, + "step": 1298 + }, + { + "epoch": 0.7052117263843648, + "grad_norm": 5.503562663348476, + "learning_rate": 1.9361006375501685e-05, + "loss": 1.1834, + "step": 1299 + }, + { + "epoch": 0.7057546145494028, + "grad_norm": 5.318422186143503, + "learning_rate": 1.9359768810689697e-05, + "loss": 1.3378, + "step": 1300 + }, + { + "epoch": 0.7062975027144408, + "grad_norm": 4.46194766163203, + "learning_rate": 1.9358530088246582e-05, + "loss": 1.3318, + "step": 1301 + }, + { + "epoch": 0.7068403908794788, + "grad_norm": 5.6623901137247765, + "learning_rate": 1.9357290208325552e-05, + "loss": 1.3512, + "step": 1302 + }, + { + "epoch": 0.7073832790445168, + "grad_norm": 5.805010217787694, + "learning_rate": 1.9356049171079957e-05, + "loss": 1.2383, + "step": 1303 + }, + { + "epoch": 0.7079261672095548, + "grad_norm": 4.847713425075124, + "learning_rate": 1.9354806976663286e-05, + "loss": 0.7329, + "step": 1304 + }, + { + "epoch": 0.7084690553745928, + "grad_norm": 4.207200307039602, + "learning_rate": 1.935356362522918e-05, + "loss": 1.0119, + "step": 1305 + }, + { + "epoch": 0.7090119435396308, + "grad_norm": 5.76731204400203, + "learning_rate": 1.9352319116931417e-05, + "loss": 1.7383, + "step": 1306 + }, + { + "epoch": 0.7095548317046688, + "grad_norm": 4.989804434976344, + "learning_rate": 1.935107345192392e-05, + "loss": 1.1958, + "step": 1307 + }, + { + "epoch": 0.7100977198697068, + "grad_norm": 4.6099552434159135, + "learning_rate": 1.9349826630360757e-05, + "loss": 0.9933, + "step": 1308 + }, + { + "epoch": 0.7106406080347448, + "grad_norm": 4.693361205108708, + "learning_rate": 1.9348578652396136e-05, + "loss": 1.3553, + "step": 1309 + }, + { + "epoch": 0.7111834961997828, + "grad_norm": 5.473441077550542, + "learning_rate": 1.9347329518184406e-05, + "loss": 1.2202, + "step": 1310 + }, + { + "epoch": 0.7117263843648208, + "grad_norm": 5.885376904454794, + "learning_rate": 1.9346079227880062e-05, + "loss": 1.3698, + "step": 1311 + }, + { + "epoch": 0.7122692725298588, + "grad_norm": 4.115405735194196, + "learning_rate": 1.9344827781637744e-05, + "loss": 1.3361, + "step": 1312 + }, + { + "epoch": 0.7128121606948968, + "grad_norm": 6.745730370060005, + "learning_rate": 1.9343575179612236e-05, + "loss": 1.6411, + "step": 1313 + }, + { + "epoch": 0.7133550488599348, + "grad_norm": 4.363445628534919, + "learning_rate": 1.9342321421958455e-05, + "loss": 0.792, + "step": 1314 + }, + { + "epoch": 0.7138979370249728, + "grad_norm": 6.167581448099132, + "learning_rate": 1.9341066508831472e-05, + "loss": 1.4264, + "step": 1315 + }, + { + "epoch": 0.7144408251900108, + "grad_norm": 4.973571230650737, + "learning_rate": 1.9339810440386495e-05, + "loss": 0.9769, + "step": 1316 + }, + { + "epoch": 0.7149837133550488, + "grad_norm": 5.2637238951774945, + "learning_rate": 1.933855321677888e-05, + "loss": 1.2478, + "step": 1317 + }, + { + "epoch": 0.7155266015200868, + "grad_norm": 5.2296234862723505, + "learning_rate": 1.9337294838164118e-05, + "loss": 1.3999, + "step": 1318 + }, + { + "epoch": 0.7160694896851248, + "grad_norm": 4.720322767524389, + "learning_rate": 1.9336035304697848e-05, + "loss": 1.3561, + "step": 1319 + }, + { + "epoch": 0.7166123778501629, + "grad_norm": 4.2513159313917015, + "learning_rate": 1.9334774616535854e-05, + "loss": 1.176, + "step": 1320 + }, + { + "epoch": 0.7171552660152009, + "grad_norm": 4.5267349508423225, + "learning_rate": 1.9333512773834057e-05, + "loss": 1.1211, + "step": 1321 + }, + { + "epoch": 0.7176981541802389, + "grad_norm": 5.384909301403834, + "learning_rate": 1.9332249776748523e-05, + "loss": 1.3904, + "step": 1322 + }, + { + "epoch": 0.7182410423452769, + "grad_norm": 4.67839864320759, + "learning_rate": 1.9330985625435468e-05, + "loss": 0.8608, + "step": 1323 + }, + { + "epoch": 0.7187839305103149, + "grad_norm": 5.193826637693982, + "learning_rate": 1.9329720320051233e-05, + "loss": 1.3172, + "step": 1324 + }, + { + "epoch": 0.7193268186753529, + "grad_norm": 4.8605293911870096, + "learning_rate": 1.9328453860752324e-05, + "loss": 1.3134, + "step": 1325 + }, + { + "epoch": 0.7198697068403909, + "grad_norm": 5.685265101122876, + "learning_rate": 1.9327186247695377e-05, + "loss": 1.2036, + "step": 1326 + }, + { + "epoch": 0.7204125950054289, + "grad_norm": 5.159715270999642, + "learning_rate": 1.9325917481037164e-05, + "loss": 1.4898, + "step": 1327 + }, + { + "epoch": 0.7209554831704669, + "grad_norm": 5.793178164657722, + "learning_rate": 1.9324647560934613e-05, + "loss": 1.3909, + "step": 1328 + }, + { + "epoch": 0.7214983713355049, + "grad_norm": 3.7442056178189267, + "learning_rate": 1.9323376487544795e-05, + "loss": 0.774, + "step": 1329 + }, + { + "epoch": 0.7220412595005429, + "grad_norm": 4.962252385233631, + "learning_rate": 1.9322104261024912e-05, + "loss": 1.3528, + "step": 1330 + }, + { + "epoch": 0.7225841476655809, + "grad_norm": 5.4534463408495, + "learning_rate": 1.9320830881532316e-05, + "loss": 1.6754, + "step": 1331 + }, + { + "epoch": 0.7231270358306189, + "grad_norm": 6.301262380555263, + "learning_rate": 1.93195563492245e-05, + "loss": 1.5583, + "step": 1332 + }, + { + "epoch": 0.7236699239956569, + "grad_norm": 3.970618104957748, + "learning_rate": 1.9318280664259103e-05, + "loss": 0.8964, + "step": 1333 + }, + { + "epoch": 0.7242128121606949, + "grad_norm": 3.490105469488415, + "learning_rate": 1.9317003826793904e-05, + "loss": 0.8098, + "step": 1334 + }, + { + "epoch": 0.7247557003257329, + "grad_norm": 5.240119278714349, + "learning_rate": 1.9315725836986822e-05, + "loss": 1.1688, + "step": 1335 + }, + { + "epoch": 0.7252985884907709, + "grad_norm": 4.550130688769963, + "learning_rate": 1.931444669499592e-05, + "loss": 1.2569, + "step": 1336 + }, + { + "epoch": 0.7258414766558089, + "grad_norm": 5.058353172472431, + "learning_rate": 1.9313166400979404e-05, + "loss": 1.3471, + "step": 1337 + }, + { + "epoch": 0.7263843648208469, + "grad_norm": 4.528967594836655, + "learning_rate": 1.931188495509563e-05, + "loss": 1.1831, + "step": 1338 + }, + { + "epoch": 0.7269272529858849, + "grad_norm": 4.493564615513058, + "learning_rate": 1.931060235750308e-05, + "loss": 1.4889, + "step": 1339 + }, + { + "epoch": 0.7274701411509229, + "grad_norm": 5.176344629525538, + "learning_rate": 1.9309318608360392e-05, + "loss": 1.6704, + "step": 1340 + }, + { + "epoch": 0.7280130293159609, + "grad_norm": 6.738195156936204, + "learning_rate": 1.930803370782634e-05, + "loss": 1.2147, + "step": 1341 + }, + { + "epoch": 0.7285559174809989, + "grad_norm": 4.248439922934747, + "learning_rate": 1.9306747656059847e-05, + "loss": 1.1943, + "step": 1342 + }, + { + "epoch": 0.7290988056460369, + "grad_norm": 5.175389864390461, + "learning_rate": 1.930546045321997e-05, + "loss": 1.4587, + "step": 1343 + }, + { + "epoch": 0.7296416938110749, + "grad_norm": 5.19934365393201, + "learning_rate": 1.9304172099465914e-05, + "loss": 0.9433, + "step": 1344 + }, + { + "epoch": 0.7301845819761129, + "grad_norm": 6.18681550522946, + "learning_rate": 1.9302882594957025e-05, + "loss": 1.5576, + "step": 1345 + }, + { + "epoch": 0.7307274701411509, + "grad_norm": 4.311510847746277, + "learning_rate": 1.930159193985279e-05, + "loss": 1.327, + "step": 1346 + }, + { + "epoch": 0.7312703583061889, + "grad_norm": 4.044345899710192, + "learning_rate": 1.9300300134312838e-05, + "loss": 1.2106, + "step": 1347 + }, + { + "epoch": 0.7318132464712269, + "grad_norm": 4.181894752624144, + "learning_rate": 1.929900717849694e-05, + "loss": 0.8703, + "step": 1348 + }, + { + "epoch": 0.7323561346362649, + "grad_norm": 5.37566626429507, + "learning_rate": 1.929771307256502e-05, + "loss": 1.446, + "step": 1349 + }, + { + "epoch": 0.7328990228013029, + "grad_norm": 5.37247658500195, + "learning_rate": 1.9296417816677123e-05, + "loss": 0.9239, + "step": 1350 + }, + { + "epoch": 0.7334419109663409, + "grad_norm": 5.749510968066806, + "learning_rate": 1.929512141099346e-05, + "loss": 1.4799, + "step": 1351 + }, + { + "epoch": 0.7339847991313789, + "grad_norm": 5.079635654383841, + "learning_rate": 1.929382385567436e-05, + "loss": 1.2571, + "step": 1352 + }, + { + "epoch": 0.7345276872964169, + "grad_norm": 4.923400923431618, + "learning_rate": 1.929252515088032e-05, + "loss": 1.1967, + "step": 1353 + }, + { + "epoch": 0.7350705754614549, + "grad_norm": 4.397561354835364, + "learning_rate": 1.9291225296771957e-05, + "loss": 0.9088, + "step": 1354 + }, + { + "epoch": 0.7356134636264929, + "grad_norm": 4.13744065040371, + "learning_rate": 1.9289924293510037e-05, + "loss": 1.023, + "step": 1355 + }, + { + "epoch": 0.7361563517915309, + "grad_norm": 4.266553618249588, + "learning_rate": 1.9288622141255477e-05, + "loss": 1.019, + "step": 1356 + }, + { + "epoch": 0.7366992399565689, + "grad_norm": 4.52588883698767, + "learning_rate": 1.928731884016933e-05, + "loss": 1.0935, + "step": 1357 + }, + { + "epoch": 0.7372421281216069, + "grad_norm": 5.46382933116543, + "learning_rate": 1.9286014390412786e-05, + "loss": 1.2833, + "step": 1358 + }, + { + "epoch": 0.737785016286645, + "grad_norm": 4.88651312061201, + "learning_rate": 1.928470879214718e-05, + "loss": 1.6064, + "step": 1359 + }, + { + "epoch": 0.738327904451683, + "grad_norm": 5.403882256899124, + "learning_rate": 1.9283402045533995e-05, + "loss": 1.1537, + "step": 1360 + }, + { + "epoch": 0.738870792616721, + "grad_norm": 6.381532932652649, + "learning_rate": 1.928209415073485e-05, + "loss": 1.527, + "step": 1361 + }, + { + "epoch": 0.739413680781759, + "grad_norm": 4.795244725313847, + "learning_rate": 1.9280785107911505e-05, + "loss": 1.4696, + "step": 1362 + }, + { + "epoch": 0.739956568946797, + "grad_norm": 5.8579061053451955, + "learning_rate": 1.9279474917225866e-05, + "loss": 1.5566, + "step": 1363 + }, + { + "epoch": 0.740499457111835, + "grad_norm": 4.5018596923036185, + "learning_rate": 1.927816357883998e-05, + "loss": 0.9842, + "step": 1364 + }, + { + "epoch": 0.741042345276873, + "grad_norm": 3.596935443077127, + "learning_rate": 1.927685109291604e-05, + "loss": 0.632, + "step": 1365 + }, + { + "epoch": 0.741585233441911, + "grad_norm": 4.535193817325199, + "learning_rate": 1.9275537459616364e-05, + "loss": 1.332, + "step": 1366 + }, + { + "epoch": 0.742128121606949, + "grad_norm": 6.678844775309246, + "learning_rate": 1.9274222679103437e-05, + "loss": 1.6826, + "step": 1367 + }, + { + "epoch": 0.742671009771987, + "grad_norm": 4.444554760823683, + "learning_rate": 1.927290675153987e-05, + "loss": 0.9597, + "step": 1368 + }, + { + "epoch": 0.743213897937025, + "grad_norm": 5.19917061501543, + "learning_rate": 1.927158967708841e-05, + "loss": 1.4333, + "step": 1369 + }, + { + "epoch": 0.743756786102063, + "grad_norm": 4.742331177279521, + "learning_rate": 1.927027145591197e-05, + "loss": 1.1659, + "step": 1370 + }, + { + "epoch": 0.744299674267101, + "grad_norm": 3.713496752478956, + "learning_rate": 1.926895208817358e-05, + "loss": 0.8688, + "step": 1371 + }, + { + "epoch": 0.744842562432139, + "grad_norm": 4.424173231823696, + "learning_rate": 1.9267631574036417e-05, + "loss": 1.0751, + "step": 1372 + }, + { + "epoch": 0.745385450597177, + "grad_norm": 4.7028219143924055, + "learning_rate": 1.9266309913663815e-05, + "loss": 1.3478, + "step": 1373 + }, + { + "epoch": 0.745928338762215, + "grad_norm": 5.498103807374834, + "learning_rate": 1.9264987107219237e-05, + "loss": 1.168, + "step": 1374 + }, + { + "epoch": 0.746471226927253, + "grad_norm": 4.326699254091958, + "learning_rate": 1.9263663154866285e-05, + "loss": 1.0713, + "step": 1375 + }, + { + "epoch": 0.747014115092291, + "grad_norm": 5.58697502498226, + "learning_rate": 1.926233805676871e-05, + "loss": 1.2474, + "step": 1376 + }, + { + "epoch": 0.747557003257329, + "grad_norm": 6.475832093356038, + "learning_rate": 1.92610118130904e-05, + "loss": 1.2805, + "step": 1377 + }, + { + "epoch": 0.748099891422367, + "grad_norm": 5.194166453296571, + "learning_rate": 1.925968442399539e-05, + "loss": 0.9736, + "step": 1378 + }, + { + "epoch": 0.748642779587405, + "grad_norm": 3.867607548943643, + "learning_rate": 1.9258355889647855e-05, + "loss": 1.2325, + "step": 1379 + }, + { + "epoch": 0.749185667752443, + "grad_norm": 5.602424635230674, + "learning_rate": 1.925702621021211e-05, + "loss": 1.0153, + "step": 1380 + }, + { + "epoch": 0.749728555917481, + "grad_norm": 5.011820196133583, + "learning_rate": 1.9255695385852604e-05, + "loss": 1.1823, + "step": 1381 + }, + { + "epoch": 0.750271444082519, + "grad_norm": 5.238515717778492, + "learning_rate": 1.9254363416733944e-05, + "loss": 1.4279, + "step": 1382 + }, + { + "epoch": 0.750814332247557, + "grad_norm": 5.600089025591521, + "learning_rate": 1.925303030302087e-05, + "loss": 0.9212, + "step": 1383 + }, + { + "epoch": 0.751357220412595, + "grad_norm": 4.318444011047328, + "learning_rate": 1.9251696044878255e-05, + "loss": 0.9623, + "step": 1384 + }, + { + "epoch": 0.751900108577633, + "grad_norm": 5.471464995723797, + "learning_rate": 1.925036064247113e-05, + "loss": 1.5446, + "step": 1385 + }, + { + "epoch": 0.752442996742671, + "grad_norm": 5.153911794581426, + "learning_rate": 1.9249024095964663e-05, + "loss": 1.7223, + "step": 1386 + }, + { + "epoch": 0.752985884907709, + "grad_norm": 3.7202924563691537, + "learning_rate": 1.924768640552415e-05, + "loss": 1.1469, + "step": 1387 + }, + { + "epoch": 0.753528773072747, + "grad_norm": 4.612135079041308, + "learning_rate": 1.9246347571315043e-05, + "loss": 0.9572, + "step": 1388 + }, + { + "epoch": 0.754071661237785, + "grad_norm": 4.618898432554816, + "learning_rate": 1.9245007593502937e-05, + "loss": 1.0443, + "step": 1389 + }, + { + "epoch": 0.754614549402823, + "grad_norm": 5.2787203720857345, + "learning_rate": 1.9243666472253554e-05, + "loss": 1.2034, + "step": 1390 + }, + { + "epoch": 0.755157437567861, + "grad_norm": 4.510719578546688, + "learning_rate": 1.9242324207732766e-05, + "loss": 0.8055, + "step": 1391 + }, + { + "epoch": 0.755700325732899, + "grad_norm": 6.256234902874701, + "learning_rate": 1.9240980800106596e-05, + "loss": 1.5985, + "step": 1392 + }, + { + "epoch": 0.756243213897937, + "grad_norm": 9.01213198061367, + "learning_rate": 1.923963624954119e-05, + "loss": 1.7921, + "step": 1393 + }, + { + "epoch": 0.756786102062975, + "grad_norm": 5.149498439254725, + "learning_rate": 1.923829055620285e-05, + "loss": 1.2406, + "step": 1394 + }, + { + "epoch": 0.757328990228013, + "grad_norm": 5.083183084254609, + "learning_rate": 1.9236943720258007e-05, + "loss": 0.9887, + "step": 1395 + }, + { + "epoch": 0.757871878393051, + "grad_norm": 4.5118989088500685, + "learning_rate": 1.9235595741873247e-05, + "loss": 1.3528, + "step": 1396 + }, + { + "epoch": 0.758414766558089, + "grad_norm": 4.583236692880461, + "learning_rate": 1.923424662121528e-05, + "loss": 1.0185, + "step": 1397 + }, + { + "epoch": 0.758957654723127, + "grad_norm": 5.3602373686113625, + "learning_rate": 1.9232896358450976e-05, + "loss": 1.1827, + "step": 1398 + }, + { + "epoch": 0.759500542888165, + "grad_norm": 4.510149132944334, + "learning_rate": 1.9231544953747336e-05, + "loss": 0.9981, + "step": 1399 + }, + { + "epoch": 0.760043431053203, + "grad_norm": 5.008678356958532, + "learning_rate": 1.9230192407271506e-05, + "loss": 1.4957, + "step": 1400 + }, + { + "epoch": 0.760586319218241, + "grad_norm": 4.6938647576746995, + "learning_rate": 1.9228838719190765e-05, + "loss": 1.361, + "step": 1401 + }, + { + "epoch": 0.761129207383279, + "grad_norm": 5.263462060803471, + "learning_rate": 1.9227483889672544e-05, + "loss": 1.1716, + "step": 1402 + }, + { + "epoch": 0.761672095548317, + "grad_norm": 4.737410394333335, + "learning_rate": 1.9226127918884407e-05, + "loss": 1.3924, + "step": 1403 + }, + { + "epoch": 0.762214983713355, + "grad_norm": 5.117476933198257, + "learning_rate": 1.9224770806994066e-05, + "loss": 1.1215, + "step": 1404 + }, + { + "epoch": 0.7627578718783931, + "grad_norm": 4.525492833460315, + "learning_rate": 1.922341255416937e-05, + "loss": 1.0626, + "step": 1405 + }, + { + "epoch": 0.7633007600434311, + "grad_norm": 4.928417801176309, + "learning_rate": 1.9222053160578312e-05, + "loss": 0.9576, + "step": 1406 + }, + { + "epoch": 0.7638436482084691, + "grad_norm": 4.44076437857558, + "learning_rate": 1.9220692626389018e-05, + "loss": 0.9186, + "step": 1407 + }, + { + "epoch": 0.7643865363735071, + "grad_norm": 4.779705057681976, + "learning_rate": 1.9219330951769763e-05, + "loss": 1.3392, + "step": 1408 + }, + { + "epoch": 0.7649294245385451, + "grad_norm": 5.37856562129718, + "learning_rate": 1.9217968136888965e-05, + "loss": 1.043, + "step": 1409 + }, + { + "epoch": 0.7654723127035831, + "grad_norm": 5.596447438931628, + "learning_rate": 1.9216604181915178e-05, + "loss": 1.3223, + "step": 1410 + }, + { + "epoch": 0.7660152008686211, + "grad_norm": 5.267245970848837, + "learning_rate": 1.9215239087017093e-05, + "loss": 1.0484, + "step": 1411 + }, + { + "epoch": 0.7665580890336591, + "grad_norm": 6.124227707475327, + "learning_rate": 1.9213872852363552e-05, + "loss": 1.5361, + "step": 1412 + }, + { + "epoch": 0.7671009771986971, + "grad_norm": 5.632804630433347, + "learning_rate": 1.9212505478123532e-05, + "loss": 1.2227, + "step": 1413 + }, + { + "epoch": 0.7676438653637351, + "grad_norm": 4.903635376409972, + "learning_rate": 1.9211136964466152e-05, + "loss": 0.9045, + "step": 1414 + }, + { + "epoch": 0.7681867535287731, + "grad_norm": 4.995962297168909, + "learning_rate": 1.9209767311560673e-05, + "loss": 1.2364, + "step": 1415 + }, + { + "epoch": 0.7687296416938111, + "grad_norm": 4.76685077713632, + "learning_rate": 1.9208396519576494e-05, + "loss": 1.4849, + "step": 1416 + }, + { + "epoch": 0.7692725298588491, + "grad_norm": 5.66289239913894, + "learning_rate": 1.9207024588683158e-05, + "loss": 1.389, + "step": 1417 + }, + { + "epoch": 0.7698154180238871, + "grad_norm": 4.044195406366437, + "learning_rate": 1.920565151905035e-05, + "loss": 0.5736, + "step": 1418 + }, + { + "epoch": 0.7703583061889251, + "grad_norm": 5.610527750585898, + "learning_rate": 1.9204277310847887e-05, + "loss": 1.5147, + "step": 1419 + }, + { + "epoch": 0.7709011943539631, + "grad_norm": 3.8629108074125424, + "learning_rate": 1.9202901964245734e-05, + "loss": 0.9184, + "step": 1420 + }, + { + "epoch": 0.7714440825190011, + "grad_norm": 4.582445031278247, + "learning_rate": 1.9201525479414e-05, + "loss": 1.071, + "step": 1421 + }, + { + "epoch": 0.7719869706840391, + "grad_norm": 5.016846104390101, + "learning_rate": 1.9200147856522933e-05, + "loss": 1.3673, + "step": 1422 + }, + { + "epoch": 0.7725298588490771, + "grad_norm": 4.798189213061551, + "learning_rate": 1.9198769095742914e-05, + "loss": 1.3483, + "step": 1423 + }, + { + "epoch": 0.7730727470141151, + "grad_norm": 5.871902023790772, + "learning_rate": 1.9197389197244473e-05, + "loss": 1.7625, + "step": 1424 + }, + { + "epoch": 0.7736156351791531, + "grad_norm": 5.895934775040147, + "learning_rate": 1.9196008161198277e-05, + "loss": 0.999, + "step": 1425 + }, + { + "epoch": 0.7741585233441911, + "grad_norm": 5.0199402476408155, + "learning_rate": 1.9194625987775138e-05, + "loss": 1.3251, + "step": 1426 + }, + { + "epoch": 0.7747014115092291, + "grad_norm": 6.650397672217608, + "learning_rate": 1.9193242677146e-05, + "loss": 1.2162, + "step": 1427 + }, + { + "epoch": 0.7752442996742671, + "grad_norm": 5.515267491505962, + "learning_rate": 1.9191858229481958e-05, + "loss": 1.3849, + "step": 1428 + }, + { + "epoch": 0.7757871878393051, + "grad_norm": 5.694611687374825, + "learning_rate": 1.9190472644954236e-05, + "loss": 1.0831, + "step": 1429 + }, + { + "epoch": 0.7763300760043431, + "grad_norm": 4.548114219835821, + "learning_rate": 1.9189085923734215e-05, + "loss": 1.2549, + "step": 1430 + }, + { + "epoch": 0.7768729641693811, + "grad_norm": 6.119997613777156, + "learning_rate": 1.9187698065993398e-05, + "loss": 1.6137, + "step": 1431 + }, + { + "epoch": 0.7774158523344191, + "grad_norm": 4.574150272616086, + "learning_rate": 1.9186309071903445e-05, + "loss": 1.3015, + "step": 1432 + }, + { + "epoch": 0.7779587404994571, + "grad_norm": 4.77854731853541, + "learning_rate": 1.9184918941636142e-05, + "loss": 0.6973, + "step": 1433 + }, + { + "epoch": 0.7785016286644951, + "grad_norm": 6.926122760031406, + "learning_rate": 1.9183527675363425e-05, + "loss": 1.4034, + "step": 1434 + }, + { + "epoch": 0.7790445168295331, + "grad_norm": 4.6748890605309645, + "learning_rate": 1.9182135273257372e-05, + "loss": 0.9854, + "step": 1435 + }, + { + "epoch": 0.7795874049945711, + "grad_norm": 5.072338861625223, + "learning_rate": 1.9180741735490194e-05, + "loss": 1.1604, + "step": 1436 + }, + { + "epoch": 0.7801302931596091, + "grad_norm": 5.995183838581222, + "learning_rate": 1.9179347062234245e-05, + "loss": 1.3913, + "step": 1437 + }, + { + "epoch": 0.7806731813246471, + "grad_norm": 5.494006248546126, + "learning_rate": 1.917795125366202e-05, + "loss": 1.2541, + "step": 1438 + }, + { + "epoch": 0.7812160694896851, + "grad_norm": 3.8223879163574694, + "learning_rate": 1.917655430994616e-05, + "loss": 0.7292, + "step": 1439 + }, + { + "epoch": 0.7817589576547231, + "grad_norm": 4.1476284953657405, + "learning_rate": 1.9175156231259434e-05, + "loss": 0.608, + "step": 1440 + }, + { + "epoch": 0.7823018458197611, + "grad_norm": 5.5144637244676495, + "learning_rate": 1.9173757017774764e-05, + "loss": 1.2674, + "step": 1441 + }, + { + "epoch": 0.7828447339847991, + "grad_norm": 5.4660654663594945, + "learning_rate": 1.9172356669665206e-05, + "loss": 1.3043, + "step": 1442 + }, + { + "epoch": 0.7833876221498371, + "grad_norm": 3.526312966998694, + "learning_rate": 1.9170955187103957e-05, + "loss": 0.8721, + "step": 1443 + }, + { + "epoch": 0.7839305103148752, + "grad_norm": 4.395429767339641, + "learning_rate": 1.9169552570264355e-05, + "loss": 0.985, + "step": 1444 + }, + { + "epoch": 0.7844733984799132, + "grad_norm": 4.608635000415997, + "learning_rate": 1.9168148819319874e-05, + "loss": 1.3492, + "step": 1445 + }, + { + "epoch": 0.7850162866449512, + "grad_norm": 5.729965497961937, + "learning_rate": 1.9166743934444137e-05, + "loss": 1.3405, + "step": 1446 + }, + { + "epoch": 0.7855591748099892, + "grad_norm": 6.60908414105679, + "learning_rate": 1.91653379158109e-05, + "loss": 1.5271, + "step": 1447 + }, + { + "epoch": 0.7861020629750272, + "grad_norm": 4.91386728282805, + "learning_rate": 1.916393076359406e-05, + "loss": 1.475, + "step": 1448 + }, + { + "epoch": 0.7866449511400652, + "grad_norm": 5.286158051966196, + "learning_rate": 1.916252247796766e-05, + "loss": 1.5235, + "step": 1449 + }, + { + "epoch": 0.7871878393051032, + "grad_norm": 4.699459928019414, + "learning_rate": 1.916111305910588e-05, + "loss": 1.151, + "step": 1450 + }, + { + "epoch": 0.7877307274701412, + "grad_norm": 5.125659016648778, + "learning_rate": 1.915970250718303e-05, + "loss": 1.2952, + "step": 1451 + }, + { + "epoch": 0.7882736156351792, + "grad_norm": 4.976170683421487, + "learning_rate": 1.915829082237358e-05, + "loss": 1.3291, + "step": 1452 + }, + { + "epoch": 0.7888165038002172, + "grad_norm": 5.237853683905863, + "learning_rate": 1.9156878004852123e-05, + "loss": 1.4775, + "step": 1453 + }, + { + "epoch": 0.7893593919652552, + "grad_norm": 4.914941394388547, + "learning_rate": 1.9155464054793404e-05, + "loss": 1.2151, + "step": 1454 + }, + { + "epoch": 0.7899022801302932, + "grad_norm": 5.050785426148085, + "learning_rate": 1.9154048972372293e-05, + "loss": 1.163, + "step": 1455 + }, + { + "epoch": 0.7904451682953312, + "grad_norm": 4.2016259300832255, + "learning_rate": 1.915263275776382e-05, + "loss": 0.9601, + "step": 1456 + }, + { + "epoch": 0.7909880564603692, + "grad_norm": 5.48804064939896, + "learning_rate": 1.915121541114314e-05, + "loss": 1.3026, + "step": 1457 + }, + { + "epoch": 0.7915309446254072, + "grad_norm": 5.230495684608947, + "learning_rate": 1.9149796932685552e-05, + "loss": 1.1923, + "step": 1458 + }, + { + "epoch": 0.7920738327904452, + "grad_norm": 4.788655104859546, + "learning_rate": 1.91483773225665e-05, + "loss": 1.2437, + "step": 1459 + }, + { + "epoch": 0.7926167209554832, + "grad_norm": 4.506216689801701, + "learning_rate": 1.9146956580961556e-05, + "loss": 0.9364, + "step": 1460 + }, + { + "epoch": 0.7931596091205212, + "grad_norm": 4.2335852812311865, + "learning_rate": 1.9145534708046446e-05, + "loss": 0.7104, + "step": 1461 + }, + { + "epoch": 0.7937024972855592, + "grad_norm": 6.204688311211956, + "learning_rate": 1.914411170399703e-05, + "loss": 1.0825, + "step": 1462 + }, + { + "epoch": 0.7942453854505972, + "grad_norm": 3.445489329210515, + "learning_rate": 1.91426875689893e-05, + "loss": 0.9921, + "step": 1463 + }, + { + "epoch": 0.7947882736156352, + "grad_norm": 5.204416925095863, + "learning_rate": 1.9141262303199403e-05, + "loss": 1.3043, + "step": 1464 + }, + { + "epoch": 0.7953311617806732, + "grad_norm": 6.3486214559668985, + "learning_rate": 1.9139835906803612e-05, + "loss": 1.3193, + "step": 1465 + }, + { + "epoch": 0.7958740499457112, + "grad_norm": 5.610159156463615, + "learning_rate": 1.913840837997835e-05, + "loss": 1.0455, + "step": 1466 + }, + { + "epoch": 0.7964169381107492, + "grad_norm": 4.625080334899242, + "learning_rate": 1.913697972290018e-05, + "loss": 0.7981, + "step": 1467 + }, + { + "epoch": 0.7969598262757872, + "grad_norm": 5.993005746484773, + "learning_rate": 1.9135549935745792e-05, + "loss": 1.1674, + "step": 1468 + }, + { + "epoch": 0.7975027144408252, + "grad_norm": 5.084136125969368, + "learning_rate": 1.913411901869203e-05, + "loss": 1.0362, + "step": 1469 + }, + { + "epoch": 0.7980456026058632, + "grad_norm": 5.701736254232889, + "learning_rate": 1.913268697191587e-05, + "loss": 1.2159, + "step": 1470 + }, + { + "epoch": 0.7985884907709012, + "grad_norm": 4.715190322082246, + "learning_rate": 1.9131253795594428e-05, + "loss": 0.9848, + "step": 1471 + }, + { + "epoch": 0.7991313789359392, + "grad_norm": 5.375233653105075, + "learning_rate": 1.9129819489904964e-05, + "loss": 1.0476, + "step": 1472 + }, + { + "epoch": 0.7996742671009772, + "grad_norm": 6.352674291337691, + "learning_rate": 1.9128384055024874e-05, + "loss": 1.2362, + "step": 1473 + }, + { + "epoch": 0.8002171552660152, + "grad_norm": 5.46716259791096, + "learning_rate": 1.91269474911317e-05, + "loss": 1.2949, + "step": 1474 + }, + { + "epoch": 0.8007600434310532, + "grad_norm": 5.3150808603597826, + "learning_rate": 1.912550979840311e-05, + "loss": 1.1587, + "step": 1475 + }, + { + "epoch": 0.8013029315960912, + "grad_norm": 4.46814887402293, + "learning_rate": 1.9124070977016926e-05, + "loss": 0.9649, + "step": 1476 + }, + { + "epoch": 0.8018458197611292, + "grad_norm": 4.916267658604107, + "learning_rate": 1.9122631027151103e-05, + "loss": 1.2117, + "step": 1477 + }, + { + "epoch": 0.8023887079261672, + "grad_norm": 5.506988713852874, + "learning_rate": 1.9121189948983733e-05, + "loss": 1.3387, + "step": 1478 + }, + { + "epoch": 0.8029315960912052, + "grad_norm": 4.7704152930487895, + "learning_rate": 1.911974774269305e-05, + "loss": 1.3379, + "step": 1479 + }, + { + "epoch": 0.8034744842562432, + "grad_norm": 5.667769720352476, + "learning_rate": 1.9118304408457435e-05, + "loss": 1.0552, + "step": 1480 + }, + { + "epoch": 0.8040173724212812, + "grad_norm": 4.9471232370904925, + "learning_rate": 1.91168599464554e-05, + "loss": 1.0013, + "step": 1481 + }, + { + "epoch": 0.8045602605863192, + "grad_norm": 4.856623754785127, + "learning_rate": 1.9115414356865594e-05, + "loss": 0.9001, + "step": 1482 + }, + { + "epoch": 0.8051031487513572, + "grad_norm": 4.004976425614515, + "learning_rate": 1.9113967639866815e-05, + "loss": 0.8114, + "step": 1483 + }, + { + "epoch": 0.8056460369163952, + "grad_norm": 5.983237168687411, + "learning_rate": 1.911251979563799e-05, + "loss": 1.1678, + "step": 1484 + }, + { + "epoch": 0.8061889250814332, + "grad_norm": 4.808701495369871, + "learning_rate": 1.9111070824358196e-05, + "loss": 0.9181, + "step": 1485 + }, + { + "epoch": 0.8067318132464713, + "grad_norm": 5.080557830592386, + "learning_rate": 1.910962072620664e-05, + "loss": 1.0751, + "step": 1486 + }, + { + "epoch": 0.8072747014115093, + "grad_norm": 5.730412968009966, + "learning_rate": 1.9108169501362674e-05, + "loss": 1.2727, + "step": 1487 + }, + { + "epoch": 0.8078175895765473, + "grad_norm": 6.47838124684283, + "learning_rate": 1.9106717150005785e-05, + "loss": 1.6491, + "step": 1488 + }, + { + "epoch": 0.8083604777415853, + "grad_norm": 5.311235841870027, + "learning_rate": 1.910526367231561e-05, + "loss": 0.8382, + "step": 1489 + }, + { + "epoch": 0.8089033659066233, + "grad_norm": 5.43072146168114, + "learning_rate": 1.9103809068471914e-05, + "loss": 1.3026, + "step": 1490 + }, + { + "epoch": 0.8094462540716613, + "grad_norm": 4.46156225654375, + "learning_rate": 1.9102353338654597e-05, + "loss": 0.8071, + "step": 1491 + }, + { + "epoch": 0.8099891422366993, + "grad_norm": 5.297520452299967, + "learning_rate": 1.9100896483043714e-05, + "loss": 1.0625, + "step": 1492 + }, + { + "epoch": 0.8105320304017373, + "grad_norm": 6.736735698357962, + "learning_rate": 1.909943850181945e-05, + "loss": 1.6497, + "step": 1493 + }, + { + "epoch": 0.8110749185667753, + "grad_norm": 4.8675910913183955, + "learning_rate": 1.9097979395162132e-05, + "loss": 1.0822, + "step": 1494 + }, + { + "epoch": 0.8116178067318133, + "grad_norm": 4.5489525578867305, + "learning_rate": 1.909651916325222e-05, + "loss": 0.7908, + "step": 1495 + }, + { + "epoch": 0.8121606948968513, + "grad_norm": 6.002121574887706, + "learning_rate": 1.909505780627032e-05, + "loss": 0.5178, + "step": 1496 + }, + { + "epoch": 0.8127035830618893, + "grad_norm": 4.34564271493731, + "learning_rate": 1.9093595324397175e-05, + "loss": 0.9818, + "step": 1497 + }, + { + "epoch": 0.8132464712269273, + "grad_norm": 5.0382562486474525, + "learning_rate": 1.9092131717813668e-05, + "loss": 1.0997, + "step": 1498 + }, + { + "epoch": 0.8137893593919653, + "grad_norm": 6.8289191133749, + "learning_rate": 1.909066698670082e-05, + "loss": 1.0046, + "step": 1499 + }, + { + "epoch": 0.8143322475570033, + "grad_norm": 6.6968746410007585, + "learning_rate": 1.908920113123979e-05, + "loss": 1.5191, + "step": 1500 + }, + { + "epoch": 0.8148751357220413, + "grad_norm": 5.548259576405749, + "learning_rate": 1.9087734151611877e-05, + "loss": 1.3272, + "step": 1501 + }, + { + "epoch": 0.8154180238870793, + "grad_norm": 5.684602861026239, + "learning_rate": 1.9086266047998522e-05, + "loss": 1.0698, + "step": 1502 + }, + { + "epoch": 0.8159609120521173, + "grad_norm": 6.832627276636624, + "learning_rate": 1.90847968205813e-05, + "loss": 1.5796, + "step": 1503 + }, + { + "epoch": 0.8165038002171553, + "grad_norm": 5.605724679057613, + "learning_rate": 1.908332646954193e-05, + "loss": 1.1709, + "step": 1504 + }, + { + "epoch": 0.8170466883821933, + "grad_norm": 4.607345782065147, + "learning_rate": 1.908185499506226e-05, + "loss": 0.8686, + "step": 1505 + }, + { + "epoch": 0.8175895765472313, + "grad_norm": 4.687252800873217, + "learning_rate": 1.9080382397324296e-05, + "loss": 0.8881, + "step": 1506 + }, + { + "epoch": 0.8181324647122693, + "grad_norm": 6.357714385620256, + "learning_rate": 1.907890867651016e-05, + "loss": 1.1948, + "step": 1507 + }, + { + "epoch": 0.8186753528773073, + "grad_norm": 4.8608334992124425, + "learning_rate": 1.9077433832802135e-05, + "loss": 1.0311, + "step": 1508 + }, + { + "epoch": 0.8192182410423453, + "grad_norm": 5.948864135911491, + "learning_rate": 1.9075957866382623e-05, + "loss": 1.1314, + "step": 1509 + }, + { + "epoch": 0.8197611292073833, + "grad_norm": 4.807850657807276, + "learning_rate": 1.9074480777434178e-05, + "loss": 0.9478, + "step": 1510 + }, + { + "epoch": 0.8203040173724213, + "grad_norm": 4.26222909817726, + "learning_rate": 1.9073002566139486e-05, + "loss": 0.8541, + "step": 1511 + }, + { + "epoch": 0.8208469055374593, + "grad_norm": 6.6130092616391005, + "learning_rate": 1.9071523232681382e-05, + "loss": 1.2754, + "step": 1512 + }, + { + "epoch": 0.8213897937024973, + "grad_norm": 5.255155673043404, + "learning_rate": 1.907004277724282e-05, + "loss": 1.0813, + "step": 1513 + }, + { + "epoch": 0.8219326818675353, + "grad_norm": 5.079098182105948, + "learning_rate": 1.9068561200006917e-05, + "loss": 1.0016, + "step": 1514 + }, + { + "epoch": 0.8224755700325733, + "grad_norm": 5.491172879434626, + "learning_rate": 1.906707850115691e-05, + "loss": 1.2884, + "step": 1515 + }, + { + "epoch": 0.8230184581976113, + "grad_norm": 6.718962166599785, + "learning_rate": 1.9065594680876182e-05, + "loss": 1.4973, + "step": 1516 + }, + { + "epoch": 0.8235613463626493, + "grad_norm": 5.348428262646105, + "learning_rate": 1.9064109739348257e-05, + "loss": 1.1113, + "step": 1517 + }, + { + "epoch": 0.8241042345276873, + "grad_norm": 4.604059650726469, + "learning_rate": 1.906262367675679e-05, + "loss": 1.1614, + "step": 1518 + }, + { + "epoch": 0.8246471226927253, + "grad_norm": 5.895860662978225, + "learning_rate": 1.9061136493285586e-05, + "loss": 1.1532, + "step": 1519 + }, + { + "epoch": 0.8251900108577633, + "grad_norm": 5.621847434524929, + "learning_rate": 1.905964818911858e-05, + "loss": 0.9277, + "step": 1520 + }, + { + "epoch": 0.8257328990228013, + "grad_norm": 4.585955492662189, + "learning_rate": 1.9058158764439844e-05, + "loss": 0.7988, + "step": 1521 + }, + { + "epoch": 0.8262757871878393, + "grad_norm": 5.784308925988881, + "learning_rate": 1.9056668219433595e-05, + "loss": 1.6078, + "step": 1522 + }, + { + "epoch": 0.8268186753528773, + "grad_norm": 5.621216253388429, + "learning_rate": 1.905517655428419e-05, + "loss": 1.389, + "step": 1523 + }, + { + "epoch": 0.8273615635179153, + "grad_norm": 5.514208208061458, + "learning_rate": 1.9053683769176115e-05, + "loss": 0.7612, + "step": 1524 + }, + { + "epoch": 0.8279044516829533, + "grad_norm": 4.6572185133859065, + "learning_rate": 1.9052189864294002e-05, + "loss": 0.744, + "step": 1525 + }, + { + "epoch": 0.8284473398479913, + "grad_norm": 5.239632270223703, + "learning_rate": 1.905069483982262e-05, + "loss": 0.939, + "step": 1526 + }, + { + "epoch": 0.8289902280130294, + "grad_norm": 3.7037650901217454, + "learning_rate": 1.9049198695946876e-05, + "loss": 1.0177, + "step": 1527 + }, + { + "epoch": 0.8295331161780674, + "grad_norm": 6.974869726679236, + "learning_rate": 1.9047701432851813e-05, + "loss": 1.2722, + "step": 1528 + }, + { + "epoch": 0.8300760043431054, + "grad_norm": 6.642878483620589, + "learning_rate": 1.904620305072262e-05, + "loss": 1.5369, + "step": 1529 + }, + { + "epoch": 0.8306188925081434, + "grad_norm": 7.99082645392899, + "learning_rate": 1.9044703549744616e-05, + "loss": 1.2245, + "step": 1530 + }, + { + "epoch": 0.8311617806731814, + "grad_norm": 6.5593948883008135, + "learning_rate": 1.904320293010326e-05, + "loss": 1.307, + "step": 1531 + }, + { + "epoch": 0.8317046688382194, + "grad_norm": 4.930764759519961, + "learning_rate": 1.9041701191984155e-05, + "loss": 0.9564, + "step": 1532 + }, + { + "epoch": 0.8322475570032574, + "grad_norm": 3.9621684331427773, + "learning_rate": 1.9040198335573033e-05, + "loss": 0.8153, + "step": 1533 + }, + { + "epoch": 0.8327904451682954, + "grad_norm": 4.859084711241092, + "learning_rate": 1.9038694361055774e-05, + "loss": 1.0967, + "step": 1534 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.9476278172327595, + "learning_rate": 1.903718926861839e-05, + "loss": 1.4743, + "step": 1535 + }, + { + "epoch": 0.8338762214983714, + "grad_norm": 4.312795088472635, + "learning_rate": 1.903568305844704e-05, + "loss": 0.8249, + "step": 1536 + }, + { + "epoch": 0.8344191096634094, + "grad_norm": 6.638027572281857, + "learning_rate": 1.9034175730728e-05, + "loss": 1.6353, + "step": 1537 + }, + { + "epoch": 0.8349619978284474, + "grad_norm": 5.709548912096025, + "learning_rate": 1.9032667285647714e-05, + "loss": 1.8512, + "step": 1538 + }, + { + "epoch": 0.8355048859934854, + "grad_norm": 5.098275426179771, + "learning_rate": 1.9031157723392738e-05, + "loss": 1.0348, + "step": 1539 + }, + { + "epoch": 0.8360477741585234, + "grad_norm": 5.05236269448673, + "learning_rate": 1.9029647044149783e-05, + "loss": 1.108, + "step": 1540 + }, + { + "epoch": 0.8365906623235614, + "grad_norm": 4.674708875536442, + "learning_rate": 1.9028135248105692e-05, + "loss": 1.0453, + "step": 1541 + }, + { + "epoch": 0.8371335504885994, + "grad_norm": 4.717321956389267, + "learning_rate": 1.902662233544744e-05, + "loss": 0.9042, + "step": 1542 + }, + { + "epoch": 0.8376764386536374, + "grad_norm": 6.217040367936212, + "learning_rate": 1.9025108306362158e-05, + "loss": 1.0762, + "step": 1543 + }, + { + "epoch": 0.8382193268186754, + "grad_norm": 5.338744288323046, + "learning_rate": 1.9023593161037094e-05, + "loss": 1.1631, + "step": 1544 + }, + { + "epoch": 0.8387622149837134, + "grad_norm": 4.8224874086687874, + "learning_rate": 1.9022076899659643e-05, + "loss": 1.3907, + "step": 1545 + }, + { + "epoch": 0.8393051031487514, + "grad_norm": 4.699800413179793, + "learning_rate": 1.9020559522417345e-05, + "loss": 0.7682, + "step": 1546 + }, + { + "epoch": 0.8398479913137894, + "grad_norm": 4.271439794022252, + "learning_rate": 1.9019041029497866e-05, + "loss": 0.8475, + "step": 1547 + }, + { + "epoch": 0.8403908794788274, + "grad_norm": 4.240562925498168, + "learning_rate": 1.9017521421089022e-05, + "loss": 0.8201, + "step": 1548 + }, + { + "epoch": 0.8409337676438654, + "grad_norm": 5.24718786534657, + "learning_rate": 1.9016000697378755e-05, + "loss": 1.0728, + "step": 1549 + }, + { + "epoch": 0.8414766558089034, + "grad_norm": 6.492220902503762, + "learning_rate": 1.9014478858555156e-05, + "loss": 0.9432, + "step": 1550 + }, + { + "epoch": 0.8420195439739414, + "grad_norm": 4.917811923935393, + "learning_rate": 1.9012955904806438e-05, + "loss": 1.1672, + "step": 1551 + }, + { + "epoch": 0.8425624321389794, + "grad_norm": 5.66709669173795, + "learning_rate": 1.9011431836320976e-05, + "loss": 1.5058, + "step": 1552 + }, + { + "epoch": 0.8431053203040174, + "grad_norm": 5.658634152413846, + "learning_rate": 1.9009906653287258e-05, + "loss": 1.3653, + "step": 1553 + }, + { + "epoch": 0.8436482084690554, + "grad_norm": 4.956250367213818, + "learning_rate": 1.9008380355893925e-05, + "loss": 0.8309, + "step": 1554 + }, + { + "epoch": 0.8441910966340934, + "grad_norm": 4.329497016180362, + "learning_rate": 1.9006852944329753e-05, + "loss": 1.1141, + "step": 1555 + }, + { + "epoch": 0.8447339847991314, + "grad_norm": 5.717864943326053, + "learning_rate": 1.9005324418783658e-05, + "loss": 1.3274, + "step": 1556 + }, + { + "epoch": 0.8452768729641694, + "grad_norm": 5.657300743895006, + "learning_rate": 1.900379477944468e-05, + "loss": 1.2432, + "step": 1557 + }, + { + "epoch": 0.8458197611292074, + "grad_norm": 5.325761071371062, + "learning_rate": 1.900226402650202e-05, + "loss": 1.3428, + "step": 1558 + }, + { + "epoch": 0.8463626492942454, + "grad_norm": 5.207304704126981, + "learning_rate": 1.9000732160144996e-05, + "loss": 1.455, + "step": 1559 + }, + { + "epoch": 0.8469055374592834, + "grad_norm": 5.248656371435267, + "learning_rate": 1.8999199180563074e-05, + "loss": 0.7851, + "step": 1560 + }, + { + "epoch": 0.8474484256243214, + "grad_norm": 6.92723307445887, + "learning_rate": 1.899766508794585e-05, + "loss": 1.5236, + "step": 1561 + }, + { + "epoch": 0.8479913137893594, + "grad_norm": 4.3434986726191935, + "learning_rate": 1.899612988248307e-05, + "loss": 0.9117, + "step": 1562 + }, + { + "epoch": 0.8485342019543974, + "grad_norm": 5.815800316925401, + "learning_rate": 1.8994593564364612e-05, + "loss": 1.0097, + "step": 1563 + }, + { + "epoch": 0.8490770901194354, + "grad_norm": 5.5494362406379905, + "learning_rate": 1.8993056133780484e-05, + "loss": 1.1829, + "step": 1564 + }, + { + "epoch": 0.8496199782844734, + "grad_norm": 4.323494913553237, + "learning_rate": 1.899151759092084e-05, + "loss": 0.9359, + "step": 1565 + }, + { + "epoch": 0.8501628664495114, + "grad_norm": 4.455911135073797, + "learning_rate": 1.898997793597597e-05, + "loss": 0.9354, + "step": 1566 + }, + { + "epoch": 0.8507057546145494, + "grad_norm": 6.684553981467422, + "learning_rate": 1.8988437169136302e-05, + "loss": 1.3693, + "step": 1567 + }, + { + "epoch": 0.8512486427795874, + "grad_norm": 6.341757024960091, + "learning_rate": 1.89868952905924e-05, + "loss": 0.9441, + "step": 1568 + }, + { + "epoch": 0.8517915309446255, + "grad_norm": 6.017605683279906, + "learning_rate": 1.8985352300534965e-05, + "loss": 1.2747, + "step": 1569 + }, + { + "epoch": 0.8523344191096635, + "grad_norm": 5.501152081837135, + "learning_rate": 1.8983808199154835e-05, + "loss": 0.8414, + "step": 1570 + }, + { + "epoch": 0.8528773072747015, + "grad_norm": 5.075164131477861, + "learning_rate": 1.8982262986642993e-05, + "loss": 1.124, + "step": 1571 + }, + { + "epoch": 0.8534201954397395, + "grad_norm": 5.7750932469486065, + "learning_rate": 1.8980716663190545e-05, + "loss": 1.3968, + "step": 1572 + }, + { + "epoch": 0.8539630836047775, + "grad_norm": 6.967872140444835, + "learning_rate": 1.897916922898875e-05, + "loss": 1.2395, + "step": 1573 + }, + { + "epoch": 0.8545059717698155, + "grad_norm": 5.037804725894728, + "learning_rate": 1.8977620684228994e-05, + "loss": 0.9741, + "step": 1574 + }, + { + "epoch": 0.8550488599348535, + "grad_norm": 4.586000206489282, + "learning_rate": 1.8976071029102802e-05, + "loss": 0.6757, + "step": 1575 + }, + { + "epoch": 0.8555917480998915, + "grad_norm": 4.686417605581938, + "learning_rate": 1.897452026380184e-05, + "loss": 0.8382, + "step": 1576 + }, + { + "epoch": 0.8561346362649295, + "grad_norm": 5.133601528061789, + "learning_rate": 1.8972968388517908e-05, + "loss": 1.2999, + "step": 1577 + }, + { + "epoch": 0.8566775244299675, + "grad_norm": 6.170796021401621, + "learning_rate": 1.8971415403442942e-05, + "loss": 1.1513, + "step": 1578 + }, + { + "epoch": 0.8572204125950055, + "grad_norm": 5.229050623246509, + "learning_rate": 1.8969861308769025e-05, + "loss": 0.9371, + "step": 1579 + }, + { + "epoch": 0.8577633007600435, + "grad_norm": 5.081229057143735, + "learning_rate": 1.8968306104688365e-05, + "loss": 0.9398, + "step": 1580 + }, + { + "epoch": 0.8583061889250815, + "grad_norm": 4.641876540865143, + "learning_rate": 1.896674979139331e-05, + "loss": 1.1401, + "step": 1581 + }, + { + "epoch": 0.8588490770901195, + "grad_norm": 4.149586485324207, + "learning_rate": 1.8965192369076356e-05, + "loss": 0.6768, + "step": 1582 + }, + { + "epoch": 0.8593919652551575, + "grad_norm": 6.422772444923944, + "learning_rate": 1.8963633837930114e-05, + "loss": 1.392, + "step": 1583 + }, + { + "epoch": 0.8599348534201955, + "grad_norm": 4.860635992388028, + "learning_rate": 1.8962074198147357e-05, + "loss": 0.9983, + "step": 1584 + }, + { + "epoch": 0.8604777415852335, + "grad_norm": 4.216877618358879, + "learning_rate": 1.8960513449920982e-05, + "loss": 0.6958, + "step": 1585 + }, + { + "epoch": 0.8610206297502715, + "grad_norm": 5.012340070254638, + "learning_rate": 1.8958951593444017e-05, + "loss": 0.9859, + "step": 1586 + }, + { + "epoch": 0.8615635179153095, + "grad_norm": 5.467321056155664, + "learning_rate": 1.8957388628909644e-05, + "loss": 1.3782, + "step": 1587 + }, + { + "epoch": 0.8621064060803475, + "grad_norm": 5.59589685839463, + "learning_rate": 1.8955824556511168e-05, + "loss": 0.8297, + "step": 1588 + }, + { + "epoch": 0.8626492942453855, + "grad_norm": 6.519624998298618, + "learning_rate": 1.895425937644204e-05, + "loss": 1.7264, + "step": 1589 + }, + { + "epoch": 0.8631921824104235, + "grad_norm": 6.6951270307269795, + "learning_rate": 1.8952693088895837e-05, + "loss": 1.2919, + "step": 1590 + }, + { + "epoch": 0.8637350705754615, + "grad_norm": 4.348741825803296, + "learning_rate": 1.895112569406629e-05, + "loss": 1.0062, + "step": 1591 + }, + { + "epoch": 0.8642779587404995, + "grad_norm": 5.0213162819553565, + "learning_rate": 1.8949557192147243e-05, + "loss": 1.2381, + "step": 1592 + }, + { + "epoch": 0.8648208469055375, + "grad_norm": 6.1637021381056885, + "learning_rate": 1.8947987583332705e-05, + "loss": 1.2561, + "step": 1593 + }, + { + "epoch": 0.8653637350705755, + "grad_norm": 6.638257635344733, + "learning_rate": 1.89464168678168e-05, + "loss": 0.9861, + "step": 1594 + }, + { + "epoch": 0.8659066232356135, + "grad_norm": 4.861237127098975, + "learning_rate": 1.89448450457938e-05, + "loss": 0.9407, + "step": 1595 + }, + { + "epoch": 0.8664495114006515, + "grad_norm": 6.4248430682281565, + "learning_rate": 1.894327211745811e-05, + "loss": 1.3079, + "step": 1596 + }, + { + "epoch": 0.8669923995656895, + "grad_norm": 5.712502208347676, + "learning_rate": 1.8941698083004265e-05, + "loss": 1.4422, + "step": 1597 + }, + { + "epoch": 0.8675352877307275, + "grad_norm": 5.634308307822415, + "learning_rate": 1.8940122942626957e-05, + "loss": 0.7869, + "step": 1598 + }, + { + "epoch": 0.8680781758957655, + "grad_norm": 7.839535783401759, + "learning_rate": 1.893854669652099e-05, + "loss": 1.3397, + "step": 1599 + }, + { + "epoch": 0.8686210640608035, + "grad_norm": 5.156814072355032, + "learning_rate": 1.8936969344881323e-05, + "loss": 1.2541, + "step": 1600 + }, + { + "epoch": 0.8691639522258415, + "grad_norm": 4.912293322601371, + "learning_rate": 1.8935390887903044e-05, + "loss": 0.8418, + "step": 1601 + }, + { + "epoch": 0.8697068403908795, + "grad_norm": 5.436641391280079, + "learning_rate": 1.8933811325781382e-05, + "loss": 1.4456, + "step": 1602 + }, + { + "epoch": 0.8702497285559175, + "grad_norm": 5.221596044730403, + "learning_rate": 1.8932230658711696e-05, + "loss": 0.9626, + "step": 1603 + }, + { + "epoch": 0.8707926167209555, + "grad_norm": 6.103653494209826, + "learning_rate": 1.8930648886889482e-05, + "loss": 1.2338, + "step": 1604 + }, + { + "epoch": 0.8713355048859935, + "grad_norm": 7.003639368124227, + "learning_rate": 1.8929066010510383e-05, + "loss": 1.2216, + "step": 1605 + }, + { + "epoch": 0.8718783930510315, + "grad_norm": 4.911905632269173, + "learning_rate": 1.8927482029770168e-05, + "loss": 0.9049, + "step": 1606 + }, + { + "epoch": 0.8724212812160695, + "grad_norm": 6.079178298711795, + "learning_rate": 1.8925896944864748e-05, + "loss": 1.6408, + "step": 1607 + }, + { + "epoch": 0.8729641693811075, + "grad_norm": 6.045568168649525, + "learning_rate": 1.892431075599017e-05, + "loss": 1.084, + "step": 1608 + }, + { + "epoch": 0.8735070575461455, + "grad_norm": 4.827633470123435, + "learning_rate": 1.892272346334261e-05, + "loss": 0.8121, + "step": 1609 + }, + { + "epoch": 0.8740499457111836, + "grad_norm": 5.314535878915851, + "learning_rate": 1.8921135067118396e-05, + "loss": 1.2657, + "step": 1610 + }, + { + "epoch": 0.8745928338762216, + "grad_norm": 4.786606110240068, + "learning_rate": 1.8919545567513976e-05, + "loss": 1.0534, + "step": 1611 + }, + { + "epoch": 0.8751357220412594, + "grad_norm": 4.075700855026623, + "learning_rate": 1.8917954964725948e-05, + "loss": 0.7757, + "step": 1612 + }, + { + "epoch": 0.8756786102062974, + "grad_norm": 6.286980562376595, + "learning_rate": 1.8916363258951033e-05, + "loss": 0.9782, + "step": 1613 + }, + { + "epoch": 0.8762214983713354, + "grad_norm": 6.745303990081763, + "learning_rate": 1.8914770450386102e-05, + "loss": 0.9488, + "step": 1614 + }, + { + "epoch": 0.8767643865363735, + "grad_norm": 6.2663007227218275, + "learning_rate": 1.8913176539228152e-05, + "loss": 1.9529, + "step": 1615 + }, + { + "epoch": 0.8773072747014115, + "grad_norm": 4.793153881320079, + "learning_rate": 1.8911581525674324e-05, + "loss": 1.0242, + "step": 1616 + }, + { + "epoch": 0.8778501628664495, + "grad_norm": 4.6844946413894, + "learning_rate": 1.890998540992189e-05, + "loss": 0.8565, + "step": 1617 + }, + { + "epoch": 0.8783930510314875, + "grad_norm": 3.993737542716379, + "learning_rate": 1.8908388192168256e-05, + "loss": 0.6107, + "step": 1618 + }, + { + "epoch": 0.8789359391965255, + "grad_norm": 6.099718972513945, + "learning_rate": 1.8906789872610977e-05, + "loss": 0.8043, + "step": 1619 + }, + { + "epoch": 0.8794788273615635, + "grad_norm": 7.113418049135383, + "learning_rate": 1.8905190451447726e-05, + "loss": 1.241, + "step": 1620 + }, + { + "epoch": 0.8800217155266015, + "grad_norm": 4.439254236407197, + "learning_rate": 1.8903589928876337e-05, + "loss": 1.0627, + "step": 1621 + }, + { + "epoch": 0.8805646036916395, + "grad_norm": 5.773948255627926, + "learning_rate": 1.8901988305094746e-05, + "loss": 1.3241, + "step": 1622 + }, + { + "epoch": 0.8811074918566775, + "grad_norm": 6.231870498005159, + "learning_rate": 1.890038558030106e-05, + "loss": 1.5241, + "step": 1623 + }, + { + "epoch": 0.8816503800217155, + "grad_norm": 5.888612961801449, + "learning_rate": 1.8898781754693495e-05, + "loss": 1.109, + "step": 1624 + }, + { + "epoch": 0.8821932681867535, + "grad_norm": 4.880160750019388, + "learning_rate": 1.8897176828470424e-05, + "loss": 1.0124, + "step": 1625 + }, + { + "epoch": 0.8827361563517915, + "grad_norm": 6.365487465081988, + "learning_rate": 1.889557080183034e-05, + "loss": 0.8999, + "step": 1626 + }, + { + "epoch": 0.8832790445168295, + "grad_norm": 4.461166628483842, + "learning_rate": 1.8893963674971883e-05, + "loss": 0.8762, + "step": 1627 + }, + { + "epoch": 0.8838219326818675, + "grad_norm": 5.675449614336963, + "learning_rate": 1.8892355448093825e-05, + "loss": 1.049, + "step": 1628 + }, + { + "epoch": 0.8843648208469055, + "grad_norm": 4.320721965444944, + "learning_rate": 1.8890746121395072e-05, + "loss": 1.1291, + "step": 1629 + }, + { + "epoch": 0.8849077090119435, + "grad_norm": 5.155835351166163, + "learning_rate": 1.8889135695074668e-05, + "loss": 1.1035, + "step": 1630 + }, + { + "epoch": 0.8854505971769815, + "grad_norm": 6.506050711845877, + "learning_rate": 1.8887524169331794e-05, + "loss": 2.1954, + "step": 1631 + }, + { + "epoch": 0.8859934853420195, + "grad_norm": 5.514989348863087, + "learning_rate": 1.8885911544365766e-05, + "loss": 1.0237, + "step": 1632 + }, + { + "epoch": 0.8865363735070575, + "grad_norm": 5.123449587076818, + "learning_rate": 1.8884297820376038e-05, + "loss": 0.9908, + "step": 1633 + }, + { + "epoch": 0.8870792616720955, + "grad_norm": 5.421671567850239, + "learning_rate": 1.8882682997562197e-05, + "loss": 1.7734, + "step": 1634 + }, + { + "epoch": 0.8876221498371335, + "grad_norm": 7.1012436118787186, + "learning_rate": 1.8881067076123963e-05, + "loss": 1.5488, + "step": 1635 + }, + { + "epoch": 0.8881650380021715, + "grad_norm": 5.328780061143714, + "learning_rate": 1.88794500562612e-05, + "loss": 1.3832, + "step": 1636 + }, + { + "epoch": 0.8887079261672095, + "grad_norm": 5.1503624666971595, + "learning_rate": 1.88778319381739e-05, + "loss": 0.8524, + "step": 1637 + }, + { + "epoch": 0.8892508143322475, + "grad_norm": 5.131553278195334, + "learning_rate": 1.88762127220622e-05, + "loss": 0.9769, + "step": 1638 + }, + { + "epoch": 0.8897937024972855, + "grad_norm": 3.1519686008526135, + "learning_rate": 1.8874592408126365e-05, + "loss": 0.472, + "step": 1639 + }, + { + "epoch": 0.8903365906623235, + "grad_norm": 4.164586140054025, + "learning_rate": 1.8872970996566794e-05, + "loss": 0.67, + "step": 1640 + }, + { + "epoch": 0.8908794788273615, + "grad_norm": 7.488659246842809, + "learning_rate": 1.8871348487584028e-05, + "loss": 1.3141, + "step": 1641 + }, + { + "epoch": 0.8914223669923995, + "grad_norm": 5.911298710361154, + "learning_rate": 1.8869724881378743e-05, + "loss": 1.084, + "step": 1642 + }, + { + "epoch": 0.8919652551574375, + "grad_norm": 5.977793625957034, + "learning_rate": 1.886810017815175e-05, + "loss": 0.9273, + "step": 1643 + }, + { + "epoch": 0.8925081433224755, + "grad_norm": 5.2773273690434515, + "learning_rate": 1.8866474378103993e-05, + "loss": 1.2594, + "step": 1644 + }, + { + "epoch": 0.8930510314875135, + "grad_norm": 8.049487951903771, + "learning_rate": 1.8864847481436554e-05, + "loss": 1.0104, + "step": 1645 + }, + { + "epoch": 0.8935939196525515, + "grad_norm": 5.428394606766542, + "learning_rate": 1.886321948835065e-05, + "loss": 0.827, + "step": 1646 + }, + { + "epoch": 0.8941368078175895, + "grad_norm": 5.687133062399138, + "learning_rate": 1.8861590399047635e-05, + "loss": 1.2759, + "step": 1647 + }, + { + "epoch": 0.8946796959826275, + "grad_norm": 5.037591954761082, + "learning_rate": 1.885996021372899e-05, + "loss": 0.8928, + "step": 1648 + }, + { + "epoch": 0.8952225841476655, + "grad_norm": 4.911533421640023, + "learning_rate": 1.8858328932596352e-05, + "loss": 0.705, + "step": 1649 + }, + { + "epoch": 0.8957654723127035, + "grad_norm": 6.054608860222289, + "learning_rate": 1.885669655585147e-05, + "loss": 1.296, + "step": 1650 + }, + { + "epoch": 0.8963083604777415, + "grad_norm": 6.221099773279408, + "learning_rate": 1.8855063083696244e-05, + "loss": 1.5253, + "step": 1651 + }, + { + "epoch": 0.8968512486427795, + "grad_norm": 5.0718329796444035, + "learning_rate": 1.8853428516332702e-05, + "loss": 1.1683, + "step": 1652 + }, + { + "epoch": 0.8973941368078175, + "grad_norm": 5.721985007773526, + "learning_rate": 1.8851792853963015e-05, + "loss": 1.1635, + "step": 1653 + }, + { + "epoch": 0.8979370249728555, + "grad_norm": 7.224045168649458, + "learning_rate": 1.8850156096789473e-05, + "loss": 1.0042, + "step": 1654 + }, + { + "epoch": 0.8984799131378935, + "grad_norm": 4.938883702947416, + "learning_rate": 1.8848518245014526e-05, + "loss": 0.7712, + "step": 1655 + }, + { + "epoch": 0.8990228013029316, + "grad_norm": 5.9261619850877, + "learning_rate": 1.8846879298840735e-05, + "loss": 1.4387, + "step": 1656 + }, + { + "epoch": 0.8995656894679696, + "grad_norm": 5.181216185736725, + "learning_rate": 1.8845239258470817e-05, + "loss": 0.9389, + "step": 1657 + }, + { + "epoch": 0.9001085776330076, + "grad_norm": 5.108249236166185, + "learning_rate": 1.8843598124107608e-05, + "loss": 1.0034, + "step": 1658 + }, + { + "epoch": 0.9006514657980456, + "grad_norm": 4.839799276438571, + "learning_rate": 1.8841955895954088e-05, + "loss": 0.8524, + "step": 1659 + }, + { + "epoch": 0.9011943539630836, + "grad_norm": 5.598018889964705, + "learning_rate": 1.8840312574213372e-05, + "loss": 1.0939, + "step": 1660 + }, + { + "epoch": 0.9017372421281216, + "grad_norm": 5.2496226470868725, + "learning_rate": 1.8838668159088707e-05, + "loss": 1.0812, + "step": 1661 + }, + { + "epoch": 0.9022801302931596, + "grad_norm": 5.2298567821168, + "learning_rate": 1.8837022650783477e-05, + "loss": 0.953, + "step": 1662 + }, + { + "epoch": 0.9028230184581976, + "grad_norm": 4.471282412199049, + "learning_rate": 1.88353760495012e-05, + "loss": 1.2652, + "step": 1663 + }, + { + "epoch": 0.9033659066232356, + "grad_norm": 5.892082884595572, + "learning_rate": 1.8833728355445534e-05, + "loss": 1.1165, + "step": 1664 + }, + { + "epoch": 0.9039087947882736, + "grad_norm": 5.49174731306936, + "learning_rate": 1.8832079568820268e-05, + "loss": 0.9262, + "step": 1665 + }, + { + "epoch": 0.9044516829533116, + "grad_norm": 5.55211658654164, + "learning_rate": 1.883042968982932e-05, + "loss": 1.3112, + "step": 1666 + }, + { + "epoch": 0.9049945711183496, + "grad_norm": 5.597392580982551, + "learning_rate": 1.8828778718676757e-05, + "loss": 0.8511, + "step": 1667 + }, + { + "epoch": 0.9055374592833876, + "grad_norm": 6.183127994440369, + "learning_rate": 1.8827126655566773e-05, + "loss": 1.2784, + "step": 1668 + }, + { + "epoch": 0.9060803474484256, + "grad_norm": 6.991796130593696, + "learning_rate": 1.882547350070369e-05, + "loss": 1.1703, + "step": 1669 + }, + { + "epoch": 0.9066232356134636, + "grad_norm": 5.7794859741652305, + "learning_rate": 1.8823819254291986e-05, + "loss": 1.0974, + "step": 1670 + }, + { + "epoch": 0.9071661237785016, + "grad_norm": 5.388338879165358, + "learning_rate": 1.8822163916536245e-05, + "loss": 1.2087, + "step": 1671 + }, + { + "epoch": 0.9077090119435396, + "grad_norm": 5.072420136637763, + "learning_rate": 1.8820507487641218e-05, + "loss": 0.9244, + "step": 1672 + }, + { + "epoch": 0.9082519001085776, + "grad_norm": 5.9571835605705115, + "learning_rate": 1.8818849967811762e-05, + "loss": 1.3299, + "step": 1673 + }, + { + "epoch": 0.9087947882736156, + "grad_norm": 5.706573186001289, + "learning_rate": 1.8817191357252892e-05, + "loss": 1.1059, + "step": 1674 + }, + { + "epoch": 0.9093376764386536, + "grad_norm": 5.4984127833712435, + "learning_rate": 1.8815531656169737e-05, + "loss": 1.1428, + "step": 1675 + }, + { + "epoch": 0.9098805646036916, + "grad_norm": 4.824306323799433, + "learning_rate": 1.8813870864767582e-05, + "loss": 0.7313, + "step": 1676 + }, + { + "epoch": 0.9104234527687296, + "grad_norm": 6.180183255961245, + "learning_rate": 1.8812208983251828e-05, + "loss": 1.0729, + "step": 1677 + }, + { + "epoch": 0.9109663409337676, + "grad_norm": 6.586598617269493, + "learning_rate": 1.8810546011828024e-05, + "loss": 1.5871, + "step": 1678 + }, + { + "epoch": 0.9115092290988056, + "grad_norm": 5.5611281281711875, + "learning_rate": 1.8808881950701845e-05, + "loss": 0.8133, + "step": 1679 + }, + { + "epoch": 0.9120521172638436, + "grad_norm": 5.214863992160379, + "learning_rate": 1.8807216800079108e-05, + "loss": 1.0437, + "step": 1680 + }, + { + "epoch": 0.9125950054288816, + "grad_norm": 6.631045200232836, + "learning_rate": 1.8805550560165763e-05, + "loss": 1.446, + "step": 1681 + }, + { + "epoch": 0.9131378935939196, + "grad_norm": 6.140822200691373, + "learning_rate": 1.8803883231167887e-05, + "loss": 1.8672, + "step": 1682 + }, + { + "epoch": 0.9136807817589576, + "grad_norm": 4.5606387908479995, + "learning_rate": 1.8802214813291708e-05, + "loss": 1.0662, + "step": 1683 + }, + { + "epoch": 0.9142236699239956, + "grad_norm": 4.166651706834997, + "learning_rate": 1.8800545306743567e-05, + "loss": 0.749, + "step": 1684 + }, + { + "epoch": 0.9147665580890336, + "grad_norm": 4.674238905131916, + "learning_rate": 1.8798874711729957e-05, + "loss": 1.1106, + "step": 1685 + }, + { + "epoch": 0.9153094462540716, + "grad_norm": 5.224600466137879, + "learning_rate": 1.8797203028457497e-05, + "loss": 0.9652, + "step": 1686 + }, + { + "epoch": 0.9158523344191096, + "grad_norm": 5.87706899393542, + "learning_rate": 1.8795530257132947e-05, + "loss": 1.2146, + "step": 1687 + }, + { + "epoch": 0.9163952225841476, + "grad_norm": 7.587853042260599, + "learning_rate": 1.87938563979632e-05, + "loss": 1.9051, + "step": 1688 + }, + { + "epoch": 0.9169381107491856, + "grad_norm": 4.988334377581133, + "learning_rate": 1.8792181451155275e-05, + "loss": 0.8502, + "step": 1689 + }, + { + "epoch": 0.9174809989142236, + "grad_norm": 5.454638378956608, + "learning_rate": 1.8790505416916338e-05, + "loss": 1.0493, + "step": 1690 + }, + { + "epoch": 0.9180238870792616, + "grad_norm": 5.95672873256444, + "learning_rate": 1.878882829545368e-05, + "loss": 1.1296, + "step": 1691 + }, + { + "epoch": 0.9185667752442996, + "grad_norm": 6.921816321230723, + "learning_rate": 1.8787150086974734e-05, + "loss": 1.6388, + "step": 1692 + }, + { + "epoch": 0.9191096634093376, + "grad_norm": 5.423998305707574, + "learning_rate": 1.878547079168706e-05, + "loss": 1.4865, + "step": 1693 + }, + { + "epoch": 0.9196525515743756, + "grad_norm": 4.479238264802495, + "learning_rate": 1.878379040979835e-05, + "loss": 1.0257, + "step": 1694 + }, + { + "epoch": 0.9201954397394136, + "grad_norm": 6.325821606507086, + "learning_rate": 1.8782108941516446e-05, + "loss": 1.1156, + "step": 1695 + }, + { + "epoch": 0.9207383279044516, + "grad_norm": 6.220712298120886, + "learning_rate": 1.8780426387049315e-05, + "loss": 1.375, + "step": 1696 + }, + { + "epoch": 0.9212812160694897, + "grad_norm": 5.813906943099404, + "learning_rate": 1.877874274660505e-05, + "loss": 1.1869, + "step": 1697 + }, + { + "epoch": 0.9218241042345277, + "grad_norm": 5.0160254446208965, + "learning_rate": 1.8777058020391893e-05, + "loss": 1.0443, + "step": 1698 + }, + { + "epoch": 0.9223669923995657, + "grad_norm": 4.273981514908416, + "learning_rate": 1.877537220861821e-05, + "loss": 1.0015, + "step": 1699 + }, + { + "epoch": 0.9229098805646037, + "grad_norm": 4.908796260576538, + "learning_rate": 1.8773685311492513e-05, + "loss": 1.0203, + "step": 1700 + }, + { + "epoch": 0.9234527687296417, + "grad_norm": 5.566794452728763, + "learning_rate": 1.8771997329223425e-05, + "loss": 1.4637, + "step": 1701 + }, + { + "epoch": 0.9239956568946797, + "grad_norm": 4.946565856604512, + "learning_rate": 1.8770308262019733e-05, + "loss": 0.8093, + "step": 1702 + }, + { + "epoch": 0.9245385450597177, + "grad_norm": 4.767566465605265, + "learning_rate": 1.8768618110090334e-05, + "loss": 0.5938, + "step": 1703 + }, + { + "epoch": 0.9250814332247557, + "grad_norm": 6.919371263621627, + "learning_rate": 1.8766926873644272e-05, + "loss": 0.9857, + "step": 1704 + }, + { + "epoch": 0.9256243213897937, + "grad_norm": 3.5297488420709944, + "learning_rate": 1.876523455289072e-05, + "loss": 0.8744, + "step": 1705 + }, + { + "epoch": 0.9261672095548317, + "grad_norm": 4.910386600622229, + "learning_rate": 1.8763541148038994e-05, + "loss": 0.9491, + "step": 1706 + }, + { + "epoch": 0.9267100977198697, + "grad_norm": 5.550892558363115, + "learning_rate": 1.876184665929853e-05, + "loss": 1.2793, + "step": 1707 + }, + { + "epoch": 0.9272529858849077, + "grad_norm": 4.470721855980947, + "learning_rate": 1.8760151086878905e-05, + "loss": 0.9706, + "step": 1708 + }, + { + "epoch": 0.9277958740499457, + "grad_norm": 7.1893201073258926, + "learning_rate": 1.8758454430989833e-05, + "loss": 1.2473, + "step": 1709 + }, + { + "epoch": 0.9283387622149837, + "grad_norm": 5.933304434057961, + "learning_rate": 1.875675669184116e-05, + "loss": 1.4917, + "step": 1710 + }, + { + "epoch": 0.9288816503800217, + "grad_norm": 5.386119916834561, + "learning_rate": 1.8755057869642857e-05, + "loss": 0.9417, + "step": 1711 + }, + { + "epoch": 0.9294245385450597, + "grad_norm": 6.121437468126703, + "learning_rate": 1.875335796460505e-05, + "loss": 1.203, + "step": 1712 + }, + { + "epoch": 0.9299674267100977, + "grad_norm": 5.34957970840911, + "learning_rate": 1.8751656976937974e-05, + "loss": 1.2557, + "step": 1713 + }, + { + "epoch": 0.9305103148751357, + "grad_norm": 7.404283054108375, + "learning_rate": 1.8749954906852023e-05, + "loss": 1.449, + "step": 1714 + }, + { + "epoch": 0.9310532030401737, + "grad_norm": 6.5511839319022585, + "learning_rate": 1.8748251754557696e-05, + "loss": 1.5548, + "step": 1715 + }, + { + "epoch": 0.9315960912052117, + "grad_norm": 5.0254557279034815, + "learning_rate": 1.8746547520265654e-05, + "loss": 0.9777, + "step": 1716 + }, + { + "epoch": 0.9321389793702497, + "grad_norm": 6.385252198006408, + "learning_rate": 1.874484220418667e-05, + "loss": 1.254, + "step": 1717 + }, + { + "epoch": 0.9326818675352877, + "grad_norm": 6.872281478717196, + "learning_rate": 1.874313580653167e-05, + "loss": 1.278, + "step": 1718 + }, + { + "epoch": 0.9332247557003257, + "grad_norm": 8.525842687889295, + "learning_rate": 1.8741428327511696e-05, + "loss": 1.3712, + "step": 1719 + }, + { + "epoch": 0.9337676438653637, + "grad_norm": 7.97420765579221, + "learning_rate": 1.8739719767337933e-05, + "loss": 1.7493, + "step": 1720 + }, + { + "epoch": 0.9343105320304017, + "grad_norm": 6.107109526814546, + "learning_rate": 1.8738010126221705e-05, + "loss": 0.9861, + "step": 1721 + }, + { + "epoch": 0.9348534201954397, + "grad_norm": 6.4666849584489645, + "learning_rate": 1.8736299404374453e-05, + "loss": 0.8041, + "step": 1722 + }, + { + "epoch": 0.9353963083604777, + "grad_norm": 7.489590627219774, + "learning_rate": 1.873458760200777e-05, + "loss": 1.2325, + "step": 1723 + }, + { + "epoch": 0.9359391965255157, + "grad_norm": 7.020182643385788, + "learning_rate": 1.8732874719333373e-05, + "loss": 1.1698, + "step": 1724 + }, + { + "epoch": 0.9364820846905537, + "grad_norm": 6.436449840675483, + "learning_rate": 1.873116075656311e-05, + "loss": 1.084, + "step": 1725 + }, + { + "epoch": 0.9370249728555917, + "grad_norm": 7.27100690341834, + "learning_rate": 1.872944571390897e-05, + "loss": 0.9793, + "step": 1726 + }, + { + "epoch": 0.9375678610206297, + "grad_norm": 6.337617737677888, + "learning_rate": 1.872772959158307e-05, + "loss": 1.0558, + "step": 1727 + }, + { + "epoch": 0.9381107491856677, + "grad_norm": 5.356493873446841, + "learning_rate": 1.8726012389797667e-05, + "loss": 1.1031, + "step": 1728 + }, + { + "epoch": 0.9386536373507057, + "grad_norm": 6.302965546816682, + "learning_rate": 1.8724294108765142e-05, + "loss": 1.1401, + "step": 1729 + }, + { + "epoch": 0.9391965255157437, + "grad_norm": 6.116650787967334, + "learning_rate": 1.872257474869802e-05, + "loss": 1.0848, + "step": 1730 + }, + { + "epoch": 0.9397394136807817, + "grad_norm": 4.3874746967683, + "learning_rate": 1.8720854309808948e-05, + "loss": 0.7463, + "step": 1731 + }, + { + "epoch": 0.9402823018458197, + "grad_norm": 5.00673534942558, + "learning_rate": 1.871913279231072e-05, + "loss": 1.2365, + "step": 1732 + }, + { + "epoch": 0.9408251900108577, + "grad_norm": 5.222845312743783, + "learning_rate": 1.871741019641625e-05, + "loss": 1.062, + "step": 1733 + }, + { + "epoch": 0.9413680781758957, + "grad_norm": 4.624403224726164, + "learning_rate": 1.871568652233859e-05, + "loss": 1.0155, + "step": 1734 + }, + { + "epoch": 0.9419109663409337, + "grad_norm": 5.386412889478532, + "learning_rate": 1.8713961770290936e-05, + "loss": 1.4316, + "step": 1735 + }, + { + "epoch": 0.9424538545059717, + "grad_norm": 6.0899860043030705, + "learning_rate": 1.87122359404866e-05, + "loss": 1.1372, + "step": 1736 + }, + { + "epoch": 0.9429967426710097, + "grad_norm": 5.306362358042622, + "learning_rate": 1.8710509033139037e-05, + "loss": 0.9353, + "step": 1737 + }, + { + "epoch": 0.9435396308360477, + "grad_norm": 6.479420699028343, + "learning_rate": 1.8708781048461832e-05, + "loss": 1.2435, + "step": 1738 + }, + { + "epoch": 0.9440825190010858, + "grad_norm": 6.205746500302448, + "learning_rate": 1.8707051986668712e-05, + "loss": 0.7872, + "step": 1739 + }, + { + "epoch": 0.9446254071661238, + "grad_norm": 4.157411034856641, + "learning_rate": 1.8705321847973523e-05, + "loss": 0.8004, + "step": 1740 + }, + { + "epoch": 0.9451682953311618, + "grad_norm": 5.5814599600578285, + "learning_rate": 1.8703590632590254e-05, + "loss": 0.7384, + "step": 1741 + }, + { + "epoch": 0.9457111834961998, + "grad_norm": 6.040759722497843, + "learning_rate": 1.8701858340733023e-05, + "loss": 1.5941, + "step": 1742 + }, + { + "epoch": 0.9462540716612378, + "grad_norm": 4.592040464655725, + "learning_rate": 1.8700124972616085e-05, + "loss": 1.0662, + "step": 1743 + }, + { + "epoch": 0.9467969598262758, + "grad_norm": 6.077319934267432, + "learning_rate": 1.8698390528453823e-05, + "loss": 1.3726, + "step": 1744 + }, + { + "epoch": 0.9473398479913138, + "grad_norm": 4.255180894541495, + "learning_rate": 1.869665500846076e-05, + "loss": 0.7001, + "step": 1745 + }, + { + "epoch": 0.9478827361563518, + "grad_norm": 6.390593673413393, + "learning_rate": 1.869491841285154e-05, + "loss": 1.2268, + "step": 1746 + }, + { + "epoch": 0.9484256243213898, + "grad_norm": 5.347324129941674, + "learning_rate": 1.8693180741840957e-05, + "loss": 1.2193, + "step": 1747 + }, + { + "epoch": 0.9489685124864278, + "grad_norm": 6.407261723862006, + "learning_rate": 1.8691441995643927e-05, + "loss": 0.868, + "step": 1748 + }, + { + "epoch": 0.9495114006514658, + "grad_norm": 4.400520878629861, + "learning_rate": 1.8689702174475496e-05, + "loss": 0.4679, + "step": 1749 + }, + { + "epoch": 0.9500542888165038, + "grad_norm": 5.6161548427455505, + "learning_rate": 1.8687961278550852e-05, + "loss": 0.8842, + "step": 1750 + }, + { + "epoch": 0.9505971769815418, + "grad_norm": 4.7163538240048, + "learning_rate": 1.8686219308085306e-05, + "loss": 0.9972, + "step": 1751 + }, + { + "epoch": 0.9511400651465798, + "grad_norm": 4.845854095585142, + "learning_rate": 1.8684476263294318e-05, + "loss": 1.1046, + "step": 1752 + }, + { + "epoch": 0.9516829533116178, + "grad_norm": 3.878984502983999, + "learning_rate": 1.8682732144393463e-05, + "loss": 0.7873, + "step": 1753 + }, + { + "epoch": 0.9522258414766558, + "grad_norm": 5.736472248757515, + "learning_rate": 1.8680986951598458e-05, + "loss": 1.2046, + "step": 1754 + }, + { + "epoch": 0.9527687296416938, + "grad_norm": 5.12988223940441, + "learning_rate": 1.867924068512515e-05, + "loss": 0.6293, + "step": 1755 + }, + { + "epoch": 0.9533116178067318, + "grad_norm": 7.822467968073818, + "learning_rate": 1.867749334518952e-05, + "loss": 1.3625, + "step": 1756 + }, + { + "epoch": 0.9538545059717698, + "grad_norm": 5.488388947029871, + "learning_rate": 1.8675744932007687e-05, + "loss": 1.078, + "step": 1757 + }, + { + "epoch": 0.9543973941368078, + "grad_norm": 4.631848438676548, + "learning_rate": 1.8673995445795894e-05, + "loss": 0.69, + "step": 1758 + }, + { + "epoch": 0.9549402823018458, + "grad_norm": 7.078171019993124, + "learning_rate": 1.8672244886770516e-05, + "loss": 1.1036, + "step": 1759 + }, + { + "epoch": 0.9554831704668838, + "grad_norm": 5.9259972832349215, + "learning_rate": 1.8670493255148073e-05, + "loss": 0.7919, + "step": 1760 + }, + { + "epoch": 0.9560260586319218, + "grad_norm": 5.846343833877514, + "learning_rate": 1.8668740551145205e-05, + "loss": 1.1653, + "step": 1761 + }, + { + "epoch": 0.9565689467969598, + "grad_norm": 5.559082125902064, + "learning_rate": 1.8666986774978685e-05, + "loss": 1.4214, + "step": 1762 + }, + { + "epoch": 0.9571118349619978, + "grad_norm": 6.511207981127819, + "learning_rate": 1.8665231926865433e-05, + "loss": 1.3552, + "step": 1763 + }, + { + "epoch": 0.9576547231270358, + "grad_norm": 5.728999474717542, + "learning_rate": 1.8663476007022482e-05, + "loss": 0.7861, + "step": 1764 + }, + { + "epoch": 0.9581976112920738, + "grad_norm": 6.152118564149309, + "learning_rate": 1.8661719015667016e-05, + "loss": 0.8679, + "step": 1765 + }, + { + "epoch": 0.9587404994571118, + "grad_norm": 6.77121947972117, + "learning_rate": 1.8659960953016334e-05, + "loss": 0.8103, + "step": 1766 + }, + { + "epoch": 0.9592833876221498, + "grad_norm": 7.822575173268641, + "learning_rate": 1.865820181928788e-05, + "loss": 1.2558, + "step": 1767 + }, + { + "epoch": 0.9598262757871878, + "grad_norm": 6.389073592671253, + "learning_rate": 1.8656441614699225e-05, + "loss": 1.0314, + "step": 1768 + }, + { + "epoch": 0.9603691639522258, + "grad_norm": 6.141689972496849, + "learning_rate": 1.8654680339468076e-05, + "loss": 1.2452, + "step": 1769 + }, + { + "epoch": 0.9609120521172638, + "grad_norm": 6.934092850250001, + "learning_rate": 1.8652917993812267e-05, + "loss": 1.1186, + "step": 1770 + }, + { + "epoch": 0.9614549402823018, + "grad_norm": 6.129320663785202, + "learning_rate": 1.865115457794977e-05, + "loss": 0.9076, + "step": 1771 + }, + { + "epoch": 0.9619978284473398, + "grad_norm": 7.008261896063745, + "learning_rate": 1.8649390092098693e-05, + "loss": 1.7152, + "step": 1772 + }, + { + "epoch": 0.9625407166123778, + "grad_norm": 6.36833090372477, + "learning_rate": 1.8647624536477255e-05, + "loss": 0.8087, + "step": 1773 + }, + { + "epoch": 0.9630836047774158, + "grad_norm": 4.7595903392838865, + "learning_rate": 1.8645857911303838e-05, + "loss": 0.6374, + "step": 1774 + }, + { + "epoch": 0.9636264929424538, + "grad_norm": 6.410325139984648, + "learning_rate": 1.8644090216796934e-05, + "loss": 1.3611, + "step": 1775 + }, + { + "epoch": 0.9641693811074918, + "grad_norm": 5.900212705071564, + "learning_rate": 1.8642321453175177e-05, + "loss": 0.891, + "step": 1776 + }, + { + "epoch": 0.9647122692725298, + "grad_norm": 6.90562183471938, + "learning_rate": 1.8640551620657326e-05, + "loss": 0.7899, + "step": 1777 + }, + { + "epoch": 0.9652551574375678, + "grad_norm": 6.476066490933977, + "learning_rate": 1.8638780719462278e-05, + "loss": 0.9614, + "step": 1778 + }, + { + "epoch": 0.9657980456026058, + "grad_norm": 6.659586852509603, + "learning_rate": 1.8637008749809065e-05, + "loss": 1.2419, + "step": 1779 + }, + { + "epoch": 0.9663409337676439, + "grad_norm": 6.587754652016677, + "learning_rate": 1.8635235711916847e-05, + "loss": 1.3664, + "step": 1780 + }, + { + "epoch": 0.9668838219326819, + "grad_norm": 6.558578755813245, + "learning_rate": 1.863346160600491e-05, + "loss": 1.1249, + "step": 1781 + }, + { + "epoch": 0.9674267100977199, + "grad_norm": 4.791301716076818, + "learning_rate": 1.8631686432292685e-05, + "loss": 0.6836, + "step": 1782 + }, + { + "epoch": 0.9679695982627579, + "grad_norm": 5.120778292537024, + "learning_rate": 1.862991019099972e-05, + "loss": 0.592, + "step": 1783 + }, + { + "epoch": 0.9685124864277959, + "grad_norm": 8.250755784410368, + "learning_rate": 1.8628132882345713e-05, + "loss": 1.2579, + "step": 1784 + }, + { + "epoch": 0.9690553745928339, + "grad_norm": 5.811263926579046, + "learning_rate": 1.862635450655048e-05, + "loss": 1.2935, + "step": 1785 + }, + { + "epoch": 0.9695982627578719, + "grad_norm": 6.042040633945999, + "learning_rate": 1.862457506383397e-05, + "loss": 0.8784, + "step": 1786 + }, + { + "epoch": 0.9701411509229099, + "grad_norm": 7.115011060082062, + "learning_rate": 1.8622794554416272e-05, + "loss": 1.0717, + "step": 1787 + }, + { + "epoch": 0.9706840390879479, + "grad_norm": 7.520775839533288, + "learning_rate": 1.8621012978517604e-05, + "loss": 1.3468, + "step": 1788 + }, + { + "epoch": 0.9712269272529859, + "grad_norm": 6.949921961354188, + "learning_rate": 1.8619230336358306e-05, + "loss": 1.5609, + "step": 1789 + }, + { + "epoch": 0.9717698154180239, + "grad_norm": 6.396218455762339, + "learning_rate": 1.8617446628158866e-05, + "loss": 1.1807, + "step": 1790 + }, + { + "epoch": 0.9723127035830619, + "grad_norm": 6.314794279037124, + "learning_rate": 1.861566185413989e-05, + "loss": 0.9255, + "step": 1791 + }, + { + "epoch": 0.9728555917480999, + "grad_norm": 5.393752442926871, + "learning_rate": 1.8613876014522128e-05, + "loss": 0.7926, + "step": 1792 + }, + { + "epoch": 0.9733984799131379, + "grad_norm": 5.908313693688933, + "learning_rate": 1.8612089109526453e-05, + "loss": 0.7984, + "step": 1793 + }, + { + "epoch": 0.9739413680781759, + "grad_norm": 4.6952509401002125, + "learning_rate": 1.8610301139373867e-05, + "loss": 1.0344, + "step": 1794 + }, + { + "epoch": 0.9744842562432139, + "grad_norm": 6.81835335105175, + "learning_rate": 1.8608512104285517e-05, + "loss": 1.1532, + "step": 1795 + }, + { + "epoch": 0.9750271444082519, + "grad_norm": 4.159411114912345, + "learning_rate": 1.860672200448267e-05, + "loss": 1.0665, + "step": 1796 + }, + { + "epoch": 0.9755700325732899, + "grad_norm": 4.911404446825004, + "learning_rate": 1.8604930840186726e-05, + "loss": 1.1784, + "step": 1797 + }, + { + "epoch": 0.9761129207383279, + "grad_norm": 5.399453855489055, + "learning_rate": 1.860313861161922e-05, + "loss": 0.8215, + "step": 1798 + }, + { + "epoch": 0.9766558089033659, + "grad_norm": 4.797137239148531, + "learning_rate": 1.860134531900182e-05, + "loss": 0.6405, + "step": 1799 + }, + { + "epoch": 0.9771986970684039, + "grad_norm": 3.937076316066272, + "learning_rate": 1.859955096255633e-05, + "loss": 0.5912, + "step": 1800 + }, + { + "epoch": 0.9777415852334419, + "grad_norm": 5.569110777302129, + "learning_rate": 1.859775554250466e-05, + "loss": 0.8089, + "step": 1801 + }, + { + "epoch": 0.9782844733984799, + "grad_norm": 6.364258901894282, + "learning_rate": 1.859595905906889e-05, + "loss": 1.5181, + "step": 1802 + }, + { + "epoch": 0.9788273615635179, + "grad_norm": 4.995224609073908, + "learning_rate": 1.85941615124712e-05, + "loss": 0.8573, + "step": 1803 + }, + { + "epoch": 0.9793702497285559, + "grad_norm": 5.041992740138782, + "learning_rate": 1.8592362902933918e-05, + "loss": 0.8982, + "step": 1804 + }, + { + "epoch": 0.9799131378935939, + "grad_norm": 5.6054574490953275, + "learning_rate": 1.8590563230679496e-05, + "loss": 0.9285, + "step": 1805 + }, + { + "epoch": 0.9804560260586319, + "grad_norm": 5.269503319178264, + "learning_rate": 1.8588762495930526e-05, + "loss": 1.0963, + "step": 1806 + }, + { + "epoch": 0.9809989142236699, + "grad_norm": 6.184787731593613, + "learning_rate": 1.8586960698909718e-05, + "loss": 1.4005, + "step": 1807 + }, + { + "epoch": 0.9815418023887079, + "grad_norm": 8.50824591479003, + "learning_rate": 1.858515783983993e-05, + "loss": 1.9607, + "step": 1808 + }, + { + "epoch": 0.9820846905537459, + "grad_norm": 5.399464732589782, + "learning_rate": 1.8583353918944134e-05, + "loss": 0.951, + "step": 1809 + }, + { + "epoch": 0.9826275787187839, + "grad_norm": 6.577634699737779, + "learning_rate": 1.8581548936445447e-05, + "loss": 1.5139, + "step": 1810 + }, + { + "epoch": 0.9831704668838219, + "grad_norm": 10.183489163585788, + "learning_rate": 1.8579742892567107e-05, + "loss": 2.0539, + "step": 1811 + }, + { + "epoch": 0.9837133550488599, + "grad_norm": 6.578605968943458, + "learning_rate": 1.8577935787532494e-05, + "loss": 1.1454, + "step": 1812 + }, + { + "epoch": 0.9842562432138979, + "grad_norm": 4.923285404215955, + "learning_rate": 1.8576127621565113e-05, + "loss": 0.8965, + "step": 1813 + }, + { + "epoch": 0.9847991313789359, + "grad_norm": 5.912445932400008, + "learning_rate": 1.85743183948886e-05, + "loss": 1.1311, + "step": 1814 + }, + { + "epoch": 0.9853420195439739, + "grad_norm": 4.5183231221646105, + "learning_rate": 1.8572508107726725e-05, + "loss": 0.7511, + "step": 1815 + }, + { + "epoch": 0.9858849077090119, + "grad_norm": 5.2218543332097935, + "learning_rate": 1.8570696760303378e-05, + "loss": 1.0063, + "step": 1816 + }, + { + "epoch": 0.9864277958740499, + "grad_norm": 7.215748803707923, + "learning_rate": 1.85688843528426e-05, + "loss": 1.3674, + "step": 1817 + }, + { + "epoch": 0.9869706840390879, + "grad_norm": 6.362354074709609, + "learning_rate": 1.8567070885568547e-05, + "loss": 1.1319, + "step": 1818 + }, + { + "epoch": 0.987513572204126, + "grad_norm": 4.680755706450828, + "learning_rate": 1.8565256358705513e-05, + "loss": 0.6613, + "step": 1819 + }, + { + "epoch": 0.988056460369164, + "grad_norm": 4.80241945107321, + "learning_rate": 1.8563440772477922e-05, + "loss": 0.8619, + "step": 1820 + }, + { + "epoch": 0.988599348534202, + "grad_norm": 5.858709563059296, + "learning_rate": 1.856162412711033e-05, + "loss": 0.9432, + "step": 1821 + }, + { + "epoch": 0.98914223669924, + "grad_norm": 6.478119299588854, + "learning_rate": 1.855980642282742e-05, + "loss": 1.1866, + "step": 1822 + }, + { + "epoch": 0.989685124864278, + "grad_norm": 7.081924554034699, + "learning_rate": 1.8557987659854006e-05, + "loss": 1.3657, + "step": 1823 + }, + { + "epoch": 0.990228013029316, + "grad_norm": 5.404410220733361, + "learning_rate": 1.855616783841504e-05, + "loss": 1.1051, + "step": 1824 + }, + { + "epoch": 0.990770901194354, + "grad_norm": 6.742207962837618, + "learning_rate": 1.8554346958735602e-05, + "loss": 1.4667, + "step": 1825 + }, + { + "epoch": 0.991313789359392, + "grad_norm": 6.164178667211635, + "learning_rate": 1.8552525021040895e-05, + "loss": 0.7861, + "step": 1826 + }, + { + "epoch": 0.99185667752443, + "grad_norm": 6.078017237490637, + "learning_rate": 1.8550702025556265e-05, + "loss": 1.3294, + "step": 1827 + }, + { + "epoch": 0.992399565689468, + "grad_norm": 4.7372949568909455, + "learning_rate": 1.8548877972507182e-05, + "loss": 0.9779, + "step": 1828 + }, + { + "epoch": 0.992942453854506, + "grad_norm": 5.25971277777658, + "learning_rate": 1.8547052862119247e-05, + "loss": 1.049, + "step": 1829 + }, + { + "epoch": 0.993485342019544, + "grad_norm": 4.950709970125391, + "learning_rate": 1.854522669461819e-05, + "loss": 0.6665, + "step": 1830 + }, + { + "epoch": 0.994028230184582, + "grad_norm": 5.856428194694733, + "learning_rate": 1.8543399470229876e-05, + "loss": 1.0954, + "step": 1831 + }, + { + "epoch": 0.99457111834962, + "grad_norm": 4.895286613591861, + "learning_rate": 1.85415711891803e-05, + "loss": 0.7069, + "step": 1832 + }, + { + "epoch": 0.995114006514658, + "grad_norm": 5.249228502887773, + "learning_rate": 1.8539741851695586e-05, + "loss": 1.0655, + "step": 1833 + }, + { + "epoch": 0.995656894679696, + "grad_norm": 4.855694799243912, + "learning_rate": 1.8537911458001988e-05, + "loss": 1.1872, + "step": 1834 + }, + { + "epoch": 0.996199782844734, + "grad_norm": 5.180821237913711, + "learning_rate": 1.8536080008325896e-05, + "loss": 0.8281, + "step": 1835 + }, + { + "epoch": 0.996742671009772, + "grad_norm": 5.073254295771654, + "learning_rate": 1.8534247502893823e-05, + "loss": 0.7002, + "step": 1836 + }, + { + "epoch": 0.99728555917481, + "grad_norm": 5.4022604097623255, + "learning_rate": 1.8532413941932416e-05, + "loss": 1.1793, + "step": 1837 + }, + { + "epoch": 0.997828447339848, + "grad_norm": 4.981761324197401, + "learning_rate": 1.8530579325668455e-05, + "loss": 0.8331, + "step": 1838 + }, + { + "epoch": 0.998371335504886, + "grad_norm": 7.325363733761658, + "learning_rate": 1.852874365432885e-05, + "loss": 1.3071, + "step": 1839 + }, + { + "epoch": 0.998914223669924, + "grad_norm": 7.5668251202705505, + "learning_rate": 1.852690692814063e-05, + "loss": 1.4955, + "step": 1840 + }, + { + "epoch": 0.999457111834962, + "grad_norm": 6.706936559053067, + "learning_rate": 1.8525069147330978e-05, + "loss": 1.0286, + "step": 1841 + }, + { + "epoch": 1.0, + "grad_norm": 6.338369283394405, + "learning_rate": 1.8523230312127183e-05, + "loss": 1.0509, + "step": 1842 + }, + { + "epoch": 1.000542888165038, + "grad_norm": 6.119502670416093, + "learning_rate": 1.8521390422756683e-05, + "loss": 1.0805, + "step": 1843 + }, + { + "epoch": 1.001085776330076, + "grad_norm": 5.074420967885639, + "learning_rate": 1.851954947944703e-05, + "loss": 1.0931, + "step": 1844 + }, + { + "epoch": 1.001628664495114, + "grad_norm": 8.194067497007946, + "learning_rate": 1.851770748242592e-05, + "loss": 2.0732, + "step": 1845 + }, + { + "epoch": 1.002171552660152, + "grad_norm": 4.389522929363591, + "learning_rate": 1.8515864431921177e-05, + "loss": 0.8684, + "step": 1846 + }, + { + "epoch": 1.00271444082519, + "grad_norm": 5.720373954211297, + "learning_rate": 1.8514020328160748e-05, + "loss": 1.0223, + "step": 1847 + }, + { + "epoch": 1.003257328990228, + "grad_norm": 5.163678744409467, + "learning_rate": 1.8512175171372713e-05, + "loss": 0.9611, + "step": 1848 + }, + { + "epoch": 1.003800217155266, + "grad_norm": 5.786056891749039, + "learning_rate": 1.8510328961785286e-05, + "loss": 1.3473, + "step": 1849 + }, + { + "epoch": 1.004343105320304, + "grad_norm": 5.9194250471944985, + "learning_rate": 1.850848169962681e-05, + "loss": 0.8218, + "step": 1850 + }, + { + "epoch": 1.004885993485342, + "grad_norm": 5.890241796399565, + "learning_rate": 1.850663338512576e-05, + "loss": 0.842, + "step": 1851 + }, + { + "epoch": 1.00542888165038, + "grad_norm": 6.939341519595904, + "learning_rate": 1.8504784018510732e-05, + "loss": 0.9263, + "step": 1852 + }, + { + "epoch": 1.005971769815418, + "grad_norm": 4.592014594714093, + "learning_rate": 1.850293360001046e-05, + "loss": 1.015, + "step": 1853 + }, + { + "epoch": 1.006514657980456, + "grad_norm": 4.575145990552055, + "learning_rate": 1.8501082129853816e-05, + "loss": 1.0505, + "step": 1854 + }, + { + "epoch": 1.007057546145494, + "grad_norm": 5.999218616951103, + "learning_rate": 1.849922960826978e-05, + "loss": 1.0945, + "step": 1855 + }, + { + "epoch": 1.007600434310532, + "grad_norm": 5.459227552994586, + "learning_rate": 1.8497376035487483e-05, + "loss": 1.0167, + "step": 1856 + }, + { + "epoch": 1.00814332247557, + "grad_norm": 4.785837605715454, + "learning_rate": 1.8495521411736173e-05, + "loss": 1.0559, + "step": 1857 + }, + { + "epoch": 1.008686210640608, + "grad_norm": 5.813312187108195, + "learning_rate": 1.8493665737245236e-05, + "loss": 1.3477, + "step": 1858 + }, + { + "epoch": 1.009229098805646, + "grad_norm": 6.512945450475729, + "learning_rate": 1.8491809012244182e-05, + "loss": 1.4416, + "step": 1859 + }, + { + "epoch": 1.009771986970684, + "grad_norm": 5.599962385405486, + "learning_rate": 1.8489951236962658e-05, + "loss": 0.8961, + "step": 1860 + }, + { + "epoch": 1.010314875135722, + "grad_norm": 4.863989016796961, + "learning_rate": 1.848809241163043e-05, + "loss": 0.639, + "step": 1861 + }, + { + "epoch": 1.01085776330076, + "grad_norm": 5.937471136238718, + "learning_rate": 1.848623253647741e-05, + "loss": 1.0548, + "step": 1862 + }, + { + "epoch": 1.011400651465798, + "grad_norm": 4.75172125273428, + "learning_rate": 1.8484371611733625e-05, + "loss": 0.6493, + "step": 1863 + }, + { + "epoch": 1.011943539630836, + "grad_norm": 5.867263430143688, + "learning_rate": 1.848250963762923e-05, + "loss": 0.8569, + "step": 1864 + }, + { + "epoch": 1.012486427795874, + "grad_norm": 5.932410925593062, + "learning_rate": 1.848064661439453e-05, + "loss": 0.8351, + "step": 1865 + }, + { + "epoch": 1.013029315960912, + "grad_norm": 5.8320858990674, + "learning_rate": 1.847878254225994e-05, + "loss": 0.8916, + "step": 1866 + }, + { + "epoch": 1.01357220412595, + "grad_norm": 5.836721791741835, + "learning_rate": 1.847691742145601e-05, + "loss": 1.1742, + "step": 1867 + }, + { + "epoch": 1.014115092290988, + "grad_norm": 6.806668722890977, + "learning_rate": 1.8475051252213423e-05, + "loss": 1.0714, + "step": 1868 + }, + { + "epoch": 1.014657980456026, + "grad_norm": 7.026347641082525, + "learning_rate": 1.8473184034762992e-05, + "loss": 1.4101, + "step": 1869 + }, + { + "epoch": 1.015200868621064, + "grad_norm": 5.9532414283206005, + "learning_rate": 1.8471315769335657e-05, + "loss": 0.7855, + "step": 1870 + }, + { + "epoch": 1.015743756786102, + "grad_norm": 5.7578706206451855, + "learning_rate": 1.846944645616248e-05, + "loss": 0.7644, + "step": 1871 + }, + { + "epoch": 1.01628664495114, + "grad_norm": 6.340541706204924, + "learning_rate": 1.846757609547467e-05, + "loss": 1.1607, + "step": 1872 + }, + { + "epoch": 1.016829533116178, + "grad_norm": 4.942991579941027, + "learning_rate": 1.8465704687503558e-05, + "loss": 0.7578, + "step": 1873 + }, + { + "epoch": 1.017372421281216, + "grad_norm": 6.42121491667965, + "learning_rate": 1.846383223248059e-05, + "loss": 0.9924, + "step": 1874 + }, + { + "epoch": 1.017915309446254, + "grad_norm": 6.252498990648505, + "learning_rate": 1.8461958730637368e-05, + "loss": 0.811, + "step": 1875 + }, + { + "epoch": 1.018458197611292, + "grad_norm": 5.941336187529878, + "learning_rate": 1.84600841822056e-05, + "loss": 0.772, + "step": 1876 + }, + { + "epoch": 1.01900108577633, + "grad_norm": 5.323895762793836, + "learning_rate": 1.845820858741714e-05, + "loss": 0.8812, + "step": 1877 + }, + { + "epoch": 1.019543973941368, + "grad_norm": 7.2574721612646105, + "learning_rate": 1.845633194650396e-05, + "loss": 1.2759, + "step": 1878 + }, + { + "epoch": 1.020086862106406, + "grad_norm": 4.307176114235153, + "learning_rate": 1.8454454259698165e-05, + "loss": 0.569, + "step": 1879 + }, + { + "epoch": 1.020629750271444, + "grad_norm": 6.504543473682945, + "learning_rate": 1.8452575527231997e-05, + "loss": 0.9355, + "step": 1880 + }, + { + "epoch": 1.021172638436482, + "grad_norm": 5.822065969327671, + "learning_rate": 1.8450695749337816e-05, + "loss": 1.0287, + "step": 1881 + }, + { + "epoch": 1.02171552660152, + "grad_norm": 8.90668319984316, + "learning_rate": 1.8448814926248112e-05, + "loss": 1.5348, + "step": 1882 + }, + { + "epoch": 1.022258414766558, + "grad_norm": 5.233217719773852, + "learning_rate": 1.844693305819552e-05, + "loss": 0.564, + "step": 1883 + }, + { + "epoch": 1.022801302931596, + "grad_norm": 5.806961114054563, + "learning_rate": 1.844505014541278e-05, + "loss": 0.9349, + "step": 1884 + }, + { + "epoch": 1.023344191096634, + "grad_norm": 6.239514346807307, + "learning_rate": 1.8443166188132777e-05, + "loss": 0.9677, + "step": 1885 + }, + { + "epoch": 1.023887079261672, + "grad_norm": 5.596402826374599, + "learning_rate": 1.8441281186588528e-05, + "loss": 0.695, + "step": 1886 + }, + { + "epoch": 1.02442996742671, + "grad_norm": 6.531429104200087, + "learning_rate": 1.8439395141013165e-05, + "loss": 1.1543, + "step": 1887 + }, + { + "epoch": 1.024972855591748, + "grad_norm": 5.129461327017003, + "learning_rate": 1.843750805163996e-05, + "loss": 1.2624, + "step": 1888 + }, + { + "epoch": 1.0255157437567861, + "grad_norm": 6.862993011679228, + "learning_rate": 1.8435619918702318e-05, + "loss": 1.2673, + "step": 1889 + }, + { + "epoch": 1.0260586319218241, + "grad_norm": 5.134671966877655, + "learning_rate": 1.8433730742433755e-05, + "loss": 0.6032, + "step": 1890 + }, + { + "epoch": 1.0266015200868621, + "grad_norm": 6.9199078267443195, + "learning_rate": 1.8431840523067932e-05, + "loss": 1.1255, + "step": 1891 + }, + { + "epoch": 1.0271444082519001, + "grad_norm": 5.808796921918892, + "learning_rate": 1.8429949260838635e-05, + "loss": 0.8488, + "step": 1892 + }, + { + "epoch": 1.0276872964169381, + "grad_norm": 7.102540931353532, + "learning_rate": 1.842805695597978e-05, + "loss": 1.0634, + "step": 1893 + }, + { + "epoch": 1.0282301845819761, + "grad_norm": 9.932212718124791, + "learning_rate": 1.8426163608725403e-05, + "loss": 1.3909, + "step": 1894 + }, + { + "epoch": 1.0287730727470141, + "grad_norm": 6.999752698080143, + "learning_rate": 1.8424269219309686e-05, + "loss": 1.1106, + "step": 1895 + }, + { + "epoch": 1.0293159609120521, + "grad_norm": 6.569873028691542, + "learning_rate": 1.842237378796693e-05, + "loss": 0.9455, + "step": 1896 + }, + { + "epoch": 1.0298588490770901, + "grad_norm": 6.418746090533204, + "learning_rate": 1.8420477314931554e-05, + "loss": 0.934, + "step": 1897 + }, + { + "epoch": 1.0304017372421281, + "grad_norm": 7.7235490740904815, + "learning_rate": 1.8418579800438125e-05, + "loss": 1.1582, + "step": 1898 + }, + { + "epoch": 1.0309446254071661, + "grad_norm": 5.4350436368749895, + "learning_rate": 1.841668124472133e-05, + "loss": 0.7726, + "step": 1899 + }, + { + "epoch": 1.0314875135722041, + "grad_norm": 5.34390520812136, + "learning_rate": 1.8414781648015983e-05, + "loss": 0.7497, + "step": 1900 + }, + { + "epoch": 1.0320304017372421, + "grad_norm": 6.563688266956639, + "learning_rate": 1.841288101055703e-05, + "loss": 1.3352, + "step": 1901 + }, + { + "epoch": 1.0325732899022801, + "grad_norm": 7.710283890330315, + "learning_rate": 1.841097933257955e-05, + "loss": 0.8804, + "step": 1902 + }, + { + "epoch": 1.0331161780673181, + "grad_norm": 5.31500431944041, + "learning_rate": 1.840907661431874e-05, + "loss": 0.8667, + "step": 1903 + }, + { + "epoch": 1.0336590662323561, + "grad_norm": 7.476982099026401, + "learning_rate": 1.840717285600993e-05, + "loss": 0.8317, + "step": 1904 + }, + { + "epoch": 1.0342019543973942, + "grad_norm": 5.1563433597221415, + "learning_rate": 1.840526805788858e-05, + "loss": 1.0108, + "step": 1905 + }, + { + "epoch": 1.0347448425624322, + "grad_norm": 6.411832164370409, + "learning_rate": 1.8403362220190284e-05, + "loss": 0.7778, + "step": 1906 + }, + { + "epoch": 1.0352877307274702, + "grad_norm": 6.688312152010462, + "learning_rate": 1.8401455343150757e-05, + "loss": 1.2438, + "step": 1907 + }, + { + "epoch": 1.0358306188925082, + "grad_norm": 6.412659783246144, + "learning_rate": 1.839954742700584e-05, + "loss": 0.8191, + "step": 1908 + }, + { + "epoch": 1.0363735070575462, + "grad_norm": 6.399618830323376, + "learning_rate": 1.839763847199151e-05, + "loss": 1.4485, + "step": 1909 + }, + { + "epoch": 1.0369163952225842, + "grad_norm": 8.041718562623554, + "learning_rate": 1.8395728478343873e-05, + "loss": 1.0043, + "step": 1910 + }, + { + "epoch": 1.0374592833876222, + "grad_norm": 6.4118854777282905, + "learning_rate": 1.8393817446299152e-05, + "loss": 0.9183, + "step": 1911 + }, + { + "epoch": 1.0380021715526602, + "grad_norm": 6.400834366436549, + "learning_rate": 1.8391905376093717e-05, + "loss": 1.1517, + "step": 1912 + }, + { + "epoch": 1.0385450597176982, + "grad_norm": 6.160364431172616, + "learning_rate": 1.8389992267964046e-05, + "loss": 0.5949, + "step": 1913 + }, + { + "epoch": 1.0390879478827362, + "grad_norm": 7.298564934931577, + "learning_rate": 1.8388078122146763e-05, + "loss": 1.0234, + "step": 1914 + }, + { + "epoch": 1.0396308360477742, + "grad_norm": 6.132796013684569, + "learning_rate": 1.838616293887861e-05, + "loss": 0.9417, + "step": 1915 + }, + { + "epoch": 1.0401737242128122, + "grad_norm": 8.083369310536504, + "learning_rate": 1.8384246718396458e-05, + "loss": 1.2617, + "step": 1916 + }, + { + "epoch": 1.0407166123778502, + "grad_norm": 5.492413210001038, + "learning_rate": 1.8382329460937306e-05, + "loss": 0.7906, + "step": 1917 + }, + { + "epoch": 1.0412595005428882, + "grad_norm": 5.936786319330656, + "learning_rate": 1.838041116673829e-05, + "loss": 0.8254, + "step": 1918 + }, + { + "epoch": 1.0418023887079262, + "grad_norm": 6.637068024744901, + "learning_rate": 1.8378491836036666e-05, + "loss": 1.1779, + "step": 1919 + }, + { + "epoch": 1.0423452768729642, + "grad_norm": 5.481354066687985, + "learning_rate": 1.8376571469069814e-05, + "loss": 0.9586, + "step": 1920 + }, + { + "epoch": 1.0428881650380022, + "grad_norm": 7.669719057080363, + "learning_rate": 1.8374650066075257e-05, + "loss": 1.2058, + "step": 1921 + }, + { + "epoch": 1.0434310532030402, + "grad_norm": 7.387160499244698, + "learning_rate": 1.8372727627290627e-05, + "loss": 1.3861, + "step": 1922 + }, + { + "epoch": 1.0439739413680782, + "grad_norm": 4.975886150156898, + "learning_rate": 1.8370804152953704e-05, + "loss": 0.5849, + "step": 1923 + }, + { + "epoch": 1.0445168295331162, + "grad_norm": 5.413290989228805, + "learning_rate": 1.8368879643302383e-05, + "loss": 0.8425, + "step": 1924 + }, + { + "epoch": 1.0450597176981542, + "grad_norm": 6.564098001301886, + "learning_rate": 1.836695409857469e-05, + "loss": 0.9658, + "step": 1925 + }, + { + "epoch": 1.0456026058631922, + "grad_norm": 8.050721283259312, + "learning_rate": 1.8365027519008774e-05, + "loss": 1.684, + "step": 1926 + }, + { + "epoch": 1.0461454940282302, + "grad_norm": 5.409117554986433, + "learning_rate": 1.836309990484293e-05, + "loss": 0.8247, + "step": 1927 + }, + { + "epoch": 1.0466883821932682, + "grad_norm": 5.108392155413988, + "learning_rate": 1.8361171256315555e-05, + "loss": 0.8244, + "step": 1928 + }, + { + "epoch": 1.0472312703583062, + "grad_norm": 6.806129539780182, + "learning_rate": 1.8359241573665194e-05, + "loss": 1.0678, + "step": 1929 + }, + { + "epoch": 1.0477741585233442, + "grad_norm": 4.43019983774416, + "learning_rate": 1.8357310857130514e-05, + "loss": 0.6213, + "step": 1930 + }, + { + "epoch": 1.0483170466883822, + "grad_norm": 6.194875830854703, + "learning_rate": 1.835537910695031e-05, + "loss": 0.8774, + "step": 1931 + }, + { + "epoch": 1.0488599348534202, + "grad_norm": 4.540451693265499, + "learning_rate": 1.8353446323363496e-05, + "loss": 0.8508, + "step": 1932 + }, + { + "epoch": 1.0494028230184582, + "grad_norm": 5.9491142341914225, + "learning_rate": 1.8351512506609133e-05, + "loss": 1.0139, + "step": 1933 + }, + { + "epoch": 1.0499457111834962, + "grad_norm": 6.436571178648067, + "learning_rate": 1.834957765692639e-05, + "loss": 0.9272, + "step": 1934 + }, + { + "epoch": 1.0504885993485342, + "grad_norm": 5.6538429161280535, + "learning_rate": 1.8347641774554573e-05, + "loss": 0.7834, + "step": 1935 + }, + { + "epoch": 1.0510314875135722, + "grad_norm": 6.481163094620501, + "learning_rate": 1.8345704859733123e-05, + "loss": 1.0916, + "step": 1936 + }, + { + "epoch": 1.0515743756786102, + "grad_norm": 6.165149195118496, + "learning_rate": 1.8343766912701588e-05, + "loss": 1.1054, + "step": 1937 + }, + { + "epoch": 1.0521172638436482, + "grad_norm": 7.778083794032891, + "learning_rate": 1.834182793369967e-05, + "loss": 1.297, + "step": 1938 + }, + { + "epoch": 1.0526601520086862, + "grad_norm": 8.871857523729659, + "learning_rate": 1.8339887922967176e-05, + "loss": 1.5749, + "step": 1939 + }, + { + "epoch": 1.0532030401737242, + "grad_norm": 6.563413115451867, + "learning_rate": 1.8337946880744047e-05, + "loss": 1.0271, + "step": 1940 + }, + { + "epoch": 1.0537459283387622, + "grad_norm": 6.5304770661428275, + "learning_rate": 1.833600480727036e-05, + "loss": 0.9236, + "step": 1941 + }, + { + "epoch": 1.0542888165038002, + "grad_norm": 6.802479270907137, + "learning_rate": 1.8334061702786317e-05, + "loss": 0.9076, + "step": 1942 + }, + { + "epoch": 1.0548317046688382, + "grad_norm": 5.92560525391674, + "learning_rate": 1.833211756753224e-05, + "loss": 0.9601, + "step": 1943 + }, + { + "epoch": 1.0553745928338762, + "grad_norm": 7.436926375817581, + "learning_rate": 1.8330172401748584e-05, + "loss": 1.1758, + "step": 1944 + }, + { + "epoch": 1.0559174809989142, + "grad_norm": 6.369742715679364, + "learning_rate": 1.8328226205675927e-05, + "loss": 0.8424, + "step": 1945 + }, + { + "epoch": 1.0564603691639523, + "grad_norm": 7.60632781901004, + "learning_rate": 1.8326278979554976e-05, + "loss": 1.7815, + "step": 1946 + }, + { + "epoch": 1.0570032573289903, + "grad_norm": 8.499242837541805, + "learning_rate": 1.8324330723626578e-05, + "loss": 1.1341, + "step": 1947 + }, + { + "epoch": 1.0575461454940283, + "grad_norm": 6.469082998862602, + "learning_rate": 1.8322381438131686e-05, + "loss": 0.7798, + "step": 1948 + }, + { + "epoch": 1.0580890336590663, + "grad_norm": 5.923622010018089, + "learning_rate": 1.8320431123311388e-05, + "loss": 0.7971, + "step": 1949 + }, + { + "epoch": 1.0586319218241043, + "grad_norm": 6.617488371294151, + "learning_rate": 1.8318479779406914e-05, + "loss": 1.1551, + "step": 1950 + }, + { + "epoch": 1.0591748099891423, + "grad_norm": 6.813827391520987, + "learning_rate": 1.8316527406659604e-05, + "loss": 1.0234, + "step": 1951 + }, + { + "epoch": 1.0597176981541803, + "grad_norm": 7.580867549521637, + "learning_rate": 1.831457400531093e-05, + "loss": 0.7896, + "step": 1952 + }, + { + "epoch": 1.0602605863192183, + "grad_norm": 5.975929968595557, + "learning_rate": 1.8312619575602486e-05, + "loss": 1.1248, + "step": 1953 + }, + { + "epoch": 1.0608034744842563, + "grad_norm": 6.2878616306091955, + "learning_rate": 1.831066411777601e-05, + "loss": 1.0833, + "step": 1954 + }, + { + "epoch": 1.0613463626492943, + "grad_norm": 6.284963948133933, + "learning_rate": 1.8308707632073345e-05, + "loss": 1.0773, + "step": 1955 + }, + { + "epoch": 1.0618892508143323, + "grad_norm": 4.009278799804313, + "learning_rate": 1.830675011873648e-05, + "loss": 0.4782, + "step": 1956 + }, + { + "epoch": 1.0624321389793703, + "grad_norm": 6.648653088879531, + "learning_rate": 1.8304791578007524e-05, + "loss": 0.9761, + "step": 1957 + }, + { + "epoch": 1.0629750271444083, + "grad_norm": 6.867797713608168, + "learning_rate": 1.830283201012871e-05, + "loss": 0.9315, + "step": 1958 + }, + { + "epoch": 1.0635179153094463, + "grad_norm": 5.989268423819992, + "learning_rate": 1.83008714153424e-05, + "loss": 0.8312, + "step": 1959 + }, + { + "epoch": 1.0640608034744843, + "grad_norm": 6.207518262073172, + "learning_rate": 1.8298909793891083e-05, + "loss": 1.1003, + "step": 1960 + }, + { + "epoch": 1.0646036916395223, + "grad_norm": 5.5435710943195815, + "learning_rate": 1.8296947146017373e-05, + "loss": 0.8012, + "step": 1961 + }, + { + "epoch": 1.0651465798045603, + "grad_norm": 8.000995164961113, + "learning_rate": 1.829498347196402e-05, + "loss": 0.9864, + "step": 1962 + }, + { + "epoch": 1.0656894679695983, + "grad_norm": 6.483090321755992, + "learning_rate": 1.829301877197389e-05, + "loss": 1.2823, + "step": 1963 + }, + { + "epoch": 1.0662323561346363, + "grad_norm": 7.147142303433649, + "learning_rate": 1.8291053046289985e-05, + "loss": 0.9185, + "step": 1964 + }, + { + "epoch": 1.0667752442996743, + "grad_norm": 6.000816189695904, + "learning_rate": 1.828908629515542e-05, + "loss": 1.0284, + "step": 1965 + }, + { + "epoch": 1.0673181324647123, + "grad_norm": 6.188793151434744, + "learning_rate": 1.8287118518813453e-05, + "loss": 0.7041, + "step": 1966 + }, + { + "epoch": 1.0678610206297503, + "grad_norm": 7.353722352349769, + "learning_rate": 1.828514971750746e-05, + "loss": 1.0707, + "step": 1967 + }, + { + "epoch": 1.0684039087947883, + "grad_norm": 7.553378116857428, + "learning_rate": 1.8283179891480944e-05, + "loss": 1.1426, + "step": 1968 + }, + { + "epoch": 1.0689467969598263, + "grad_norm": 7.134594743515001, + "learning_rate": 1.828120904097754e-05, + "loss": 0.9583, + "step": 1969 + }, + { + "epoch": 1.0694896851248643, + "grad_norm": 5.680617024729498, + "learning_rate": 1.8279237166241004e-05, + "loss": 0.9632, + "step": 1970 + }, + { + "epoch": 1.0700325732899023, + "grad_norm": 5.932910663063727, + "learning_rate": 1.8277264267515218e-05, + "loss": 0.9984, + "step": 1971 + }, + { + "epoch": 1.0705754614549403, + "grad_norm": 7.833159555663921, + "learning_rate": 1.8275290345044198e-05, + "loss": 0.8038, + "step": 1972 + }, + { + "epoch": 1.0711183496199783, + "grad_norm": 6.429988603781813, + "learning_rate": 1.8273315399072076e-05, + "loss": 0.9125, + "step": 1973 + }, + { + "epoch": 1.0716612377850163, + "grad_norm": 7.955595430340166, + "learning_rate": 1.827133942984312e-05, + "loss": 0.8697, + "step": 1974 + }, + { + "epoch": 1.0722041259500543, + "grad_norm": 5.983514211221058, + "learning_rate": 1.826936243760172e-05, + "loss": 0.8155, + "step": 1975 + }, + { + "epoch": 1.0727470141150923, + "grad_norm": 6.19913176090901, + "learning_rate": 1.8267384422592398e-05, + "loss": 1.3207, + "step": 1976 + }, + { + "epoch": 1.0732899022801303, + "grad_norm": 6.467400855860047, + "learning_rate": 1.8265405385059792e-05, + "loss": 1.1279, + "step": 1977 + }, + { + "epoch": 1.0738327904451683, + "grad_norm": 6.515150741776156, + "learning_rate": 1.8263425325248675e-05, + "loss": 1.372, + "step": 1978 + }, + { + "epoch": 1.0743756786102063, + "grad_norm": 6.181497786496646, + "learning_rate": 1.8261444243403945e-05, + "loss": 0.8571, + "step": 1979 + }, + { + "epoch": 1.0749185667752443, + "grad_norm": 7.637816753034266, + "learning_rate": 1.8259462139770624e-05, + "loss": 1.4861, + "step": 1980 + }, + { + "epoch": 1.0754614549402823, + "grad_norm": 7.0094523855293485, + "learning_rate": 1.825747901459386e-05, + "loss": 1.0113, + "step": 1981 + }, + { + "epoch": 1.0760043431053203, + "grad_norm": 7.6230550634556655, + "learning_rate": 1.8255494868118933e-05, + "loss": 1.3818, + "step": 1982 + }, + { + "epoch": 1.0765472312703583, + "grad_norm": 6.946643864799888, + "learning_rate": 1.8253509700591242e-05, + "loss": 1.1384, + "step": 1983 + }, + { + "epoch": 1.0770901194353963, + "grad_norm": 7.288235795183878, + "learning_rate": 1.825152351225632e-05, + "loss": 1.1288, + "step": 1984 + }, + { + "epoch": 1.0776330076004343, + "grad_norm": 8.872287367521855, + "learning_rate": 1.8249536303359816e-05, + "loss": 1.6016, + "step": 1985 + }, + { + "epoch": 1.0781758957654723, + "grad_norm": 5.75562973378046, + "learning_rate": 1.8247548074147515e-05, + "loss": 1.1803, + "step": 1986 + }, + { + "epoch": 1.0787187839305103, + "grad_norm": 6.377752945158632, + "learning_rate": 1.824555882486532e-05, + "loss": 0.6991, + "step": 1987 + }, + { + "epoch": 1.0792616720955484, + "grad_norm": 7.178965937765225, + "learning_rate": 1.8243568555759274e-05, + "loss": 0.988, + "step": 1988 + }, + { + "epoch": 1.0798045602605864, + "grad_norm": 6.0869015160234525, + "learning_rate": 1.824157726707553e-05, + "loss": 0.8587, + "step": 1989 + }, + { + "epoch": 1.0803474484256244, + "grad_norm": 5.9113747723637395, + "learning_rate": 1.823958495906037e-05, + "loss": 0.8815, + "step": 1990 + }, + { + "epoch": 1.0808903365906624, + "grad_norm": 7.84172234151508, + "learning_rate": 1.8237591631960218e-05, + "loss": 1.2621, + "step": 1991 + }, + { + "epoch": 1.0814332247557004, + "grad_norm": 7.188505727546616, + "learning_rate": 1.8235597286021597e-05, + "loss": 0.5991, + "step": 1992 + }, + { + "epoch": 1.0819761129207384, + "grad_norm": 6.358900329219305, + "learning_rate": 1.823360192149118e-05, + "loss": 1.3325, + "step": 1993 + }, + { + "epoch": 1.0825190010857764, + "grad_norm": 6.545296131842583, + "learning_rate": 1.8231605538615756e-05, + "loss": 1.1091, + "step": 1994 + }, + { + "epoch": 1.0830618892508144, + "grad_norm": 6.708966367161533, + "learning_rate": 1.8229608137642238e-05, + "loss": 1.1352, + "step": 1995 + }, + { + "epoch": 1.0836047774158524, + "grad_norm": 6.651776068492304, + "learning_rate": 1.822760971881767e-05, + "loss": 0.6818, + "step": 1996 + }, + { + "epoch": 1.0841476655808904, + "grad_norm": 4.745518059898734, + "learning_rate": 1.8225610282389222e-05, + "loss": 0.9403, + "step": 1997 + }, + { + "epoch": 1.0846905537459284, + "grad_norm": 6.526942367542791, + "learning_rate": 1.8223609828604184e-05, + "loss": 1.0543, + "step": 1998 + }, + { + "epoch": 1.0852334419109664, + "grad_norm": 5.331318824902818, + "learning_rate": 1.8221608357709973e-05, + "loss": 0.6626, + "step": 1999 + }, + { + "epoch": 1.0857763300760044, + "grad_norm": 6.640324946305343, + "learning_rate": 1.8219605869954134e-05, + "loss": 1.0166, + "step": 2000 + }, + { + "epoch": 1.0863192182410424, + "grad_norm": 5.684199608008352, + "learning_rate": 1.8217602365584352e-05, + "loss": 0.6796, + "step": 2001 + }, + { + "epoch": 1.0868621064060804, + "grad_norm": 5.429738239817468, + "learning_rate": 1.8215597844848403e-05, + "loss": 0.6783, + "step": 2002 + }, + { + "epoch": 1.0874049945711184, + "grad_norm": 6.1832795338939235, + "learning_rate": 1.821359230799422e-05, + "loss": 1.4244, + "step": 2003 + }, + { + "epoch": 1.0879478827361564, + "grad_norm": 7.118410329483546, + "learning_rate": 1.8211585755269852e-05, + "loss": 1.0274, + "step": 2004 + }, + { + "epoch": 1.0884907709011944, + "grad_norm": 5.5061218005078265, + "learning_rate": 1.820957818692347e-05, + "loss": 0.7143, + "step": 2005 + }, + { + "epoch": 1.0890336590662324, + "grad_norm": 5.518412565715972, + "learning_rate": 1.8207569603203373e-05, + "loss": 1.3158, + "step": 2006 + }, + { + "epoch": 1.0895765472312704, + "grad_norm": 5.47000403426222, + "learning_rate": 1.820556000435799e-05, + "loss": 1.0694, + "step": 2007 + }, + { + "epoch": 1.0901194353963084, + "grad_norm": 7.469961154104622, + "learning_rate": 1.820354939063586e-05, + "loss": 1.0702, + "step": 2008 + }, + { + "epoch": 1.0906623235613464, + "grad_norm": 5.889961089255285, + "learning_rate": 1.8201537762285674e-05, + "loss": 1.0598, + "step": 2009 + }, + { + "epoch": 1.0912052117263844, + "grad_norm": 6.659595389259375, + "learning_rate": 1.8199525119556226e-05, + "loss": 0.8456, + "step": 2010 + }, + { + "epoch": 1.0917480998914224, + "grad_norm": 6.328157257432415, + "learning_rate": 1.8197511462696443e-05, + "loss": 0.9763, + "step": 2011 + }, + { + "epoch": 1.0922909880564604, + "grad_norm": 5.916734142619132, + "learning_rate": 1.8195496791955373e-05, + "loss": 1.1194, + "step": 2012 + }, + { + "epoch": 1.0928338762214984, + "grad_norm": 5.904087624332602, + "learning_rate": 1.8193481107582203e-05, + "loss": 0.756, + "step": 2013 + }, + { + "epoch": 1.0933767643865364, + "grad_norm": 6.284745154848233, + "learning_rate": 1.8191464409826227e-05, + "loss": 0.7931, + "step": 2014 + }, + { + "epoch": 1.0939196525515744, + "grad_norm": 7.838306243079945, + "learning_rate": 1.8189446698936878e-05, + "loss": 1.3635, + "step": 2015 + }, + { + "epoch": 1.0944625407166124, + "grad_norm": 6.3405758352546995, + "learning_rate": 1.818742797516371e-05, + "loss": 0.5365, + "step": 2016 + }, + { + "epoch": 1.0950054288816504, + "grad_norm": 8.315910084823138, + "learning_rate": 1.8185408238756405e-05, + "loss": 1.6592, + "step": 2017 + }, + { + "epoch": 1.0955483170466884, + "grad_norm": 5.205053142209431, + "learning_rate": 1.8183387489964762e-05, + "loss": 0.6416, + "step": 2018 + }, + { + "epoch": 1.0960912052117264, + "grad_norm": 7.018373673519811, + "learning_rate": 1.8181365729038706e-05, + "loss": 1.0814, + "step": 2019 + }, + { + "epoch": 1.0966340933767644, + "grad_norm": 5.705860306884454, + "learning_rate": 1.8179342956228307e-05, + "loss": 0.5806, + "step": 2020 + }, + { + "epoch": 1.0971769815418024, + "grad_norm": 7.077662290720525, + "learning_rate": 1.8177319171783728e-05, + "loss": 0.8969, + "step": 2021 + }, + { + "epoch": 1.0977198697068404, + "grad_norm": 6.540901861387866, + "learning_rate": 1.8175294375955284e-05, + "loss": 0.7982, + "step": 2022 + }, + { + "epoch": 1.0982627578718784, + "grad_norm": 7.283213833491601, + "learning_rate": 1.81732685689934e-05, + "loss": 1.2659, + "step": 2023 + }, + { + "epoch": 1.0988056460369164, + "grad_norm": 7.016589270114781, + "learning_rate": 1.8171241751148633e-05, + "loss": 1.0011, + "step": 2024 + }, + { + "epoch": 1.0993485342019544, + "grad_norm": 5.724817250167612, + "learning_rate": 1.8169213922671666e-05, + "loss": 0.8967, + "step": 2025 + }, + { + "epoch": 1.0998914223669924, + "grad_norm": 7.431207902407107, + "learning_rate": 1.81671850838133e-05, + "loss": 0.9445, + "step": 2026 + }, + { + "epoch": 1.1004343105320304, + "grad_norm": 6.9152342645144955, + "learning_rate": 1.816515523482447e-05, + "loss": 0.8187, + "step": 2027 + }, + { + "epoch": 1.1009771986970684, + "grad_norm": 7.045028624610342, + "learning_rate": 1.816312437595622e-05, + "loss": 0.7323, + "step": 2028 + }, + { + "epoch": 1.1015200868621065, + "grad_norm": 6.63265573559803, + "learning_rate": 1.816109250745974e-05, + "loss": 0.6689, + "step": 2029 + }, + { + "epoch": 1.1020629750271445, + "grad_norm": 8.605675554379312, + "learning_rate": 1.8159059629586333e-05, + "loss": 1.3733, + "step": 2030 + }, + { + "epoch": 1.1026058631921825, + "grad_norm": 7.481808301387036, + "learning_rate": 1.8157025742587426e-05, + "loss": 0.9767, + "step": 2031 + }, + { + "epoch": 1.1031487513572205, + "grad_norm": 7.113214270917786, + "learning_rate": 1.8154990846714575e-05, + "loss": 0.6365, + "step": 2032 + }, + { + "epoch": 1.1036916395222585, + "grad_norm": 7.672152203903941, + "learning_rate": 1.8152954942219462e-05, + "loss": 1.3314, + "step": 2033 + }, + { + "epoch": 1.1042345276872965, + "grad_norm": 7.533257156973554, + "learning_rate": 1.8150918029353885e-05, + "loss": 0.967, + "step": 2034 + }, + { + "epoch": 1.1047774158523345, + "grad_norm": 6.469759228237502, + "learning_rate": 1.8148880108369775e-05, + "loss": 1.348, + "step": 2035 + }, + { + "epoch": 1.1053203040173725, + "grad_norm": 6.620692148711997, + "learning_rate": 1.8146841179519186e-05, + "loss": 0.7538, + "step": 2036 + }, + { + "epoch": 1.1058631921824105, + "grad_norm": 6.974705446587608, + "learning_rate": 1.8144801243054297e-05, + "loss": 1.3664, + "step": 2037 + }, + { + "epoch": 1.1064060803474485, + "grad_norm": 8.047700738112871, + "learning_rate": 1.8142760299227408e-05, + "loss": 1.0466, + "step": 2038 + }, + { + "epoch": 1.1069489685124865, + "grad_norm": 5.469412128245213, + "learning_rate": 1.814071834829095e-05, + "loss": 0.5464, + "step": 2039 + }, + { + "epoch": 1.1074918566775245, + "grad_norm": 5.732997692484992, + "learning_rate": 1.813867539049747e-05, + "loss": 0.7307, + "step": 2040 + }, + { + "epoch": 1.1080347448425625, + "grad_norm": 8.89935072196041, + "learning_rate": 1.8136631426099646e-05, + "loss": 1.8231, + "step": 2041 + }, + { + "epoch": 1.1085776330076005, + "grad_norm": 7.0036806828899465, + "learning_rate": 1.813458645535028e-05, + "loss": 1.063, + "step": 2042 + }, + { + "epoch": 1.1091205211726385, + "grad_norm": 6.013891951698058, + "learning_rate": 1.8132540478502297e-05, + "loss": 0.9836, + "step": 2043 + }, + { + "epoch": 1.1096634093376765, + "grad_norm": 4.234811309290956, + "learning_rate": 1.813049349580875e-05, + "loss": 0.3708, + "step": 2044 + }, + { + "epoch": 1.1102062975027145, + "grad_norm": 6.259459841046158, + "learning_rate": 1.8128445507522806e-05, + "loss": 0.8463, + "step": 2045 + }, + { + "epoch": 1.1107491856677525, + "grad_norm": 7.9568342619649775, + "learning_rate": 1.8126396513897764e-05, + "loss": 1.0563, + "step": 2046 + }, + { + "epoch": 1.1112920738327905, + "grad_norm": 6.937848428371049, + "learning_rate": 1.8124346515187056e-05, + "loss": 0.6359, + "step": 2047 + }, + { + "epoch": 1.1118349619978285, + "grad_norm": 7.672632972051955, + "learning_rate": 1.8122295511644218e-05, + "loss": 0.657, + "step": 2048 + }, + { + "epoch": 1.1123778501628665, + "grad_norm": 6.450214026534184, + "learning_rate": 1.8120243503522924e-05, + "loss": 0.7279, + "step": 2049 + }, + { + "epoch": 1.1129207383279045, + "grad_norm": 6.628216836060006, + "learning_rate": 1.8118190491076978e-05, + "loss": 0.5804, + "step": 2050 + }, + { + "epoch": 1.1134636264929425, + "grad_norm": 6.698074316145226, + "learning_rate": 1.8116136474560288e-05, + "loss": 1.011, + "step": 2051 + }, + { + "epoch": 1.1140065146579805, + "grad_norm": 5.4126983291571795, + "learning_rate": 1.8114081454226905e-05, + "loss": 0.8564, + "step": 2052 + }, + { + "epoch": 1.1145494028230185, + "grad_norm": 8.596797841329312, + "learning_rate": 1.8112025430331e-05, + "loss": 1.1924, + "step": 2053 + }, + { + "epoch": 1.1150922909880565, + "grad_norm": 7.416912053391948, + "learning_rate": 1.8109968403126856e-05, + "loss": 0.9894, + "step": 2054 + }, + { + "epoch": 1.1156351791530945, + "grad_norm": 7.434308789484601, + "learning_rate": 1.8107910372868898e-05, + "loss": 1.1544, + "step": 2055 + }, + { + "epoch": 1.1161780673181325, + "grad_norm": 8.325689788099726, + "learning_rate": 1.8105851339811663e-05, + "loss": 0.9994, + "step": 2056 + }, + { + "epoch": 1.1167209554831705, + "grad_norm": 8.384964960696278, + "learning_rate": 1.8103791304209813e-05, + "loss": 1.1768, + "step": 2057 + }, + { + "epoch": 1.1172638436482085, + "grad_norm": 7.044535285708613, + "learning_rate": 1.810173026631814e-05, + "loss": 1.1801, + "step": 2058 + }, + { + "epoch": 1.1178067318132465, + "grad_norm": 6.665810359686943, + "learning_rate": 1.8099668226391552e-05, + "loss": 0.7865, + "step": 2059 + }, + { + "epoch": 1.1183496199782845, + "grad_norm": 7.395697719004186, + "learning_rate": 1.8097605184685093e-05, + "loss": 1.1357, + "step": 2060 + }, + { + "epoch": 1.1188925081433225, + "grad_norm": 7.1495559268215185, + "learning_rate": 1.809554114145392e-05, + "loss": 1.1507, + "step": 2061 + }, + { + "epoch": 1.1194353963083605, + "grad_norm": 7.877063869804663, + "learning_rate": 1.8093476096953315e-05, + "loss": 1.634, + "step": 2062 + }, + { + "epoch": 1.1199782844733985, + "grad_norm": 6.528060830825324, + "learning_rate": 1.809141005143869e-05, + "loss": 0.697, + "step": 2063 + }, + { + "epoch": 1.1205211726384365, + "grad_norm": 6.5949576624011, + "learning_rate": 1.808934300516557e-05, + "loss": 1.1506, + "step": 2064 + }, + { + "epoch": 1.1210640608034745, + "grad_norm": 6.787348272435354, + "learning_rate": 1.8087274958389612e-05, + "loss": 0.9986, + "step": 2065 + }, + { + "epoch": 1.1216069489685125, + "grad_norm": 6.310955809912108, + "learning_rate": 1.8085205911366602e-05, + "loss": 0.89, + "step": 2066 + }, + { + "epoch": 1.1221498371335505, + "grad_norm": 6.8487503881448895, + "learning_rate": 1.8083135864352442e-05, + "loss": 1.1061, + "step": 2067 + }, + { + "epoch": 1.1226927252985885, + "grad_norm": 5.329494375939551, + "learning_rate": 1.808106481760315e-05, + "loss": 1.0011, + "step": 2068 + }, + { + "epoch": 1.1232356134636265, + "grad_norm": 6.286005313228098, + "learning_rate": 1.8078992771374886e-05, + "loss": 0.8399, + "step": 2069 + }, + { + "epoch": 1.1237785016286646, + "grad_norm": 6.324896503266866, + "learning_rate": 1.8076919725923917e-05, + "loss": 1.2716, + "step": 2070 + }, + { + "epoch": 1.1243213897937026, + "grad_norm": 5.663105266452551, + "learning_rate": 1.8074845681506644e-05, + "loss": 0.9943, + "step": 2071 + }, + { + "epoch": 1.1248642779587406, + "grad_norm": 5.175572461557621, + "learning_rate": 1.807277063837959e-05, + "loss": 0.6319, + "step": 2072 + }, + { + "epoch": 1.1254071661237786, + "grad_norm": 7.297487212372251, + "learning_rate": 1.8070694596799397e-05, + "loss": 0.7533, + "step": 2073 + }, + { + "epoch": 1.1259500542888166, + "grad_norm": 9.046578233072974, + "learning_rate": 1.806861755702283e-05, + "loss": 0.9991, + "step": 2074 + }, + { + "epoch": 1.1264929424538546, + "grad_norm": 4.516511292735119, + "learning_rate": 1.8066539519306786e-05, + "loss": 0.6393, + "step": 2075 + }, + { + "epoch": 1.1270358306188926, + "grad_norm": 5.16623460840236, + "learning_rate": 1.8064460483908283e-05, + "loss": 0.7083, + "step": 2076 + }, + { + "epoch": 1.1275787187839306, + "grad_norm": 10.7323945119629, + "learning_rate": 1.8062380451084445e-05, + "loss": 2.4193, + "step": 2077 + }, + { + "epoch": 1.1281216069489686, + "grad_norm": 6.438876475868993, + "learning_rate": 1.8060299421092554e-05, + "loss": 0.8468, + "step": 2078 + }, + { + "epoch": 1.1286644951140066, + "grad_norm": 8.437791617086905, + "learning_rate": 1.8058217394189976e-05, + "loss": 1.4895, + "step": 2079 + }, + { + "epoch": 1.1292073832790446, + "grad_norm": 5.237528459736948, + "learning_rate": 1.805613437063423e-05, + "loss": 0.643, + "step": 2080 + }, + { + "epoch": 1.1297502714440826, + "grad_norm": 6.010151293762264, + "learning_rate": 1.8054050350682947e-05, + "loss": 0.9001, + "step": 2081 + }, + { + "epoch": 1.1302931596091206, + "grad_norm": 4.816849804237718, + "learning_rate": 1.805196533459388e-05, + "loss": 0.6349, + "step": 2082 + }, + { + "epoch": 1.1308360477741586, + "grad_norm": 6.872496495387926, + "learning_rate": 1.8049879322624906e-05, + "loss": 0.8808, + "step": 2083 + }, + { + "epoch": 1.1313789359391966, + "grad_norm": 5.86067647581003, + "learning_rate": 1.804779231503403e-05, + "loss": 0.6811, + "step": 2084 + }, + { + "epoch": 1.1319218241042346, + "grad_norm": 6.073217718554074, + "learning_rate": 1.8045704312079376e-05, + "loss": 1.0086, + "step": 2085 + }, + { + "epoch": 1.1324647122692726, + "grad_norm": 7.519215180241719, + "learning_rate": 1.804361531401918e-05, + "loss": 0.8815, + "step": 2086 + }, + { + "epoch": 1.1330076004343106, + "grad_norm": 6.49012361077388, + "learning_rate": 1.8041525321111835e-05, + "loss": 1.0129, + "step": 2087 + }, + { + "epoch": 1.1335504885993486, + "grad_norm": 5.938773374319115, + "learning_rate": 1.8039434333615814e-05, + "loss": 0.8485, + "step": 2088 + }, + { + "epoch": 1.1340933767643866, + "grad_norm": 5.607636698742016, + "learning_rate": 1.8037342351789743e-05, + "loss": 0.6209, + "step": 2089 + }, + { + "epoch": 1.1346362649294246, + "grad_norm": 5.522935966563286, + "learning_rate": 1.803524937589236e-05, + "loss": 0.6576, + "step": 2090 + }, + { + "epoch": 1.1351791530944626, + "grad_norm": 8.130974168982206, + "learning_rate": 1.8033155406182533e-05, + "loss": 1.0676, + "step": 2091 + }, + { + "epoch": 1.1357220412595006, + "grad_norm": 6.027276275319545, + "learning_rate": 1.803106044291924e-05, + "loss": 0.9853, + "step": 2092 + }, + { + "epoch": 1.1362649294245386, + "grad_norm": 4.645985748184023, + "learning_rate": 1.8028964486361586e-05, + "loss": 0.7505, + "step": 2093 + }, + { + "epoch": 1.1368078175895766, + "grad_norm": 6.820894579082682, + "learning_rate": 1.8026867536768816e-05, + "loss": 0.9358, + "step": 2094 + }, + { + "epoch": 1.1373507057546146, + "grad_norm": 3.9530727606720872, + "learning_rate": 1.802476959440027e-05, + "loss": 0.7736, + "step": 2095 + }, + { + "epoch": 1.1378935939196526, + "grad_norm": 7.650275740883737, + "learning_rate": 1.8022670659515432e-05, + "loss": 1.4832, + "step": 2096 + }, + { + "epoch": 1.1384364820846906, + "grad_norm": 6.94341164115752, + "learning_rate": 1.80205707323739e-05, + "loss": 1.0336, + "step": 2097 + }, + { + "epoch": 1.1389793702497286, + "grad_norm": 5.608528053150328, + "learning_rate": 1.8018469813235403e-05, + "loss": 0.829, + "step": 2098 + }, + { + "epoch": 1.1395222584147666, + "grad_norm": 5.2191336182398995, + "learning_rate": 1.8016367902359776e-05, + "loss": 0.9708, + "step": 2099 + }, + { + "epoch": 1.1400651465798046, + "grad_norm": 6.773346810380977, + "learning_rate": 1.801426500000699e-05, + "loss": 0.6755, + "step": 2100 + }, + { + "epoch": 1.1406080347448426, + "grad_norm": 7.292992073718992, + "learning_rate": 1.8012161106437137e-05, + "loss": 0.8842, + "step": 2101 + }, + { + "epoch": 1.1411509229098806, + "grad_norm": 6.699434744454233, + "learning_rate": 1.8010056221910427e-05, + "loss": 1.0385, + "step": 2102 + }, + { + "epoch": 1.1416938110749186, + "grad_norm": 6.565299107509569, + "learning_rate": 1.8007950346687198e-05, + "loss": 0.6557, + "step": 2103 + }, + { + "epoch": 1.1422366992399566, + "grad_norm": 7.2741514068627025, + "learning_rate": 1.800584348102791e-05, + "loss": 1.0347, + "step": 2104 + }, + { + "epoch": 1.1427795874049946, + "grad_norm": 6.467241522225306, + "learning_rate": 1.800373562519314e-05, + "loss": 0.8026, + "step": 2105 + }, + { + "epoch": 1.1433224755700326, + "grad_norm": 6.109318081193457, + "learning_rate": 1.800162677944359e-05, + "loss": 0.9182, + "step": 2106 + }, + { + "epoch": 1.1438653637350706, + "grad_norm": 7.136347555199431, + "learning_rate": 1.7999516944040087e-05, + "loss": 1.2743, + "step": 2107 + }, + { + "epoch": 1.1444082519001086, + "grad_norm": 7.791284332349926, + "learning_rate": 1.7997406119243582e-05, + "loss": 0.9738, + "step": 2108 + }, + { + "epoch": 1.1449511400651466, + "grad_norm": 7.315957477694666, + "learning_rate": 1.7995294305315137e-05, + "loss": 1.0674, + "step": 2109 + }, + { + "epoch": 1.1454940282301846, + "grad_norm": 6.305984776426275, + "learning_rate": 1.7993181502515957e-05, + "loss": 0.8204, + "step": 2110 + }, + { + "epoch": 1.1460369163952226, + "grad_norm": 5.824114865699788, + "learning_rate": 1.7991067711107345e-05, + "loss": 0.7953, + "step": 2111 + }, + { + "epoch": 1.1465798045602607, + "grad_norm": 7.4284096250312786, + "learning_rate": 1.798895293135074e-05, + "loss": 0.7349, + "step": 2112 + }, + { + "epoch": 1.1471226927252987, + "grad_norm": 6.415949338549894, + "learning_rate": 1.798683716350771e-05, + "loss": 0.8346, + "step": 2113 + }, + { + "epoch": 1.1476655808903367, + "grad_norm": 6.360883520037482, + "learning_rate": 1.7984720407839925e-05, + "loss": 0.7966, + "step": 2114 + }, + { + "epoch": 1.1482084690553747, + "grad_norm": 6.703196393264975, + "learning_rate": 1.7982602664609198e-05, + "loss": 1.3341, + "step": 2115 + }, + { + "epoch": 1.1487513572204127, + "grad_norm": 8.115384707506347, + "learning_rate": 1.798048393407745e-05, + "loss": 1.0284, + "step": 2116 + }, + { + "epoch": 1.1492942453854507, + "grad_norm": 8.519166176194911, + "learning_rate": 1.797836421650673e-05, + "loss": 1.4075, + "step": 2117 + }, + { + "epoch": 1.1498371335504887, + "grad_norm": 7.120345871314324, + "learning_rate": 1.7976243512159207e-05, + "loss": 1.1933, + "step": 2118 + }, + { + "epoch": 1.1503800217155267, + "grad_norm": 5.887743000051609, + "learning_rate": 1.7974121821297178e-05, + "loss": 0.9143, + "step": 2119 + }, + { + "epoch": 1.1509229098805647, + "grad_norm": 6.264158167802545, + "learning_rate": 1.797199914418305e-05, + "loss": 0.805, + "step": 2120 + }, + { + "epoch": 1.1514657980456027, + "grad_norm": 7.747345696360016, + "learning_rate": 1.7969875481079363e-05, + "loss": 1.1374, + "step": 2121 + }, + { + "epoch": 1.1520086862106407, + "grad_norm": 5.932381639003466, + "learning_rate": 1.7967750832248774e-05, + "loss": 0.7801, + "step": 2122 + }, + { + "epoch": 1.1525515743756787, + "grad_norm": 5.646019326027894, + "learning_rate": 1.7965625197954064e-05, + "loss": 0.5053, + "step": 2123 + }, + { + "epoch": 1.1530944625407167, + "grad_norm": 6.085604242778934, + "learning_rate": 1.7963498578458135e-05, + "loss": 0.6711, + "step": 2124 + }, + { + "epoch": 1.1536373507057547, + "grad_norm": 6.390132028289734, + "learning_rate": 1.7961370974024008e-05, + "loss": 0.9897, + "step": 2125 + }, + { + "epoch": 1.1541802388707927, + "grad_norm": 5.052558465006266, + "learning_rate": 1.795924238491483e-05, + "loss": 0.8742, + "step": 2126 + }, + { + "epoch": 1.1547231270358307, + "grad_norm": 5.86713723099216, + "learning_rate": 1.7957112811393867e-05, + "loss": 0.965, + "step": 2127 + }, + { + "epoch": 1.1552660152008687, + "grad_norm": 6.092033075473227, + "learning_rate": 1.7954982253724515e-05, + "loss": 1.104, + "step": 2128 + }, + { + "epoch": 1.1558089033659067, + "grad_norm": 7.051795930424219, + "learning_rate": 1.7952850712170278e-05, + "loss": 1.0918, + "step": 2129 + }, + { + "epoch": 1.1563517915309447, + "grad_norm": 5.613114568974867, + "learning_rate": 1.7950718186994788e-05, + "loss": 0.6698, + "step": 2130 + }, + { + "epoch": 1.1568946796959827, + "grad_norm": 5.533439816776461, + "learning_rate": 1.7948584678461802e-05, + "loss": 1.4395, + "step": 2131 + }, + { + "epoch": 1.1574375678610207, + "grad_norm": 7.73758559868109, + "learning_rate": 1.7946450186835195e-05, + "loss": 1.0806, + "step": 2132 + }, + { + "epoch": 1.1579804560260587, + "grad_norm": 5.982378084297589, + "learning_rate": 1.7944314712378966e-05, + "loss": 1.0428, + "step": 2133 + }, + { + "epoch": 1.1585233441910967, + "grad_norm": 5.604421383798244, + "learning_rate": 1.7942178255357227e-05, + "loss": 0.9411, + "step": 2134 + }, + { + "epoch": 1.1590662323561347, + "grad_norm": 6.5637069585312435, + "learning_rate": 1.794004081603423e-05, + "loss": 0.8285, + "step": 2135 + }, + { + "epoch": 1.1596091205211727, + "grad_norm": 5.567535422744373, + "learning_rate": 1.7937902394674326e-05, + "loss": 1.0185, + "step": 2136 + }, + { + "epoch": 1.1601520086862107, + "grad_norm": 7.898198443147329, + "learning_rate": 1.7935762991542004e-05, + "loss": 0.8878, + "step": 2137 + }, + { + "epoch": 1.1606948968512487, + "grad_norm": 5.12407458777131, + "learning_rate": 1.7933622606901865e-05, + "loss": 0.9174, + "step": 2138 + }, + { + "epoch": 1.1612377850162867, + "grad_norm": 4.480165604953252, + "learning_rate": 1.7931481241018642e-05, + "loss": 0.9424, + "step": 2139 + }, + { + "epoch": 1.1617806731813247, + "grad_norm": 7.6708161194386495, + "learning_rate": 1.7929338894157173e-05, + "loss": 0.9875, + "step": 2140 + }, + { + "epoch": 1.1623235613463627, + "grad_norm": 9.749916559657342, + "learning_rate": 1.7927195566582435e-05, + "loss": 1.7168, + "step": 2141 + }, + { + "epoch": 1.1628664495114007, + "grad_norm": 5.50355689991458, + "learning_rate": 1.7925051258559516e-05, + "loss": 0.7957, + "step": 2142 + }, + { + "epoch": 1.1634093376764387, + "grad_norm": 5.910752363178181, + "learning_rate": 1.7922905970353627e-05, + "loss": 1.0525, + "step": 2143 + }, + { + "epoch": 1.1639522258414767, + "grad_norm": 6.523454927461212, + "learning_rate": 1.7920759702230098e-05, + "loss": 0.9938, + "step": 2144 + }, + { + "epoch": 1.1644951140065147, + "grad_norm": 6.963057538162332, + "learning_rate": 1.7918612454454387e-05, + "loss": 0.8677, + "step": 2145 + }, + { + "epoch": 1.1650380021715527, + "grad_norm": 6.55011455398606, + "learning_rate": 1.7916464227292067e-05, + "loss": 0.7268, + "step": 2146 + }, + { + "epoch": 1.1655808903365907, + "grad_norm": 6.47957363507043, + "learning_rate": 1.7914315021008836e-05, + "loss": 0.7449, + "step": 2147 + }, + { + "epoch": 1.1661237785016287, + "grad_norm": 6.574809242111077, + "learning_rate": 1.7912164835870506e-05, + "loss": 0.6999, + "step": 2148 + }, + { + "epoch": 1.1666666666666667, + "grad_norm": 8.307297080129953, + "learning_rate": 1.791001367214302e-05, + "loss": 1.0402, + "step": 2149 + }, + { + "epoch": 1.1672095548317047, + "grad_norm": 8.01469917888599, + "learning_rate": 1.7907861530092438e-05, + "loss": 1.1024, + "step": 2150 + }, + { + "epoch": 1.1677524429967427, + "grad_norm": 7.336381076899337, + "learning_rate": 1.7905708409984938e-05, + "loss": 0.9112, + "step": 2151 + }, + { + "epoch": 1.1682953311617807, + "grad_norm": 7.872105686379811, + "learning_rate": 1.7903554312086823e-05, + "loss": 0.9002, + "step": 2152 + }, + { + "epoch": 1.1688382193268188, + "grad_norm": 5.842761146463113, + "learning_rate": 1.7901399236664514e-05, + "loss": 0.9457, + "step": 2153 + }, + { + "epoch": 1.1693811074918568, + "grad_norm": 5.8100183378475005, + "learning_rate": 1.7899243183984552e-05, + "loss": 0.9675, + "step": 2154 + }, + { + "epoch": 1.1699239956568948, + "grad_norm": 5.208689262592845, + "learning_rate": 1.7897086154313604e-05, + "loss": 0.7562, + "step": 2155 + }, + { + "epoch": 1.1704668838219328, + "grad_norm": 7.695958654055186, + "learning_rate": 1.7894928147918458e-05, + "loss": 0.9995, + "step": 2156 + }, + { + "epoch": 1.1710097719869708, + "grad_norm": 5.8108200610880525, + "learning_rate": 1.789276916506601e-05, + "loss": 0.8179, + "step": 2157 + }, + { + "epoch": 1.1715526601520088, + "grad_norm": 5.776078861908261, + "learning_rate": 1.7890609206023298e-05, + "loss": 0.6612, + "step": 2158 + }, + { + "epoch": 1.1720955483170468, + "grad_norm": 6.458218841014727, + "learning_rate": 1.7888448271057463e-05, + "loss": 0.8419, + "step": 2159 + }, + { + "epoch": 1.1726384364820848, + "grad_norm": 7.236506499419343, + "learning_rate": 1.7886286360435772e-05, + "loss": 0.6336, + "step": 2160 + }, + { + "epoch": 1.1731813246471228, + "grad_norm": 7.20552304926134, + "learning_rate": 1.7884123474425614e-05, + "loss": 0.9157, + "step": 2161 + }, + { + "epoch": 1.1737242128121608, + "grad_norm": 8.090329675463652, + "learning_rate": 1.78819596132945e-05, + "loss": 1.6894, + "step": 2162 + }, + { + "epoch": 1.1742671009771988, + "grad_norm": 7.500510561892478, + "learning_rate": 1.787979477731006e-05, + "loss": 0.9469, + "step": 2163 + }, + { + "epoch": 1.1748099891422368, + "grad_norm": 7.6126459328076415, + "learning_rate": 1.7877628966740044e-05, + "loss": 0.9179, + "step": 2164 + }, + { + "epoch": 1.1753528773072748, + "grad_norm": 7.065751172246334, + "learning_rate": 1.787546218185232e-05, + "loss": 0.9649, + "step": 2165 + }, + { + "epoch": 1.1758957654723128, + "grad_norm": 6.923487399525626, + "learning_rate": 1.787329442291488e-05, + "loss": 1.0791, + "step": 2166 + }, + { + "epoch": 1.1764386536373508, + "grad_norm": 7.614292041669914, + "learning_rate": 1.7871125690195843e-05, + "loss": 1.2367, + "step": 2167 + }, + { + "epoch": 1.1769815418023888, + "grad_norm": 8.560468932577576, + "learning_rate": 1.7868955983963434e-05, + "loss": 1.0876, + "step": 2168 + }, + { + "epoch": 1.1775244299674268, + "grad_norm": 6.209836424501219, + "learning_rate": 1.786678530448601e-05, + "loss": 0.6251, + "step": 2169 + }, + { + "epoch": 1.1780673181324648, + "grad_norm": 5.926645140708837, + "learning_rate": 1.7864613652032035e-05, + "loss": 0.5939, + "step": 2170 + }, + { + "epoch": 1.1786102062975028, + "grad_norm": 5.460593849347875, + "learning_rate": 1.7862441026870114e-05, + "loss": 0.8769, + "step": 2171 + }, + { + "epoch": 1.1791530944625408, + "grad_norm": 8.063368903057139, + "learning_rate": 1.7860267429268954e-05, + "loss": 1.1492, + "step": 2172 + }, + { + "epoch": 1.1796959826275788, + "grad_norm": 6.246398264149539, + "learning_rate": 1.7858092859497392e-05, + "loss": 0.9146, + "step": 2173 + }, + { + "epoch": 1.1802388707926168, + "grad_norm": 9.483282452267977, + "learning_rate": 1.7855917317824383e-05, + "loss": 1.3714, + "step": 2174 + }, + { + "epoch": 1.1807817589576548, + "grad_norm": 7.377757446359213, + "learning_rate": 1.7853740804519e-05, + "loss": 0.7914, + "step": 2175 + }, + { + "epoch": 1.1813246471226928, + "grad_norm": 6.555194448443568, + "learning_rate": 1.7851563319850435e-05, + "loss": 1.0025, + "step": 2176 + }, + { + "epoch": 1.1818675352877308, + "grad_norm": 5.968820734002013, + "learning_rate": 1.7849384864088005e-05, + "loss": 1.0449, + "step": 2177 + }, + { + "epoch": 1.1824104234527688, + "grad_norm": 8.050731398373367, + "learning_rate": 1.7847205437501145e-05, + "loss": 1.2944, + "step": 2178 + }, + { + "epoch": 1.1829533116178068, + "grad_norm": 6.479093872062473, + "learning_rate": 1.784502504035941e-05, + "loss": 0.8312, + "step": 2179 + }, + { + "epoch": 1.1834961997828448, + "grad_norm": 9.645172625032446, + "learning_rate": 1.7842843672932473e-05, + "loss": 1.4877, + "step": 2180 + }, + { + "epoch": 1.1840390879478828, + "grad_norm": 7.811721210097083, + "learning_rate": 1.7840661335490133e-05, + "loss": 1.0861, + "step": 2181 + }, + { + "epoch": 1.1845819761129208, + "grad_norm": 5.939133632197409, + "learning_rate": 1.7838478028302303e-05, + "loss": 0.7359, + "step": 2182 + }, + { + "epoch": 1.1851248642779588, + "grad_norm": 7.121597254982603, + "learning_rate": 1.7836293751639017e-05, + "loss": 0.626, + "step": 2183 + }, + { + "epoch": 1.1856677524429968, + "grad_norm": 7.225734005859657, + "learning_rate": 1.783410850577043e-05, + "loss": 1.0651, + "step": 2184 + }, + { + "epoch": 1.1862106406080348, + "grad_norm": 6.3733995711762965, + "learning_rate": 1.783192229096682e-05, + "loss": 1.0399, + "step": 2185 + }, + { + "epoch": 1.1867535287730728, + "grad_norm": 7.215840642904109, + "learning_rate": 1.7829735107498576e-05, + "loss": 1.0958, + "step": 2186 + }, + { + "epoch": 1.1872964169381108, + "grad_norm": 7.2349637750528135, + "learning_rate": 1.7827546955636216e-05, + "loss": 0.8005, + "step": 2187 + }, + { + "epoch": 1.1878393051031488, + "grad_norm": 7.592927799049675, + "learning_rate": 1.7825357835650376e-05, + "loss": 1.0981, + "step": 2188 + }, + { + "epoch": 1.1883821932681868, + "grad_norm": 8.010034188435748, + "learning_rate": 1.7823167747811805e-05, + "loss": 1.3088, + "step": 2189 + }, + { + "epoch": 1.1889250814332248, + "grad_norm": 5.674359355487917, + "learning_rate": 1.7820976692391377e-05, + "loss": 0.8725, + "step": 2190 + }, + { + "epoch": 1.1894679695982628, + "grad_norm": 5.236415159647034, + "learning_rate": 1.781878466966009e-05, + "loss": 0.7162, + "step": 2191 + }, + { + "epoch": 1.1900108577633008, + "grad_norm": 5.677268305787393, + "learning_rate": 1.781659167988905e-05, + "loss": 0.7101, + "step": 2192 + }, + { + "epoch": 1.1905537459283388, + "grad_norm": 6.370730170332413, + "learning_rate": 1.7814397723349496e-05, + "loss": 0.9087, + "step": 2193 + }, + { + "epoch": 1.1910966340933768, + "grad_norm": 6.045624314512159, + "learning_rate": 1.7812202800312776e-05, + "loss": 1.1714, + "step": 2194 + }, + { + "epoch": 1.1916395222584149, + "grad_norm": 10.798495850925244, + "learning_rate": 1.7810006911050366e-05, + "loss": 1.1997, + "step": 2195 + }, + { + "epoch": 1.1921824104234529, + "grad_norm": 5.89010386817935, + "learning_rate": 1.780781005583385e-05, + "loss": 1.0702, + "step": 2196 + }, + { + "epoch": 1.1927252985884909, + "grad_norm": 6.09421937705502, + "learning_rate": 1.7805612234934946e-05, + "loss": 0.825, + "step": 2197 + }, + { + "epoch": 1.1932681867535289, + "grad_norm": 5.582107046729241, + "learning_rate": 1.780341344862548e-05, + "loss": 0.8548, + "step": 2198 + }, + { + "epoch": 1.1938110749185669, + "grad_norm": 7.598179521038406, + "learning_rate": 1.78012136971774e-05, + "loss": 1.0015, + "step": 2199 + }, + { + "epoch": 1.1943539630836049, + "grad_norm": 7.4068666684121585, + "learning_rate": 1.7799012980862777e-05, + "loss": 1.0645, + "step": 2200 + }, + { + "epoch": 1.1948968512486429, + "grad_norm": 6.087565282375561, + "learning_rate": 1.7796811299953796e-05, + "loss": 0.9169, + "step": 2201 + }, + { + "epoch": 1.1954397394136809, + "grad_norm": 6.440325481735819, + "learning_rate": 1.7794608654722772e-05, + "loss": 1.0707, + "step": 2202 + }, + { + "epoch": 1.1959826275787189, + "grad_norm": 7.562903613167714, + "learning_rate": 1.7792405045442125e-05, + "loss": 1.0086, + "step": 2203 + }, + { + "epoch": 1.1965255157437569, + "grad_norm": 6.2377346284888, + "learning_rate": 1.77902004723844e-05, + "loss": 0.6111, + "step": 2204 + }, + { + "epoch": 1.1970684039087949, + "grad_norm": 5.978299124706355, + "learning_rate": 1.7787994935822268e-05, + "loss": 0.5782, + "step": 2205 + }, + { + "epoch": 1.1976112920738329, + "grad_norm": 5.9766366336106715, + "learning_rate": 1.778578843602851e-05, + "loss": 0.8409, + "step": 2206 + }, + { + "epoch": 1.1981541802388709, + "grad_norm": 7.700010929432555, + "learning_rate": 1.7783580973276027e-05, + "loss": 1.091, + "step": 2207 + }, + { + "epoch": 1.1986970684039089, + "grad_norm": 8.043903647823342, + "learning_rate": 1.778137254783785e-05, + "loss": 1.2894, + "step": 2208 + }, + { + "epoch": 1.1992399565689469, + "grad_norm": 8.709529168595635, + "learning_rate": 1.777916315998711e-05, + "loss": 1.5601, + "step": 2209 + }, + { + "epoch": 1.1997828447339849, + "grad_norm": 6.234520197542074, + "learning_rate": 1.7776952809997073e-05, + "loss": 1.0414, + "step": 2210 + }, + { + "epoch": 1.200325732899023, + "grad_norm": 5.8108372705767, + "learning_rate": 1.7774741498141116e-05, + "loss": 0.8811, + "step": 2211 + }, + { + "epoch": 1.200868621064061, + "grad_norm": 6.423388338643525, + "learning_rate": 1.7772529224692744e-05, + "loss": 0.8371, + "step": 2212 + }, + { + "epoch": 1.201411509229099, + "grad_norm": 7.740996825838731, + "learning_rate": 1.7770315989925565e-05, + "loss": 0.9993, + "step": 2213 + }, + { + "epoch": 1.201954397394137, + "grad_norm": 6.33257469290407, + "learning_rate": 1.7768101794113323e-05, + "loss": 0.8563, + "step": 2214 + }, + { + "epoch": 1.202497285559175, + "grad_norm": 8.126954888789392, + "learning_rate": 1.776588663752987e-05, + "loss": 0.9479, + "step": 2215 + }, + { + "epoch": 1.203040173724213, + "grad_norm": 7.781450888823224, + "learning_rate": 1.7763670520449178e-05, + "loss": 1.0559, + "step": 2216 + }, + { + "epoch": 1.203583061889251, + "grad_norm": 6.615763983931515, + "learning_rate": 1.7761453443145348e-05, + "loss": 0.7463, + "step": 2217 + }, + { + "epoch": 1.204125950054289, + "grad_norm": 10.193394757123674, + "learning_rate": 1.7759235405892584e-05, + "loss": 1.3597, + "step": 2218 + }, + { + "epoch": 1.204668838219327, + "grad_norm": 7.365896137130193, + "learning_rate": 1.7757016408965217e-05, + "loss": 0.5822, + "step": 2219 + }, + { + "epoch": 1.205211726384365, + "grad_norm": 8.530190774406401, + "learning_rate": 1.77547964526377e-05, + "loss": 1.0075, + "step": 2220 + }, + { + "epoch": 1.205754614549403, + "grad_norm": 7.143571088845206, + "learning_rate": 1.7752575537184597e-05, + "loss": 0.9395, + "step": 2221 + }, + { + "epoch": 1.206297502714441, + "grad_norm": 7.790375269151535, + "learning_rate": 1.7750353662880595e-05, + "loss": 1.0556, + "step": 2222 + }, + { + "epoch": 1.206840390879479, + "grad_norm": 6.5033471136457255, + "learning_rate": 1.77481308300005e-05, + "loss": 0.7464, + "step": 2223 + }, + { + "epoch": 1.207383279044517, + "grad_norm": 7.06026002468489, + "learning_rate": 1.774590703881924e-05, + "loss": 0.6315, + "step": 2224 + }, + { + "epoch": 1.207926167209555, + "grad_norm": 8.773109114641178, + "learning_rate": 1.774368228961185e-05, + "loss": 0.9884, + "step": 2225 + }, + { + "epoch": 1.208469055374593, + "grad_norm": 7.343364313617542, + "learning_rate": 1.774145658265349e-05, + "loss": 1.1542, + "step": 2226 + }, + { + "epoch": 1.209011943539631, + "grad_norm": 5.953950803365371, + "learning_rate": 1.773922991821944e-05, + "loss": 0.8485, + "step": 2227 + }, + { + "epoch": 1.209554831704669, + "grad_norm": 8.905484324028157, + "learning_rate": 1.7737002296585105e-05, + "loss": 1.0171, + "step": 2228 + }, + { + "epoch": 1.210097719869707, + "grad_norm": 7.158929117808072, + "learning_rate": 1.7734773718025992e-05, + "loss": 0.6134, + "step": 2229 + }, + { + "epoch": 1.210640608034745, + "grad_norm": 8.342029051568577, + "learning_rate": 1.7732544182817737e-05, + "loss": 0.9775, + "step": 2230 + }, + { + "epoch": 1.211183496199783, + "grad_norm": 5.587278578636139, + "learning_rate": 1.7730313691236098e-05, + "loss": 0.9395, + "step": 2231 + }, + { + "epoch": 1.211726384364821, + "grad_norm": 6.1948883535007475, + "learning_rate": 1.7728082243556936e-05, + "loss": 0.9977, + "step": 2232 + }, + { + "epoch": 1.212269272529859, + "grad_norm": 6.095377323439518, + "learning_rate": 1.7725849840056248e-05, + "loss": 0.9141, + "step": 2233 + }, + { + "epoch": 1.212812160694897, + "grad_norm": 8.71522968216411, + "learning_rate": 1.7723616481010137e-05, + "loss": 1.4356, + "step": 2234 + }, + { + "epoch": 1.213355048859935, + "grad_norm": 5.416863970522961, + "learning_rate": 1.7721382166694834e-05, + "loss": 0.6594, + "step": 2235 + }, + { + "epoch": 1.213897937024973, + "grad_norm": 8.046753742934493, + "learning_rate": 1.7719146897386674e-05, + "loss": 1.1187, + "step": 2236 + }, + { + "epoch": 1.214440825190011, + "grad_norm": 6.625446441699582, + "learning_rate": 1.7716910673362123e-05, + "loss": 0.8731, + "step": 2237 + }, + { + "epoch": 1.214983713355049, + "grad_norm": 6.47539429072121, + "learning_rate": 1.7714673494897767e-05, + "loss": 0.9473, + "step": 2238 + }, + { + "epoch": 1.215526601520087, + "grad_norm": 7.458434209050833, + "learning_rate": 1.7712435362270292e-05, + "loss": 0.7844, + "step": 2239 + }, + { + "epoch": 1.216069489685125, + "grad_norm": 6.935847404573923, + "learning_rate": 1.7710196275756524e-05, + "loss": 0.8431, + "step": 2240 + }, + { + "epoch": 1.216612377850163, + "grad_norm": 8.015017144687471, + "learning_rate": 1.770795623563339e-05, + "loss": 1.2329, + "step": 2241 + }, + { + "epoch": 1.217155266015201, + "grad_norm": 6.914919733505818, + "learning_rate": 1.7705715242177944e-05, + "loss": 1.1335, + "step": 2242 + }, + { + "epoch": 1.217698154180239, + "grad_norm": 5.900061304870694, + "learning_rate": 1.7703473295667354e-05, + "loss": 0.7453, + "step": 2243 + }, + { + "epoch": 1.218241042345277, + "grad_norm": 4.959614765477869, + "learning_rate": 1.770123039637891e-05, + "loss": 0.6093, + "step": 2244 + }, + { + "epoch": 1.2187839305103148, + "grad_norm": 8.038465432918807, + "learning_rate": 1.769898654459002e-05, + "loss": 1.0029, + "step": 2245 + }, + { + "epoch": 1.219326818675353, + "grad_norm": 8.08635156220498, + "learning_rate": 1.7696741740578204e-05, + "loss": 1.0839, + "step": 2246 + }, + { + "epoch": 1.2198697068403908, + "grad_norm": 7.985335036181301, + "learning_rate": 1.7694495984621097e-05, + "loss": 1.262, + "step": 2247 + }, + { + "epoch": 1.220412595005429, + "grad_norm": 8.31037377117703, + "learning_rate": 1.769224927699647e-05, + "loss": 1.0265, + "step": 2248 + }, + { + "epoch": 1.2209554831704668, + "grad_norm": 6.241752286447289, + "learning_rate": 1.769000161798219e-05, + "loss": 0.6179, + "step": 2249 + }, + { + "epoch": 1.221498371335505, + "grad_norm": 6.767956890107702, + "learning_rate": 1.7687753007856253e-05, + "loss": 0.9897, + "step": 2250 + }, + { + "epoch": 1.2220412595005428, + "grad_norm": 7.119621854687504, + "learning_rate": 1.7685503446896772e-05, + "loss": 1.0529, + "step": 2251 + }, + { + "epoch": 1.222584147665581, + "grad_norm": 6.37786350129091, + "learning_rate": 1.7683252935381976e-05, + "loss": 0.8575, + "step": 2252 + }, + { + "epoch": 1.2231270358306188, + "grad_norm": 9.280258739154695, + "learning_rate": 1.768100147359021e-05, + "loss": 0.959, + "step": 2253 + }, + { + "epoch": 1.223669923995657, + "grad_norm": 7.172898793975323, + "learning_rate": 1.7678749061799942e-05, + "loss": 1.3745, + "step": 2254 + }, + { + "epoch": 1.2242128121606948, + "grad_norm": 5.3297508420409825, + "learning_rate": 1.7676495700289753e-05, + "loss": 0.6697, + "step": 2255 + }, + { + "epoch": 1.224755700325733, + "grad_norm": 6.518884308754971, + "learning_rate": 1.767424138933834e-05, + "loss": 0.9854, + "step": 2256 + }, + { + "epoch": 1.2252985884907708, + "grad_norm": 7.939498291049385, + "learning_rate": 1.767198612922452e-05, + "loss": 0.8918, + "step": 2257 + }, + { + "epoch": 1.225841476655809, + "grad_norm": 6.309602610605377, + "learning_rate": 1.7669729920227226e-05, + "loss": 0.9939, + "step": 2258 + }, + { + "epoch": 1.2263843648208468, + "grad_norm": 7.462197154770194, + "learning_rate": 1.7667472762625518e-05, + "loss": 1.1299, + "step": 2259 + }, + { + "epoch": 1.226927252985885, + "grad_norm": 6.700031938849371, + "learning_rate": 1.7665214656698555e-05, + "loss": 0.8655, + "step": 2260 + }, + { + "epoch": 1.2274701411509228, + "grad_norm": 6.72124660471967, + "learning_rate": 1.7662955602725627e-05, + "loss": 0.7383, + "step": 2261 + }, + { + "epoch": 1.228013029315961, + "grad_norm": 6.071348581517431, + "learning_rate": 1.7660695600986137e-05, + "loss": 0.8571, + "step": 2262 + }, + { + "epoch": 1.2285559174809988, + "grad_norm": 5.914184858865564, + "learning_rate": 1.7658434651759605e-05, + "loss": 0.7494, + "step": 2263 + }, + { + "epoch": 1.229098805646037, + "grad_norm": 12.685614201985125, + "learning_rate": 1.7656172755325675e-05, + "loss": 1.608, + "step": 2264 + }, + { + "epoch": 1.2296416938110748, + "grad_norm": 7.001237249554534, + "learning_rate": 1.765390991196409e-05, + "loss": 1.0437, + "step": 2265 + }, + { + "epoch": 1.230184581976113, + "grad_norm": 8.232094967642231, + "learning_rate": 1.765164612195473e-05, + "loss": 0.8574, + "step": 2266 + }, + { + "epoch": 1.2307274701411508, + "grad_norm": 6.367083948330669, + "learning_rate": 1.7649381385577582e-05, + "loss": 0.7106, + "step": 2267 + }, + { + "epoch": 1.231270358306189, + "grad_norm": 6.2062819181758195, + "learning_rate": 1.7647115703112756e-05, + "loss": 0.6838, + "step": 2268 + }, + { + "epoch": 1.2318132464712268, + "grad_norm": 6.896096048797596, + "learning_rate": 1.7644849074840472e-05, + "loss": 0.9595, + "step": 2269 + }, + { + "epoch": 1.232356134636265, + "grad_norm": 5.915145487902616, + "learning_rate": 1.7642581501041067e-05, + "loss": 0.5741, + "step": 2270 + }, + { + "epoch": 1.2328990228013028, + "grad_norm": 5.966871119524657, + "learning_rate": 1.7640312981995004e-05, + "loss": 0.8686, + "step": 2271 + }, + { + "epoch": 1.233441910966341, + "grad_norm": 6.835680775363237, + "learning_rate": 1.7638043517982854e-05, + "loss": 0.9532, + "step": 2272 + }, + { + "epoch": 1.2339847991313788, + "grad_norm": 8.228029492129801, + "learning_rate": 1.7635773109285306e-05, + "loss": 1.0834, + "step": 2273 + }, + { + "epoch": 1.234527687296417, + "grad_norm": 8.272927259623488, + "learning_rate": 1.7633501756183174e-05, + "loss": 1.2591, + "step": 2274 + }, + { + "epoch": 1.2350705754614548, + "grad_norm": 6.755748928056399, + "learning_rate": 1.7631229458957377e-05, + "loss": 0.997, + "step": 2275 + }, + { + "epoch": 1.235613463626493, + "grad_norm": 7.335027842642818, + "learning_rate": 1.762895621788896e-05, + "loss": 0.8695, + "step": 2276 + }, + { + "epoch": 1.2361563517915308, + "grad_norm": 5.237971036796901, + "learning_rate": 1.7626682033259077e-05, + "loss": 0.6846, + "step": 2277 + }, + { + "epoch": 1.236699239956569, + "grad_norm": 6.333455514649705, + "learning_rate": 1.7624406905349004e-05, + "loss": 0.7532, + "step": 2278 + }, + { + "epoch": 1.2372421281216068, + "grad_norm": 6.255556726847127, + "learning_rate": 1.7622130834440136e-05, + "loss": 0.7801, + "step": 2279 + }, + { + "epoch": 1.237785016286645, + "grad_norm": 5.602832960919912, + "learning_rate": 1.761985382081398e-05, + "loss": 0.5105, + "step": 2280 + }, + { + "epoch": 1.2383279044516828, + "grad_norm": 5.814619917758832, + "learning_rate": 1.7617575864752155e-05, + "loss": 0.9034, + "step": 2281 + }, + { + "epoch": 1.238870792616721, + "grad_norm": 6.743402665368447, + "learning_rate": 1.7615296966536407e-05, + "loss": 0.7713, + "step": 2282 + }, + { + "epoch": 1.2394136807817588, + "grad_norm": 7.176469994889228, + "learning_rate": 1.7613017126448597e-05, + "loss": 0.957, + "step": 2283 + }, + { + "epoch": 1.239956568946797, + "grad_norm": 8.5427155590311, + "learning_rate": 1.7610736344770693e-05, + "loss": 0.8789, + "step": 2284 + }, + { + "epoch": 1.2404994571118348, + "grad_norm": 6.674205629131359, + "learning_rate": 1.7608454621784786e-05, + "loss": 1.0239, + "step": 2285 + }, + { + "epoch": 1.241042345276873, + "grad_norm": 7.687894188099568, + "learning_rate": 1.7606171957773088e-05, + "loss": 1.3164, + "step": 2286 + }, + { + "epoch": 1.2415852334419109, + "grad_norm": 6.331613648212601, + "learning_rate": 1.760388835301792e-05, + "loss": 0.8047, + "step": 2287 + }, + { + "epoch": 1.242128121606949, + "grad_norm": 7.0776252458357725, + "learning_rate": 1.760160380780172e-05, + "loss": 1.1893, + "step": 2288 + }, + { + "epoch": 1.2426710097719869, + "grad_norm": 5.434346705496786, + "learning_rate": 1.7599318322407047e-05, + "loss": 0.763, + "step": 2289 + }, + { + "epoch": 1.243213897937025, + "grad_norm": 6.0271106768536775, + "learning_rate": 1.759703189711657e-05, + "loss": 1.1129, + "step": 2290 + }, + { + "epoch": 1.2437567861020629, + "grad_norm": 7.286292908201965, + "learning_rate": 1.7594744532213088e-05, + "loss": 1.1136, + "step": 2291 + }, + { + "epoch": 1.244299674267101, + "grad_norm": 6.685086643433707, + "learning_rate": 1.7592456227979488e-05, + "loss": 0.724, + "step": 2292 + }, + { + "epoch": 1.2448425624321389, + "grad_norm": 7.184174802374185, + "learning_rate": 1.7590166984698807e-05, + "loss": 1.1471, + "step": 2293 + }, + { + "epoch": 1.245385450597177, + "grad_norm": 7.787563610771557, + "learning_rate": 1.7587876802654176e-05, + "loss": 1.01, + "step": 2294 + }, + { + "epoch": 1.2459283387622149, + "grad_norm": 7.036440767647583, + "learning_rate": 1.7585585682128848e-05, + "loss": 0.8612, + "step": 2295 + }, + { + "epoch": 1.246471226927253, + "grad_norm": 6.6372154433448705, + "learning_rate": 1.7583293623406192e-05, + "loss": 1.3394, + "step": 2296 + }, + { + "epoch": 1.2470141150922909, + "grad_norm": 7.707986168491473, + "learning_rate": 1.7581000626769697e-05, + "loss": 0.8688, + "step": 2297 + }, + { + "epoch": 1.247557003257329, + "grad_norm": 7.862359368033938, + "learning_rate": 1.757870669250296e-05, + "loss": 1.1999, + "step": 2298 + }, + { + "epoch": 1.2480998914223669, + "grad_norm": 7.942280078814929, + "learning_rate": 1.75764118208897e-05, + "loss": 1.241, + "step": 2299 + }, + { + "epoch": 1.248642779587405, + "grad_norm": 7.258825765708901, + "learning_rate": 1.7574116012213753e-05, + "loss": 0.8336, + "step": 2300 + }, + { + "epoch": 1.2491856677524429, + "grad_norm": 7.0149158036864065, + "learning_rate": 1.7571819266759066e-05, + "loss": 0.8637, + "step": 2301 + }, + { + "epoch": 1.249728555917481, + "grad_norm": 7.069351546856077, + "learning_rate": 1.7569521584809703e-05, + "loss": 0.8146, + "step": 2302 + }, + { + "epoch": 1.250271444082519, + "grad_norm": 4.6424312647112425, + "learning_rate": 1.756722296664985e-05, + "loss": 0.5056, + "step": 2303 + }, + { + "epoch": 1.2508143322475571, + "grad_norm": 7.857412677197525, + "learning_rate": 1.7564923412563797e-05, + "loss": 1.3072, + "step": 2304 + }, + { + "epoch": 1.251357220412595, + "grad_norm": 6.081845076053873, + "learning_rate": 1.7562622922835962e-05, + "loss": 1.0212, + "step": 2305 + }, + { + "epoch": 1.2519001085776331, + "grad_norm": 5.244697868381976, + "learning_rate": 1.7560321497750867e-05, + "loss": 0.9677, + "step": 2306 + }, + { + "epoch": 1.252442996742671, + "grad_norm": 11.027915321544773, + "learning_rate": 1.7558019137593162e-05, + "loss": 0.8984, + "step": 2307 + }, + { + "epoch": 1.2529858849077091, + "grad_norm": 5.955816485468965, + "learning_rate": 1.7555715842647603e-05, + "loss": 0.6282, + "step": 2308 + }, + { + "epoch": 1.253528773072747, + "grad_norm": 8.925991341226974, + "learning_rate": 1.755341161319907e-05, + "loss": 1.2121, + "step": 2309 + }, + { + "epoch": 1.2540716612377851, + "grad_norm": 7.947406679428915, + "learning_rate": 1.7551106449532547e-05, + "loss": 1.0545, + "step": 2310 + }, + { + "epoch": 1.254614549402823, + "grad_norm": 7.329950391674132, + "learning_rate": 1.7548800351933147e-05, + "loss": 1.0585, + "step": 2311 + }, + { + "epoch": 1.2551574375678611, + "grad_norm": 5.0931347466751955, + "learning_rate": 1.7546493320686084e-05, + "loss": 0.5533, + "step": 2312 + }, + { + "epoch": 1.255700325732899, + "grad_norm": 7.125279667319334, + "learning_rate": 1.7544185356076703e-05, + "loss": 1.1011, + "step": 2313 + }, + { + "epoch": 1.2562432138979371, + "grad_norm": 8.060697991979964, + "learning_rate": 1.7541876458390453e-05, + "loss": 0.8732, + "step": 2314 + }, + { + "epoch": 1.256786102062975, + "grad_norm": 9.732797197135078, + "learning_rate": 1.75395666279129e-05, + "loss": 1.5663, + "step": 2315 + }, + { + "epoch": 1.2573289902280131, + "grad_norm": 5.3673655570709435, + "learning_rate": 1.7537255864929733e-05, + "loss": 0.7057, + "step": 2316 + }, + { + "epoch": 1.257871878393051, + "grad_norm": 5.997561584131186, + "learning_rate": 1.7534944169726752e-05, + "loss": 0.8526, + "step": 2317 + }, + { + "epoch": 1.2584147665580891, + "grad_norm": 7.236766116786611, + "learning_rate": 1.753263154258986e-05, + "loss": 0.8921, + "step": 2318 + }, + { + "epoch": 1.258957654723127, + "grad_norm": 7.663917187467161, + "learning_rate": 1.7530317983805096e-05, + "loss": 0.9309, + "step": 2319 + }, + { + "epoch": 1.2595005428881652, + "grad_norm": 5.704193869412544, + "learning_rate": 1.75280034936586e-05, + "loss": 0.6291, + "step": 2320 + }, + { + "epoch": 1.260043431053203, + "grad_norm": 7.365455665485428, + "learning_rate": 1.7525688072436636e-05, + "loss": 0.9946, + "step": 2321 + }, + { + "epoch": 1.2605863192182412, + "grad_norm": 4.8267670121648205, + "learning_rate": 1.752337172042558e-05, + "loss": 0.8254, + "step": 2322 + }, + { + "epoch": 1.261129207383279, + "grad_norm": 7.231116660749922, + "learning_rate": 1.752105443791192e-05, + "loss": 1.3835, + "step": 2323 + }, + { + "epoch": 1.2616720955483172, + "grad_norm": 6.02849395913157, + "learning_rate": 1.7518736225182253e-05, + "loss": 0.6229, + "step": 2324 + }, + { + "epoch": 1.262214983713355, + "grad_norm": 6.458823304206971, + "learning_rate": 1.7516417082523314e-05, + "loss": 0.7707, + "step": 2325 + }, + { + "epoch": 1.2627578718783932, + "grad_norm": 7.163598438660781, + "learning_rate": 1.7514097010221927e-05, + "loss": 1.0007, + "step": 2326 + }, + { + "epoch": 1.263300760043431, + "grad_norm": 6.607394766457111, + "learning_rate": 1.7511776008565043e-05, + "loss": 1.0289, + "step": 2327 + }, + { + "epoch": 1.2638436482084692, + "grad_norm": 8.20658895752514, + "learning_rate": 1.7509454077839736e-05, + "loss": 0.7553, + "step": 2328 + }, + { + "epoch": 1.264386536373507, + "grad_norm": 8.963409997263136, + "learning_rate": 1.750713121833318e-05, + "loss": 0.9558, + "step": 2329 + }, + { + "epoch": 1.2649294245385452, + "grad_norm": 6.463632425962391, + "learning_rate": 1.7504807430332668e-05, + "loss": 0.6553, + "step": 2330 + }, + { + "epoch": 1.265472312703583, + "grad_norm": 6.936173778617146, + "learning_rate": 1.7502482714125612e-05, + "loss": 0.7861, + "step": 2331 + }, + { + "epoch": 1.2660152008686212, + "grad_norm": 6.975529559460763, + "learning_rate": 1.7500157069999536e-05, + "loss": 0.665, + "step": 2332 + }, + { + "epoch": 1.266558089033659, + "grad_norm": 6.768480922885786, + "learning_rate": 1.7497830498242082e-05, + "loss": 1.0163, + "step": 2333 + }, + { + "epoch": 1.2671009771986972, + "grad_norm": 6.357421024100016, + "learning_rate": 1.7495502999141004e-05, + "loss": 0.5897, + "step": 2334 + }, + { + "epoch": 1.267643865363735, + "grad_norm": 7.82539792030223, + "learning_rate": 1.7493174572984168e-05, + "loss": 1.0509, + "step": 2335 + }, + { + "epoch": 1.2681867535287732, + "grad_norm": 7.6523935004894215, + "learning_rate": 1.7490845220059554e-05, + "loss": 0.7603, + "step": 2336 + }, + { + "epoch": 1.268729641693811, + "grad_norm": 6.677360283026087, + "learning_rate": 1.7488514940655267e-05, + "loss": 1.1797, + "step": 2337 + }, + { + "epoch": 1.2692725298588492, + "grad_norm": 7.837606738097467, + "learning_rate": 1.7486183735059517e-05, + "loss": 1.1995, + "step": 2338 + }, + { + "epoch": 1.269815418023887, + "grad_norm": 8.913928805014676, + "learning_rate": 1.7483851603560634e-05, + "loss": 1.6007, + "step": 2339 + }, + { + "epoch": 1.2703583061889252, + "grad_norm": 8.467363918664116, + "learning_rate": 1.7481518546447058e-05, + "loss": 0.9515, + "step": 2340 + }, + { + "epoch": 1.270901194353963, + "grad_norm": 6.758882762498143, + "learning_rate": 1.747918456400734e-05, + "loss": 0.7935, + "step": 2341 + }, + { + "epoch": 1.2714440825190012, + "grad_norm": 5.455764070482765, + "learning_rate": 1.7476849656530158e-05, + "loss": 0.4623, + "step": 2342 + }, + { + "epoch": 1.271986970684039, + "grad_norm": 8.417334291043444, + "learning_rate": 1.7474513824304294e-05, + "loss": 1.1563, + "step": 2343 + }, + { + "epoch": 1.2725298588490772, + "grad_norm": 5.61084844590966, + "learning_rate": 1.7472177067618646e-05, + "loss": 0.8566, + "step": 2344 + }, + { + "epoch": 1.273072747014115, + "grad_norm": 7.726478476188633, + "learning_rate": 1.746983938676223e-05, + "loss": 1.0568, + "step": 2345 + }, + { + "epoch": 1.2736156351791532, + "grad_norm": 8.882013629978706, + "learning_rate": 1.7467500782024177e-05, + "loss": 1.0921, + "step": 2346 + }, + { + "epoch": 1.274158523344191, + "grad_norm": 6.976574380776494, + "learning_rate": 1.746516125369372e-05, + "loss": 1.1222, + "step": 2347 + }, + { + "epoch": 1.2747014115092292, + "grad_norm": 7.641347465211616, + "learning_rate": 1.7462820802060224e-05, + "loss": 1.0672, + "step": 2348 + }, + { + "epoch": 1.275244299674267, + "grad_norm": 6.473243122755322, + "learning_rate": 1.7460479427413156e-05, + "loss": 0.9019, + "step": 2349 + }, + { + "epoch": 1.2757871878393052, + "grad_norm": 6.494958002593054, + "learning_rate": 1.7458137130042105e-05, + "loss": 0.7862, + "step": 2350 + }, + { + "epoch": 1.276330076004343, + "grad_norm": 6.54459648554074, + "learning_rate": 1.7455793910236764e-05, + "loss": 0.9572, + "step": 2351 + }, + { + "epoch": 1.2768729641693812, + "grad_norm": 6.249218254368162, + "learning_rate": 1.7453449768286952e-05, + "loss": 0.9415, + "step": 2352 + }, + { + "epoch": 1.277415852334419, + "grad_norm": 6.335956299458767, + "learning_rate": 1.7451104704482592e-05, + "loss": 0.7647, + "step": 2353 + }, + { + "epoch": 1.2779587404994572, + "grad_norm": 6.296362619304004, + "learning_rate": 1.744875871911373e-05, + "loss": 0.6293, + "step": 2354 + }, + { + "epoch": 1.278501628664495, + "grad_norm": 6.629983181103369, + "learning_rate": 1.7446411812470516e-05, + "loss": 0.7816, + "step": 2355 + }, + { + "epoch": 1.2790445168295332, + "grad_norm": 6.921460108918273, + "learning_rate": 1.744406398484322e-05, + "loss": 0.6373, + "step": 2356 + }, + { + "epoch": 1.279587404994571, + "grad_norm": 9.4421099603369, + "learning_rate": 1.7441715236522228e-05, + "loss": 1.5291, + "step": 2357 + }, + { + "epoch": 1.2801302931596092, + "grad_norm": 7.838644313955947, + "learning_rate": 1.7439365567798035e-05, + "loss": 0.7934, + "step": 2358 + }, + { + "epoch": 1.280673181324647, + "grad_norm": 5.7659414273416, + "learning_rate": 1.743701497896125e-05, + "loss": 0.8358, + "step": 2359 + }, + { + "epoch": 1.2812160694896852, + "grad_norm": 4.731198780322981, + "learning_rate": 1.7434663470302602e-05, + "loss": 0.7357, + "step": 2360 + }, + { + "epoch": 1.281758957654723, + "grad_norm": 8.993227122901354, + "learning_rate": 1.7432311042112926e-05, + "loss": 1.0679, + "step": 2361 + }, + { + "epoch": 1.2823018458197613, + "grad_norm": 7.193753575449645, + "learning_rate": 1.7429957694683175e-05, + "loss": 0.9257, + "step": 2362 + }, + { + "epoch": 1.282844733984799, + "grad_norm": 7.1027186942584635, + "learning_rate": 1.7427603428304416e-05, + "loss": 0.64, + "step": 2363 + }, + { + "epoch": 1.2833876221498373, + "grad_norm": 9.441739719572269, + "learning_rate": 1.7425248243267824e-05, + "loss": 1.3236, + "step": 2364 + }, + { + "epoch": 1.283930510314875, + "grad_norm": 6.478965109994284, + "learning_rate": 1.7422892139864696e-05, + "loss": 0.7726, + "step": 2365 + }, + { + "epoch": 1.2844733984799133, + "grad_norm": 6.372951649068815, + "learning_rate": 1.742053511838644e-05, + "loss": 0.7785, + "step": 2366 + }, + { + "epoch": 1.285016286644951, + "grad_norm": 10.052356612127076, + "learning_rate": 1.7418177179124574e-05, + "loss": 1.7877, + "step": 2367 + }, + { + "epoch": 1.2855591748099893, + "grad_norm": 5.24633801429042, + "learning_rate": 1.741581832237073e-05, + "loss": 0.4195, + "step": 2368 + }, + { + "epoch": 1.286102062975027, + "grad_norm": 7.958362658884473, + "learning_rate": 1.7413458548416656e-05, + "loss": 1.1782, + "step": 2369 + }, + { + "epoch": 1.2866449511400653, + "grad_norm": 7.923031881378813, + "learning_rate": 1.7411097857554216e-05, + "loss": 1.0861, + "step": 2370 + }, + { + "epoch": 1.287187839305103, + "grad_norm": 6.53099588563486, + "learning_rate": 1.7408736250075378e-05, + "loss": 0.8717, + "step": 2371 + }, + { + "epoch": 1.2877307274701413, + "grad_norm": 7.364911313605863, + "learning_rate": 1.740637372627224e-05, + "loss": 0.9341, + "step": 2372 + }, + { + "epoch": 1.288273615635179, + "grad_norm": 8.083622859612968, + "learning_rate": 1.740401028643699e-05, + "loss": 1.1654, + "step": 2373 + }, + { + "epoch": 1.2888165038002173, + "grad_norm": 6.561315990238234, + "learning_rate": 1.7401645930861944e-05, + "loss": 0.8088, + "step": 2374 + }, + { + "epoch": 1.289359391965255, + "grad_norm": 6.926193447339149, + "learning_rate": 1.739928065983954e-05, + "loss": 0.8335, + "step": 2375 + }, + { + "epoch": 1.2899022801302933, + "grad_norm": 7.876730774519387, + "learning_rate": 1.7396914473662306e-05, + "loss": 0.9969, + "step": 2376 + }, + { + "epoch": 1.290445168295331, + "grad_norm": 6.513645878383447, + "learning_rate": 1.73945473726229e-05, + "loss": 0.9302, + "step": 2377 + }, + { + "epoch": 1.2909880564603693, + "grad_norm": 6.441562626276137, + "learning_rate": 1.73921793570141e-05, + "loss": 1.0318, + "step": 2378 + }, + { + "epoch": 1.291530944625407, + "grad_norm": 5.226443422910506, + "learning_rate": 1.7389810427128766e-05, + "loss": 0.6142, + "step": 2379 + }, + { + "epoch": 1.2920738327904453, + "grad_norm": 7.436131751002477, + "learning_rate": 1.7387440583259906e-05, + "loss": 0.9907, + "step": 2380 + }, + { + "epoch": 1.292616720955483, + "grad_norm": 6.335395136223753, + "learning_rate": 1.7385069825700615e-05, + "loss": 0.8431, + "step": 2381 + }, + { + "epoch": 1.2931596091205213, + "grad_norm": 8.93013417703781, + "learning_rate": 1.738269815474412e-05, + "loss": 1.3162, + "step": 2382 + }, + { + "epoch": 1.293702497285559, + "grad_norm": 6.616000665685319, + "learning_rate": 1.7380325570683754e-05, + "loss": 0.81, + "step": 2383 + }, + { + "epoch": 1.2942453854505973, + "grad_norm": 9.126678309827314, + "learning_rate": 1.737795207381296e-05, + "loss": 1.4976, + "step": 2384 + }, + { + "epoch": 1.294788273615635, + "grad_norm": 9.726273416685123, + "learning_rate": 1.737557766442529e-05, + "loss": 1.0284, + "step": 2385 + }, + { + "epoch": 1.2953311617806733, + "grad_norm": 6.207571191669611, + "learning_rate": 1.737320234281442e-05, + "loss": 0.9409, + "step": 2386 + }, + { + "epoch": 1.295874049945711, + "grad_norm": 6.150632184961992, + "learning_rate": 1.7370826109274136e-05, + "loss": 0.7558, + "step": 2387 + }, + { + "epoch": 1.2964169381107493, + "grad_norm": 6.555861715157727, + "learning_rate": 1.736844896409833e-05, + "loss": 1.4217, + "step": 2388 + }, + { + "epoch": 1.296959826275787, + "grad_norm": 7.291607947091272, + "learning_rate": 1.7366070907581015e-05, + "loss": 0.9969, + "step": 2389 + }, + { + "epoch": 1.2975027144408253, + "grad_norm": 7.003873752430508, + "learning_rate": 1.7363691940016307e-05, + "loss": 0.9569, + "step": 2390 + }, + { + "epoch": 1.298045602605863, + "grad_norm": 6.355975398274633, + "learning_rate": 1.7361312061698444e-05, + "loss": 0.9244, + "step": 2391 + }, + { + "epoch": 1.2985884907709013, + "grad_norm": 5.093818720596728, + "learning_rate": 1.7358931272921773e-05, + "loss": 0.5581, + "step": 2392 + }, + { + "epoch": 1.299131378935939, + "grad_norm": 4.532047411652702, + "learning_rate": 1.7356549573980753e-05, + "loss": 0.5153, + "step": 2393 + }, + { + "epoch": 1.2996742671009773, + "grad_norm": 7.378796483754393, + "learning_rate": 1.735416696516996e-05, + "loss": 0.9631, + "step": 2394 + }, + { + "epoch": 1.3002171552660151, + "grad_norm": 6.186232096125507, + "learning_rate": 1.7351783446784075e-05, + "loss": 0.7876, + "step": 2395 + }, + { + "epoch": 1.3007600434310533, + "grad_norm": 7.851387526494335, + "learning_rate": 1.7349399019117897e-05, + "loss": 0.8983, + "step": 2396 + }, + { + "epoch": 1.3013029315960911, + "grad_norm": 7.275506442499215, + "learning_rate": 1.7347013682466335e-05, + "loss": 0.8334, + "step": 2397 + }, + { + "epoch": 1.3018458197611293, + "grad_norm": 6.264072188984264, + "learning_rate": 1.7344627437124407e-05, + "loss": 0.6585, + "step": 2398 + }, + { + "epoch": 1.3023887079261671, + "grad_norm": 7.537073303434296, + "learning_rate": 1.734224028338726e-05, + "loss": 1.1033, + "step": 2399 + }, + { + "epoch": 1.3029315960912053, + "grad_norm": 6.850281208830108, + "learning_rate": 1.7339852221550126e-05, + "loss": 0.8133, + "step": 2400 + }, + { + "epoch": 1.3034744842562431, + "grad_norm": 8.089656198012262, + "learning_rate": 1.7337463251908374e-05, + "loss": 1.1342, + "step": 2401 + }, + { + "epoch": 1.3040173724212814, + "grad_norm": 6.2919884156176105, + "learning_rate": 1.733507337475747e-05, + "loss": 0.6533, + "step": 2402 + }, + { + "epoch": 1.3045602605863191, + "grad_norm": 5.114953151143873, + "learning_rate": 1.7332682590393008e-05, + "loss": 0.7451, + "step": 2403 + }, + { + "epoch": 1.3051031487513574, + "grad_norm": 6.18470588828327, + "learning_rate": 1.733029089911067e-05, + "loss": 1.1299, + "step": 2404 + }, + { + "epoch": 1.3056460369163951, + "grad_norm": 5.950129534147113, + "learning_rate": 1.7327898301206273e-05, + "loss": 0.7329, + "step": 2405 + }, + { + "epoch": 1.3061889250814334, + "grad_norm": 6.293661255114301, + "learning_rate": 1.7325504796975732e-05, + "loss": 0.7535, + "step": 2406 + }, + { + "epoch": 1.3067318132464711, + "grad_norm": 7.682262802579054, + "learning_rate": 1.732311038671509e-05, + "loss": 1.2399, + "step": 2407 + }, + { + "epoch": 1.3072747014115094, + "grad_norm": 7.314299514485556, + "learning_rate": 1.7320715070720478e-05, + "loss": 0.9298, + "step": 2408 + }, + { + "epoch": 1.3078175895765471, + "grad_norm": 8.921126356976833, + "learning_rate": 1.7318318849288158e-05, + "loss": 1.4254, + "step": 2409 + }, + { + "epoch": 1.3083604777415854, + "grad_norm": 7.060288459365394, + "learning_rate": 1.7315921722714503e-05, + "loss": 0.9428, + "step": 2410 + }, + { + "epoch": 1.3089033659066232, + "grad_norm": 8.568251970472561, + "learning_rate": 1.7313523691295988e-05, + "loss": 0.9011, + "step": 2411 + }, + { + "epoch": 1.3094462540716614, + "grad_norm": 6.049363733159396, + "learning_rate": 1.7311124755329206e-05, + "loss": 1.0718, + "step": 2412 + }, + { + "epoch": 1.3099891422366992, + "grad_norm": 8.287809242433424, + "learning_rate": 1.7308724915110864e-05, + "loss": 0.9137, + "step": 2413 + }, + { + "epoch": 1.3105320304017374, + "grad_norm": 8.436233795197621, + "learning_rate": 1.7306324170937774e-05, + "loss": 1.2427, + "step": 2414 + }, + { + "epoch": 1.3110749185667752, + "grad_norm": 7.035710404708695, + "learning_rate": 1.7303922523106863e-05, + "loss": 0.8206, + "step": 2415 + }, + { + "epoch": 1.3116178067318134, + "grad_norm": 5.9728457186472275, + "learning_rate": 1.730151997191518e-05, + "loss": 0.7307, + "step": 2416 + }, + { + "epoch": 1.3121606948968512, + "grad_norm": 6.296494925379207, + "learning_rate": 1.7299116517659865e-05, + "loss": 0.9808, + "step": 2417 + }, + { + "epoch": 1.3127035830618892, + "grad_norm": 7.672949954931739, + "learning_rate": 1.7296712160638192e-05, + "loss": 1.0289, + "step": 2418 + }, + { + "epoch": 1.3132464712269272, + "grad_norm": 5.958844838187463, + "learning_rate": 1.7294306901147525e-05, + "loss": 0.9657, + "step": 2419 + }, + { + "epoch": 1.3137893593919652, + "grad_norm": 6.5874736925476745, + "learning_rate": 1.7291900739485356e-05, + "loss": 0.8701, + "step": 2420 + }, + { + "epoch": 1.3143322475570032, + "grad_norm": 8.273467490572564, + "learning_rate": 1.7289493675949282e-05, + "loss": 0.9353, + "step": 2421 + }, + { + "epoch": 1.3148751357220412, + "grad_norm": 8.115143215098843, + "learning_rate": 1.7287085710837013e-05, + "loss": 0.6241, + "step": 2422 + }, + { + "epoch": 1.3154180238870792, + "grad_norm": 7.004703854660018, + "learning_rate": 1.7284676844446368e-05, + "loss": 0.801, + "step": 2423 + }, + { + "epoch": 1.3159609120521172, + "grad_norm": 6.039976470078682, + "learning_rate": 1.728226707707528e-05, + "loss": 1.085, + "step": 2424 + }, + { + "epoch": 1.3165038002171552, + "grad_norm": 5.318299059101813, + "learning_rate": 1.72798564090218e-05, + "loss": 0.4945, + "step": 2425 + }, + { + "epoch": 1.3170466883821932, + "grad_norm": 6.095578475774351, + "learning_rate": 1.727744484058407e-05, + "loss": 0.7696, + "step": 2426 + }, + { + "epoch": 1.3175895765472312, + "grad_norm": 6.4886080927006695, + "learning_rate": 1.7275032372060368e-05, + "loss": 0.5364, + "step": 2427 + }, + { + "epoch": 1.3181324647122692, + "grad_norm": 8.681669485105683, + "learning_rate": 1.7272619003749066e-05, + "loss": 0.9132, + "step": 2428 + }, + { + "epoch": 1.3186753528773072, + "grad_norm": 6.819144122945338, + "learning_rate": 1.7270204735948653e-05, + "loss": 0.7867, + "step": 2429 + }, + { + "epoch": 1.3192182410423452, + "grad_norm": 6.65338033222052, + "learning_rate": 1.7267789568957734e-05, + "loss": 0.7236, + "step": 2430 + }, + { + "epoch": 1.3197611292073832, + "grad_norm": 6.622862530496735, + "learning_rate": 1.7265373503075014e-05, + "loss": 0.9038, + "step": 2431 + }, + { + "epoch": 1.3203040173724212, + "grad_norm": 6.284216416307323, + "learning_rate": 1.7262956538599323e-05, + "loss": 0.677, + "step": 2432 + }, + { + "epoch": 1.3208469055374592, + "grad_norm": 9.150803536423062, + "learning_rate": 1.7260538675829593e-05, + "loss": 1.7953, + "step": 2433 + }, + { + "epoch": 1.3213897937024972, + "grad_norm": 9.236250071710021, + "learning_rate": 1.7258119915064867e-05, + "loss": 0.8568, + "step": 2434 + }, + { + "epoch": 1.3219326818675352, + "grad_norm": 7.061272817814159, + "learning_rate": 1.72557002566043e-05, + "loss": 0.9355, + "step": 2435 + }, + { + "epoch": 1.3224755700325732, + "grad_norm": 6.786187626307571, + "learning_rate": 1.7253279700747164e-05, + "loss": 0.7707, + "step": 2436 + }, + { + "epoch": 1.3230184581976112, + "grad_norm": 5.592273858747665, + "learning_rate": 1.725085824779283e-05, + "loss": 0.5932, + "step": 2437 + }, + { + "epoch": 1.3235613463626492, + "grad_norm": 8.137496643840409, + "learning_rate": 1.72484358980408e-05, + "loss": 1.0244, + "step": 2438 + }, + { + "epoch": 1.3241042345276872, + "grad_norm": 5.981122138792478, + "learning_rate": 1.724601265179066e-05, + "loss": 0.9807, + "step": 2439 + }, + { + "epoch": 1.3246471226927252, + "grad_norm": 6.912027532729551, + "learning_rate": 1.7243588509342127e-05, + "loss": 0.8673, + "step": 2440 + }, + { + "epoch": 1.3251900108577632, + "grad_norm": 5.836398731450734, + "learning_rate": 1.7241163470995024e-05, + "loss": 0.5738, + "step": 2441 + }, + { + "epoch": 1.3257328990228012, + "grad_norm": 5.0689326588185795, + "learning_rate": 1.7238737537049283e-05, + "loss": 0.4449, + "step": 2442 + }, + { + "epoch": 1.3262757871878392, + "grad_norm": 6.628337387569613, + "learning_rate": 1.7236310707804943e-05, + "loss": 0.7975, + "step": 2443 + }, + { + "epoch": 1.3268186753528772, + "grad_norm": 9.359654484357105, + "learning_rate": 1.7233882983562168e-05, + "loss": 0.8996, + "step": 2444 + }, + { + "epoch": 1.3273615635179152, + "grad_norm": 5.979606493030731, + "learning_rate": 1.723145436462121e-05, + "loss": 0.732, + "step": 2445 + }, + { + "epoch": 1.3279044516829532, + "grad_norm": 6.393853922089353, + "learning_rate": 1.7229024851282453e-05, + "loss": 0.9713, + "step": 2446 + }, + { + "epoch": 1.3284473398479912, + "grad_norm": 9.783386290725945, + "learning_rate": 1.722659444384638e-05, + "loss": 1.2991, + "step": 2447 + }, + { + "epoch": 1.3289902280130292, + "grad_norm": 8.471557152983001, + "learning_rate": 1.722416314261359e-05, + "loss": 0.8517, + "step": 2448 + }, + { + "epoch": 1.3295331161780672, + "grad_norm": 7.20996107804807, + "learning_rate": 1.7221730947884793e-05, + "loss": 1.0892, + "step": 2449 + }, + { + "epoch": 1.3300760043431052, + "grad_norm": 9.97079093211024, + "learning_rate": 1.7219297859960796e-05, + "loss": 0.8866, + "step": 2450 + }, + { + "epoch": 1.3306188925081432, + "grad_norm": 5.88899239355285, + "learning_rate": 1.7216863879142536e-05, + "loss": 0.7286, + "step": 2451 + }, + { + "epoch": 1.3311617806731812, + "grad_norm": 8.317968838525326, + "learning_rate": 1.7214429005731054e-05, + "loss": 0.9374, + "step": 2452 + }, + { + "epoch": 1.3317046688382193, + "grad_norm": 8.278980033154179, + "learning_rate": 1.721199324002749e-05, + "loss": 1.2977, + "step": 2453 + }, + { + "epoch": 1.3322475570032573, + "grad_norm": 7.50073001311901, + "learning_rate": 1.7209556582333106e-05, + "loss": 0.9479, + "step": 2454 + }, + { + "epoch": 1.3327904451682953, + "grad_norm": 9.34140520001984, + "learning_rate": 1.720711903294928e-05, + "loss": 1.1834, + "step": 2455 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 6.155163014555799, + "learning_rate": 1.720468059217748e-05, + "loss": 1.0674, + "step": 2456 + }, + { + "epoch": 1.3338762214983713, + "grad_norm": 8.207152122618293, + "learning_rate": 1.7202241260319305e-05, + "loss": 1.1623, + "step": 2457 + }, + { + "epoch": 1.3344191096634093, + "grad_norm": 8.477127177193351, + "learning_rate": 1.719980103767645e-05, + "loss": 1.0812, + "step": 2458 + }, + { + "epoch": 1.3349619978284473, + "grad_norm": 9.352009366934986, + "learning_rate": 1.7197359924550726e-05, + "loss": 0.9593, + "step": 2459 + }, + { + "epoch": 1.3355048859934853, + "grad_norm": 6.377220189803284, + "learning_rate": 1.719491792124406e-05, + "loss": 0.8569, + "step": 2460 + }, + { + "epoch": 1.3360477741585233, + "grad_norm": 10.372722210417326, + "learning_rate": 1.7192475028058475e-05, + "loss": 0.8778, + "step": 2461 + }, + { + "epoch": 1.3365906623235613, + "grad_norm": 6.943692981514271, + "learning_rate": 1.7190031245296118e-05, + "loss": 0.8143, + "step": 2462 + }, + { + "epoch": 1.3371335504885993, + "grad_norm": 8.53588364120296, + "learning_rate": 1.7187586573259237e-05, + "loss": 1.2608, + "step": 2463 + }, + { + "epoch": 1.3376764386536373, + "grad_norm": 6.85497949710551, + "learning_rate": 1.7185141012250195e-05, + "loss": 0.621, + "step": 2464 + }, + { + "epoch": 1.3382193268186753, + "grad_norm": 9.575890345813113, + "learning_rate": 1.7182694562571458e-05, + "loss": 1.5686, + "step": 2465 + }, + { + "epoch": 1.3387622149837133, + "grad_norm": 8.637844905126164, + "learning_rate": 1.718024722452561e-05, + "loss": 0.8287, + "step": 2466 + }, + { + "epoch": 1.3393051031487513, + "grad_norm": 7.464168837468054, + "learning_rate": 1.7177798998415344e-05, + "loss": 0.9837, + "step": 2467 + }, + { + "epoch": 1.3398479913137893, + "grad_norm": 9.27927006463804, + "learning_rate": 1.7175349884543458e-05, + "loss": 1.3479, + "step": 2468 + }, + { + "epoch": 1.3403908794788273, + "grad_norm": 6.9150180643503765, + "learning_rate": 1.717289988321286e-05, + "loss": 0.7731, + "step": 2469 + }, + { + "epoch": 1.3409337676438653, + "grad_norm": 8.264286643294954, + "learning_rate": 1.7170448994726574e-05, + "loss": 1.1252, + "step": 2470 + }, + { + "epoch": 1.3414766558089033, + "grad_norm": 6.1769106952218715, + "learning_rate": 1.7167997219387728e-05, + "loss": 0.6077, + "step": 2471 + }, + { + "epoch": 1.3420195439739413, + "grad_norm": 6.00046554758586, + "learning_rate": 1.7165544557499563e-05, + "loss": 0.9807, + "step": 2472 + }, + { + "epoch": 1.3425624321389793, + "grad_norm": 5.684026968717126, + "learning_rate": 1.7163091009365427e-05, + "loss": 0.8755, + "step": 2473 + }, + { + "epoch": 1.3431053203040173, + "grad_norm": 8.910694398982464, + "learning_rate": 1.7160636575288777e-05, + "loss": 1.1781, + "step": 2474 + }, + { + "epoch": 1.3436482084690553, + "grad_norm": 6.090059050660724, + "learning_rate": 1.7158181255573184e-05, + "loss": 0.8757, + "step": 2475 + }, + { + "epoch": 1.3441910966340933, + "grad_norm": 5.3693699493967575, + "learning_rate": 1.7155725050522325e-05, + "loss": 0.4111, + "step": 2476 + }, + { + "epoch": 1.3447339847991313, + "grad_norm": 6.4883376045754675, + "learning_rate": 1.7153267960439987e-05, + "loss": 1.1667, + "step": 2477 + }, + { + "epoch": 1.3452768729641693, + "grad_norm": 6.9649527868325585, + "learning_rate": 1.7150809985630065e-05, + "loss": 1.2642, + "step": 2478 + }, + { + "epoch": 1.3458197611292073, + "grad_norm": 7.458967124353787, + "learning_rate": 1.714835112639657e-05, + "loss": 1.1198, + "step": 2479 + }, + { + "epoch": 1.3463626492942453, + "grad_norm": 7.4820540779231015, + "learning_rate": 1.7145891383043613e-05, + "loss": 1.069, + "step": 2480 + }, + { + "epoch": 1.3469055374592833, + "grad_norm": 6.8442083181868165, + "learning_rate": 1.7143430755875422e-05, + "loss": 1.1187, + "step": 2481 + }, + { + "epoch": 1.3474484256243213, + "grad_norm": 8.389465670773259, + "learning_rate": 1.7140969245196332e-05, + "loss": 1.216, + "step": 2482 + }, + { + "epoch": 1.3479913137893593, + "grad_norm": 8.099481746156716, + "learning_rate": 1.713850685131078e-05, + "loss": 1.1365, + "step": 2483 + }, + { + "epoch": 1.3485342019543973, + "grad_norm": 7.005684437567773, + "learning_rate": 1.713604357452333e-05, + "loss": 0.7665, + "step": 2484 + }, + { + "epoch": 1.3490770901194353, + "grad_norm": 6.032904961736961, + "learning_rate": 1.7133579415138634e-05, + "loss": 0.7527, + "step": 2485 + }, + { + "epoch": 1.3496199782844733, + "grad_norm": 7.644359802639278, + "learning_rate": 1.7131114373461466e-05, + "loss": 0.7293, + "step": 2486 + }, + { + "epoch": 1.3501628664495113, + "grad_norm": 9.681466854129468, + "learning_rate": 1.7128648449796706e-05, + "loss": 1.0718, + "step": 2487 + }, + { + "epoch": 1.3507057546145493, + "grad_norm": 7.703820713192501, + "learning_rate": 1.7126181644449348e-05, + "loss": 1.0049, + "step": 2488 + }, + { + "epoch": 1.3512486427795873, + "grad_norm": 8.313943023532504, + "learning_rate": 1.7123713957724482e-05, + "loss": 1.1647, + "step": 2489 + }, + { + "epoch": 1.3517915309446253, + "grad_norm": 6.869318843676937, + "learning_rate": 1.7121245389927326e-05, + "loss": 0.7056, + "step": 2490 + }, + { + "epoch": 1.3523344191096633, + "grad_norm": 8.354366894281007, + "learning_rate": 1.7118775941363186e-05, + "loss": 1.4107, + "step": 2491 + }, + { + "epoch": 1.3528773072747013, + "grad_norm": 7.594996874862703, + "learning_rate": 1.7116305612337493e-05, + "loss": 1.3947, + "step": 2492 + }, + { + "epoch": 1.3534201954397393, + "grad_norm": 5.912640331169303, + "learning_rate": 1.7113834403155782e-05, + "loss": 0.678, + "step": 2493 + }, + { + "epoch": 1.3539630836047774, + "grad_norm": 8.603089873660597, + "learning_rate": 1.7111362314123693e-05, + "loss": 1.1846, + "step": 2494 + }, + { + "epoch": 1.3545059717698154, + "grad_norm": 5.899587064063632, + "learning_rate": 1.710888934554698e-05, + "loss": 0.7241, + "step": 2495 + }, + { + "epoch": 1.3550488599348534, + "grad_norm": 5.1366064982711315, + "learning_rate": 1.7106415497731502e-05, + "loss": 0.6714, + "step": 2496 + }, + { + "epoch": 1.3555917480998914, + "grad_norm": 6.55484035818926, + "learning_rate": 1.710394077098323e-05, + "loss": 1.0163, + "step": 2497 + }, + { + "epoch": 1.3561346362649294, + "grad_norm": 7.060748735507699, + "learning_rate": 1.710146516560824e-05, + "loss": 0.8407, + "step": 2498 + }, + { + "epoch": 1.3566775244299674, + "grad_norm": 7.488417378829399, + "learning_rate": 1.709898868191272e-05, + "loss": 1.0936, + "step": 2499 + }, + { + "epoch": 1.3572204125950054, + "grad_norm": 5.616560879457259, + "learning_rate": 1.7096511320202965e-05, + "loss": 0.6962, + "step": 2500 + }, + { + "epoch": 1.3577633007600434, + "grad_norm": 6.5457002772857775, + "learning_rate": 1.7094033080785384e-05, + "loss": 0.6338, + "step": 2501 + }, + { + "epoch": 1.3583061889250814, + "grad_norm": 5.788382651374094, + "learning_rate": 1.709155396396648e-05, + "loss": 0.9251, + "step": 2502 + }, + { + "epoch": 1.3588490770901194, + "grad_norm": 5.94278314769549, + "learning_rate": 1.7089073970052883e-05, + "loss": 0.5738, + "step": 2503 + }, + { + "epoch": 1.3593919652551574, + "grad_norm": 8.987615230603502, + "learning_rate": 1.7086593099351318e-05, + "loss": 1.1561, + "step": 2504 + }, + { + "epoch": 1.3599348534201954, + "grad_norm": 6.25509019167934, + "learning_rate": 1.7084111352168627e-05, + "loss": 0.5584, + "step": 2505 + }, + { + "epoch": 1.3604777415852334, + "grad_norm": 6.42626784615777, + "learning_rate": 1.708162872881175e-05, + "loss": 0.6385, + "step": 2506 + }, + { + "epoch": 1.3610206297502714, + "grad_norm": 7.342680004921898, + "learning_rate": 1.707914522958775e-05, + "loss": 1.0729, + "step": 2507 + }, + { + "epoch": 1.3615635179153094, + "grad_norm": 6.858864236878158, + "learning_rate": 1.707666085480378e-05, + "loss": 0.8641, + "step": 2508 + }, + { + "epoch": 1.3621064060803474, + "grad_norm": 6.176457712037886, + "learning_rate": 1.707417560476712e-05, + "loss": 0.9472, + "step": 2509 + }, + { + "epoch": 1.3626492942453854, + "grad_norm": 7.346954574146698, + "learning_rate": 1.7071689479785145e-05, + "loss": 1.0253, + "step": 2510 + }, + { + "epoch": 1.3631921824104234, + "grad_norm": 4.9331066104603085, + "learning_rate": 1.7069202480165344e-05, + "loss": 0.7307, + "step": 2511 + }, + { + "epoch": 1.3637350705754614, + "grad_norm": 8.316711829847868, + "learning_rate": 1.7066714606215316e-05, + "loss": 0.8078, + "step": 2512 + }, + { + "epoch": 1.3642779587404994, + "grad_norm": 8.079444940270628, + "learning_rate": 1.706422585824276e-05, + "loss": 1.2922, + "step": 2513 + }, + { + "epoch": 1.3648208469055374, + "grad_norm": 7.503741374806181, + "learning_rate": 1.7061736236555494e-05, + "loss": 0.7943, + "step": 2514 + }, + { + "epoch": 1.3653637350705754, + "grad_norm": 6.4023293864360635, + "learning_rate": 1.7059245741461435e-05, + "loss": 1.0111, + "step": 2515 + }, + { + "epoch": 1.3659066232356134, + "grad_norm": 5.7325920409358595, + "learning_rate": 1.705675437326861e-05, + "loss": 1.1201, + "step": 2516 + }, + { + "epoch": 1.3664495114006514, + "grad_norm": 8.452660578831315, + "learning_rate": 1.705426213228516e-05, + "loss": 0.8818, + "step": 2517 + }, + { + "epoch": 1.3669923995656894, + "grad_norm": 7.8301031574049755, + "learning_rate": 1.7051769018819327e-05, + "loss": 0.8803, + "step": 2518 + }, + { + "epoch": 1.3675352877307274, + "grad_norm": 10.77827779179992, + "learning_rate": 1.704927503317946e-05, + "loss": 1.6667, + "step": 2519 + }, + { + "epoch": 1.3680781758957654, + "grad_norm": 4.579120603442138, + "learning_rate": 1.7046780175674023e-05, + "loss": 0.6101, + "step": 2520 + }, + { + "epoch": 1.3686210640608034, + "grad_norm": 7.443733277623912, + "learning_rate": 1.7044284446611584e-05, + "loss": 0.8174, + "step": 2521 + }, + { + "epoch": 1.3691639522258414, + "grad_norm": 10.860855146774309, + "learning_rate": 1.7041787846300817e-05, + "loss": 1.2849, + "step": 2522 + }, + { + "epoch": 1.3697068403908794, + "grad_norm": 9.311283837684355, + "learning_rate": 1.7039290375050503e-05, + "loss": 1.4923, + "step": 2523 + }, + { + "epoch": 1.3702497285559174, + "grad_norm": 7.171226901795553, + "learning_rate": 1.703679203316954e-05, + "loss": 0.9493, + "step": 2524 + }, + { + "epoch": 1.3707926167209554, + "grad_norm": 8.658053683953712, + "learning_rate": 1.7034292820966923e-05, + "loss": 1.1366, + "step": 2525 + }, + { + "epoch": 1.3713355048859934, + "grad_norm": 6.400723985129287, + "learning_rate": 1.703179273875176e-05, + "loss": 1.0484, + "step": 2526 + }, + { + "epoch": 1.3718783930510314, + "grad_norm": 6.2215780206112274, + "learning_rate": 1.702929178683326e-05, + "loss": 1.1058, + "step": 2527 + }, + { + "epoch": 1.3724212812160694, + "grad_norm": 7.548758658264992, + "learning_rate": 1.7026789965520755e-05, + "loss": 1.094, + "step": 2528 + }, + { + "epoch": 1.3729641693811074, + "grad_norm": 7.250992484079327, + "learning_rate": 1.7024287275123664e-05, + "loss": 0.8424, + "step": 2529 + }, + { + "epoch": 1.3735070575461454, + "grad_norm": 6.664109303085317, + "learning_rate": 1.702178371595153e-05, + "loss": 1.0498, + "step": 2530 + }, + { + "epoch": 1.3740499457111834, + "grad_norm": 6.379180216401721, + "learning_rate": 1.7019279288313994e-05, + "loss": 0.9038, + "step": 2531 + }, + { + "epoch": 1.3745928338762214, + "grad_norm": 7.524913368583668, + "learning_rate": 1.7016773992520806e-05, + "loss": 0.8787, + "step": 2532 + }, + { + "epoch": 1.3751357220412594, + "grad_norm": 5.722778911929788, + "learning_rate": 1.701426782888183e-05, + "loss": 0.8199, + "step": 2533 + }, + { + "epoch": 1.3756786102062974, + "grad_norm": 6.26381521799836, + "learning_rate": 1.701176079770703e-05, + "loss": 0.6989, + "step": 2534 + }, + { + "epoch": 1.3762214983713354, + "grad_norm": 7.2300935887481765, + "learning_rate": 1.700925289930648e-05, + "loss": 1.094, + "step": 2535 + }, + { + "epoch": 1.3767643865363735, + "grad_norm": 7.385614391862118, + "learning_rate": 1.7006744133990358e-05, + "loss": 0.7753, + "step": 2536 + }, + { + "epoch": 1.3773072747014115, + "grad_norm": 5.52011787904346, + "learning_rate": 1.7004234502068952e-05, + "loss": 0.9232, + "step": 2537 + }, + { + "epoch": 1.3778501628664495, + "grad_norm": 6.523028453950784, + "learning_rate": 1.7001724003852668e-05, + "loss": 1.1181, + "step": 2538 + }, + { + "epoch": 1.3783930510314875, + "grad_norm": 6.334090281128193, + "learning_rate": 1.6999212639651995e-05, + "loss": 0.9384, + "step": 2539 + }, + { + "epoch": 1.3789359391965255, + "grad_norm": 7.3687722051890265, + "learning_rate": 1.6996700409777548e-05, + "loss": 0.9678, + "step": 2540 + }, + { + "epoch": 1.3794788273615635, + "grad_norm": 9.49113126498042, + "learning_rate": 1.6994187314540042e-05, + "loss": 0.7875, + "step": 2541 + }, + { + "epoch": 1.3800217155266015, + "grad_norm": 4.885613246464195, + "learning_rate": 1.69916733542503e-05, + "loss": 0.4488, + "step": 2542 + }, + { + "epoch": 1.3805646036916395, + "grad_norm": 7.362521236541499, + "learning_rate": 1.6989158529219262e-05, + "loss": 0.9808, + "step": 2543 + }, + { + "epoch": 1.3811074918566775, + "grad_norm": 8.6488909724118, + "learning_rate": 1.6986642839757953e-05, + "loss": 1.0599, + "step": 2544 + }, + { + "epoch": 1.3816503800217155, + "grad_norm": 6.054541159551389, + "learning_rate": 1.698412628617752e-05, + "loss": 0.8252, + "step": 2545 + }, + { + "epoch": 1.3821932681867535, + "grad_norm": 7.77360294326381, + "learning_rate": 1.698160886878922e-05, + "loss": 1.0718, + "step": 2546 + }, + { + "epoch": 1.3827361563517915, + "grad_norm": 9.493714306510151, + "learning_rate": 1.697909058790441e-05, + "loss": 1.2784, + "step": 2547 + }, + { + "epoch": 1.3832790445168295, + "grad_norm": 7.7307988823000455, + "learning_rate": 1.6976571443834555e-05, + "loss": 1.0308, + "step": 2548 + }, + { + "epoch": 1.3838219326818675, + "grad_norm": 7.149534093427325, + "learning_rate": 1.697405143689122e-05, + "loss": 0.9702, + "step": 2549 + }, + { + "epoch": 1.3843648208469055, + "grad_norm": 6.16619254042592, + "learning_rate": 1.6971530567386087e-05, + "loss": 0.6922, + "step": 2550 + }, + { + "epoch": 1.3849077090119435, + "grad_norm": 6.555255328428672, + "learning_rate": 1.6969008835630947e-05, + "loss": 0.8771, + "step": 2551 + }, + { + "epoch": 1.3854505971769815, + "grad_norm": 7.652524555180243, + "learning_rate": 1.6966486241937685e-05, + "loss": 1.0635, + "step": 2552 + }, + { + "epoch": 1.3859934853420195, + "grad_norm": 9.675328775672968, + "learning_rate": 1.69639627866183e-05, + "loss": 1.1616, + "step": 2553 + }, + { + "epoch": 1.3865363735070575, + "grad_norm": 6.5667057730439256, + "learning_rate": 1.69614384699849e-05, + "loss": 0.7181, + "step": 2554 + }, + { + "epoch": 1.3870792616720955, + "grad_norm": 7.374476144001139, + "learning_rate": 1.6958913292349698e-05, + "loss": 1.123, + "step": 2555 + }, + { + "epoch": 1.3876221498371335, + "grad_norm": 8.873033944832025, + "learning_rate": 1.6956387254025005e-05, + "loss": 1.2651, + "step": 2556 + }, + { + "epoch": 1.3881650380021715, + "grad_norm": 6.877978468812787, + "learning_rate": 1.695386035532325e-05, + "loss": 1.0039, + "step": 2557 + }, + { + "epoch": 1.3887079261672095, + "grad_norm": 4.746919583421645, + "learning_rate": 1.6951332596556966e-05, + "loss": 0.7222, + "step": 2558 + }, + { + "epoch": 1.3892508143322475, + "grad_norm": 5.538693020619686, + "learning_rate": 1.6948803978038787e-05, + "loss": 0.8463, + "step": 2559 + }, + { + "epoch": 1.3897937024972855, + "grad_norm": 6.636448326582328, + "learning_rate": 1.6946274500081455e-05, + "loss": 0.7594, + "step": 2560 + }, + { + "epoch": 1.3903365906623235, + "grad_norm": 8.74658695496399, + "learning_rate": 1.6943744162997825e-05, + "loss": 1.224, + "step": 2561 + }, + { + "epoch": 1.3908794788273615, + "grad_norm": 7.305739264312674, + "learning_rate": 1.694121296710085e-05, + "loss": 1.0755, + "step": 2562 + }, + { + "epoch": 1.3914223669923995, + "grad_norm": 4.6475508332537165, + "learning_rate": 1.693868091270359e-05, + "loss": 0.6127, + "step": 2563 + }, + { + "epoch": 1.3919652551574375, + "grad_norm": 8.484116710765257, + "learning_rate": 1.6936148000119218e-05, + "loss": 1.029, + "step": 2564 + }, + { + "epoch": 1.3925081433224755, + "grad_norm": 6.532276109161297, + "learning_rate": 1.6933614229661008e-05, + "loss": 1.0517, + "step": 2565 + }, + { + "epoch": 1.3930510314875135, + "grad_norm": 5.43723324019315, + "learning_rate": 1.6931079601642336e-05, + "loss": 0.6915, + "step": 2566 + }, + { + "epoch": 1.3935939196525515, + "grad_norm": 7.562306350008547, + "learning_rate": 1.69285441163767e-05, + "loss": 0.5726, + "step": 2567 + }, + { + "epoch": 1.3941368078175895, + "grad_norm": 8.500602954507755, + "learning_rate": 1.6926007774177678e-05, + "loss": 1.0297, + "step": 2568 + }, + { + "epoch": 1.3946796959826275, + "grad_norm": 6.432151204429844, + "learning_rate": 1.6923470575358977e-05, + "loss": 1.0016, + "step": 2569 + }, + { + "epoch": 1.3952225841476655, + "grad_norm": 9.552257208599151, + "learning_rate": 1.6920932520234402e-05, + "loss": 1.0814, + "step": 2570 + }, + { + "epoch": 1.3957654723127035, + "grad_norm": 8.075918082294125, + "learning_rate": 1.6918393609117863e-05, + "loss": 1.1699, + "step": 2571 + }, + { + "epoch": 1.3963083604777415, + "grad_norm": 6.296577538864887, + "learning_rate": 1.6915853842323373e-05, + "loss": 1.0211, + "step": 2572 + }, + { + "epoch": 1.3968512486427795, + "grad_norm": 7.785873884786678, + "learning_rate": 1.691331322016506e-05, + "loss": 0.9225, + "step": 2573 + }, + { + "epoch": 1.3973941368078175, + "grad_norm": 6.768130802749829, + "learning_rate": 1.691077174295715e-05, + "loss": 0.8092, + "step": 2574 + }, + { + "epoch": 1.3979370249728555, + "grad_norm": 5.566346022945414, + "learning_rate": 1.6908229411013977e-05, + "loss": 0.8312, + "step": 2575 + }, + { + "epoch": 1.3984799131378935, + "grad_norm": 7.2860384645139815, + "learning_rate": 1.6905686224649978e-05, + "loss": 0.9357, + "step": 2576 + }, + { + "epoch": 1.3990228013029316, + "grad_norm": 7.072468614016554, + "learning_rate": 1.6903142184179704e-05, + "loss": 0.7362, + "step": 2577 + }, + { + "epoch": 1.3995656894679696, + "grad_norm": 6.421225685489366, + "learning_rate": 1.6900597289917803e-05, + "loss": 0.6828, + "step": 2578 + }, + { + "epoch": 1.4001085776330076, + "grad_norm": 8.167489592155093, + "learning_rate": 1.689805154217903e-05, + "loss": 1.1237, + "step": 2579 + }, + { + "epoch": 1.4006514657980456, + "grad_norm": 11.676833397587332, + "learning_rate": 1.6895504941278246e-05, + "loss": 0.9532, + "step": 2580 + }, + { + "epoch": 1.4011943539630836, + "grad_norm": 6.04293054204543, + "learning_rate": 1.689295748753042e-05, + "loss": 0.6541, + "step": 2581 + }, + { + "epoch": 1.4017372421281216, + "grad_norm": 8.432023560966908, + "learning_rate": 1.6890409181250632e-05, + "loss": 1.0167, + "step": 2582 + }, + { + "epoch": 1.4022801302931596, + "grad_norm": 7.494983940960417, + "learning_rate": 1.688786002275405e-05, + "loss": 0.8913, + "step": 2583 + }, + { + "epoch": 1.4028230184581976, + "grad_norm": 6.475912046078606, + "learning_rate": 1.6885310012355964e-05, + "loss": 0.8599, + "step": 2584 + }, + { + "epoch": 1.4033659066232356, + "grad_norm": 5.145369710998311, + "learning_rate": 1.6882759150371765e-05, + "loss": 0.8079, + "step": 2585 + }, + { + "epoch": 1.4039087947882736, + "grad_norm": 6.1855027756804155, + "learning_rate": 1.688020743711694e-05, + "loss": 0.8046, + "step": 2586 + }, + { + "epoch": 1.4044516829533116, + "grad_norm": 8.409570707749053, + "learning_rate": 1.68776548729071e-05, + "loss": 0.75, + "step": 2587 + }, + { + "epoch": 1.4049945711183496, + "grad_norm": 6.150666283054011, + "learning_rate": 1.687510145805794e-05, + "loss": 0.8339, + "step": 2588 + }, + { + "epoch": 1.4055374592833876, + "grad_norm": 5.4082205122977625, + "learning_rate": 1.6872547192885272e-05, + "loss": 0.7878, + "step": 2589 + }, + { + "epoch": 1.4060803474484256, + "grad_norm": 7.011727228130495, + "learning_rate": 1.686999207770502e-05, + "loss": 0.6415, + "step": 2590 + }, + { + "epoch": 1.4066232356134636, + "grad_norm": 7.384524419917869, + "learning_rate": 1.6867436112833193e-05, + "loss": 0.8489, + "step": 2591 + }, + { + "epoch": 1.4071661237785016, + "grad_norm": 6.846659366372164, + "learning_rate": 1.6864879298585925e-05, + "loss": 1.128, + "step": 2592 + }, + { + "epoch": 1.4077090119435396, + "grad_norm": 6.142948047083128, + "learning_rate": 1.6862321635279444e-05, + "loss": 0.7886, + "step": 2593 + }, + { + "epoch": 1.4082519001085776, + "grad_norm": 10.23557881641297, + "learning_rate": 1.6859763123230086e-05, + "loss": 0.8323, + "step": 2594 + }, + { + "epoch": 1.4087947882736156, + "grad_norm": 6.995135696669204, + "learning_rate": 1.6857203762754294e-05, + "loss": 1.0283, + "step": 2595 + }, + { + "epoch": 1.4093376764386536, + "grad_norm": 7.709014920487576, + "learning_rate": 1.685464355416861e-05, + "loss": 0.8586, + "step": 2596 + }, + { + "epoch": 1.4098805646036916, + "grad_norm": 8.153794053960306, + "learning_rate": 1.6852082497789684e-05, + "loss": 1.1797, + "step": 2597 + }, + { + "epoch": 1.4104234527687296, + "grad_norm": 7.721286185320884, + "learning_rate": 1.684952059393428e-05, + "loss": 0.9196, + "step": 2598 + }, + { + "epoch": 1.4109663409337676, + "grad_norm": 8.024349951964256, + "learning_rate": 1.684695784291925e-05, + "loss": 0.7769, + "step": 2599 + }, + { + "epoch": 1.4115092290988056, + "grad_norm": 8.135395328882872, + "learning_rate": 1.684439424506156e-05, + "loss": 1.1642, + "step": 2600 + }, + { + "epoch": 1.4120521172638436, + "grad_norm": 6.057801451822921, + "learning_rate": 1.684182980067828e-05, + "loss": 0.7352, + "step": 2601 + }, + { + "epoch": 1.4125950054288816, + "grad_norm": 7.808755828672067, + "learning_rate": 1.683926451008659e-05, + "loss": 0.9788, + "step": 2602 + }, + { + "epoch": 1.4131378935939196, + "grad_norm": 7.426450515083944, + "learning_rate": 1.6836698373603765e-05, + "loss": 0.7696, + "step": 2603 + }, + { + "epoch": 1.4136807817589576, + "grad_norm": 9.382792907740685, + "learning_rate": 1.6834131391547187e-05, + "loss": 1.0872, + "step": 2604 + }, + { + "epoch": 1.4142236699239956, + "grad_norm": 6.171700145892427, + "learning_rate": 1.6831563564234347e-05, + "loss": 0.7233, + "step": 2605 + }, + { + "epoch": 1.4147665580890336, + "grad_norm": 11.270930151197227, + "learning_rate": 1.682899489198284e-05, + "loss": 0.9739, + "step": 2606 + }, + { + "epoch": 1.4153094462540716, + "grad_norm": 9.050544540368543, + "learning_rate": 1.6826425375110357e-05, + "loss": 0.9436, + "step": 2607 + }, + { + "epoch": 1.4158523344191096, + "grad_norm": 6.969930349987723, + "learning_rate": 1.6823855013934705e-05, + "loss": 0.7632, + "step": 2608 + }, + { + "epoch": 1.4163952225841476, + "grad_norm": 5.791480821515424, + "learning_rate": 1.682128380877379e-05, + "loss": 0.8202, + "step": 2609 + }, + { + "epoch": 1.4169381107491856, + "grad_norm": 9.905757175497309, + "learning_rate": 1.6818711759945623e-05, + "loss": 1.2161, + "step": 2610 + }, + { + "epoch": 1.4174809989142236, + "grad_norm": 6.082684402096193, + "learning_rate": 1.6816138867768318e-05, + "loss": 0.7245, + "step": 2611 + }, + { + "epoch": 1.4180238870792616, + "grad_norm": 10.15499139340414, + "learning_rate": 1.6813565132560092e-05, + "loss": 1.1514, + "step": 2612 + }, + { + "epoch": 1.4185667752442996, + "grad_norm": 6.978513243855861, + "learning_rate": 1.6810990554639276e-05, + "loss": 0.5331, + "step": 2613 + }, + { + "epoch": 1.4191096634093376, + "grad_norm": 8.037739371135425, + "learning_rate": 1.6808415134324288e-05, + "loss": 0.7089, + "step": 2614 + }, + { + "epoch": 1.4196525515743756, + "grad_norm": 6.925791024229076, + "learning_rate": 1.6805838871933664e-05, + "loss": 0.8981, + "step": 2615 + }, + { + "epoch": 1.4201954397394136, + "grad_norm": 6.352165521713976, + "learning_rate": 1.6803261767786048e-05, + "loss": 0.5644, + "step": 2616 + }, + { + "epoch": 1.4207383279044516, + "grad_norm": 10.267497627922179, + "learning_rate": 1.680068382220017e-05, + "loss": 0.9607, + "step": 2617 + }, + { + "epoch": 1.4212812160694897, + "grad_norm": 7.696468630335302, + "learning_rate": 1.679810503549488e-05, + "loss": 0.729, + "step": 2618 + }, + { + "epoch": 1.4218241042345277, + "grad_norm": 6.868981955235214, + "learning_rate": 1.679552540798912e-05, + "loss": 0.9152, + "step": 2619 + }, + { + "epoch": 1.4223669923995657, + "grad_norm": 8.909790381587262, + "learning_rate": 1.6792944940001952e-05, + "loss": 0.9335, + "step": 2620 + }, + { + "epoch": 1.4229098805646037, + "grad_norm": 9.62415635421133, + "learning_rate": 1.6790363631852524e-05, + "loss": 1.1012, + "step": 2621 + }, + { + "epoch": 1.4234527687296417, + "grad_norm": 7.307385111665146, + "learning_rate": 1.67877814838601e-05, + "loss": 0.6963, + "step": 2622 + }, + { + "epoch": 1.4239956568946797, + "grad_norm": 9.018642559605246, + "learning_rate": 1.678519849634405e-05, + "loss": 1.0612, + "step": 2623 + }, + { + "epoch": 1.4245385450597177, + "grad_norm": 6.929244675286548, + "learning_rate": 1.6782614669623827e-05, + "loss": 0.9462, + "step": 2624 + }, + { + "epoch": 1.4250814332247557, + "grad_norm": 8.577419731292203, + "learning_rate": 1.6780030004019016e-05, + "loss": 1.112, + "step": 2625 + }, + { + "epoch": 1.4256243213897937, + "grad_norm": 7.125694710248634, + "learning_rate": 1.677744449984929e-05, + "loss": 0.851, + "step": 2626 + }, + { + "epoch": 1.4261672095548317, + "grad_norm": 9.488635353984538, + "learning_rate": 1.6774858157434425e-05, + "loss": 1.3064, + "step": 2627 + }, + { + "epoch": 1.4267100977198697, + "grad_norm": 7.500971938576216, + "learning_rate": 1.6772270977094307e-05, + "loss": 1.3675, + "step": 2628 + }, + { + "epoch": 1.4272529858849077, + "grad_norm": 6.979855088071626, + "learning_rate": 1.676968295914892e-05, + "loss": 0.8896, + "step": 2629 + }, + { + "epoch": 1.4277958740499457, + "grad_norm": 7.795843207695432, + "learning_rate": 1.6767094103918357e-05, + "loss": 0.7758, + "step": 2630 + }, + { + "epoch": 1.4283387622149837, + "grad_norm": 6.131884468994572, + "learning_rate": 1.6764504411722806e-05, + "loss": 0.7744, + "step": 2631 + }, + { + "epoch": 1.4288816503800217, + "grad_norm": 10.33693974618067, + "learning_rate": 1.676191388288257e-05, + "loss": 1.5364, + "step": 2632 + }, + { + "epoch": 1.4294245385450597, + "grad_norm": 10.005426546410897, + "learning_rate": 1.6759322517718048e-05, + "loss": 0.9713, + "step": 2633 + }, + { + "epoch": 1.4299674267100977, + "grad_norm": 9.480236119835816, + "learning_rate": 1.6756730316549745e-05, + "loss": 1.2573, + "step": 2634 + }, + { + "epoch": 1.4305103148751357, + "grad_norm": 10.874101086577593, + "learning_rate": 1.675413727969827e-05, + "loss": 1.0308, + "step": 2635 + }, + { + "epoch": 1.4310532030401737, + "grad_norm": 5.34984499382639, + "learning_rate": 1.675154340748433e-05, + "loss": 0.6013, + "step": 2636 + }, + { + "epoch": 1.4315960912052117, + "grad_norm": 7.464386702227026, + "learning_rate": 1.674894870022874e-05, + "loss": 0.9918, + "step": 2637 + }, + { + "epoch": 1.4321389793702497, + "grad_norm": 8.280480153626241, + "learning_rate": 1.674635315825242e-05, + "loss": 0.8745, + "step": 2638 + }, + { + "epoch": 1.4326818675352877, + "grad_norm": 6.169575731939732, + "learning_rate": 1.6743756781876385e-05, + "loss": 0.6984, + "step": 2639 + }, + { + "epoch": 1.4332247557003257, + "grad_norm": 5.390920772379431, + "learning_rate": 1.6741159571421768e-05, + "loss": 0.8886, + "step": 2640 + }, + { + "epoch": 1.4337676438653637, + "grad_norm": 7.003336357340687, + "learning_rate": 1.6738561527209792e-05, + "loss": 1.0335, + "step": 2641 + }, + { + "epoch": 1.4343105320304017, + "grad_norm": 7.722166856340297, + "learning_rate": 1.6735962649561784e-05, + "loss": 0.9804, + "step": 2642 + }, + { + "epoch": 1.4348534201954397, + "grad_norm": 7.328242752900802, + "learning_rate": 1.673336293879918e-05, + "loss": 0.8675, + "step": 2643 + }, + { + "epoch": 1.4353963083604777, + "grad_norm": 7.344577484351048, + "learning_rate": 1.6730762395243515e-05, + "loss": 1.2518, + "step": 2644 + }, + { + "epoch": 1.4359391965255157, + "grad_norm": 7.560893408391802, + "learning_rate": 1.6728161019216433e-05, + "loss": 1.0196, + "step": 2645 + }, + { + "epoch": 1.4364820846905537, + "grad_norm": 7.509821569775923, + "learning_rate": 1.6725558811039674e-05, + "loss": 1.1112, + "step": 2646 + }, + { + "epoch": 1.4370249728555917, + "grad_norm": 7.672412677248201, + "learning_rate": 1.672295577103508e-05, + "loss": 0.9075, + "step": 2647 + }, + { + "epoch": 1.4375678610206297, + "grad_norm": 8.562389123203614, + "learning_rate": 1.67203518995246e-05, + "loss": 1.2123, + "step": 2648 + }, + { + "epoch": 1.4381107491856677, + "grad_norm": 8.172008773566091, + "learning_rate": 1.671774719683029e-05, + "loss": 0.8064, + "step": 2649 + }, + { + "epoch": 1.4386536373507057, + "grad_norm": 7.5145061670971405, + "learning_rate": 1.6715141663274297e-05, + "loss": 1.3513, + "step": 2650 + }, + { + "epoch": 1.4391965255157437, + "grad_norm": 5.675981880534199, + "learning_rate": 1.6712535299178883e-05, + "loss": 0.5029, + "step": 2651 + }, + { + "epoch": 1.4397394136807817, + "grad_norm": 8.195158801834715, + "learning_rate": 1.6709928104866403e-05, + "loss": 0.9414, + "step": 2652 + }, + { + "epoch": 1.4402823018458197, + "grad_norm": 7.374510239535514, + "learning_rate": 1.6707320080659322e-05, + "loss": 0.8103, + "step": 2653 + }, + { + "epoch": 1.4408251900108577, + "grad_norm": 5.703232221762084, + "learning_rate": 1.6704711226880204e-05, + "loss": 1.1646, + "step": 2654 + }, + { + "epoch": 1.4413680781758957, + "grad_norm": 6.981749095787225, + "learning_rate": 1.6702101543851714e-05, + "loss": 0.7174, + "step": 2655 + }, + { + "epoch": 1.4419109663409337, + "grad_norm": 6.456866833246808, + "learning_rate": 1.6699491031896625e-05, + "loss": 0.9795, + "step": 2656 + }, + { + "epoch": 1.4424538545059717, + "grad_norm": 10.193228844116039, + "learning_rate": 1.6696879691337807e-05, + "loss": 1.1288, + "step": 2657 + }, + { + "epoch": 1.4429967426710097, + "grad_norm": 9.920534268729181, + "learning_rate": 1.6694267522498237e-05, + "loss": 0.8379, + "step": 2658 + }, + { + "epoch": 1.4435396308360477, + "grad_norm": 6.153776792682404, + "learning_rate": 1.669165452570099e-05, + "loss": 0.8818, + "step": 2659 + }, + { + "epoch": 1.4440825190010858, + "grad_norm": 5.650871346064418, + "learning_rate": 1.6689040701269245e-05, + "loss": 0.732, + "step": 2660 + }, + { + "epoch": 1.4446254071661238, + "grad_norm": 7.829649689315363, + "learning_rate": 1.668642604952629e-05, + "loss": 0.8119, + "step": 2661 + }, + { + "epoch": 1.4451682953311618, + "grad_norm": 5.959350361834891, + "learning_rate": 1.6683810570795498e-05, + "loss": 0.6829, + "step": 2662 + }, + { + "epoch": 1.4457111834961998, + "grad_norm": 8.76687690585514, + "learning_rate": 1.6681194265400365e-05, + "loss": 1.4888, + "step": 2663 + }, + { + "epoch": 1.4462540716612378, + "grad_norm": 6.129763847493397, + "learning_rate": 1.6678577133664476e-05, + "loss": 0.8372, + "step": 2664 + }, + { + "epoch": 1.4467969598262758, + "grad_norm": 6.529235674005995, + "learning_rate": 1.6675959175911527e-05, + "loss": 0.9655, + "step": 2665 + }, + { + "epoch": 1.4473398479913138, + "grad_norm": 7.741547370588863, + "learning_rate": 1.6673340392465304e-05, + "loss": 1.0772, + "step": 2666 + }, + { + "epoch": 1.4478827361563518, + "grad_norm": 6.674042204787818, + "learning_rate": 1.6670720783649706e-05, + "loss": 0.9681, + "step": 2667 + }, + { + "epoch": 1.4484256243213898, + "grad_norm": 7.920578917847642, + "learning_rate": 1.666810034978873e-05, + "loss": 1.1068, + "step": 2668 + }, + { + "epoch": 1.4489685124864278, + "grad_norm": 7.249023283604718, + "learning_rate": 1.6665479091206476e-05, + "loss": 1.1025, + "step": 2669 + }, + { + "epoch": 1.4495114006514658, + "grad_norm": 6.407638163624444, + "learning_rate": 1.6662857008227145e-05, + "loss": 0.8236, + "step": 2670 + }, + { + "epoch": 1.4500542888165038, + "grad_norm": 7.289336395940614, + "learning_rate": 1.6660234101175036e-05, + "loss": 0.9386, + "step": 2671 + }, + { + "epoch": 1.4505971769815418, + "grad_norm": 6.973959594036107, + "learning_rate": 1.665761037037456e-05, + "loss": 0.8435, + "step": 2672 + }, + { + "epoch": 1.4511400651465798, + "grad_norm": 6.971928597160326, + "learning_rate": 1.665498581615023e-05, + "loss": 0.9065, + "step": 2673 + }, + { + "epoch": 1.4516829533116178, + "grad_norm": 6.1608850694063735, + "learning_rate": 1.665236043882664e-05, + "loss": 0.9289, + "step": 2674 + }, + { + "epoch": 1.4522258414766558, + "grad_norm": 6.62873967155635, + "learning_rate": 1.6649734238728512e-05, + "loss": 0.7797, + "step": 2675 + }, + { + "epoch": 1.4527687296416938, + "grad_norm": 7.089586063125632, + "learning_rate": 1.6647107216180655e-05, + "loss": 0.6306, + "step": 2676 + }, + { + "epoch": 1.4533116178067318, + "grad_norm": 10.986990416571043, + "learning_rate": 1.6644479371507985e-05, + "loss": 0.8764, + "step": 2677 + }, + { + "epoch": 1.4538545059717698, + "grad_norm": 5.865852703001031, + "learning_rate": 1.664185070503551e-05, + "loss": 0.9106, + "step": 2678 + }, + { + "epoch": 1.4543973941368078, + "grad_norm": 7.303323841843, + "learning_rate": 1.663922121708836e-05, + "loss": 0.8188, + "step": 2679 + }, + { + "epoch": 1.4549402823018458, + "grad_norm": 5.715280365147904, + "learning_rate": 1.663659090799175e-05, + "loss": 1.0573, + "step": 2680 + }, + { + "epoch": 1.4554831704668838, + "grad_norm": 7.084313061900368, + "learning_rate": 1.6633959778070992e-05, + "loss": 1.0749, + "step": 2681 + }, + { + "epoch": 1.4560260586319218, + "grad_norm": 7.902311581251679, + "learning_rate": 1.6631327827651524e-05, + "loss": 0.8677, + "step": 2682 + }, + { + "epoch": 1.4565689467969598, + "grad_norm": 8.1594384957164, + "learning_rate": 1.6628695057058855e-05, + "loss": 0.9999, + "step": 2683 + }, + { + "epoch": 1.4571118349619978, + "grad_norm": 5.580141751242927, + "learning_rate": 1.6626061466618623e-05, + "loss": 0.6387, + "step": 2684 + }, + { + "epoch": 1.4576547231270358, + "grad_norm": 8.145754034329466, + "learning_rate": 1.6623427056656544e-05, + "loss": 0.8984, + "step": 2685 + }, + { + "epoch": 1.4581976112920738, + "grad_norm": 6.453637820501138, + "learning_rate": 1.6620791827498454e-05, + "loss": 0.6922, + "step": 2686 + }, + { + "epoch": 1.4587404994571118, + "grad_norm": 7.3511834666776075, + "learning_rate": 1.6618155779470275e-05, + "loss": 0.8462, + "step": 2687 + }, + { + "epoch": 1.4592833876221498, + "grad_norm": 7.962874909840381, + "learning_rate": 1.6615518912898043e-05, + "loss": 0.9639, + "step": 2688 + }, + { + "epoch": 1.4598262757871878, + "grad_norm": 6.975602073560851, + "learning_rate": 1.6612881228107886e-05, + "loss": 0.9364, + "step": 2689 + }, + { + "epoch": 1.4603691639522258, + "grad_norm": 6.916712660881488, + "learning_rate": 1.6610242725426044e-05, + "loss": 0.8033, + "step": 2690 + }, + { + "epoch": 1.4609120521172638, + "grad_norm": 6.268208953592866, + "learning_rate": 1.6607603405178842e-05, + "loss": 0.7542, + "step": 2691 + }, + { + "epoch": 1.4614549402823018, + "grad_norm": 9.047138051843492, + "learning_rate": 1.660496326769272e-05, + "loss": 1.5956, + "step": 2692 + }, + { + "epoch": 1.4619978284473398, + "grad_norm": 7.228556514923675, + "learning_rate": 1.6602322313294216e-05, + "loss": 0.8461, + "step": 2693 + }, + { + "epoch": 1.4625407166123778, + "grad_norm": 6.195288408674038, + "learning_rate": 1.659968054230997e-05, + "loss": 0.9851, + "step": 2694 + }, + { + "epoch": 1.4630836047774158, + "grad_norm": 7.413116367868534, + "learning_rate": 1.6597037955066713e-05, + "loss": 0.9693, + "step": 2695 + }, + { + "epoch": 1.4636264929424538, + "grad_norm": 7.014031390984585, + "learning_rate": 1.6594394551891288e-05, + "loss": 1.0539, + "step": 2696 + }, + { + "epoch": 1.4641693811074918, + "grad_norm": 6.8196013668965225, + "learning_rate": 1.6591750333110634e-05, + "loss": 0.712, + "step": 2697 + }, + { + "epoch": 1.4647122692725298, + "grad_norm": 8.160690900976675, + "learning_rate": 1.658910529905179e-05, + "loss": 0.6551, + "step": 2698 + }, + { + "epoch": 1.4652551574375678, + "grad_norm": 5.174705288640052, + "learning_rate": 1.6586459450041906e-05, + "loss": 0.8285, + "step": 2699 + }, + { + "epoch": 1.4657980456026058, + "grad_norm": 7.029482771999071, + "learning_rate": 1.6583812786408216e-05, + "loss": 0.923, + "step": 2700 + }, + { + "epoch": 1.4663409337676439, + "grad_norm": 7.8877107275923946, + "learning_rate": 1.658116530847807e-05, + "loss": 1.1915, + "step": 2701 + }, + { + "epoch": 1.4668838219326819, + "grad_norm": 9.269305925663183, + "learning_rate": 1.657851701657891e-05, + "loss": 1.098, + "step": 2702 + }, + { + "epoch": 1.4674267100977199, + "grad_norm": 10.478546362437942, + "learning_rate": 1.657586791103828e-05, + "loss": 1.2416, + "step": 2703 + }, + { + "epoch": 1.4679695982627579, + "grad_norm": 8.207786471433751, + "learning_rate": 1.6573217992183826e-05, + "loss": 0.8658, + "step": 2704 + }, + { + "epoch": 1.4685124864277959, + "grad_norm": 7.862118410069405, + "learning_rate": 1.6570567260343294e-05, + "loss": 0.8714, + "step": 2705 + }, + { + "epoch": 1.4690553745928339, + "grad_norm": 7.744957075186954, + "learning_rate": 1.6567915715844534e-05, + "loss": 0.9374, + "step": 2706 + }, + { + "epoch": 1.4695982627578719, + "grad_norm": 6.97424297060321, + "learning_rate": 1.6565263359015488e-05, + "loss": 0.7183, + "step": 2707 + }, + { + "epoch": 1.4701411509229099, + "grad_norm": 7.349900880491086, + "learning_rate": 1.6562610190184206e-05, + "loss": 0.6066, + "step": 2708 + }, + { + "epoch": 1.4706840390879479, + "grad_norm": 6.982564531951699, + "learning_rate": 1.655995620967884e-05, + "loss": 1.2218, + "step": 2709 + }, + { + "epoch": 1.4712269272529859, + "grad_norm": 7.572476514419708, + "learning_rate": 1.6557301417827632e-05, + "loss": 0.89, + "step": 2710 + }, + { + "epoch": 1.4717698154180239, + "grad_norm": 8.654751087453503, + "learning_rate": 1.6554645814958932e-05, + "loss": 0.7188, + "step": 2711 + }, + { + "epoch": 1.4723127035830619, + "grad_norm": 6.282711052010393, + "learning_rate": 1.6551989401401196e-05, + "loss": 0.7586, + "step": 2712 + }, + { + "epoch": 1.4728555917480999, + "grad_norm": 8.732262593936317, + "learning_rate": 1.6549332177482966e-05, + "loss": 1.0178, + "step": 2713 + }, + { + "epoch": 1.4733984799131379, + "grad_norm": 6.989470691687101, + "learning_rate": 1.6546674143532895e-05, + "loss": 0.6981, + "step": 2714 + }, + { + "epoch": 1.4739413680781759, + "grad_norm": 7.179372740296565, + "learning_rate": 1.6544015299879734e-05, + "loss": 0.7617, + "step": 2715 + }, + { + "epoch": 1.4744842562432139, + "grad_norm": 9.117294639273645, + "learning_rate": 1.6541355646852327e-05, + "loss": 0.8822, + "step": 2716 + }, + { + "epoch": 1.475027144408252, + "grad_norm": 7.875727317127435, + "learning_rate": 1.653869518477963e-05, + "loss": 0.9149, + "step": 2717 + }, + { + "epoch": 1.47557003257329, + "grad_norm": 7.926910105501724, + "learning_rate": 1.6536033913990687e-05, + "loss": 0.8797, + "step": 2718 + }, + { + "epoch": 1.476112920738328, + "grad_norm": 5.14878192977047, + "learning_rate": 1.6533371834814657e-05, + "loss": 0.4491, + "step": 2719 + }, + { + "epoch": 1.476655808903366, + "grad_norm": 5.327596982546826, + "learning_rate": 1.6530708947580785e-05, + "loss": 0.4335, + "step": 2720 + }, + { + "epoch": 1.477198697068404, + "grad_norm": 9.158336314582998, + "learning_rate": 1.6528045252618423e-05, + "loss": 1.1208, + "step": 2721 + }, + { + "epoch": 1.477741585233442, + "grad_norm": 10.630652314970348, + "learning_rate": 1.6525380750257022e-05, + "loss": 1.097, + "step": 2722 + }, + { + "epoch": 1.47828447339848, + "grad_norm": 8.97220182564547, + "learning_rate": 1.652271544082613e-05, + "loss": 0.7838, + "step": 2723 + }, + { + "epoch": 1.478827361563518, + "grad_norm": 10.609251261516084, + "learning_rate": 1.652004932465539e-05, + "loss": 0.8861, + "step": 2724 + }, + { + "epoch": 1.479370249728556, + "grad_norm": 7.303801635012744, + "learning_rate": 1.6517382402074563e-05, + "loss": 0.5437, + "step": 2725 + }, + { + "epoch": 1.479913137893594, + "grad_norm": 11.770584674671793, + "learning_rate": 1.651471467341349e-05, + "loss": 1.4947, + "step": 2726 + }, + { + "epoch": 1.480456026058632, + "grad_norm": 6.824299484966605, + "learning_rate": 1.6512046139002128e-05, + "loss": 0.7144, + "step": 2727 + }, + { + "epoch": 1.48099891422367, + "grad_norm": 7.595224046166877, + "learning_rate": 1.650937679917052e-05, + "loss": 1.0259, + "step": 2728 + }, + { + "epoch": 1.481541802388708, + "grad_norm": 8.979875171863954, + "learning_rate": 1.6506706654248813e-05, + "loss": 1.0951, + "step": 2729 + }, + { + "epoch": 1.482084690553746, + "grad_norm": 7.700209331252355, + "learning_rate": 1.650403570456726e-05, + "loss": 1.0228, + "step": 2730 + }, + { + "epoch": 1.482627578718784, + "grad_norm": 9.090566362046857, + "learning_rate": 1.65013639504562e-05, + "loss": 1.0199, + "step": 2731 + }, + { + "epoch": 1.483170466883822, + "grad_norm": 6.644791741009562, + "learning_rate": 1.6498691392246088e-05, + "loss": 0.5478, + "step": 2732 + }, + { + "epoch": 1.48371335504886, + "grad_norm": 7.067718277939746, + "learning_rate": 1.6496018030267467e-05, + "loss": 0.864, + "step": 2733 + }, + { + "epoch": 1.484256243213898, + "grad_norm": 8.017737461642762, + "learning_rate": 1.6493343864850984e-05, + "loss": 1.1688, + "step": 2734 + }, + { + "epoch": 1.484799131378936, + "grad_norm": 6.479930584435578, + "learning_rate": 1.6490668896327382e-05, + "loss": 0.7227, + "step": 2735 + }, + { + "epoch": 1.485342019543974, + "grad_norm": 8.392289588009671, + "learning_rate": 1.6487993125027504e-05, + "loss": 0.7213, + "step": 2736 + }, + { + "epoch": 1.485884907709012, + "grad_norm": 7.817891248616158, + "learning_rate": 1.64853165512823e-05, + "loss": 0.7521, + "step": 2737 + }, + { + "epoch": 1.48642779587405, + "grad_norm": 8.824885297850575, + "learning_rate": 1.6482639175422804e-05, + "loss": 1.016, + "step": 2738 + }, + { + "epoch": 1.486970684039088, + "grad_norm": 11.48655873033237, + "learning_rate": 1.6479960997780165e-05, + "loss": 0.9814, + "step": 2739 + }, + { + "epoch": 1.487513572204126, + "grad_norm": 10.88221555712033, + "learning_rate": 1.6477282018685628e-05, + "loss": 1.464, + "step": 2740 + }, + { + "epoch": 1.488056460369164, + "grad_norm": 8.701247869952772, + "learning_rate": 1.6474602238470524e-05, + "loss": 0.918, + "step": 2741 + }, + { + "epoch": 1.488599348534202, + "grad_norm": 6.4327269926453505, + "learning_rate": 1.6471921657466294e-05, + "loss": 0.7949, + "step": 2742 + }, + { + "epoch": 1.48914223669924, + "grad_norm": 7.592315420975657, + "learning_rate": 1.6469240276004477e-05, + "loss": 0.9485, + "step": 2743 + }, + { + "epoch": 1.489685124864278, + "grad_norm": 7.682698616258038, + "learning_rate": 1.6466558094416717e-05, + "loss": 0.8439, + "step": 2744 + }, + { + "epoch": 1.490228013029316, + "grad_norm": 6.127626205001148, + "learning_rate": 1.6463875113034743e-05, + "loss": 0.8652, + "step": 2745 + }, + { + "epoch": 1.490770901194354, + "grad_norm": 9.65248236696942, + "learning_rate": 1.6461191332190397e-05, + "loss": 1.6278, + "step": 2746 + }, + { + "epoch": 1.491313789359392, + "grad_norm": 8.46309710838681, + "learning_rate": 1.6458506752215603e-05, + "loss": 1.5697, + "step": 2747 + }, + { + "epoch": 1.49185667752443, + "grad_norm": 6.578253971678976, + "learning_rate": 1.6455821373442407e-05, + "loss": 0.8836, + "step": 2748 + }, + { + "epoch": 1.492399565689468, + "grad_norm": 5.927853369454149, + "learning_rate": 1.645313519620293e-05, + "loss": 0.7182, + "step": 2749 + }, + { + "epoch": 1.492942453854506, + "grad_norm": 7.725891960915116, + "learning_rate": 1.645044822082941e-05, + "loss": 1.1794, + "step": 2750 + }, + { + "epoch": 1.493485342019544, + "grad_norm": 8.975220195142995, + "learning_rate": 1.644776044765417e-05, + "loss": 0.8909, + "step": 2751 + }, + { + "epoch": 1.494028230184582, + "grad_norm": 7.34643792387778, + "learning_rate": 1.6445071877009643e-05, + "loss": 1.0047, + "step": 2752 + }, + { + "epoch": 1.49457111834962, + "grad_norm": 7.561064198594243, + "learning_rate": 1.6442382509228355e-05, + "loss": 0.7533, + "step": 2753 + }, + { + "epoch": 1.495114006514658, + "grad_norm": 9.327085390245353, + "learning_rate": 1.6439692344642933e-05, + "loss": 0.8048, + "step": 2754 + }, + { + "epoch": 1.495656894679696, + "grad_norm": 8.662401438640156, + "learning_rate": 1.6437001383586095e-05, + "loss": 1.2513, + "step": 2755 + }, + { + "epoch": 1.496199782844734, + "grad_norm": 7.683989985812991, + "learning_rate": 1.6434309626390667e-05, + "loss": 0.8005, + "step": 2756 + }, + { + "epoch": 1.496742671009772, + "grad_norm": 6.295413547214335, + "learning_rate": 1.6431617073389574e-05, + "loss": 0.8778, + "step": 2757 + }, + { + "epoch": 1.49728555917481, + "grad_norm": 10.390849259618786, + "learning_rate": 1.6428923724915825e-05, + "loss": 1.6234, + "step": 2758 + }, + { + "epoch": 1.497828447339848, + "grad_norm": 6.815996692584369, + "learning_rate": 1.6426229581302545e-05, + "loss": 0.8142, + "step": 2759 + }, + { + "epoch": 1.498371335504886, + "grad_norm": 7.203388917066917, + "learning_rate": 1.642353464288295e-05, + "loss": 1.1273, + "step": 2760 + }, + { + "epoch": 1.498914223669924, + "grad_norm": 6.45381693158517, + "learning_rate": 1.6420838909990356e-05, + "loss": 0.7989, + "step": 2761 + }, + { + "epoch": 1.499457111834962, + "grad_norm": 7.135095705156932, + "learning_rate": 1.6418142382958167e-05, + "loss": 0.6753, + "step": 2762 + }, + { + "epoch": 1.5, + "grad_norm": 8.263302638252904, + "learning_rate": 1.64154450621199e-05, + "loss": 1.0357, + "step": 2763 + }, + { + "epoch": 1.500542888165038, + "grad_norm": 8.19537793615059, + "learning_rate": 1.6412746947809165e-05, + "loss": 0.8135, + "step": 2764 + }, + { + "epoch": 1.501085776330076, + "grad_norm": 5.084112201229359, + "learning_rate": 1.6410048040359665e-05, + "loss": 0.3466, + "step": 2765 + }, + { + "epoch": 1.501628664495114, + "grad_norm": 6.621661501175659, + "learning_rate": 1.6407348340105208e-05, + "loss": 1.1468, + "step": 2766 + }, + { + "epoch": 1.502171552660152, + "grad_norm": 7.5339066223692495, + "learning_rate": 1.6404647847379696e-05, + "loss": 0.8808, + "step": 2767 + }, + { + "epoch": 1.50271444082519, + "grad_norm": 6.501647846893153, + "learning_rate": 1.6401946562517134e-05, + "loss": 0.7589, + "step": 2768 + }, + { + "epoch": 1.503257328990228, + "grad_norm": 8.29356168583173, + "learning_rate": 1.6399244485851614e-05, + "loss": 1.3969, + "step": 2769 + }, + { + "epoch": 1.503800217155266, + "grad_norm": 6.518146966456392, + "learning_rate": 1.6396541617717337e-05, + "loss": 0.7951, + "step": 2770 + }, + { + "epoch": 1.504343105320304, + "grad_norm": 6.51134736256057, + "learning_rate": 1.63938379584486e-05, + "loss": 0.9073, + "step": 2771 + }, + { + "epoch": 1.504885993485342, + "grad_norm": 6.682459044312825, + "learning_rate": 1.6391133508379797e-05, + "loss": 0.733, + "step": 2772 + }, + { + "epoch": 1.50542888165038, + "grad_norm": 6.390687464891663, + "learning_rate": 1.638842826784541e-05, + "loss": 0.6934, + "step": 2773 + }, + { + "epoch": 1.505971769815418, + "grad_norm": 8.130197298030478, + "learning_rate": 1.6385722237180038e-05, + "loss": 0.8567, + "step": 2774 + }, + { + "epoch": 1.506514657980456, + "grad_norm": 7.959636836831885, + "learning_rate": 1.6383015416718356e-05, + "loss": 0.9926, + "step": 2775 + }, + { + "epoch": 1.507057546145494, + "grad_norm": 7.732366564125716, + "learning_rate": 1.638030780679516e-05, + "loss": 1.2228, + "step": 2776 + }, + { + "epoch": 1.507600434310532, + "grad_norm": 4.633651796670195, + "learning_rate": 1.6377599407745324e-05, + "loss": 0.4633, + "step": 2777 + }, + { + "epoch": 1.50814332247557, + "grad_norm": 7.016748827947926, + "learning_rate": 1.6374890219903828e-05, + "loss": 0.9593, + "step": 2778 + }, + { + "epoch": 1.508686210640608, + "grad_norm": 8.428067394567327, + "learning_rate": 1.637218024360575e-05, + "loss": 1.0493, + "step": 2779 + }, + { + "epoch": 1.509229098805646, + "grad_norm": 8.297567044395613, + "learning_rate": 1.6369469479186266e-05, + "loss": 1.2335, + "step": 2780 + }, + { + "epoch": 1.509771986970684, + "grad_norm": 5.8755950768313205, + "learning_rate": 1.6366757926980643e-05, + "loss": 0.6556, + "step": 2781 + }, + { + "epoch": 1.510314875135722, + "grad_norm": 6.837923725489964, + "learning_rate": 1.6364045587324254e-05, + "loss": 0.7031, + "step": 2782 + }, + { + "epoch": 1.51085776330076, + "grad_norm": 6.718777411432131, + "learning_rate": 1.6361332460552565e-05, + "loss": 0.8982, + "step": 2783 + }, + { + "epoch": 1.511400651465798, + "grad_norm": 8.788667184000067, + "learning_rate": 1.6358618547001137e-05, + "loss": 0.8548, + "step": 2784 + }, + { + "epoch": 1.511943539630836, + "grad_norm": 6.559897969720367, + "learning_rate": 1.635590384700563e-05, + "loss": 0.9627, + "step": 2785 + }, + { + "epoch": 1.512486427795874, + "grad_norm": 6.538968751920423, + "learning_rate": 1.635318836090181e-05, + "loss": 1.0731, + "step": 2786 + }, + { + "epoch": 1.513029315960912, + "grad_norm": 8.149646454407613, + "learning_rate": 1.6350472089025523e-05, + "loss": 0.9898, + "step": 2787 + }, + { + "epoch": 1.51357220412595, + "grad_norm": 8.544281989077668, + "learning_rate": 1.6347755031712734e-05, + "loss": 1.1125, + "step": 2788 + }, + { + "epoch": 1.514115092290988, + "grad_norm": 6.930736082842491, + "learning_rate": 1.634503718929948e-05, + "loss": 0.7617, + "step": 2789 + }, + { + "epoch": 1.514657980456026, + "grad_norm": 8.527931500643358, + "learning_rate": 1.6342318562121916e-05, + "loss": 1.2179, + "step": 2790 + }, + { + "epoch": 1.515200868621064, + "grad_norm": 7.509610271006853, + "learning_rate": 1.6339599150516283e-05, + "loss": 0.908, + "step": 2791 + }, + { + "epoch": 1.515743756786102, + "grad_norm": 6.0902820346277045, + "learning_rate": 1.6336878954818926e-05, + "loss": 0.829, + "step": 2792 + }, + { + "epoch": 1.51628664495114, + "grad_norm": 7.3274444131512935, + "learning_rate": 1.6334157975366278e-05, + "loss": 0.9667, + "step": 2793 + }, + { + "epoch": 1.516829533116178, + "grad_norm": 7.753641450951333, + "learning_rate": 1.633143621249488e-05, + "loss": 0.7961, + "step": 2794 + }, + { + "epoch": 1.517372421281216, + "grad_norm": 14.786266283108317, + "learning_rate": 1.6328713666541357e-05, + "loss": 1.4119, + "step": 2795 + }, + { + "epoch": 1.517915309446254, + "grad_norm": 6.1388168060810875, + "learning_rate": 1.632599033784244e-05, + "loss": 0.7726, + "step": 2796 + }, + { + "epoch": 1.518458197611292, + "grad_norm": 9.134190897524304, + "learning_rate": 1.632326622673496e-05, + "loss": 1.208, + "step": 2797 + }, + { + "epoch": 1.51900108577633, + "grad_norm": 6.898277731717493, + "learning_rate": 1.632054133355583e-05, + "loss": 0.6929, + "step": 2798 + }, + { + "epoch": 1.519543973941368, + "grad_norm": 6.90513711625334, + "learning_rate": 1.631781565864208e-05, + "loss": 0.7279, + "step": 2799 + }, + { + "epoch": 1.520086862106406, + "grad_norm": 8.332173251791465, + "learning_rate": 1.6315089202330817e-05, + "loss": 0.987, + "step": 2800 + }, + { + "epoch": 1.520629750271444, + "grad_norm": 6.783078863499027, + "learning_rate": 1.631236196495926e-05, + "loss": 0.6793, + "step": 2801 + }, + { + "epoch": 1.521172638436482, + "grad_norm": 7.841204229725534, + "learning_rate": 1.6309633946864712e-05, + "loss": 0.832, + "step": 2802 + }, + { + "epoch": 1.52171552660152, + "grad_norm": 6.4931904342190405, + "learning_rate": 1.630690514838458e-05, + "loss": 0.6662, + "step": 2803 + }, + { + "epoch": 1.522258414766558, + "grad_norm": 8.50187086749661, + "learning_rate": 1.6304175569856368e-05, + "loss": 0.8359, + "step": 2804 + }, + { + "epoch": 1.522801302931596, + "grad_norm": 6.196174007418195, + "learning_rate": 1.6301445211617676e-05, + "loss": 0.8437, + "step": 2805 + }, + { + "epoch": 1.523344191096634, + "grad_norm": 8.232047070633076, + "learning_rate": 1.6298714074006196e-05, + "loss": 0.9134, + "step": 2806 + }, + { + "epoch": 1.523887079261672, + "grad_norm": 6.594428796101092, + "learning_rate": 1.629598215735972e-05, + "loss": 0.6703, + "step": 2807 + }, + { + "epoch": 1.52442996742671, + "grad_norm": 7.634546429927066, + "learning_rate": 1.629324946201614e-05, + "loss": 0.8208, + "step": 2808 + }, + { + "epoch": 1.524972855591748, + "grad_norm": 8.225245989940067, + "learning_rate": 1.6290515988313432e-05, + "loss": 1.0002, + "step": 2809 + }, + { + "epoch": 1.5255157437567861, + "grad_norm": 8.334512060802671, + "learning_rate": 1.628778173658968e-05, + "loss": 0.8883, + "step": 2810 + }, + { + "epoch": 1.5260586319218241, + "grad_norm": 7.733244019009635, + "learning_rate": 1.6285046707183068e-05, + "loss": 0.7155, + "step": 2811 + }, + { + "epoch": 1.5266015200868621, + "grad_norm": 9.607081900344873, + "learning_rate": 1.628231090043186e-05, + "loss": 0.8321, + "step": 2812 + }, + { + "epoch": 1.5271444082519001, + "grad_norm": 8.390046905361096, + "learning_rate": 1.6279574316674426e-05, + "loss": 0.8818, + "step": 2813 + }, + { + "epoch": 1.5276872964169381, + "grad_norm": 7.608191285481891, + "learning_rate": 1.6276836956249235e-05, + "loss": 0.7382, + "step": 2814 + }, + { + "epoch": 1.5282301845819761, + "grad_norm": 8.998269112751723, + "learning_rate": 1.6274098819494844e-05, + "loss": 0.9285, + "step": 2815 + }, + { + "epoch": 1.5287730727470141, + "grad_norm": 5.877268783539307, + "learning_rate": 1.627135990674991e-05, + "loss": 0.8434, + "step": 2816 + }, + { + "epoch": 1.5293159609120521, + "grad_norm": 6.956730702284698, + "learning_rate": 1.6268620218353188e-05, + "loss": 0.5463, + "step": 2817 + }, + { + "epoch": 1.5298588490770901, + "grad_norm": 6.516430093047641, + "learning_rate": 1.626587975464353e-05, + "loss": 0.4886, + "step": 2818 + }, + { + "epoch": 1.5304017372421281, + "grad_norm": 6.751333342823053, + "learning_rate": 1.626313851595987e-05, + "loss": 0.6279, + "step": 2819 + }, + { + "epoch": 1.5309446254071661, + "grad_norm": 8.120808986494415, + "learning_rate": 1.6260396502641264e-05, + "loss": 0.8192, + "step": 2820 + }, + { + "epoch": 1.5314875135722041, + "grad_norm": 7.570228095805488, + "learning_rate": 1.6257653715026837e-05, + "loss": 0.5813, + "step": 2821 + }, + { + "epoch": 1.5320304017372421, + "grad_norm": 8.436330907807918, + "learning_rate": 1.625491015345583e-05, + "loss": 0.9115, + "step": 2822 + }, + { + "epoch": 1.5325732899022801, + "grad_norm": 8.384048301051982, + "learning_rate": 1.6252165818267564e-05, + "loss": 1.0131, + "step": 2823 + }, + { + "epoch": 1.5331161780673181, + "grad_norm": 6.665532827172479, + "learning_rate": 1.6249420709801462e-05, + "loss": 0.5454, + "step": 2824 + }, + { + "epoch": 1.5336590662323561, + "grad_norm": 7.096056294730765, + "learning_rate": 1.624667482839705e-05, + "loss": 1.043, + "step": 2825 + }, + { + "epoch": 1.5342019543973942, + "grad_norm": 5.384161272714602, + "learning_rate": 1.6243928174393935e-05, + "loss": 0.8345, + "step": 2826 + }, + { + "epoch": 1.5347448425624322, + "grad_norm": 8.398527018753486, + "learning_rate": 1.6241180748131834e-05, + "loss": 0.6911, + "step": 2827 + }, + { + "epoch": 1.5352877307274702, + "grad_norm": 6.880924484125559, + "learning_rate": 1.6238432549950552e-05, + "loss": 0.6039, + "step": 2828 + }, + { + "epoch": 1.5358306188925082, + "grad_norm": 9.23669884568822, + "learning_rate": 1.623568358018999e-05, + "loss": 1.1145, + "step": 2829 + }, + { + "epoch": 1.5363735070575462, + "grad_norm": 9.838376476241027, + "learning_rate": 1.6232933839190146e-05, + "loss": 1.0435, + "step": 2830 + }, + { + "epoch": 1.5369163952225842, + "grad_norm": 8.64475205930121, + "learning_rate": 1.6230183327291108e-05, + "loss": 1.0886, + "step": 2831 + }, + { + "epoch": 1.5374592833876222, + "grad_norm": 8.446847258487077, + "learning_rate": 1.6227432044833072e-05, + "loss": 0.8922, + "step": 2832 + }, + { + "epoch": 1.5380021715526602, + "grad_norm": 7.373724851347831, + "learning_rate": 1.622467999215631e-05, + "loss": 0.7823, + "step": 2833 + }, + { + "epoch": 1.5385450597176982, + "grad_norm": 9.661226436708493, + "learning_rate": 1.622192716960121e-05, + "loss": 1.2651, + "step": 2834 + }, + { + "epoch": 1.5390879478827362, + "grad_norm": 9.316115050127708, + "learning_rate": 1.6219173577508237e-05, + "loss": 0.897, + "step": 2835 + }, + { + "epoch": 1.5396308360477742, + "grad_norm": 12.011440258586251, + "learning_rate": 1.621641921621797e-05, + "loss": 0.868, + "step": 2836 + }, + { + "epoch": 1.5401737242128122, + "grad_norm": 7.737580132404622, + "learning_rate": 1.6213664086071058e-05, + "loss": 1.0727, + "step": 2837 + }, + { + "epoch": 1.5407166123778502, + "grad_norm": 6.696935187657473, + "learning_rate": 1.6210908187408275e-05, + "loss": 0.7404, + "step": 2838 + }, + { + "epoch": 1.5412595005428882, + "grad_norm": 6.460846604904407, + "learning_rate": 1.6208151520570465e-05, + "loss": 0.5892, + "step": 2839 + }, + { + "epoch": 1.5418023887079262, + "grad_norm": 8.244871763303658, + "learning_rate": 1.6205394085898586e-05, + "loss": 1.0692, + "step": 2840 + }, + { + "epoch": 1.5423452768729642, + "grad_norm": 6.140143281113326, + "learning_rate": 1.620263588373367e-05, + "loss": 0.7648, + "step": 2841 + }, + { + "epoch": 1.5428881650380022, + "grad_norm": 6.3369705303866235, + "learning_rate": 1.619987691441687e-05, + "loss": 0.6665, + "step": 2842 + }, + { + "epoch": 1.5434310532030402, + "grad_norm": 8.864356355873657, + "learning_rate": 1.6197117178289405e-05, + "loss": 0.9855, + "step": 2843 + }, + { + "epoch": 1.5439739413680782, + "grad_norm": 8.487808912288623, + "learning_rate": 1.6194356675692614e-05, + "loss": 1.3813, + "step": 2844 + }, + { + "epoch": 1.5445168295331162, + "grad_norm": 6.5782618837581435, + "learning_rate": 1.619159540696792e-05, + "loss": 1.1941, + "step": 2845 + }, + { + "epoch": 1.5450597176981542, + "grad_norm": 8.668580644259926, + "learning_rate": 1.6188833372456833e-05, + "loss": 1.1044, + "step": 2846 + }, + { + "epoch": 1.5456026058631922, + "grad_norm": 8.677506532765584, + "learning_rate": 1.6186070572500972e-05, + "loss": 1.3984, + "step": 2847 + }, + { + "epoch": 1.5461454940282302, + "grad_norm": 6.822422879140709, + "learning_rate": 1.6183307007442046e-05, + "loss": 0.6903, + "step": 2848 + }, + { + "epoch": 1.5466883821932682, + "grad_norm": 8.480828664812217, + "learning_rate": 1.6180542677621852e-05, + "loss": 1.2466, + "step": 2849 + }, + { + "epoch": 1.5472312703583062, + "grad_norm": 7.804300661419004, + "learning_rate": 1.617777758338229e-05, + "loss": 0.8833, + "step": 2850 + }, + { + "epoch": 1.5477741585233442, + "grad_norm": 6.418809144902363, + "learning_rate": 1.617501172506535e-05, + "loss": 0.9174, + "step": 2851 + }, + { + "epoch": 1.5483170466883822, + "grad_norm": 6.627083812958508, + "learning_rate": 1.617224510301312e-05, + "loss": 0.7867, + "step": 2852 + }, + { + "epoch": 1.5488599348534202, + "grad_norm": 7.792114190076133, + "learning_rate": 1.616947771756778e-05, + "loss": 0.6242, + "step": 2853 + }, + { + "epoch": 1.5494028230184582, + "grad_norm": 6.965108246318713, + "learning_rate": 1.6166709569071598e-05, + "loss": 0.5844, + "step": 2854 + }, + { + "epoch": 1.5499457111834962, + "grad_norm": 6.9966373703206655, + "learning_rate": 1.616394065786695e-05, + "loss": 1.1688, + "step": 2855 + }, + { + "epoch": 1.5504885993485342, + "grad_norm": 8.466603441198636, + "learning_rate": 1.6161170984296298e-05, + "loss": 0.9506, + "step": 2856 + }, + { + "epoch": 1.5510314875135722, + "grad_norm": 6.361127259105106, + "learning_rate": 1.61584005487022e-05, + "loss": 0.5573, + "step": 2857 + }, + { + "epoch": 1.5515743756786102, + "grad_norm": 6.137955961198825, + "learning_rate": 1.6155629351427306e-05, + "loss": 1.0474, + "step": 2858 + }, + { + "epoch": 1.5521172638436482, + "grad_norm": 6.262211837195136, + "learning_rate": 1.6152857392814367e-05, + "loss": 0.6786, + "step": 2859 + }, + { + "epoch": 1.5526601520086862, + "grad_norm": 9.3980797837583, + "learning_rate": 1.6150084673206214e-05, + "loss": 1.2893, + "step": 2860 + }, + { + "epoch": 1.5532030401737242, + "grad_norm": 9.600537200283373, + "learning_rate": 1.614731119294579e-05, + "loss": 0.8668, + "step": 2861 + }, + { + "epoch": 1.5537459283387622, + "grad_norm": 8.65304941049683, + "learning_rate": 1.614453695237612e-05, + "loss": 1.0377, + "step": 2862 + }, + { + "epoch": 1.5542888165038002, + "grad_norm": 6.075747510405264, + "learning_rate": 1.6141761951840327e-05, + "loss": 0.8203, + "step": 2863 + }, + { + "epoch": 1.5548317046688382, + "grad_norm": 6.578770531031146, + "learning_rate": 1.6138986191681626e-05, + "loss": 0.6808, + "step": 2864 + }, + { + "epoch": 1.5553745928338762, + "grad_norm": 9.206515322081769, + "learning_rate": 1.6136209672243332e-05, + "loss": 0.9111, + "step": 2865 + }, + { + "epoch": 1.5559174809989142, + "grad_norm": 8.829723041453656, + "learning_rate": 1.613343239386884e-05, + "loss": 1.7392, + "step": 2866 + }, + { + "epoch": 1.5564603691639523, + "grad_norm": 9.489367814866748, + "learning_rate": 1.613065435690166e-05, + "loss": 0.9186, + "step": 2867 + }, + { + "epoch": 1.5570032573289903, + "grad_norm": 7.7196736231335175, + "learning_rate": 1.6127875561685376e-05, + "loss": 0.769, + "step": 2868 + }, + { + "epoch": 1.5575461454940283, + "grad_norm": 7.668511006453606, + "learning_rate": 1.6125096008563677e-05, + "loss": 0.7572, + "step": 2869 + }, + { + "epoch": 1.5580890336590663, + "grad_norm": 8.31551289254719, + "learning_rate": 1.6122315697880343e-05, + "loss": 1.0657, + "step": 2870 + }, + { + "epoch": 1.5586319218241043, + "grad_norm": 11.828690633927348, + "learning_rate": 1.6119534629979244e-05, + "loss": 1.0095, + "step": 2871 + }, + { + "epoch": 1.5591748099891423, + "grad_norm": 7.137823181609677, + "learning_rate": 1.611675280520435e-05, + "loss": 0.7246, + "step": 2872 + }, + { + "epoch": 1.5597176981541803, + "grad_norm": 5.051313327297567, + "learning_rate": 1.611397022389972e-05, + "loss": 0.5905, + "step": 2873 + }, + { + "epoch": 1.5602605863192183, + "grad_norm": 8.165805991671068, + "learning_rate": 1.6111186886409504e-05, + "loss": 1.0678, + "step": 2874 + }, + { + "epoch": 1.5608034744842563, + "grad_norm": 11.597649604311457, + "learning_rate": 1.6108402793077957e-05, + "loss": 1.4836, + "step": 2875 + }, + { + "epoch": 1.5613463626492943, + "grad_norm": 7.953203184026488, + "learning_rate": 1.610561794424942e-05, + "loss": 0.9216, + "step": 2876 + }, + { + "epoch": 1.5618892508143323, + "grad_norm": 7.397930684647979, + "learning_rate": 1.6102832340268322e-05, + "loss": 0.6688, + "step": 2877 + }, + { + "epoch": 1.5624321389793703, + "grad_norm": 8.999084750858366, + "learning_rate": 1.6100045981479195e-05, + "loss": 1.254, + "step": 2878 + }, + { + "epoch": 1.5629750271444083, + "grad_norm": 7.445594920489225, + "learning_rate": 1.6097258868226658e-05, + "loss": 0.6406, + "step": 2879 + }, + { + "epoch": 1.5635179153094463, + "grad_norm": 10.469355169581702, + "learning_rate": 1.609447100085543e-05, + "loss": 0.7928, + "step": 2880 + }, + { + "epoch": 1.5640608034744843, + "grad_norm": 8.853634937458189, + "learning_rate": 1.6091682379710313e-05, + "loss": 1.0788, + "step": 2881 + }, + { + "epoch": 1.5646036916395223, + "grad_norm": 7.909620213759866, + "learning_rate": 1.6088893005136206e-05, + "loss": 1.011, + "step": 2882 + }, + { + "epoch": 1.5651465798045603, + "grad_norm": 10.379252728708625, + "learning_rate": 1.6086102877478117e-05, + "loss": 0.8675, + "step": 2883 + }, + { + "epoch": 1.5656894679695983, + "grad_norm": 8.143888785034273, + "learning_rate": 1.6083311997081116e-05, + "loss": 0.8935, + "step": 2884 + }, + { + "epoch": 1.5662323561346363, + "grad_norm": 8.577520797942114, + "learning_rate": 1.6080520364290396e-05, + "loss": 0.6763, + "step": 2885 + }, + { + "epoch": 1.5667752442996743, + "grad_norm": 9.226094142154105, + "learning_rate": 1.6077727979451228e-05, + "loss": 0.6679, + "step": 2886 + }, + { + "epoch": 1.5673181324647123, + "grad_norm": 11.197536431412058, + "learning_rate": 1.607493484290897e-05, + "loss": 1.0525, + "step": 2887 + }, + { + "epoch": 1.5678610206297503, + "grad_norm": 8.43012225245798, + "learning_rate": 1.6072140955009093e-05, + "loss": 1.0317, + "step": 2888 + }, + { + "epoch": 1.5684039087947883, + "grad_norm": 6.69358812320454, + "learning_rate": 1.606934631609715e-05, + "loss": 0.7703, + "step": 2889 + }, + { + "epoch": 1.5689467969598263, + "grad_norm": 9.36327291463899, + "learning_rate": 1.6066550926518776e-05, + "loss": 1.0844, + "step": 2890 + }, + { + "epoch": 1.5694896851248643, + "grad_norm": 8.40518653624583, + "learning_rate": 1.6063754786619716e-05, + "loss": 0.6863, + "step": 2891 + }, + { + "epoch": 1.5700325732899023, + "grad_norm": 9.552486022227878, + "learning_rate": 1.60609578967458e-05, + "loss": 1.0342, + "step": 2892 + }, + { + "epoch": 1.5705754614549403, + "grad_norm": 6.76445711513327, + "learning_rate": 1.6058160257242953e-05, + "loss": 0.8777, + "step": 2893 + }, + { + "epoch": 1.5711183496199783, + "grad_norm": 9.040602694092291, + "learning_rate": 1.6055361868457188e-05, + "loss": 0.9634, + "step": 2894 + }, + { + "epoch": 1.5716612377850163, + "grad_norm": 7.00278619777266, + "learning_rate": 1.6052562730734614e-05, + "loss": 0.8995, + "step": 2895 + }, + { + "epoch": 1.5722041259500543, + "grad_norm": 8.533327622490077, + "learning_rate": 1.604976284442144e-05, + "loss": 1.0896, + "step": 2896 + }, + { + "epoch": 1.5727470141150923, + "grad_norm": 9.895218231842426, + "learning_rate": 1.6046962209863953e-05, + "loss": 0.9309, + "step": 2897 + }, + { + "epoch": 1.5732899022801303, + "grad_norm": 5.945823958044379, + "learning_rate": 1.604416082740854e-05, + "loss": 0.6512, + "step": 2898 + }, + { + "epoch": 1.5738327904451683, + "grad_norm": 10.822429137471252, + "learning_rate": 1.6041358697401687e-05, + "loss": 1.0744, + "step": 2899 + }, + { + "epoch": 1.5743756786102063, + "grad_norm": 7.728147258511819, + "learning_rate": 1.603855582018996e-05, + "loss": 0.9566, + "step": 2900 + }, + { + "epoch": 1.5749185667752443, + "grad_norm": 8.643748761880973, + "learning_rate": 1.603575219612003e-05, + "loss": 0.9188, + "step": 2901 + }, + { + "epoch": 1.5754614549402823, + "grad_norm": 9.357572097357403, + "learning_rate": 1.603294782553864e-05, + "loss": 0.9045, + "step": 2902 + }, + { + "epoch": 1.5760043431053203, + "grad_norm": 6.76207345266047, + "learning_rate": 1.6030142708792653e-05, + "loss": 0.7658, + "step": 2903 + }, + { + "epoch": 1.5765472312703583, + "grad_norm": 6.47603683733189, + "learning_rate": 1.6027336846229005e-05, + "loss": 0.5406, + "step": 2904 + }, + { + "epoch": 1.5770901194353963, + "grad_norm": 6.030033648190992, + "learning_rate": 1.602453023819473e-05, + "loss": 0.5141, + "step": 2905 + }, + { + "epoch": 1.5776330076004343, + "grad_norm": 8.344821553681937, + "learning_rate": 1.6021722885036954e-05, + "loss": 1.255, + "step": 2906 + }, + { + "epoch": 1.5781758957654723, + "grad_norm": 7.2392259468937885, + "learning_rate": 1.601891478710289e-05, + "loss": 0.8358, + "step": 2907 + }, + { + "epoch": 1.5787187839305103, + "grad_norm": 8.011046574978268, + "learning_rate": 1.6016105944739856e-05, + "loss": 1.0316, + "step": 2908 + }, + { + "epoch": 1.5792616720955484, + "grad_norm": 6.609140115647256, + "learning_rate": 1.601329635829525e-05, + "loss": 0.7324, + "step": 2909 + }, + { + "epoch": 1.5798045602605864, + "grad_norm": 6.672864395642082, + "learning_rate": 1.6010486028116568e-05, + "loss": 0.6386, + "step": 2910 + }, + { + "epoch": 1.5803474484256244, + "grad_norm": 7.270447246601273, + "learning_rate": 1.600767495455139e-05, + "loss": 0.6699, + "step": 2911 + }, + { + "epoch": 1.5808903365906624, + "grad_norm": 9.7186017734922, + "learning_rate": 1.6004863137947405e-05, + "loss": 0.9604, + "step": 2912 + }, + { + "epoch": 1.5814332247557004, + "grad_norm": 7.18305646407285, + "learning_rate": 1.6002050578652374e-05, + "loss": 0.7286, + "step": 2913 + }, + { + "epoch": 1.5819761129207384, + "grad_norm": 8.171408196684215, + "learning_rate": 1.5999237277014162e-05, + "loss": 0.9121, + "step": 2914 + }, + { + "epoch": 1.5825190010857764, + "grad_norm": 7.710673421134947, + "learning_rate": 1.599642323338072e-05, + "loss": 0.7287, + "step": 2915 + }, + { + "epoch": 1.5830618892508144, + "grad_norm": 8.264934044661224, + "learning_rate": 1.5993608448100095e-05, + "loss": 0.6537, + "step": 2916 + }, + { + "epoch": 1.5836047774158524, + "grad_norm": 7.874914961028911, + "learning_rate": 1.599079292152043e-05, + "loss": 0.6511, + "step": 2917 + }, + { + "epoch": 1.5841476655808904, + "grad_norm": 7.399378371785218, + "learning_rate": 1.5987976653989945e-05, + "loss": 0.6236, + "step": 2918 + }, + { + "epoch": 1.5846905537459284, + "grad_norm": 7.837764322959507, + "learning_rate": 1.5985159645856966e-05, + "loss": 0.8184, + "step": 2919 + }, + { + "epoch": 1.5852334419109664, + "grad_norm": 8.64272103945922, + "learning_rate": 1.5982341897469903e-05, + "loss": 0.9182, + "step": 2920 + }, + { + "epoch": 1.5857763300760044, + "grad_norm": 9.994158036669248, + "learning_rate": 1.5979523409177254e-05, + "loss": 1.3466, + "step": 2921 + }, + { + "epoch": 1.5863192182410424, + "grad_norm": 6.671268632879107, + "learning_rate": 1.5976704181327626e-05, + "loss": 0.7506, + "step": 2922 + }, + { + "epoch": 1.5868621064060804, + "grad_norm": 8.161895609948308, + "learning_rate": 1.59738842142697e-05, + "loss": 0.9683, + "step": 2923 + }, + { + "epoch": 1.5874049945711184, + "grad_norm": 7.236619510875819, + "learning_rate": 1.597106350835225e-05, + "loss": 0.798, + "step": 2924 + }, + { + "epoch": 1.5879478827361564, + "grad_norm": 10.263209688443808, + "learning_rate": 1.5968242063924152e-05, + "loss": 1.0799, + "step": 2925 + }, + { + "epoch": 1.5884907709011944, + "grad_norm": 9.433308260958457, + "learning_rate": 1.596541988133436e-05, + "loss": 0.8755, + "step": 2926 + }, + { + "epoch": 1.5890336590662324, + "grad_norm": 9.48995441422385, + "learning_rate": 1.5962596960931927e-05, + "loss": 1.4712, + "step": 2927 + }, + { + "epoch": 1.5895765472312704, + "grad_norm": 7.317297215230638, + "learning_rate": 1.5959773303066005e-05, + "loss": 1.0421, + "step": 2928 + }, + { + "epoch": 1.5901194353963084, + "grad_norm": 7.491426794362114, + "learning_rate": 1.595694890808582e-05, + "loss": 0.9468, + "step": 2929 + }, + { + "epoch": 1.5906623235613464, + "grad_norm": 7.320440852081977, + "learning_rate": 1.5954123776340702e-05, + "loss": 0.7459, + "step": 2930 + }, + { + "epoch": 1.5912052117263844, + "grad_norm": 8.334158307527314, + "learning_rate": 1.5951297908180062e-05, + "loss": 1.3182, + "step": 2931 + }, + { + "epoch": 1.5917480998914224, + "grad_norm": 7.02613629768153, + "learning_rate": 1.5948471303953418e-05, + "loss": 0.919, + "step": 2932 + }, + { + "epoch": 1.5922909880564604, + "grad_norm": 6.904443487786306, + "learning_rate": 1.594564396401036e-05, + "loss": 0.5737, + "step": 2933 + }, + { + "epoch": 1.5928338762214984, + "grad_norm": 8.041576521446402, + "learning_rate": 1.594281588870058e-05, + "loss": 0.7191, + "step": 2934 + }, + { + "epoch": 1.5933767643865364, + "grad_norm": 8.557472225459021, + "learning_rate": 1.5939987078373856e-05, + "loss": 0.8936, + "step": 2935 + }, + { + "epoch": 1.5939196525515744, + "grad_norm": 10.051674589973098, + "learning_rate": 1.5937157533380065e-05, + "loss": 1.1777, + "step": 2936 + }, + { + "epoch": 1.5944625407166124, + "grad_norm": 7.1580018473739235, + "learning_rate": 1.5934327254069167e-05, + "loss": 0.8628, + "step": 2937 + }, + { + "epoch": 1.5950054288816504, + "grad_norm": 8.689338133459966, + "learning_rate": 1.593149624079122e-05, + "loss": 1.1877, + "step": 2938 + }, + { + "epoch": 1.5955483170466884, + "grad_norm": 8.40904043481009, + "learning_rate": 1.5928664493896364e-05, + "loss": 0.9815, + "step": 2939 + }, + { + "epoch": 1.5960912052117264, + "grad_norm": 7.778305354024487, + "learning_rate": 1.5925832013734832e-05, + "loss": 0.8282, + "step": 2940 + }, + { + "epoch": 1.5966340933767644, + "grad_norm": 7.074281482828989, + "learning_rate": 1.5922998800656956e-05, + "loss": 0.8347, + "step": 2941 + }, + { + "epoch": 1.5971769815418024, + "grad_norm": 9.188840060350078, + "learning_rate": 1.5920164855013145e-05, + "loss": 0.8566, + "step": 2942 + }, + { + "epoch": 1.5977198697068404, + "grad_norm": 7.021116417649334, + "learning_rate": 1.591733017715391e-05, + "loss": 0.8367, + "step": 2943 + }, + { + "epoch": 1.5982627578718784, + "grad_norm": 8.039041768789302, + "learning_rate": 1.5914494767429846e-05, + "loss": 0.9889, + "step": 2944 + }, + { + "epoch": 1.5988056460369164, + "grad_norm": 6.252701385804125, + "learning_rate": 1.5911658626191645e-05, + "loss": 1.0478, + "step": 2945 + }, + { + "epoch": 1.5993485342019544, + "grad_norm": 8.223409826369243, + "learning_rate": 1.5908821753790083e-05, + "loss": 1.2458, + "step": 2946 + }, + { + "epoch": 1.5998914223669924, + "grad_norm": 8.972690403999096, + "learning_rate": 1.590598415057603e-05, + "loss": 0.8471, + "step": 2947 + }, + { + "epoch": 1.6004343105320304, + "grad_norm": 8.163032020074416, + "learning_rate": 1.5903145816900445e-05, + "loss": 0.8213, + "step": 2948 + }, + { + "epoch": 1.6009771986970684, + "grad_norm": 5.667868413023322, + "learning_rate": 1.5900306753114375e-05, + "loss": 0.8426, + "step": 2949 + }, + { + "epoch": 1.6015200868621065, + "grad_norm": 6.960744978365833, + "learning_rate": 1.5897466959568967e-05, + "loss": 0.7224, + "step": 2950 + }, + { + "epoch": 1.6020629750271445, + "grad_norm": 7.648210267430006, + "learning_rate": 1.589462643661544e-05, + "loss": 0.7451, + "step": 2951 + }, + { + "epoch": 1.6026058631921825, + "grad_norm": 5.9568596208122875, + "learning_rate": 1.5891785184605123e-05, + "loss": 0.5252, + "step": 2952 + }, + { + "epoch": 1.6031487513572205, + "grad_norm": 6.41676218105273, + "learning_rate": 1.5888943203889427e-05, + "loss": 0.6502, + "step": 2953 + }, + { + "epoch": 1.6036916395222585, + "grad_norm": 7.704871607753619, + "learning_rate": 1.5886100494819846e-05, + "loss": 1.1762, + "step": 2954 + }, + { + "epoch": 1.6042345276872965, + "grad_norm": 7.656219757403813, + "learning_rate": 1.5883257057747975e-05, + "loss": 0.8076, + "step": 2955 + }, + { + "epoch": 1.6047774158523345, + "grad_norm": 6.897030921112048, + "learning_rate": 1.58804128930255e-05, + "loss": 0.883, + "step": 2956 + }, + { + "epoch": 1.6053203040173725, + "grad_norm": 6.718724904599124, + "learning_rate": 1.5877568001004182e-05, + "loss": 0.8533, + "step": 2957 + }, + { + "epoch": 1.6058631921824105, + "grad_norm": 8.458861423777194, + "learning_rate": 1.5874722382035887e-05, + "loss": 1.2017, + "step": 2958 + }, + { + "epoch": 1.6064060803474485, + "grad_norm": 9.16622725543938, + "learning_rate": 1.5871876036472565e-05, + "loss": 1.0074, + "step": 2959 + }, + { + "epoch": 1.6069489685124865, + "grad_norm": 7.053786350768386, + "learning_rate": 1.5869028964666254e-05, + "loss": 0.8193, + "step": 2960 + }, + { + "epoch": 1.6074918566775245, + "grad_norm": 7.670976634760575, + "learning_rate": 1.5866181166969088e-05, + "loss": 0.8331, + "step": 2961 + }, + { + "epoch": 1.6080347448425625, + "grad_norm": 9.942186867470006, + "learning_rate": 1.586333264373329e-05, + "loss": 0.8672, + "step": 2962 + }, + { + "epoch": 1.6085776330076005, + "grad_norm": 7.744463789862724, + "learning_rate": 1.586048339531116e-05, + "loss": 0.8859, + "step": 2963 + }, + { + "epoch": 1.6091205211726385, + "grad_norm": 7.181553947846623, + "learning_rate": 1.5857633422055104e-05, + "loss": 1.0376, + "step": 2964 + }, + { + "epoch": 1.6096634093376765, + "grad_norm": 7.184586780377298, + "learning_rate": 1.5854782724317616e-05, + "loss": 0.694, + "step": 2965 + }, + { + "epoch": 1.6102062975027145, + "grad_norm": 6.9154974167720855, + "learning_rate": 1.5851931302451262e-05, + "loss": 0.8573, + "step": 2966 + }, + { + "epoch": 1.6107491856677525, + "grad_norm": 5.056944401416877, + "learning_rate": 1.5849079156808726e-05, + "loss": 0.4972, + "step": 2967 + }, + { + "epoch": 1.6112920738327905, + "grad_norm": 6.98635366168859, + "learning_rate": 1.584622628774275e-05, + "loss": 0.9002, + "step": 2968 + }, + { + "epoch": 1.6118349619978285, + "grad_norm": 9.22418255248469, + "learning_rate": 1.5843372695606196e-05, + "loss": 1.4143, + "step": 2969 + }, + { + "epoch": 1.6123778501628665, + "grad_norm": 9.49766654897752, + "learning_rate": 1.584051838075199e-05, + "loss": 1.0145, + "step": 2970 + }, + { + "epoch": 1.6129207383279045, + "grad_norm": 9.681135461557052, + "learning_rate": 1.5837663343533166e-05, + "loss": 0.8841, + "step": 2971 + }, + { + "epoch": 1.6134636264929425, + "grad_norm": 6.00661489729409, + "learning_rate": 1.583480758430283e-05, + "loss": 0.677, + "step": 2972 + }, + { + "epoch": 1.6140065146579805, + "grad_norm": 7.843885336692957, + "learning_rate": 1.5831951103414194e-05, + "loss": 0.4729, + "step": 2973 + }, + { + "epoch": 1.6145494028230185, + "grad_norm": 9.391971866087287, + "learning_rate": 1.5829093901220557e-05, + "loss": 1.0833, + "step": 2974 + }, + { + "epoch": 1.6150922909880565, + "grad_norm": 7.202123614778415, + "learning_rate": 1.582623597807529e-05, + "loss": 0.9755, + "step": 2975 + }, + { + "epoch": 1.6156351791530945, + "grad_norm": 8.050976157508988, + "learning_rate": 1.5823377334331875e-05, + "loss": 0.8981, + "step": 2976 + }, + { + "epoch": 1.6161780673181325, + "grad_norm": 6.409986462897389, + "learning_rate": 1.5820517970343867e-05, + "loss": 0.3994, + "step": 2977 + }, + { + "epoch": 1.6167209554831705, + "grad_norm": 6.001133896600598, + "learning_rate": 1.581765788646492e-05, + "loss": 0.4701, + "step": 2978 + }, + { + "epoch": 1.6172638436482085, + "grad_norm": 12.858193971885504, + "learning_rate": 1.581479708304878e-05, + "loss": 1.4327, + "step": 2979 + }, + { + "epoch": 1.6178067318132465, + "grad_norm": 9.559044098763156, + "learning_rate": 1.5811935560449262e-05, + "loss": 0.9058, + "step": 2980 + }, + { + "epoch": 1.6183496199782845, + "grad_norm": 8.041859101089647, + "learning_rate": 1.5809073319020293e-05, + "loss": 0.6251, + "step": 2981 + }, + { + "epoch": 1.6188925081433225, + "grad_norm": 7.413714357196275, + "learning_rate": 1.580621035911588e-05, + "loss": 0.7887, + "step": 2982 + }, + { + "epoch": 1.6194353963083605, + "grad_norm": 11.941805543277479, + "learning_rate": 1.5803346681090113e-05, + "loss": 1.2547, + "step": 2983 + }, + { + "epoch": 1.6199782844733985, + "grad_norm": 5.728003242952451, + "learning_rate": 1.580048228529718e-05, + "loss": 0.4941, + "step": 2984 + }, + { + "epoch": 1.6205211726384365, + "grad_norm": 8.67398306592389, + "learning_rate": 1.5797617172091354e-05, + "loss": 0.7758, + "step": 2985 + }, + { + "epoch": 1.6210640608034745, + "grad_norm": 6.198163559825324, + "learning_rate": 1.5794751341826996e-05, + "loss": 0.4601, + "step": 2986 + }, + { + "epoch": 1.6216069489685125, + "grad_norm": 8.257469675141675, + "learning_rate": 1.5791884794858557e-05, + "loss": 0.8679, + "step": 2987 + }, + { + "epoch": 1.6221498371335505, + "grad_norm": 12.953868836673164, + "learning_rate": 1.5789017531540575e-05, + "loss": 1.572, + "step": 2988 + }, + { + "epoch": 1.6226927252985885, + "grad_norm": 9.232767459675694, + "learning_rate": 1.5786149552227682e-05, + "loss": 0.7511, + "step": 2989 + }, + { + "epoch": 1.6232356134636265, + "grad_norm": 7.465499623540017, + "learning_rate": 1.5783280857274586e-05, + "loss": 0.5754, + "step": 2990 + }, + { + "epoch": 1.6237785016286646, + "grad_norm": 5.087952971939767, + "learning_rate": 1.5780411447036097e-05, + "loss": 0.532, + "step": 2991 + }, + { + "epoch": 1.6243213897937026, + "grad_norm": 11.064786965172743, + "learning_rate": 1.577754132186711e-05, + "loss": 0.7195, + "step": 2992 + }, + { + "epoch": 1.6248642779587406, + "grad_norm": 10.268274657097892, + "learning_rate": 1.57746704821226e-05, + "loss": 1.1258, + "step": 2993 + }, + { + "epoch": 1.6254071661237783, + "grad_norm": 7.650078703788778, + "learning_rate": 1.5771798928157645e-05, + "loss": 0.5949, + "step": 2994 + }, + { + "epoch": 1.6259500542888166, + "grad_norm": 9.330939436907839, + "learning_rate": 1.5768926660327396e-05, + "loss": 0.9243, + "step": 2995 + }, + { + "epoch": 1.6264929424538543, + "grad_norm": 11.566902464205237, + "learning_rate": 1.576605367898711e-05, + "loss": 1.5588, + "step": 2996 + }, + { + "epoch": 1.6270358306188926, + "grad_norm": 9.91827217636312, + "learning_rate": 1.576317998449211e-05, + "loss": 1.1418, + "step": 2997 + }, + { + "epoch": 1.6275787187839303, + "grad_norm": 5.842910545051031, + "learning_rate": 1.5760305577197824e-05, + "loss": 0.5064, + "step": 2998 + }, + { + "epoch": 1.6281216069489686, + "grad_norm": 7.20637691224012, + "learning_rate": 1.5757430457459765e-05, + "loss": 0.8129, + "step": 2999 + }, + { + "epoch": 1.6286644951140063, + "grad_norm": 7.5338669003432965, + "learning_rate": 1.5754554625633535e-05, + "loss": 0.7244, + "step": 3000 + }, + { + "epoch": 1.6292073832790446, + "grad_norm": 7.489311206471124, + "learning_rate": 1.5751678082074813e-05, + "loss": 1.1409, + "step": 3001 + }, + { + "epoch": 1.6297502714440824, + "grad_norm": 7.26299222123934, + "learning_rate": 1.574880082713938e-05, + "loss": 0.7061, + "step": 3002 + }, + { + "epoch": 1.6302931596091206, + "grad_norm": 6.292246912680897, + "learning_rate": 1.5745922861183095e-05, + "loss": 0.6383, + "step": 3003 + }, + { + "epoch": 1.6308360477741584, + "grad_norm": 7.483086031387996, + "learning_rate": 1.574304418456192e-05, + "loss": 0.8194, + "step": 3004 + }, + { + "epoch": 1.6313789359391966, + "grad_norm": 8.062451557552306, + "learning_rate": 1.5740164797631882e-05, + "loss": 0.9474, + "step": 3005 + }, + { + "epoch": 1.6319218241042344, + "grad_norm": 7.62904254588456, + "learning_rate": 1.5737284700749116e-05, + "loss": 1.2324, + "step": 3006 + }, + { + "epoch": 1.6324647122692726, + "grad_norm": 8.533361938490982, + "learning_rate": 1.573440389426983e-05, + "loss": 0.6182, + "step": 3007 + }, + { + "epoch": 1.6330076004343104, + "grad_norm": 7.0451736765954855, + "learning_rate": 1.5731522378550337e-05, + "loss": 0.7464, + "step": 3008 + }, + { + "epoch": 1.6335504885993486, + "grad_norm": 8.121336222871497, + "learning_rate": 1.572864015394702e-05, + "loss": 1.1087, + "step": 3009 + }, + { + "epoch": 1.6340933767643864, + "grad_norm": 7.2112540180194955, + "learning_rate": 1.5725757220816356e-05, + "loss": 0.7756, + "step": 3010 + }, + { + "epoch": 1.6346362649294246, + "grad_norm": 8.092755806422884, + "learning_rate": 1.5722873579514915e-05, + "loss": 1.1184, + "step": 3011 + }, + { + "epoch": 1.6351791530944624, + "grad_norm": 7.485157971844339, + "learning_rate": 1.5719989230399347e-05, + "loss": 0.6839, + "step": 3012 + }, + { + "epoch": 1.6357220412595006, + "grad_norm": 8.516980380341119, + "learning_rate": 1.5717104173826397e-05, + "loss": 0.948, + "step": 3013 + }, + { + "epoch": 1.6362649294245384, + "grad_norm": 7.602547351124631, + "learning_rate": 1.571421841015289e-05, + "loss": 0.8479, + "step": 3014 + }, + { + "epoch": 1.6368078175895766, + "grad_norm": 7.0470960880026245, + "learning_rate": 1.5711331939735744e-05, + "loss": 0.9014, + "step": 3015 + }, + { + "epoch": 1.6373507057546144, + "grad_norm": 6.760810209806174, + "learning_rate": 1.570844476293196e-05, + "loss": 0.9316, + "step": 3016 + }, + { + "epoch": 1.6378935939196526, + "grad_norm": 8.02053578699558, + "learning_rate": 1.570555688009863e-05, + "loss": 0.7837, + "step": 3017 + }, + { + "epoch": 1.6384364820846904, + "grad_norm": 4.624578610614632, + "learning_rate": 1.5702668291592936e-05, + "loss": 0.4706, + "step": 3018 + }, + { + "epoch": 1.6389793702497286, + "grad_norm": 8.011712265192347, + "learning_rate": 1.569977899777213e-05, + "loss": 1.0338, + "step": 3019 + }, + { + "epoch": 1.6395222584147664, + "grad_norm": 7.773921575637797, + "learning_rate": 1.569688899899358e-05, + "loss": 0.7512, + "step": 3020 + }, + { + "epoch": 1.6400651465798046, + "grad_norm": 12.19433508686472, + "learning_rate": 1.569399829561472e-05, + "loss": 1.3502, + "step": 3021 + }, + { + "epoch": 1.6406080347448424, + "grad_norm": 11.319107865931453, + "learning_rate": 1.569110688799307e-05, + "loss": 0.9996, + "step": 3022 + }, + { + "epoch": 1.6411509229098806, + "grad_norm": 7.1911299711947585, + "learning_rate": 1.5688214776486255e-05, + "loss": 0.9037, + "step": 3023 + }, + { + "epoch": 1.6416938110749184, + "grad_norm": 8.326843836043436, + "learning_rate": 1.5685321961451968e-05, + "loss": 1.1179, + "step": 3024 + }, + { + "epoch": 1.6422366992399566, + "grad_norm": 8.257534801333424, + "learning_rate": 1.5682428443248002e-05, + "loss": 0.7197, + "step": 3025 + }, + { + "epoch": 1.6427795874049944, + "grad_norm": 6.669062512754087, + "learning_rate": 1.567953422223223e-05, + "loss": 0.6292, + "step": 3026 + }, + { + "epoch": 1.6433224755700326, + "grad_norm": 7.99477616152165, + "learning_rate": 1.567663929876261e-05, + "loss": 0.926, + "step": 3027 + }, + { + "epoch": 1.6438653637350704, + "grad_norm": 5.3717184947082615, + "learning_rate": 1.56737436731972e-05, + "loss": 0.5378, + "step": 3028 + }, + { + "epoch": 1.6444082519001086, + "grad_norm": 6.522826762176589, + "learning_rate": 1.5670847345894125e-05, + "loss": 0.7393, + "step": 3029 + }, + { + "epoch": 1.6449511400651464, + "grad_norm": 9.503371738808127, + "learning_rate": 1.5667950317211612e-05, + "loss": 1.2727, + "step": 3030 + }, + { + "epoch": 1.6454940282301846, + "grad_norm": 8.208077816222845, + "learning_rate": 1.5665052587507974e-05, + "loss": 0.7032, + "step": 3031 + }, + { + "epoch": 1.6460369163952224, + "grad_norm": 9.101101018529372, + "learning_rate": 1.56621541571416e-05, + "loss": 0.6983, + "step": 3032 + }, + { + "epoch": 1.6465798045602607, + "grad_norm": 7.254188682380924, + "learning_rate": 1.565925502647098e-05, + "loss": 0.7196, + "step": 3033 + }, + { + "epoch": 1.6471226927252984, + "grad_norm": 7.837018641144827, + "learning_rate": 1.5656355195854676e-05, + "loss": 1.2035, + "step": 3034 + }, + { + "epoch": 1.6476655808903367, + "grad_norm": 8.949191823273495, + "learning_rate": 1.5653454665651344e-05, + "loss": 1.0169, + "step": 3035 + }, + { + "epoch": 1.6482084690553744, + "grad_norm": 8.166317934132502, + "learning_rate": 1.5650553436219732e-05, + "loss": 0.672, + "step": 3036 + }, + { + "epoch": 1.6487513572204127, + "grad_norm": 6.443983188038021, + "learning_rate": 1.564765150791866e-05, + "loss": 0.5552, + "step": 3037 + }, + { + "epoch": 1.6492942453854504, + "grad_norm": 10.403060348162544, + "learning_rate": 1.5644748881107057e-05, + "loss": 1.0709, + "step": 3038 + }, + { + "epoch": 1.6498371335504887, + "grad_norm": 8.319421686790951, + "learning_rate": 1.564184555614391e-05, + "loss": 0.7923, + "step": 3039 + }, + { + "epoch": 1.6503800217155264, + "grad_norm": 9.65474819234395, + "learning_rate": 1.5638941533388318e-05, + "loss": 0.8519, + "step": 3040 + }, + { + "epoch": 1.6509229098805647, + "grad_norm": 8.635959327913778, + "learning_rate": 1.5636036813199445e-05, + "loss": 0.8447, + "step": 3041 + }, + { + "epoch": 1.6514657980456025, + "grad_norm": 7.674155625456285, + "learning_rate": 1.563313139593656e-05, + "loss": 0.7205, + "step": 3042 + }, + { + "epoch": 1.6520086862106407, + "grad_norm": 7.249478120616782, + "learning_rate": 1.5630225281959003e-05, + "loss": 0.6632, + "step": 3043 + }, + { + "epoch": 1.6525515743756785, + "grad_norm": 8.90384172612564, + "learning_rate": 1.5627318471626208e-05, + "loss": 0.8243, + "step": 3044 + }, + { + "epoch": 1.6530944625407167, + "grad_norm": 8.274462585020828, + "learning_rate": 1.5624410965297703e-05, + "loss": 0.7381, + "step": 3045 + }, + { + "epoch": 1.6536373507057545, + "grad_norm": 7.266215459792981, + "learning_rate": 1.562150276333308e-05, + "loss": 0.4989, + "step": 3046 + }, + { + "epoch": 1.6541802388707927, + "grad_norm": 4.816150056433193, + "learning_rate": 1.5618593866092036e-05, + "loss": 0.5382, + "step": 3047 + }, + { + "epoch": 1.6547231270358305, + "grad_norm": 6.835417235425346, + "learning_rate": 1.561568427393435e-05, + "loss": 1.044, + "step": 3048 + }, + { + "epoch": 1.6552660152008687, + "grad_norm": 7.2539344971377435, + "learning_rate": 1.5612773987219885e-05, + "loss": 0.5901, + "step": 3049 + }, + { + "epoch": 1.6558089033659065, + "grad_norm": 11.003646663954257, + "learning_rate": 1.5609863006308586e-05, + "loss": 0.9407, + "step": 3050 + }, + { + "epoch": 1.6563517915309447, + "grad_norm": 8.268114443948154, + "learning_rate": 1.560695133156049e-05, + "loss": 1.0164, + "step": 3051 + }, + { + "epoch": 1.6568946796959825, + "grad_norm": 7.824779774533612, + "learning_rate": 1.5604038963335716e-05, + "loss": 0.7111, + "step": 3052 + }, + { + "epoch": 1.6574375678610207, + "grad_norm": 8.542948361290183, + "learning_rate": 1.560112590199447e-05, + "loss": 0.6501, + "step": 3053 + }, + { + "epoch": 1.6579804560260585, + "grad_norm": 7.85824626614946, + "learning_rate": 1.5598212147897047e-05, + "loss": 0.752, + "step": 3054 + }, + { + "epoch": 1.6585233441910967, + "grad_norm": 6.588621135611242, + "learning_rate": 1.559529770140382e-05, + "loss": 0.507, + "step": 3055 + }, + { + "epoch": 1.6590662323561345, + "grad_norm": 6.418764525989286, + "learning_rate": 1.559238256287526e-05, + "loss": 0.5947, + "step": 3056 + }, + { + "epoch": 1.6596091205211727, + "grad_norm": 7.681071515192806, + "learning_rate": 1.5589466732671913e-05, + "loss": 0.8854, + "step": 3057 + }, + { + "epoch": 1.6601520086862105, + "grad_norm": 5.918230957225734, + "learning_rate": 1.558655021115441e-05, + "loss": 0.5546, + "step": 3058 + }, + { + "epoch": 1.6606948968512487, + "grad_norm": 10.669803193569846, + "learning_rate": 1.5583632998683475e-05, + "loss": 1.1548, + "step": 3059 + }, + { + "epoch": 1.6612377850162865, + "grad_norm": 9.271579004183549, + "learning_rate": 1.558071509561991e-05, + "loss": 0.8049, + "step": 3060 + }, + { + "epoch": 1.6617806731813247, + "grad_norm": 8.337085337830533, + "learning_rate": 1.557779650232461e-05, + "loss": 0.954, + "step": 3061 + }, + { + "epoch": 1.6623235613463625, + "grad_norm": 11.145075141347627, + "learning_rate": 1.5574877219158543e-05, + "loss": 1.1592, + "step": 3062 + }, + { + "epoch": 1.6628664495114007, + "grad_norm": 13.953049521423079, + "learning_rate": 1.557195724648278e-05, + "loss": 1.5699, + "step": 3063 + }, + { + "epoch": 1.6634093376764385, + "grad_norm": 8.654584332416588, + "learning_rate": 1.5569036584658466e-05, + "loss": 1.0185, + "step": 3064 + }, + { + "epoch": 1.6639522258414767, + "grad_norm": 7.520260829640502, + "learning_rate": 1.556611523404683e-05, + "loss": 0.7898, + "step": 3065 + }, + { + "epoch": 1.6644951140065145, + "grad_norm": 6.7373805901465875, + "learning_rate": 1.5563193195009188e-05, + "loss": 0.8392, + "step": 3066 + }, + { + "epoch": 1.6650380021715527, + "grad_norm": 6.732598633253056, + "learning_rate": 1.556027046790695e-05, + "loss": 0.6672, + "step": 3067 + }, + { + "epoch": 1.6655808903365905, + "grad_norm": 9.217725100287737, + "learning_rate": 1.55573470531016e-05, + "loss": 0.9874, + "step": 3068 + }, + { + "epoch": 1.6661237785016287, + "grad_norm": 6.963043776606882, + "learning_rate": 1.5554422950954706e-05, + "loss": 0.812, + "step": 3069 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 6.449074540999346, + "learning_rate": 1.5551498161827928e-05, + "loss": 0.5035, + "step": 3070 + }, + { + "epoch": 1.6672095548317047, + "grad_norm": 7.793693453975787, + "learning_rate": 1.554857268608301e-05, + "loss": 0.7974, + "step": 3071 + }, + { + "epoch": 1.6677524429967425, + "grad_norm": 10.124707886106192, + "learning_rate": 1.554564652408178e-05, + "loss": 1.4946, + "step": 3072 + }, + { + "epoch": 1.6682953311617807, + "grad_norm": 5.481775138355068, + "learning_rate": 1.5542719676186153e-05, + "loss": 0.502, + "step": 3073 + }, + { + "epoch": 1.6688382193268185, + "grad_norm": 6.055974601989459, + "learning_rate": 1.5539792142758114e-05, + "loss": 0.7463, + "step": 3074 + }, + { + "epoch": 1.6693811074918568, + "grad_norm": 7.452121005302326, + "learning_rate": 1.5536863924159762e-05, + "loss": 0.7282, + "step": 3075 + }, + { + "epoch": 1.6699239956568945, + "grad_norm": 8.991453230181806, + "learning_rate": 1.5533935020753252e-05, + "loss": 0.9327, + "step": 3076 + }, + { + "epoch": 1.6704668838219328, + "grad_norm": 8.138196256677595, + "learning_rate": 1.5531005432900838e-05, + "loss": 0.658, + "step": 3077 + }, + { + "epoch": 1.6710097719869705, + "grad_norm": 6.674313540160648, + "learning_rate": 1.552807516096486e-05, + "loss": 0.569, + "step": 3078 + }, + { + "epoch": 1.6715526601520088, + "grad_norm": 7.2303438904479105, + "learning_rate": 1.5525144205307734e-05, + "loss": 0.5186, + "step": 3079 + }, + { + "epoch": 1.6720955483170465, + "grad_norm": 6.872945037718312, + "learning_rate": 1.5522212566291966e-05, + "loss": 0.7282, + "step": 3080 + }, + { + "epoch": 1.6726384364820848, + "grad_norm": 7.913543579080126, + "learning_rate": 1.5519280244280146e-05, + "loss": 0.7015, + "step": 3081 + }, + { + "epoch": 1.6731813246471225, + "grad_norm": 6.144062932600708, + "learning_rate": 1.551634723963495e-05, + "loss": 0.7829, + "step": 3082 + }, + { + "epoch": 1.6737242128121608, + "grad_norm": 7.111349951384809, + "learning_rate": 1.5513413552719137e-05, + "loss": 0.6904, + "step": 3083 + }, + { + "epoch": 1.6742671009771986, + "grad_norm": 10.453360973997349, + "learning_rate": 1.5510479183895545e-05, + "loss": 1.4584, + "step": 3084 + }, + { + "epoch": 1.6748099891422368, + "grad_norm": 13.976555611119428, + "learning_rate": 1.5507544133527108e-05, + "loss": 1.542, + "step": 3085 + }, + { + "epoch": 1.6753528773072746, + "grad_norm": 5.894489690173581, + "learning_rate": 1.5504608401976835e-05, + "loss": 0.6792, + "step": 3086 + }, + { + "epoch": 1.6758957654723128, + "grad_norm": 8.937162315762803, + "learning_rate": 1.550167198960782e-05, + "loss": 0.8243, + "step": 3087 + }, + { + "epoch": 1.6764386536373506, + "grad_norm": 9.787723535687237, + "learning_rate": 1.549873489678325e-05, + "loss": 1.0071, + "step": 3088 + }, + { + "epoch": 1.6769815418023888, + "grad_norm": 6.0371827348978755, + "learning_rate": 1.5495797123866378e-05, + "loss": 0.6407, + "step": 3089 + }, + { + "epoch": 1.6775244299674266, + "grad_norm": 5.997673025227573, + "learning_rate": 1.5492858671220563e-05, + "loss": 0.7129, + "step": 3090 + }, + { + "epoch": 1.6780673181324648, + "grad_norm": 11.08813931574187, + "learning_rate": 1.5489919539209232e-05, + "loss": 0.8435, + "step": 3091 + }, + { + "epoch": 1.6786102062975026, + "grad_norm": 8.730951161091724, + "learning_rate": 1.54869797281959e-05, + "loss": 0.9159, + "step": 3092 + }, + { + "epoch": 1.6791530944625408, + "grad_norm": 7.293119927853279, + "learning_rate": 1.5484039238544173e-05, + "loss": 1.1001, + "step": 3093 + }, + { + "epoch": 1.6796959826275786, + "grad_norm": 7.091810627268002, + "learning_rate": 1.5481098070617734e-05, + "loss": 0.747, + "step": 3094 + }, + { + "epoch": 1.6802388707926168, + "grad_norm": 9.079124871383598, + "learning_rate": 1.547815622478035e-05, + "loss": 1.2292, + "step": 3095 + }, + { + "epoch": 1.6807817589576546, + "grad_norm": 7.238161650182196, + "learning_rate": 1.5475213701395867e-05, + "loss": 0.6657, + "step": 3096 + }, + { + "epoch": 1.6813246471226928, + "grad_norm": 9.385885763398004, + "learning_rate": 1.5472270500828236e-05, + "loss": 1.305, + "step": 3097 + }, + { + "epoch": 1.6818675352877306, + "grad_norm": 8.98902696621929, + "learning_rate": 1.5469326623441463e-05, + "loss": 0.8641, + "step": 3098 + }, + { + "epoch": 1.6824104234527688, + "grad_norm": 6.519773416264212, + "learning_rate": 1.5466382069599656e-05, + "loss": 0.4567, + "step": 3099 + }, + { + "epoch": 1.6829533116178066, + "grad_norm": 6.775917495491388, + "learning_rate": 1.5463436839667007e-05, + "loss": 0.6874, + "step": 3100 + }, + { + "epoch": 1.6834961997828448, + "grad_norm": 7.82569199258335, + "learning_rate": 1.5460490934007776e-05, + "loss": 0.819, + "step": 3101 + }, + { + "epoch": 1.6840390879478826, + "grad_norm": 6.885216490352139, + "learning_rate": 1.5457544352986326e-05, + "loss": 0.7891, + "step": 3102 + }, + { + "epoch": 1.6845819761129208, + "grad_norm": 6.733247427994312, + "learning_rate": 1.5454597096967093e-05, + "loss": 0.6664, + "step": 3103 + }, + { + "epoch": 1.6851248642779586, + "grad_norm": 6.96769902934235, + "learning_rate": 1.5451649166314598e-05, + "loss": 0.6578, + "step": 3104 + }, + { + "epoch": 1.6856677524429968, + "grad_norm": 6.941713803652074, + "learning_rate": 1.5448700561393444e-05, + "loss": 0.5855, + "step": 3105 + }, + { + "epoch": 1.6862106406080346, + "grad_norm": 8.203921934361416, + "learning_rate": 1.5445751282568324e-05, + "loss": 0.9843, + "step": 3106 + }, + { + "epoch": 1.6867535287730728, + "grad_norm": 8.822160140984163, + "learning_rate": 1.5442801330204004e-05, + "loss": 1.1003, + "step": 3107 + }, + { + "epoch": 1.6872964169381106, + "grad_norm": 8.330851912949297, + "learning_rate": 1.5439850704665338e-05, + "loss": 1.0992, + "step": 3108 + }, + { + "epoch": 1.6878393051031488, + "grad_norm": 6.552171568470756, + "learning_rate": 1.543689940631727e-05, + "loss": 0.5434, + "step": 3109 + }, + { + "epoch": 1.6883821932681866, + "grad_norm": 5.772049074992556, + "learning_rate": 1.5433947435524822e-05, + "loss": 0.7515, + "step": 3110 + }, + { + "epoch": 1.6889250814332248, + "grad_norm": 9.575440058135785, + "learning_rate": 1.543099479265309e-05, + "loss": 0.9691, + "step": 3111 + }, + { + "epoch": 1.6894679695982626, + "grad_norm": 6.008162314259628, + "learning_rate": 1.5428041478067263e-05, + "loss": 0.6137, + "step": 3112 + }, + { + "epoch": 1.6900108577633008, + "grad_norm": 8.993171311918207, + "learning_rate": 1.542508749213262e-05, + "loss": 0.7608, + "step": 3113 + }, + { + "epoch": 1.6905537459283386, + "grad_norm": 8.042583538721042, + "learning_rate": 1.542213283521451e-05, + "loss": 0.8188, + "step": 3114 + }, + { + "epoch": 1.6910966340933768, + "grad_norm": 9.290959640283983, + "learning_rate": 1.541917750767837e-05, + "loss": 1.1737, + "step": 3115 + }, + { + "epoch": 1.6916395222584146, + "grad_norm": 6.3883435583657535, + "learning_rate": 1.5416221509889718e-05, + "loss": 0.4376, + "step": 3116 + }, + { + "epoch": 1.6921824104234529, + "grad_norm": 8.943225081508276, + "learning_rate": 1.541326484221416e-05, + "loss": 0.9983, + "step": 3117 + }, + { + "epoch": 1.6927252985884906, + "grad_norm": 7.583355336491315, + "learning_rate": 1.541030750501737e-05, + "loss": 0.6078, + "step": 3118 + }, + { + "epoch": 1.6932681867535289, + "grad_norm": 7.318011543277784, + "learning_rate": 1.5407349498665133e-05, + "loss": 0.745, + "step": 3119 + }, + { + "epoch": 1.6938110749185666, + "grad_norm": 6.212465291912028, + "learning_rate": 1.5404390823523287e-05, + "loss": 0.4739, + "step": 3120 + }, + { + "epoch": 1.6943539630836049, + "grad_norm": 8.223585888775384, + "learning_rate": 1.5401431479957775e-05, + "loss": 1.1819, + "step": 3121 + }, + { + "epoch": 1.6948968512486426, + "grad_norm": 7.957232407413796, + "learning_rate": 1.5398471468334605e-05, + "loss": 0.7071, + "step": 3122 + }, + { + "epoch": 1.6954397394136809, + "grad_norm": 5.52597775233848, + "learning_rate": 1.5395510789019884e-05, + "loss": 0.6171, + "step": 3123 + }, + { + "epoch": 1.6959826275787186, + "grad_norm": 8.015216509333372, + "learning_rate": 1.5392549442379785e-05, + "loss": 0.9051, + "step": 3124 + }, + { + "epoch": 1.6965255157437569, + "grad_norm": 8.330470177699917, + "learning_rate": 1.538958742878058e-05, + "loss": 0.5037, + "step": 3125 + }, + { + "epoch": 1.6970684039087947, + "grad_norm": 6.1063862074204245, + "learning_rate": 1.538662474858861e-05, + "loss": 0.6509, + "step": 3126 + }, + { + "epoch": 1.6976112920738329, + "grad_norm": 8.11378323370098, + "learning_rate": 1.5383661402170308e-05, + "loss": 1.0356, + "step": 3127 + }, + { + "epoch": 1.6981541802388707, + "grad_norm": 5.995828293111682, + "learning_rate": 1.5380697389892185e-05, + "loss": 0.5297, + "step": 3128 + }, + { + "epoch": 1.6986970684039089, + "grad_norm": 6.962108385144619, + "learning_rate": 1.537773271212083e-05, + "loss": 0.8711, + "step": 3129 + }, + { + "epoch": 1.6992399565689467, + "grad_norm": 8.69611973474616, + "learning_rate": 1.5374767369222922e-05, + "loss": 0.911, + "step": 3130 + }, + { + "epoch": 1.6997828447339849, + "grad_norm": 11.047832664292853, + "learning_rate": 1.5371801361565223e-05, + "loss": 0.8499, + "step": 3131 + }, + { + "epoch": 1.7003257328990227, + "grad_norm": 10.490928296052365, + "learning_rate": 1.5368834689514568e-05, + "loss": 0.7617, + "step": 3132 + }, + { + "epoch": 1.700868621064061, + "grad_norm": 8.723324926351989, + "learning_rate": 1.536586735343788e-05, + "loss": 0.7158, + "step": 3133 + }, + { + "epoch": 1.7014115092290987, + "grad_norm": 7.780646168464086, + "learning_rate": 1.536289935370217e-05, + "loss": 0.6452, + "step": 3134 + }, + { + "epoch": 1.701954397394137, + "grad_norm": 8.037748179528911, + "learning_rate": 1.5359930690674518e-05, + "loss": 0.8368, + "step": 3135 + }, + { + "epoch": 1.7024972855591747, + "grad_norm": 7.539822669872962, + "learning_rate": 1.5356961364722096e-05, + "loss": 0.7224, + "step": 3136 + }, + { + "epoch": 1.703040173724213, + "grad_norm": 10.929660828744602, + "learning_rate": 1.5353991376212155e-05, + "loss": 1.2459, + "step": 3137 + }, + { + "epoch": 1.7035830618892507, + "grad_norm": 11.118929306326729, + "learning_rate": 1.5351020725512028e-05, + "loss": 1.2769, + "step": 3138 + }, + { + "epoch": 1.704125950054289, + "grad_norm": 8.820097629927263, + "learning_rate": 1.534804941298913e-05, + "loss": 1.0722, + "step": 3139 + }, + { + "epoch": 1.7046688382193267, + "grad_norm": 8.79991079654506, + "learning_rate": 1.5345077439010956e-05, + "loss": 0.7596, + "step": 3140 + }, + { + "epoch": 1.705211726384365, + "grad_norm": 7.5009896900263096, + "learning_rate": 1.5342104803945087e-05, + "loss": 0.6457, + "step": 3141 + }, + { + "epoch": 1.7057546145494027, + "grad_norm": 9.190121352339093, + "learning_rate": 1.533913150815918e-05, + "loss": 0.7152, + "step": 3142 + }, + { + "epoch": 1.706297502714441, + "grad_norm": 7.209026897095485, + "learning_rate": 1.5336157552020977e-05, + "loss": 0.534, + "step": 3143 + }, + { + "epoch": 1.7068403908794787, + "grad_norm": 8.752247270964865, + "learning_rate": 1.5333182935898306e-05, + "loss": 1.0953, + "step": 3144 + }, + { + "epoch": 1.707383279044517, + "grad_norm": 8.714563086288697, + "learning_rate": 1.5330207660159068e-05, + "loss": 0.8016, + "step": 3145 + }, + { + "epoch": 1.7079261672095547, + "grad_norm": 9.540320185420228, + "learning_rate": 1.5327231725171255e-05, + "loss": 1.2679, + "step": 3146 + }, + { + "epoch": 1.708469055374593, + "grad_norm": 8.091263210809633, + "learning_rate": 1.532425513130293e-05, + "loss": 0.7521, + "step": 3147 + }, + { + "epoch": 1.7090119435396307, + "grad_norm": 10.062104841108379, + "learning_rate": 1.5321277878922246e-05, + "loss": 1.0696, + "step": 3148 + }, + { + "epoch": 1.709554831704669, + "grad_norm": 8.184716628635337, + "learning_rate": 1.531829996839743e-05, + "loss": 0.643, + "step": 3149 + }, + { + "epoch": 1.7100977198697067, + "grad_norm": 11.130045309860368, + "learning_rate": 1.53153214000968e-05, + "loss": 1.2565, + "step": 3150 + }, + { + "epoch": 1.710640608034745, + "grad_norm": 8.856470164658607, + "learning_rate": 1.5312342174388746e-05, + "loss": 0.8376, + "step": 3151 + }, + { + "epoch": 1.7111834961997827, + "grad_norm": 11.958863336707164, + "learning_rate": 1.5309362291641747e-05, + "loss": 0.978, + "step": 3152 + }, + { + "epoch": 1.711726384364821, + "grad_norm": 10.369392130776264, + "learning_rate": 1.5306381752224357e-05, + "loss": 1.1822, + "step": 3153 + }, + { + "epoch": 1.7122692725298587, + "grad_norm": 7.346196971786586, + "learning_rate": 1.5303400556505213e-05, + "loss": 0.9284, + "step": 3154 + }, + { + "epoch": 1.712812160694897, + "grad_norm": 6.652184209108613, + "learning_rate": 1.5300418704853042e-05, + "loss": 0.5787, + "step": 3155 + }, + { + "epoch": 1.7133550488599347, + "grad_norm": 9.745597378440513, + "learning_rate": 1.5297436197636634e-05, + "loss": 1.0194, + "step": 3156 + }, + { + "epoch": 1.713897937024973, + "grad_norm": 6.17200933356452, + "learning_rate": 1.5294453035224874e-05, + "loss": 0.6315, + "step": 3157 + }, + { + "epoch": 1.7144408251900107, + "grad_norm": 8.373044157038668, + "learning_rate": 1.5291469217986724e-05, + "loss": 0.9127, + "step": 3158 + }, + { + "epoch": 1.714983713355049, + "grad_norm": 9.811623309823384, + "learning_rate": 1.5288484746291227e-05, + "loss": 0.7791, + "step": 3159 + }, + { + "epoch": 1.7155266015200867, + "grad_norm": 9.264521529810194, + "learning_rate": 1.5285499620507513e-05, + "loss": 0.6764, + "step": 3160 + }, + { + "epoch": 1.716069489685125, + "grad_norm": 8.688562646594697, + "learning_rate": 1.5282513841004777e-05, + "loss": 1.0202, + "step": 3161 + }, + { + "epoch": 1.7166123778501627, + "grad_norm": 8.522452395448054, + "learning_rate": 1.527952740815231e-05, + "loss": 1.0043, + "step": 3162 + }, + { + "epoch": 1.717155266015201, + "grad_norm": 6.330008909603884, + "learning_rate": 1.527654032231948e-05, + "loss": 0.6651, + "step": 3163 + }, + { + "epoch": 1.7176981541802387, + "grad_norm": 9.314495623946685, + "learning_rate": 1.5273552583875736e-05, + "loss": 0.8312, + "step": 3164 + }, + { + "epoch": 1.718241042345277, + "grad_norm": 9.157480690732486, + "learning_rate": 1.52705641931906e-05, + "loss": 1.0017, + "step": 3165 + }, + { + "epoch": 1.7187839305103148, + "grad_norm": 8.300251884517943, + "learning_rate": 1.5267575150633687e-05, + "loss": 0.9929, + "step": 3166 + }, + { + "epoch": 1.719326818675353, + "grad_norm": 8.328644452020212, + "learning_rate": 1.5264585456574684e-05, + "loss": 0.8105, + "step": 3167 + }, + { + "epoch": 1.7198697068403908, + "grad_norm": 9.499775808882656, + "learning_rate": 1.526159511138336e-05, + "loss": 0.8593, + "step": 3168 + }, + { + "epoch": 1.720412595005429, + "grad_norm": 7.684886832360556, + "learning_rate": 1.5258604115429567e-05, + "loss": 0.5919, + "step": 3169 + }, + { + "epoch": 1.7209554831704668, + "grad_norm": 8.666869336691208, + "learning_rate": 1.5255612469083239e-05, + "loss": 0.6537, + "step": 3170 + }, + { + "epoch": 1.721498371335505, + "grad_norm": 11.871617471990923, + "learning_rate": 1.5252620172714378e-05, + "loss": 1.3826, + "step": 3171 + }, + { + "epoch": 1.7220412595005428, + "grad_norm": 6.088995555337644, + "learning_rate": 1.5249627226693089e-05, + "loss": 0.5085, + "step": 3172 + }, + { + "epoch": 1.722584147665581, + "grad_norm": 9.255760547122362, + "learning_rate": 1.5246633631389536e-05, + "loss": 0.8823, + "step": 3173 + }, + { + "epoch": 1.7231270358306188, + "grad_norm": 9.91965748862537, + "learning_rate": 1.5243639387173974e-05, + "loss": 0.8752, + "step": 3174 + }, + { + "epoch": 1.723669923995657, + "grad_norm": 6.397561552165108, + "learning_rate": 1.5240644494416734e-05, + "loss": 0.6673, + "step": 3175 + }, + { + "epoch": 1.7242128121606948, + "grad_norm": 7.066817871358342, + "learning_rate": 1.523764895348823e-05, + "loss": 1.041, + "step": 3176 + }, + { + "epoch": 1.724755700325733, + "grad_norm": 8.282830238008364, + "learning_rate": 1.5234652764758959e-05, + "loss": 0.9521, + "step": 3177 + }, + { + "epoch": 1.7252985884907708, + "grad_norm": 7.611547937947443, + "learning_rate": 1.523165592859949e-05, + "loss": 0.7182, + "step": 3178 + }, + { + "epoch": 1.725841476655809, + "grad_norm": 8.068698205768472, + "learning_rate": 1.5228658445380475e-05, + "loss": 0.9078, + "step": 3179 + }, + { + "epoch": 1.7263843648208468, + "grad_norm": 8.731772184014808, + "learning_rate": 1.5225660315472652e-05, + "loss": 1.2148, + "step": 3180 + }, + { + "epoch": 1.726927252985885, + "grad_norm": 7.785982141134894, + "learning_rate": 1.5222661539246832e-05, + "loss": 0.8653, + "step": 3181 + }, + { + "epoch": 1.7274701411509228, + "grad_norm": 7.6267653051373285, + "learning_rate": 1.521966211707391e-05, + "loss": 0.7454, + "step": 3182 + }, + { + "epoch": 1.728013029315961, + "grad_norm": 7.440711922406893, + "learning_rate": 1.521666204932486e-05, + "loss": 0.6576, + "step": 3183 + }, + { + "epoch": 1.7285559174809988, + "grad_norm": 7.69035147644333, + "learning_rate": 1.521366133637073e-05, + "loss": 0.8269, + "step": 3184 + }, + { + "epoch": 1.729098805646037, + "grad_norm": 8.387844627258719, + "learning_rate": 1.5210659978582662e-05, + "loss": 1.1049, + "step": 3185 + }, + { + "epoch": 1.7296416938110748, + "grad_norm": 9.08085381726355, + "learning_rate": 1.5207657976331862e-05, + "loss": 0.97, + "step": 3186 + }, + { + "epoch": 1.730184581976113, + "grad_norm": 7.621613071814458, + "learning_rate": 1.5204655329989617e-05, + "loss": 0.7113, + "step": 3187 + }, + { + "epoch": 1.7307274701411508, + "grad_norm": 6.743838306403499, + "learning_rate": 1.5201652039927313e-05, + "loss": 0.7307, + "step": 3188 + }, + { + "epoch": 1.731270358306189, + "grad_norm": 9.602658786805504, + "learning_rate": 1.5198648106516392e-05, + "loss": 1.0515, + "step": 3189 + }, + { + "epoch": 1.7318132464712268, + "grad_norm": 8.153765987898234, + "learning_rate": 1.5195643530128387e-05, + "loss": 1.0879, + "step": 3190 + }, + { + "epoch": 1.732356134636265, + "grad_norm": 9.745244927474419, + "learning_rate": 1.519263831113491e-05, + "loss": 0.9542, + "step": 3191 + }, + { + "epoch": 1.7328990228013028, + "grad_norm": 8.908425729214914, + "learning_rate": 1.5189632449907654e-05, + "loss": 0.8936, + "step": 3192 + }, + { + "epoch": 1.733441910966341, + "grad_norm": 7.409065356665995, + "learning_rate": 1.5186625946818382e-05, + "loss": 0.7645, + "step": 3193 + }, + { + "epoch": 1.7339847991313788, + "grad_norm": 7.958784484679066, + "learning_rate": 1.5183618802238949e-05, + "loss": 0.8512, + "step": 3194 + }, + { + "epoch": 1.734527687296417, + "grad_norm": 5.808082449761269, + "learning_rate": 1.5180611016541278e-05, + "loss": 0.7241, + "step": 3195 + }, + { + "epoch": 1.7350705754614548, + "grad_norm": 8.88976614296984, + "learning_rate": 1.5177602590097382e-05, + "loss": 1.0506, + "step": 3196 + }, + { + "epoch": 1.735613463626493, + "grad_norm": 8.197925091983134, + "learning_rate": 1.5174593523279346e-05, + "loss": 0.7959, + "step": 3197 + }, + { + "epoch": 1.7361563517915308, + "grad_norm": 10.2752921216035, + "learning_rate": 1.5171583816459334e-05, + "loss": 1.2859, + "step": 3198 + }, + { + "epoch": 1.736699239956569, + "grad_norm": 9.43498245507836, + "learning_rate": 1.5168573470009596e-05, + "loss": 1.0459, + "step": 3199 + }, + { + "epoch": 1.7372421281216068, + "grad_norm": 9.033995794822129, + "learning_rate": 1.516556248430245e-05, + "loss": 1.0003, + "step": 3200 + }, + { + "epoch": 1.737785016286645, + "grad_norm": 9.577836032406582, + "learning_rate": 1.5162550859710306e-05, + "loss": 1.2394, + "step": 3201 + }, + { + "epoch": 1.7383279044516828, + "grad_norm": 5.2063230201693775, + "learning_rate": 1.5159538596605642e-05, + "loss": 0.3625, + "step": 3202 + }, + { + "epoch": 1.738870792616721, + "grad_norm": 7.741298457702819, + "learning_rate": 1.5156525695361022e-05, + "loss": 0.6744, + "step": 3203 + }, + { + "epoch": 1.7394136807817588, + "grad_norm": 8.718382461609222, + "learning_rate": 1.5153512156349086e-05, + "loss": 1.0423, + "step": 3204 + }, + { + "epoch": 1.739956568946797, + "grad_norm": 6.706656901306365, + "learning_rate": 1.5150497979942556e-05, + "loss": 0.5121, + "step": 3205 + }, + { + "epoch": 1.7404994571118348, + "grad_norm": 9.934397785927217, + "learning_rate": 1.5147483166514225e-05, + "loss": 1.5269, + "step": 3206 + }, + { + "epoch": 1.741042345276873, + "grad_norm": 8.069224944733923, + "learning_rate": 1.514446771643697e-05, + "loss": 0.7801, + "step": 3207 + }, + { + "epoch": 1.7415852334419109, + "grad_norm": 9.193161021474104, + "learning_rate": 1.514145163008375e-05, + "loss": 1.2144, + "step": 3208 + }, + { + "epoch": 1.742128121606949, + "grad_norm": 6.5647281089901, + "learning_rate": 1.5138434907827599e-05, + "loss": 0.6692, + "step": 3209 + }, + { + "epoch": 1.7426710097719869, + "grad_norm": 9.387942500806554, + "learning_rate": 1.513541755004163e-05, + "loss": 1.1655, + "step": 3210 + }, + { + "epoch": 1.743213897937025, + "grad_norm": 6.936398185140981, + "learning_rate": 1.5132399557099031e-05, + "loss": 0.7761, + "step": 3211 + }, + { + "epoch": 1.7437567861020629, + "grad_norm": 8.699280554503527, + "learning_rate": 1.5129380929373076e-05, + "loss": 1.2891, + "step": 3212 + }, + { + "epoch": 1.744299674267101, + "grad_norm": 11.932192326740894, + "learning_rate": 1.5126361667237116e-05, + "loss": 0.9868, + "step": 3213 + }, + { + "epoch": 1.7448425624321389, + "grad_norm": 8.456694582402795, + "learning_rate": 1.512334177106457e-05, + "loss": 1.199, + "step": 3214 + }, + { + "epoch": 1.745385450597177, + "grad_norm": 8.284979161281417, + "learning_rate": 1.512032124122895e-05, + "loss": 0.658, + "step": 3215 + }, + { + "epoch": 1.7459283387622149, + "grad_norm": 8.490755188379255, + "learning_rate": 1.5117300078103841e-05, + "loss": 0.6974, + "step": 3216 + }, + { + "epoch": 1.746471226927253, + "grad_norm": 14.621603534181931, + "learning_rate": 1.5114278282062898e-05, + "loss": 1.4386, + "step": 3217 + }, + { + "epoch": 1.7470141150922909, + "grad_norm": 8.35853855400652, + "learning_rate": 1.511125585347987e-05, + "loss": 0.7991, + "step": 3218 + }, + { + "epoch": 1.747557003257329, + "grad_norm": 10.14192231306983, + "learning_rate": 1.5108232792728567e-05, + "loss": 0.9942, + "step": 3219 + }, + { + "epoch": 1.7480998914223669, + "grad_norm": 6.003113555943506, + "learning_rate": 1.5105209100182893e-05, + "loss": 0.5614, + "step": 3220 + }, + { + "epoch": 1.748642779587405, + "grad_norm": 7.420823995167257, + "learning_rate": 1.5102184776216824e-05, + "loss": 0.6918, + "step": 3221 + }, + { + "epoch": 1.7491856677524429, + "grad_norm": 7.3724702610256, + "learning_rate": 1.5099159821204406e-05, + "loss": 0.815, + "step": 3222 + }, + { + "epoch": 1.749728555917481, + "grad_norm": 8.539056209113738, + "learning_rate": 1.509613423551978e-05, + "loss": 1.0859, + "step": 3223 + }, + { + "epoch": 1.750271444082519, + "grad_norm": 10.518635534148203, + "learning_rate": 1.5093108019537147e-05, + "loss": 1.1873, + "step": 3224 + }, + { + "epoch": 1.7508143322475571, + "grad_norm": 8.049010262043616, + "learning_rate": 1.5090081173630795e-05, + "loss": 0.971, + "step": 3225 + }, + { + "epoch": 1.751357220412595, + "grad_norm": 7.874724964543097, + "learning_rate": 1.5087053698175092e-05, + "loss": 1.06, + "step": 3226 + }, + { + "epoch": 1.7519001085776331, + "grad_norm": 8.018813791724389, + "learning_rate": 1.508402559354448e-05, + "loss": 0.7192, + "step": 3227 + }, + { + "epoch": 1.752442996742671, + "grad_norm": 11.748757549078345, + "learning_rate": 1.508099686011348e-05, + "loss": 1.4476, + "step": 3228 + }, + { + "epoch": 1.7529858849077091, + "grad_norm": 7.885344513181394, + "learning_rate": 1.507796749825669e-05, + "loss": 0.7493, + "step": 3229 + }, + { + "epoch": 1.753528773072747, + "grad_norm": 6.808213775155858, + "learning_rate": 1.507493750834879e-05, + "loss": 0.6646, + "step": 3230 + }, + { + "epoch": 1.7540716612377851, + "grad_norm": 6.304489454276083, + "learning_rate": 1.5071906890764527e-05, + "loss": 0.7375, + "step": 3231 + }, + { + "epoch": 1.754614549402823, + "grad_norm": 8.217636823407403, + "learning_rate": 1.5068875645878739e-05, + "loss": 1.0069, + "step": 3232 + }, + { + "epoch": 1.7551574375678611, + "grad_norm": 10.706707510999644, + "learning_rate": 1.5065843774066329e-05, + "loss": 0.689, + "step": 3233 + }, + { + "epoch": 1.755700325732899, + "grad_norm": 7.600175930285094, + "learning_rate": 1.5062811275702291e-05, + "loss": 0.8645, + "step": 3234 + }, + { + "epoch": 1.7562432138979371, + "grad_norm": 8.463001355349205, + "learning_rate": 1.5059778151161684e-05, + "loss": 0.7185, + "step": 3235 + }, + { + "epoch": 1.756786102062975, + "grad_norm": 5.834156873881953, + "learning_rate": 1.5056744400819651e-05, + "loss": 0.6486, + "step": 3236 + }, + { + "epoch": 1.7573289902280131, + "grad_norm": 7.152665898997485, + "learning_rate": 1.5053710025051411e-05, + "loss": 0.9116, + "step": 3237 + }, + { + "epoch": 1.757871878393051, + "grad_norm": 8.037482464894001, + "learning_rate": 1.5050675024232262e-05, + "loss": 0.8231, + "step": 3238 + }, + { + "epoch": 1.7584147665580891, + "grad_norm": 10.416359782719013, + "learning_rate": 1.5047639398737573e-05, + "loss": 1.2555, + "step": 3239 + }, + { + "epoch": 1.758957654723127, + "grad_norm": 6.789336798229921, + "learning_rate": 1.50446031489428e-05, + "loss": 0.7295, + "step": 3240 + }, + { + "epoch": 1.7595005428881652, + "grad_norm": 9.927450801451293, + "learning_rate": 1.5041566275223472e-05, + "loss": 1.5417, + "step": 3241 + }, + { + "epoch": 1.760043431053203, + "grad_norm": 7.040644062299268, + "learning_rate": 1.5038528777955188e-05, + "loss": 0.4495, + "step": 3242 + }, + { + "epoch": 1.7605863192182412, + "grad_norm": 8.848733486454524, + "learning_rate": 1.5035490657513638e-05, + "loss": 1.0123, + "step": 3243 + }, + { + "epoch": 1.761129207383279, + "grad_norm": 8.109736914130819, + "learning_rate": 1.503245191427458e-05, + "loss": 0.8984, + "step": 3244 + }, + { + "epoch": 1.7616720955483172, + "grad_norm": 7.210785707597705, + "learning_rate": 1.5029412548613845e-05, + "loss": 0.6947, + "step": 3245 + }, + { + "epoch": 1.762214983713355, + "grad_norm": 10.368251730052483, + "learning_rate": 1.502637256090735e-05, + "loss": 0.8413, + "step": 3246 + }, + { + "epoch": 1.7627578718783932, + "grad_norm": 6.770955437649876, + "learning_rate": 1.5023331951531086e-05, + "loss": 0.6386, + "step": 3247 + }, + { + "epoch": 1.763300760043431, + "grad_norm": 10.818384833705863, + "learning_rate": 1.5020290720861122e-05, + "loss": 1.2544, + "step": 3248 + }, + { + "epoch": 1.7638436482084692, + "grad_norm": 7.9104385792951195, + "learning_rate": 1.50172488692736e-05, + "loss": 0.6538, + "step": 3249 + }, + { + "epoch": 1.764386536373507, + "grad_norm": 9.565689306609565, + "learning_rate": 1.5014206397144742e-05, + "loss": 1.1232, + "step": 3250 + }, + { + "epoch": 1.7649294245385452, + "grad_norm": 11.651444088862487, + "learning_rate": 1.5011163304850844e-05, + "loss": 1.2054, + "step": 3251 + }, + { + "epoch": 1.765472312703583, + "grad_norm": 10.326063989708704, + "learning_rate": 1.500811959276828e-05, + "loss": 0.9441, + "step": 3252 + }, + { + "epoch": 1.7660152008686212, + "grad_norm": 8.0529225996154, + "learning_rate": 1.5005075261273506e-05, + "loss": 0.9431, + "step": 3253 + }, + { + "epoch": 1.766558089033659, + "grad_norm": 7.391183931496241, + "learning_rate": 1.500203031074305e-05, + "loss": 0.8619, + "step": 3254 + }, + { + "epoch": 1.7671009771986972, + "grad_norm": 6.892445188811786, + "learning_rate": 1.4998984741553508e-05, + "loss": 0.7039, + "step": 3255 + }, + { + "epoch": 1.767643865363735, + "grad_norm": 6.541771639151763, + "learning_rate": 1.4995938554081568e-05, + "loss": 0.6359, + "step": 3256 + }, + { + "epoch": 1.7681867535287732, + "grad_norm": 7.690176361123399, + "learning_rate": 1.4992891748703985e-05, + "loss": 0.8998, + "step": 3257 + }, + { + "epoch": 1.768729641693811, + "grad_norm": 6.497301879748706, + "learning_rate": 1.498984432579759e-05, + "loss": 0.5743, + "step": 3258 + }, + { + "epoch": 1.7692725298588492, + "grad_norm": 6.656723951695686, + "learning_rate": 1.4986796285739298e-05, + "loss": 0.7002, + "step": 3259 + }, + { + "epoch": 1.769815418023887, + "grad_norm": 7.9129397206513685, + "learning_rate": 1.4983747628906095e-05, + "loss": 0.9239, + "step": 3260 + }, + { + "epoch": 1.7703583061889252, + "grad_norm": 4.544544755639526, + "learning_rate": 1.4980698355675043e-05, + "loss": 0.3431, + "step": 3261 + }, + { + "epoch": 1.770901194353963, + "grad_norm": 8.243743590575315, + "learning_rate": 1.4977648466423278e-05, + "loss": 1.1386, + "step": 3262 + }, + { + "epoch": 1.7714440825190012, + "grad_norm": 9.404041237838763, + "learning_rate": 1.4974597961528021e-05, + "loss": 0.9087, + "step": 3263 + }, + { + "epoch": 1.771986970684039, + "grad_norm": 8.534670509056685, + "learning_rate": 1.4971546841366556e-05, + "loss": 0.8769, + "step": 3264 + }, + { + "epoch": 1.7725298588490772, + "grad_norm": 9.442139570892468, + "learning_rate": 1.4968495106316254e-05, + "loss": 0.6942, + "step": 3265 + }, + { + "epoch": 1.773072747014115, + "grad_norm": 8.449365139837813, + "learning_rate": 1.496544275675456e-05, + "loss": 1.0743, + "step": 3266 + }, + { + "epoch": 1.7736156351791532, + "grad_norm": 9.461545898435599, + "learning_rate": 1.4962389793058994e-05, + "loss": 0.7473, + "step": 3267 + }, + { + "epoch": 1.774158523344191, + "grad_norm": 11.871672085414737, + "learning_rate": 1.4959336215607147e-05, + "loss": 1.5433, + "step": 3268 + }, + { + "epoch": 1.7747014115092292, + "grad_norm": 6.417686676489462, + "learning_rate": 1.4956282024776693e-05, + "loss": 0.5006, + "step": 3269 + }, + { + "epoch": 1.775244299674267, + "grad_norm": 10.49161190534651, + "learning_rate": 1.4953227220945381e-05, + "loss": 0.9918, + "step": 3270 + }, + { + "epoch": 1.7757871878393052, + "grad_norm": 9.298743913365943, + "learning_rate": 1.4950171804491031e-05, + "loss": 1.3149, + "step": 3271 + }, + { + "epoch": 1.776330076004343, + "grad_norm": 7.775473503246929, + "learning_rate": 1.4947115775791541e-05, + "loss": 0.6873, + "step": 3272 + }, + { + "epoch": 1.7768729641693812, + "grad_norm": 6.75591368497042, + "learning_rate": 1.4944059135224891e-05, + "loss": 0.7799, + "step": 3273 + }, + { + "epoch": 1.777415852334419, + "grad_norm": 8.234180005996857, + "learning_rate": 1.4941001883169124e-05, + "loss": 1.0393, + "step": 3274 + }, + { + "epoch": 1.7779587404994572, + "grad_norm": 8.419005181659063, + "learning_rate": 1.4937944020002371e-05, + "loss": 0.6798, + "step": 3275 + }, + { + "epoch": 1.778501628664495, + "grad_norm": 7.0982600255490205, + "learning_rate": 1.4934885546102833e-05, + "loss": 0.4936, + "step": 3276 + }, + { + "epoch": 1.7790445168295332, + "grad_norm": 8.404626943584326, + "learning_rate": 1.4931826461848785e-05, + "loss": 1.0507, + "step": 3277 + }, + { + "epoch": 1.779587404994571, + "grad_norm": 5.861882830901275, + "learning_rate": 1.4928766767618576e-05, + "loss": 0.6785, + "step": 3278 + }, + { + "epoch": 1.7801302931596092, + "grad_norm": 8.114055102204162, + "learning_rate": 1.4925706463790642e-05, + "loss": 0.765, + "step": 3279 + }, + { + "epoch": 1.780673181324647, + "grad_norm": 10.033269670592238, + "learning_rate": 1.492264555074348e-05, + "loss": 0.9163, + "step": 3280 + }, + { + "epoch": 1.7812160694896852, + "grad_norm": 7.040755628619449, + "learning_rate": 1.4919584028855671e-05, + "loss": 0.5962, + "step": 3281 + }, + { + "epoch": 1.781758957654723, + "grad_norm": 12.093987616559675, + "learning_rate": 1.4916521898505872e-05, + "loss": 0.8999, + "step": 3282 + }, + { + "epoch": 1.7823018458197613, + "grad_norm": 9.400200287574112, + "learning_rate": 1.4913459160072805e-05, + "loss": 1.0203, + "step": 3283 + }, + { + "epoch": 1.782844733984799, + "grad_norm": 8.605701710844945, + "learning_rate": 1.4910395813935279e-05, + "loss": 0.7874, + "step": 3284 + }, + { + "epoch": 1.7833876221498373, + "grad_norm": 10.279849472300688, + "learning_rate": 1.4907331860472174e-05, + "loss": 1.4306, + "step": 3285 + }, + { + "epoch": 1.783930510314875, + "grad_norm": 10.582881297295359, + "learning_rate": 1.4904267300062443e-05, + "loss": 0.6536, + "step": 3286 + }, + { + "epoch": 1.7844733984799133, + "grad_norm": 6.960885594125151, + "learning_rate": 1.4901202133085115e-05, + "loss": 0.7256, + "step": 3287 + }, + { + "epoch": 1.785016286644951, + "grad_norm": 9.088589913670967, + "learning_rate": 1.4898136359919298e-05, + "loss": 0.9299, + "step": 3288 + }, + { + "epoch": 1.7855591748099893, + "grad_norm": 8.922355254191798, + "learning_rate": 1.4895069980944168e-05, + "loss": 0.8159, + "step": 3289 + }, + { + "epoch": 1.786102062975027, + "grad_norm": 9.459478678151763, + "learning_rate": 1.4892002996538982e-05, + "loss": 1.132, + "step": 3290 + }, + { + "epoch": 1.7866449511400653, + "grad_norm": 9.746458814865415, + "learning_rate": 1.4888935407083065e-05, + "loss": 0.9874, + "step": 3291 + }, + { + "epoch": 1.787187839305103, + "grad_norm": 8.13065950431162, + "learning_rate": 1.4885867212955827e-05, + "loss": 0.8013, + "step": 3292 + }, + { + "epoch": 1.7877307274701413, + "grad_norm": 8.469962208509646, + "learning_rate": 1.4882798414536749e-05, + "loss": 1.0408, + "step": 3293 + }, + { + "epoch": 1.788273615635179, + "grad_norm": 9.018264603768202, + "learning_rate": 1.4879729012205379e-05, + "loss": 0.8324, + "step": 3294 + }, + { + "epoch": 1.7888165038002173, + "grad_norm": 8.934092671110381, + "learning_rate": 1.4876659006341347e-05, + "loss": 0.8842, + "step": 3295 + }, + { + "epoch": 1.789359391965255, + "grad_norm": 7.612464888658943, + "learning_rate": 1.4873588397324356e-05, + "loss": 0.9356, + "step": 3296 + }, + { + "epoch": 1.7899022801302933, + "grad_norm": 7.303452266512586, + "learning_rate": 1.4870517185534184e-05, + "loss": 1.0291, + "step": 3297 + }, + { + "epoch": 1.790445168295331, + "grad_norm": 9.805172223903986, + "learning_rate": 1.4867445371350687e-05, + "loss": 1.0243, + "step": 3298 + }, + { + "epoch": 1.7909880564603693, + "grad_norm": 7.742224062288237, + "learning_rate": 1.4864372955153785e-05, + "loss": 0.7173, + "step": 3299 + }, + { + "epoch": 1.791530944625407, + "grad_norm": 9.174714938761847, + "learning_rate": 1.4861299937323485e-05, + "loss": 0.9721, + "step": 3300 + }, + { + "epoch": 1.7920738327904453, + "grad_norm": 8.116980821679304, + "learning_rate": 1.4858226318239864e-05, + "loss": 0.7776, + "step": 3301 + }, + { + "epoch": 1.792616720955483, + "grad_norm": 5.885386861657401, + "learning_rate": 1.4855152098283068e-05, + "loss": 0.3833, + "step": 3302 + }, + { + "epoch": 1.7931596091205213, + "grad_norm": 6.804177263707199, + "learning_rate": 1.4852077277833319e-05, + "loss": 0.7972, + "step": 3303 + }, + { + "epoch": 1.793702497285559, + "grad_norm": 8.874804659745028, + "learning_rate": 1.4849001857270922e-05, + "loss": 1.063, + "step": 3304 + }, + { + "epoch": 1.7942453854505973, + "grad_norm": 10.231899453600786, + "learning_rate": 1.4845925836976247e-05, + "loss": 1.6617, + "step": 3305 + }, + { + "epoch": 1.794788273615635, + "grad_norm": 9.067079935451718, + "learning_rate": 1.4842849217329742e-05, + "loss": 1.1062, + "step": 3306 + }, + { + "epoch": 1.7953311617806733, + "grad_norm": 9.5261315220339, + "learning_rate": 1.4839771998711925e-05, + "loss": 1.5991, + "step": 3307 + }, + { + "epoch": 1.795874049945711, + "grad_norm": 9.53442504374803, + "learning_rate": 1.4836694181503397e-05, + "loss": 0.9267, + "step": 3308 + }, + { + "epoch": 1.7964169381107493, + "grad_norm": 8.231497735162955, + "learning_rate": 1.4833615766084821e-05, + "loss": 0.9802, + "step": 3309 + }, + { + "epoch": 1.796959826275787, + "grad_norm": 7.2171054452276815, + "learning_rate": 1.4830536752836944e-05, + "loss": 0.6116, + "step": 3310 + }, + { + "epoch": 1.7975027144408253, + "grad_norm": 8.412637332813022, + "learning_rate": 1.4827457142140583e-05, + "loss": 0.7482, + "step": 3311 + }, + { + "epoch": 1.798045602605863, + "grad_norm": 6.24239912297862, + "learning_rate": 1.482437693437663e-05, + "loss": 0.4333, + "step": 3312 + }, + { + "epoch": 1.7985884907709013, + "grad_norm": 9.706212881828902, + "learning_rate": 1.4821296129926049e-05, + "loss": 1.3862, + "step": 3313 + }, + { + "epoch": 1.799131378935939, + "grad_norm": 8.275124350006633, + "learning_rate": 1.4818214729169878e-05, + "loss": 0.812, + "step": 3314 + }, + { + "epoch": 1.7996742671009773, + "grad_norm": 7.0810692617964275, + "learning_rate": 1.4815132732489228e-05, + "loss": 0.8919, + "step": 3315 + }, + { + "epoch": 1.8002171552660151, + "grad_norm": 7.225811411306636, + "learning_rate": 1.4812050140265288e-05, + "loss": 0.7512, + "step": 3316 + }, + { + "epoch": 1.8007600434310533, + "grad_norm": 6.458874811050634, + "learning_rate": 1.4808966952879318e-05, + "loss": 0.6914, + "step": 3317 + }, + { + "epoch": 1.8013029315960911, + "grad_norm": 6.537680447766501, + "learning_rate": 1.4805883170712653e-05, + "loss": 0.5511, + "step": 3318 + }, + { + "epoch": 1.8018458197611293, + "grad_norm": 8.023777979063889, + "learning_rate": 1.4802798794146695e-05, + "loss": 0.6814, + "step": 3319 + }, + { + "epoch": 1.8023887079261671, + "grad_norm": 13.062126505855263, + "learning_rate": 1.4799713823562932e-05, + "loss": 1.0526, + "step": 3320 + }, + { + "epoch": 1.8029315960912053, + "grad_norm": 8.173027381817493, + "learning_rate": 1.479662825934291e-05, + "loss": 0.8939, + "step": 3321 + }, + { + "epoch": 1.8034744842562431, + "grad_norm": 7.465360921787222, + "learning_rate": 1.4793542101868261e-05, + "loss": 0.5965, + "step": 3322 + }, + { + "epoch": 1.8040173724212814, + "grad_norm": 6.5098081549027835, + "learning_rate": 1.4790455351520684e-05, + "loss": 0.5963, + "step": 3323 + }, + { + "epoch": 1.8045602605863191, + "grad_norm": 11.071700718231854, + "learning_rate": 1.4787368008681956e-05, + "loss": 0.9728, + "step": 3324 + }, + { + "epoch": 1.8051031487513574, + "grad_norm": 9.251168051579464, + "learning_rate": 1.4784280073733924e-05, + "loss": 0.8816, + "step": 3325 + }, + { + "epoch": 1.8056460369163951, + "grad_norm": 7.876824650963893, + "learning_rate": 1.4781191547058505e-05, + "loss": 1.0365, + "step": 3326 + }, + { + "epoch": 1.8061889250814334, + "grad_norm": 7.984254709859244, + "learning_rate": 1.4778102429037696e-05, + "loss": 0.7252, + "step": 3327 + }, + { + "epoch": 1.8067318132464711, + "grad_norm": 8.162664097932598, + "learning_rate": 1.4775012720053563e-05, + "loss": 0.8297, + "step": 3328 + }, + { + "epoch": 1.8072747014115094, + "grad_norm": 6.631733087479191, + "learning_rate": 1.4771922420488245e-05, + "loss": 0.6962, + "step": 3329 + }, + { + "epoch": 1.8078175895765471, + "grad_norm": 8.80250228426254, + "learning_rate": 1.476883153072396e-05, + "loss": 1.0576, + "step": 3330 + }, + { + "epoch": 1.8083604777415854, + "grad_norm": 10.637063152950418, + "learning_rate": 1.476574005114299e-05, + "loss": 0.8866, + "step": 3331 + }, + { + "epoch": 1.8089033659066232, + "grad_norm": 9.58332512337787, + "learning_rate": 1.4762647982127696e-05, + "loss": 1.6148, + "step": 3332 + }, + { + "epoch": 1.8094462540716614, + "grad_norm": 7.826567769007031, + "learning_rate": 1.4759555324060508e-05, + "loss": 0.9256, + "step": 3333 + }, + { + "epoch": 1.8099891422366992, + "grad_norm": 9.445307614493151, + "learning_rate": 1.4756462077323931e-05, + "loss": 1.1389, + "step": 3334 + }, + { + "epoch": 1.8105320304017374, + "grad_norm": 7.504505604214132, + "learning_rate": 1.4753368242300543e-05, + "loss": 1.1333, + "step": 3335 + }, + { + "epoch": 1.8110749185667752, + "grad_norm": 6.996150668976886, + "learning_rate": 1.4750273819372996e-05, + "loss": 0.6721, + "step": 3336 + }, + { + "epoch": 1.8116178067318134, + "grad_norm": 7.593185686378431, + "learning_rate": 1.4747178808924012e-05, + "loss": 0.6856, + "step": 3337 + }, + { + "epoch": 1.8121606948968512, + "grad_norm": 7.277948846492979, + "learning_rate": 1.474408321133639e-05, + "loss": 0.5487, + "step": 3338 + }, + { + "epoch": 1.8127035830618894, + "grad_norm": 9.500801329454344, + "learning_rate": 1.474098702699299e-05, + "loss": 1.063, + "step": 3339 + }, + { + "epoch": 1.8132464712269272, + "grad_norm": 7.732920511283457, + "learning_rate": 1.4737890256276763e-05, + "loss": 0.8875, + "step": 3340 + }, + { + "epoch": 1.8137893593919654, + "grad_norm": 9.56546309643246, + "learning_rate": 1.4734792899570716e-05, + "loss": 0.7724, + "step": 3341 + }, + { + "epoch": 1.8143322475570032, + "grad_norm": 7.735046028758256, + "learning_rate": 1.4731694957257938e-05, + "loss": 0.8498, + "step": 3342 + }, + { + "epoch": 1.8148751357220414, + "grad_norm": 9.077413763001829, + "learning_rate": 1.4728596429721587e-05, + "loss": 1.1604, + "step": 3343 + }, + { + "epoch": 1.8154180238870792, + "grad_norm": 8.33530084886159, + "learning_rate": 1.4725497317344892e-05, + "loss": 1.0126, + "step": 3344 + }, + { + "epoch": 1.8159609120521174, + "grad_norm": 11.938749223535789, + "learning_rate": 1.4722397620511158e-05, + "loss": 0.9538, + "step": 3345 + }, + { + "epoch": 1.8165038002171552, + "grad_norm": 4.946988727080576, + "learning_rate": 1.471929733960376e-05, + "loss": 0.3007, + "step": 3346 + }, + { + "epoch": 1.8170466883821934, + "grad_norm": 6.934881579248824, + "learning_rate": 1.4716196475006145e-05, + "loss": 0.5286, + "step": 3347 + }, + { + "epoch": 1.8175895765472312, + "grad_norm": 12.504257989709076, + "learning_rate": 1.4713095027101833e-05, + "loss": 0.9948, + "step": 3348 + }, + { + "epoch": 1.8181324647122694, + "grad_norm": 6.299717892798866, + "learning_rate": 1.4709992996274416e-05, + "loss": 0.6503, + "step": 3349 + }, + { + "epoch": 1.8186753528773072, + "grad_norm": 8.48347618555589, + "learning_rate": 1.4706890382907562e-05, + "loss": 0.7195, + "step": 3350 + }, + { + "epoch": 1.8192182410423454, + "grad_norm": 7.992565784639017, + "learning_rate": 1.4703787187385002e-05, + "loss": 0.7349, + "step": 3351 + }, + { + "epoch": 1.8197611292073832, + "grad_norm": 7.567197193038979, + "learning_rate": 1.4700683410090546e-05, + "loss": 0.4719, + "step": 3352 + }, + { + "epoch": 1.8203040173724214, + "grad_norm": 7.276389893333727, + "learning_rate": 1.4697579051408073e-05, + "loss": 0.93, + "step": 3353 + }, + { + "epoch": 1.8208469055374592, + "grad_norm": 8.13251629417801, + "learning_rate": 1.4694474111721537e-05, + "loss": 0.9865, + "step": 3354 + }, + { + "epoch": 1.8213897937024974, + "grad_norm": 6.788010338042944, + "learning_rate": 1.469136859141496e-05, + "loss": 0.5671, + "step": 3355 + }, + { + "epoch": 1.8219326818675352, + "grad_norm": 6.83194614525194, + "learning_rate": 1.4688262490872438e-05, + "loss": 0.5486, + "step": 3356 + }, + { + "epoch": 1.8224755700325734, + "grad_norm": 12.201391760355081, + "learning_rate": 1.4685155810478143e-05, + "loss": 0.9178, + "step": 3357 + }, + { + "epoch": 1.8230184581976112, + "grad_norm": 7.955977866907339, + "learning_rate": 1.4682048550616306e-05, + "loss": 0.9671, + "step": 3358 + }, + { + "epoch": 1.8235613463626494, + "grad_norm": 9.699881747283202, + "learning_rate": 1.4678940711671245e-05, + "loss": 0.7699, + "step": 3359 + }, + { + "epoch": 1.8241042345276872, + "grad_norm": 7.703855587772515, + "learning_rate": 1.4675832294027339e-05, + "loss": 0.6471, + "step": 3360 + }, + { + "epoch": 1.8246471226927254, + "grad_norm": 11.230217201340418, + "learning_rate": 1.4672723298069043e-05, + "loss": 1.1256, + "step": 3361 + }, + { + "epoch": 1.8251900108577632, + "grad_norm": 11.964156332533788, + "learning_rate": 1.4669613724180883e-05, + "loss": 1.1829, + "step": 3362 + }, + { + "epoch": 1.8257328990228014, + "grad_norm": 8.221701868139569, + "learning_rate": 1.4666503572747452e-05, + "loss": 0.6004, + "step": 3363 + }, + { + "epoch": 1.8262757871878392, + "grad_norm": 6.179895131215982, + "learning_rate": 1.4663392844153426e-05, + "loss": 0.6923, + "step": 3364 + }, + { + "epoch": 1.8268186753528775, + "grad_norm": 8.333959237389427, + "learning_rate": 1.4660281538783538e-05, + "loss": 0.8785, + "step": 3365 + }, + { + "epoch": 1.8273615635179152, + "grad_norm": 9.742872979143373, + "learning_rate": 1.4657169657022603e-05, + "loss": 1.0129, + "step": 3366 + }, + { + "epoch": 1.8279044516829535, + "grad_norm": 8.019435507225802, + "learning_rate": 1.46540571992555e-05, + "loss": 1.0615, + "step": 3367 + }, + { + "epoch": 1.8284473398479912, + "grad_norm": 5.872559756186675, + "learning_rate": 1.4650944165867187e-05, + "loss": 0.6391, + "step": 3368 + }, + { + "epoch": 1.8289902280130295, + "grad_norm": 7.878073300625923, + "learning_rate": 1.4647830557242685e-05, + "loss": 1.0317, + "step": 3369 + }, + { + "epoch": 1.8295331161780672, + "grad_norm": 10.290400619980408, + "learning_rate": 1.4644716373767096e-05, + "loss": 1.3108, + "step": 3370 + }, + { + "epoch": 1.8300760043431055, + "grad_norm": 7.655417729944032, + "learning_rate": 1.4641601615825582e-05, + "loss": 0.8434, + "step": 3371 + }, + { + "epoch": 1.8306188925081432, + "grad_norm": 8.894875963607664, + "learning_rate": 1.4638486283803384e-05, + "loss": 1.4973, + "step": 3372 + }, + { + "epoch": 1.8311617806731815, + "grad_norm": 7.70346849256546, + "learning_rate": 1.4635370378085809e-05, + "loss": 0.9202, + "step": 3373 + }, + { + "epoch": 1.8317046688382193, + "grad_norm": 7.428334777443758, + "learning_rate": 1.4632253899058233e-05, + "loss": 1.005, + "step": 3374 + }, + { + "epoch": 1.8322475570032575, + "grad_norm": 6.905061230055496, + "learning_rate": 1.4629136847106118e-05, + "loss": 0.7045, + "step": 3375 + }, + { + "epoch": 1.8327904451682953, + "grad_norm": 9.092059147160219, + "learning_rate": 1.4626019222614977e-05, + "loss": 1.1478, + "step": 3376 + }, + { + "epoch": 1.8333333333333335, + "grad_norm": 8.171417716573657, + "learning_rate": 1.4622901025970406e-05, + "loss": 0.8098, + "step": 3377 + }, + { + "epoch": 1.8338762214983713, + "grad_norm": 8.935159598593566, + "learning_rate": 1.461978225755807e-05, + "loss": 0.9121, + "step": 3378 + }, + { + "epoch": 1.8344191096634095, + "grad_norm": 9.059802045574934, + "learning_rate": 1.46166629177637e-05, + "loss": 0.952, + "step": 3379 + }, + { + "epoch": 1.8349619978284473, + "grad_norm": 6.404641837192718, + "learning_rate": 1.4613543006973103e-05, + "loss": 0.7178, + "step": 3380 + }, + { + "epoch": 1.8355048859934855, + "grad_norm": 7.714640842582964, + "learning_rate": 1.4610422525572155e-05, + "loss": 0.7053, + "step": 3381 + }, + { + "epoch": 1.8360477741585233, + "grad_norm": 9.232659815202982, + "learning_rate": 1.46073014739468e-05, + "loss": 0.7707, + "step": 3382 + }, + { + "epoch": 1.8365906623235615, + "grad_norm": 7.597273066110282, + "learning_rate": 1.4604179852483056e-05, + "loss": 0.9593, + "step": 3383 + }, + { + "epoch": 1.8371335504885993, + "grad_norm": 7.246167686662127, + "learning_rate": 1.460105766156701e-05, + "loss": 0.8489, + "step": 3384 + }, + { + "epoch": 1.8376764386536375, + "grad_norm": 9.724005322694314, + "learning_rate": 1.4597934901584818e-05, + "loss": 1.4838, + "step": 3385 + }, + { + "epoch": 1.8382193268186753, + "grad_norm": 5.623806840885622, + "learning_rate": 1.459481157292271e-05, + "loss": 0.4218, + "step": 3386 + }, + { + "epoch": 1.8387622149837135, + "grad_norm": 10.505590233471066, + "learning_rate": 1.4591687675966984e-05, + "loss": 1.3634, + "step": 3387 + }, + { + "epoch": 1.8393051031487513, + "grad_norm": 7.225825527188378, + "learning_rate": 1.458856321110401e-05, + "loss": 1.0878, + "step": 3388 + }, + { + "epoch": 1.8398479913137895, + "grad_norm": 6.92346369662784, + "learning_rate": 1.4585438178720221e-05, + "loss": 0.8579, + "step": 3389 + }, + { + "epoch": 1.8403908794788273, + "grad_norm": 6.78903317106508, + "learning_rate": 1.4582312579202134e-05, + "loss": 0.9986, + "step": 3390 + }, + { + "epoch": 1.8409337676438655, + "grad_norm": 7.493115074862596, + "learning_rate": 1.4579186412936323e-05, + "loss": 0.6403, + "step": 3391 + }, + { + "epoch": 1.8414766558089033, + "grad_norm": 7.292140639462925, + "learning_rate": 1.457605968030944e-05, + "loss": 0.7705, + "step": 3392 + }, + { + "epoch": 1.8420195439739415, + "grad_norm": 8.413656799764597, + "learning_rate": 1.4572932381708198e-05, + "loss": 0.7362, + "step": 3393 + }, + { + "epoch": 1.8425624321389793, + "grad_norm": 8.092404119538978, + "learning_rate": 1.4569804517519392e-05, + "loss": 0.8419, + "step": 3394 + }, + { + "epoch": 1.8431053203040175, + "grad_norm": 8.746713556642208, + "learning_rate": 1.456667608812988e-05, + "loss": 0.8992, + "step": 3395 + }, + { + "epoch": 1.8436482084690553, + "grad_norm": 9.315011907817567, + "learning_rate": 1.4563547093926595e-05, + "loss": 1.1811, + "step": 3396 + }, + { + "epoch": 1.8441910966340935, + "grad_norm": 7.921580843874425, + "learning_rate": 1.4560417535296529e-05, + "loss": 0.9896, + "step": 3397 + }, + { + "epoch": 1.8447339847991313, + "grad_norm": 5.324625495868631, + "learning_rate": 1.4557287412626755e-05, + "loss": 0.5757, + "step": 3398 + }, + { + "epoch": 1.8452768729641695, + "grad_norm": 7.061326556489923, + "learning_rate": 1.4554156726304411e-05, + "loss": 0.7016, + "step": 3399 + }, + { + "epoch": 1.8458197611292073, + "grad_norm": 7.833004564597817, + "learning_rate": 1.4551025476716704e-05, + "loss": 0.6112, + "step": 3400 + }, + { + "epoch": 1.8463626492942455, + "grad_norm": 6.762019290951184, + "learning_rate": 1.4547893664250912e-05, + "loss": 0.6197, + "step": 3401 + }, + { + "epoch": 1.8469055374592833, + "grad_norm": 8.02135944193365, + "learning_rate": 1.4544761289294384e-05, + "loss": 0.7665, + "step": 3402 + }, + { + "epoch": 1.8474484256243215, + "grad_norm": 9.145981041987138, + "learning_rate": 1.4541628352234538e-05, + "loss": 1.4491, + "step": 3403 + }, + { + "epoch": 1.8479913137893593, + "grad_norm": 10.127072377919164, + "learning_rate": 1.4538494853458858e-05, + "loss": 1.1929, + "step": 3404 + }, + { + "epoch": 1.8485342019543975, + "grad_norm": 9.041637256167727, + "learning_rate": 1.45353607933549e-05, + "loss": 0.8471, + "step": 3405 + }, + { + "epoch": 1.8490770901194353, + "grad_norm": 7.08038124220055, + "learning_rate": 1.453222617231029e-05, + "loss": 0.5152, + "step": 3406 + }, + { + "epoch": 1.8496199782844736, + "grad_norm": 8.460108981189723, + "learning_rate": 1.4529090990712726e-05, + "loss": 0.9495, + "step": 3407 + }, + { + "epoch": 1.8501628664495113, + "grad_norm": 8.939702203617836, + "learning_rate": 1.452595524894997e-05, + "loss": 1.0959, + "step": 3408 + }, + { + "epoch": 1.8507057546145496, + "grad_norm": 7.5616663713048595, + "learning_rate": 1.4522818947409855e-05, + "loss": 1.0391, + "step": 3409 + }, + { + "epoch": 1.8512486427795873, + "grad_norm": 8.75531450227851, + "learning_rate": 1.4519682086480287e-05, + "loss": 0.7298, + "step": 3410 + }, + { + "epoch": 1.8517915309446256, + "grad_norm": 7.003753183958329, + "learning_rate": 1.4516544666549233e-05, + "loss": 0.4788, + "step": 3411 + }, + { + "epoch": 1.8523344191096633, + "grad_norm": 10.952441700747523, + "learning_rate": 1.4513406688004734e-05, + "loss": 1.0903, + "step": 3412 + }, + { + "epoch": 1.8528773072747016, + "grad_norm": 8.61536651212902, + "learning_rate": 1.4510268151234903e-05, + "loss": 0.6924, + "step": 3413 + }, + { + "epoch": 1.8534201954397393, + "grad_norm": 9.237257682023538, + "learning_rate": 1.4507129056627922e-05, + "loss": 0.8686, + "step": 3414 + }, + { + "epoch": 1.8539630836047776, + "grad_norm": 7.792081331119547, + "learning_rate": 1.4503989404572031e-05, + "loss": 0.9818, + "step": 3415 + }, + { + "epoch": 1.8545059717698154, + "grad_norm": 8.622336937881878, + "learning_rate": 1.4500849195455557e-05, + "loss": 0.9444, + "step": 3416 + }, + { + "epoch": 1.8550488599348536, + "grad_norm": 7.607437519589793, + "learning_rate": 1.4497708429666882e-05, + "loss": 0.846, + "step": 3417 + }, + { + "epoch": 1.8555917480998914, + "grad_norm": 8.358779281223866, + "learning_rate": 1.4494567107594457e-05, + "loss": 0.7125, + "step": 3418 + }, + { + "epoch": 1.8561346362649296, + "grad_norm": 6.8176708905901044, + "learning_rate": 1.449142522962681e-05, + "loss": 0.8921, + "step": 3419 + }, + { + "epoch": 1.8566775244299674, + "grad_norm": 8.931750092418099, + "learning_rate": 1.448828279615253e-05, + "loss": 0.7989, + "step": 3420 + }, + { + "epoch": 1.8572204125950056, + "grad_norm": 7.481066786036915, + "learning_rate": 1.4485139807560281e-05, + "loss": 0.6642, + "step": 3421 + }, + { + "epoch": 1.8577633007600434, + "grad_norm": 7.460924029721227, + "learning_rate": 1.4481996264238796e-05, + "loss": 0.6509, + "step": 3422 + }, + { + "epoch": 1.8583061889250816, + "grad_norm": 7.224784834236484, + "learning_rate": 1.4478852166576869e-05, + "loss": 0.7244, + "step": 3423 + }, + { + "epoch": 1.8588490770901194, + "grad_norm": 5.912102755785491, + "learning_rate": 1.4475707514963367e-05, + "loss": 0.5713, + "step": 3424 + }, + { + "epoch": 1.8593919652551576, + "grad_norm": 6.994884495015334, + "learning_rate": 1.4472562309787224e-05, + "loss": 0.4875, + "step": 3425 + }, + { + "epoch": 1.8599348534201954, + "grad_norm": 6.38845062016643, + "learning_rate": 1.4469416551437445e-05, + "loss": 0.4708, + "step": 3426 + }, + { + "epoch": 1.8604777415852336, + "grad_norm": 11.250341421225603, + "learning_rate": 1.4466270240303109e-05, + "loss": 1.4338, + "step": 3427 + }, + { + "epoch": 1.8610206297502714, + "grad_norm": 9.828853213376544, + "learning_rate": 1.4463123376773348e-05, + "loss": 0.7307, + "step": 3428 + }, + { + "epoch": 1.8615635179153096, + "grad_norm": 7.96498248613748, + "learning_rate": 1.4459975961237375e-05, + "loss": 0.6727, + "step": 3429 + }, + { + "epoch": 1.8621064060803474, + "grad_norm": 9.736354023562699, + "learning_rate": 1.4456827994084473e-05, + "loss": 0.8091, + "step": 3430 + }, + { + "epoch": 1.8626492942453856, + "grad_norm": 10.317215804231997, + "learning_rate": 1.4453679475703974e-05, + "loss": 1.0997, + "step": 3431 + }, + { + "epoch": 1.8631921824104234, + "grad_norm": 10.00211430988322, + "learning_rate": 1.4450530406485301e-05, + "loss": 1.4007, + "step": 3432 + }, + { + "epoch": 1.8637350705754616, + "grad_norm": 10.665872978467544, + "learning_rate": 1.4447380786817934e-05, + "loss": 0.9221, + "step": 3433 + }, + { + "epoch": 1.8642779587404994, + "grad_norm": 8.8489121199226, + "learning_rate": 1.4444230617091424e-05, + "loss": 0.7745, + "step": 3434 + }, + { + "epoch": 1.8648208469055376, + "grad_norm": 7.083707820329958, + "learning_rate": 1.444107989769539e-05, + "loss": 0.7152, + "step": 3435 + }, + { + "epoch": 1.8653637350705754, + "grad_norm": 7.496522090143239, + "learning_rate": 1.4437928629019515e-05, + "loss": 0.5901, + "step": 3436 + }, + { + "epoch": 1.8659066232356136, + "grad_norm": 10.576310319834935, + "learning_rate": 1.4434776811453552e-05, + "loss": 0.7752, + "step": 3437 + }, + { + "epoch": 1.8664495114006514, + "grad_norm": 8.686730112665717, + "learning_rate": 1.4431624445387327e-05, + "loss": 0.8748, + "step": 3438 + }, + { + "epoch": 1.8669923995656896, + "grad_norm": 9.435491336462407, + "learning_rate": 1.4428471531210726e-05, + "loss": 1.1954, + "step": 3439 + }, + { + "epoch": 1.8675352877307274, + "grad_norm": 9.012159097359817, + "learning_rate": 1.4425318069313712e-05, + "loss": 0.7353, + "step": 3440 + }, + { + "epoch": 1.8680781758957656, + "grad_norm": 9.617409218513528, + "learning_rate": 1.4422164060086304e-05, + "loss": 0.6611, + "step": 3441 + }, + { + "epoch": 1.8686210640608034, + "grad_norm": 9.612516235298202, + "learning_rate": 1.4419009503918598e-05, + "loss": 0.8101, + "step": 3442 + }, + { + "epoch": 1.8691639522258416, + "grad_norm": 6.542519560792006, + "learning_rate": 1.4415854401200751e-05, + "loss": 0.473, + "step": 3443 + }, + { + "epoch": 1.8697068403908794, + "grad_norm": 10.696166776054064, + "learning_rate": 1.4412698752322998e-05, + "loss": 1.0001, + "step": 3444 + }, + { + "epoch": 1.8702497285559176, + "grad_norm": 10.087499378994824, + "learning_rate": 1.4409542557675625e-05, + "loss": 0.9323, + "step": 3445 + }, + { + "epoch": 1.8707926167209554, + "grad_norm": 7.829085764895678, + "learning_rate": 1.4406385817649002e-05, + "loss": 0.8624, + "step": 3446 + }, + { + "epoch": 1.8713355048859937, + "grad_norm": 11.554825371147222, + "learning_rate": 1.4403228532633561e-05, + "loss": 0.7735, + "step": 3447 + }, + { + "epoch": 1.8718783930510314, + "grad_norm": 9.43390979006707, + "learning_rate": 1.4400070703019797e-05, + "loss": 0.9128, + "step": 3448 + }, + { + "epoch": 1.8724212812160697, + "grad_norm": 11.851717520188132, + "learning_rate": 1.4396912329198276e-05, + "loss": 1.2379, + "step": 3449 + }, + { + "epoch": 1.8729641693811074, + "grad_norm": 8.408242915866566, + "learning_rate": 1.4393753411559632e-05, + "loss": 0.7549, + "step": 3450 + }, + { + "epoch": 1.8735070575461457, + "grad_norm": 8.144856314123112, + "learning_rate": 1.4390593950494557e-05, + "loss": 0.7281, + "step": 3451 + }, + { + "epoch": 1.8740499457111834, + "grad_norm": 8.202846690293642, + "learning_rate": 1.438743394639383e-05, + "loss": 0.6773, + "step": 3452 + }, + { + "epoch": 1.8745928338762217, + "grad_norm": 9.423562458268494, + "learning_rate": 1.438427339964828e-05, + "loss": 1.1, + "step": 3453 + }, + { + "epoch": 1.8751357220412594, + "grad_norm": 7.196483568403804, + "learning_rate": 1.4381112310648806e-05, + "loss": 0.5809, + "step": 3454 + }, + { + "epoch": 1.8756786102062974, + "grad_norm": 7.26770042201738, + "learning_rate": 1.4377950679786382e-05, + "loss": 0.8053, + "step": 3455 + }, + { + "epoch": 1.8762214983713354, + "grad_norm": 9.472969211327861, + "learning_rate": 1.4374788507452039e-05, + "loss": 0.7998, + "step": 3456 + }, + { + "epoch": 1.8767643865363735, + "grad_norm": 9.514618174868874, + "learning_rate": 1.437162579403688e-05, + "loss": 0.9572, + "step": 3457 + }, + { + "epoch": 1.8773072747014115, + "grad_norm": 9.363034597703797, + "learning_rate": 1.4368462539932073e-05, + "loss": 0.8528, + "step": 3458 + }, + { + "epoch": 1.8778501628664495, + "grad_norm": 7.113386807438533, + "learning_rate": 1.4365298745528863e-05, + "loss": 0.733, + "step": 3459 + }, + { + "epoch": 1.8783930510314875, + "grad_norm": 9.09654372206602, + "learning_rate": 1.4362134411218545e-05, + "loss": 0.7779, + "step": 3460 + }, + { + "epoch": 1.8789359391965255, + "grad_norm": 7.449242636613164, + "learning_rate": 1.4358969537392489e-05, + "loss": 0.899, + "step": 3461 + }, + { + "epoch": 1.8794788273615635, + "grad_norm": 10.755480560042127, + "learning_rate": 1.4355804124442133e-05, + "loss": 1.099, + "step": 3462 + }, + { + "epoch": 1.8800217155266015, + "grad_norm": 9.091218073930861, + "learning_rate": 1.4352638172758984e-05, + "loss": 0.7489, + "step": 3463 + }, + { + "epoch": 1.8805646036916395, + "grad_norm": 9.701260384390375, + "learning_rate": 1.4349471682734603e-05, + "loss": 0.8495, + "step": 3464 + }, + { + "epoch": 1.8811074918566775, + "grad_norm": 9.008809447502522, + "learning_rate": 1.4346304654760637e-05, + "loss": 0.8984, + "step": 3465 + }, + { + "epoch": 1.8816503800217155, + "grad_norm": 8.176404536031864, + "learning_rate": 1.4343137089228783e-05, + "loss": 0.8278, + "step": 3466 + }, + { + "epoch": 1.8821932681867535, + "grad_norm": 9.197307290475337, + "learning_rate": 1.4339968986530813e-05, + "loss": 0.7142, + "step": 3467 + }, + { + "epoch": 1.8827361563517915, + "grad_norm": 5.505228455971892, + "learning_rate": 1.433680034705856e-05, + "loss": 0.4204, + "step": 3468 + }, + { + "epoch": 1.8832790445168295, + "grad_norm": 9.014964705534542, + "learning_rate": 1.433363117120393e-05, + "loss": 0.8988, + "step": 3469 + }, + { + "epoch": 1.8838219326818675, + "grad_norm": 7.290687267145188, + "learning_rate": 1.4330461459358889e-05, + "loss": 0.7683, + "step": 3470 + }, + { + "epoch": 1.8843648208469055, + "grad_norm": 10.834879009670003, + "learning_rate": 1.4327291211915473e-05, + "loss": 1.0446, + "step": 3471 + }, + { + "epoch": 1.8849077090119435, + "grad_norm": 8.240452054600638, + "learning_rate": 1.4324120429265781e-05, + "loss": 0.893, + "step": 3472 + }, + { + "epoch": 1.8854505971769815, + "grad_norm": 8.17919491677486, + "learning_rate": 1.4320949111801987e-05, + "loss": 0.9846, + "step": 3473 + }, + { + "epoch": 1.8859934853420195, + "grad_norm": 7.693481880154859, + "learning_rate": 1.431777725991632e-05, + "loss": 0.6054, + "step": 3474 + }, + { + "epoch": 1.8865363735070575, + "grad_norm": 7.929110152467141, + "learning_rate": 1.4314604874001076e-05, + "loss": 0.7299, + "step": 3475 + }, + { + "epoch": 1.8870792616720955, + "grad_norm": 9.650029930461764, + "learning_rate": 1.4311431954448629e-05, + "loss": 1.0855, + "step": 3476 + }, + { + "epoch": 1.8876221498371335, + "grad_norm": 8.487868906333745, + "learning_rate": 1.4308258501651404e-05, + "loss": 0.7117, + "step": 3477 + }, + { + "epoch": 1.8881650380021715, + "grad_norm": 7.111157851383161, + "learning_rate": 1.43050845160019e-05, + "loss": 0.6477, + "step": 3478 + }, + { + "epoch": 1.8887079261672095, + "grad_norm": 8.2908796123436, + "learning_rate": 1.4301909997892684e-05, + "loss": 0.9819, + "step": 3479 + }, + { + "epoch": 1.8892508143322475, + "grad_norm": 10.322900044501562, + "learning_rate": 1.4298734947716385e-05, + "loss": 0.9368, + "step": 3480 + }, + { + "epoch": 1.8897937024972855, + "grad_norm": 6.902927389625658, + "learning_rate": 1.4295559365865694e-05, + "loss": 0.6201, + "step": 3481 + }, + { + "epoch": 1.8903365906623235, + "grad_norm": 8.909574060437844, + "learning_rate": 1.4292383252733375e-05, + "loss": 0.9393, + "step": 3482 + }, + { + "epoch": 1.8908794788273615, + "grad_norm": 7.643883869480783, + "learning_rate": 1.4289206608712252e-05, + "loss": 0.9349, + "step": 3483 + }, + { + "epoch": 1.8914223669923995, + "grad_norm": 7.103928943926094, + "learning_rate": 1.428602943419522e-05, + "loss": 0.6511, + "step": 3484 + }, + { + "epoch": 1.8919652551574375, + "grad_norm": 8.843983286039231, + "learning_rate": 1.428285172957524e-05, + "loss": 0.6958, + "step": 3485 + }, + { + "epoch": 1.8925081433224755, + "grad_norm": 5.044738479793456, + "learning_rate": 1.427967349524533e-05, + "loss": 0.59, + "step": 3486 + }, + { + "epoch": 1.8930510314875135, + "grad_norm": 6.9998626206641195, + "learning_rate": 1.4276494731598582e-05, + "loss": 0.5973, + "step": 3487 + }, + { + "epoch": 1.8935939196525515, + "grad_norm": 8.316471360458495, + "learning_rate": 1.4273315439028151e-05, + "loss": 0.8941, + "step": 3488 + }, + { + "epoch": 1.8941368078175895, + "grad_norm": 12.713883241844957, + "learning_rate": 1.4270135617927254e-05, + "loss": 1.4921, + "step": 3489 + }, + { + "epoch": 1.8946796959826275, + "grad_norm": 10.043253133675226, + "learning_rate": 1.426695526868918e-05, + "loss": 1.0242, + "step": 3490 + }, + { + "epoch": 1.8952225841476655, + "grad_norm": 11.142761788047633, + "learning_rate": 1.4263774391707274e-05, + "loss": 1.265, + "step": 3491 + }, + { + "epoch": 1.8957654723127035, + "grad_norm": 7.12450280478247, + "learning_rate": 1.4260592987374962e-05, + "loss": 0.484, + "step": 3492 + }, + { + "epoch": 1.8963083604777415, + "grad_norm": 7.017938163026924, + "learning_rate": 1.4257411056085712e-05, + "loss": 0.6549, + "step": 3493 + }, + { + "epoch": 1.8968512486427795, + "grad_norm": 7.643680244898419, + "learning_rate": 1.4254228598233082e-05, + "loss": 0.614, + "step": 3494 + }, + { + "epoch": 1.8973941368078175, + "grad_norm": 9.484247292353862, + "learning_rate": 1.4251045614210678e-05, + "loss": 0.9667, + "step": 3495 + }, + { + "epoch": 1.8979370249728555, + "grad_norm": 10.846455465412243, + "learning_rate": 1.4247862104412175e-05, + "loss": 1.4425, + "step": 3496 + }, + { + "epoch": 1.8984799131378935, + "grad_norm": 7.211983265810251, + "learning_rate": 1.4244678069231319e-05, + "loss": 0.7508, + "step": 3497 + }, + { + "epoch": 1.8990228013029316, + "grad_norm": 8.283019454355765, + "learning_rate": 1.4241493509061912e-05, + "loss": 0.9747, + "step": 3498 + }, + { + "epoch": 1.8995656894679696, + "grad_norm": 6.678762476220025, + "learning_rate": 1.4238308424297833e-05, + "loss": 0.4702, + "step": 3499 + }, + { + "epoch": 1.9001085776330076, + "grad_norm": 8.750314729983545, + "learning_rate": 1.423512281533301e-05, + "loss": 0.7636, + "step": 3500 + }, + { + "epoch": 1.9006514657980456, + "grad_norm": 7.304249105438763, + "learning_rate": 1.4231936682561446e-05, + "loss": 0.7198, + "step": 3501 + }, + { + "epoch": 1.9011943539630836, + "grad_norm": 7.676960806676722, + "learning_rate": 1.4228750026377212e-05, + "loss": 0.8639, + "step": 3502 + }, + { + "epoch": 1.9017372421281216, + "grad_norm": 10.327339388528987, + "learning_rate": 1.4225562847174431e-05, + "loss": 0.9386, + "step": 3503 + }, + { + "epoch": 1.9022801302931596, + "grad_norm": 8.828430779713369, + "learning_rate": 1.4222375145347304e-05, + "loss": 1.0821, + "step": 3504 + }, + { + "epoch": 1.9028230184581976, + "grad_norm": 8.099300889556567, + "learning_rate": 1.421918692129009e-05, + "loss": 0.6251, + "step": 3505 + }, + { + "epoch": 1.9033659066232356, + "grad_norm": 9.197832742387652, + "learning_rate": 1.4215998175397115e-05, + "loss": 1.1794, + "step": 3506 + }, + { + "epoch": 1.9039087947882736, + "grad_norm": 9.305579433655693, + "learning_rate": 1.4212808908062763e-05, + "loss": 1.3447, + "step": 3507 + }, + { + "epoch": 1.9044516829533116, + "grad_norm": 6.897677914233707, + "learning_rate": 1.4209619119681497e-05, + "loss": 0.5178, + "step": 3508 + }, + { + "epoch": 1.9049945711183496, + "grad_norm": 9.359746343598495, + "learning_rate": 1.420642881064782e-05, + "loss": 0.985, + "step": 3509 + }, + { + "epoch": 1.9055374592833876, + "grad_norm": 5.855418324834113, + "learning_rate": 1.4203237981356331e-05, + "loss": 0.5144, + "step": 3510 + }, + { + "epoch": 1.9060803474484256, + "grad_norm": 7.408991833255532, + "learning_rate": 1.4200046632201665e-05, + "loss": 0.7358, + "step": 3511 + }, + { + "epoch": 1.9066232356134636, + "grad_norm": 9.024544122531433, + "learning_rate": 1.419685476357854e-05, + "loss": 0.778, + "step": 3512 + }, + { + "epoch": 1.9071661237785016, + "grad_norm": 7.874409137592207, + "learning_rate": 1.419366237588173e-05, + "loss": 0.8703, + "step": 3513 + }, + { + "epoch": 1.9077090119435396, + "grad_norm": 8.221027989379657, + "learning_rate": 1.4190469469506073e-05, + "loss": 0.7436, + "step": 3514 + }, + { + "epoch": 1.9082519001085776, + "grad_norm": 10.797659549408706, + "learning_rate": 1.4187276044846473e-05, + "loss": 1.0635, + "step": 3515 + }, + { + "epoch": 1.9087947882736156, + "grad_norm": 9.718910477411724, + "learning_rate": 1.4184082102297896e-05, + "loss": 0.8802, + "step": 3516 + }, + { + "epoch": 1.9093376764386536, + "grad_norm": 7.365463419520298, + "learning_rate": 1.4180887642255376e-05, + "loss": 0.5919, + "step": 3517 + }, + { + "epoch": 1.9098805646036916, + "grad_norm": 6.487376521266865, + "learning_rate": 1.4177692665114014e-05, + "loss": 0.7098, + "step": 3518 + }, + { + "epoch": 1.9104234527687296, + "grad_norm": 8.328541010595174, + "learning_rate": 1.4174497171268962e-05, + "loss": 0.6124, + "step": 3519 + }, + { + "epoch": 1.9109663409337676, + "grad_norm": 7.888322073346588, + "learning_rate": 1.4171301161115447e-05, + "loss": 0.827, + "step": 3520 + }, + { + "epoch": 1.9115092290988056, + "grad_norm": 10.407728423373431, + "learning_rate": 1.4168104635048756e-05, + "loss": 1.0511, + "step": 3521 + }, + { + "epoch": 1.9120521172638436, + "grad_norm": 8.154984984105669, + "learning_rate": 1.4164907593464239e-05, + "loss": 0.8156, + "step": 3522 + }, + { + "epoch": 1.9125950054288816, + "grad_norm": 9.142522776167024, + "learning_rate": 1.4161710036757314e-05, + "loss": 0.9046, + "step": 3523 + }, + { + "epoch": 1.9131378935939196, + "grad_norm": 8.610209396948733, + "learning_rate": 1.415851196532346e-05, + "loss": 0.5673, + "step": 3524 + }, + { + "epoch": 1.9136807817589576, + "grad_norm": 8.072698611315632, + "learning_rate": 1.415531337955822e-05, + "loss": 0.7375, + "step": 3525 + }, + { + "epoch": 1.9142236699239956, + "grad_norm": 9.878888535422037, + "learning_rate": 1.4152114279857197e-05, + "loss": 0.9677, + "step": 3526 + }, + { + "epoch": 1.9147665580890336, + "grad_norm": 8.524331697287977, + "learning_rate": 1.4148914666616062e-05, + "loss": 0.7617, + "step": 3527 + }, + { + "epoch": 1.9153094462540716, + "grad_norm": 8.04939882692512, + "learning_rate": 1.4145714540230549e-05, + "loss": 0.8767, + "step": 3528 + }, + { + "epoch": 1.9158523344191096, + "grad_norm": 10.69782500673993, + "learning_rate": 1.4142513901096453e-05, + "loss": 0.6357, + "step": 3529 + }, + { + "epoch": 1.9163952225841476, + "grad_norm": 7.279779735048647, + "learning_rate": 1.4139312749609637e-05, + "loss": 0.6474, + "step": 3530 + }, + { + "epoch": 1.9169381107491856, + "grad_norm": 7.290241222497209, + "learning_rate": 1.4136111086166024e-05, + "loss": 0.6929, + "step": 3531 + }, + { + "epoch": 1.9174809989142236, + "grad_norm": 8.3613074350218, + "learning_rate": 1.4132908911161598e-05, + "loss": 0.8199, + "step": 3532 + }, + { + "epoch": 1.9180238870792616, + "grad_norm": 9.015830183265928, + "learning_rate": 1.4129706224992413e-05, + "loss": 0.9811, + "step": 3533 + }, + { + "epoch": 1.9185667752442996, + "grad_norm": 9.621458967761185, + "learning_rate": 1.4126503028054579e-05, + "loss": 0.7498, + "step": 3534 + }, + { + "epoch": 1.9191096634093376, + "grad_norm": 11.105255757566027, + "learning_rate": 1.412329932074427e-05, + "loss": 0.7587, + "step": 3535 + }, + { + "epoch": 1.9196525515743756, + "grad_norm": 8.599145695443788, + "learning_rate": 1.4120095103457734e-05, + "loss": 0.934, + "step": 3536 + }, + { + "epoch": 1.9201954397394136, + "grad_norm": 7.727889210926642, + "learning_rate": 1.4116890376591268e-05, + "loss": 0.8305, + "step": 3537 + }, + { + "epoch": 1.9207383279044516, + "grad_norm": 12.029529213228301, + "learning_rate": 1.4113685140541242e-05, + "loss": 1.494, + "step": 3538 + }, + { + "epoch": 1.9212812160694897, + "grad_norm": 7.404838471492053, + "learning_rate": 1.411047939570408e-05, + "loss": 0.5168, + "step": 3539 + }, + { + "epoch": 1.9218241042345277, + "grad_norm": 7.540024472508733, + "learning_rate": 1.4107273142476272e-05, + "loss": 0.7162, + "step": 3540 + }, + { + "epoch": 1.9223669923995657, + "grad_norm": 7.885769945154571, + "learning_rate": 1.4104066381254378e-05, + "loss": 0.5964, + "step": 3541 + }, + { + "epoch": 1.9229098805646037, + "grad_norm": 8.287652724014645, + "learning_rate": 1.4100859112435013e-05, + "loss": 0.6175, + "step": 3542 + }, + { + "epoch": 1.9234527687296417, + "grad_norm": 8.335133777216624, + "learning_rate": 1.4097651336414857e-05, + "loss": 0.7163, + "step": 3543 + }, + { + "epoch": 1.9239956568946797, + "grad_norm": 7.765701865468595, + "learning_rate": 1.4094443053590652e-05, + "loss": 0.4534, + "step": 3544 + }, + { + "epoch": 1.9245385450597177, + "grad_norm": 7.716709219714201, + "learning_rate": 1.4091234264359206e-05, + "loss": 0.4425, + "step": 3545 + }, + { + "epoch": 1.9250814332247557, + "grad_norm": 11.979255681165771, + "learning_rate": 1.4088024969117387e-05, + "loss": 1.0249, + "step": 3546 + }, + { + "epoch": 1.9256243213897937, + "grad_norm": 9.424753754574079, + "learning_rate": 1.4084815168262123e-05, + "loss": 0.9663, + "step": 3547 + }, + { + "epoch": 1.9261672095548317, + "grad_norm": 10.101408847774994, + "learning_rate": 1.4081604862190407e-05, + "loss": 0.8136, + "step": 3548 + }, + { + "epoch": 1.9267100977198697, + "grad_norm": 11.755751498543427, + "learning_rate": 1.4078394051299298e-05, + "loss": 0.9451, + "step": 3549 + }, + { + "epoch": 1.9272529858849077, + "grad_norm": 12.251136923575924, + "learning_rate": 1.4075182735985913e-05, + "loss": 1.5716, + "step": 3550 + }, + { + "epoch": 1.9277958740499457, + "grad_norm": 10.087922707091717, + "learning_rate": 1.4071970916647432e-05, + "loss": 0.7026, + "step": 3551 + }, + { + "epoch": 1.9283387622149837, + "grad_norm": 8.507872704310849, + "learning_rate": 1.40687585936811e-05, + "loss": 0.7016, + "step": 3552 + }, + { + "epoch": 1.9288816503800217, + "grad_norm": 8.98138340766033, + "learning_rate": 1.4065545767484218e-05, + "loss": 0.6834, + "step": 3553 + }, + { + "epoch": 1.9294245385450597, + "grad_norm": 7.1085330003672045, + "learning_rate": 1.4062332438454156e-05, + "loss": 0.6813, + "step": 3554 + }, + { + "epoch": 1.9299674267100977, + "grad_norm": 13.499346110936985, + "learning_rate": 1.4059118606988345e-05, + "loss": 1.0068, + "step": 3555 + }, + { + "epoch": 1.9305103148751357, + "grad_norm": 7.698407375447584, + "learning_rate": 1.4055904273484275e-05, + "loss": 0.8418, + "step": 3556 + }, + { + "epoch": 1.9310532030401737, + "grad_norm": 8.424294783097983, + "learning_rate": 1.40526894383395e-05, + "loss": 0.8087, + "step": 3557 + }, + { + "epoch": 1.9315960912052117, + "grad_norm": 8.962780671623113, + "learning_rate": 1.4049474101951639e-05, + "loss": 0.8644, + "step": 3558 + }, + { + "epoch": 1.9321389793702497, + "grad_norm": 8.580581507909985, + "learning_rate": 1.4046258264718363e-05, + "loss": 1.0325, + "step": 3559 + }, + { + "epoch": 1.9326818675352877, + "grad_norm": 10.417878574145533, + "learning_rate": 1.4043041927037418e-05, + "loss": 0.7705, + "step": 3560 + }, + { + "epoch": 1.9332247557003257, + "grad_norm": 8.250628073154727, + "learning_rate": 1.4039825089306603e-05, + "loss": 0.794, + "step": 3561 + }, + { + "epoch": 1.9337676438653637, + "grad_norm": 8.228275233396532, + "learning_rate": 1.4036607751923783e-05, + "loss": 0.8759, + "step": 3562 + }, + { + "epoch": 1.9343105320304017, + "grad_norm": 7.760702739983814, + "learning_rate": 1.4033389915286884e-05, + "loss": 0.6458, + "step": 3563 + }, + { + "epoch": 1.9348534201954397, + "grad_norm": 6.508974748032192, + "learning_rate": 1.403017157979389e-05, + "loss": 0.6348, + "step": 3564 + }, + { + "epoch": 1.9353963083604777, + "grad_norm": 11.570824985629717, + "learning_rate": 1.4026952745842853e-05, + "loss": 1.0935, + "step": 3565 + }, + { + "epoch": 1.9359391965255157, + "grad_norm": 9.901046454739335, + "learning_rate": 1.4023733413831881e-05, + "loss": 0.7458, + "step": 3566 + }, + { + "epoch": 1.9364820846905537, + "grad_norm": 9.103951624002182, + "learning_rate": 1.402051358415915e-05, + "loss": 0.8253, + "step": 3567 + }, + { + "epoch": 1.9370249728555917, + "grad_norm": 9.293062225974824, + "learning_rate": 1.4017293257222887e-05, + "loss": 0.5801, + "step": 3568 + }, + { + "epoch": 1.9375678610206297, + "grad_norm": 7.180133735690273, + "learning_rate": 1.4014072433421398e-05, + "loss": 0.6063, + "step": 3569 + }, + { + "epoch": 1.9381107491856677, + "grad_norm": 9.028614205338394, + "learning_rate": 1.4010851113153028e-05, + "loss": 0.9412, + "step": 3570 + }, + { + "epoch": 1.9386536373507057, + "grad_norm": 9.51192022384654, + "learning_rate": 1.4007629296816202e-05, + "loss": 0.6459, + "step": 3571 + }, + { + "epoch": 1.9391965255157437, + "grad_norm": 9.883078787028563, + "learning_rate": 1.4004406984809396e-05, + "loss": 1.0324, + "step": 3572 + }, + { + "epoch": 1.9397394136807817, + "grad_norm": 7.433318187249177, + "learning_rate": 1.4001184177531154e-05, + "loss": 0.6152, + "step": 3573 + }, + { + "epoch": 1.9402823018458197, + "grad_norm": 10.165800890508075, + "learning_rate": 1.399796087538007e-05, + "loss": 0.8751, + "step": 3574 + }, + { + "epoch": 1.9408251900108577, + "grad_norm": 7.668138219396951, + "learning_rate": 1.3994737078754819e-05, + "loss": 0.43, + "step": 3575 + }, + { + "epoch": 1.9413680781758957, + "grad_norm": 8.929591410847367, + "learning_rate": 1.3991512788054115e-05, + "loss": 0.6173, + "step": 3576 + }, + { + "epoch": 1.9419109663409337, + "grad_norm": 8.069899029370944, + "learning_rate": 1.3988288003676755e-05, + "loss": 0.6782, + "step": 3577 + }, + { + "epoch": 1.9424538545059717, + "grad_norm": 7.635775018643444, + "learning_rate": 1.3985062726021574e-05, + "loss": 0.7503, + "step": 3578 + }, + { + "epoch": 1.9429967426710097, + "grad_norm": 11.649932328973133, + "learning_rate": 1.3981836955487485e-05, + "loss": 0.9154, + "step": 3579 + }, + { + "epoch": 1.9435396308360477, + "grad_norm": 9.11865472044189, + "learning_rate": 1.397861069247345e-05, + "loss": 0.8799, + "step": 3580 + }, + { + "epoch": 1.9440825190010858, + "grad_norm": 7.628563459317163, + "learning_rate": 1.3975383937378508e-05, + "loss": 0.7199, + "step": 3581 + }, + { + "epoch": 1.9446254071661238, + "grad_norm": 10.640127088045624, + "learning_rate": 1.3972156690601747e-05, + "loss": 0.7885, + "step": 3582 + }, + { + "epoch": 1.9451682953311618, + "grad_norm": 10.192406733466123, + "learning_rate": 1.3968928952542313e-05, + "loss": 0.6036, + "step": 3583 + }, + { + "epoch": 1.9457111834961998, + "grad_norm": 12.020287450972278, + "learning_rate": 1.396570072359942e-05, + "loss": 0.9477, + "step": 3584 + }, + { + "epoch": 1.9462540716612378, + "grad_norm": 8.855072111194623, + "learning_rate": 1.3962472004172343e-05, + "loss": 0.4384, + "step": 3585 + }, + { + "epoch": 1.9467969598262758, + "grad_norm": 7.173236338345466, + "learning_rate": 1.3959242794660412e-05, + "loss": 0.4969, + "step": 3586 + }, + { + "epoch": 1.9473398479913138, + "grad_norm": 6.505005692892076, + "learning_rate": 1.3956013095463024e-05, + "loss": 0.4585, + "step": 3587 + }, + { + "epoch": 1.9478827361563518, + "grad_norm": 7.041517208688192, + "learning_rate": 1.395278290697963e-05, + "loss": 0.5439, + "step": 3588 + }, + { + "epoch": 1.9484256243213898, + "grad_norm": 9.730465359550896, + "learning_rate": 1.3949552229609746e-05, + "loss": 0.8865, + "step": 3589 + }, + { + "epoch": 1.9489685124864278, + "grad_norm": 7.73678668778, + "learning_rate": 1.3946321063752948e-05, + "loss": 0.5094, + "step": 3590 + }, + { + "epoch": 1.9495114006514658, + "grad_norm": 7.095011334397053, + "learning_rate": 1.3943089409808872e-05, + "loss": 0.5224, + "step": 3591 + }, + { + "epoch": 1.9500542888165038, + "grad_norm": 8.086388468165458, + "learning_rate": 1.393985726817721e-05, + "loss": 0.7798, + "step": 3592 + }, + { + "epoch": 1.9505971769815418, + "grad_norm": 10.262604564335476, + "learning_rate": 1.3936624639257726e-05, + "loss": 1.2595, + "step": 3593 + }, + { + "epoch": 1.9511400651465798, + "grad_norm": 8.535907967530573, + "learning_rate": 1.393339152345023e-05, + "loss": 0.8205, + "step": 3594 + }, + { + "epoch": 1.9516829533116178, + "grad_norm": 7.864477340620307, + "learning_rate": 1.3930157921154601e-05, + "loss": 0.5967, + "step": 3595 + }, + { + "epoch": 1.9522258414766558, + "grad_norm": 6.739275201436143, + "learning_rate": 1.392692383277078e-05, + "loss": 0.5836, + "step": 3596 + }, + { + "epoch": 1.9527687296416938, + "grad_norm": 9.533970605538912, + "learning_rate": 1.392368925869876e-05, + "loss": 0.5761, + "step": 3597 + }, + { + "epoch": 1.9533116178067318, + "grad_norm": 11.376343280359167, + "learning_rate": 1.3920454199338598e-05, + "loss": 0.7596, + "step": 3598 + }, + { + "epoch": 1.9538545059717698, + "grad_norm": 7.280453649201749, + "learning_rate": 1.391721865509041e-05, + "loss": 0.4949, + "step": 3599 + }, + { + "epoch": 1.9543973941368078, + "grad_norm": 7.734190007799969, + "learning_rate": 1.391398262635438e-05, + "loss": 0.9177, + "step": 3600 + }, + { + "epoch": 1.9549402823018458, + "grad_norm": 13.106124045444247, + "learning_rate": 1.3910746113530738e-05, + "loss": 1.1884, + "step": 3601 + }, + { + "epoch": 1.9554831704668838, + "grad_norm": 11.868912051178873, + "learning_rate": 1.3907509117019783e-05, + "loss": 0.9245, + "step": 3602 + }, + { + "epoch": 1.9560260586319218, + "grad_norm": 10.307498071834987, + "learning_rate": 1.3904271637221876e-05, + "loss": 1.0802, + "step": 3603 + }, + { + "epoch": 1.9565689467969598, + "grad_norm": 8.308697157532139, + "learning_rate": 1.390103367453743e-05, + "loss": 0.4979, + "step": 3604 + }, + { + "epoch": 1.9571118349619978, + "grad_norm": 8.447604525372762, + "learning_rate": 1.3897795229366919e-05, + "loss": 0.8942, + "step": 3605 + }, + { + "epoch": 1.9576547231270358, + "grad_norm": 7.864013969756676, + "learning_rate": 1.3894556302110883e-05, + "loss": 0.642, + "step": 3606 + }, + { + "epoch": 1.9581976112920738, + "grad_norm": 7.903822544650853, + "learning_rate": 1.389131689316992e-05, + "loss": 0.8251, + "step": 3607 + }, + { + "epoch": 1.9587404994571118, + "grad_norm": 9.744275276467558, + "learning_rate": 1.3888077002944678e-05, + "loss": 1.0052, + "step": 3608 + }, + { + "epoch": 1.9592833876221498, + "grad_norm": 7.731164092689308, + "learning_rate": 1.3884836631835877e-05, + "loss": 0.6883, + "step": 3609 + }, + { + "epoch": 1.9598262757871878, + "grad_norm": 8.849291385162568, + "learning_rate": 1.3881595780244288e-05, + "loss": 0.9985, + "step": 3610 + }, + { + "epoch": 1.9603691639522258, + "grad_norm": 9.531929458494986, + "learning_rate": 1.3878354448570748e-05, + "loss": 0.7643, + "step": 3611 + }, + { + "epoch": 1.9609120521172638, + "grad_norm": 10.32443732773501, + "learning_rate": 1.3875112637216145e-05, + "loss": 0.848, + "step": 3612 + }, + { + "epoch": 1.9614549402823018, + "grad_norm": 7.417155617136626, + "learning_rate": 1.387187034658144e-05, + "loss": 0.5682, + "step": 3613 + }, + { + "epoch": 1.9619978284473398, + "grad_norm": 14.072831276291689, + "learning_rate": 1.386862757706764e-05, + "loss": 0.8279, + "step": 3614 + }, + { + "epoch": 1.9625407166123778, + "grad_norm": 7.764451928370161, + "learning_rate": 1.3865384329075812e-05, + "loss": 0.6687, + "step": 3615 + }, + { + "epoch": 1.9630836047774158, + "grad_norm": 8.686094567618806, + "learning_rate": 1.3862140603007095e-05, + "loss": 1.1034, + "step": 3616 + }, + { + "epoch": 1.9636264929424538, + "grad_norm": 10.753407337858437, + "learning_rate": 1.3858896399262669e-05, + "loss": 1.1692, + "step": 3617 + }, + { + "epoch": 1.9641693811074918, + "grad_norm": 9.096005648867607, + "learning_rate": 1.3855651718243786e-05, + "loss": 0.6191, + "step": 3618 + }, + { + "epoch": 1.9647122692725298, + "grad_norm": 10.387647049254962, + "learning_rate": 1.3852406560351752e-05, + "loss": 1.3573, + "step": 3619 + }, + { + "epoch": 1.9652551574375678, + "grad_norm": 7.173958455599766, + "learning_rate": 1.3849160925987936e-05, + "loss": 0.8106, + "step": 3620 + }, + { + "epoch": 1.9657980456026058, + "grad_norm": 9.434903281870886, + "learning_rate": 1.3845914815553765e-05, + "loss": 0.9581, + "step": 3621 + }, + { + "epoch": 1.9663409337676439, + "grad_norm": 7.671221453670278, + "learning_rate": 1.3842668229450717e-05, + "loss": 0.6989, + "step": 3622 + }, + { + "epoch": 1.9668838219326819, + "grad_norm": 7.360534692664803, + "learning_rate": 1.3839421168080338e-05, + "loss": 0.4912, + "step": 3623 + }, + { + "epoch": 1.9674267100977199, + "grad_norm": 6.853527063108637, + "learning_rate": 1.3836173631844231e-05, + "loss": 0.7485, + "step": 3624 + }, + { + "epoch": 1.9679695982627579, + "grad_norm": 8.619792011062271, + "learning_rate": 1.3832925621144057e-05, + "loss": 0.712, + "step": 3625 + }, + { + "epoch": 1.9685124864277959, + "grad_norm": 8.062990147577946, + "learning_rate": 1.3829677136381532e-05, + "loss": 0.7126, + "step": 3626 + }, + { + "epoch": 1.9690553745928339, + "grad_norm": 6.35400375341867, + "learning_rate": 1.3826428177958433e-05, + "loss": 0.5262, + "step": 3627 + }, + { + "epoch": 1.9695982627578719, + "grad_norm": 7.322594621519839, + "learning_rate": 1.3823178746276603e-05, + "loss": 0.6644, + "step": 3628 + }, + { + "epoch": 1.9701411509229099, + "grad_norm": 9.976906946653425, + "learning_rate": 1.3819928841737929e-05, + "loss": 0.7411, + "step": 3629 + }, + { + "epoch": 1.9706840390879479, + "grad_norm": 11.052922199516026, + "learning_rate": 1.3816678464744368e-05, + "loss": 0.9998, + "step": 3630 + }, + { + "epoch": 1.9712269272529859, + "grad_norm": 7.695988647358978, + "learning_rate": 1.381342761569793e-05, + "loss": 0.6805, + "step": 3631 + }, + { + "epoch": 1.9717698154180239, + "grad_norm": 10.387946538947885, + "learning_rate": 1.381017629500069e-05, + "loss": 0.8954, + "step": 3632 + }, + { + "epoch": 1.9723127035830619, + "grad_norm": 8.215553841008688, + "learning_rate": 1.380692450305477e-05, + "loss": 1.0488, + "step": 3633 + }, + { + "epoch": 1.9728555917480999, + "grad_norm": 8.550898170705977, + "learning_rate": 1.3803672240262364e-05, + "loss": 0.9301, + "step": 3634 + }, + { + "epoch": 1.9733984799131379, + "grad_norm": 9.978410770263547, + "learning_rate": 1.380041950702571e-05, + "loss": 0.9725, + "step": 3635 + }, + { + "epoch": 1.9739413680781759, + "grad_norm": 10.211160158432088, + "learning_rate": 1.3797166303747119e-05, + "loss": 1.1789, + "step": 3636 + }, + { + "epoch": 1.9744842562432139, + "grad_norm": 7.47091491709439, + "learning_rate": 1.3793912630828942e-05, + "loss": 1.2311, + "step": 3637 + }, + { + "epoch": 1.975027144408252, + "grad_norm": 9.966534074490724, + "learning_rate": 1.3790658488673607e-05, + "loss": 1.0057, + "step": 3638 + }, + { + "epoch": 1.97557003257329, + "grad_norm": 9.341517091004942, + "learning_rate": 1.378740387768359e-05, + "loss": 0.7708, + "step": 3639 + }, + { + "epoch": 1.976112920738328, + "grad_norm": 10.066755071795829, + "learning_rate": 1.3784148798261422e-05, + "loss": 1.3146, + "step": 3640 + }, + { + "epoch": 1.976655808903366, + "grad_norm": 8.306959980081418, + "learning_rate": 1.3780893250809705e-05, + "loss": 0.6926, + "step": 3641 + }, + { + "epoch": 1.977198697068404, + "grad_norm": 5.980548702977229, + "learning_rate": 1.377763723573108e-05, + "loss": 0.6776, + "step": 3642 + }, + { + "epoch": 1.977741585233442, + "grad_norm": 10.796897781225315, + "learning_rate": 1.3774380753428263e-05, + "loss": 1.0422, + "step": 3643 + }, + { + "epoch": 1.97828447339848, + "grad_norm": 9.650854393967004, + "learning_rate": 1.3771123804304018e-05, + "loss": 1.2566, + "step": 3644 + }, + { + "epoch": 1.978827361563518, + "grad_norm": 6.718663747121732, + "learning_rate": 1.3767866388761168e-05, + "loss": 0.5018, + "step": 3645 + }, + { + "epoch": 1.979370249728556, + "grad_norm": 8.345918705516796, + "learning_rate": 1.3764608507202604e-05, + "loss": 0.8501, + "step": 3646 + }, + { + "epoch": 1.979913137893594, + "grad_norm": 11.065999609001787, + "learning_rate": 1.3761350160031258e-05, + "loss": 1.2627, + "step": 3647 + }, + { + "epoch": 1.980456026058632, + "grad_norm": 8.567448073304528, + "learning_rate": 1.3758091347650126e-05, + "loss": 0.4477, + "step": 3648 + }, + { + "epoch": 1.98099891422367, + "grad_norm": 11.930236219703469, + "learning_rate": 1.3754832070462269e-05, + "loss": 1.5386, + "step": 3649 + }, + { + "epoch": 1.981541802388708, + "grad_norm": 7.334841489157127, + "learning_rate": 1.3751572328870797e-05, + "loss": 0.5759, + "step": 3650 + }, + { + "epoch": 1.982084690553746, + "grad_norm": 6.392865090419794, + "learning_rate": 1.3748312123278879e-05, + "loss": 0.6078, + "step": 3651 + }, + { + "epoch": 1.982627578718784, + "grad_norm": 8.598388892264492, + "learning_rate": 1.3745051454089744e-05, + "loss": 0.7682, + "step": 3652 + }, + { + "epoch": 1.983170466883822, + "grad_norm": 9.310542957266444, + "learning_rate": 1.3741790321706678e-05, + "loss": 0.9833, + "step": 3653 + }, + { + "epoch": 1.98371335504886, + "grad_norm": 8.917570196452148, + "learning_rate": 1.3738528726533021e-05, + "loss": 0.5915, + "step": 3654 + }, + { + "epoch": 1.984256243213898, + "grad_norm": 9.503202436394638, + "learning_rate": 1.3735266668972174e-05, + "loss": 0.9347, + "step": 3655 + }, + { + "epoch": 1.984799131378936, + "grad_norm": 9.187024367480415, + "learning_rate": 1.3732004149427592e-05, + "loss": 0.6669, + "step": 3656 + }, + { + "epoch": 1.985342019543974, + "grad_norm": 10.497812893691144, + "learning_rate": 1.3728741168302785e-05, + "loss": 1.0663, + "step": 3657 + }, + { + "epoch": 1.985884907709012, + "grad_norm": 9.646030688042543, + "learning_rate": 1.3725477726001332e-05, + "loss": 0.9913, + "step": 3658 + }, + { + "epoch": 1.98642779587405, + "grad_norm": 9.475431343889321, + "learning_rate": 1.3722213822926855e-05, + "loss": 0.8377, + "step": 3659 + }, + { + "epoch": 1.986970684039088, + "grad_norm": 9.449591675048769, + "learning_rate": 1.371894945948304e-05, + "loss": 0.8763, + "step": 3660 + }, + { + "epoch": 1.987513572204126, + "grad_norm": 7.3204142422253025, + "learning_rate": 1.3715684636073628e-05, + "loss": 0.7384, + "step": 3661 + }, + { + "epoch": 1.988056460369164, + "grad_norm": 10.339298014193695, + "learning_rate": 1.371241935310242e-05, + "loss": 1.0206, + "step": 3662 + }, + { + "epoch": 1.988599348534202, + "grad_norm": 7.083819102334954, + "learning_rate": 1.3709153610973266e-05, + "loss": 0.6905, + "step": 3663 + }, + { + "epoch": 1.98914223669924, + "grad_norm": 6.318629816327554, + "learning_rate": 1.3705887410090085e-05, + "loss": 0.7946, + "step": 3664 + }, + { + "epoch": 1.989685124864278, + "grad_norm": 7.109392074745231, + "learning_rate": 1.3702620750856843e-05, + "loss": 0.7125, + "step": 3665 + }, + { + "epoch": 1.990228013029316, + "grad_norm": 6.7832401368423545, + "learning_rate": 1.3699353633677565e-05, + "loss": 0.6091, + "step": 3666 + }, + { + "epoch": 1.990770901194354, + "grad_norm": 9.301471508167843, + "learning_rate": 1.3696086058956333e-05, + "loss": 0.9335, + "step": 3667 + }, + { + "epoch": 1.991313789359392, + "grad_norm": 8.167349191349723, + "learning_rate": 1.3692818027097288e-05, + "loss": 0.7148, + "step": 3668 + }, + { + "epoch": 1.99185667752443, + "grad_norm": 7.582331201930941, + "learning_rate": 1.3689549538504622e-05, + "loss": 0.5952, + "step": 3669 + }, + { + "epoch": 1.992399565689468, + "grad_norm": 8.743965167551632, + "learning_rate": 1.3686280593582588e-05, + "loss": 0.9268, + "step": 3670 + }, + { + "epoch": 1.992942453854506, + "grad_norm": 10.291738284427778, + "learning_rate": 1.3683011192735496e-05, + "loss": 0.9934, + "step": 3671 + }, + { + "epoch": 1.993485342019544, + "grad_norm": 9.808936601876956, + "learning_rate": 1.3679741336367711e-05, + "loss": 1.0506, + "step": 3672 + }, + { + "epoch": 1.994028230184582, + "grad_norm": 10.002691567142412, + "learning_rate": 1.3676471024883654e-05, + "loss": 1.1439, + "step": 3673 + }, + { + "epoch": 1.99457111834962, + "grad_norm": 7.558103827510508, + "learning_rate": 1.36732002586878e-05, + "loss": 1.0059, + "step": 3674 + }, + { + "epoch": 1.995114006514658, + "grad_norm": 7.958890420546252, + "learning_rate": 1.3669929038184684e-05, + "loss": 1.0301, + "step": 3675 + }, + { + "epoch": 1.995656894679696, + "grad_norm": 8.042682701022132, + "learning_rate": 1.3666657363778895e-05, + "loss": 0.6881, + "step": 3676 + }, + { + "epoch": 1.996199782844734, + "grad_norm": 7.54271762136155, + "learning_rate": 1.366338523587508e-05, + "loss": 0.5227, + "step": 3677 + }, + { + "epoch": 1.996742671009772, + "grad_norm": 9.060117135659972, + "learning_rate": 1.3660112654877939e-05, + "loss": 0.9847, + "step": 3678 + }, + { + "epoch": 1.99728555917481, + "grad_norm": 11.00270077841202, + "learning_rate": 1.3656839621192233e-05, + "loss": 1.0918, + "step": 3679 + }, + { + "epoch": 1.997828447339848, + "grad_norm": 9.156434005073692, + "learning_rate": 1.3653566135222774e-05, + "loss": 0.8104, + "step": 3680 + }, + { + "epoch": 1.998371335504886, + "grad_norm": 8.405413514743108, + "learning_rate": 1.3650292197374433e-05, + "loss": 0.9414, + "step": 3681 + }, + { + "epoch": 1.998914223669924, + "grad_norm": 7.9594907003419495, + "learning_rate": 1.3647017808052135e-05, + "loss": 0.9257, + "step": 3682 + }, + { + "epoch": 1.999457111834962, + "grad_norm": 9.821777823228977, + "learning_rate": 1.3643742967660859e-05, + "loss": 0.8787, + "step": 3683 + }, + { + "epoch": 2.0, + "grad_norm": 8.766507481527489, + "learning_rate": 1.3640467676605648e-05, + "loss": 0.8472, + "step": 3684 + }, + { + "epoch": 2.000542888165038, + "grad_norm": 9.493942149196295, + "learning_rate": 1.3637191935291596e-05, + "loss": 0.8147, + "step": 3685 + }, + { + "epoch": 2.001085776330076, + "grad_norm": 9.782638508357579, + "learning_rate": 1.3633915744123844e-05, + "loss": 0.8788, + "step": 3686 + }, + { + "epoch": 2.001628664495114, + "grad_norm": 8.46210638839138, + "learning_rate": 1.3630639103507604e-05, + "loss": 1.078, + "step": 3687 + }, + { + "epoch": 2.002171552660152, + "grad_norm": 7.345658753018104, + "learning_rate": 1.3627362013848134e-05, + "loss": 0.6838, + "step": 3688 + }, + { + "epoch": 2.00271444082519, + "grad_norm": 7.329182601936008, + "learning_rate": 1.3624084475550743e-05, + "loss": 0.7249, + "step": 3689 + }, + { + "epoch": 2.003257328990228, + "grad_norm": 9.89510466995638, + "learning_rate": 1.3620806489020813e-05, + "loss": 0.8018, + "step": 3690 + }, + { + "epoch": 2.003800217155266, + "grad_norm": 6.836942603575798, + "learning_rate": 1.3617528054663764e-05, + "loss": 0.6645, + "step": 3691 + }, + { + "epoch": 2.004343105320304, + "grad_norm": 8.842045073299985, + "learning_rate": 1.3614249172885081e-05, + "loss": 0.5143, + "step": 3692 + }, + { + "epoch": 2.004885993485342, + "grad_norm": 7.671200527248645, + "learning_rate": 1.36109698440903e-05, + "loss": 0.8038, + "step": 3693 + }, + { + "epoch": 2.00542888165038, + "grad_norm": 9.544464428119456, + "learning_rate": 1.3607690068685013e-05, + "loss": 0.6396, + "step": 3694 + }, + { + "epoch": 2.005971769815418, + "grad_norm": 8.537202930183216, + "learning_rate": 1.3604409847074868e-05, + "loss": 0.7104, + "step": 3695 + }, + { + "epoch": 2.006514657980456, + "grad_norm": 7.121091005738386, + "learning_rate": 1.3601129179665572e-05, + "loss": 0.4337, + "step": 3696 + }, + { + "epoch": 2.007057546145494, + "grad_norm": 9.740494539210806, + "learning_rate": 1.3597848066862875e-05, + "loss": 0.9114, + "step": 3697 + }, + { + "epoch": 2.007600434310532, + "grad_norm": 8.100951122902112, + "learning_rate": 1.3594566509072599e-05, + "loss": 0.722, + "step": 3698 + }, + { + "epoch": 2.00814332247557, + "grad_norm": 6.959347625774498, + "learning_rate": 1.3591284506700606e-05, + "loss": 0.5429, + "step": 3699 + }, + { + "epoch": 2.008686210640608, + "grad_norm": 9.993493206941414, + "learning_rate": 1.3588002060152822e-05, + "loss": 0.8251, + "step": 3700 + }, + { + "epoch": 2.009229098805646, + "grad_norm": 9.966526343412045, + "learning_rate": 1.3584719169835226e-05, + "loss": 0.4847, + "step": 3701 + }, + { + "epoch": 2.009771986970684, + "grad_norm": 7.888503709599375, + "learning_rate": 1.3581435836153847e-05, + "loss": 0.7941, + "step": 3702 + }, + { + "epoch": 2.010314875135722, + "grad_norm": 9.057972693252164, + "learning_rate": 1.3578152059514778e-05, + "loss": 0.8963, + "step": 3703 + }, + { + "epoch": 2.01085776330076, + "grad_norm": 10.717910066603423, + "learning_rate": 1.3574867840324157e-05, + "loss": 0.6485, + "step": 3704 + }, + { + "epoch": 2.011400651465798, + "grad_norm": 7.257931643777971, + "learning_rate": 1.3571583178988188e-05, + "loss": 0.4569, + "step": 3705 + }, + { + "epoch": 2.011943539630836, + "grad_norm": 16.546104792345954, + "learning_rate": 1.3568298075913119e-05, + "loss": 1.0473, + "step": 3706 + }, + { + "epoch": 2.012486427795874, + "grad_norm": 9.882729561546991, + "learning_rate": 1.3565012531505252e-05, + "loss": 0.9729, + "step": 3707 + }, + { + "epoch": 2.013029315960912, + "grad_norm": 7.060037163975289, + "learning_rate": 1.3561726546170956e-05, + "loss": 0.5935, + "step": 3708 + }, + { + "epoch": 2.01357220412595, + "grad_norm": 7.081117948590212, + "learning_rate": 1.3558440120316644e-05, + "loss": 0.6676, + "step": 3709 + }, + { + "epoch": 2.014115092290988, + "grad_norm": 10.048606586375149, + "learning_rate": 1.3555153254348788e-05, + "loss": 0.6307, + "step": 3710 + }, + { + "epoch": 2.014657980456026, + "grad_norm": 7.720688031890385, + "learning_rate": 1.3551865948673912e-05, + "loss": 0.7882, + "step": 3711 + }, + { + "epoch": 2.015200868621064, + "grad_norm": 7.644495073412038, + "learning_rate": 1.3548578203698592e-05, + "loss": 0.8904, + "step": 3712 + }, + { + "epoch": 2.015743756786102, + "grad_norm": 10.13216917020047, + "learning_rate": 1.3545290019829466e-05, + "loss": 0.7832, + "step": 3713 + }, + { + "epoch": 2.01628664495114, + "grad_norm": 6.321686691568546, + "learning_rate": 1.3542001397473219e-05, + "loss": 0.5682, + "step": 3714 + }, + { + "epoch": 2.016829533116178, + "grad_norm": 7.966895819853924, + "learning_rate": 1.3538712337036594e-05, + "loss": 0.7704, + "step": 3715 + }, + { + "epoch": 2.017372421281216, + "grad_norm": 8.827713158118389, + "learning_rate": 1.3535422838926389e-05, + "loss": 0.6741, + "step": 3716 + }, + { + "epoch": 2.017915309446254, + "grad_norm": 7.404288938912365, + "learning_rate": 1.3532132903549453e-05, + "loss": 0.5325, + "step": 3717 + }, + { + "epoch": 2.018458197611292, + "grad_norm": 9.869324924245625, + "learning_rate": 1.352884253131269e-05, + "loss": 0.7326, + "step": 3718 + }, + { + "epoch": 2.01900108577633, + "grad_norm": 7.824783976445888, + "learning_rate": 1.3525551722623056e-05, + "loss": 0.4683, + "step": 3719 + }, + { + "epoch": 2.019543973941368, + "grad_norm": 10.516375854488775, + "learning_rate": 1.3522260477887566e-05, + "loss": 0.839, + "step": 3720 + }, + { + "epoch": 2.020086862106406, + "grad_norm": 10.698678918246005, + "learning_rate": 1.3518968797513288e-05, + "loss": 1.0513, + "step": 3721 + }, + { + "epoch": 2.020629750271444, + "grad_norm": 9.430167728870957, + "learning_rate": 1.351567668190734e-05, + "loss": 0.6696, + "step": 3722 + }, + { + "epoch": 2.021172638436482, + "grad_norm": 9.619270238605017, + "learning_rate": 1.3512384131476897e-05, + "loss": 0.4608, + "step": 3723 + }, + { + "epoch": 2.02171552660152, + "grad_norm": 6.9328636261250525, + "learning_rate": 1.350909114662919e-05, + "loss": 0.3952, + "step": 3724 + }, + { + "epoch": 2.022258414766558, + "grad_norm": 11.694555068035188, + "learning_rate": 1.3505797727771493e-05, + "loss": 0.7526, + "step": 3725 + }, + { + "epoch": 2.022801302931596, + "grad_norm": 7.417427417334792, + "learning_rate": 1.3502503875311149e-05, + "loss": 0.5239, + "step": 3726 + }, + { + "epoch": 2.023344191096634, + "grad_norm": 10.99931861711161, + "learning_rate": 1.349920958965554e-05, + "loss": 0.5504, + "step": 3727 + }, + { + "epoch": 2.023887079261672, + "grad_norm": 9.418315087755559, + "learning_rate": 1.3495914871212113e-05, + "loss": 0.6791, + "step": 3728 + }, + { + "epoch": 2.02442996742671, + "grad_norm": 12.589472190346669, + "learning_rate": 1.3492619720388363e-05, + "loss": 0.815, + "step": 3729 + }, + { + "epoch": 2.024972855591748, + "grad_norm": 6.60691672927324, + "learning_rate": 1.348932413759184e-05, + "loss": 0.4112, + "step": 3730 + }, + { + "epoch": 2.025515743756786, + "grad_norm": 10.970405529985833, + "learning_rate": 1.3486028123230145e-05, + "loss": 0.6164, + "step": 3731 + }, + { + "epoch": 2.026058631921824, + "grad_norm": 9.419473243249211, + "learning_rate": 1.3482731677710938e-05, + "loss": 0.829, + "step": 3732 + }, + { + "epoch": 2.026601520086862, + "grad_norm": 8.809487998193466, + "learning_rate": 1.3479434801441925e-05, + "loss": 0.6386, + "step": 3733 + }, + { + "epoch": 2.0271444082519, + "grad_norm": 8.586869223296738, + "learning_rate": 1.3476137494830872e-05, + "loss": 0.4586, + "step": 3734 + }, + { + "epoch": 2.027687296416938, + "grad_norm": 10.792985384487755, + "learning_rate": 1.3472839758285595e-05, + "loss": 0.68, + "step": 3735 + }, + { + "epoch": 2.028230184581976, + "grad_norm": 11.494014730485675, + "learning_rate": 1.346954159221396e-05, + "loss": 0.9465, + "step": 3736 + }, + { + "epoch": 2.028773072747014, + "grad_norm": 9.047474116723855, + "learning_rate": 1.3466242997023891e-05, + "loss": 0.637, + "step": 3737 + }, + { + "epoch": 2.029315960912052, + "grad_norm": 7.32591412143244, + "learning_rate": 1.3462943973123362e-05, + "loss": 0.5071, + "step": 3738 + }, + { + "epoch": 2.02985884907709, + "grad_norm": 8.462579835722188, + "learning_rate": 1.3459644520920405e-05, + "loss": 0.5568, + "step": 3739 + }, + { + "epoch": 2.030401737242128, + "grad_norm": 11.667085873745426, + "learning_rate": 1.34563446408231e-05, + "loss": 0.579, + "step": 3740 + }, + { + "epoch": 2.030944625407166, + "grad_norm": 6.778347309469529, + "learning_rate": 1.3453044333239577e-05, + "loss": 0.3399, + "step": 3741 + }, + { + "epoch": 2.031487513572204, + "grad_norm": 7.280157509234148, + "learning_rate": 1.3449743598578033e-05, + "loss": 0.4236, + "step": 3742 + }, + { + "epoch": 2.032030401737242, + "grad_norm": 10.902366947251025, + "learning_rate": 1.34464424372467e-05, + "loss": 0.7984, + "step": 3743 + }, + { + "epoch": 2.03257328990228, + "grad_norm": 7.729938792603154, + "learning_rate": 1.344314084965388e-05, + "loss": 0.5115, + "step": 3744 + }, + { + "epoch": 2.033116178067318, + "grad_norm": 9.034690624481534, + "learning_rate": 1.3439838836207905e-05, + "loss": 0.3528, + "step": 3745 + }, + { + "epoch": 2.033659066232356, + "grad_norm": 10.813371264470662, + "learning_rate": 1.3436536397317183e-05, + "loss": 0.4972, + "step": 3746 + }, + { + "epoch": 2.034201954397394, + "grad_norm": 12.116428074423327, + "learning_rate": 1.343323353339016e-05, + "loss": 0.9414, + "step": 3747 + }, + { + "epoch": 2.034744842562432, + "grad_norm": 11.714995187138708, + "learning_rate": 1.3429930244835343e-05, + "loss": 0.7842, + "step": 3748 + }, + { + "epoch": 2.03528773072747, + "grad_norm": 9.956102297397003, + "learning_rate": 1.3426626532061287e-05, + "loss": 0.7304, + "step": 3749 + }, + { + "epoch": 2.035830618892508, + "grad_norm": 10.670014393137768, + "learning_rate": 1.34233223954766e-05, + "loss": 0.746, + "step": 3750 + }, + { + "epoch": 2.036373507057546, + "grad_norm": 10.266437034084714, + "learning_rate": 1.3420017835489945e-05, + "loss": 1.0871, + "step": 3751 + }, + { + "epoch": 2.036916395222584, + "grad_norm": 11.75190588784515, + "learning_rate": 1.3416712852510033e-05, + "loss": 0.7606, + "step": 3752 + }, + { + "epoch": 2.037459283387622, + "grad_norm": 11.54288048263274, + "learning_rate": 1.3413407446945627e-05, + "loss": 0.5887, + "step": 3753 + }, + { + "epoch": 2.03800217155266, + "grad_norm": 18.002644323614724, + "learning_rate": 1.3410101619205552e-05, + "loss": 1.0593, + "step": 3754 + }, + { + "epoch": 2.038545059717698, + "grad_norm": 8.97759172274721, + "learning_rate": 1.3406795369698671e-05, + "loss": 0.7194, + "step": 3755 + }, + { + "epoch": 2.039087947882736, + "grad_norm": 8.793507308104033, + "learning_rate": 1.3403488698833912e-05, + "loss": 0.4891, + "step": 3756 + }, + { + "epoch": 2.039630836047774, + "grad_norm": 9.263725202719241, + "learning_rate": 1.3400181607020243e-05, + "loss": 0.8258, + "step": 3757 + }, + { + "epoch": 2.040173724212812, + "grad_norm": 10.568214203997627, + "learning_rate": 1.3396874094666694e-05, + "loss": 0.5841, + "step": 3758 + }, + { + "epoch": 2.04071661237785, + "grad_norm": 10.237711379932081, + "learning_rate": 1.3393566162182346e-05, + "loss": 0.6901, + "step": 3759 + }, + { + "epoch": 2.041259500542888, + "grad_norm": 11.264856346485661, + "learning_rate": 1.3390257809976322e-05, + "loss": 0.8029, + "step": 3760 + }, + { + "epoch": 2.041802388707926, + "grad_norm": 7.504173750209935, + "learning_rate": 1.3386949038457813e-05, + "loss": 0.3739, + "step": 3761 + }, + { + "epoch": 2.042345276872964, + "grad_norm": 8.099570741006275, + "learning_rate": 1.3383639848036044e-05, + "loss": 0.7457, + "step": 3762 + }, + { + "epoch": 2.042888165038002, + "grad_norm": 9.87563463161881, + "learning_rate": 1.3380330239120313e-05, + "loss": 0.632, + "step": 3763 + }, + { + "epoch": 2.04343105320304, + "grad_norm": 9.304965814712034, + "learning_rate": 1.3377020212119946e-05, + "loss": 0.657, + "step": 3764 + }, + { + "epoch": 2.043973941368078, + "grad_norm": 7.738549848691211, + "learning_rate": 1.3373709767444339e-05, + "loss": 0.4198, + "step": 3765 + }, + { + "epoch": 2.044516829533116, + "grad_norm": 11.805824611419341, + "learning_rate": 1.3370398905502928e-05, + "loss": 0.8173, + "step": 3766 + }, + { + "epoch": 2.045059717698154, + "grad_norm": 10.425380865754251, + "learning_rate": 1.3367087626705211e-05, + "loss": 0.7149, + "step": 3767 + }, + { + "epoch": 2.045602605863192, + "grad_norm": 11.884153232685467, + "learning_rate": 1.336377593146073e-05, + "loss": 0.723, + "step": 3768 + }, + { + "epoch": 2.04614549402823, + "grad_norm": 9.663581488842127, + "learning_rate": 1.336046382017908e-05, + "loss": 0.6457, + "step": 3769 + }, + { + "epoch": 2.046688382193268, + "grad_norm": 9.184284700313228, + "learning_rate": 1.335715129326991e-05, + "loss": 0.6645, + "step": 3770 + }, + { + "epoch": 2.047231270358306, + "grad_norm": 7.398657595562234, + "learning_rate": 1.3353838351142915e-05, + "loss": 0.7288, + "step": 3771 + }, + { + "epoch": 2.047774158523344, + "grad_norm": 10.416946087073072, + "learning_rate": 1.335052499420785e-05, + "loss": 0.4974, + "step": 3772 + }, + { + "epoch": 2.048317046688382, + "grad_norm": 8.676282075018099, + "learning_rate": 1.3347211222874514e-05, + "loss": 0.7128, + "step": 3773 + }, + { + "epoch": 2.04885993485342, + "grad_norm": 10.339910800835435, + "learning_rate": 1.3343897037552758e-05, + "loss": 0.6674, + "step": 3774 + }, + { + "epoch": 2.049402823018458, + "grad_norm": 10.976308424624598, + "learning_rate": 1.3340582438652488e-05, + "loss": 0.7683, + "step": 3775 + }, + { + "epoch": 2.049945711183496, + "grad_norm": 11.874329810000683, + "learning_rate": 1.3337267426583658e-05, + "loss": 1.2199, + "step": 3776 + }, + { + "epoch": 2.050488599348534, + "grad_norm": 9.306473243935873, + "learning_rate": 1.3333952001756272e-05, + "loss": 0.8681, + "step": 3777 + }, + { + "epoch": 2.0510314875135722, + "grad_norm": 9.083116427983501, + "learning_rate": 1.333063616458039e-05, + "loss": 0.7512, + "step": 3778 + }, + { + "epoch": 2.05157437567861, + "grad_norm": 5.762711234835306, + "learning_rate": 1.3327319915466119e-05, + "loss": 0.3384, + "step": 3779 + }, + { + "epoch": 2.0521172638436482, + "grad_norm": 7.431272084519616, + "learning_rate": 1.332400325482362e-05, + "loss": 0.5188, + "step": 3780 + }, + { + "epoch": 2.052660152008686, + "grad_norm": 5.736965263179954, + "learning_rate": 1.3320686183063096e-05, + "loss": 0.3476, + "step": 3781 + }, + { + "epoch": 2.0532030401737242, + "grad_norm": 7.130870597882685, + "learning_rate": 1.3317368700594815e-05, + "loss": 0.7929, + "step": 3782 + }, + { + "epoch": 2.053745928338762, + "grad_norm": 7.713657404988891, + "learning_rate": 1.3314050807829088e-05, + "loss": 0.422, + "step": 3783 + }, + { + "epoch": 2.0542888165038002, + "grad_norm": 6.383542293487279, + "learning_rate": 1.3310732505176276e-05, + "loss": 0.568, + "step": 3784 + }, + { + "epoch": 2.054831704668838, + "grad_norm": 12.326747003957852, + "learning_rate": 1.3307413793046787e-05, + "loss": 0.8061, + "step": 3785 + }, + { + "epoch": 2.0553745928338762, + "grad_norm": 10.668098962086818, + "learning_rate": 1.330409467185109e-05, + "loss": 1.4344, + "step": 3786 + }, + { + "epoch": 2.055917480998914, + "grad_norm": 9.844537720560512, + "learning_rate": 1.3300775141999698e-05, + "loss": 1.0089, + "step": 3787 + }, + { + "epoch": 2.0564603691639523, + "grad_norm": 9.714672703692647, + "learning_rate": 1.3297455203903176e-05, + "loss": 0.6069, + "step": 3788 + }, + { + "epoch": 2.05700325732899, + "grad_norm": 10.27882170416886, + "learning_rate": 1.3294134857972139e-05, + "loss": 0.7279, + "step": 3789 + }, + { + "epoch": 2.0575461454940283, + "grad_norm": 10.99914755541959, + "learning_rate": 1.3290814104617253e-05, + "loss": 0.6787, + "step": 3790 + }, + { + "epoch": 2.058089033659066, + "grad_norm": 13.627823006472678, + "learning_rate": 1.3287492944249234e-05, + "loss": 0.9052, + "step": 3791 + }, + { + "epoch": 2.0586319218241043, + "grad_norm": 7.734783593910536, + "learning_rate": 1.3284171377278849e-05, + "loss": 0.7141, + "step": 3792 + }, + { + "epoch": 2.059174809989142, + "grad_norm": 9.247896244835857, + "learning_rate": 1.3280849404116913e-05, + "loss": 0.6071, + "step": 3793 + }, + { + "epoch": 2.0597176981541803, + "grad_norm": 9.004013607687236, + "learning_rate": 1.3277527025174295e-05, + "loss": 0.5887, + "step": 3794 + }, + { + "epoch": 2.060260586319218, + "grad_norm": 9.62029412051341, + "learning_rate": 1.3274204240861908e-05, + "loss": 0.8847, + "step": 3795 + }, + { + "epoch": 2.0608034744842563, + "grad_norm": 7.500348953035017, + "learning_rate": 1.3270881051590725e-05, + "loss": 0.4769, + "step": 3796 + }, + { + "epoch": 2.061346362649294, + "grad_norm": 10.153144580789224, + "learning_rate": 1.326755745777176e-05, + "loss": 0.645, + "step": 3797 + }, + { + "epoch": 2.0618892508143323, + "grad_norm": 6.813155627292939, + "learning_rate": 1.326423345981608e-05, + "loss": 0.6711, + "step": 3798 + }, + { + "epoch": 2.06243213897937, + "grad_norm": 9.362898780009592, + "learning_rate": 1.32609090581348e-05, + "loss": 0.5464, + "step": 3799 + }, + { + "epoch": 2.0629750271444083, + "grad_norm": 8.540638734313639, + "learning_rate": 1.3257584253139096e-05, + "loss": 0.6523, + "step": 3800 + }, + { + "epoch": 2.063517915309446, + "grad_norm": 9.000753471311711, + "learning_rate": 1.3254259045240176e-05, + "loss": 0.5362, + "step": 3801 + }, + { + "epoch": 2.0640608034744843, + "grad_norm": 10.275669325762056, + "learning_rate": 1.3250933434849316e-05, + "loss": 0.6061, + "step": 3802 + }, + { + "epoch": 2.064603691639522, + "grad_norm": 11.615294601299743, + "learning_rate": 1.3247607422377823e-05, + "loss": 0.673, + "step": 3803 + }, + { + "epoch": 2.0651465798045603, + "grad_norm": 10.316289286330115, + "learning_rate": 1.324428100823707e-05, + "loss": 0.7997, + "step": 3804 + }, + { + "epoch": 2.065689467969598, + "grad_norm": 14.555740784089037, + "learning_rate": 1.3240954192838472e-05, + "loss": 0.9957, + "step": 3805 + }, + { + "epoch": 2.0662323561346363, + "grad_norm": 7.674873455902871, + "learning_rate": 1.323762697659349e-05, + "loss": 0.4771, + "step": 3806 + }, + { + "epoch": 2.066775244299674, + "grad_norm": 8.530555441214027, + "learning_rate": 1.3234299359913647e-05, + "loss": 0.8666, + "step": 3807 + }, + { + "epoch": 2.0673181324647123, + "grad_norm": 6.463731480876421, + "learning_rate": 1.3230971343210503e-05, + "loss": 0.5742, + "step": 3808 + }, + { + "epoch": 2.06786102062975, + "grad_norm": 8.074017879990592, + "learning_rate": 1.3227642926895676e-05, + "loss": 0.3637, + "step": 3809 + }, + { + "epoch": 2.0684039087947883, + "grad_norm": 9.989752857060338, + "learning_rate": 1.3224314111380828e-05, + "loss": 0.7389, + "step": 3810 + }, + { + "epoch": 2.068946796959826, + "grad_norm": 12.892065053280843, + "learning_rate": 1.3220984897077669e-05, + "loss": 1.2959, + "step": 3811 + }, + { + "epoch": 2.0694896851248643, + "grad_norm": 7.497956491794191, + "learning_rate": 1.3217655284397965e-05, + "loss": 0.4384, + "step": 3812 + }, + { + "epoch": 2.070032573289902, + "grad_norm": 14.719876212901973, + "learning_rate": 1.3214325273753528e-05, + "loss": 1.3054, + "step": 3813 + }, + { + "epoch": 2.0705754614549403, + "grad_norm": 10.767865599283661, + "learning_rate": 1.3210994865556219e-05, + "loss": 0.9961, + "step": 3814 + }, + { + "epoch": 2.071118349619978, + "grad_norm": 9.198818655810543, + "learning_rate": 1.3207664060217946e-05, + "loss": 0.5916, + "step": 3815 + }, + { + "epoch": 2.0716612377850163, + "grad_norm": 9.598428080121797, + "learning_rate": 1.320433285815067e-05, + "loss": 0.7907, + "step": 3816 + }, + { + "epoch": 2.072204125950054, + "grad_norm": 9.48389213525687, + "learning_rate": 1.32010012597664e-05, + "loss": 0.8385, + "step": 3817 + }, + { + "epoch": 2.0727470141150923, + "grad_norm": 7.562927002528845, + "learning_rate": 1.3197669265477191e-05, + "loss": 0.6748, + "step": 3818 + }, + { + "epoch": 2.07328990228013, + "grad_norm": 7.5241368226931185, + "learning_rate": 1.319433687569515e-05, + "loss": 0.5605, + "step": 3819 + }, + { + "epoch": 2.0738327904451683, + "grad_norm": 7.952800933557159, + "learning_rate": 1.3191004090832436e-05, + "loss": 0.6624, + "step": 3820 + }, + { + "epoch": 2.074375678610206, + "grad_norm": 8.02814530035842, + "learning_rate": 1.318767091130125e-05, + "loss": 0.488, + "step": 3821 + }, + { + "epoch": 2.0749185667752443, + "grad_norm": 7.000417608550192, + "learning_rate": 1.3184337337513849e-05, + "loss": 0.7194, + "step": 3822 + }, + { + "epoch": 2.075461454940282, + "grad_norm": 8.965861432524546, + "learning_rate": 1.3181003369882527e-05, + "loss": 0.4823, + "step": 3823 + }, + { + "epoch": 2.0760043431053203, + "grad_norm": 8.717885349674289, + "learning_rate": 1.3177669008819635e-05, + "loss": 0.6482, + "step": 3824 + }, + { + "epoch": 2.076547231270358, + "grad_norm": 9.217767101253152, + "learning_rate": 1.317433425473758e-05, + "loss": 0.5276, + "step": 3825 + }, + { + "epoch": 2.0770901194353963, + "grad_norm": 8.84305940258517, + "learning_rate": 1.3170999108048804e-05, + "loss": 0.6579, + "step": 3826 + }, + { + "epoch": 2.077633007600434, + "grad_norm": 7.501762934992334, + "learning_rate": 1.3167663569165803e-05, + "loss": 0.6633, + "step": 3827 + }, + { + "epoch": 2.0781758957654723, + "grad_norm": 7.815925996775423, + "learning_rate": 1.3164327638501126e-05, + "loss": 0.6881, + "step": 3828 + }, + { + "epoch": 2.07871878393051, + "grad_norm": 11.789396888417095, + "learning_rate": 1.3160991316467362e-05, + "loss": 0.5761, + "step": 3829 + }, + { + "epoch": 2.0792616720955484, + "grad_norm": 8.450719528350264, + "learning_rate": 1.3157654603477155e-05, + "loss": 0.7263, + "step": 3830 + }, + { + "epoch": 2.079804560260586, + "grad_norm": 9.544428310813693, + "learning_rate": 1.315431749994319e-05, + "loss": 0.6408, + "step": 3831 + }, + { + "epoch": 2.0803474484256244, + "grad_norm": 7.454950880630089, + "learning_rate": 1.3150980006278219e-05, + "loss": 0.4425, + "step": 3832 + }, + { + "epoch": 2.080890336590662, + "grad_norm": 11.462923993924347, + "learning_rate": 1.314764212289501e-05, + "loss": 0.5349, + "step": 3833 + }, + { + "epoch": 2.0814332247557004, + "grad_norm": 8.900231872199269, + "learning_rate": 1.314430385020641e-05, + "loss": 0.8022, + "step": 3834 + }, + { + "epoch": 2.081976112920738, + "grad_norm": 7.035528648373333, + "learning_rate": 1.3140965188625299e-05, + "loss": 0.6513, + "step": 3835 + }, + { + "epoch": 2.0825190010857764, + "grad_norm": 12.327229999200162, + "learning_rate": 1.3137626138564606e-05, + "loss": 1.0623, + "step": 3836 + }, + { + "epoch": 2.083061889250814, + "grad_norm": 8.942692654829559, + "learning_rate": 1.3134286700437308e-05, + "loss": 0.4689, + "step": 3837 + }, + { + "epoch": 2.0836047774158524, + "grad_norm": 11.092455380734007, + "learning_rate": 1.313094687465644e-05, + "loss": 0.6808, + "step": 3838 + }, + { + "epoch": 2.08414766558089, + "grad_norm": 8.918018917563646, + "learning_rate": 1.3127606661635075e-05, + "loss": 0.523, + "step": 3839 + }, + { + "epoch": 2.0846905537459284, + "grad_norm": 13.878014157853947, + "learning_rate": 1.312426606178633e-05, + "loss": 1.0231, + "step": 3840 + }, + { + "epoch": 2.085233441910966, + "grad_norm": 10.432596711564004, + "learning_rate": 1.3120925075523379e-05, + "loss": 0.8011, + "step": 3841 + }, + { + "epoch": 2.0857763300760044, + "grad_norm": 8.126214858686065, + "learning_rate": 1.3117583703259445e-05, + "loss": 0.5272, + "step": 3842 + }, + { + "epoch": 2.086319218241042, + "grad_norm": 9.596541824389169, + "learning_rate": 1.3114241945407783e-05, + "loss": 0.6585, + "step": 3843 + }, + { + "epoch": 2.0868621064060804, + "grad_norm": 10.438282441705661, + "learning_rate": 1.3110899802381718e-05, + "loss": 0.8474, + "step": 3844 + }, + { + "epoch": 2.087404994571118, + "grad_norm": 11.98868629313521, + "learning_rate": 1.3107557274594607e-05, + "loss": 1.0492, + "step": 3845 + }, + { + "epoch": 2.0879478827361564, + "grad_norm": 9.734289423556795, + "learning_rate": 1.310421436245986e-05, + "loss": 0.6573, + "step": 3846 + }, + { + "epoch": 2.088490770901194, + "grad_norm": 10.196913105458549, + "learning_rate": 1.310087106639093e-05, + "loss": 0.8964, + "step": 3847 + }, + { + "epoch": 2.0890336590662324, + "grad_norm": 9.155222462665467, + "learning_rate": 1.3097527386801327e-05, + "loss": 0.6668, + "step": 3848 + }, + { + "epoch": 2.08957654723127, + "grad_norm": 8.512351137441192, + "learning_rate": 1.3094183324104602e-05, + "loss": 0.6084, + "step": 3849 + }, + { + "epoch": 2.0901194353963084, + "grad_norm": 7.62701351090082, + "learning_rate": 1.3090838878714349e-05, + "loss": 0.4199, + "step": 3850 + }, + { + "epoch": 2.090662323561346, + "grad_norm": 8.57862015121728, + "learning_rate": 1.3087494051044218e-05, + "loss": 0.3694, + "step": 3851 + }, + { + "epoch": 2.0912052117263844, + "grad_norm": 11.407572249271688, + "learning_rate": 1.3084148841507904e-05, + "loss": 0.8861, + "step": 3852 + }, + { + "epoch": 2.091748099891422, + "grad_norm": 9.297247466948287, + "learning_rate": 1.3080803250519142e-05, + "loss": 0.8432, + "step": 3853 + }, + { + "epoch": 2.0922909880564604, + "grad_norm": 7.909660106539073, + "learning_rate": 1.3077457278491728e-05, + "loss": 0.5103, + "step": 3854 + }, + { + "epoch": 2.092833876221498, + "grad_norm": 11.013011688454522, + "learning_rate": 1.3074110925839491e-05, + "loss": 1.0117, + "step": 3855 + }, + { + "epoch": 2.0933767643865364, + "grad_norm": 10.249008356283914, + "learning_rate": 1.3070764192976315e-05, + "loss": 0.5248, + "step": 3856 + }, + { + "epoch": 2.093919652551574, + "grad_norm": 8.240438304204865, + "learning_rate": 1.306741708031613e-05, + "loss": 0.6811, + "step": 3857 + }, + { + "epoch": 2.0944625407166124, + "grad_norm": 9.542880797186081, + "learning_rate": 1.3064069588272913e-05, + "loss": 0.4746, + "step": 3858 + }, + { + "epoch": 2.09500542888165, + "grad_norm": 9.251214264026915, + "learning_rate": 1.3060721717260685e-05, + "loss": 0.6405, + "step": 3859 + }, + { + "epoch": 2.0955483170466884, + "grad_norm": 12.366199541130733, + "learning_rate": 1.3057373467693515e-05, + "loss": 0.8378, + "step": 3860 + }, + { + "epoch": 2.096091205211726, + "grad_norm": 10.566968498265949, + "learning_rate": 1.3054024839985526e-05, + "loss": 0.5975, + "step": 3861 + }, + { + "epoch": 2.0966340933767644, + "grad_norm": 9.58134406953771, + "learning_rate": 1.3050675834550872e-05, + "loss": 0.6846, + "step": 3862 + }, + { + "epoch": 2.097176981541802, + "grad_norm": 8.437576527492594, + "learning_rate": 1.3047326451803772e-05, + "loss": 0.6301, + "step": 3863 + }, + { + "epoch": 2.0977198697068404, + "grad_norm": 8.68957421381677, + "learning_rate": 1.304397669215848e-05, + "loss": 0.6121, + "step": 3864 + }, + { + "epoch": 2.098262757871878, + "grad_norm": 12.281098753982102, + "learning_rate": 1.30406265560293e-05, + "loss": 1.174, + "step": 3865 + }, + { + "epoch": 2.0988056460369164, + "grad_norm": 12.64390360725308, + "learning_rate": 1.303727604383058e-05, + "loss": 0.727, + "step": 3866 + }, + { + "epoch": 2.099348534201954, + "grad_norm": 8.982715547555616, + "learning_rate": 1.3033925155976718e-05, + "loss": 0.4946, + "step": 3867 + }, + { + "epoch": 2.0998914223669924, + "grad_norm": 8.77924533690129, + "learning_rate": 1.303057389288216e-05, + "loss": 0.5027, + "step": 3868 + }, + { + "epoch": 2.1004343105320302, + "grad_norm": 12.246059108799978, + "learning_rate": 1.302722225496139e-05, + "loss": 0.6522, + "step": 3869 + }, + { + "epoch": 2.1009771986970684, + "grad_norm": 13.95610327170378, + "learning_rate": 1.3023870242628944e-05, + "loss": 1.043, + "step": 3870 + }, + { + "epoch": 2.1015200868621062, + "grad_norm": 10.355142745873641, + "learning_rate": 1.3020517856299413e-05, + "loss": 0.9167, + "step": 3871 + }, + { + "epoch": 2.1020629750271445, + "grad_norm": 8.43332484970558, + "learning_rate": 1.3017165096387419e-05, + "loss": 0.5952, + "step": 3872 + }, + { + "epoch": 2.1026058631921822, + "grad_norm": 9.224204311836482, + "learning_rate": 1.3013811963307634e-05, + "loss": 0.5631, + "step": 3873 + }, + { + "epoch": 2.1031487513572205, + "grad_norm": 9.719137418718985, + "learning_rate": 1.3010458457474784e-05, + "loss": 0.9293, + "step": 3874 + }, + { + "epoch": 2.1036916395222582, + "grad_norm": 8.234100299546002, + "learning_rate": 1.3007104579303636e-05, + "loss": 0.7041, + "step": 3875 + }, + { + "epoch": 2.1042345276872965, + "grad_norm": 11.0180135357339, + "learning_rate": 1.3003750329208995e-05, + "loss": 0.6068, + "step": 3876 + }, + { + "epoch": 2.1047774158523342, + "grad_norm": 8.790070650912016, + "learning_rate": 1.300039570760573e-05, + "loss": 0.4956, + "step": 3877 + }, + { + "epoch": 2.1053203040173725, + "grad_norm": 8.43754180724615, + "learning_rate": 1.2997040714908742e-05, + "loss": 0.3965, + "step": 3878 + }, + { + "epoch": 2.1058631921824102, + "grad_norm": 11.815520632897917, + "learning_rate": 1.299368535153298e-05, + "loss": 0.5963, + "step": 3879 + }, + { + "epoch": 2.1064060803474485, + "grad_norm": 14.168765746771772, + "learning_rate": 1.2990329617893445e-05, + "loss": 0.8111, + "step": 3880 + }, + { + "epoch": 2.1069489685124863, + "grad_norm": 5.706418305089391, + "learning_rate": 1.2986973514405176e-05, + "loss": 0.41, + "step": 3881 + }, + { + "epoch": 2.1074918566775245, + "grad_norm": 10.870432547178469, + "learning_rate": 1.2983617041483261e-05, + "loss": 0.7805, + "step": 3882 + }, + { + "epoch": 2.1080347448425623, + "grad_norm": 9.440648088759033, + "learning_rate": 1.2980260199542838e-05, + "loss": 0.6047, + "step": 3883 + }, + { + "epoch": 2.1085776330076005, + "grad_norm": 12.145529985557056, + "learning_rate": 1.2976902988999081e-05, + "loss": 0.8366, + "step": 3884 + }, + { + "epoch": 2.1091205211726383, + "grad_norm": 9.17886258135476, + "learning_rate": 1.2973545410267218e-05, + "loss": 0.6402, + "step": 3885 + }, + { + "epoch": 2.1096634093376765, + "grad_norm": 9.613840291682566, + "learning_rate": 1.297018746376252e-05, + "loss": 0.8175, + "step": 3886 + }, + { + "epoch": 2.1102062975027143, + "grad_norm": 12.898243756363112, + "learning_rate": 1.2966829149900304e-05, + "loss": 0.8514, + "step": 3887 + }, + { + "epoch": 2.1107491856677525, + "grad_norm": 8.825284004643853, + "learning_rate": 1.2963470469095928e-05, + "loss": 0.4508, + "step": 3888 + }, + { + "epoch": 2.1112920738327903, + "grad_norm": 9.953568896257451, + "learning_rate": 1.2960111421764803e-05, + "loss": 0.5657, + "step": 3889 + }, + { + "epoch": 2.1118349619978285, + "grad_norm": 8.988870757227582, + "learning_rate": 1.2956752008322378e-05, + "loss": 0.7242, + "step": 3890 + }, + { + "epoch": 2.1123778501628663, + "grad_norm": 9.68770492429417, + "learning_rate": 1.2953392229184156e-05, + "loss": 0.9403, + "step": 3891 + }, + { + "epoch": 2.1129207383279045, + "grad_norm": 11.903522492661367, + "learning_rate": 1.2950032084765674e-05, + "loss": 0.8743, + "step": 3892 + }, + { + "epoch": 2.1134636264929423, + "grad_norm": 12.24198259147124, + "learning_rate": 1.294667157548252e-05, + "loss": 0.9861, + "step": 3893 + }, + { + "epoch": 2.1140065146579805, + "grad_norm": 13.105368254945933, + "learning_rate": 1.2943310701750331e-05, + "loss": 0.8815, + "step": 3894 + }, + { + "epoch": 2.1145494028230183, + "grad_norm": 9.322992215038694, + "learning_rate": 1.2939949463984782e-05, + "loss": 0.4986, + "step": 3895 + }, + { + "epoch": 2.1150922909880565, + "grad_norm": 7.59330739725767, + "learning_rate": 1.29365878626016e-05, + "loss": 0.4343, + "step": 3896 + }, + { + "epoch": 2.1156351791530943, + "grad_norm": 9.782670203398764, + "learning_rate": 1.293322589801655e-05, + "loss": 0.4843, + "step": 3897 + }, + { + "epoch": 2.1161780673181325, + "grad_norm": 7.130500958222279, + "learning_rate": 1.2929863570645446e-05, + "loss": 0.3784, + "step": 3898 + }, + { + "epoch": 2.1167209554831703, + "grad_norm": 7.336707110452157, + "learning_rate": 1.2926500880904147e-05, + "loss": 0.4938, + "step": 3899 + }, + { + "epoch": 2.1172638436482085, + "grad_norm": 7.047478722817777, + "learning_rate": 1.2923137829208555e-05, + "loss": 0.3002, + "step": 3900 + }, + { + "epoch": 2.1178067318132463, + "grad_norm": 9.533093539252569, + "learning_rate": 1.2919774415974616e-05, + "loss": 0.711, + "step": 3901 + }, + { + "epoch": 2.1183496199782845, + "grad_norm": 8.468755965185016, + "learning_rate": 1.2916410641618324e-05, + "loss": 0.5491, + "step": 3902 + }, + { + "epoch": 2.1188925081433223, + "grad_norm": 10.351769959990103, + "learning_rate": 1.2913046506555715e-05, + "loss": 0.6269, + "step": 3903 + }, + { + "epoch": 2.1194353963083605, + "grad_norm": 10.2212813809041, + "learning_rate": 1.2909682011202875e-05, + "loss": 0.7249, + "step": 3904 + }, + { + "epoch": 2.1199782844733983, + "grad_norm": 9.344205468199693, + "learning_rate": 1.2906317155975922e-05, + "loss": 0.6277, + "step": 3905 + }, + { + "epoch": 2.1205211726384365, + "grad_norm": 8.704990474328149, + "learning_rate": 1.2902951941291035e-05, + "loss": 0.6317, + "step": 3906 + }, + { + "epoch": 2.1210640608034743, + "grad_norm": 11.644285794605782, + "learning_rate": 1.2899586367564422e-05, + "loss": 0.8585, + "step": 3907 + }, + { + "epoch": 2.1216069489685125, + "grad_norm": 8.462034849045626, + "learning_rate": 1.2896220435212347e-05, + "loss": 0.5174, + "step": 3908 + }, + { + "epoch": 2.1221498371335503, + "grad_norm": 10.023149647380059, + "learning_rate": 1.2892854144651112e-05, + "loss": 0.6044, + "step": 3909 + }, + { + "epoch": 2.1226927252985885, + "grad_norm": 9.437366874691593, + "learning_rate": 1.2889487496297068e-05, + "loss": 0.5081, + "step": 3910 + }, + { + "epoch": 2.1232356134636263, + "grad_norm": 11.06235798573597, + "learning_rate": 1.2886120490566607e-05, + "loss": 0.7385, + "step": 3911 + }, + { + "epoch": 2.1237785016286646, + "grad_norm": 9.45153797141243, + "learning_rate": 1.288275312787616e-05, + "loss": 0.6165, + "step": 3912 + }, + { + "epoch": 2.1243213897937023, + "grad_norm": 8.907808417684745, + "learning_rate": 1.2879385408642214e-05, + "loss": 0.3956, + "step": 3913 + }, + { + "epoch": 2.1248642779587406, + "grad_norm": 12.022235830361943, + "learning_rate": 1.2876017333281288e-05, + "loss": 0.7492, + "step": 3914 + }, + { + "epoch": 2.1254071661237783, + "grad_norm": 10.18229463165741, + "learning_rate": 1.2872648902209957e-05, + "loss": 1.0188, + "step": 3915 + }, + { + "epoch": 2.1259500542888166, + "grad_norm": 8.068860605172112, + "learning_rate": 1.2869280115844831e-05, + "loss": 0.4258, + "step": 3916 + }, + { + "epoch": 2.1264929424538543, + "grad_norm": 7.300141006448243, + "learning_rate": 1.2865910974602568e-05, + "loss": 0.414, + "step": 3917 + }, + { + "epoch": 2.1270358306188926, + "grad_norm": 9.447874959421453, + "learning_rate": 1.2862541478899868e-05, + "loss": 0.6323, + "step": 3918 + }, + { + "epoch": 2.1275787187839303, + "grad_norm": 11.845175678539553, + "learning_rate": 1.2859171629153476e-05, + "loss": 0.8539, + "step": 3919 + }, + { + "epoch": 2.1281216069489686, + "grad_norm": 8.29306868561595, + "learning_rate": 1.2855801425780179e-05, + "loss": 0.7127, + "step": 3920 + }, + { + "epoch": 2.1286644951140063, + "grad_norm": 8.040765832896788, + "learning_rate": 1.285243086919681e-05, + "loss": 0.495, + "step": 3921 + }, + { + "epoch": 2.1292073832790446, + "grad_norm": 13.042470321623425, + "learning_rate": 1.2849059959820248e-05, + "loss": 0.616, + "step": 3922 + }, + { + "epoch": 2.1297502714440824, + "grad_norm": 8.585317997471494, + "learning_rate": 1.2845688698067406e-05, + "loss": 0.5599, + "step": 3923 + }, + { + "epoch": 2.1302931596091206, + "grad_norm": 9.356896896219936, + "learning_rate": 1.2842317084355251e-05, + "loss": 0.605, + "step": 3924 + }, + { + "epoch": 2.1308360477741584, + "grad_norm": 17.057557558601207, + "learning_rate": 1.283894511910079e-05, + "loss": 0.9396, + "step": 3925 + }, + { + "epoch": 2.1313789359391966, + "grad_norm": 10.659876178353805, + "learning_rate": 1.2835572802721072e-05, + "loss": 0.765, + "step": 3926 + }, + { + "epoch": 2.1319218241042344, + "grad_norm": 9.903600273604221, + "learning_rate": 1.2832200135633191e-05, + "loss": 0.4745, + "step": 3927 + }, + { + "epoch": 2.1324647122692726, + "grad_norm": 12.075983241730247, + "learning_rate": 1.2828827118254279e-05, + "loss": 0.8613, + "step": 3928 + }, + { + "epoch": 2.1330076004343104, + "grad_norm": 10.13381566988187, + "learning_rate": 1.2825453751001526e-05, + "loss": 0.7342, + "step": 3929 + }, + { + "epoch": 2.1335504885993486, + "grad_norm": 12.671225163451464, + "learning_rate": 1.282208003429215e-05, + "loss": 0.7765, + "step": 3930 + }, + { + "epoch": 2.1340933767643864, + "grad_norm": 10.503393267034513, + "learning_rate": 1.2818705968543417e-05, + "loss": 0.7042, + "step": 3931 + }, + { + "epoch": 2.1346362649294246, + "grad_norm": 7.8720699552641635, + "learning_rate": 1.2815331554172634e-05, + "loss": 0.4393, + "step": 3932 + }, + { + "epoch": 2.1351791530944624, + "grad_norm": 14.057115653429742, + "learning_rate": 1.281195679159716e-05, + "loss": 0.8676, + "step": 3933 + }, + { + "epoch": 2.1357220412595006, + "grad_norm": 12.006814229853843, + "learning_rate": 1.2808581681234387e-05, + "loss": 0.8764, + "step": 3934 + }, + { + "epoch": 2.1362649294245384, + "grad_norm": 11.511042628784308, + "learning_rate": 1.2805206223501756e-05, + "loss": 0.6034, + "step": 3935 + }, + { + "epoch": 2.1368078175895766, + "grad_norm": 6.922027915468186, + "learning_rate": 1.2801830418816749e-05, + "loss": 0.4303, + "step": 3936 + }, + { + "epoch": 2.1373507057546144, + "grad_norm": 8.583211919241508, + "learning_rate": 1.2798454267596892e-05, + "loss": 0.6059, + "step": 3937 + }, + { + "epoch": 2.1378935939196526, + "grad_norm": 10.400540758946061, + "learning_rate": 1.2795077770259749e-05, + "loss": 0.7327, + "step": 3938 + }, + { + "epoch": 2.1384364820846904, + "grad_norm": 11.253848098269156, + "learning_rate": 1.2791700927222932e-05, + "loss": 0.7247, + "step": 3939 + }, + { + "epoch": 2.1389793702497286, + "grad_norm": 10.650957738366092, + "learning_rate": 1.2788323738904098e-05, + "loss": 0.8589, + "step": 3940 + }, + { + "epoch": 2.1395222584147664, + "grad_norm": 7.284559079353519, + "learning_rate": 1.2784946205720936e-05, + "loss": 0.3361, + "step": 3941 + }, + { + "epoch": 2.1400651465798046, + "grad_norm": 10.532878125436458, + "learning_rate": 1.2781568328091192e-05, + "loss": 0.8466, + "step": 3942 + }, + { + "epoch": 2.1406080347448424, + "grad_norm": 16.5717179039462, + "learning_rate": 1.2778190106432643e-05, + "loss": 1.1453, + "step": 3943 + }, + { + "epoch": 2.1411509229098806, + "grad_norm": 8.591001506273539, + "learning_rate": 1.2774811541163114e-05, + "loss": 0.5379, + "step": 3944 + }, + { + "epoch": 2.1416938110749184, + "grad_norm": 11.173192436567485, + "learning_rate": 1.2771432632700471e-05, + "loss": 1.0715, + "step": 3945 + }, + { + "epoch": 2.1422366992399566, + "grad_norm": 11.381968938636014, + "learning_rate": 1.2768053381462625e-05, + "loss": 0.8903, + "step": 3946 + }, + { + "epoch": 2.1427795874049944, + "grad_norm": 9.847481766271917, + "learning_rate": 1.2764673787867519e-05, + "loss": 0.5874, + "step": 3947 + }, + { + "epoch": 2.1433224755700326, + "grad_norm": 9.183549346287741, + "learning_rate": 1.2761293852333156e-05, + "loss": 0.6745, + "step": 3948 + }, + { + "epoch": 2.1438653637350704, + "grad_norm": 9.610890909648008, + "learning_rate": 1.2757913575277572e-05, + "loss": 0.9781, + "step": 3949 + }, + { + "epoch": 2.1444082519001086, + "grad_norm": 9.814885217738354, + "learning_rate": 1.275453295711884e-05, + "loss": 0.9977, + "step": 3950 + }, + { + "epoch": 2.1449511400651464, + "grad_norm": 10.984041832783662, + "learning_rate": 1.275115199827508e-05, + "loss": 0.7822, + "step": 3951 + }, + { + "epoch": 2.1454940282301846, + "grad_norm": 9.634528803708669, + "learning_rate": 1.2747770699164457e-05, + "loss": 1.1719, + "step": 3952 + }, + { + "epoch": 2.1460369163952224, + "grad_norm": 9.767304548901404, + "learning_rate": 1.2744389060205173e-05, + "loss": 0.6284, + "step": 3953 + }, + { + "epoch": 2.1465798045602607, + "grad_norm": 10.373642481495464, + "learning_rate": 1.2741007081815478e-05, + "loss": 0.6742, + "step": 3954 + }, + { + "epoch": 2.1471226927252984, + "grad_norm": 10.390187789918086, + "learning_rate": 1.2737624764413659e-05, + "loss": 0.6647, + "step": 3955 + }, + { + "epoch": 2.1476655808903367, + "grad_norm": 10.617840626931814, + "learning_rate": 1.2734242108418044e-05, + "loss": 0.877, + "step": 3956 + }, + { + "epoch": 2.1482084690553744, + "grad_norm": 9.898468473630299, + "learning_rate": 1.273085911424701e-05, + "loss": 0.7335, + "step": 3957 + }, + { + "epoch": 2.1487513572204127, + "grad_norm": 10.409230729626815, + "learning_rate": 1.2727475782318966e-05, + "loss": 0.7169, + "step": 3958 + }, + { + "epoch": 2.1492942453854504, + "grad_norm": 10.425915655294723, + "learning_rate": 1.2724092113052372e-05, + "loss": 0.6578, + "step": 3959 + }, + { + "epoch": 2.1498371335504887, + "grad_norm": 6.3375473265246995, + "learning_rate": 1.2720708106865722e-05, + "loss": 0.3471, + "step": 3960 + }, + { + "epoch": 2.1503800217155264, + "grad_norm": 9.396081485808933, + "learning_rate": 1.2717323764177559e-05, + "loss": 0.5232, + "step": 3961 + }, + { + "epoch": 2.1509229098805647, + "grad_norm": 13.419504298012779, + "learning_rate": 1.2713939085406461e-05, + "loss": 1.8096, + "step": 3962 + }, + { + "epoch": 2.1514657980456025, + "grad_norm": 9.635837991132963, + "learning_rate": 1.2710554070971053e-05, + "loss": 0.5949, + "step": 3963 + }, + { + "epoch": 2.1520086862106407, + "grad_norm": 10.111061783908927, + "learning_rate": 1.2707168721289995e-05, + "loss": 0.8731, + "step": 3964 + }, + { + "epoch": 2.1525515743756785, + "grad_norm": 8.256139809775819, + "learning_rate": 1.2703783036781998e-05, + "loss": 0.726, + "step": 3965 + }, + { + "epoch": 2.1530944625407167, + "grad_norm": 8.385511442711055, + "learning_rate": 1.2700397017865802e-05, + "loss": 0.6209, + "step": 3966 + }, + { + "epoch": 2.1536373507057545, + "grad_norm": 8.556891915186654, + "learning_rate": 1.2697010664960201e-05, + "loss": 0.6746, + "step": 3967 + }, + { + "epoch": 2.1541802388707927, + "grad_norm": 8.219611855039698, + "learning_rate": 1.2693623978484022e-05, + "loss": 0.7542, + "step": 3968 + }, + { + "epoch": 2.1547231270358305, + "grad_norm": 9.425031325511556, + "learning_rate": 1.2690236958856139e-05, + "loss": 0.6735, + "step": 3969 + }, + { + "epoch": 2.1552660152008687, + "grad_norm": 9.21346190912602, + "learning_rate": 1.2686849606495461e-05, + "loss": 0.5311, + "step": 3970 + }, + { + "epoch": 2.1558089033659065, + "grad_norm": 12.261438418537729, + "learning_rate": 1.268346192182094e-05, + "loss": 0.8556, + "step": 3971 + }, + { + "epoch": 2.1563517915309447, + "grad_norm": 8.630907790930149, + "learning_rate": 1.268007390525157e-05, + "loss": 0.7789, + "step": 3972 + }, + { + "epoch": 2.1568946796959825, + "grad_norm": 10.654367663061162, + "learning_rate": 1.267668555720639e-05, + "loss": 1.1708, + "step": 3973 + }, + { + "epoch": 2.1574375678610207, + "grad_norm": 15.113883332023352, + "learning_rate": 1.2673296878104472e-05, + "loss": 1.2273, + "step": 3974 + }, + { + "epoch": 2.1579804560260585, + "grad_norm": 10.205014286438582, + "learning_rate": 1.266990786836494e-05, + "loss": 1.2711, + "step": 3975 + }, + { + "epoch": 2.1585233441910967, + "grad_norm": 8.32550465973343, + "learning_rate": 1.2666518528406944e-05, + "loss": 0.6387, + "step": 3976 + }, + { + "epoch": 2.1590662323561345, + "grad_norm": 8.80250087611585, + "learning_rate": 1.2663128858649686e-05, + "loss": 0.5669, + "step": 3977 + }, + { + "epoch": 2.1596091205211727, + "grad_norm": 7.977255679001254, + "learning_rate": 1.2659738859512411e-05, + "loss": 0.4443, + "step": 3978 + }, + { + "epoch": 2.1601520086862105, + "grad_norm": 6.086622373481402, + "learning_rate": 1.265634853141439e-05, + "loss": 0.3029, + "step": 3979 + }, + { + "epoch": 2.1606948968512487, + "grad_norm": 16.46377636281654, + "learning_rate": 1.2652957874774953e-05, + "loss": 1.09, + "step": 3980 + }, + { + "epoch": 2.1612377850162865, + "grad_norm": 10.805039141935024, + "learning_rate": 1.2649566890013456e-05, + "loss": 0.672, + "step": 3981 + }, + { + "epoch": 2.1617806731813247, + "grad_norm": 10.731407400272554, + "learning_rate": 1.2646175577549303e-05, + "loss": 0.6618, + "step": 3982 + }, + { + "epoch": 2.1623235613463625, + "grad_norm": 9.572301578766334, + "learning_rate": 1.2642783937801937e-05, + "loss": 0.6608, + "step": 3983 + }, + { + "epoch": 2.1628664495114007, + "grad_norm": 8.415534879772189, + "learning_rate": 1.2639391971190842e-05, + "loss": 0.5314, + "step": 3984 + }, + { + "epoch": 2.1634093376764385, + "grad_norm": 8.290688286613944, + "learning_rate": 1.263599967813554e-05, + "loss": 0.6736, + "step": 3985 + }, + { + "epoch": 2.1639522258414767, + "grad_norm": 8.92112434155926, + "learning_rate": 1.2632607059055597e-05, + "loss": 0.5572, + "step": 3986 + }, + { + "epoch": 2.1644951140065145, + "grad_norm": 9.82202342503724, + "learning_rate": 1.2629214114370618e-05, + "loss": 0.7838, + "step": 3987 + }, + { + "epoch": 2.1650380021715527, + "grad_norm": 7.20619661733841, + "learning_rate": 1.262582084450025e-05, + "loss": 0.5577, + "step": 3988 + }, + { + "epoch": 2.1655808903365905, + "grad_norm": 8.640482408220107, + "learning_rate": 1.2622427249864172e-05, + "loss": 0.819, + "step": 3989 + }, + { + "epoch": 2.1661237785016287, + "grad_norm": 9.245400712044432, + "learning_rate": 1.2619033330882114e-05, + "loss": 0.6009, + "step": 3990 + }, + { + "epoch": 2.1666666666666665, + "grad_norm": 8.891547209594252, + "learning_rate": 1.2615639087973835e-05, + "loss": 0.6085, + "step": 3991 + }, + { + "epoch": 2.1672095548317047, + "grad_norm": 12.196798951360252, + "learning_rate": 1.2612244521559148e-05, + "loss": 0.745, + "step": 3992 + }, + { + "epoch": 2.1677524429967425, + "grad_norm": 9.610928616001752, + "learning_rate": 1.2608849632057895e-05, + "loss": 0.7709, + "step": 3993 + }, + { + "epoch": 2.1682953311617807, + "grad_norm": 9.454740121591225, + "learning_rate": 1.2605454419889962e-05, + "loss": 0.6455, + "step": 3994 + }, + { + "epoch": 2.1688382193268185, + "grad_norm": 13.245987794421069, + "learning_rate": 1.2602058885475273e-05, + "loss": 0.7104, + "step": 3995 + }, + { + "epoch": 2.1693811074918568, + "grad_norm": 10.51068899063252, + "learning_rate": 1.2598663029233794e-05, + "loss": 0.5747, + "step": 3996 + }, + { + "epoch": 2.1699239956568945, + "grad_norm": 10.589584490390159, + "learning_rate": 1.2595266851585532e-05, + "loss": 0.7878, + "step": 3997 + }, + { + "epoch": 2.1704668838219328, + "grad_norm": 7.818930212111502, + "learning_rate": 1.2591870352950528e-05, + "loss": 0.5179, + "step": 3998 + }, + { + "epoch": 2.1710097719869705, + "grad_norm": 8.222376749362372, + "learning_rate": 1.2588473533748868e-05, + "loss": 0.4722, + "step": 3999 + }, + { + "epoch": 2.1715526601520088, + "grad_norm": 12.451645756942625, + "learning_rate": 1.2585076394400675e-05, + "loss": 0.7742, + "step": 4000 + }, + { + "epoch": 2.1720955483170465, + "grad_norm": 11.33368393512842, + "learning_rate": 1.2581678935326117e-05, + "loss": 0.3945, + "step": 4001 + }, + { + "epoch": 2.1726384364820848, + "grad_norm": 10.914311466768305, + "learning_rate": 1.2578281156945389e-05, + "loss": 0.6345, + "step": 4002 + }, + { + "epoch": 2.1731813246471225, + "grad_norm": 12.867661375339019, + "learning_rate": 1.2574883059678742e-05, + "loss": 1.1569, + "step": 4003 + }, + { + "epoch": 2.1737242128121608, + "grad_norm": 11.4420969393128, + "learning_rate": 1.2571484643946452e-05, + "loss": 0.4927, + "step": 4004 + }, + { + "epoch": 2.1742671009771986, + "grad_norm": 10.234468502706436, + "learning_rate": 1.2568085910168842e-05, + "loss": 0.463, + "step": 4005 + }, + { + "epoch": 2.1748099891422368, + "grad_norm": 7.649699643480241, + "learning_rate": 1.2564686858766275e-05, + "loss": 0.3981, + "step": 4006 + }, + { + "epoch": 2.1753528773072746, + "grad_norm": 9.503268181536553, + "learning_rate": 1.2561287490159151e-05, + "loss": 0.4162, + "step": 4007 + }, + { + "epoch": 2.175895765472313, + "grad_norm": 8.81726258724681, + "learning_rate": 1.2557887804767907e-05, + "loss": 0.92, + "step": 4008 + }, + { + "epoch": 2.1764386536373506, + "grad_norm": 9.625823227183314, + "learning_rate": 1.2554487803013027e-05, + "loss": 0.7527, + "step": 4009 + }, + { + "epoch": 2.176981541802389, + "grad_norm": 11.452435192265499, + "learning_rate": 1.2551087485315017e-05, + "loss": 0.8611, + "step": 4010 + }, + { + "epoch": 2.1775244299674266, + "grad_norm": 9.123700040373722, + "learning_rate": 1.2547686852094445e-05, + "loss": 0.475, + "step": 4011 + }, + { + "epoch": 2.178067318132465, + "grad_norm": 9.306092992521913, + "learning_rate": 1.2544285903771902e-05, + "loss": 0.6063, + "step": 4012 + }, + { + "epoch": 2.1786102062975026, + "grad_norm": 14.283804972037375, + "learning_rate": 1.2540884640768022e-05, + "loss": 0.9142, + "step": 4013 + }, + { + "epoch": 2.179153094462541, + "grad_norm": 14.299740055929474, + "learning_rate": 1.2537483063503483e-05, + "loss": 0.8124, + "step": 4014 + }, + { + "epoch": 2.1796959826275786, + "grad_norm": 9.637221226214642, + "learning_rate": 1.2534081172398993e-05, + "loss": 0.6278, + "step": 4015 + }, + { + "epoch": 2.180238870792617, + "grad_norm": 6.674062415604907, + "learning_rate": 1.2530678967875304e-05, + "loss": 0.5078, + "step": 4016 + }, + { + "epoch": 2.1807817589576546, + "grad_norm": 11.68250334366375, + "learning_rate": 1.252727645035321e-05, + "loss": 0.7393, + "step": 4017 + }, + { + "epoch": 2.181324647122693, + "grad_norm": 9.642668530719417, + "learning_rate": 1.2523873620253535e-05, + "loss": 0.6218, + "step": 4018 + }, + { + "epoch": 2.1818675352877306, + "grad_norm": 9.955103488106618, + "learning_rate": 1.2520470477997146e-05, + "loss": 0.899, + "step": 4019 + }, + { + "epoch": 2.182410423452769, + "grad_norm": 13.184397199493825, + "learning_rate": 1.2517067024004955e-05, + "loss": 0.8437, + "step": 4020 + }, + { + "epoch": 2.1829533116178066, + "grad_norm": 11.142111068408965, + "learning_rate": 1.2513663258697901e-05, + "loss": 0.6064, + "step": 4021 + }, + { + "epoch": 2.183496199782845, + "grad_norm": 9.336503366824324, + "learning_rate": 1.251025918249697e-05, + "loss": 0.6649, + "step": 4022 + }, + { + "epoch": 2.1840390879478826, + "grad_norm": 11.35476662379509, + "learning_rate": 1.250685479582318e-05, + "loss": 0.8573, + "step": 4023 + }, + { + "epoch": 2.184581976112921, + "grad_norm": 9.049806197407708, + "learning_rate": 1.2503450099097594e-05, + "loss": 0.3839, + "step": 4024 + }, + { + "epoch": 2.1851248642779586, + "grad_norm": 12.739046308083323, + "learning_rate": 1.250004509274131e-05, + "loss": 0.9344, + "step": 4025 + }, + { + "epoch": 2.185667752442997, + "grad_norm": 7.539200963039068, + "learning_rate": 1.2496639777175465e-05, + "loss": 0.3691, + "step": 4026 + }, + { + "epoch": 2.1862106406080346, + "grad_norm": 10.659038181144831, + "learning_rate": 1.2493234152821234e-05, + "loss": 0.9462, + "step": 4027 + }, + { + "epoch": 2.186753528773073, + "grad_norm": 8.047186696422612, + "learning_rate": 1.2489828220099831e-05, + "loss": 0.4974, + "step": 4028 + }, + { + "epoch": 2.1872964169381106, + "grad_norm": 11.048564474051437, + "learning_rate": 1.2486421979432503e-05, + "loss": 0.6221, + "step": 4029 + }, + { + "epoch": 2.187839305103149, + "grad_norm": 9.808559304060044, + "learning_rate": 1.2483015431240542e-05, + "loss": 0.5221, + "step": 4030 + }, + { + "epoch": 2.1883821932681866, + "grad_norm": 7.720481766274275, + "learning_rate": 1.2479608575945274e-05, + "loss": 0.5248, + "step": 4031 + }, + { + "epoch": 2.188925081433225, + "grad_norm": 7.791803772948659, + "learning_rate": 1.2476201413968068e-05, + "loss": 0.5313, + "step": 4032 + }, + { + "epoch": 2.1894679695982626, + "grad_norm": 8.280464830936529, + "learning_rate": 1.2472793945730323e-05, + "loss": 0.5155, + "step": 4033 + }, + { + "epoch": 2.190010857763301, + "grad_norm": 6.500045291722395, + "learning_rate": 1.2469386171653483e-05, + "loss": 0.6057, + "step": 4034 + }, + { + "epoch": 2.1905537459283386, + "grad_norm": 6.735518823264981, + "learning_rate": 1.2465978092159025e-05, + "loss": 0.5207, + "step": 4035 + }, + { + "epoch": 2.191096634093377, + "grad_norm": 12.977966385021237, + "learning_rate": 1.2462569707668468e-05, + "loss": 1.0472, + "step": 4036 + }, + { + "epoch": 2.1916395222584146, + "grad_norm": 12.805295393279883, + "learning_rate": 1.245916101860336e-05, + "loss": 0.7486, + "step": 4037 + }, + { + "epoch": 2.192182410423453, + "grad_norm": 11.04531935760714, + "learning_rate": 1.2455752025385304e-05, + "loss": 0.6107, + "step": 4038 + }, + { + "epoch": 2.1927252985884906, + "grad_norm": 12.12096270238078, + "learning_rate": 1.245234272843592e-05, + "loss": 0.6264, + "step": 4039 + }, + { + "epoch": 2.193268186753529, + "grad_norm": 14.202650747934204, + "learning_rate": 1.244893312817688e-05, + "loss": 0.7965, + "step": 4040 + }, + { + "epoch": 2.1938110749185666, + "grad_norm": 10.409453866661186, + "learning_rate": 1.2445523225029887e-05, + "loss": 0.6012, + "step": 4041 + }, + { + "epoch": 2.194353963083605, + "grad_norm": 8.68329891299571, + "learning_rate": 1.2442113019416683e-05, + "loss": 0.4393, + "step": 4042 + }, + { + "epoch": 2.1948968512486426, + "grad_norm": 11.77894618341781, + "learning_rate": 1.2438702511759049e-05, + "loss": 0.7838, + "step": 4043 + }, + { + "epoch": 2.195439739413681, + "grad_norm": 8.31374036003515, + "learning_rate": 1.2435291702478802e-05, + "loss": 0.6043, + "step": 4044 + }, + { + "epoch": 2.1959826275787186, + "grad_norm": 10.219603175661236, + "learning_rate": 1.2431880591997799e-05, + "loss": 0.9292, + "step": 4045 + }, + { + "epoch": 2.196525515743757, + "grad_norm": 10.662478372576002, + "learning_rate": 1.2428469180737923e-05, + "loss": 0.5951, + "step": 4046 + }, + { + "epoch": 2.1970684039087947, + "grad_norm": 15.414471746340716, + "learning_rate": 1.2425057469121113e-05, + "loss": 1.0091, + "step": 4047 + }, + { + "epoch": 2.197611292073833, + "grad_norm": 8.553095532374297, + "learning_rate": 1.242164545756933e-05, + "loss": 0.4494, + "step": 4048 + }, + { + "epoch": 2.1981541802388707, + "grad_norm": 10.520590758759942, + "learning_rate": 1.2418233146504575e-05, + "loss": 0.5277, + "step": 4049 + }, + { + "epoch": 2.198697068403909, + "grad_norm": 10.792986797196212, + "learning_rate": 1.241482053634889e-05, + "loss": 0.7036, + "step": 4050 + }, + { + "epoch": 2.1992399565689467, + "grad_norm": 10.425002068049416, + "learning_rate": 1.2411407627524354e-05, + "loss": 0.8283, + "step": 4051 + }, + { + "epoch": 2.199782844733985, + "grad_norm": 10.421667203553737, + "learning_rate": 1.240799442045308e-05, + "loss": 0.8689, + "step": 4052 + }, + { + "epoch": 2.2003257328990227, + "grad_norm": 9.866413466645723, + "learning_rate": 1.2404580915557217e-05, + "loss": 0.517, + "step": 4053 + }, + { + "epoch": 2.200868621064061, + "grad_norm": 9.806755744401848, + "learning_rate": 1.2401167113258954e-05, + "loss": 0.7176, + "step": 4054 + }, + { + "epoch": 2.2014115092290987, + "grad_norm": 12.103200009209074, + "learning_rate": 1.2397753013980516e-05, + "loss": 0.8563, + "step": 4055 + }, + { + "epoch": 2.201954397394137, + "grad_norm": 9.310089002310463, + "learning_rate": 1.2394338618144162e-05, + "loss": 0.5478, + "step": 4056 + }, + { + "epoch": 2.2024972855591747, + "grad_norm": 8.147499111803365, + "learning_rate": 1.2390923926172194e-05, + "loss": 0.4616, + "step": 4057 + }, + { + "epoch": 2.203040173724213, + "grad_norm": 10.363627124726912, + "learning_rate": 1.2387508938486945e-05, + "loss": 0.6039, + "step": 4058 + }, + { + "epoch": 2.2035830618892507, + "grad_norm": 11.09580557172493, + "learning_rate": 1.2384093655510785e-05, + "loss": 0.6772, + "step": 4059 + }, + { + "epoch": 2.204125950054289, + "grad_norm": 10.762880100606296, + "learning_rate": 1.238067807766612e-05, + "loss": 0.8104, + "step": 4060 + }, + { + "epoch": 2.2046688382193267, + "grad_norm": 10.590539714326667, + "learning_rate": 1.2377262205375398e-05, + "loss": 0.6766, + "step": 4061 + }, + { + "epoch": 2.205211726384365, + "grad_norm": 10.060689580140192, + "learning_rate": 1.2373846039061095e-05, + "loss": 1.0282, + "step": 4062 + }, + { + "epoch": 2.2057546145494027, + "grad_norm": 7.612672548635284, + "learning_rate": 1.237042957914573e-05, + "loss": 0.449, + "step": 4063 + }, + { + "epoch": 2.206297502714441, + "grad_norm": 12.239451929254052, + "learning_rate": 1.2367012826051861e-05, + "loss": 0.5385, + "step": 4064 + }, + { + "epoch": 2.2068403908794787, + "grad_norm": 14.199356139948698, + "learning_rate": 1.236359578020207e-05, + "loss": 0.7867, + "step": 4065 + }, + { + "epoch": 2.207383279044517, + "grad_norm": 8.357180617953833, + "learning_rate": 1.2360178442018989e-05, + "loss": 0.6803, + "step": 4066 + }, + { + "epoch": 2.2079261672095547, + "grad_norm": 9.203584379654275, + "learning_rate": 1.2356760811925277e-05, + "loss": 0.677, + "step": 4067 + }, + { + "epoch": 2.208469055374593, + "grad_norm": 8.063224350440471, + "learning_rate": 1.2353342890343626e-05, + "loss": 0.644, + "step": 4068 + }, + { + "epoch": 2.2090119435396307, + "grad_norm": 9.352085551954278, + "learning_rate": 1.2349924677696781e-05, + "loss": 0.9888, + "step": 4069 + }, + { + "epoch": 2.209554831704669, + "grad_norm": 10.54696840423902, + "learning_rate": 1.2346506174407505e-05, + "loss": 0.5895, + "step": 4070 + }, + { + "epoch": 2.2100977198697067, + "grad_norm": 8.568183164428444, + "learning_rate": 1.2343087380898604e-05, + "loss": 0.6287, + "step": 4071 + }, + { + "epoch": 2.210640608034745, + "grad_norm": 11.191021916852304, + "learning_rate": 1.2339668297592924e-05, + "loss": 0.8216, + "step": 4072 + }, + { + "epoch": 2.2111834961997827, + "grad_norm": 7.67795379285754, + "learning_rate": 1.233624892491334e-05, + "loss": 0.423, + "step": 4073 + }, + { + "epoch": 2.211726384364821, + "grad_norm": 8.108373133947762, + "learning_rate": 1.2332829263282764e-05, + "loss": 0.6548, + "step": 4074 + }, + { + "epoch": 2.2122692725298587, + "grad_norm": 9.0962841262343, + "learning_rate": 1.2329409313124146e-05, + "loss": 0.5506, + "step": 4075 + }, + { + "epoch": 2.212812160694897, + "grad_norm": 8.85949097192624, + "learning_rate": 1.2325989074860472e-05, + "loss": 0.5258, + "step": 4076 + }, + { + "epoch": 2.2133550488599347, + "grad_norm": 7.051914022367066, + "learning_rate": 1.2322568548914764e-05, + "loss": 0.5242, + "step": 4077 + }, + { + "epoch": 2.213897937024973, + "grad_norm": 8.287257693111446, + "learning_rate": 1.2319147735710076e-05, + "loss": 0.4769, + "step": 4078 + }, + { + "epoch": 2.2144408251900107, + "grad_norm": 9.473240892632637, + "learning_rate": 1.2315726635669498e-05, + "loss": 1.0853, + "step": 4079 + }, + { + "epoch": 2.214983713355049, + "grad_norm": 10.460874830183508, + "learning_rate": 1.231230524921616e-05, + "loss": 0.8907, + "step": 4080 + }, + { + "epoch": 2.2155266015200867, + "grad_norm": 9.17618869023913, + "learning_rate": 1.2308883576773223e-05, + "loss": 0.6355, + "step": 4081 + }, + { + "epoch": 2.216069489685125, + "grad_norm": 10.389188046047401, + "learning_rate": 1.2305461618763886e-05, + "loss": 0.9455, + "step": 4082 + }, + { + "epoch": 2.2166123778501627, + "grad_norm": 8.41220072964392, + "learning_rate": 1.230203937561138e-05, + "loss": 0.5574, + "step": 4083 + }, + { + "epoch": 2.217155266015201, + "grad_norm": 8.193062738430084, + "learning_rate": 1.2298616847738978e-05, + "loss": 0.548, + "step": 4084 + }, + { + "epoch": 2.2176981541802387, + "grad_norm": 8.55082220874788, + "learning_rate": 1.2295194035569979e-05, + "loss": 0.6087, + "step": 4085 + }, + { + "epoch": 2.218241042345277, + "grad_norm": 10.482752217531827, + "learning_rate": 1.2291770939527725e-05, + "loss": 0.5404, + "step": 4086 + }, + { + "epoch": 2.2187839305103148, + "grad_norm": 9.786686689665721, + "learning_rate": 1.2288347560035587e-05, + "loss": 0.6795, + "step": 4087 + }, + { + "epoch": 2.219326818675353, + "grad_norm": 9.696178175990593, + "learning_rate": 1.2284923897516978e-05, + "loss": 1.2474, + "step": 4088 + }, + { + "epoch": 2.2198697068403908, + "grad_norm": 9.135935376096977, + "learning_rate": 1.2281499952395336e-05, + "loss": 0.6746, + "step": 4089 + }, + { + "epoch": 2.220412595005429, + "grad_norm": 10.341327194770198, + "learning_rate": 1.2278075725094147e-05, + "loss": 0.6501, + "step": 4090 + }, + { + "epoch": 2.2209554831704668, + "grad_norm": 11.750926749777816, + "learning_rate": 1.2274651216036921e-05, + "loss": 0.6421, + "step": 4091 + }, + { + "epoch": 2.221498371335505, + "grad_norm": 15.264249369051182, + "learning_rate": 1.2271226425647207e-05, + "loss": 0.9053, + "step": 4092 + }, + { + "epoch": 2.2220412595005428, + "grad_norm": 10.996946312108877, + "learning_rate": 1.2267801354348589e-05, + "loss": 0.6776, + "step": 4093 + }, + { + "epoch": 2.222584147665581, + "grad_norm": 9.01852590664764, + "learning_rate": 1.2264376002564687e-05, + "loss": 0.7342, + "step": 4094 + }, + { + "epoch": 2.2231270358306188, + "grad_norm": 9.776717004980412, + "learning_rate": 1.2260950370719146e-05, + "loss": 0.7353, + "step": 4095 + }, + { + "epoch": 2.223669923995657, + "grad_norm": 12.928031483014212, + "learning_rate": 1.2257524459235666e-05, + "loss": 0.8649, + "step": 4096 + }, + { + "epoch": 2.2242128121606948, + "grad_norm": 9.549545586005916, + "learning_rate": 1.2254098268537962e-05, + "loss": 0.8056, + "step": 4097 + }, + { + "epoch": 2.224755700325733, + "grad_norm": 10.729492841694832, + "learning_rate": 1.2250671799049791e-05, + "loss": 0.5817, + "step": 4098 + }, + { + "epoch": 2.225298588490771, + "grad_norm": 14.13032585380651, + "learning_rate": 1.2247245051194944e-05, + "loss": 0.5998, + "step": 4099 + }, + { + "epoch": 2.225841476655809, + "grad_norm": 9.143101156244217, + "learning_rate": 1.2243818025397247e-05, + "loss": 0.5633, + "step": 4100 + }, + { + "epoch": 2.226384364820847, + "grad_norm": 9.010721857216959, + "learning_rate": 1.224039072208056e-05, + "loss": 0.5855, + "step": 4101 + }, + { + "epoch": 2.226927252985885, + "grad_norm": 11.257457693848512, + "learning_rate": 1.2236963141668778e-05, + "loss": 0.5672, + "step": 4102 + }, + { + "epoch": 2.227470141150923, + "grad_norm": 9.238231248800124, + "learning_rate": 1.2233535284585831e-05, + "loss": 0.4633, + "step": 4103 + }, + { + "epoch": 2.228013029315961, + "grad_norm": 8.93957445803232, + "learning_rate": 1.223010715125568e-05, + "loss": 0.8019, + "step": 4104 + }, + { + "epoch": 2.228555917480999, + "grad_norm": 10.869791394117428, + "learning_rate": 1.2226678742102322e-05, + "loss": 0.4235, + "step": 4105 + }, + { + "epoch": 2.229098805646037, + "grad_norm": 11.71960282550904, + "learning_rate": 1.2223250057549789e-05, + "loss": 0.8825, + "step": 4106 + }, + { + "epoch": 2.229641693811075, + "grad_norm": 9.133591574750724, + "learning_rate": 1.2219821098022146e-05, + "loss": 0.456, + "step": 4107 + }, + { + "epoch": 2.230184581976113, + "grad_norm": 7.713253727965175, + "learning_rate": 1.2216391863943492e-05, + "loss": 0.407, + "step": 4108 + }, + { + "epoch": 2.230727470141151, + "grad_norm": 8.351720726424757, + "learning_rate": 1.2212962355737959e-05, + "loss": 0.6003, + "step": 4109 + }, + { + "epoch": 2.231270358306189, + "grad_norm": 12.012727761361187, + "learning_rate": 1.2209532573829714e-05, + "loss": 0.731, + "step": 4110 + }, + { + "epoch": 2.231813246471227, + "grad_norm": 12.902339895008394, + "learning_rate": 1.220610251864296e-05, + "loss": 0.9873, + "step": 4111 + }, + { + "epoch": 2.232356134636265, + "grad_norm": 8.935941303438996, + "learning_rate": 1.220267219060193e-05, + "loss": 0.7549, + "step": 4112 + }, + { + "epoch": 2.232899022801303, + "grad_norm": 9.69521468611751, + "learning_rate": 1.2199241590130892e-05, + "loss": 0.6099, + "step": 4113 + }, + { + "epoch": 2.233441910966341, + "grad_norm": 7.437030925496973, + "learning_rate": 1.2195810717654148e-05, + "loss": 0.4743, + "step": 4114 + }, + { + "epoch": 2.233984799131379, + "grad_norm": 7.435823614079823, + "learning_rate": 1.2192379573596036e-05, + "loss": 0.5019, + "step": 4115 + }, + { + "epoch": 2.234527687296417, + "grad_norm": 7.563032718583443, + "learning_rate": 1.2188948158380927e-05, + "loss": 0.7176, + "step": 4116 + }, + { + "epoch": 2.235070575461455, + "grad_norm": 10.449779515276235, + "learning_rate": 1.2185516472433218e-05, + "loss": 0.7921, + "step": 4117 + }, + { + "epoch": 2.235613463626493, + "grad_norm": 9.088046167756731, + "learning_rate": 1.2182084516177351e-05, + "loss": 0.811, + "step": 4118 + }, + { + "epoch": 2.236156351791531, + "grad_norm": 9.664269830378425, + "learning_rate": 1.217865229003779e-05, + "loss": 0.8024, + "step": 4119 + }, + { + "epoch": 2.236699239956569, + "grad_norm": 8.055354430405984, + "learning_rate": 1.2175219794439044e-05, + "loss": 0.5747, + "step": 4120 + }, + { + "epoch": 2.237242128121607, + "grad_norm": 6.428342674450092, + "learning_rate": 1.2171787029805646e-05, + "loss": 0.5673, + "step": 4121 + }, + { + "epoch": 2.237785016286645, + "grad_norm": 7.502074776894542, + "learning_rate": 1.2168353996562167e-05, + "loss": 0.566, + "step": 4122 + }, + { + "epoch": 2.238327904451683, + "grad_norm": 10.380110541434405, + "learning_rate": 1.2164920695133212e-05, + "loss": 1.0609, + "step": 4123 + }, + { + "epoch": 2.238870792616721, + "grad_norm": 13.199317832444493, + "learning_rate": 1.2161487125943413e-05, + "loss": 0.7662, + "step": 4124 + }, + { + "epoch": 2.239413680781759, + "grad_norm": 9.067564902950002, + "learning_rate": 1.2158053289417439e-05, + "loss": 0.6325, + "step": 4125 + }, + { + "epoch": 2.239956568946797, + "grad_norm": 8.074724230493178, + "learning_rate": 1.215461918598e-05, + "loss": 0.6854, + "step": 4126 + }, + { + "epoch": 2.240499457111835, + "grad_norm": 7.916509466850638, + "learning_rate": 1.2151184816055822e-05, + "loss": 0.3852, + "step": 4127 + }, + { + "epoch": 2.241042345276873, + "grad_norm": 8.647089691838135, + "learning_rate": 1.214775018006968e-05, + "loss": 0.6291, + "step": 4128 + }, + { + "epoch": 2.241585233441911, + "grad_norm": 9.238872667328287, + "learning_rate": 1.214431527844637e-05, + "loss": 0.6117, + "step": 4129 + }, + { + "epoch": 2.242128121606949, + "grad_norm": 12.537117099920934, + "learning_rate": 1.214088011161073e-05, + "loss": 0.745, + "step": 4130 + }, + { + "epoch": 2.242671009771987, + "grad_norm": 9.99529796933279, + "learning_rate": 1.2137444679987627e-05, + "loss": 0.612, + "step": 4131 + }, + { + "epoch": 2.243213897937025, + "grad_norm": 6.673816908475189, + "learning_rate": 1.2134008984001959e-05, + "loss": 0.3219, + "step": 4132 + }, + { + "epoch": 2.243756786102063, + "grad_norm": 13.936566760747976, + "learning_rate": 1.2130573024078656e-05, + "loss": 0.8689, + "step": 4133 + }, + { + "epoch": 2.244299674267101, + "grad_norm": 7.800578548383095, + "learning_rate": 1.2127136800642689e-05, + "loss": 0.3804, + "step": 4134 + }, + { + "epoch": 2.244842562432139, + "grad_norm": 9.283785411998215, + "learning_rate": 1.2123700314119052e-05, + "loss": 0.699, + "step": 4135 + }, + { + "epoch": 2.245385450597177, + "grad_norm": 8.613596910195628, + "learning_rate": 1.2120263564932775e-05, + "loss": 0.7155, + "step": 4136 + }, + { + "epoch": 2.245928338762215, + "grad_norm": 8.522662049603209, + "learning_rate": 1.2116826553508923e-05, + "loss": 0.5705, + "step": 4137 + }, + { + "epoch": 2.246471226927253, + "grad_norm": 9.023598184433123, + "learning_rate": 1.2113389280272587e-05, + "loss": 0.4392, + "step": 4138 + }, + { + "epoch": 2.247014115092291, + "grad_norm": 10.959276458237868, + "learning_rate": 1.2109951745648894e-05, + "loss": 0.7827, + "step": 4139 + }, + { + "epoch": 2.247557003257329, + "grad_norm": 11.660193747264707, + "learning_rate": 1.210651395006301e-05, + "loss": 0.9824, + "step": 4140 + }, + { + "epoch": 2.248099891422367, + "grad_norm": 18.121219443467986, + "learning_rate": 1.2103075893940122e-05, + "loss": 1.2339, + "step": 4141 + }, + { + "epoch": 2.248642779587405, + "grad_norm": 15.059306090767514, + "learning_rate": 1.2099637577705457e-05, + "loss": 0.7434, + "step": 4142 + }, + { + "epoch": 2.249185667752443, + "grad_norm": 10.999940106944315, + "learning_rate": 1.2096199001784268e-05, + "loss": 0.6422, + "step": 4143 + }, + { + "epoch": 2.249728555917481, + "grad_norm": 6.704480729179984, + "learning_rate": 1.2092760166601848e-05, + "loss": 0.4829, + "step": 4144 + }, + { + "epoch": 2.250271444082519, + "grad_norm": 11.849496386091012, + "learning_rate": 1.2089321072583512e-05, + "loss": 0.8969, + "step": 4145 + }, + { + "epoch": 2.250814332247557, + "grad_norm": 8.596596068581539, + "learning_rate": 1.2085881720154617e-05, + "loss": 0.6148, + "step": 4146 + }, + { + "epoch": 2.251357220412595, + "grad_norm": 7.50181808267645, + "learning_rate": 1.2082442109740548e-05, + "loss": 0.5499, + "step": 4147 + }, + { + "epoch": 2.251900108577633, + "grad_norm": 10.178327453067736, + "learning_rate": 1.207900224176672e-05, + "loss": 0.5582, + "step": 4148 + }, + { + "epoch": 2.252442996742671, + "grad_norm": 10.74128468394649, + "learning_rate": 1.2075562116658581e-05, + "loss": 0.8023, + "step": 4149 + }, + { + "epoch": 2.252985884907709, + "grad_norm": 10.195557872174708, + "learning_rate": 1.207212173484161e-05, + "loss": 0.6092, + "step": 4150 + }, + { + "epoch": 2.253528773072747, + "grad_norm": 9.843275181764184, + "learning_rate": 1.2068681096741322e-05, + "loss": 0.8451, + "step": 4151 + }, + { + "epoch": 2.254071661237785, + "grad_norm": 7.859122806337064, + "learning_rate": 1.2065240202783259e-05, + "loss": 0.6158, + "step": 4152 + }, + { + "epoch": 2.254614549402823, + "grad_norm": 11.4233860908698, + "learning_rate": 1.2061799053392994e-05, + "loss": 0.6552, + "step": 4153 + }, + { + "epoch": 2.255157437567861, + "grad_norm": 7.339673441521232, + "learning_rate": 1.2058357648996138e-05, + "loss": 0.5103, + "step": 4154 + }, + { + "epoch": 2.255700325732899, + "grad_norm": 11.603888679814519, + "learning_rate": 1.2054915990018326e-05, + "loss": 0.8117, + "step": 4155 + }, + { + "epoch": 2.256243213897937, + "grad_norm": 9.757430713637703, + "learning_rate": 1.2051474076885232e-05, + "loss": 0.6942, + "step": 4156 + }, + { + "epoch": 2.256786102062975, + "grad_norm": 10.286384353132442, + "learning_rate": 1.2048031910022553e-05, + "loss": 0.8958, + "step": 4157 + }, + { + "epoch": 2.257328990228013, + "grad_norm": 7.6945320206947185, + "learning_rate": 1.2044589489856023e-05, + "loss": 0.4508, + "step": 4158 + }, + { + "epoch": 2.257871878393051, + "grad_norm": 9.546205961324182, + "learning_rate": 1.2041146816811406e-05, + "loss": 0.6889, + "step": 4159 + }, + { + "epoch": 2.258414766558089, + "grad_norm": 6.440627468414342, + "learning_rate": 1.2037703891314498e-05, + "loss": 0.5146, + "step": 4160 + }, + { + "epoch": 2.258957654723127, + "grad_norm": 8.769926030437105, + "learning_rate": 1.2034260713791124e-05, + "loss": 0.539, + "step": 4161 + }, + { + "epoch": 2.259500542888165, + "grad_norm": 8.385311572142552, + "learning_rate": 1.2030817284667145e-05, + "loss": 0.6471, + "step": 4162 + }, + { + "epoch": 2.260043431053203, + "grad_norm": 10.687000132395355, + "learning_rate": 1.2027373604368446e-05, + "loss": 0.4047, + "step": 4163 + }, + { + "epoch": 2.260586319218241, + "grad_norm": 9.22667613364525, + "learning_rate": 1.2023929673320952e-05, + "loss": 0.6039, + "step": 4164 + }, + { + "epoch": 2.261129207383279, + "grad_norm": 8.822764152626888, + "learning_rate": 1.2020485491950607e-05, + "loss": 0.5333, + "step": 4165 + }, + { + "epoch": 2.261672095548317, + "grad_norm": 8.505680622817092, + "learning_rate": 1.2017041060683394e-05, + "loss": 0.6679, + "step": 4166 + }, + { + "epoch": 2.262214983713355, + "grad_norm": 12.980914094830798, + "learning_rate": 1.2013596379945331e-05, + "loss": 0.7265, + "step": 4167 + }, + { + "epoch": 2.262757871878393, + "grad_norm": 12.278007785084847, + "learning_rate": 1.201015145016246e-05, + "loss": 0.7175, + "step": 4168 + }, + { + "epoch": 2.263300760043431, + "grad_norm": 8.871985541095814, + "learning_rate": 1.2006706271760851e-05, + "loss": 0.4937, + "step": 4169 + }, + { + "epoch": 2.263843648208469, + "grad_norm": 10.070308144714824, + "learning_rate": 1.2003260845166613e-05, + "loss": 0.5994, + "step": 4170 + }, + { + "epoch": 2.264386536373507, + "grad_norm": 8.846376302708403, + "learning_rate": 1.1999815170805882e-05, + "loss": 0.5795, + "step": 4171 + }, + { + "epoch": 2.264929424538545, + "grad_norm": 11.984661783852376, + "learning_rate": 1.199636924910482e-05, + "loss": 1.1718, + "step": 4172 + }, + { + "epoch": 2.265472312703583, + "grad_norm": 10.035244756765978, + "learning_rate": 1.1992923080489629e-05, + "loss": 0.8101, + "step": 4173 + }, + { + "epoch": 2.266015200868621, + "grad_norm": 7.429999917349929, + "learning_rate": 1.1989476665386537e-05, + "loss": 0.4121, + "step": 4174 + }, + { + "epoch": 2.266558089033659, + "grad_norm": 8.464944728590373, + "learning_rate": 1.1986030004221802e-05, + "loss": 0.5553, + "step": 4175 + }, + { + "epoch": 2.267100977198697, + "grad_norm": 11.414355640267063, + "learning_rate": 1.1982583097421707e-05, + "loss": 0.8705, + "step": 4176 + }, + { + "epoch": 2.267643865363735, + "grad_norm": 10.094383752442985, + "learning_rate": 1.1979135945412576e-05, + "loss": 0.5512, + "step": 4177 + }, + { + "epoch": 2.268186753528773, + "grad_norm": 10.131657295101125, + "learning_rate": 1.1975688548620756e-05, + "loss": 0.5678, + "step": 4178 + }, + { + "epoch": 2.268729641693811, + "grad_norm": 11.675191701181866, + "learning_rate": 1.197224090747263e-05, + "loss": 0.5967, + "step": 4179 + }, + { + "epoch": 2.269272529858849, + "grad_norm": 9.879570041096912, + "learning_rate": 1.1968793022394603e-05, + "loss": 1.0245, + "step": 4180 + }, + { + "epoch": 2.269815418023887, + "grad_norm": 12.255119190489571, + "learning_rate": 1.1965344893813117e-05, + "loss": 0.9963, + "step": 4181 + }, + { + "epoch": 2.270358306188925, + "grad_norm": 8.870271458097042, + "learning_rate": 1.1961896522154643e-05, + "loss": 0.3658, + "step": 4182 + }, + { + "epoch": 2.270901194353963, + "grad_norm": 9.50086592138889, + "learning_rate": 1.1958447907845684e-05, + "loss": 0.588, + "step": 4183 + }, + { + "epoch": 2.271444082519001, + "grad_norm": 8.92062913157286, + "learning_rate": 1.1954999051312762e-05, + "loss": 0.6038, + "step": 4184 + }, + { + "epoch": 2.271986970684039, + "grad_norm": 11.511359098427171, + "learning_rate": 1.1951549952982442e-05, + "loss": 0.7837, + "step": 4185 + }, + { + "epoch": 2.272529858849077, + "grad_norm": 7.7230189119336075, + "learning_rate": 1.1948100613281315e-05, + "loss": 0.5235, + "step": 4186 + }, + { + "epoch": 2.273072747014115, + "grad_norm": 8.85303744288786, + "learning_rate": 1.1944651032635997e-05, + "loss": 0.5194, + "step": 4187 + }, + { + "epoch": 2.273615635179153, + "grad_norm": 9.931552012214697, + "learning_rate": 1.1941201211473142e-05, + "loss": 0.6419, + "step": 4188 + }, + { + "epoch": 2.274158523344191, + "grad_norm": 11.635637385284054, + "learning_rate": 1.1937751150219428e-05, + "loss": 0.7948, + "step": 4189 + }, + { + "epoch": 2.274701411509229, + "grad_norm": 7.783494952264655, + "learning_rate": 1.1934300849301561e-05, + "loss": 0.568, + "step": 4190 + }, + { + "epoch": 2.275244299674267, + "grad_norm": 7.979443655042872, + "learning_rate": 1.1930850309146283e-05, + "loss": 0.531, + "step": 4191 + }, + { + "epoch": 2.2757871878393052, + "grad_norm": 12.081321074807914, + "learning_rate": 1.1927399530180359e-05, + "loss": 0.9917, + "step": 4192 + }, + { + "epoch": 2.276330076004343, + "grad_norm": 9.376910406413142, + "learning_rate": 1.1923948512830595e-05, + "loss": 0.6587, + "step": 4193 + }, + { + "epoch": 2.2768729641693812, + "grad_norm": 9.576589478578342, + "learning_rate": 1.1920497257523809e-05, + "loss": 0.5703, + "step": 4194 + }, + { + "epoch": 2.277415852334419, + "grad_norm": 9.129784891039465, + "learning_rate": 1.1917045764686863e-05, + "loss": 0.668, + "step": 4195 + }, + { + "epoch": 2.2779587404994572, + "grad_norm": 11.723388678968943, + "learning_rate": 1.191359403474664e-05, + "loss": 0.6293, + "step": 4196 + }, + { + "epoch": 2.278501628664495, + "grad_norm": 10.662032565840782, + "learning_rate": 1.1910142068130058e-05, + "loss": 1.0251, + "step": 4197 + }, + { + "epoch": 2.2790445168295332, + "grad_norm": 5.976705265828664, + "learning_rate": 1.1906689865264058e-05, + "loss": 0.3892, + "step": 4198 + }, + { + "epoch": 2.279587404994571, + "grad_norm": 9.77428172632025, + "learning_rate": 1.190323742657562e-05, + "loss": 0.8023, + "step": 4199 + }, + { + "epoch": 2.2801302931596092, + "grad_norm": 9.701826347676784, + "learning_rate": 1.1899784752491743e-05, + "loss": 0.6668, + "step": 4200 + }, + { + "epoch": 2.280673181324647, + "grad_norm": 10.939483572398276, + "learning_rate": 1.1896331843439459e-05, + "loss": 0.8751, + "step": 4201 + }, + { + "epoch": 2.2812160694896852, + "grad_norm": 12.656192710489524, + "learning_rate": 1.1892878699845832e-05, + "loss": 0.7874, + "step": 4202 + }, + { + "epoch": 2.281758957654723, + "grad_norm": 8.821175958313784, + "learning_rate": 1.1889425322137948e-05, + "loss": 0.7599, + "step": 4203 + }, + { + "epoch": 2.2823018458197613, + "grad_norm": 11.535196460357659, + "learning_rate": 1.188597171074293e-05, + "loss": 0.7448, + "step": 4204 + }, + { + "epoch": 2.282844733984799, + "grad_norm": 9.221866003570257, + "learning_rate": 1.1882517866087926e-05, + "loss": 0.7176, + "step": 4205 + }, + { + "epoch": 2.2833876221498373, + "grad_norm": 10.791249990779047, + "learning_rate": 1.1879063788600113e-05, + "loss": 0.6905, + "step": 4206 + }, + { + "epoch": 2.283930510314875, + "grad_norm": 11.953403018554845, + "learning_rate": 1.1875609478706694e-05, + "loss": 0.7006, + "step": 4207 + }, + { + "epoch": 2.2844733984799133, + "grad_norm": 10.830659589867402, + "learning_rate": 1.1872154936834905e-05, + "loss": 0.6872, + "step": 4208 + }, + { + "epoch": 2.285016286644951, + "grad_norm": 10.8147041656128, + "learning_rate": 1.186870016341201e-05, + "loss": 0.6429, + "step": 4209 + }, + { + "epoch": 2.2855591748099893, + "grad_norm": 10.70508214004579, + "learning_rate": 1.18652451588653e-05, + "loss": 0.9063, + "step": 4210 + }, + { + "epoch": 2.286102062975027, + "grad_norm": 8.333893553672915, + "learning_rate": 1.1861789923622099e-05, + "loss": 0.5245, + "step": 4211 + }, + { + "epoch": 2.2866449511400653, + "grad_norm": 9.62415991156623, + "learning_rate": 1.1858334458109747e-05, + "loss": 0.734, + "step": 4212 + }, + { + "epoch": 2.287187839305103, + "grad_norm": 15.893278058161348, + "learning_rate": 1.1854878762755633e-05, + "loss": 0.898, + "step": 4213 + }, + { + "epoch": 2.2877307274701413, + "grad_norm": 8.12705480823925, + "learning_rate": 1.1851422837987159e-05, + "loss": 0.5586, + "step": 4214 + }, + { + "epoch": 2.288273615635179, + "grad_norm": 19.440766463719637, + "learning_rate": 1.1847966684231754e-05, + "loss": 1.0309, + "step": 4215 + }, + { + "epoch": 2.2888165038002173, + "grad_norm": 10.590183829770972, + "learning_rate": 1.1844510301916883e-05, + "loss": 0.8545, + "step": 4216 + }, + { + "epoch": 2.289359391965255, + "grad_norm": 9.825370158957117, + "learning_rate": 1.184105369147004e-05, + "loss": 0.7426, + "step": 4217 + }, + { + "epoch": 2.2899022801302933, + "grad_norm": 10.341219416998992, + "learning_rate": 1.1837596853318742e-05, + "loss": 0.8584, + "step": 4218 + }, + { + "epoch": 2.290445168295331, + "grad_norm": 11.121322184848584, + "learning_rate": 1.1834139787890538e-05, + "loss": 0.6598, + "step": 4219 + }, + { + "epoch": 2.2909880564603693, + "grad_norm": 9.600608599739235, + "learning_rate": 1.1830682495613e-05, + "loss": 0.5885, + "step": 4220 + }, + { + "epoch": 2.291530944625407, + "grad_norm": 10.517374576113433, + "learning_rate": 1.1827224976913732e-05, + "loss": 0.751, + "step": 4221 + }, + { + "epoch": 2.2920738327904453, + "grad_norm": 9.666522143402908, + "learning_rate": 1.1823767232220369e-05, + "loss": 0.7602, + "step": 4222 + }, + { + "epoch": 2.292616720955483, + "grad_norm": 9.390811834354404, + "learning_rate": 1.1820309261960567e-05, + "loss": 0.6187, + "step": 4223 + }, + { + "epoch": 2.2931596091205213, + "grad_norm": 8.772653626817414, + "learning_rate": 1.1816851066562017e-05, + "loss": 0.5221, + "step": 4224 + }, + { + "epoch": 2.293702497285559, + "grad_norm": 11.957278593216312, + "learning_rate": 1.1813392646452429e-05, + "loss": 0.9231, + "step": 4225 + }, + { + "epoch": 2.2942453854505973, + "grad_norm": 9.58191568158383, + "learning_rate": 1.1809934002059547e-05, + "loss": 0.6747, + "step": 4226 + }, + { + "epoch": 2.294788273615635, + "grad_norm": 8.242154702381846, + "learning_rate": 1.1806475133811146e-05, + "loss": 0.4053, + "step": 4227 + }, + { + "epoch": 2.2953311617806733, + "grad_norm": 8.83768826569311, + "learning_rate": 1.1803016042135021e-05, + "loss": 0.4177, + "step": 4228 + }, + { + "epoch": 2.295874049945711, + "grad_norm": 7.496802017426867, + "learning_rate": 1.1799556727458996e-05, + "loss": 0.6402, + "step": 4229 + }, + { + "epoch": 2.2964169381107493, + "grad_norm": 10.15598151517949, + "learning_rate": 1.1796097190210927e-05, + "loss": 0.6301, + "step": 4230 + }, + { + "epoch": 2.296959826275787, + "grad_norm": 9.100517578174307, + "learning_rate": 1.1792637430818695e-05, + "loss": 0.6059, + "step": 4231 + }, + { + "epoch": 2.2975027144408253, + "grad_norm": 9.341766570492698, + "learning_rate": 1.178917744971021e-05, + "loss": 0.7462, + "step": 4232 + }, + { + "epoch": 2.298045602605863, + "grad_norm": 9.790098160117006, + "learning_rate": 1.1785717247313407e-05, + "loss": 0.7422, + "step": 4233 + }, + { + "epoch": 2.2985884907709013, + "grad_norm": 11.785065190423747, + "learning_rate": 1.1782256824056249e-05, + "loss": 0.8544, + "step": 4234 + }, + { + "epoch": 2.299131378935939, + "grad_norm": 11.165339576639699, + "learning_rate": 1.1778796180366722e-05, + "loss": 0.6492, + "step": 4235 + }, + { + "epoch": 2.2996742671009773, + "grad_norm": 9.026810304485423, + "learning_rate": 1.1775335316672854e-05, + "loss": 0.751, + "step": 4236 + }, + { + "epoch": 2.300217155266015, + "grad_norm": 10.085898624762818, + "learning_rate": 1.1771874233402682e-05, + "loss": 0.6052, + "step": 4237 + }, + { + "epoch": 2.3007600434310533, + "grad_norm": 9.302471978496024, + "learning_rate": 1.176841293098428e-05, + "loss": 0.7269, + "step": 4238 + }, + { + "epoch": 2.301302931596091, + "grad_norm": 11.406933536282311, + "learning_rate": 1.176495140984575e-05, + "loss": 0.7287, + "step": 4239 + }, + { + "epoch": 2.3018458197611293, + "grad_norm": 8.172755818720033, + "learning_rate": 1.1761489670415218e-05, + "loss": 0.4692, + "step": 4240 + }, + { + "epoch": 2.302388707926167, + "grad_norm": 11.729275499093248, + "learning_rate": 1.1758027713120833e-05, + "loss": 0.8301, + "step": 4241 + }, + { + "epoch": 2.3029315960912053, + "grad_norm": 7.31457707003035, + "learning_rate": 1.1754565538390782e-05, + "loss": 0.5372, + "step": 4242 + }, + { + "epoch": 2.303474484256243, + "grad_norm": 10.15517896282683, + "learning_rate": 1.1751103146653265e-05, + "loss": 0.8744, + "step": 4243 + }, + { + "epoch": 2.3040173724212814, + "grad_norm": 7.780368081203029, + "learning_rate": 1.1747640538336525e-05, + "loss": 0.4743, + "step": 4244 + }, + { + "epoch": 2.304560260586319, + "grad_norm": 7.814816774582812, + "learning_rate": 1.1744177713868814e-05, + "loss": 0.628, + "step": 4245 + }, + { + "epoch": 2.3051031487513574, + "grad_norm": 8.531245398152228, + "learning_rate": 1.1740714673678425e-05, + "loss": 0.4419, + "step": 4246 + }, + { + "epoch": 2.305646036916395, + "grad_norm": 8.091385303458143, + "learning_rate": 1.1737251418193669e-05, + "loss": 0.6249, + "step": 4247 + }, + { + "epoch": 2.3061889250814334, + "grad_norm": 10.253051982681455, + "learning_rate": 1.1733787947842892e-05, + "loss": 0.8892, + "step": 4248 + }, + { + "epoch": 2.306731813246471, + "grad_norm": 8.95406502670429, + "learning_rate": 1.1730324263054453e-05, + "loss": 0.7452, + "step": 4249 + }, + { + "epoch": 2.3072747014115094, + "grad_norm": 10.491159901998326, + "learning_rate": 1.1726860364256756e-05, + "loss": 0.4995, + "step": 4250 + }, + { + "epoch": 2.307817589576547, + "grad_norm": 9.241135791509704, + "learning_rate": 1.1723396251878214e-05, + "loss": 0.7167, + "step": 4251 + }, + { + "epoch": 2.3083604777415854, + "grad_norm": 9.466621171091207, + "learning_rate": 1.1719931926347279e-05, + "loss": 0.5849, + "step": 4252 + }, + { + "epoch": 2.308903365906623, + "grad_norm": 10.933415330757905, + "learning_rate": 1.1716467388092423e-05, + "loss": 1.0411, + "step": 4253 + }, + { + "epoch": 2.3094462540716614, + "grad_norm": 10.009779873858768, + "learning_rate": 1.1713002637542146e-05, + "loss": 0.5456, + "step": 4254 + }, + { + "epoch": 2.309989142236699, + "grad_norm": 11.678038797864954, + "learning_rate": 1.1709537675124966e-05, + "loss": 0.5537, + "step": 4255 + }, + { + "epoch": 2.3105320304017374, + "grad_norm": 11.578035258644414, + "learning_rate": 1.1706072501269448e-05, + "loss": 0.7804, + "step": 4256 + }, + { + "epoch": 2.311074918566775, + "grad_norm": 8.678307181037598, + "learning_rate": 1.1702607116404163e-05, + "loss": 0.5378, + "step": 4257 + }, + { + "epoch": 2.3116178067318134, + "grad_norm": 8.487670920129334, + "learning_rate": 1.1699141520957716e-05, + "loss": 0.558, + "step": 4258 + }, + { + "epoch": 2.312160694896851, + "grad_norm": 10.92371973791016, + "learning_rate": 1.1695675715358737e-05, + "loss": 0.6615, + "step": 4259 + }, + { + "epoch": 2.3127035830618894, + "grad_norm": 9.360985487621843, + "learning_rate": 1.169220970003588e-05, + "loss": 0.5896, + "step": 4260 + }, + { + "epoch": 2.313246471226927, + "grad_norm": 10.382914656511591, + "learning_rate": 1.1688743475417836e-05, + "loss": 0.8869, + "step": 4261 + }, + { + "epoch": 2.3137893593919654, + "grad_norm": 10.27408683638607, + "learning_rate": 1.1685277041933303e-05, + "loss": 1.0929, + "step": 4262 + }, + { + "epoch": 2.314332247557003, + "grad_norm": 10.738574683790112, + "learning_rate": 1.168181040001102e-05, + "loss": 0.845, + "step": 4263 + }, + { + "epoch": 2.3148751357220414, + "grad_norm": 11.783082834204276, + "learning_rate": 1.1678343550079751e-05, + "loss": 0.8358, + "step": 4264 + }, + { + "epoch": 2.315418023887079, + "grad_norm": 9.277577078134165, + "learning_rate": 1.1674876492568272e-05, + "loss": 0.4774, + "step": 4265 + }, + { + "epoch": 2.3159609120521174, + "grad_norm": 10.232983577395474, + "learning_rate": 1.1671409227905401e-05, + "loss": 0.6358, + "step": 4266 + }, + { + "epoch": 2.316503800217155, + "grad_norm": 7.971516257815676, + "learning_rate": 1.1667941756519973e-05, + "loss": 0.5724, + "step": 4267 + }, + { + "epoch": 2.3170466883821934, + "grad_norm": 6.657326238165826, + "learning_rate": 1.1664474078840848e-05, + "loss": 0.3377, + "step": 4268 + }, + { + "epoch": 2.317589576547231, + "grad_norm": 9.831416988328497, + "learning_rate": 1.1661006195296918e-05, + "loss": 0.7115, + "step": 4269 + }, + { + "epoch": 2.3181324647122694, + "grad_norm": 7.2473297381695225, + "learning_rate": 1.1657538106317094e-05, + "loss": 0.4828, + "step": 4270 + }, + { + "epoch": 2.318675352877307, + "grad_norm": 11.474212650892378, + "learning_rate": 1.1654069812330315e-05, + "loss": 0.66, + "step": 4271 + }, + { + "epoch": 2.3192182410423454, + "grad_norm": 9.354650178037069, + "learning_rate": 1.1650601313765547e-05, + "loss": 0.6651, + "step": 4272 + }, + { + "epoch": 2.319761129207383, + "grad_norm": 8.945160340560053, + "learning_rate": 1.1647132611051779e-05, + "loss": 0.6422, + "step": 4273 + }, + { + "epoch": 2.3203040173724214, + "grad_norm": 8.790681463696455, + "learning_rate": 1.164366370461802e-05, + "loss": 0.6387, + "step": 4274 + }, + { + "epoch": 2.320846905537459, + "grad_norm": 8.950872212457588, + "learning_rate": 1.1640194594893317e-05, + "loss": 0.4455, + "step": 4275 + }, + { + "epoch": 2.3213897937024974, + "grad_norm": 9.879748686678086, + "learning_rate": 1.1636725282306733e-05, + "loss": 0.5539, + "step": 4276 + }, + { + "epoch": 2.321932681867535, + "grad_norm": 8.351885716181597, + "learning_rate": 1.1633255767287356e-05, + "loss": 0.4273, + "step": 4277 + }, + { + "epoch": 2.3224755700325734, + "grad_norm": 12.165689696691171, + "learning_rate": 1.1629786050264304e-05, + "loss": 0.6877, + "step": 4278 + }, + { + "epoch": 2.323018458197611, + "grad_norm": 10.70545900286579, + "learning_rate": 1.1626316131666714e-05, + "loss": 0.6667, + "step": 4279 + }, + { + "epoch": 2.3235613463626494, + "grad_norm": 7.521776321190906, + "learning_rate": 1.1622846011923756e-05, + "loss": 0.6538, + "step": 4280 + }, + { + "epoch": 2.324104234527687, + "grad_norm": 10.429691195494106, + "learning_rate": 1.1619375691464614e-05, + "loss": 0.7918, + "step": 4281 + }, + { + "epoch": 2.3246471226927254, + "grad_norm": 13.40241359482942, + "learning_rate": 1.1615905170718507e-05, + "loss": 0.8492, + "step": 4282 + }, + { + "epoch": 2.3251900108577632, + "grad_norm": 9.180074174181755, + "learning_rate": 1.1612434450114674e-05, + "loss": 0.9197, + "step": 4283 + }, + { + "epoch": 2.3257328990228014, + "grad_norm": 8.662388876238891, + "learning_rate": 1.1608963530082375e-05, + "loss": 0.4911, + "step": 4284 + }, + { + "epoch": 2.3262757871878392, + "grad_norm": 12.915716152146215, + "learning_rate": 1.1605492411050907e-05, + "loss": 1.2342, + "step": 4285 + }, + { + "epoch": 2.3268186753528775, + "grad_norm": 15.064512853420336, + "learning_rate": 1.1602021093449577e-05, + "loss": 0.9441, + "step": 4286 + }, + { + "epoch": 2.3273615635179152, + "grad_norm": 8.152078540931088, + "learning_rate": 1.159854957770772e-05, + "loss": 0.4407, + "step": 4287 + }, + { + "epoch": 2.3279044516829535, + "grad_norm": 13.440334362031807, + "learning_rate": 1.1595077864254708e-05, + "loss": 1.0186, + "step": 4288 + }, + { + "epoch": 2.3284473398479912, + "grad_norm": 14.960404464650807, + "learning_rate": 1.1591605953519925e-05, + "loss": 0.854, + "step": 4289 + }, + { + "epoch": 2.3289902280130295, + "grad_norm": 10.871106066518651, + "learning_rate": 1.1588133845932779e-05, + "loss": 0.8522, + "step": 4290 + }, + { + "epoch": 2.3295331161780672, + "grad_norm": 9.218798486570465, + "learning_rate": 1.1584661541922706e-05, + "loss": 0.5529, + "step": 4291 + }, + { + "epoch": 2.3300760043431055, + "grad_norm": 9.644924391242567, + "learning_rate": 1.158118904191917e-05, + "loss": 0.619, + "step": 4292 + }, + { + "epoch": 2.3306188925081432, + "grad_norm": 7.573505575520159, + "learning_rate": 1.1577716346351655e-05, + "loss": 0.5931, + "step": 4293 + }, + { + "epoch": 2.3311617806731815, + "grad_norm": 11.909869183875244, + "learning_rate": 1.1574243455649666e-05, + "loss": 0.8922, + "step": 4294 + }, + { + "epoch": 2.3317046688382193, + "grad_norm": 13.963856423636752, + "learning_rate": 1.1570770370242737e-05, + "loss": 0.9243, + "step": 4295 + }, + { + "epoch": 2.3322475570032575, + "grad_norm": 8.732100079298183, + "learning_rate": 1.1567297090560425e-05, + "loss": 0.5075, + "step": 4296 + }, + { + "epoch": 2.3327904451682953, + "grad_norm": 8.823180159543773, + "learning_rate": 1.1563823617032313e-05, + "loss": 0.5471, + "step": 4297 + }, + { + "epoch": 2.3333333333333335, + "grad_norm": 8.566225161651246, + "learning_rate": 1.1560349950088001e-05, + "loss": 0.5719, + "step": 4298 + }, + { + "epoch": 2.3338762214983713, + "grad_norm": 10.549198011052681, + "learning_rate": 1.155687609015712e-05, + "loss": 0.4775, + "step": 4299 + }, + { + "epoch": 2.3344191096634095, + "grad_norm": 9.79276879932382, + "learning_rate": 1.1553402037669326e-05, + "loss": 0.5429, + "step": 4300 + }, + { + "epoch": 2.3349619978284473, + "grad_norm": 12.065070131397079, + "learning_rate": 1.154992779305429e-05, + "loss": 0.9434, + "step": 4301 + }, + { + "epoch": 2.3355048859934855, + "grad_norm": 11.007790011549671, + "learning_rate": 1.1546453356741714e-05, + "loss": 0.6199, + "step": 4302 + }, + { + "epoch": 2.3360477741585233, + "grad_norm": 12.466762411503309, + "learning_rate": 1.1542978729161326e-05, + "loss": 0.7941, + "step": 4303 + }, + { + "epoch": 2.3365906623235615, + "grad_norm": 11.55197462123976, + "learning_rate": 1.1539503910742866e-05, + "loss": 0.6766, + "step": 4304 + }, + { + "epoch": 2.3371335504885993, + "grad_norm": 8.462143561329682, + "learning_rate": 1.153602890191611e-05, + "loss": 0.5183, + "step": 4305 + }, + { + "epoch": 2.3376764386536375, + "grad_norm": 13.818285051654057, + "learning_rate": 1.1532553703110852e-05, + "loss": 0.8363, + "step": 4306 + }, + { + "epoch": 2.3382193268186753, + "grad_norm": 8.676120343621662, + "learning_rate": 1.1529078314756905e-05, + "loss": 0.6976, + "step": 4307 + }, + { + "epoch": 2.3387622149837135, + "grad_norm": 10.024443385726268, + "learning_rate": 1.1525602737284119e-05, + "loss": 0.4879, + "step": 4308 + }, + { + "epoch": 2.3393051031487513, + "grad_norm": 11.965265440206402, + "learning_rate": 1.1522126971122355e-05, + "loss": 0.8655, + "step": 4309 + }, + { + "epoch": 2.3398479913137895, + "grad_norm": 9.503916256652449, + "learning_rate": 1.15186510167015e-05, + "loss": 0.8966, + "step": 4310 + }, + { + "epoch": 2.3403908794788273, + "grad_norm": 11.942859003506952, + "learning_rate": 1.1515174874451466e-05, + "loss": 0.7769, + "step": 4311 + }, + { + "epoch": 2.3409337676438655, + "grad_norm": 9.50234633554061, + "learning_rate": 1.1511698544802193e-05, + "loss": 0.552, + "step": 4312 + }, + { + "epoch": 2.3414766558089033, + "grad_norm": 8.495987313736284, + "learning_rate": 1.1508222028183632e-05, + "loss": 0.8566, + "step": 4313 + }, + { + "epoch": 2.3420195439739415, + "grad_norm": 11.854978708479615, + "learning_rate": 1.1504745325025766e-05, + "loss": 0.6736, + "step": 4314 + }, + { + "epoch": 2.3425624321389793, + "grad_norm": 11.169148481350696, + "learning_rate": 1.1501268435758602e-05, + "loss": 0.8075, + "step": 4315 + }, + { + "epoch": 2.3431053203040175, + "grad_norm": 11.983506203130691, + "learning_rate": 1.1497791360812163e-05, + "loss": 0.6407, + "step": 4316 + }, + { + "epoch": 2.3436482084690553, + "grad_norm": 8.507852970222672, + "learning_rate": 1.1494314100616502e-05, + "loss": 0.6247, + "step": 4317 + }, + { + "epoch": 2.3441910966340935, + "grad_norm": 19.964294777139003, + "learning_rate": 1.1490836655601692e-05, + "loss": 0.9219, + "step": 4318 + }, + { + "epoch": 2.3447339847991313, + "grad_norm": 8.202074828225685, + "learning_rate": 1.1487359026197828e-05, + "loss": 0.4887, + "step": 4319 + }, + { + "epoch": 2.3452768729641695, + "grad_norm": 10.22404099258637, + "learning_rate": 1.1483881212835029e-05, + "loss": 0.4028, + "step": 4320 + }, + { + "epoch": 2.3458197611292073, + "grad_norm": 8.30587529218613, + "learning_rate": 1.1480403215943436e-05, + "loss": 0.4047, + "step": 4321 + }, + { + "epoch": 2.3463626492942455, + "grad_norm": 7.121924947407594, + "learning_rate": 1.1476925035953217e-05, + "loss": 0.4922, + "step": 4322 + }, + { + "epoch": 2.3469055374592833, + "grad_norm": 10.58793763221932, + "learning_rate": 1.1473446673294552e-05, + "loss": 0.9691, + "step": 4323 + }, + { + "epoch": 2.3474484256243215, + "grad_norm": 11.51770882137354, + "learning_rate": 1.1469968128397656e-05, + "loss": 0.6713, + "step": 4324 + }, + { + "epoch": 2.3479913137893593, + "grad_norm": 9.960125063324698, + "learning_rate": 1.146648940169276e-05, + "loss": 0.8266, + "step": 4325 + }, + { + "epoch": 2.3485342019543975, + "grad_norm": 14.574982915069748, + "learning_rate": 1.1463010493610114e-05, + "loss": 1.4129, + "step": 4326 + }, + { + "epoch": 2.3490770901194353, + "grad_norm": 7.9866843222052, + "learning_rate": 1.145953140458e-05, + "loss": 0.6368, + "step": 4327 + }, + { + "epoch": 2.3496199782844736, + "grad_norm": 14.879386143441858, + "learning_rate": 1.1456052135032717e-05, + "loss": 0.579, + "step": 4328 + }, + { + "epoch": 2.3501628664495113, + "grad_norm": 9.951154243493171, + "learning_rate": 1.1452572685398586e-05, + "loss": 0.6989, + "step": 4329 + }, + { + "epoch": 2.3507057546145496, + "grad_norm": 8.242511521892395, + "learning_rate": 1.144909305610795e-05, + "loss": 0.5213, + "step": 4330 + }, + { + "epoch": 2.3512486427795873, + "grad_norm": 10.259159837692009, + "learning_rate": 1.1445613247591174e-05, + "loss": 0.9849, + "step": 4331 + }, + { + "epoch": 2.3517915309446256, + "grad_norm": 7.97479592417325, + "learning_rate": 1.1442133260278648e-05, + "loss": 0.5867, + "step": 4332 + }, + { + "epoch": 2.3523344191096633, + "grad_norm": 8.493125656661944, + "learning_rate": 1.143865309460078e-05, + "loss": 0.9095, + "step": 4333 + }, + { + "epoch": 2.3528773072747016, + "grad_norm": 12.49067508129187, + "learning_rate": 1.1435172750988009e-05, + "loss": 0.8484, + "step": 4334 + }, + { + "epoch": 2.3534201954397393, + "grad_norm": 11.134649233778442, + "learning_rate": 1.1431692229870782e-05, + "loss": 0.8422, + "step": 4335 + }, + { + "epoch": 2.3539630836047776, + "grad_norm": 6.99064506687852, + "learning_rate": 1.1428211531679576e-05, + "loss": 0.5388, + "step": 4336 + }, + { + "epoch": 2.3545059717698154, + "grad_norm": 7.798255243454909, + "learning_rate": 1.1424730656844895e-05, + "loss": 0.593, + "step": 4337 + }, + { + "epoch": 2.3550488599348536, + "grad_norm": 8.178641043550261, + "learning_rate": 1.1421249605797255e-05, + "loss": 0.4778, + "step": 4338 + }, + { + "epoch": 2.3555917480998914, + "grad_norm": 8.917024765042555, + "learning_rate": 1.1417768378967196e-05, + "loss": 0.5782, + "step": 4339 + }, + { + "epoch": 2.3561346362649296, + "grad_norm": 9.742472427719518, + "learning_rate": 1.1414286976785286e-05, + "loss": 0.8709, + "step": 4340 + }, + { + "epoch": 2.3566775244299674, + "grad_norm": 10.08457903035175, + "learning_rate": 1.141080539968211e-05, + "loss": 0.5582, + "step": 4341 + }, + { + "epoch": 2.3572204125950056, + "grad_norm": 10.74958408516018, + "learning_rate": 1.1407323648088275e-05, + "loss": 0.7498, + "step": 4342 + }, + { + "epoch": 2.3577633007600434, + "grad_norm": 12.356830179446728, + "learning_rate": 1.1403841722434405e-05, + "loss": 0.9357, + "step": 4343 + }, + { + "epoch": 2.3583061889250816, + "grad_norm": 8.220220873503408, + "learning_rate": 1.1400359623151153e-05, + "loss": 0.5554, + "step": 4344 + }, + { + "epoch": 2.3588490770901194, + "grad_norm": 10.916598567149965, + "learning_rate": 1.1396877350669194e-05, + "loss": 0.8239, + "step": 4345 + }, + { + "epoch": 2.3593919652551576, + "grad_norm": 10.51236366107464, + "learning_rate": 1.1393394905419215e-05, + "loss": 0.5945, + "step": 4346 + }, + { + "epoch": 2.3599348534201954, + "grad_norm": 10.534153676139686, + "learning_rate": 1.1389912287831935e-05, + "loss": 0.575, + "step": 4347 + }, + { + "epoch": 2.3604777415852336, + "grad_norm": 10.06758846687261, + "learning_rate": 1.1386429498338088e-05, + "loss": 0.702, + "step": 4348 + }, + { + "epoch": 2.3610206297502714, + "grad_norm": 9.282335539362101, + "learning_rate": 1.1382946537368432e-05, + "loss": 0.5645, + "step": 4349 + }, + { + "epoch": 2.3615635179153096, + "grad_norm": 8.137108078430895, + "learning_rate": 1.1379463405353741e-05, + "loss": 0.5057, + "step": 4350 + }, + { + "epoch": 2.3621064060803474, + "grad_norm": 10.651520296934317, + "learning_rate": 1.137598010272482e-05, + "loss": 0.762, + "step": 4351 + }, + { + "epoch": 2.3626492942453856, + "grad_norm": 10.001083727438939, + "learning_rate": 1.1372496629912486e-05, + "loss": 0.7255, + "step": 4352 + }, + { + "epoch": 2.3631921824104234, + "grad_norm": 12.180927670825834, + "learning_rate": 1.1369012987347582e-05, + "loss": 0.7295, + "step": 4353 + }, + { + "epoch": 2.3637350705754616, + "grad_norm": 10.697210588401258, + "learning_rate": 1.136552917546097e-05, + "loss": 0.5159, + "step": 4354 + }, + { + "epoch": 2.3642779587404994, + "grad_norm": 13.551389197850602, + "learning_rate": 1.1362045194683534e-05, + "loss": 0.9265, + "step": 4355 + }, + { + "epoch": 2.3648208469055376, + "grad_norm": 12.38857946888456, + "learning_rate": 1.1358561045446175e-05, + "loss": 0.7577, + "step": 4356 + }, + { + "epoch": 2.3653637350705754, + "grad_norm": 10.138108604403497, + "learning_rate": 1.1355076728179824e-05, + "loss": 0.5328, + "step": 4357 + }, + { + "epoch": 2.3659066232356136, + "grad_norm": 9.823785539850283, + "learning_rate": 1.1351592243315416e-05, + "loss": 0.6116, + "step": 4358 + }, + { + "epoch": 2.3664495114006514, + "grad_norm": 8.238880239030014, + "learning_rate": 1.1348107591283932e-05, + "loss": 0.3863, + "step": 4359 + }, + { + "epoch": 2.3669923995656896, + "grad_norm": 11.34031175123055, + "learning_rate": 1.1344622772516352e-05, + "loss": 0.9601, + "step": 4360 + }, + { + "epoch": 2.3675352877307274, + "grad_norm": 11.270983899241404, + "learning_rate": 1.1341137787443683e-05, + "loss": 0.7598, + "step": 4361 + }, + { + "epoch": 2.3680781758957656, + "grad_norm": 10.649130568340542, + "learning_rate": 1.1337652636496959e-05, + "loss": 0.6547, + "step": 4362 + }, + { + "epoch": 2.3686210640608034, + "grad_norm": 9.462979596192575, + "learning_rate": 1.1334167320107224e-05, + "loss": 0.5247, + "step": 4363 + }, + { + "epoch": 2.3691639522258416, + "grad_norm": 11.496313370817235, + "learning_rate": 1.1330681838705544e-05, + "loss": 0.5465, + "step": 4364 + }, + { + "epoch": 2.3697068403908794, + "grad_norm": 10.125563730301518, + "learning_rate": 1.1327196192723016e-05, + "loss": 0.5687, + "step": 4365 + }, + { + "epoch": 2.3702497285559176, + "grad_norm": 14.511657264928447, + "learning_rate": 1.1323710382590747e-05, + "loss": 0.7274, + "step": 4366 + }, + { + "epoch": 2.3707926167209554, + "grad_norm": 10.08856219742556, + "learning_rate": 1.1320224408739868e-05, + "loss": 0.4633, + "step": 4367 + }, + { + "epoch": 2.3713355048859937, + "grad_norm": 9.943874647753937, + "learning_rate": 1.131673827160153e-05, + "loss": 0.7042, + "step": 4368 + }, + { + "epoch": 2.3718783930510314, + "grad_norm": 10.550654748070723, + "learning_rate": 1.1313251971606906e-05, + "loss": 0.6336, + "step": 4369 + }, + { + "epoch": 2.3724212812160697, + "grad_norm": 14.84394771322204, + "learning_rate": 1.1309765509187182e-05, + "loss": 0.6591, + "step": 4370 + }, + { + "epoch": 2.3729641693811074, + "grad_norm": 10.962774929698096, + "learning_rate": 1.1306278884773576e-05, + "loss": 0.9648, + "step": 4371 + }, + { + "epoch": 2.3735070575461457, + "grad_norm": 9.063486875880555, + "learning_rate": 1.1302792098797312e-05, + "loss": 0.4618, + "step": 4372 + }, + { + "epoch": 2.3740499457111834, + "grad_norm": 10.458070080982525, + "learning_rate": 1.1299305151689648e-05, + "loss": 0.6216, + "step": 4373 + }, + { + "epoch": 2.3745928338762217, + "grad_norm": 10.041421519621146, + "learning_rate": 1.1295818043881852e-05, + "loss": 0.6642, + "step": 4374 + }, + { + "epoch": 2.3751357220412594, + "grad_norm": 8.817266669804289, + "learning_rate": 1.1292330775805214e-05, + "loss": 0.4725, + "step": 4375 + }, + { + "epoch": 2.3756786102062977, + "grad_norm": 12.823614653687773, + "learning_rate": 1.1288843347891045e-05, + "loss": 0.553, + "step": 4376 + }, + { + "epoch": 2.3762214983713354, + "grad_norm": 9.281557160359569, + "learning_rate": 1.1285355760570677e-05, + "loss": 0.6436, + "step": 4377 + }, + { + "epoch": 2.3767643865363737, + "grad_norm": 13.234297666234115, + "learning_rate": 1.1281868014275458e-05, + "loss": 0.7665, + "step": 4378 + }, + { + "epoch": 2.3773072747014115, + "grad_norm": 13.013112559574942, + "learning_rate": 1.1278380109436764e-05, + "loss": 0.9087, + "step": 4379 + }, + { + "epoch": 2.3778501628664497, + "grad_norm": 9.413433369666237, + "learning_rate": 1.1274892046485979e-05, + "loss": 0.5248, + "step": 4380 + }, + { + "epoch": 2.3783930510314875, + "grad_norm": 10.865873661981059, + "learning_rate": 1.1271403825854516e-05, + "loss": 0.7398, + "step": 4381 + }, + { + "epoch": 2.3789359391965257, + "grad_norm": 8.516596816721957, + "learning_rate": 1.12679154479738e-05, + "loss": 0.4652, + "step": 4382 + }, + { + "epoch": 2.3794788273615635, + "grad_norm": 7.9464803374429085, + "learning_rate": 1.1264426913275277e-05, + "loss": 0.6952, + "step": 4383 + }, + { + "epoch": 2.3800217155266017, + "grad_norm": 9.171877282099603, + "learning_rate": 1.1260938222190423e-05, + "loss": 0.7534, + "step": 4384 + }, + { + "epoch": 2.3805646036916395, + "grad_norm": 10.833526250938174, + "learning_rate": 1.1257449375150717e-05, + "loss": 0.7966, + "step": 4385 + }, + { + "epoch": 2.3811074918566777, + "grad_norm": 7.087015821603363, + "learning_rate": 1.1253960372587669e-05, + "loss": 0.5145, + "step": 4386 + }, + { + "epoch": 2.3816503800217155, + "grad_norm": 9.355325600419727, + "learning_rate": 1.1250471214932805e-05, + "loss": 0.6084, + "step": 4387 + }, + { + "epoch": 2.3821932681867537, + "grad_norm": 11.679312988029317, + "learning_rate": 1.1246981902617666e-05, + "loss": 0.7734, + "step": 4388 + }, + { + "epoch": 2.3827361563517915, + "grad_norm": 11.696985760230557, + "learning_rate": 1.124349243607382e-05, + "loss": 0.7843, + "step": 4389 + }, + { + "epoch": 2.3832790445168297, + "grad_norm": 11.045261072625825, + "learning_rate": 1.1240002815732847e-05, + "loss": 0.8102, + "step": 4390 + }, + { + "epoch": 2.3838219326818675, + "grad_norm": 13.853716237422205, + "learning_rate": 1.123651304202635e-05, + "loss": 0.6548, + "step": 4391 + }, + { + "epoch": 2.3843648208469057, + "grad_norm": 9.256284103603257, + "learning_rate": 1.1233023115385949e-05, + "loss": 0.5769, + "step": 4392 + }, + { + "epoch": 2.3849077090119435, + "grad_norm": 11.719289709502613, + "learning_rate": 1.1229533036243283e-05, + "loss": 0.9445, + "step": 4393 + }, + { + "epoch": 2.3854505971769817, + "grad_norm": 13.181528127455582, + "learning_rate": 1.1226042805030014e-05, + "loss": 0.78, + "step": 4394 + }, + { + "epoch": 2.3859934853420195, + "grad_norm": 10.610212858388932, + "learning_rate": 1.1222552422177813e-05, + "loss": 0.9325, + "step": 4395 + }, + { + "epoch": 2.3865363735070577, + "grad_norm": 8.750069163521957, + "learning_rate": 1.1219061888118381e-05, + "loss": 0.5576, + "step": 4396 + }, + { + "epoch": 2.3870792616720955, + "grad_norm": 11.697241415039246, + "learning_rate": 1.121557120328343e-05, + "loss": 0.6042, + "step": 4397 + }, + { + "epoch": 2.3876221498371337, + "grad_norm": 14.150909202475516, + "learning_rate": 1.1212080368104699e-05, + "loss": 0.8888, + "step": 4398 + }, + { + "epoch": 2.3881650380021715, + "grad_norm": 8.45338495407854, + "learning_rate": 1.1208589383013934e-05, + "loss": 0.5394, + "step": 4399 + }, + { + "epoch": 2.3887079261672097, + "grad_norm": 8.581855639244504, + "learning_rate": 1.120509824844291e-05, + "loss": 0.4231, + "step": 4400 + }, + { + "epoch": 2.3892508143322475, + "grad_norm": 9.723553293714806, + "learning_rate": 1.1201606964823415e-05, + "loss": 0.5184, + "step": 4401 + }, + { + "epoch": 2.3897937024972857, + "grad_norm": 11.437338247112033, + "learning_rate": 1.1198115532587254e-05, + "loss": 0.9033, + "step": 4402 + }, + { + "epoch": 2.3903365906623235, + "grad_norm": 11.185644591472458, + "learning_rate": 1.1194623952166252e-05, + "loss": 0.7607, + "step": 4403 + }, + { + "epoch": 2.3908794788273617, + "grad_norm": 11.125764861054968, + "learning_rate": 1.1191132223992258e-05, + "loss": 0.7844, + "step": 4404 + }, + { + "epoch": 2.3914223669923995, + "grad_norm": 10.915426846928037, + "learning_rate": 1.1187640348497131e-05, + "loss": 0.5862, + "step": 4405 + }, + { + "epoch": 2.3919652551574377, + "grad_norm": 8.032417531111763, + "learning_rate": 1.1184148326112753e-05, + "loss": 0.5992, + "step": 4406 + }, + { + "epoch": 2.3925081433224755, + "grad_norm": 10.322050440175012, + "learning_rate": 1.1180656157271026e-05, + "loss": 0.6275, + "step": 4407 + }, + { + "epoch": 2.3930510314875137, + "grad_norm": 5.756218508975568, + "learning_rate": 1.117716384240386e-05, + "loss": 0.3605, + "step": 4408 + }, + { + "epoch": 2.3935939196525515, + "grad_norm": 13.072320066260373, + "learning_rate": 1.1173671381943197e-05, + "loss": 1.1771, + "step": 4409 + }, + { + "epoch": 2.3941368078175898, + "grad_norm": 8.390483377503248, + "learning_rate": 1.1170178776320982e-05, + "loss": 0.484, + "step": 4410 + }, + { + "epoch": 2.3946796959826275, + "grad_norm": 9.168734569918955, + "learning_rate": 1.1166686025969196e-05, + "loss": 0.4637, + "step": 4411 + }, + { + "epoch": 2.3952225841476658, + "grad_norm": 11.945726197992448, + "learning_rate": 1.116319313131982e-05, + "loss": 0.6501, + "step": 4412 + }, + { + "epoch": 2.3957654723127035, + "grad_norm": 9.925140956274975, + "learning_rate": 1.1159700092804866e-05, + "loss": 0.5583, + "step": 4413 + }, + { + "epoch": 2.3963083604777418, + "grad_norm": 10.0377125619045, + "learning_rate": 1.1156206910856353e-05, + "loss": 0.5783, + "step": 4414 + }, + { + "epoch": 2.3968512486427795, + "grad_norm": 13.878494510161136, + "learning_rate": 1.115271358590633e-05, + "loss": 0.7029, + "step": 4415 + }, + { + "epoch": 2.3973941368078178, + "grad_norm": 8.414419089776398, + "learning_rate": 1.114922011838685e-05, + "loss": 0.4398, + "step": 4416 + }, + { + "epoch": 2.3979370249728555, + "grad_norm": 10.455474671185716, + "learning_rate": 1.1145726508729994e-05, + "loss": 0.5462, + "step": 4417 + }, + { + "epoch": 2.3984799131378938, + "grad_norm": 8.80305307358664, + "learning_rate": 1.1142232757367861e-05, + "loss": 0.3885, + "step": 4418 + }, + { + "epoch": 2.3990228013029316, + "grad_norm": 9.381551586854757, + "learning_rate": 1.1138738864732559e-05, + "loss": 0.6646, + "step": 4419 + }, + { + "epoch": 2.3995656894679698, + "grad_norm": 9.409317053539013, + "learning_rate": 1.113524483125622e-05, + "loss": 0.7772, + "step": 4420 + }, + { + "epoch": 2.4001085776330076, + "grad_norm": 7.414798924114551, + "learning_rate": 1.113175065737099e-05, + "loss": 0.4508, + "step": 4421 + }, + { + "epoch": 2.400651465798046, + "grad_norm": 10.931216387643186, + "learning_rate": 1.1128256343509032e-05, + "loss": 0.6434, + "step": 4422 + }, + { + "epoch": 2.4011943539630836, + "grad_norm": 9.811349462515302, + "learning_rate": 1.1124761890102535e-05, + "loss": 0.5104, + "step": 4423 + }, + { + "epoch": 2.401737242128122, + "grad_norm": 10.856233458965448, + "learning_rate": 1.1121267297583694e-05, + "loss": 0.6535, + "step": 4424 + }, + { + "epoch": 2.4022801302931596, + "grad_norm": 10.2119482085421, + "learning_rate": 1.1117772566384725e-05, + "loss": 0.593, + "step": 4425 + }, + { + "epoch": 2.402823018458198, + "grad_norm": 11.327455944037217, + "learning_rate": 1.1114277696937865e-05, + "loss": 0.9199, + "step": 4426 + }, + { + "epoch": 2.4033659066232356, + "grad_norm": 10.64004734338448, + "learning_rate": 1.1110782689675365e-05, + "loss": 0.793, + "step": 4427 + }, + { + "epoch": 2.403908794788274, + "grad_norm": 5.71867532259649, + "learning_rate": 1.110728754502949e-05, + "loss": 0.2726, + "step": 4428 + }, + { + "epoch": 2.4044516829533116, + "grad_norm": 10.33820400695821, + "learning_rate": 1.1103792263432525e-05, + "loss": 0.787, + "step": 4429 + }, + { + "epoch": 2.40499457111835, + "grad_norm": 10.02381386485252, + "learning_rate": 1.1100296845316777e-05, + "loss": 0.7701, + "step": 4430 + }, + { + "epoch": 2.4055374592833876, + "grad_norm": 12.111852512562214, + "learning_rate": 1.1096801291114561e-05, + "loss": 0.6123, + "step": 4431 + }, + { + "epoch": 2.406080347448426, + "grad_norm": 8.432777542688768, + "learning_rate": 1.1093305601258213e-05, + "loss": 0.4817, + "step": 4432 + }, + { + "epoch": 2.4066232356134636, + "grad_norm": 12.648769211741241, + "learning_rate": 1.1089809776180084e-05, + "loss": 0.7328, + "step": 4433 + }, + { + "epoch": 2.407166123778502, + "grad_norm": 9.260727895387426, + "learning_rate": 1.1086313816312546e-05, + "loss": 0.2951, + "step": 4434 + }, + { + "epoch": 2.4077090119435396, + "grad_norm": 7.801968230318713, + "learning_rate": 1.1082817722087981e-05, + "loss": 0.6204, + "step": 4435 + }, + { + "epoch": 2.408251900108578, + "grad_norm": 18.28104514211005, + "learning_rate": 1.10793214939388e-05, + "loss": 1.2717, + "step": 4436 + }, + { + "epoch": 2.4087947882736156, + "grad_norm": 8.074911353186037, + "learning_rate": 1.1075825132297414e-05, + "loss": 0.391, + "step": 4437 + }, + { + "epoch": 2.409337676438654, + "grad_norm": 10.223302145592744, + "learning_rate": 1.1072328637596259e-05, + "loss": 0.8764, + "step": 4438 + }, + { + "epoch": 2.4098805646036916, + "grad_norm": 8.909676311441572, + "learning_rate": 1.1068832010267792e-05, + "loss": 0.5866, + "step": 4439 + }, + { + "epoch": 2.41042345276873, + "grad_norm": 8.270875830864977, + "learning_rate": 1.1065335250744478e-05, + "loss": 0.4935, + "step": 4440 + }, + { + "epoch": 2.4109663409337676, + "grad_norm": 8.125694150069458, + "learning_rate": 1.1061838359458801e-05, + "loss": 0.5855, + "step": 4441 + }, + { + "epoch": 2.411509229098806, + "grad_norm": 9.00158598137614, + "learning_rate": 1.1058341336843264e-05, + "loss": 0.4464, + "step": 4442 + }, + { + "epoch": 2.4120521172638436, + "grad_norm": 10.568261973334824, + "learning_rate": 1.1054844183330385e-05, + "loss": 0.844, + "step": 4443 + }, + { + "epoch": 2.412595005428882, + "grad_norm": 10.29648068473729, + "learning_rate": 1.1051346899352693e-05, + "loss": 0.7083, + "step": 4444 + }, + { + "epoch": 2.4131378935939196, + "grad_norm": 10.658053637486233, + "learning_rate": 1.1047849485342744e-05, + "loss": 0.917, + "step": 4445 + }, + { + "epoch": 2.413680781758958, + "grad_norm": 10.85927844636501, + "learning_rate": 1.1044351941733097e-05, + "loss": 0.7737, + "step": 4446 + }, + { + "epoch": 2.4142236699239956, + "grad_norm": 13.105614857394716, + "learning_rate": 1.1040854268956343e-05, + "loss": 0.7626, + "step": 4447 + }, + { + "epoch": 2.414766558089034, + "grad_norm": 12.088913830608213, + "learning_rate": 1.1037356467445071e-05, + "loss": 0.7439, + "step": 4448 + }, + { + "epoch": 2.4153094462540716, + "grad_norm": 13.524031933483466, + "learning_rate": 1.1033858537631894e-05, + "loss": 0.6172, + "step": 4449 + }, + { + "epoch": 2.41585233441911, + "grad_norm": 12.25037908572634, + "learning_rate": 1.1030360479949452e-05, + "loss": 0.7959, + "step": 4450 + }, + { + "epoch": 2.4163952225841476, + "grad_norm": 10.26766943217781, + "learning_rate": 1.102686229483038e-05, + "loss": 0.8928, + "step": 4451 + }, + { + "epoch": 2.416938110749186, + "grad_norm": 10.66247714207056, + "learning_rate": 1.1023363982707345e-05, + "loss": 0.8498, + "step": 4452 + }, + { + "epoch": 2.4174809989142236, + "grad_norm": 9.37397925289073, + "learning_rate": 1.101986554401302e-05, + "loss": 0.5633, + "step": 4453 + }, + { + "epoch": 2.418023887079262, + "grad_norm": 8.656197915403546, + "learning_rate": 1.10163669791801e-05, + "loss": 0.6403, + "step": 4454 + }, + { + "epoch": 2.4185667752442996, + "grad_norm": 10.199878069425685, + "learning_rate": 1.101286828864129e-05, + "loss": 0.7139, + "step": 4455 + }, + { + "epoch": 2.419109663409338, + "grad_norm": 8.176125775602177, + "learning_rate": 1.1009369472829316e-05, + "loss": 0.5322, + "step": 4456 + }, + { + "epoch": 2.4196525515743756, + "grad_norm": 9.441655350085792, + "learning_rate": 1.1005870532176918e-05, + "loss": 0.8574, + "step": 4457 + }, + { + "epoch": 2.420195439739414, + "grad_norm": 10.487844747641686, + "learning_rate": 1.100237146711685e-05, + "loss": 0.7399, + "step": 4458 + }, + { + "epoch": 2.4207383279044516, + "grad_norm": 8.456961156540515, + "learning_rate": 1.0998872278081883e-05, + "loss": 0.5795, + "step": 4459 + }, + { + "epoch": 2.42128121606949, + "grad_norm": 10.247575624097415, + "learning_rate": 1.0995372965504798e-05, + "loss": 0.7048, + "step": 4460 + }, + { + "epoch": 2.4218241042345277, + "grad_norm": 8.127194271974078, + "learning_rate": 1.09918735298184e-05, + "loss": 0.6339, + "step": 4461 + }, + { + "epoch": 2.422366992399566, + "grad_norm": 11.080151666019892, + "learning_rate": 1.0988373971455502e-05, + "loss": 0.6, + "step": 4462 + }, + { + "epoch": 2.4229098805646037, + "grad_norm": 9.249791609753641, + "learning_rate": 1.0984874290848935e-05, + "loss": 0.5823, + "step": 4463 + }, + { + "epoch": 2.423452768729642, + "grad_norm": 8.673439204574109, + "learning_rate": 1.0981374488431546e-05, + "loss": 0.4486, + "step": 4464 + }, + { + "epoch": 2.4239956568946797, + "grad_norm": 8.530441262511525, + "learning_rate": 1.0977874564636199e-05, + "loss": 0.4685, + "step": 4465 + }, + { + "epoch": 2.424538545059718, + "grad_norm": 11.975099667402777, + "learning_rate": 1.0974374519895765e-05, + "loss": 0.5608, + "step": 4466 + }, + { + "epoch": 2.4250814332247557, + "grad_norm": 12.28650683165383, + "learning_rate": 1.097087435464314e-05, + "loss": 0.6428, + "step": 4467 + }, + { + "epoch": 2.425624321389794, + "grad_norm": 11.443904338187624, + "learning_rate": 1.0967374069311224e-05, + "loss": 0.5093, + "step": 4468 + }, + { + "epoch": 2.4261672095548317, + "grad_norm": 12.11642038793139, + "learning_rate": 1.0963873664332944e-05, + "loss": 0.7573, + "step": 4469 + }, + { + "epoch": 2.42671009771987, + "grad_norm": 12.716855768902285, + "learning_rate": 1.0960373140141232e-05, + "loss": 1.0742, + "step": 4470 + }, + { + "epoch": 2.4272529858849077, + "grad_norm": 8.060615842022353, + "learning_rate": 1.095687249716904e-05, + "loss": 0.5299, + "step": 4471 + }, + { + "epoch": 2.427795874049946, + "grad_norm": 10.663178608260651, + "learning_rate": 1.0953371735849329e-05, + "loss": 0.6575, + "step": 4472 + }, + { + "epoch": 2.4283387622149837, + "grad_norm": 11.883215571850693, + "learning_rate": 1.0949870856615088e-05, + "loss": 0.7516, + "step": 4473 + }, + { + "epoch": 2.428881650380022, + "grad_norm": 8.725361323128087, + "learning_rate": 1.0946369859899299e-05, + "loss": 0.5543, + "step": 4474 + }, + { + "epoch": 2.4294245385450597, + "grad_norm": 7.964297132373328, + "learning_rate": 1.0942868746134983e-05, + "loss": 0.3799, + "step": 4475 + }, + { + "epoch": 2.429967426710098, + "grad_norm": 11.129935043446988, + "learning_rate": 1.0939367515755155e-05, + "loss": 0.5274, + "step": 4476 + }, + { + "epoch": 2.4305103148751357, + "grad_norm": 8.84506268400313, + "learning_rate": 1.0935866169192858e-05, + "loss": 0.6314, + "step": 4477 + }, + { + "epoch": 2.431053203040174, + "grad_norm": 9.898346545614544, + "learning_rate": 1.0932364706881138e-05, + "loss": 0.4733, + "step": 4478 + }, + { + "epoch": 2.4315960912052117, + "grad_norm": 12.708528067939865, + "learning_rate": 1.092886312925307e-05, + "loss": 1.1591, + "step": 4479 + }, + { + "epoch": 2.43213897937025, + "grad_norm": 8.852913949596426, + "learning_rate": 1.0925361436741726e-05, + "loss": 0.4934, + "step": 4480 + }, + { + "epoch": 2.4326818675352877, + "grad_norm": 9.976618553528468, + "learning_rate": 1.0921859629780205e-05, + "loss": 0.5726, + "step": 4481 + }, + { + "epoch": 2.433224755700326, + "grad_norm": 10.591185923462703, + "learning_rate": 1.0918357708801617e-05, + "loss": 0.9099, + "step": 4482 + }, + { + "epoch": 2.4337676438653637, + "grad_norm": 9.782319605062739, + "learning_rate": 1.0914855674239083e-05, + "loss": 0.7016, + "step": 4483 + }, + { + "epoch": 2.434310532030402, + "grad_norm": 7.962337485760134, + "learning_rate": 1.0911353526525741e-05, + "loss": 0.3899, + "step": 4484 + }, + { + "epoch": 2.4348534201954397, + "grad_norm": 10.926665567713723, + "learning_rate": 1.0907851266094742e-05, + "loss": 1.0416, + "step": 4485 + }, + { + "epoch": 2.435396308360478, + "grad_norm": 10.774706653267238, + "learning_rate": 1.0904348893379252e-05, + "loss": 0.8539, + "step": 4486 + }, + { + "epoch": 2.4359391965255157, + "grad_norm": 9.962550119554402, + "learning_rate": 1.0900846408812449e-05, + "loss": 0.6202, + "step": 4487 + }, + { + "epoch": 2.436482084690554, + "grad_norm": 7.425101710467455, + "learning_rate": 1.0897343812827528e-05, + "loss": 0.4138, + "step": 4488 + }, + { + "epoch": 2.4370249728555917, + "grad_norm": 11.159990654051901, + "learning_rate": 1.0893841105857695e-05, + "loss": 0.7927, + "step": 4489 + }, + { + "epoch": 2.4375678610206295, + "grad_norm": 8.584412414000917, + "learning_rate": 1.0890338288336168e-05, + "loss": 0.4596, + "step": 4490 + }, + { + "epoch": 2.4381107491856677, + "grad_norm": 13.182969196780135, + "learning_rate": 1.0886835360696182e-05, + "loss": 0.7052, + "step": 4491 + }, + { + "epoch": 2.438653637350706, + "grad_norm": 12.643144928400368, + "learning_rate": 1.0883332323370986e-05, + "loss": 0.7273, + "step": 4492 + }, + { + "epoch": 2.4391965255157437, + "grad_norm": 11.758017246786887, + "learning_rate": 1.087982917679384e-05, + "loss": 0.6389, + "step": 4493 + }, + { + "epoch": 2.4397394136807815, + "grad_norm": 10.332831381965816, + "learning_rate": 1.0876325921398016e-05, + "loss": 0.49, + "step": 4494 + }, + { + "epoch": 2.4402823018458197, + "grad_norm": 12.79909493300538, + "learning_rate": 1.0872822557616811e-05, + "loss": 1.027, + "step": 4495 + }, + { + "epoch": 2.440825190010858, + "grad_norm": 10.225359261113429, + "learning_rate": 1.0869319085883518e-05, + "loss": 0.6319, + "step": 4496 + }, + { + "epoch": 2.4413680781758957, + "grad_norm": 10.600390533226815, + "learning_rate": 1.0865815506631456e-05, + "loss": 0.5513, + "step": 4497 + }, + { + "epoch": 2.4419109663409335, + "grad_norm": 9.069709981087057, + "learning_rate": 1.0862311820293951e-05, + "loss": 0.4358, + "step": 4498 + }, + { + "epoch": 2.4424538545059717, + "grad_norm": 10.228930326876846, + "learning_rate": 1.0858808027304346e-05, + "loss": 1.0149, + "step": 4499 + }, + { + "epoch": 2.44299674267101, + "grad_norm": 6.981580424693794, + "learning_rate": 1.0855304128095997e-05, + "loss": 0.3823, + "step": 4500 + }, + { + "epoch": 2.4435396308360477, + "grad_norm": 12.347328486710586, + "learning_rate": 1.085180012310227e-05, + "loss": 0.8202, + "step": 4501 + }, + { + "epoch": 2.4440825190010855, + "grad_norm": 7.794475194787575, + "learning_rate": 1.0848296012756545e-05, + "loss": 0.4972, + "step": 4502 + }, + { + "epoch": 2.4446254071661238, + "grad_norm": 8.640639888495457, + "learning_rate": 1.0844791797492218e-05, + "loss": 0.4174, + "step": 4503 + }, + { + "epoch": 2.445168295331162, + "grad_norm": 9.363787480425108, + "learning_rate": 1.0841287477742695e-05, + "loss": 0.5783, + "step": 4504 + }, + { + "epoch": 2.4457111834961998, + "grad_norm": 9.455569437628846, + "learning_rate": 1.0837783053941396e-05, + "loss": 0.5611, + "step": 4505 + }, + { + "epoch": 2.4462540716612375, + "grad_norm": 7.587727857256139, + "learning_rate": 1.083427852652175e-05, + "loss": 0.4129, + "step": 4506 + }, + { + "epoch": 2.4467969598262758, + "grad_norm": 11.579317957056839, + "learning_rate": 1.083077389591721e-05, + "loss": 0.6628, + "step": 4507 + }, + { + "epoch": 2.447339847991314, + "grad_norm": 9.130000079525795, + "learning_rate": 1.0827269162561229e-05, + "loss": 0.4626, + "step": 4508 + }, + { + "epoch": 2.4478827361563518, + "grad_norm": 8.791608822057356, + "learning_rate": 1.0823764326887283e-05, + "loss": 0.7441, + "step": 4509 + }, + { + "epoch": 2.4484256243213895, + "grad_norm": 8.698379661142646, + "learning_rate": 1.0820259389328852e-05, + "loss": 0.4924, + "step": 4510 + }, + { + "epoch": 2.4489685124864278, + "grad_norm": 12.811407852860862, + "learning_rate": 1.0816754350319427e-05, + "loss": 0.9829, + "step": 4511 + }, + { + "epoch": 2.449511400651466, + "grad_norm": 11.588653893201888, + "learning_rate": 1.0813249210292525e-05, + "loss": 0.7218, + "step": 4512 + }, + { + "epoch": 2.450054288816504, + "grad_norm": 12.334703169006005, + "learning_rate": 1.0809743969681665e-05, + "loss": 0.8737, + "step": 4513 + }, + { + "epoch": 2.4505971769815416, + "grad_norm": 12.583313163331177, + "learning_rate": 1.080623862892038e-05, + "loss": 0.7226, + "step": 4514 + }, + { + "epoch": 2.45114006514658, + "grad_norm": 12.11103490873811, + "learning_rate": 1.0802733188442216e-05, + "loss": 0.7867, + "step": 4515 + }, + { + "epoch": 2.451682953311618, + "grad_norm": 11.222788743446143, + "learning_rate": 1.0799227648680732e-05, + "loss": 0.5457, + "step": 4516 + }, + { + "epoch": 2.452225841476656, + "grad_norm": 7.685503695913846, + "learning_rate": 1.07957220100695e-05, + "loss": 0.6483, + "step": 4517 + }, + { + "epoch": 2.4527687296416936, + "grad_norm": 8.407197016108878, + "learning_rate": 1.0792216273042099e-05, + "loss": 0.4895, + "step": 4518 + }, + { + "epoch": 2.453311617806732, + "grad_norm": 9.426708870733849, + "learning_rate": 1.0788710438032131e-05, + "loss": 1.0719, + "step": 4519 + }, + { + "epoch": 2.45385450597177, + "grad_norm": 13.638918388289357, + "learning_rate": 1.0785204505473197e-05, + "loss": 0.8828, + "step": 4520 + }, + { + "epoch": 2.454397394136808, + "grad_norm": 12.478693804182608, + "learning_rate": 1.0781698475798917e-05, + "loss": 0.6527, + "step": 4521 + }, + { + "epoch": 2.4549402823018456, + "grad_norm": 10.20424250017949, + "learning_rate": 1.0778192349442923e-05, + "loss": 0.9303, + "step": 4522 + }, + { + "epoch": 2.455483170466884, + "grad_norm": 12.175885097025716, + "learning_rate": 1.0774686126838863e-05, + "loss": 0.6391, + "step": 4523 + }, + { + "epoch": 2.456026058631922, + "grad_norm": 10.568732307533892, + "learning_rate": 1.0771179808420385e-05, + "loss": 0.5524, + "step": 4524 + }, + { + "epoch": 2.45656894679696, + "grad_norm": 7.314447097275954, + "learning_rate": 1.0767673394621159e-05, + "loss": 0.4976, + "step": 4525 + }, + { + "epoch": 2.4571118349619976, + "grad_norm": 11.397811881243696, + "learning_rate": 1.0764166885874862e-05, + "loss": 0.4918, + "step": 4526 + }, + { + "epoch": 2.457654723127036, + "grad_norm": 9.691957338181856, + "learning_rate": 1.0760660282615191e-05, + "loss": 0.6725, + "step": 4527 + }, + { + "epoch": 2.458197611292074, + "grad_norm": 7.510379855172465, + "learning_rate": 1.0757153585275844e-05, + "loss": 0.4684, + "step": 4528 + }, + { + "epoch": 2.458740499457112, + "grad_norm": 7.751570484202345, + "learning_rate": 1.0753646794290534e-05, + "loss": 0.5495, + "step": 4529 + }, + { + "epoch": 2.4592833876221496, + "grad_norm": 7.390182612219503, + "learning_rate": 1.0750139910092987e-05, + "loss": 0.4538, + "step": 4530 + }, + { + "epoch": 2.459826275787188, + "grad_norm": 10.180132351193016, + "learning_rate": 1.074663293311694e-05, + "loss": 0.5579, + "step": 4531 + }, + { + "epoch": 2.460369163952226, + "grad_norm": 8.909847677295557, + "learning_rate": 1.074312586379614e-05, + "loss": 0.5023, + "step": 4532 + }, + { + "epoch": 2.460912052117264, + "grad_norm": 10.091297535262147, + "learning_rate": 1.073961870256435e-05, + "loss": 0.7068, + "step": 4533 + }, + { + "epoch": 2.4614549402823016, + "grad_norm": 11.778234555569338, + "learning_rate": 1.0736111449855341e-05, + "loss": 0.7687, + "step": 4534 + }, + { + "epoch": 2.46199782844734, + "grad_norm": 11.871444812620142, + "learning_rate": 1.0732604106102895e-05, + "loss": 0.8333, + "step": 4535 + }, + { + "epoch": 2.462540716612378, + "grad_norm": 6.164790402663993, + "learning_rate": 1.0729096671740806e-05, + "loss": 0.3601, + "step": 4536 + }, + { + "epoch": 2.463083604777416, + "grad_norm": 9.524060445013603, + "learning_rate": 1.0725589147202877e-05, + "loss": 0.6859, + "step": 4537 + }, + { + "epoch": 2.4636264929424536, + "grad_norm": 11.266171306725775, + "learning_rate": 1.0722081532922925e-05, + "loss": 0.7459, + "step": 4538 + }, + { + "epoch": 2.464169381107492, + "grad_norm": 8.945835373858307, + "learning_rate": 1.0718573829334782e-05, + "loss": 0.3787, + "step": 4539 + }, + { + "epoch": 2.46471226927253, + "grad_norm": 6.838806611638338, + "learning_rate": 1.071506603687228e-05, + "loss": 0.4995, + "step": 4540 + }, + { + "epoch": 2.465255157437568, + "grad_norm": 11.246624755560871, + "learning_rate": 1.071155815596927e-05, + "loss": 0.8699, + "step": 4541 + }, + { + "epoch": 2.4657980456026056, + "grad_norm": 7.324342918236285, + "learning_rate": 1.0708050187059615e-05, + "loss": 0.4778, + "step": 4542 + }, + { + "epoch": 2.466340933767644, + "grad_norm": 10.063475303955434, + "learning_rate": 1.0704542130577185e-05, + "loss": 0.5774, + "step": 4543 + }, + { + "epoch": 2.466883821932682, + "grad_norm": 12.644825932835714, + "learning_rate": 1.070103398695586e-05, + "loss": 0.8158, + "step": 4544 + }, + { + "epoch": 2.46742671009772, + "grad_norm": 10.332286780133, + "learning_rate": 1.0697525756629531e-05, + "loss": 0.6518, + "step": 4545 + }, + { + "epoch": 2.4679695982627576, + "grad_norm": 9.98899797230979, + "learning_rate": 1.069401744003211e-05, + "loss": 0.6167, + "step": 4546 + }, + { + "epoch": 2.468512486427796, + "grad_norm": 13.445015438013797, + "learning_rate": 1.0690509037597502e-05, + "loss": 0.8229, + "step": 4547 + }, + { + "epoch": 2.469055374592834, + "grad_norm": 13.637667335089724, + "learning_rate": 1.0687000549759641e-05, + "loss": 0.7186, + "step": 4548 + }, + { + "epoch": 2.469598262757872, + "grad_norm": 10.021629994296763, + "learning_rate": 1.0683491976952454e-05, + "loss": 0.6209, + "step": 4549 + }, + { + "epoch": 2.4701411509229096, + "grad_norm": 10.510874896443969, + "learning_rate": 1.067998331960989e-05, + "loss": 0.9009, + "step": 4550 + }, + { + "epoch": 2.470684039087948, + "grad_norm": 9.53841539843291, + "learning_rate": 1.0676474578165903e-05, + "loss": 0.7187, + "step": 4551 + }, + { + "epoch": 2.471226927252986, + "grad_norm": 11.693787891306421, + "learning_rate": 1.0672965753054466e-05, + "loss": 0.5568, + "step": 4552 + }, + { + "epoch": 2.471769815418024, + "grad_norm": 11.523811853237843, + "learning_rate": 1.0669456844709549e-05, + "loss": 0.8465, + "step": 4553 + }, + { + "epoch": 2.4723127035830617, + "grad_norm": 9.165552336797202, + "learning_rate": 1.0665947853565144e-05, + "loss": 0.7148, + "step": 4554 + }, + { + "epoch": 2.4728555917481, + "grad_norm": 10.390907422337587, + "learning_rate": 1.0662438780055247e-05, + "loss": 0.5271, + "step": 4555 + }, + { + "epoch": 2.473398479913138, + "grad_norm": 10.514660404488053, + "learning_rate": 1.0658929624613866e-05, + "loss": 0.7972, + "step": 4556 + }, + { + "epoch": 2.473941368078176, + "grad_norm": 8.804604897853023, + "learning_rate": 1.065542038767502e-05, + "loss": 0.4391, + "step": 4557 + }, + { + "epoch": 2.4744842562432137, + "grad_norm": 12.58202482779834, + "learning_rate": 1.0651911069672733e-05, + "loss": 0.601, + "step": 4558 + }, + { + "epoch": 2.475027144408252, + "grad_norm": 8.510520469826693, + "learning_rate": 1.0648401671041048e-05, + "loss": 0.6638, + "step": 4559 + }, + { + "epoch": 2.47557003257329, + "grad_norm": 10.132731434552191, + "learning_rate": 1.0644892192214012e-05, + "loss": 0.547, + "step": 4560 + }, + { + "epoch": 2.476112920738328, + "grad_norm": 9.486912806484277, + "learning_rate": 1.0641382633625678e-05, + "loss": 0.6903, + "step": 4561 + }, + { + "epoch": 2.4766558089033657, + "grad_norm": 8.656529900389968, + "learning_rate": 1.0637872995710122e-05, + "loss": 0.6445, + "step": 4562 + }, + { + "epoch": 2.477198697068404, + "grad_norm": 8.377471382601675, + "learning_rate": 1.0634363278901413e-05, + "loss": 0.475, + "step": 4563 + }, + { + "epoch": 2.477741585233442, + "grad_norm": 11.941789514685023, + "learning_rate": 1.0630853483633643e-05, + "loss": 0.8705, + "step": 4564 + }, + { + "epoch": 2.47828447339848, + "grad_norm": 8.551632836218364, + "learning_rate": 1.0627343610340912e-05, + "loss": 0.3577, + "step": 4565 + }, + { + "epoch": 2.4788273615635177, + "grad_norm": 10.220220103336683, + "learning_rate": 1.0623833659457322e-05, + "loss": 0.7002, + "step": 4566 + }, + { + "epoch": 2.479370249728556, + "grad_norm": 10.783849224421678, + "learning_rate": 1.0620323631416994e-05, + "loss": 0.6969, + "step": 4567 + }, + { + "epoch": 2.479913137893594, + "grad_norm": 9.018283325771, + "learning_rate": 1.0616813526654047e-05, + "loss": 0.6844, + "step": 4568 + }, + { + "epoch": 2.480456026058632, + "grad_norm": 9.570551013210757, + "learning_rate": 1.061330334560262e-05, + "loss": 0.5063, + "step": 4569 + }, + { + "epoch": 2.4809989142236697, + "grad_norm": 10.106289919073877, + "learning_rate": 1.0609793088696856e-05, + "loss": 0.6131, + "step": 4570 + }, + { + "epoch": 2.481541802388708, + "grad_norm": 10.842801269455231, + "learning_rate": 1.060628275637091e-05, + "loss": 0.6637, + "step": 4571 + }, + { + "epoch": 2.482084690553746, + "grad_norm": 11.953941591792438, + "learning_rate": 1.0602772349058952e-05, + "loss": 0.7631, + "step": 4572 + }, + { + "epoch": 2.482627578718784, + "grad_norm": 8.194683405468492, + "learning_rate": 1.0599261867195146e-05, + "loss": 0.5936, + "step": 4573 + }, + { + "epoch": 2.4831704668838217, + "grad_norm": 10.216659522745752, + "learning_rate": 1.0595751311213678e-05, + "loss": 0.7017, + "step": 4574 + }, + { + "epoch": 2.48371335504886, + "grad_norm": 11.0782985653065, + "learning_rate": 1.0592240681548738e-05, + "loss": 0.7285, + "step": 4575 + }, + { + "epoch": 2.484256243213898, + "grad_norm": 7.740854250685081, + "learning_rate": 1.0588729978634527e-05, + "loss": 0.4544, + "step": 4576 + }, + { + "epoch": 2.484799131378936, + "grad_norm": 6.9863990778515745, + "learning_rate": 1.0585219202905255e-05, + "loss": 0.2831, + "step": 4577 + }, + { + "epoch": 2.4853420195439737, + "grad_norm": 9.687327462597597, + "learning_rate": 1.0581708354795136e-05, + "loss": 0.5028, + "step": 4578 + }, + { + "epoch": 2.485884907709012, + "grad_norm": 8.009180785791076, + "learning_rate": 1.0578197434738405e-05, + "loss": 0.342, + "step": 4579 + }, + { + "epoch": 2.48642779587405, + "grad_norm": 8.101662777133065, + "learning_rate": 1.057468644316929e-05, + "loss": 0.4146, + "step": 4580 + }, + { + "epoch": 2.486970684039088, + "grad_norm": 9.370936142076715, + "learning_rate": 1.0571175380522041e-05, + "loss": 0.6006, + "step": 4581 + }, + { + "epoch": 2.4875135722041257, + "grad_norm": 11.164770907813724, + "learning_rate": 1.0567664247230911e-05, + "loss": 0.6787, + "step": 4582 + }, + { + "epoch": 2.488056460369164, + "grad_norm": 12.573930877103525, + "learning_rate": 1.056415304373016e-05, + "loss": 0.8051, + "step": 4583 + }, + { + "epoch": 2.488599348534202, + "grad_norm": 8.395060833285095, + "learning_rate": 1.0560641770454063e-05, + "loss": 0.5596, + "step": 4584 + }, + { + "epoch": 2.48914223669924, + "grad_norm": 11.015478890741145, + "learning_rate": 1.0557130427836899e-05, + "loss": 0.5442, + "step": 4585 + }, + { + "epoch": 2.4896851248642777, + "grad_norm": 9.420432906299153, + "learning_rate": 1.0553619016312955e-05, + "loss": 0.4917, + "step": 4586 + }, + { + "epoch": 2.490228013029316, + "grad_norm": 11.026358328990858, + "learning_rate": 1.0550107536316532e-05, + "loss": 0.4291, + "step": 4587 + }, + { + "epoch": 2.490770901194354, + "grad_norm": 16.545112067555284, + "learning_rate": 1.0546595988281928e-05, + "loss": 0.9417, + "step": 4588 + }, + { + "epoch": 2.491313789359392, + "grad_norm": 7.252757680637398, + "learning_rate": 1.0543084372643462e-05, + "loss": 0.7908, + "step": 4589 + }, + { + "epoch": 2.4918566775244297, + "grad_norm": 7.598562688636616, + "learning_rate": 1.0539572689835455e-05, + "loss": 0.3229, + "step": 4590 + }, + { + "epoch": 2.492399565689468, + "grad_norm": 12.449529704914275, + "learning_rate": 1.0536060940292238e-05, + "loss": 0.7373, + "step": 4591 + }, + { + "epoch": 2.492942453854506, + "grad_norm": 11.706159828307792, + "learning_rate": 1.0532549124448151e-05, + "loss": 0.6525, + "step": 4592 + }, + { + "epoch": 2.493485342019544, + "grad_norm": 8.000677691767947, + "learning_rate": 1.0529037242737538e-05, + "loss": 0.3613, + "step": 4593 + }, + { + "epoch": 2.4940282301845818, + "grad_norm": 9.529463393752437, + "learning_rate": 1.0525525295594755e-05, + "loss": 0.5921, + "step": 4594 + }, + { + "epoch": 2.49457111834962, + "grad_norm": 11.71769497271457, + "learning_rate": 1.0522013283454169e-05, + "loss": 0.5507, + "step": 4595 + }, + { + "epoch": 2.495114006514658, + "grad_norm": 8.546908524296436, + "learning_rate": 1.0518501206750145e-05, + "loss": 0.4564, + "step": 4596 + }, + { + "epoch": 2.495656894679696, + "grad_norm": 14.968617219210655, + "learning_rate": 1.0514989065917068e-05, + "loss": 0.5593, + "step": 4597 + }, + { + "epoch": 2.4961997828447338, + "grad_norm": 13.0682164313941, + "learning_rate": 1.051147686138932e-05, + "loss": 0.8812, + "step": 4598 + }, + { + "epoch": 2.496742671009772, + "grad_norm": 15.01014358585534, + "learning_rate": 1.05079645936013e-05, + "loss": 1.0756, + "step": 4599 + }, + { + "epoch": 2.49728555917481, + "grad_norm": 7.598240677419264, + "learning_rate": 1.0504452262987407e-05, + "loss": 0.3936, + "step": 4600 + }, + { + "epoch": 2.497828447339848, + "grad_norm": 10.39263281879832, + "learning_rate": 1.0500939869982058e-05, + "loss": 0.4944, + "step": 4601 + }, + { + "epoch": 2.4983713355048858, + "grad_norm": 12.104880157441473, + "learning_rate": 1.0497427415019665e-05, + "loss": 0.7498, + "step": 4602 + }, + { + "epoch": 2.498914223669924, + "grad_norm": 9.18522035093318, + "learning_rate": 1.0493914898534656e-05, + "loss": 0.6921, + "step": 4603 + }, + { + "epoch": 2.499457111834962, + "grad_norm": 15.812637613525052, + "learning_rate": 1.0490402320961466e-05, + "loss": 0.937, + "step": 4604 + }, + { + "epoch": 2.5, + "grad_norm": 11.400921058050775, + "learning_rate": 1.0486889682734535e-05, + "loss": 0.6453, + "step": 4605 + }, + { + "epoch": 2.500542888165038, + "grad_norm": 8.270129268701748, + "learning_rate": 1.0483376984288314e-05, + "loss": 0.3921, + "step": 4606 + }, + { + "epoch": 2.501085776330076, + "grad_norm": 15.017103371329522, + "learning_rate": 1.0479864226057258e-05, + "loss": 1.0397, + "step": 4607 + }, + { + "epoch": 2.5016286644951142, + "grad_norm": 8.346121650934089, + "learning_rate": 1.047635140847583e-05, + "loss": 0.3872, + "step": 4608 + }, + { + "epoch": 2.502171552660152, + "grad_norm": 14.17791420098929, + "learning_rate": 1.0472838531978495e-05, + "loss": 0.6932, + "step": 4609 + }, + { + "epoch": 2.50271444082519, + "grad_norm": 12.940166215438651, + "learning_rate": 1.046932559699974e-05, + "loss": 0.7045, + "step": 4610 + }, + { + "epoch": 2.503257328990228, + "grad_norm": 8.154813733233013, + "learning_rate": 1.0465812603974048e-05, + "loss": 0.4225, + "step": 4611 + }, + { + "epoch": 2.5038002171552662, + "grad_norm": 13.356750387494248, + "learning_rate": 1.0462299553335911e-05, + "loss": 0.7951, + "step": 4612 + }, + { + "epoch": 2.504343105320304, + "grad_norm": 8.683404141286427, + "learning_rate": 1.0458786445519827e-05, + "loss": 0.5871, + "step": 4613 + }, + { + "epoch": 2.504885993485342, + "grad_norm": 5.750758875120564, + "learning_rate": 1.0455273280960306e-05, + "loss": 0.3815, + "step": 4614 + }, + { + "epoch": 2.50542888165038, + "grad_norm": 9.328766968136149, + "learning_rate": 1.0451760060091857e-05, + "loss": 0.6451, + "step": 4615 + }, + { + "epoch": 2.5059717698154182, + "grad_norm": 9.704349389930364, + "learning_rate": 1.0448246783349002e-05, + "loss": 0.5136, + "step": 4616 + }, + { + "epoch": 2.506514657980456, + "grad_norm": 10.438938086465326, + "learning_rate": 1.0444733451166277e-05, + "loss": 0.6278, + "step": 4617 + }, + { + "epoch": 2.507057546145494, + "grad_norm": 11.019760442005925, + "learning_rate": 1.0441220063978208e-05, + "loss": 0.7477, + "step": 4618 + }, + { + "epoch": 2.507600434310532, + "grad_norm": 11.738067010881544, + "learning_rate": 1.0437706622219337e-05, + "loss": 0.5768, + "step": 4619 + }, + { + "epoch": 2.5081433224755703, + "grad_norm": 9.601242936349621, + "learning_rate": 1.0434193126324215e-05, + "loss": 0.4236, + "step": 4620 + }, + { + "epoch": 2.508686210640608, + "grad_norm": 8.041097216565952, + "learning_rate": 1.0430679576727392e-05, + "loss": 0.494, + "step": 4621 + }, + { + "epoch": 2.509229098805646, + "grad_norm": 11.144286090694623, + "learning_rate": 1.0427165973863434e-05, + "loss": 0.7882, + "step": 4622 + }, + { + "epoch": 2.509771986970684, + "grad_norm": 14.369854225910128, + "learning_rate": 1.0423652318166905e-05, + "loss": 0.7112, + "step": 4623 + }, + { + "epoch": 2.5103148751357223, + "grad_norm": 9.688103222900846, + "learning_rate": 1.0420138610072386e-05, + "loss": 0.4564, + "step": 4624 + }, + { + "epoch": 2.51085776330076, + "grad_norm": 8.2914031160658, + "learning_rate": 1.0416624850014454e-05, + "loss": 0.4106, + "step": 4625 + }, + { + "epoch": 2.511400651465798, + "grad_norm": 13.145683230031937, + "learning_rate": 1.0413111038427696e-05, + "loss": 0.905, + "step": 4626 + }, + { + "epoch": 2.511943539630836, + "grad_norm": 10.544391990984453, + "learning_rate": 1.0409597175746709e-05, + "loss": 0.5791, + "step": 4627 + }, + { + "epoch": 2.5124864277958743, + "grad_norm": 8.830789770145405, + "learning_rate": 1.0406083262406084e-05, + "loss": 0.4806, + "step": 4628 + }, + { + "epoch": 2.513029315960912, + "grad_norm": 9.186020814962909, + "learning_rate": 1.0402569298840437e-05, + "loss": 0.7227, + "step": 4629 + }, + { + "epoch": 2.51357220412595, + "grad_norm": 8.395511535015437, + "learning_rate": 1.0399055285484378e-05, + "loss": 0.5372, + "step": 4630 + }, + { + "epoch": 2.514115092290988, + "grad_norm": 9.969741133621088, + "learning_rate": 1.0395541222772524e-05, + "loss": 0.5151, + "step": 4631 + }, + { + "epoch": 2.5146579804560263, + "grad_norm": 11.64674593889149, + "learning_rate": 1.0392027111139505e-05, + "loss": 0.6159, + "step": 4632 + }, + { + "epoch": 2.515200868621064, + "grad_norm": 12.194306814356107, + "learning_rate": 1.0388512951019945e-05, + "loss": 0.6977, + "step": 4633 + }, + { + "epoch": 2.515743756786102, + "grad_norm": 9.063328798639105, + "learning_rate": 1.0384998742848485e-05, + "loss": 0.6518, + "step": 4634 + }, + { + "epoch": 2.51628664495114, + "grad_norm": 8.709419510130173, + "learning_rate": 1.0381484487059764e-05, + "loss": 0.5571, + "step": 4635 + }, + { + "epoch": 2.5168295331161783, + "grad_norm": 10.875178380627338, + "learning_rate": 1.0377970184088438e-05, + "loss": 0.8384, + "step": 4636 + }, + { + "epoch": 2.517372421281216, + "grad_norm": 8.42478873312803, + "learning_rate": 1.0374455834369155e-05, + "loss": 0.6209, + "step": 4637 + }, + { + "epoch": 2.517915309446254, + "grad_norm": 13.875461781038107, + "learning_rate": 1.0370941438336577e-05, + "loss": 0.8787, + "step": 4638 + }, + { + "epoch": 2.518458197611292, + "grad_norm": 11.372446219871874, + "learning_rate": 1.036742699642537e-05, + "loss": 0.569, + "step": 4639 + }, + { + "epoch": 2.5190010857763303, + "grad_norm": 10.725098253853016, + "learning_rate": 1.036391250907021e-05, + "loss": 0.8431, + "step": 4640 + }, + { + "epoch": 2.519543973941368, + "grad_norm": 8.268642654673096, + "learning_rate": 1.0360397976705767e-05, + "loss": 0.5833, + "step": 4641 + }, + { + "epoch": 2.520086862106406, + "grad_norm": 10.560802722537995, + "learning_rate": 1.0356883399766731e-05, + "loss": 0.7495, + "step": 4642 + }, + { + "epoch": 2.520629750271444, + "grad_norm": 11.789754393456334, + "learning_rate": 1.0353368778687784e-05, + "loss": 0.644, + "step": 4643 + }, + { + "epoch": 2.5211726384364823, + "grad_norm": 10.236138559776574, + "learning_rate": 1.0349854113903625e-05, + "loss": 0.6544, + "step": 4644 + }, + { + "epoch": 2.52171552660152, + "grad_norm": 8.552964039446971, + "learning_rate": 1.0346339405848953e-05, + "loss": 0.5093, + "step": 4645 + }, + { + "epoch": 2.522258414766558, + "grad_norm": 8.301860195209034, + "learning_rate": 1.0342824654958471e-05, + "loss": 0.6285, + "step": 4646 + }, + { + "epoch": 2.522801302931596, + "grad_norm": 9.238558119037732, + "learning_rate": 1.0339309861666889e-05, + "loss": 0.6519, + "step": 4647 + }, + { + "epoch": 2.5233441910966343, + "grad_norm": 9.840812219132841, + "learning_rate": 1.0335795026408922e-05, + "loss": 0.5874, + "step": 4648 + }, + { + "epoch": 2.523887079261672, + "grad_norm": 8.108461441633077, + "learning_rate": 1.0332280149619291e-05, + "loss": 0.5793, + "step": 4649 + }, + { + "epoch": 2.52442996742671, + "grad_norm": 9.446694020377457, + "learning_rate": 1.032876523173272e-05, + "loss": 0.607, + "step": 4650 + }, + { + "epoch": 2.524972855591748, + "grad_norm": 11.88681485670853, + "learning_rate": 1.0325250273183944e-05, + "loss": 1.0745, + "step": 4651 + }, + { + "epoch": 2.5255157437567863, + "grad_norm": 8.347193432264572, + "learning_rate": 1.0321735274407694e-05, + "loss": 0.8431, + "step": 4652 + }, + { + "epoch": 2.526058631921824, + "grad_norm": 9.639244837832804, + "learning_rate": 1.0318220235838715e-05, + "loss": 0.7827, + "step": 4653 + }, + { + "epoch": 2.526601520086862, + "grad_norm": 12.239750504764533, + "learning_rate": 1.0314705157911744e-05, + "loss": 0.6735, + "step": 4654 + }, + { + "epoch": 2.5271444082519, + "grad_norm": 15.826940901953726, + "learning_rate": 1.0311190041061543e-05, + "loss": 0.9182, + "step": 4655 + }, + { + "epoch": 2.5276872964169383, + "grad_norm": 11.015508331511823, + "learning_rate": 1.0307674885722863e-05, + "loss": 0.7834, + "step": 4656 + }, + { + "epoch": 2.528230184581976, + "grad_norm": 11.163753697057714, + "learning_rate": 1.030415969233046e-05, + "loss": 0.6613, + "step": 4657 + }, + { + "epoch": 2.528773072747014, + "grad_norm": 8.747908439433298, + "learning_rate": 1.03006444613191e-05, + "loss": 0.6974, + "step": 4658 + }, + { + "epoch": 2.529315960912052, + "grad_norm": 8.370456762868022, + "learning_rate": 1.0297129193123555e-05, + "loss": 0.5343, + "step": 4659 + }, + { + "epoch": 2.5298588490770904, + "grad_norm": 13.46061984772866, + "learning_rate": 1.0293613888178597e-05, + "loss": 1.3054, + "step": 4660 + }, + { + "epoch": 2.530401737242128, + "grad_norm": 10.948007938187232, + "learning_rate": 1.0290098546919007e-05, + "loss": 0.7028, + "step": 4661 + }, + { + "epoch": 2.530944625407166, + "grad_norm": 8.833095994203507, + "learning_rate": 1.0286583169779568e-05, + "loss": 0.7663, + "step": 4662 + }, + { + "epoch": 2.531487513572204, + "grad_norm": 10.283493007620729, + "learning_rate": 1.0283067757195062e-05, + "loss": 0.7804, + "step": 4663 + }, + { + "epoch": 2.5320304017372424, + "grad_norm": 13.55508996688301, + "learning_rate": 1.0279552309600288e-05, + "loss": 0.656, + "step": 4664 + }, + { + "epoch": 2.53257328990228, + "grad_norm": 8.738397704085456, + "learning_rate": 1.0276036827430036e-05, + "loss": 0.4902, + "step": 4665 + }, + { + "epoch": 2.533116178067318, + "grad_norm": 10.260011622402263, + "learning_rate": 1.0272521311119113e-05, + "loss": 0.4798, + "step": 4666 + }, + { + "epoch": 2.533659066232356, + "grad_norm": 10.743076748133591, + "learning_rate": 1.0269005761102315e-05, + "loss": 0.9671, + "step": 4667 + }, + { + "epoch": 2.5342019543973944, + "grad_norm": 11.846945314251228, + "learning_rate": 1.0265490177814458e-05, + "loss": 0.7168, + "step": 4668 + }, + { + "epoch": 2.534744842562432, + "grad_norm": 9.129703791341635, + "learning_rate": 1.0261974561690356e-05, + "loss": 0.6008, + "step": 4669 + }, + { + "epoch": 2.53528773072747, + "grad_norm": 8.464074733631888, + "learning_rate": 1.025845891316482e-05, + "loss": 0.4327, + "step": 4670 + }, + { + "epoch": 2.535830618892508, + "grad_norm": 10.894552922917367, + "learning_rate": 1.0254943232672671e-05, + "loss": 0.5627, + "step": 4671 + }, + { + "epoch": 2.5363735070575464, + "grad_norm": 12.365454725760225, + "learning_rate": 1.025142752064874e-05, + "loss": 0.606, + "step": 4672 + }, + { + "epoch": 2.536916395222584, + "grad_norm": 10.066952200044973, + "learning_rate": 1.0247911777527852e-05, + "loss": 0.7503, + "step": 4673 + }, + { + "epoch": 2.537459283387622, + "grad_norm": 8.386860314534095, + "learning_rate": 1.0244396003744836e-05, + "loss": 0.7663, + "step": 4674 + }, + { + "epoch": 2.53800217155266, + "grad_norm": 9.309667750090998, + "learning_rate": 1.024088019973454e-05, + "loss": 0.5658, + "step": 4675 + }, + { + "epoch": 2.5385450597176984, + "grad_norm": 13.530278392237364, + "learning_rate": 1.0237364365931794e-05, + "loss": 0.799, + "step": 4676 + }, + { + "epoch": 2.539087947882736, + "grad_norm": 11.344034257453236, + "learning_rate": 1.0233848502771444e-05, + "loss": 0.5931, + "step": 4677 + }, + { + "epoch": 2.539630836047774, + "grad_norm": 9.26197757297852, + "learning_rate": 1.0230332610688338e-05, + "loss": 0.5419, + "step": 4678 + }, + { + "epoch": 2.540173724212812, + "grad_norm": 12.620486559012823, + "learning_rate": 1.022681669011733e-05, + "loss": 0.6746, + "step": 4679 + }, + { + "epoch": 2.5407166123778504, + "grad_norm": 13.640092752737832, + "learning_rate": 1.0223300741493268e-05, + "loss": 0.8806, + "step": 4680 + }, + { + "epoch": 2.541259500542888, + "grad_norm": 10.435820181571753, + "learning_rate": 1.0219784765251018e-05, + "loss": 0.6876, + "step": 4681 + }, + { + "epoch": 2.541802388707926, + "grad_norm": 11.624678093656978, + "learning_rate": 1.0216268761825436e-05, + "loss": 0.7377, + "step": 4682 + }, + { + "epoch": 2.542345276872964, + "grad_norm": 10.361224010713974, + "learning_rate": 1.021275273165139e-05, + "loss": 0.5911, + "step": 4683 + }, + { + "epoch": 2.5428881650380024, + "grad_norm": 6.869744257762391, + "learning_rate": 1.0209236675163747e-05, + "loss": 0.4204, + "step": 4684 + }, + { + "epoch": 2.54343105320304, + "grad_norm": 16.790885412103638, + "learning_rate": 1.020572059279738e-05, + "loss": 0.8837, + "step": 4685 + }, + { + "epoch": 2.543973941368078, + "grad_norm": 9.226978433564653, + "learning_rate": 1.0202204484987159e-05, + "loss": 0.4686, + "step": 4686 + }, + { + "epoch": 2.544516829533116, + "grad_norm": 7.391909229375343, + "learning_rate": 1.019868835216797e-05, + "loss": 0.6996, + "step": 4687 + }, + { + "epoch": 2.5450597176981544, + "grad_norm": 9.103089555285077, + "learning_rate": 1.0195172194774684e-05, + "loss": 0.7243, + "step": 4688 + }, + { + "epoch": 2.545602605863192, + "grad_norm": 7.27509790187114, + "learning_rate": 1.0191656013242192e-05, + "loss": 0.4612, + "step": 4689 + }, + { + "epoch": 2.54614549402823, + "grad_norm": 6.939999239823949, + "learning_rate": 1.0188139808005379e-05, + "loss": 0.3483, + "step": 4690 + }, + { + "epoch": 2.546688382193268, + "grad_norm": 9.036119044447904, + "learning_rate": 1.0184623579499133e-05, + "loss": 0.5822, + "step": 4691 + }, + { + "epoch": 2.5472312703583064, + "grad_norm": 15.325233133671158, + "learning_rate": 1.018110732815835e-05, + "loss": 0.9375, + "step": 4692 + }, + { + "epoch": 2.547774158523344, + "grad_norm": 13.335480222530155, + "learning_rate": 1.017759105441792e-05, + "loss": 0.8165, + "step": 4693 + }, + { + "epoch": 2.548317046688382, + "grad_norm": 10.097959894991517, + "learning_rate": 1.0174074758712751e-05, + "loss": 0.7773, + "step": 4694 + }, + { + "epoch": 2.54885993485342, + "grad_norm": 9.7484700487623, + "learning_rate": 1.0170558441477738e-05, + "loss": 0.7805, + "step": 4695 + }, + { + "epoch": 2.5494028230184584, + "grad_norm": 9.687849482731272, + "learning_rate": 1.0167042103147784e-05, + "loss": 0.6675, + "step": 4696 + }, + { + "epoch": 2.549945711183496, + "grad_norm": 12.578144763595915, + "learning_rate": 1.0163525744157796e-05, + "loss": 0.7757, + "step": 4697 + }, + { + "epoch": 2.550488599348534, + "grad_norm": 9.03070780679602, + "learning_rate": 1.0160009364942684e-05, + "loss": 0.5321, + "step": 4698 + }, + { + "epoch": 2.5510314875135722, + "grad_norm": 10.632448946778698, + "learning_rate": 1.0156492965937357e-05, + "loss": 0.5589, + "step": 4699 + }, + { + "epoch": 2.5515743756786105, + "grad_norm": 10.053170332804227, + "learning_rate": 1.015297654757673e-05, + "loss": 0.4974, + "step": 4700 + }, + { + "epoch": 2.5521172638436482, + "grad_norm": 9.833516973903698, + "learning_rate": 1.0149460110295722e-05, + "loss": 0.6022, + "step": 4701 + }, + { + "epoch": 2.552660152008686, + "grad_norm": 16.704821665027783, + "learning_rate": 1.0145943654529248e-05, + "loss": 0.776, + "step": 4702 + }, + { + "epoch": 2.5532030401737242, + "grad_norm": 12.589948540180313, + "learning_rate": 1.0142427180712231e-05, + "loss": 0.8549, + "step": 4703 + }, + { + "epoch": 2.5537459283387625, + "grad_norm": 10.414828597120934, + "learning_rate": 1.0138910689279596e-05, + "loss": 0.7209, + "step": 4704 + }, + { + "epoch": 2.5542888165038002, + "grad_norm": 10.158714848928406, + "learning_rate": 1.0135394180666261e-05, + "loss": 0.6835, + "step": 4705 + }, + { + "epoch": 2.554831704668838, + "grad_norm": 8.255885711809917, + "learning_rate": 1.013187765530716e-05, + "loss": 0.4542, + "step": 4706 + }, + { + "epoch": 2.5553745928338762, + "grad_norm": 10.386471672795576, + "learning_rate": 1.012836111363722e-05, + "loss": 0.9478, + "step": 4707 + }, + { + "epoch": 2.5559174809989145, + "grad_norm": 11.828466617801858, + "learning_rate": 1.0124844556091374e-05, + "loss": 0.7975, + "step": 4708 + }, + { + "epoch": 2.5564603691639523, + "grad_norm": 10.742130491540307, + "learning_rate": 1.0121327983104555e-05, + "loss": 0.9712, + "step": 4709 + }, + { + "epoch": 2.55700325732899, + "grad_norm": 13.314461670532042, + "learning_rate": 1.0117811395111695e-05, + "loss": 0.9351, + "step": 4710 + }, + { + "epoch": 2.5575461454940283, + "grad_norm": 10.15371491897671, + "learning_rate": 1.0114294792547733e-05, + "loss": 0.6056, + "step": 4711 + }, + { + "epoch": 2.5580890336590665, + "grad_norm": 8.317356068473407, + "learning_rate": 1.011077817584761e-05, + "loss": 0.6411, + "step": 4712 + }, + { + "epoch": 2.5586319218241043, + "grad_norm": 11.215419343504372, + "learning_rate": 1.0107261545446267e-05, + "loss": 0.5411, + "step": 4713 + }, + { + "epoch": 2.559174809989142, + "grad_norm": 10.343171151673559, + "learning_rate": 1.0103744901778645e-05, + "loss": 0.5395, + "step": 4714 + }, + { + "epoch": 2.5597176981541803, + "grad_norm": 14.089833480392985, + "learning_rate": 1.0100228245279688e-05, + "loss": 0.9855, + "step": 4715 + }, + { + "epoch": 2.5602605863192185, + "grad_norm": 14.836069456421422, + "learning_rate": 1.0096711576384344e-05, + "loss": 0.8219, + "step": 4716 + }, + { + "epoch": 2.5608034744842563, + "grad_norm": 16.933747543735457, + "learning_rate": 1.0093194895527556e-05, + "loss": 0.9341, + "step": 4717 + }, + { + "epoch": 2.561346362649294, + "grad_norm": 13.574017237717102, + "learning_rate": 1.008967820314427e-05, + "loss": 1.1551, + "step": 4718 + }, + { + "epoch": 2.5618892508143323, + "grad_norm": 8.877246747020687, + "learning_rate": 1.0086161499669447e-05, + "loss": 0.6341, + "step": 4719 + }, + { + "epoch": 2.5624321389793705, + "grad_norm": 9.391996392121968, + "learning_rate": 1.0082644785538031e-05, + "loss": 0.5285, + "step": 4720 + }, + { + "epoch": 2.5629750271444083, + "grad_norm": 10.876272607307573, + "learning_rate": 1.0079128061184977e-05, + "loss": 0.5678, + "step": 4721 + }, + { + "epoch": 2.563517915309446, + "grad_norm": 12.516236640792583, + "learning_rate": 1.0075611327045236e-05, + "loss": 0.8382, + "step": 4722 + }, + { + "epoch": 2.5640608034744843, + "grad_norm": 6.395711230996481, + "learning_rate": 1.0072094583553769e-05, + "loss": 0.3785, + "step": 4723 + }, + { + "epoch": 2.5646036916395225, + "grad_norm": 8.671727273867468, + "learning_rate": 1.0068577831145526e-05, + "loss": 0.5039, + "step": 4724 + }, + { + "epoch": 2.5651465798045603, + "grad_norm": 12.77911107984442, + "learning_rate": 1.0065061070255469e-05, + "loss": 0.7647, + "step": 4725 + }, + { + "epoch": 2.565689467969598, + "grad_norm": 9.108473588917152, + "learning_rate": 1.0061544301318556e-05, + "loss": 0.798, + "step": 4726 + }, + { + "epoch": 2.5662323561346363, + "grad_norm": 8.776410877174934, + "learning_rate": 1.0058027524769742e-05, + "loss": 0.6388, + "step": 4727 + }, + { + "epoch": 2.5667752442996745, + "grad_norm": 10.392346204038525, + "learning_rate": 1.0054510741043995e-05, + "loss": 0.8178, + "step": 4728 + }, + { + "epoch": 2.5673181324647123, + "grad_norm": 8.611264791773579, + "learning_rate": 1.005099395057627e-05, + "loss": 0.5022, + "step": 4729 + }, + { + "epoch": 2.56786102062975, + "grad_norm": 8.228466341722912, + "learning_rate": 1.0047477153801534e-05, + "loss": 0.6572, + "step": 4730 + }, + { + "epoch": 2.5684039087947883, + "grad_norm": 8.410912812963819, + "learning_rate": 1.0043960351154745e-05, + "loss": 0.4526, + "step": 4731 + }, + { + "epoch": 2.5689467969598265, + "grad_norm": 8.24917306612946, + "learning_rate": 1.0040443543070872e-05, + "loss": 0.5492, + "step": 4732 + }, + { + "epoch": 2.5694896851248643, + "grad_norm": 7.578877723490418, + "learning_rate": 1.0036926729984878e-05, + "loss": 0.5482, + "step": 4733 + }, + { + "epoch": 2.570032573289902, + "grad_norm": 8.464719447395414, + "learning_rate": 1.003340991233173e-05, + "loss": 0.5236, + "step": 4734 + }, + { + "epoch": 2.5705754614549403, + "grad_norm": 10.297009632330814, + "learning_rate": 1.0029893090546385e-05, + "loss": 0.6747, + "step": 4735 + }, + { + "epoch": 2.5711183496199785, + "grad_norm": 9.895183624345862, + "learning_rate": 1.0026376265063818e-05, + "loss": 0.7036, + "step": 4736 + }, + { + "epoch": 2.5716612377850163, + "grad_norm": 16.034084078355107, + "learning_rate": 1.0022859436318988e-05, + "loss": 0.8583, + "step": 4737 + }, + { + "epoch": 2.572204125950054, + "grad_norm": 10.881395091863821, + "learning_rate": 1.0019342604746871e-05, + "loss": 0.7214, + "step": 4738 + }, + { + "epoch": 2.5727470141150923, + "grad_norm": 9.051906164481618, + "learning_rate": 1.0015825770782428e-05, + "loss": 0.36, + "step": 4739 + }, + { + "epoch": 2.5732899022801305, + "grad_norm": 11.71081741433651, + "learning_rate": 1.001230893486063e-05, + "loss": 0.8467, + "step": 4740 + }, + { + "epoch": 2.5738327904451683, + "grad_norm": 8.4727788666991, + "learning_rate": 1.0008792097416442e-05, + "loss": 0.5312, + "step": 4741 + }, + { + "epoch": 2.574375678610206, + "grad_norm": 10.308791039753308, + "learning_rate": 1.0005275258884835e-05, + "loss": 0.7132, + "step": 4742 + }, + { + "epoch": 2.5749185667752443, + "grad_norm": 13.591326014197278, + "learning_rate": 1.0001758419700773e-05, + "loss": 0.7087, + "step": 4743 + }, + { + "epoch": 2.5754614549402826, + "grad_norm": 9.610601406060551, + "learning_rate": 9.998241580299229e-06, + "loss": 0.6708, + "step": 4744 + }, + { + "epoch": 2.5760043431053203, + "grad_norm": 10.04740706892737, + "learning_rate": 9.994724741115169e-06, + "loss": 0.6644, + "step": 4745 + }, + { + "epoch": 2.576547231270358, + "grad_norm": 7.760395046863199, + "learning_rate": 9.991207902583558e-06, + "loss": 0.4361, + "step": 4746 + }, + { + "epoch": 2.5770901194353963, + "grad_norm": 9.097495186854097, + "learning_rate": 9.987691065139373e-06, + "loss": 0.5613, + "step": 4747 + }, + { + "epoch": 2.5776330076004346, + "grad_norm": 9.584167359092984, + "learning_rate": 9.984174229217572e-06, + "loss": 0.6408, + "step": 4748 + }, + { + "epoch": 2.5781758957654723, + "grad_norm": 10.329246788617224, + "learning_rate": 9.980657395253132e-06, + "loss": 0.6776, + "step": 4749 + }, + { + "epoch": 2.57871878393051, + "grad_norm": 10.000264678013584, + "learning_rate": 9.977140563681015e-06, + "loss": 0.7489, + "step": 4750 + }, + { + "epoch": 2.5792616720955484, + "grad_norm": 6.80240569067885, + "learning_rate": 9.973623734936185e-06, + "loss": 0.4639, + "step": 4751 + }, + { + "epoch": 2.5798045602605866, + "grad_norm": 7.578175941165148, + "learning_rate": 9.97010690945362e-06, + "loss": 0.38, + "step": 4752 + }, + { + "epoch": 2.5803474484256244, + "grad_norm": 12.48681779678407, + "learning_rate": 9.966590087668274e-06, + "loss": 0.8046, + "step": 4753 + }, + { + "epoch": 2.580890336590662, + "grad_norm": 13.31142746407252, + "learning_rate": 9.963073270015126e-06, + "loss": 0.7671, + "step": 4754 + }, + { + "epoch": 2.5814332247557004, + "grad_norm": 8.987640426318515, + "learning_rate": 9.959556456929128e-06, + "loss": 0.4691, + "step": 4755 + }, + { + "epoch": 2.5819761129207386, + "grad_norm": 9.551368066054666, + "learning_rate": 9.956039648845257e-06, + "loss": 0.6207, + "step": 4756 + }, + { + "epoch": 2.5825190010857764, + "grad_norm": 17.060876787385865, + "learning_rate": 9.952522846198468e-06, + "loss": 0.9276, + "step": 4757 + }, + { + "epoch": 2.583061889250814, + "grad_norm": 9.760247190255836, + "learning_rate": 9.949006049423731e-06, + "loss": 0.539, + "step": 4758 + }, + { + "epoch": 2.5836047774158524, + "grad_norm": 14.929770896239804, + "learning_rate": 9.94548925895601e-06, + "loss": 1.0896, + "step": 4759 + }, + { + "epoch": 2.5841476655808906, + "grad_norm": 14.902014520576888, + "learning_rate": 9.94197247523026e-06, + "loss": 1.0207, + "step": 4760 + }, + { + "epoch": 2.5846905537459284, + "grad_norm": 9.630889650448959, + "learning_rate": 9.93845569868145e-06, + "loss": 0.6322, + "step": 4761 + }, + { + "epoch": 2.585233441910966, + "grad_norm": 6.534627620002312, + "learning_rate": 9.934938929744535e-06, + "loss": 0.2627, + "step": 4762 + }, + { + "epoch": 2.5857763300760044, + "grad_norm": 10.701147725996508, + "learning_rate": 9.931422168854476e-06, + "loss": 0.6789, + "step": 4763 + }, + { + "epoch": 2.5863192182410426, + "grad_norm": 9.539331065006966, + "learning_rate": 9.927905416446233e-06, + "loss": 0.7109, + "step": 4764 + }, + { + "epoch": 2.5868621064060804, + "grad_norm": 9.979058002114003, + "learning_rate": 9.924388672954766e-06, + "loss": 0.6497, + "step": 4765 + }, + { + "epoch": 2.587404994571118, + "grad_norm": 9.638452627150874, + "learning_rate": 9.920871938815024e-06, + "loss": 0.6695, + "step": 4766 + }, + { + "epoch": 2.5879478827361564, + "grad_norm": 12.326422327522844, + "learning_rate": 9.91735521446197e-06, + "loss": 0.6354, + "step": 4767 + }, + { + "epoch": 2.5884907709011946, + "grad_norm": 8.571541287165285, + "learning_rate": 9.913838500330553e-06, + "loss": 0.3939, + "step": 4768 + }, + { + "epoch": 2.5890336590662324, + "grad_norm": 9.709310867338898, + "learning_rate": 9.910321796855732e-06, + "loss": 0.5866, + "step": 4769 + }, + { + "epoch": 2.58957654723127, + "grad_norm": 8.48642736575235, + "learning_rate": 9.90680510447245e-06, + "loss": 0.8167, + "step": 4770 + }, + { + "epoch": 2.5901194353963084, + "grad_norm": 9.153780615106163, + "learning_rate": 9.90328842361566e-06, + "loss": 0.6551, + "step": 4771 + }, + { + "epoch": 2.5906623235613466, + "grad_norm": 9.331431330636146, + "learning_rate": 9.899771754720315e-06, + "loss": 0.598, + "step": 4772 + }, + { + "epoch": 2.5912052117263844, + "grad_norm": 10.71941498109364, + "learning_rate": 9.896255098221357e-06, + "loss": 0.6996, + "step": 4773 + }, + { + "epoch": 2.591748099891422, + "grad_norm": 11.415708336876241, + "learning_rate": 9.892738454553736e-06, + "loss": 0.8247, + "step": 4774 + }, + { + "epoch": 2.5922909880564604, + "grad_norm": 9.526809613609492, + "learning_rate": 9.889221824152391e-06, + "loss": 0.5088, + "step": 4775 + }, + { + "epoch": 2.5928338762214986, + "grad_norm": 8.911159845652058, + "learning_rate": 9.885705207452268e-06, + "loss": 0.3377, + "step": 4776 + }, + { + "epoch": 2.5933767643865364, + "grad_norm": 13.451504414556261, + "learning_rate": 9.882188604888307e-06, + "loss": 0.9247, + "step": 4777 + }, + { + "epoch": 2.593919652551574, + "grad_norm": 10.251509263282227, + "learning_rate": 9.87867201689545e-06, + "loss": 0.4384, + "step": 4778 + }, + { + "epoch": 2.5944625407166124, + "grad_norm": 9.955424126921221, + "learning_rate": 9.875155443908631e-06, + "loss": 0.6318, + "step": 4779 + }, + { + "epoch": 2.5950054288816506, + "grad_norm": 8.584868222198555, + "learning_rate": 9.871638886362782e-06, + "loss": 0.5148, + "step": 4780 + }, + { + "epoch": 2.5955483170466884, + "grad_norm": 16.858447088677025, + "learning_rate": 9.868122344692846e-06, + "loss": 1.2564, + "step": 4781 + }, + { + "epoch": 2.596091205211726, + "grad_norm": 9.7339544501354, + "learning_rate": 9.86460581933374e-06, + "loss": 0.7158, + "step": 4782 + }, + { + "epoch": 2.5966340933767644, + "grad_norm": 11.03551156903236, + "learning_rate": 9.861089310720409e-06, + "loss": 0.5136, + "step": 4783 + }, + { + "epoch": 2.5971769815418027, + "grad_norm": 11.173351784755749, + "learning_rate": 9.857572819287768e-06, + "loss": 0.6439, + "step": 4784 + }, + { + "epoch": 2.5977198697068404, + "grad_norm": 14.318584084756125, + "learning_rate": 9.854056345470754e-06, + "loss": 1.2827, + "step": 4785 + }, + { + "epoch": 2.598262757871878, + "grad_norm": 11.132635136137834, + "learning_rate": 9.850539889704278e-06, + "loss": 0.499, + "step": 4786 + }, + { + "epoch": 2.5988056460369164, + "grad_norm": 11.710993594625656, + "learning_rate": 9.847023452423272e-06, + "loss": 0.466, + "step": 4787 + }, + { + "epoch": 2.5993485342019547, + "grad_norm": 12.05431003500401, + "learning_rate": 9.843507034062646e-06, + "loss": 1.002, + "step": 4788 + }, + { + "epoch": 2.5998914223669924, + "grad_norm": 12.651344078635795, + "learning_rate": 9.83999063505732e-06, + "loss": 0.6906, + "step": 4789 + }, + { + "epoch": 2.6004343105320302, + "grad_norm": 11.592009871908612, + "learning_rate": 9.83647425584221e-06, + "loss": 0.908, + "step": 4790 + }, + { + "epoch": 2.6009771986970684, + "grad_norm": 11.164823794029719, + "learning_rate": 9.83295789685222e-06, + "loss": 1.1006, + "step": 4791 + }, + { + "epoch": 2.6015200868621067, + "grad_norm": 18.55453736068596, + "learning_rate": 9.829441558522267e-06, + "loss": 0.6774, + "step": 4792 + }, + { + "epoch": 2.6020629750271445, + "grad_norm": 9.004957503789628, + "learning_rate": 9.825925241287249e-06, + "loss": 0.612, + "step": 4793 + }, + { + "epoch": 2.6026058631921822, + "grad_norm": 15.331324393349687, + "learning_rate": 9.822408945582081e-06, + "loss": 0.588, + "step": 4794 + }, + { + "epoch": 2.6031487513572205, + "grad_norm": 11.30698173094918, + "learning_rate": 9.818892671841653e-06, + "loss": 0.8413, + "step": 4795 + }, + { + "epoch": 2.6036916395222587, + "grad_norm": 10.970237049386885, + "learning_rate": 9.815376420500869e-06, + "loss": 0.5687, + "step": 4796 + }, + { + "epoch": 2.6042345276872965, + "grad_norm": 12.062898313733237, + "learning_rate": 9.811860191994623e-06, + "loss": 0.88, + "step": 4797 + }, + { + "epoch": 2.6047774158523342, + "grad_norm": 13.760783812579094, + "learning_rate": 9.808343986757811e-06, + "loss": 0.7319, + "step": 4798 + }, + { + "epoch": 2.6053203040173725, + "grad_norm": 10.142158775926303, + "learning_rate": 9.80482780522532e-06, + "loss": 0.5404, + "step": 4799 + }, + { + "epoch": 2.6058631921824107, + "grad_norm": 14.235611160734313, + "learning_rate": 9.801311647832035e-06, + "loss": 0.6849, + "step": 4800 + }, + { + "epoch": 2.6064060803474485, + "grad_norm": 14.33484497964673, + "learning_rate": 9.797795515012843e-06, + "loss": 1.0677, + "step": 4801 + }, + { + "epoch": 2.6069489685124863, + "grad_norm": 12.238534475065178, + "learning_rate": 9.794279407202624e-06, + "loss": 0.8279, + "step": 4802 + }, + { + "epoch": 2.6074918566775245, + "grad_norm": 13.51950414574984, + "learning_rate": 9.790763324836256e-06, + "loss": 0.8372, + "step": 4803 + }, + { + "epoch": 2.6080347448425627, + "grad_norm": 9.360760584245645, + "learning_rate": 9.78724726834861e-06, + "loss": 0.666, + "step": 4804 + }, + { + "epoch": 2.6085776330076005, + "grad_norm": 10.793352407977682, + "learning_rate": 9.783731238174566e-06, + "loss": 0.8308, + "step": 4805 + }, + { + "epoch": 2.6091205211726383, + "grad_norm": 13.862491939748782, + "learning_rate": 9.780215234748984e-06, + "loss": 0.8018, + "step": 4806 + }, + { + "epoch": 2.6096634093376765, + "grad_norm": 8.019652931555422, + "learning_rate": 9.776699258506734e-06, + "loss": 0.3801, + "step": 4807 + }, + { + "epoch": 2.6102062975027147, + "grad_norm": 10.161487802712038, + "learning_rate": 9.773183309882675e-06, + "loss": 0.6723, + "step": 4808 + }, + { + "epoch": 2.6107491856677525, + "grad_norm": 9.64703000959591, + "learning_rate": 9.769667389311664e-06, + "loss": 0.6294, + "step": 4809 + }, + { + "epoch": 2.6112920738327903, + "grad_norm": 9.792927662736604, + "learning_rate": 9.766151497228561e-06, + "loss": 0.4596, + "step": 4810 + }, + { + "epoch": 2.6118349619978285, + "grad_norm": 10.336511067045754, + "learning_rate": 9.76263563406821e-06, + "loss": 0.6494, + "step": 4811 + }, + { + "epoch": 2.6123778501628667, + "grad_norm": 12.121823172890695, + "learning_rate": 9.759119800265464e-06, + "loss": 1.0394, + "step": 4812 + }, + { + "epoch": 2.6129207383279045, + "grad_norm": 10.891668605270631, + "learning_rate": 9.755603996255164e-06, + "loss": 0.6212, + "step": 4813 + }, + { + "epoch": 2.6134636264929423, + "grad_norm": 9.599793914114507, + "learning_rate": 9.752088222472151e-06, + "loss": 0.6182, + "step": 4814 + }, + { + "epoch": 2.6140065146579805, + "grad_norm": 10.448317247236231, + "learning_rate": 9.748572479351263e-06, + "loss": 0.6779, + "step": 4815 + }, + { + "epoch": 2.6145494028230187, + "grad_norm": 12.703205416876392, + "learning_rate": 9.74505676732733e-06, + "loss": 0.8093, + "step": 4816 + }, + { + "epoch": 2.6150922909880565, + "grad_norm": 10.14294888595903, + "learning_rate": 9.741541086835182e-06, + "loss": 0.6493, + "step": 4817 + }, + { + "epoch": 2.6156351791530943, + "grad_norm": 11.455318908063175, + "learning_rate": 9.738025438309649e-06, + "loss": 0.9375, + "step": 4818 + }, + { + "epoch": 2.6161780673181325, + "grad_norm": 8.90202557317791, + "learning_rate": 9.734509822185545e-06, + "loss": 0.5978, + "step": 4819 + }, + { + "epoch": 2.6167209554831707, + "grad_norm": 10.38484629007874, + "learning_rate": 9.730994238897687e-06, + "loss": 0.9772, + "step": 4820 + }, + { + "epoch": 2.6172638436482085, + "grad_norm": 10.650781833239714, + "learning_rate": 9.72747868888089e-06, + "loss": 0.7354, + "step": 4821 + }, + { + "epoch": 2.6178067318132463, + "grad_norm": 11.462972783418381, + "learning_rate": 9.723963172569964e-06, + "loss": 0.8748, + "step": 4822 + }, + { + "epoch": 2.6183496199782845, + "grad_norm": 7.4804704893307505, + "learning_rate": 9.720447690399716e-06, + "loss": 0.4853, + "step": 4823 + }, + { + "epoch": 2.6188925081433228, + "grad_norm": 8.032622743372611, + "learning_rate": 9.716932242804938e-06, + "loss": 0.6484, + "step": 4824 + }, + { + "epoch": 2.6194353963083605, + "grad_norm": 10.141962204854094, + "learning_rate": 9.713416830220436e-06, + "loss": 0.669, + "step": 4825 + }, + { + "epoch": 2.6199782844733983, + "grad_norm": 11.48618310533414, + "learning_rate": 9.709901453080993e-06, + "loss": 0.6279, + "step": 4826 + }, + { + "epoch": 2.6205211726384365, + "grad_norm": 9.789449860925128, + "learning_rate": 9.706386111821406e-06, + "loss": 0.6072, + "step": 4827 + }, + { + "epoch": 2.6210640608034748, + "grad_norm": 8.10222084311329, + "learning_rate": 9.702870806876448e-06, + "loss": 0.6476, + "step": 4828 + }, + { + "epoch": 2.6216069489685125, + "grad_norm": 10.30951278757575, + "learning_rate": 9.699355538680902e-06, + "loss": 0.4691, + "step": 4829 + }, + { + "epoch": 2.6221498371335503, + "grad_norm": 10.960835467770405, + "learning_rate": 9.695840307669547e-06, + "loss": 0.7962, + "step": 4830 + }, + { + "epoch": 2.6226927252985885, + "grad_norm": 6.691408537531251, + "learning_rate": 9.69232511427714e-06, + "loss": 0.3644, + "step": 4831 + }, + { + "epoch": 2.6232356134636268, + "grad_norm": 7.019904615334962, + "learning_rate": 9.68880995893846e-06, + "loss": 0.4594, + "step": 4832 + }, + { + "epoch": 2.6237785016286646, + "grad_norm": 11.780644240657415, + "learning_rate": 9.685294842088256e-06, + "loss": 1.0625, + "step": 4833 + }, + { + "epoch": 2.6243213897937023, + "grad_norm": 7.34603375318287, + "learning_rate": 9.681779764161289e-06, + "loss": 0.3502, + "step": 4834 + }, + { + "epoch": 2.6248642779587406, + "grad_norm": 9.787656886442962, + "learning_rate": 9.678264725592306e-06, + "loss": 0.5054, + "step": 4835 + }, + { + "epoch": 2.6254071661237783, + "grad_norm": 9.687484513116981, + "learning_rate": 9.674749726816058e-06, + "loss": 0.8678, + "step": 4836 + }, + { + "epoch": 2.6259500542888166, + "grad_norm": 7.620468702427845, + "learning_rate": 9.671234768267284e-06, + "loss": 0.4672, + "step": 4837 + }, + { + "epoch": 2.6264929424538543, + "grad_norm": 14.051025442281276, + "learning_rate": 9.667719850380712e-06, + "loss": 0.6462, + "step": 4838 + }, + { + "epoch": 2.6270358306188926, + "grad_norm": 11.579748119704858, + "learning_rate": 9.664204973591081e-06, + "loss": 0.8856, + "step": 4839 + }, + { + "epoch": 2.6275787187839303, + "grad_norm": 8.348102354642592, + "learning_rate": 9.660690138333114e-06, + "loss": 0.5623, + "step": 4840 + }, + { + "epoch": 2.6281216069489686, + "grad_norm": 9.351950984929493, + "learning_rate": 9.657175345041532e-06, + "loss": 0.5809, + "step": 4841 + }, + { + "epoch": 2.6286644951140063, + "grad_norm": 12.421448876984805, + "learning_rate": 9.653660594151047e-06, + "loss": 0.7831, + "step": 4842 + }, + { + "epoch": 2.6292073832790446, + "grad_norm": 11.981212578199576, + "learning_rate": 9.650145886096376e-06, + "loss": 0.8767, + "step": 4843 + }, + { + "epoch": 2.6297502714440824, + "grad_norm": 10.659578351567276, + "learning_rate": 9.646631221312216e-06, + "loss": 0.7313, + "step": 4844 + }, + { + "epoch": 2.6302931596091206, + "grad_norm": 12.480207661204222, + "learning_rate": 9.643116600233274e-06, + "loss": 0.7412, + "step": 4845 + }, + { + "epoch": 2.6308360477741584, + "grad_norm": 9.131072461148838, + "learning_rate": 9.639602023294233e-06, + "loss": 0.5518, + "step": 4846 + }, + { + "epoch": 2.6313789359391966, + "grad_norm": 7.489089256780018, + "learning_rate": 9.636087490929793e-06, + "loss": 0.4321, + "step": 4847 + }, + { + "epoch": 2.6319218241042344, + "grad_norm": 12.12610683915495, + "learning_rate": 9.632573003574634e-06, + "loss": 0.8265, + "step": 4848 + }, + { + "epoch": 2.6324647122692726, + "grad_norm": 7.965586818160769, + "learning_rate": 9.629058561663426e-06, + "loss": 0.373, + "step": 4849 + }, + { + "epoch": 2.6330076004343104, + "grad_norm": 9.348866579305472, + "learning_rate": 9.62554416563085e-06, + "loss": 0.4274, + "step": 4850 + }, + { + "epoch": 2.6335504885993486, + "grad_norm": 10.13050086693858, + "learning_rate": 9.622029815911566e-06, + "loss": 0.5276, + "step": 4851 + }, + { + "epoch": 2.6340933767643864, + "grad_norm": 10.19180911105006, + "learning_rate": 9.61851551294024e-06, + "loss": 0.5871, + "step": 4852 + }, + { + "epoch": 2.6346362649294246, + "grad_norm": 11.184621464328854, + "learning_rate": 9.61500125715152e-06, + "loss": 0.6047, + "step": 4853 + }, + { + "epoch": 2.6351791530944624, + "grad_norm": 14.484355041392742, + "learning_rate": 9.611487048980058e-06, + "loss": 0.8628, + "step": 4854 + }, + { + "epoch": 2.6357220412595006, + "grad_norm": 13.070986825901516, + "learning_rate": 9.607972888860497e-06, + "loss": 0.7793, + "step": 4855 + }, + { + "epoch": 2.6362649294245384, + "grad_norm": 9.640708751060075, + "learning_rate": 9.604458777227477e-06, + "loss": 0.6016, + "step": 4856 + }, + { + "epoch": 2.6368078175895766, + "grad_norm": 11.05523237140924, + "learning_rate": 9.600944714515627e-06, + "loss": 0.5999, + "step": 4857 + }, + { + "epoch": 2.6373507057546144, + "grad_norm": 13.240640169530616, + "learning_rate": 9.597430701159565e-06, + "loss": 0.6779, + "step": 4858 + }, + { + "epoch": 2.6378935939196526, + "grad_norm": 10.878893821343203, + "learning_rate": 9.593916737593919e-06, + "loss": 0.6119, + "step": 4859 + }, + { + "epoch": 2.6384364820846904, + "grad_norm": 10.820526634571284, + "learning_rate": 9.590402824253295e-06, + "loss": 0.839, + "step": 4860 + }, + { + "epoch": 2.6389793702497286, + "grad_norm": 12.938750622065529, + "learning_rate": 9.586888961572307e-06, + "loss": 0.517, + "step": 4861 + }, + { + "epoch": 2.6395222584147664, + "grad_norm": 9.725430374953143, + "learning_rate": 9.583375149985547e-06, + "loss": 0.7604, + "step": 4862 + }, + { + "epoch": 2.6400651465798046, + "grad_norm": 11.840312760690757, + "learning_rate": 9.579861389927615e-06, + "loss": 0.7614, + "step": 4863 + }, + { + "epoch": 2.6406080347448424, + "grad_norm": 11.047185391206618, + "learning_rate": 9.576347681833093e-06, + "loss": 0.6351, + "step": 4864 + }, + { + "epoch": 2.6411509229098806, + "grad_norm": 9.986349266550256, + "learning_rate": 9.572834026136571e-06, + "loss": 0.5549, + "step": 4865 + }, + { + "epoch": 2.6416938110749184, + "grad_norm": 14.455847384105025, + "learning_rate": 9.56932042327261e-06, + "loss": 0.8206, + "step": 4866 + }, + { + "epoch": 2.6422366992399566, + "grad_norm": 10.377988559117977, + "learning_rate": 9.56580687367579e-06, + "loss": 0.5275, + "step": 4867 + }, + { + "epoch": 2.6427795874049944, + "grad_norm": 10.71910613137521, + "learning_rate": 9.562293377780668e-06, + "loss": 0.6958, + "step": 4868 + }, + { + "epoch": 2.6433224755700326, + "grad_norm": 8.635819291033194, + "learning_rate": 9.558779936021795e-06, + "loss": 0.3946, + "step": 4869 + }, + { + "epoch": 2.6438653637350704, + "grad_norm": 16.60173101533996, + "learning_rate": 9.555266548833728e-06, + "loss": 1.066, + "step": 4870 + }, + { + "epoch": 2.6444082519001086, + "grad_norm": 14.726517961222527, + "learning_rate": 9.551753216650998e-06, + "loss": 1.1356, + "step": 4871 + }, + { + "epoch": 2.6449511400651464, + "grad_norm": 11.123184991281361, + "learning_rate": 9.548239939908146e-06, + "loss": 0.757, + "step": 4872 + }, + { + "epoch": 2.6454940282301846, + "grad_norm": 11.666285426731331, + "learning_rate": 9.544726719039699e-06, + "loss": 0.7109, + "step": 4873 + }, + { + "epoch": 2.6460369163952224, + "grad_norm": 9.657892010966115, + "learning_rate": 9.541213554480176e-06, + "loss": 0.6191, + "step": 4874 + }, + { + "epoch": 2.6465798045602607, + "grad_norm": 9.984832056257446, + "learning_rate": 9.53770044666409e-06, + "loss": 0.8649, + "step": 4875 + }, + { + "epoch": 2.6471226927252984, + "grad_norm": 9.143852173958573, + "learning_rate": 9.534187396025955e-06, + "loss": 0.5282, + "step": 4876 + }, + { + "epoch": 2.6476655808903367, + "grad_norm": 8.129441981336987, + "learning_rate": 9.530674403000266e-06, + "loss": 0.3783, + "step": 4877 + }, + { + "epoch": 2.6482084690553744, + "grad_norm": 10.147175407433437, + "learning_rate": 9.527161468021508e-06, + "loss": 0.6676, + "step": 4878 + }, + { + "epoch": 2.6487513572204127, + "grad_norm": 11.015226101188722, + "learning_rate": 9.523648591524176e-06, + "loss": 0.5218, + "step": 4879 + }, + { + "epoch": 2.6492942453854504, + "grad_norm": 9.374140703922862, + "learning_rate": 9.520135773942743e-06, + "loss": 0.6754, + "step": 4880 + }, + { + "epoch": 2.6498371335504887, + "grad_norm": 8.84122429163583, + "learning_rate": 9.516623015711688e-06, + "loss": 0.5236, + "step": 4881 + }, + { + "epoch": 2.6503800217155264, + "grad_norm": 10.659047956662723, + "learning_rate": 9.513110317265463e-06, + "loss": 0.4742, + "step": 4882 + }, + { + "epoch": 2.6509229098805647, + "grad_norm": 8.798273983585764, + "learning_rate": 9.509597679038536e-06, + "loss": 0.4949, + "step": 4883 + }, + { + "epoch": 2.6514657980456025, + "grad_norm": 14.827321458575321, + "learning_rate": 9.506085101465344e-06, + "loss": 0.967, + "step": 4884 + }, + { + "epoch": 2.6520086862106407, + "grad_norm": 13.212543627120823, + "learning_rate": 9.502572584980338e-06, + "loss": 0.904, + "step": 4885 + }, + { + "epoch": 2.6525515743756785, + "grad_norm": 12.151700192108333, + "learning_rate": 9.499060130017947e-06, + "loss": 0.6512, + "step": 4886 + }, + { + "epoch": 2.6530944625407167, + "grad_norm": 9.051653631541132, + "learning_rate": 9.495547737012594e-06, + "loss": 0.5264, + "step": 4887 + }, + { + "epoch": 2.6536373507057545, + "grad_norm": 11.286333388065833, + "learning_rate": 9.492035406398706e-06, + "loss": 0.5864, + "step": 4888 + }, + { + "epoch": 2.6541802388707927, + "grad_norm": 8.303712468523708, + "learning_rate": 9.488523138610684e-06, + "loss": 0.5887, + "step": 4889 + }, + { + "epoch": 2.6547231270358305, + "grad_norm": 8.762643235318839, + "learning_rate": 9.485010934082939e-06, + "loss": 0.5079, + "step": 4890 + }, + { + "epoch": 2.6552660152008687, + "grad_norm": 13.211351989906888, + "learning_rate": 9.481498793249859e-06, + "loss": 0.8983, + "step": 4891 + }, + { + "epoch": 2.6558089033659065, + "grad_norm": 11.087558391384889, + "learning_rate": 9.477986716545834e-06, + "loss": 1.1671, + "step": 4892 + }, + { + "epoch": 2.6563517915309447, + "grad_norm": 6.825801048730381, + "learning_rate": 9.474474704405245e-06, + "loss": 0.3664, + "step": 4893 + }, + { + "epoch": 2.6568946796959825, + "grad_norm": 6.583440546159338, + "learning_rate": 9.470962757262465e-06, + "loss": 0.5464, + "step": 4894 + }, + { + "epoch": 2.6574375678610207, + "grad_norm": 8.987234989797157, + "learning_rate": 9.46745087555185e-06, + "loss": 0.5954, + "step": 4895 + }, + { + "epoch": 2.6579804560260585, + "grad_norm": 10.699367568639158, + "learning_rate": 9.463939059707763e-06, + "loss": 0.632, + "step": 4896 + }, + { + "epoch": 2.6585233441910967, + "grad_norm": 10.246298042329498, + "learning_rate": 9.460427310164548e-06, + "loss": 0.6766, + "step": 4897 + }, + { + "epoch": 2.6590662323561345, + "grad_norm": 12.300177238674, + "learning_rate": 9.456915627356542e-06, + "loss": 0.7277, + "step": 4898 + }, + { + "epoch": 2.6596091205211727, + "grad_norm": 9.893644041111745, + "learning_rate": 9.453404011718077e-06, + "loss": 0.644, + "step": 4899 + }, + { + "epoch": 2.6601520086862105, + "grad_norm": 10.80727748895235, + "learning_rate": 9.449892463683471e-06, + "loss": 0.5844, + "step": 4900 + }, + { + "epoch": 2.6606948968512487, + "grad_norm": 8.350836469669698, + "learning_rate": 9.446380983687046e-06, + "loss": 0.4745, + "step": 4901 + }, + { + "epoch": 2.6612377850162865, + "grad_norm": 10.697471057638213, + "learning_rate": 9.442869572163101e-06, + "loss": 0.5749, + "step": 4902 + }, + { + "epoch": 2.6617806731813247, + "grad_norm": 7.582465634030485, + "learning_rate": 9.43935822954594e-06, + "loss": 0.4481, + "step": 4903 + }, + { + "epoch": 2.6623235613463625, + "grad_norm": 10.01801356389796, + "learning_rate": 9.435846956269841e-06, + "loss": 0.5817, + "step": 4904 + }, + { + "epoch": 2.6628664495114007, + "grad_norm": 9.383471785643394, + "learning_rate": 9.432335752769092e-06, + "loss": 0.526, + "step": 4905 + }, + { + "epoch": 2.6634093376764385, + "grad_norm": 9.555536887097055, + "learning_rate": 9.428824619477964e-06, + "loss": 0.5056, + "step": 4906 + }, + { + "epoch": 2.6639522258414767, + "grad_norm": 7.636054626405326, + "learning_rate": 9.425313556830713e-06, + "loss": 0.4383, + "step": 4907 + }, + { + "epoch": 2.6644951140065145, + "grad_norm": 9.842725681691716, + "learning_rate": 9.421802565261602e-06, + "loss": 0.548, + "step": 4908 + }, + { + "epoch": 2.6650380021715527, + "grad_norm": 12.891942531084513, + "learning_rate": 9.418291645204865e-06, + "loss": 0.6489, + "step": 4909 + }, + { + "epoch": 2.6655808903365905, + "grad_norm": 15.298821348423894, + "learning_rate": 9.41478079709475e-06, + "loss": 0.8286, + "step": 4910 + }, + { + "epoch": 2.6661237785016287, + "grad_norm": 10.80698538838323, + "learning_rate": 9.411270021365475e-06, + "loss": 0.7704, + "step": 4911 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 10.837498285149955, + "learning_rate": 9.407759318451264e-06, + "loss": 0.5384, + "step": 4912 + }, + { + "epoch": 2.6672095548317047, + "grad_norm": 11.666848847579812, + "learning_rate": 9.404248688786322e-06, + "loss": 0.6402, + "step": 4913 + }, + { + "epoch": 2.6677524429967425, + "grad_norm": 12.038932116915525, + "learning_rate": 9.400738132804856e-06, + "loss": 0.559, + "step": 4914 + }, + { + "epoch": 2.6682953311617807, + "grad_norm": 13.411348889811451, + "learning_rate": 9.397227650941048e-06, + "loss": 0.9587, + "step": 4915 + }, + { + "epoch": 2.6688382193268185, + "grad_norm": 8.79354274091956, + "learning_rate": 9.393717243629091e-06, + "loss": 0.5627, + "step": 4916 + }, + { + "epoch": 2.6693811074918568, + "grad_norm": 8.956388294197298, + "learning_rate": 9.390206911303148e-06, + "loss": 0.5694, + "step": 4917 + }, + { + "epoch": 2.6699239956568945, + "grad_norm": 11.296789676569125, + "learning_rate": 9.386696654397384e-06, + "loss": 0.8356, + "step": 4918 + }, + { + "epoch": 2.6704668838219328, + "grad_norm": 8.778967068719545, + "learning_rate": 9.38318647334596e-06, + "loss": 0.5283, + "step": 4919 + }, + { + "epoch": 2.6710097719869705, + "grad_norm": 10.041165955923448, + "learning_rate": 9.379676368583011e-06, + "loss": 0.7492, + "step": 4920 + }, + { + "epoch": 2.6715526601520088, + "grad_norm": 7.152774382953927, + "learning_rate": 9.376166340542682e-06, + "loss": 0.4563, + "step": 4921 + }, + { + "epoch": 2.6720955483170465, + "grad_norm": 10.189040304576451, + "learning_rate": 9.37265638965909e-06, + "loss": 0.5869, + "step": 4922 + }, + { + "epoch": 2.6726384364820848, + "grad_norm": 8.47728562416046, + "learning_rate": 9.369146516366359e-06, + "loss": 0.6168, + "step": 4923 + }, + { + "epoch": 2.6731813246471225, + "grad_norm": 9.47866332824828, + "learning_rate": 9.365636721098588e-06, + "loss": 0.5233, + "step": 4924 + }, + { + "epoch": 2.6737242128121608, + "grad_norm": 11.264475439058582, + "learning_rate": 9.362127004289883e-06, + "loss": 0.693, + "step": 4925 + }, + { + "epoch": 2.6742671009771986, + "grad_norm": 11.495205201206721, + "learning_rate": 9.358617366374327e-06, + "loss": 0.587, + "step": 4926 + }, + { + "epoch": 2.6748099891422368, + "grad_norm": 10.081706772004408, + "learning_rate": 9.355107807785993e-06, + "loss": 0.65, + "step": 4927 + }, + { + "epoch": 2.6753528773072746, + "grad_norm": 8.912371373305913, + "learning_rate": 9.351598328958957e-06, + "loss": 0.6475, + "step": 4928 + }, + { + "epoch": 2.675895765472313, + "grad_norm": 8.434539074965906, + "learning_rate": 9.348088930327269e-06, + "loss": 0.4321, + "step": 4929 + }, + { + "epoch": 2.6764386536373506, + "grad_norm": 10.356652945785703, + "learning_rate": 9.344579612324984e-06, + "loss": 0.4243, + "step": 4930 + }, + { + "epoch": 2.676981541802389, + "grad_norm": 10.975361657654028, + "learning_rate": 9.341070375386134e-06, + "loss": 0.6042, + "step": 4931 + }, + { + "epoch": 2.6775244299674266, + "grad_norm": 10.912795228418068, + "learning_rate": 9.337561219944756e-06, + "loss": 0.6655, + "step": 4932 + }, + { + "epoch": 2.678067318132465, + "grad_norm": 9.852947274867201, + "learning_rate": 9.334052146434857e-06, + "loss": 0.7696, + "step": 4933 + }, + { + "epoch": 2.6786102062975026, + "grad_norm": 7.323342130507832, + "learning_rate": 9.330543155290454e-06, + "loss": 0.3453, + "step": 4934 + }, + { + "epoch": 2.679153094462541, + "grad_norm": 7.335485317101141, + "learning_rate": 9.32703424694554e-06, + "loss": 0.4291, + "step": 4935 + }, + { + "epoch": 2.6796959826275786, + "grad_norm": 12.700387251668076, + "learning_rate": 9.3235254218341e-06, + "loss": 0.7156, + "step": 4936 + }, + { + "epoch": 2.680238870792617, + "grad_norm": 10.828431765432423, + "learning_rate": 9.320016680390115e-06, + "loss": 0.7293, + "step": 4937 + }, + { + "epoch": 2.6807817589576546, + "grad_norm": 9.224454671275938, + "learning_rate": 9.316508023047548e-06, + "loss": 0.4157, + "step": 4938 + }, + { + "epoch": 2.681324647122693, + "grad_norm": 13.095737517917204, + "learning_rate": 9.312999450240362e-06, + "loss": 0.7928, + "step": 4939 + }, + { + "epoch": 2.6818675352877306, + "grad_norm": 10.338830489426515, + "learning_rate": 9.309490962402498e-06, + "loss": 0.6826, + "step": 4940 + }, + { + "epoch": 2.682410423452769, + "grad_norm": 11.650643384237476, + "learning_rate": 9.305982559967894e-06, + "loss": 0.6709, + "step": 4941 + }, + { + "epoch": 2.6829533116178066, + "grad_norm": 9.153121673948778, + "learning_rate": 9.302474243370469e-06, + "loss": 0.4998, + "step": 4942 + }, + { + "epoch": 2.683496199782845, + "grad_norm": 11.760747525125, + "learning_rate": 9.298966013044144e-06, + "loss": 1.0389, + "step": 4943 + }, + { + "epoch": 2.6840390879478826, + "grad_norm": 13.657927099528628, + "learning_rate": 9.295457869422818e-06, + "loss": 0.9807, + "step": 4944 + }, + { + "epoch": 2.684581976112921, + "grad_norm": 12.726310962899635, + "learning_rate": 9.291949812940387e-06, + "loss": 0.6495, + "step": 4945 + }, + { + "epoch": 2.6851248642779586, + "grad_norm": 11.439120086220523, + "learning_rate": 9.288441844030735e-06, + "loss": 0.6616, + "step": 4946 + }, + { + "epoch": 2.685667752442997, + "grad_norm": 10.384831951074661, + "learning_rate": 9.284933963127724e-06, + "loss": 0.536, + "step": 4947 + }, + { + "epoch": 2.6862106406080346, + "grad_norm": 8.803102808116847, + "learning_rate": 9.281426170665225e-06, + "loss": 0.6621, + "step": 4948 + }, + { + "epoch": 2.686753528773073, + "grad_norm": 14.516361098495818, + "learning_rate": 9.277918467077077e-06, + "loss": 1.0579, + "step": 4949 + }, + { + "epoch": 2.6872964169381106, + "grad_norm": 11.08498885907847, + "learning_rate": 9.274410852797126e-06, + "loss": 0.463, + "step": 4950 + }, + { + "epoch": 2.687839305103149, + "grad_norm": 10.827937520590183, + "learning_rate": 9.270903328259195e-06, + "loss": 0.5015, + "step": 4951 + }, + { + "epoch": 2.6883821932681866, + "grad_norm": 11.572829724801146, + "learning_rate": 9.267395893897107e-06, + "loss": 0.663, + "step": 4952 + }, + { + "epoch": 2.688925081433225, + "grad_norm": 9.3016488903525, + "learning_rate": 9.26388855014466e-06, + "loss": 0.6876, + "step": 4953 + }, + { + "epoch": 2.6894679695982626, + "grad_norm": 17.324535315422423, + "learning_rate": 9.260381297435652e-06, + "loss": 1.1741, + "step": 4954 + }, + { + "epoch": 2.690010857763301, + "grad_norm": 11.465003825461562, + "learning_rate": 9.256874136203864e-06, + "loss": 0.6867, + "step": 4955 + }, + { + "epoch": 2.6905537459283386, + "grad_norm": 8.846935307480056, + "learning_rate": 9.253367066883063e-06, + "loss": 0.4518, + "step": 4956 + }, + { + "epoch": 2.691096634093377, + "grad_norm": 11.007513424878196, + "learning_rate": 9.249860089907018e-06, + "loss": 0.7987, + "step": 4957 + }, + { + "epoch": 2.6916395222584146, + "grad_norm": 12.297659996678936, + "learning_rate": 9.246353205709468e-06, + "loss": 0.588, + "step": 4958 + }, + { + "epoch": 2.692182410423453, + "grad_norm": 10.549862062131211, + "learning_rate": 9.24284641472416e-06, + "loss": 0.737, + "step": 4959 + }, + { + "epoch": 2.6927252985884906, + "grad_norm": 10.712460113916158, + "learning_rate": 9.23933971738481e-06, + "loss": 0.6092, + "step": 4960 + }, + { + "epoch": 2.693268186753529, + "grad_norm": 14.104109743551868, + "learning_rate": 9.235833114125141e-06, + "loss": 1.008, + "step": 4961 + }, + { + "epoch": 2.6938110749185666, + "grad_norm": 7.755916952898445, + "learning_rate": 9.232326605378843e-06, + "loss": 0.4708, + "step": 4962 + }, + { + "epoch": 2.694353963083605, + "grad_norm": 11.677016459785008, + "learning_rate": 9.228820191579618e-06, + "loss": 0.7099, + "step": 4963 + }, + { + "epoch": 2.6948968512486426, + "grad_norm": 10.070884129580872, + "learning_rate": 9.225313873161139e-06, + "loss": 0.6546, + "step": 4964 + }, + { + "epoch": 2.695439739413681, + "grad_norm": 8.108710814711873, + "learning_rate": 9.221807650557078e-06, + "loss": 0.6102, + "step": 4965 + }, + { + "epoch": 2.6959826275787186, + "grad_norm": 10.279519502018594, + "learning_rate": 9.218301524201087e-06, + "loss": 0.5783, + "step": 4966 + }, + { + "epoch": 2.696525515743757, + "grad_norm": 11.882105089319197, + "learning_rate": 9.214795494526806e-06, + "loss": 1.1489, + "step": 4967 + }, + { + "epoch": 2.6970684039087947, + "grad_norm": 9.767493447568013, + "learning_rate": 9.211289561967872e-06, + "loss": 0.6051, + "step": 4968 + }, + { + "epoch": 2.697611292073833, + "grad_norm": 11.690415242614575, + "learning_rate": 9.207783726957903e-06, + "loss": 0.592, + "step": 4969 + }, + { + "epoch": 2.6981541802388707, + "grad_norm": 7.427254360689941, + "learning_rate": 9.204277989930502e-06, + "loss": 0.4672, + "step": 4970 + }, + { + "epoch": 2.698697068403909, + "grad_norm": 9.41305388818184, + "learning_rate": 9.200772351319266e-06, + "loss": 0.515, + "step": 4971 + }, + { + "epoch": 2.6992399565689467, + "grad_norm": 11.02585368985108, + "learning_rate": 9.197266811557787e-06, + "loss": 0.7867, + "step": 4972 + }, + { + "epoch": 2.699782844733985, + "grad_norm": 10.462642998204657, + "learning_rate": 9.193761371079622e-06, + "loss": 0.6689, + "step": 4973 + }, + { + "epoch": 2.7003257328990227, + "grad_norm": 9.964585703249949, + "learning_rate": 9.190256030318339e-06, + "loss": 0.6676, + "step": 4974 + }, + { + "epoch": 2.700868621064061, + "grad_norm": 12.383650152937822, + "learning_rate": 9.186750789707478e-06, + "loss": 0.9124, + "step": 4975 + }, + { + "epoch": 2.7014115092290987, + "grad_norm": 7.428651587888086, + "learning_rate": 9.183245649680574e-06, + "loss": 0.6241, + "step": 4976 + }, + { + "epoch": 2.701954397394137, + "grad_norm": 8.171702461332401, + "learning_rate": 9.179740610671155e-06, + "loss": 0.468, + "step": 4977 + }, + { + "epoch": 2.7024972855591747, + "grad_norm": 15.630545262276021, + "learning_rate": 9.176235673112719e-06, + "loss": 1.4177, + "step": 4978 + }, + { + "epoch": 2.703040173724213, + "grad_norm": 9.04535779717474, + "learning_rate": 9.172730837438774e-06, + "loss": 0.504, + "step": 4979 + }, + { + "epoch": 2.7035830618892507, + "grad_norm": 9.263908210238581, + "learning_rate": 9.169226104082792e-06, + "loss": 0.4043, + "step": 4980 + }, + { + "epoch": 2.704125950054289, + "grad_norm": 10.022418469090024, + "learning_rate": 9.165721473478253e-06, + "loss": 0.6372, + "step": 4981 + }, + { + "epoch": 2.7046688382193267, + "grad_norm": 12.924869609819282, + "learning_rate": 9.16221694605861e-06, + "loss": 0.9433, + "step": 4982 + }, + { + "epoch": 2.705211726384365, + "grad_norm": 8.79512920135937, + "learning_rate": 9.158712522257309e-06, + "loss": 0.4725, + "step": 4983 + }, + { + "epoch": 2.7057546145494027, + "grad_norm": 11.753381756239166, + "learning_rate": 9.155208202507789e-06, + "loss": 0.7395, + "step": 4984 + }, + { + "epoch": 2.706297502714441, + "grad_norm": 6.454949124720575, + "learning_rate": 9.151703987243459e-06, + "loss": 0.3108, + "step": 4985 + }, + { + "epoch": 2.7068403908794787, + "grad_norm": 15.244281036714051, + "learning_rate": 9.148199876897737e-06, + "loss": 0.802, + "step": 4986 + }, + { + "epoch": 2.707383279044517, + "grad_norm": 11.093235618680396, + "learning_rate": 9.144695871904005e-06, + "loss": 0.6252, + "step": 4987 + }, + { + "epoch": 2.7079261672095547, + "grad_norm": 10.703835311171614, + "learning_rate": 9.141191972695655e-06, + "loss": 0.6642, + "step": 4988 + }, + { + "epoch": 2.708469055374593, + "grad_norm": 11.327820449600674, + "learning_rate": 9.137688179706049e-06, + "loss": 1.1613, + "step": 4989 + }, + { + "epoch": 2.7090119435396307, + "grad_norm": 7.527861958807383, + "learning_rate": 9.134184493368548e-06, + "loss": 0.4991, + "step": 4990 + }, + { + "epoch": 2.709554831704669, + "grad_norm": 10.199776895453677, + "learning_rate": 9.130680914116484e-06, + "loss": 0.977, + "step": 4991 + }, + { + "epoch": 2.7100977198697067, + "grad_norm": 9.778998506495286, + "learning_rate": 9.127177442383192e-06, + "loss": 0.6729, + "step": 4992 + }, + { + "epoch": 2.710640608034745, + "grad_norm": 10.831437132716289, + "learning_rate": 9.123674078601984e-06, + "loss": 0.6252, + "step": 4993 + }, + { + "epoch": 2.7111834961997827, + "grad_norm": 9.722012246624114, + "learning_rate": 9.120170823206165e-06, + "loss": 0.6169, + "step": 4994 + }, + { + "epoch": 2.711726384364821, + "grad_norm": 9.696836781330795, + "learning_rate": 9.116667676629019e-06, + "loss": 0.7824, + "step": 4995 + }, + { + "epoch": 2.7122692725298587, + "grad_norm": 8.532030770830737, + "learning_rate": 9.11316463930382e-06, + "loss": 0.7649, + "step": 4996 + }, + { + "epoch": 2.712812160694897, + "grad_norm": 10.155581062168379, + "learning_rate": 9.109661711663837e-06, + "loss": 0.5581, + "step": 4997 + }, + { + "epoch": 2.7133550488599347, + "grad_norm": 11.913092610845165, + "learning_rate": 9.106158894142307e-06, + "loss": 0.6256, + "step": 4998 + }, + { + "epoch": 2.713897937024973, + "grad_norm": 10.326179923874385, + "learning_rate": 9.102656187172475e-06, + "loss": 0.8703, + "step": 4999 + }, + { + "epoch": 2.7144408251900107, + "grad_norm": 10.51344451860866, + "learning_rate": 9.099153591187553e-06, + "loss": 0.8281, + "step": 5000 + }, + { + "epoch": 2.714983713355049, + "grad_norm": 9.101512118782772, + "learning_rate": 9.09565110662075e-06, + "loss": 0.4127, + "step": 5001 + }, + { + "epoch": 2.7155266015200867, + "grad_norm": 14.912595390103256, + "learning_rate": 9.092148733905257e-06, + "loss": 0.828, + "step": 5002 + }, + { + "epoch": 2.716069489685125, + "grad_norm": 10.567445107841705, + "learning_rate": 9.088646473474262e-06, + "loss": 0.8542, + "step": 5003 + }, + { + "epoch": 2.7166123778501627, + "grad_norm": 11.727344925695448, + "learning_rate": 9.085144325760922e-06, + "loss": 0.7593, + "step": 5004 + }, + { + "epoch": 2.717155266015201, + "grad_norm": 10.964390265146204, + "learning_rate": 9.081642291198387e-06, + "loss": 0.696, + "step": 5005 + }, + { + "epoch": 2.7176981541802387, + "grad_norm": 8.594728435704301, + "learning_rate": 9.0781403702198e-06, + "loss": 0.3582, + "step": 5006 + }, + { + "epoch": 2.718241042345277, + "grad_norm": 10.662428429321805, + "learning_rate": 9.074638563258279e-06, + "loss": 0.6213, + "step": 5007 + }, + { + "epoch": 2.7187839305103148, + "grad_norm": 13.853700163509458, + "learning_rate": 9.071136870746934e-06, + "loss": 0.653, + "step": 5008 + }, + { + "epoch": 2.719326818675353, + "grad_norm": 9.421600748659138, + "learning_rate": 9.067635293118862e-06, + "loss": 0.4653, + "step": 5009 + }, + { + "epoch": 2.7198697068403908, + "grad_norm": 15.40532102462592, + "learning_rate": 9.064133830807147e-06, + "loss": 0.8502, + "step": 5010 + }, + { + "epoch": 2.720412595005429, + "grad_norm": 14.246891389640892, + "learning_rate": 9.060632484244845e-06, + "loss": 0.711, + "step": 5011 + }, + { + "epoch": 2.7209554831704668, + "grad_norm": 13.657358501095473, + "learning_rate": 9.057131253865022e-06, + "loss": 1.0586, + "step": 5012 + }, + { + "epoch": 2.721498371335505, + "grad_norm": 11.08278909637767, + "learning_rate": 9.053630140100701e-06, + "loss": 0.5658, + "step": 5013 + }, + { + "epoch": 2.7220412595005428, + "grad_norm": 11.922765678431421, + "learning_rate": 9.050129143384917e-06, + "loss": 0.6929, + "step": 5014 + }, + { + "epoch": 2.722584147665581, + "grad_norm": 10.93390692730677, + "learning_rate": 9.046628264150674e-06, + "loss": 1.417, + "step": 5015 + }, + { + "epoch": 2.7231270358306188, + "grad_norm": 12.86011863530373, + "learning_rate": 9.043127502830964e-06, + "loss": 1.0364, + "step": 5016 + }, + { + "epoch": 2.723669923995657, + "grad_norm": 7.848749292562499, + "learning_rate": 9.039626859858773e-06, + "loss": 0.6696, + "step": 5017 + }, + { + "epoch": 2.7242128121606948, + "grad_norm": 11.212923870670298, + "learning_rate": 9.036126335667059e-06, + "loss": 0.5751, + "step": 5018 + }, + { + "epoch": 2.724755700325733, + "grad_norm": 9.666644723437658, + "learning_rate": 9.032625930688781e-06, + "loss": 0.5699, + "step": 5019 + }, + { + "epoch": 2.725298588490771, + "grad_norm": 8.860936490743164, + "learning_rate": 9.029125645356864e-06, + "loss": 0.4262, + "step": 5020 + }, + { + "epoch": 2.725841476655809, + "grad_norm": 8.597880980516297, + "learning_rate": 9.025625480104238e-06, + "loss": 0.5646, + "step": 5021 + }, + { + "epoch": 2.726384364820847, + "grad_norm": 11.334362425216222, + "learning_rate": 9.022125435363803e-06, + "loss": 0.7424, + "step": 5022 + }, + { + "epoch": 2.726927252985885, + "grad_norm": 10.551777615550284, + "learning_rate": 9.018625511568456e-06, + "loss": 0.4961, + "step": 5023 + }, + { + "epoch": 2.727470141150923, + "grad_norm": 9.061655577456406, + "learning_rate": 9.015125709151069e-06, + "loss": 0.6466, + "step": 5024 + }, + { + "epoch": 2.728013029315961, + "grad_norm": 12.51368822693772, + "learning_rate": 9.011626028544502e-06, + "loss": 0.8461, + "step": 5025 + }, + { + "epoch": 2.728555917480999, + "grad_norm": 9.972383117450178, + "learning_rate": 9.008126470181605e-06, + "loss": 0.7941, + "step": 5026 + }, + { + "epoch": 2.729098805646037, + "grad_norm": 9.372092186313578, + "learning_rate": 9.004627034495204e-06, + "loss": 0.5291, + "step": 5027 + }, + { + "epoch": 2.729641693811075, + "grad_norm": 11.649348346812042, + "learning_rate": 9.00112772191812e-06, + "loss": 0.658, + "step": 5028 + }, + { + "epoch": 2.730184581976113, + "grad_norm": 9.4217070731629, + "learning_rate": 8.997628532883149e-06, + "loss": 0.7391, + "step": 5029 + }, + { + "epoch": 2.730727470141151, + "grad_norm": 9.294886195505088, + "learning_rate": 8.994129467823083e-06, + "loss": 0.7955, + "step": 5030 + }, + { + "epoch": 2.731270358306189, + "grad_norm": 9.046521616766688, + "learning_rate": 8.990630527170684e-06, + "loss": 0.6163, + "step": 5031 + }, + { + "epoch": 2.731813246471227, + "grad_norm": 13.078763070358596, + "learning_rate": 8.987131711358714e-06, + "loss": 0.6352, + "step": 5032 + }, + { + "epoch": 2.732356134636265, + "grad_norm": 8.372819066707777, + "learning_rate": 8.983633020819906e-06, + "loss": 0.5139, + "step": 5033 + }, + { + "epoch": 2.732899022801303, + "grad_norm": 7.908976790945395, + "learning_rate": 8.980134455986982e-06, + "loss": 0.4263, + "step": 5034 + }, + { + "epoch": 2.733441910966341, + "grad_norm": 9.508847688216756, + "learning_rate": 8.97663601729266e-06, + "loss": 0.614, + "step": 5035 + }, + { + "epoch": 2.733984799131379, + "grad_norm": 9.135340041370506, + "learning_rate": 8.973137705169621e-06, + "loss": 0.5803, + "step": 5036 + }, + { + "epoch": 2.734527687296417, + "grad_norm": 10.900299047123902, + "learning_rate": 8.969639520050553e-06, + "loss": 0.4833, + "step": 5037 + }, + { + "epoch": 2.735070575461455, + "grad_norm": 9.243890767443602, + "learning_rate": 8.966141462368106e-06, + "loss": 0.4405, + "step": 5038 + }, + { + "epoch": 2.735613463626493, + "grad_norm": 11.372121051541363, + "learning_rate": 8.962643532554934e-06, + "loss": 0.6618, + "step": 5039 + }, + { + "epoch": 2.736156351791531, + "grad_norm": 11.318259460716959, + "learning_rate": 8.95914573104366e-06, + "loss": 1.0367, + "step": 5040 + }, + { + "epoch": 2.736699239956569, + "grad_norm": 9.055918919260264, + "learning_rate": 8.955648058266904e-06, + "loss": 0.5463, + "step": 5041 + }, + { + "epoch": 2.737242128121607, + "grad_norm": 12.452665937187591, + "learning_rate": 8.952150514657258e-06, + "loss": 0.9772, + "step": 5042 + }, + { + "epoch": 2.737785016286645, + "grad_norm": 11.532813733829784, + "learning_rate": 8.94865310064731e-06, + "loss": 0.6372, + "step": 5043 + }, + { + "epoch": 2.738327904451683, + "grad_norm": 9.800145767151978, + "learning_rate": 8.945155816669622e-06, + "loss": 0.6384, + "step": 5044 + }, + { + "epoch": 2.738870792616721, + "grad_norm": 13.619886494303222, + "learning_rate": 8.94165866315674e-06, + "loss": 0.7295, + "step": 5045 + }, + { + "epoch": 2.739413680781759, + "grad_norm": 13.097331078127143, + "learning_rate": 8.938161640541202e-06, + "loss": 1.1075, + "step": 5046 + }, + { + "epoch": 2.739956568946797, + "grad_norm": 17.458162898454454, + "learning_rate": 8.934664749255524e-06, + "loss": 1.0174, + "step": 5047 + }, + { + "epoch": 2.740499457111835, + "grad_norm": 7.51493556495456, + "learning_rate": 8.931167989732212e-06, + "loss": 0.6407, + "step": 5048 + }, + { + "epoch": 2.741042345276873, + "grad_norm": 11.717405055058808, + "learning_rate": 8.927671362403741e-06, + "loss": 0.6899, + "step": 5049 + }, + { + "epoch": 2.741585233441911, + "grad_norm": 11.433736258508228, + "learning_rate": 8.924174867702591e-06, + "loss": 0.8481, + "step": 5050 + }, + { + "epoch": 2.742128121606949, + "grad_norm": 11.800039466500984, + "learning_rate": 8.920678506061202e-06, + "loss": 0.6044, + "step": 5051 + }, + { + "epoch": 2.742671009771987, + "grad_norm": 8.499003916350551, + "learning_rate": 8.91718227791202e-06, + "loss": 0.475, + "step": 5052 + }, + { + "epoch": 2.743213897937025, + "grad_norm": 10.586652955325345, + "learning_rate": 8.913686183687459e-06, + "loss": 0.5853, + "step": 5053 + }, + { + "epoch": 2.743756786102063, + "grad_norm": 14.066921486210767, + "learning_rate": 8.910190223819919e-06, + "loss": 0.8164, + "step": 5054 + }, + { + "epoch": 2.744299674267101, + "grad_norm": 9.77990149419258, + "learning_rate": 8.906694398741792e-06, + "loss": 0.6053, + "step": 5055 + }, + { + "epoch": 2.744842562432139, + "grad_norm": 15.328290145831401, + "learning_rate": 8.903198708885442e-06, + "loss": 1.3555, + "step": 5056 + }, + { + "epoch": 2.745385450597177, + "grad_norm": 11.626317894758118, + "learning_rate": 8.899703154683228e-06, + "loss": 0.6692, + "step": 5057 + }, + { + "epoch": 2.745928338762215, + "grad_norm": 11.566102899846406, + "learning_rate": 8.896207736567476e-06, + "loss": 0.6589, + "step": 5058 + }, + { + "epoch": 2.746471226927253, + "grad_norm": 9.521589192518187, + "learning_rate": 8.892712454970512e-06, + "loss": 0.7559, + "step": 5059 + }, + { + "epoch": 2.747014115092291, + "grad_norm": 8.719647496284242, + "learning_rate": 8.889217310324636e-06, + "loss": 0.4123, + "step": 5060 + }, + { + "epoch": 2.747557003257329, + "grad_norm": 6.709750992678039, + "learning_rate": 8.885722303062136e-06, + "loss": 0.3492, + "step": 5061 + }, + { + "epoch": 2.748099891422367, + "grad_norm": 11.273080989021693, + "learning_rate": 8.882227433615275e-06, + "loss": 0.7416, + "step": 5062 + }, + { + "epoch": 2.748642779587405, + "grad_norm": 13.864388014548679, + "learning_rate": 8.87873270241631e-06, + "loss": 0.8197, + "step": 5063 + }, + { + "epoch": 2.749185667752443, + "grad_norm": 9.492558149681301, + "learning_rate": 8.875238109897468e-06, + "loss": 0.6517, + "step": 5064 + }, + { + "epoch": 2.749728555917481, + "grad_norm": 10.634433093222732, + "learning_rate": 8.87174365649097e-06, + "loss": 0.5431, + "step": 5065 + }, + { + "epoch": 2.750271444082519, + "grad_norm": 8.469745230774542, + "learning_rate": 8.868249342629015e-06, + "loss": 0.5189, + "step": 5066 + }, + { + "epoch": 2.750814332247557, + "grad_norm": 8.182617911488233, + "learning_rate": 8.864755168743783e-06, + "loss": 0.4322, + "step": 5067 + }, + { + "epoch": 2.751357220412595, + "grad_norm": 7.6335690540689685, + "learning_rate": 8.861261135267444e-06, + "loss": 0.5268, + "step": 5068 + }, + { + "epoch": 2.751900108577633, + "grad_norm": 11.52493444073719, + "learning_rate": 8.85776724263214e-06, + "loss": 0.9757, + "step": 5069 + }, + { + "epoch": 2.752442996742671, + "grad_norm": 9.591556669015672, + "learning_rate": 8.854273491270008e-06, + "loss": 0.5527, + "step": 5070 + }, + { + "epoch": 2.752985884907709, + "grad_norm": 11.178077098634855, + "learning_rate": 8.850779881613151e-06, + "loss": 0.581, + "step": 5071 + }, + { + "epoch": 2.753528773072747, + "grad_norm": 8.921814933147093, + "learning_rate": 8.847286414093673e-06, + "loss": 0.5604, + "step": 5072 + }, + { + "epoch": 2.754071661237785, + "grad_norm": 11.887605361931898, + "learning_rate": 8.84379308914365e-06, + "loss": 0.6055, + "step": 5073 + }, + { + "epoch": 2.754614549402823, + "grad_norm": 14.12418845494581, + "learning_rate": 8.840299907195137e-06, + "loss": 0.9614, + "step": 5074 + }, + { + "epoch": 2.755157437567861, + "grad_norm": 10.094252933928091, + "learning_rate": 8.836806868680185e-06, + "loss": 0.6836, + "step": 5075 + }, + { + "epoch": 2.755700325732899, + "grad_norm": 10.048762466683991, + "learning_rate": 8.833313974030807e-06, + "loss": 0.5722, + "step": 5076 + }, + { + "epoch": 2.756243213897937, + "grad_norm": 12.51909069586841, + "learning_rate": 8.829821223679022e-06, + "loss": 1.2238, + "step": 5077 + }, + { + "epoch": 2.756786102062975, + "grad_norm": 9.50434691504725, + "learning_rate": 8.826328618056808e-06, + "loss": 0.6393, + "step": 5078 + }, + { + "epoch": 2.757328990228013, + "grad_norm": 10.737059961404412, + "learning_rate": 8.822836157596141e-06, + "loss": 0.7401, + "step": 5079 + }, + { + "epoch": 2.757871878393051, + "grad_norm": 9.16480609854369, + "learning_rate": 8.819343842728976e-06, + "loss": 0.5289, + "step": 5080 + }, + { + "epoch": 2.758414766558089, + "grad_norm": 10.849970138344764, + "learning_rate": 8.815851673887248e-06, + "loss": 0.6157, + "step": 5081 + }, + { + "epoch": 2.758957654723127, + "grad_norm": 8.609887529945185, + "learning_rate": 8.812359651502872e-06, + "loss": 0.5961, + "step": 5082 + }, + { + "epoch": 2.759500542888165, + "grad_norm": 10.765964781119916, + "learning_rate": 8.808867776007745e-06, + "loss": 1.1942, + "step": 5083 + }, + { + "epoch": 2.760043431053203, + "grad_norm": 10.466671227569803, + "learning_rate": 8.80537604783375e-06, + "loss": 0.7117, + "step": 5084 + }, + { + "epoch": 2.760586319218241, + "grad_norm": 10.079228354003783, + "learning_rate": 8.801884467412747e-06, + "loss": 0.4238, + "step": 5085 + }, + { + "epoch": 2.761129207383279, + "grad_norm": 11.422859396303908, + "learning_rate": 8.798393035176588e-06, + "loss": 0.8841, + "step": 5086 + }, + { + "epoch": 2.761672095548317, + "grad_norm": 9.736302153055926, + "learning_rate": 8.79490175155709e-06, + "loss": 0.6414, + "step": 5087 + }, + { + "epoch": 2.762214983713355, + "grad_norm": 8.513897139101958, + "learning_rate": 8.791410616986067e-06, + "loss": 0.3052, + "step": 5088 + }, + { + "epoch": 2.762757871878393, + "grad_norm": 10.101572440599828, + "learning_rate": 8.787919631895301e-06, + "loss": 0.5305, + "step": 5089 + }, + { + "epoch": 2.763300760043431, + "grad_norm": 12.17319240146357, + "learning_rate": 8.784428796716571e-06, + "loss": 0.6585, + "step": 5090 + }, + { + "epoch": 2.763843648208469, + "grad_norm": 10.930394499045825, + "learning_rate": 8.78093811188162e-06, + "loss": 0.5984, + "step": 5091 + }, + { + "epoch": 2.764386536373507, + "grad_norm": 9.550475415502788, + "learning_rate": 8.77744757782219e-06, + "loss": 0.5634, + "step": 5092 + }, + { + "epoch": 2.764929424538545, + "grad_norm": 12.890085749205962, + "learning_rate": 8.773957194969993e-06, + "loss": 0.9274, + "step": 5093 + }, + { + "epoch": 2.765472312703583, + "grad_norm": 12.38000962026181, + "learning_rate": 8.77046696375672e-06, + "loss": 0.6988, + "step": 5094 + }, + { + "epoch": 2.766015200868621, + "grad_norm": 9.523692226826551, + "learning_rate": 8.766976884614056e-06, + "loss": 0.4856, + "step": 5095 + }, + { + "epoch": 2.766558089033659, + "grad_norm": 9.26829606100349, + "learning_rate": 8.763486957973652e-06, + "loss": 0.6863, + "step": 5096 + }, + { + "epoch": 2.767100977198697, + "grad_norm": 11.227870751984687, + "learning_rate": 8.759997184267155e-06, + "loss": 0.6344, + "step": 5097 + }, + { + "epoch": 2.767643865363735, + "grad_norm": 11.653609442391305, + "learning_rate": 8.756507563926182e-06, + "loss": 0.5884, + "step": 5098 + }, + { + "epoch": 2.768186753528773, + "grad_norm": 11.59192980688709, + "learning_rate": 8.753018097382336e-06, + "loss": 0.8214, + "step": 5099 + }, + { + "epoch": 2.768729641693811, + "grad_norm": 8.668298021688548, + "learning_rate": 8.749528785067196e-06, + "loss": 0.6487, + "step": 5100 + }, + { + "epoch": 2.769272529858849, + "grad_norm": 5.456725122843202, + "learning_rate": 8.746039627412333e-06, + "loss": 0.2868, + "step": 5101 + }, + { + "epoch": 2.769815418023887, + "grad_norm": 12.29265212176352, + "learning_rate": 8.742550624849288e-06, + "loss": 0.512, + "step": 5102 + }, + { + "epoch": 2.770358306188925, + "grad_norm": 14.125650271042089, + "learning_rate": 8.73906177780958e-06, + "loss": 0.8903, + "step": 5103 + }, + { + "epoch": 2.770901194353963, + "grad_norm": 12.046023026774584, + "learning_rate": 8.735573086724725e-06, + "loss": 0.801, + "step": 5104 + }, + { + "epoch": 2.771444082519001, + "grad_norm": 10.153944591634119, + "learning_rate": 8.732084552026203e-06, + "loss": 0.6888, + "step": 5105 + }, + { + "epoch": 2.771986970684039, + "grad_norm": 10.581737293617769, + "learning_rate": 8.72859617414549e-06, + "loss": 0.7775, + "step": 5106 + }, + { + "epoch": 2.772529858849077, + "grad_norm": 8.304371234513637, + "learning_rate": 8.725107953514021e-06, + "loss": 0.4704, + "step": 5107 + }, + { + "epoch": 2.773072747014115, + "grad_norm": 12.579288268640527, + "learning_rate": 8.72161989056324e-06, + "loss": 0.5861, + "step": 5108 + }, + { + "epoch": 2.773615635179153, + "grad_norm": 10.68813978428288, + "learning_rate": 8.718131985724542e-06, + "loss": 0.6367, + "step": 5109 + }, + { + "epoch": 2.774158523344191, + "grad_norm": 8.921584649335454, + "learning_rate": 8.714644239429326e-06, + "loss": 0.5597, + "step": 5110 + }, + { + "epoch": 2.774701411509229, + "grad_norm": 10.473739328784887, + "learning_rate": 8.711156652108957e-06, + "loss": 0.8587, + "step": 5111 + }, + { + "epoch": 2.775244299674267, + "grad_norm": 10.905504725358476, + "learning_rate": 8.70766922419479e-06, + "loss": 0.5729, + "step": 5112 + }, + { + "epoch": 2.7757871878393052, + "grad_norm": 7.88310879226585, + "learning_rate": 8.704181956118153e-06, + "loss": 0.4998, + "step": 5113 + }, + { + "epoch": 2.776330076004343, + "grad_norm": 8.595679686197345, + "learning_rate": 8.700694848310354e-06, + "loss": 0.699, + "step": 5114 + }, + { + "epoch": 2.7768729641693812, + "grad_norm": 9.784504057993846, + "learning_rate": 8.697207901202691e-06, + "loss": 0.6509, + "step": 5115 + }, + { + "epoch": 2.777415852334419, + "grad_norm": 9.537890155200703, + "learning_rate": 8.693721115226427e-06, + "loss": 0.4484, + "step": 5116 + }, + { + "epoch": 2.7779587404994572, + "grad_norm": 10.110968424420188, + "learning_rate": 8.69023449081282e-06, + "loss": 0.6746, + "step": 5117 + }, + { + "epoch": 2.778501628664495, + "grad_norm": 14.837427287717714, + "learning_rate": 8.686748028393096e-06, + "loss": 1.3601, + "step": 5118 + }, + { + "epoch": 2.7790445168295332, + "grad_norm": 11.553555350476039, + "learning_rate": 8.683261728398472e-06, + "loss": 0.6054, + "step": 5119 + }, + { + "epoch": 2.779587404994571, + "grad_norm": 12.684609984344696, + "learning_rate": 8.679775591260132e-06, + "loss": 0.947, + "step": 5120 + }, + { + "epoch": 2.7801302931596092, + "grad_norm": 9.50144977278699, + "learning_rate": 8.676289617409256e-06, + "loss": 0.6401, + "step": 5121 + }, + { + "epoch": 2.780673181324647, + "grad_norm": 9.588585115620386, + "learning_rate": 8.672803807276988e-06, + "loss": 0.7214, + "step": 5122 + }, + { + "epoch": 2.7812160694896852, + "grad_norm": 14.597044499431716, + "learning_rate": 8.66931816129446e-06, + "loss": 1.0037, + "step": 5123 + }, + { + "epoch": 2.781758957654723, + "grad_norm": 10.552996924508232, + "learning_rate": 8.665832679892783e-06, + "loss": 0.5097, + "step": 5124 + }, + { + "epoch": 2.7823018458197613, + "grad_norm": 8.368460915764464, + "learning_rate": 8.662347363503043e-06, + "loss": 0.5289, + "step": 5125 + }, + { + "epoch": 2.782844733984799, + "grad_norm": 7.412774968306893, + "learning_rate": 8.658862212556318e-06, + "loss": 0.4117, + "step": 5126 + }, + { + "epoch": 2.7833876221498373, + "grad_norm": 7.965326482678092, + "learning_rate": 8.655377227483648e-06, + "loss": 0.4492, + "step": 5127 + }, + { + "epoch": 2.783930510314875, + "grad_norm": 10.16388699920939, + "learning_rate": 8.65189240871607e-06, + "loss": 0.6847, + "step": 5128 + }, + { + "epoch": 2.7844733984799133, + "grad_norm": 8.606167058304848, + "learning_rate": 8.648407756684582e-06, + "loss": 0.5952, + "step": 5129 + }, + { + "epoch": 2.785016286644951, + "grad_norm": 16.346182786192927, + "learning_rate": 8.64492327182018e-06, + "loss": 1.0626, + "step": 5130 + }, + { + "epoch": 2.7855591748099893, + "grad_norm": 9.397161435588243, + "learning_rate": 8.64143895455383e-06, + "loss": 0.6416, + "step": 5131 + }, + { + "epoch": 2.786102062975027, + "grad_norm": 10.215541582389136, + "learning_rate": 8.63795480531647e-06, + "loss": 0.5311, + "step": 5132 + }, + { + "epoch": 2.7866449511400653, + "grad_norm": 13.362954088182768, + "learning_rate": 8.634470824539035e-06, + "loss": 1.0249, + "step": 5133 + }, + { + "epoch": 2.787187839305103, + "grad_norm": 13.843076908870096, + "learning_rate": 8.630987012652421e-06, + "loss": 0.767, + "step": 5134 + }, + { + "epoch": 2.7877307274701413, + "grad_norm": 10.581731921492887, + "learning_rate": 8.627503370087519e-06, + "loss": 0.6027, + "step": 5135 + }, + { + "epoch": 2.788273615635179, + "grad_norm": 11.161057087658808, + "learning_rate": 8.624019897275184e-06, + "loss": 0.8193, + "step": 5136 + }, + { + "epoch": 2.7888165038002173, + "grad_norm": 7.972214509867566, + "learning_rate": 8.620536594646262e-06, + "loss": 0.4776, + "step": 5137 + }, + { + "epoch": 2.789359391965255, + "grad_norm": 12.57635432127455, + "learning_rate": 8.61705346263157e-06, + "loss": 0.7194, + "step": 5138 + }, + { + "epoch": 2.7899022801302933, + "grad_norm": 7.077115343788877, + "learning_rate": 8.613570501661915e-06, + "loss": 0.3571, + "step": 5139 + }, + { + "epoch": 2.790445168295331, + "grad_norm": 13.158625298591346, + "learning_rate": 8.610087712168065e-06, + "loss": 0.5872, + "step": 5140 + }, + { + "epoch": 2.7909880564603693, + "grad_norm": 8.652483474203775, + "learning_rate": 8.606605094580788e-06, + "loss": 0.624, + "step": 5141 + }, + { + "epoch": 2.791530944625407, + "grad_norm": 8.73276991998644, + "learning_rate": 8.603122649330811e-06, + "loss": 0.4358, + "step": 5142 + }, + { + "epoch": 2.7920738327904453, + "grad_norm": 12.749713905257105, + "learning_rate": 8.599640376848849e-06, + "loss": 0.7574, + "step": 5143 + }, + { + "epoch": 2.792616720955483, + "grad_norm": 12.054416741196698, + "learning_rate": 8.5961582775656e-06, + "loss": 0.886, + "step": 5144 + }, + { + "epoch": 2.7931596091205213, + "grad_norm": 10.958538797565106, + "learning_rate": 8.592676351911728e-06, + "loss": 0.7312, + "step": 5145 + }, + { + "epoch": 2.793702497285559, + "grad_norm": 10.124234819891349, + "learning_rate": 8.589194600317894e-06, + "loss": 0.5059, + "step": 5146 + }, + { + "epoch": 2.7942453854505973, + "grad_norm": 12.84934263063866, + "learning_rate": 8.585713023214715e-06, + "loss": 0.7016, + "step": 5147 + }, + { + "epoch": 2.794788273615635, + "grad_norm": 11.859965090765057, + "learning_rate": 8.582231621032807e-06, + "loss": 0.7158, + "step": 5148 + }, + { + "epoch": 2.7953311617806733, + "grad_norm": 11.57843340583913, + "learning_rate": 8.578750394202749e-06, + "loss": 0.7103, + "step": 5149 + }, + { + "epoch": 2.795874049945711, + "grad_norm": 7.143782607145388, + "learning_rate": 8.575269343155108e-06, + "loss": 0.5275, + "step": 5150 + }, + { + "epoch": 2.7964169381107493, + "grad_norm": 12.095108158973952, + "learning_rate": 8.571788468320427e-06, + "loss": 0.7308, + "step": 5151 + }, + { + "epoch": 2.796959826275787, + "grad_norm": 11.153004271537196, + "learning_rate": 8.568307770129223e-06, + "loss": 0.7576, + "step": 5152 + }, + { + "epoch": 2.7975027144408253, + "grad_norm": 8.326263454089252, + "learning_rate": 8.564827249011998e-06, + "loss": 0.408, + "step": 5153 + }, + { + "epoch": 2.798045602605863, + "grad_norm": 13.46902871500607, + "learning_rate": 8.561346905399221e-06, + "loss": 0.6624, + "step": 5154 + }, + { + "epoch": 2.7985884907709013, + "grad_norm": 8.925567749596054, + "learning_rate": 8.557866739721356e-06, + "loss": 0.4557, + "step": 5155 + }, + { + "epoch": 2.799131378935939, + "grad_norm": 11.700305884529763, + "learning_rate": 8.554386752408827e-06, + "loss": 0.6818, + "step": 5156 + }, + { + "epoch": 2.7996742671009773, + "grad_norm": 9.49767460620272, + "learning_rate": 8.550906943892054e-06, + "loss": 0.7256, + "step": 5157 + }, + { + "epoch": 2.800217155266015, + "grad_norm": 11.749650388906772, + "learning_rate": 8.547427314601416e-06, + "loss": 0.6933, + "step": 5158 + }, + { + "epoch": 2.8007600434310533, + "grad_norm": 11.830065328109, + "learning_rate": 8.543947864967286e-06, + "loss": 0.4282, + "step": 5159 + }, + { + "epoch": 2.801302931596091, + "grad_norm": 12.11100989733807, + "learning_rate": 8.540468595419999e-06, + "loss": 0.3617, + "step": 5160 + }, + { + "epoch": 2.8018458197611293, + "grad_norm": 9.923764364092053, + "learning_rate": 8.536989506389889e-06, + "loss": 0.6469, + "step": 5161 + }, + { + "epoch": 2.802388707926167, + "grad_norm": 10.555923377692814, + "learning_rate": 8.533510598307244e-06, + "loss": 0.9672, + "step": 5162 + }, + { + "epoch": 2.8029315960912053, + "grad_norm": 12.544866122081839, + "learning_rate": 8.530031871602345e-06, + "loss": 0.6885, + "step": 5163 + }, + { + "epoch": 2.803474484256243, + "grad_norm": 11.680102099467787, + "learning_rate": 8.526553326705452e-06, + "loss": 0.5571, + "step": 5164 + }, + { + "epoch": 2.8040173724212814, + "grad_norm": 11.769615311559438, + "learning_rate": 8.523074964046785e-06, + "loss": 0.6176, + "step": 5165 + }, + { + "epoch": 2.804560260586319, + "grad_norm": 10.249168095085986, + "learning_rate": 8.519596784056567e-06, + "loss": 0.7974, + "step": 5166 + }, + { + "epoch": 2.8051031487513574, + "grad_norm": 7.602995989085923, + "learning_rate": 8.516118787164973e-06, + "loss": 0.4132, + "step": 5167 + }, + { + "epoch": 2.805646036916395, + "grad_norm": 9.32078805379704, + "learning_rate": 8.512640973802175e-06, + "loss": 0.6492, + "step": 5168 + }, + { + "epoch": 2.8061889250814334, + "grad_norm": 13.80437475122657, + "learning_rate": 8.50916334439831e-06, + "loss": 0.6663, + "step": 5169 + }, + { + "epoch": 2.806731813246471, + "grad_norm": 9.121579447999185, + "learning_rate": 8.5056858993835e-06, + "loss": 0.4546, + "step": 5170 + }, + { + "epoch": 2.8072747014115094, + "grad_norm": 8.073186873408805, + "learning_rate": 8.502208639187842e-06, + "loss": 0.4467, + "step": 5171 + }, + { + "epoch": 2.807817589576547, + "grad_norm": 10.498644561860338, + "learning_rate": 8.498731564241403e-06, + "loss": 0.4907, + "step": 5172 + }, + { + "epoch": 2.8083604777415854, + "grad_norm": 9.027361205852865, + "learning_rate": 8.495254674974239e-06, + "loss": 0.4581, + "step": 5173 + }, + { + "epoch": 2.808903365906623, + "grad_norm": 9.692775527105594, + "learning_rate": 8.491777971816372e-06, + "loss": 0.5898, + "step": 5174 + }, + { + "epoch": 2.8094462540716614, + "grad_norm": 9.171182553280271, + "learning_rate": 8.48830145519781e-06, + "loss": 0.4129, + "step": 5175 + }, + { + "epoch": 2.809989142236699, + "grad_norm": 11.635096708230561, + "learning_rate": 8.484825125548532e-06, + "loss": 0.7632, + "step": 5176 + }, + { + "epoch": 2.8105320304017374, + "grad_norm": 12.18174333972076, + "learning_rate": 8.481348983298503e-06, + "loss": 0.6158, + "step": 5177 + }, + { + "epoch": 2.811074918566775, + "grad_norm": 8.880576943532684, + "learning_rate": 8.477873028877645e-06, + "loss": 0.374, + "step": 5178 + }, + { + "epoch": 2.8116178067318134, + "grad_norm": 11.735757011642823, + "learning_rate": 8.474397262715884e-06, + "loss": 0.6772, + "step": 5179 + }, + { + "epoch": 2.812160694896851, + "grad_norm": 9.50535973508058, + "learning_rate": 8.470921685243098e-06, + "loss": 0.6021, + "step": 5180 + }, + { + "epoch": 2.8127035830618894, + "grad_norm": 11.780091573431722, + "learning_rate": 8.467446296889151e-06, + "loss": 0.6944, + "step": 5181 + }, + { + "epoch": 2.813246471226927, + "grad_norm": 10.177875512109438, + "learning_rate": 8.463971098083896e-06, + "loss": 0.4312, + "step": 5182 + }, + { + "epoch": 2.8137893593919654, + "grad_norm": 16.588608629478436, + "learning_rate": 8.460496089257136e-06, + "loss": 0.818, + "step": 5183 + }, + { + "epoch": 2.814332247557003, + "grad_norm": 12.416295723031903, + "learning_rate": 8.45702127083868e-06, + "loss": 0.5531, + "step": 5184 + }, + { + "epoch": 2.8148751357220414, + "grad_norm": 13.657036134609482, + "learning_rate": 8.453546643258286e-06, + "loss": 0.7343, + "step": 5185 + }, + { + "epoch": 2.815418023887079, + "grad_norm": 9.833034529955514, + "learning_rate": 8.450072206945715e-06, + "loss": 0.6592, + "step": 5186 + }, + { + "epoch": 2.8159609120521174, + "grad_norm": 10.036112921946827, + "learning_rate": 8.446597962330675e-06, + "loss": 0.8127, + "step": 5187 + }, + { + "epoch": 2.816503800217155, + "grad_norm": 12.901123274951528, + "learning_rate": 8.443123909842882e-06, + "loss": 1.0507, + "step": 5188 + }, + { + "epoch": 2.8170466883821934, + "grad_norm": 10.692152694001132, + "learning_rate": 8.439650049911999e-06, + "loss": 0.5602, + "step": 5189 + }, + { + "epoch": 2.817589576547231, + "grad_norm": 19.731345695292617, + "learning_rate": 8.436176382967692e-06, + "loss": 1.0473, + "step": 5190 + }, + { + "epoch": 2.8181324647122694, + "grad_norm": 14.567179107047348, + "learning_rate": 8.432702909439579e-06, + "loss": 0.6359, + "step": 5191 + }, + { + "epoch": 2.818675352877307, + "grad_norm": 13.853200529175792, + "learning_rate": 8.429229629757266e-06, + "loss": 0.8277, + "step": 5192 + }, + { + "epoch": 2.8192182410423454, + "grad_norm": 9.811701731637578, + "learning_rate": 8.425756544350338e-06, + "loss": 0.533, + "step": 5193 + }, + { + "epoch": 2.819761129207383, + "grad_norm": 11.486229326022617, + "learning_rate": 8.422283653648348e-06, + "loss": 0.9078, + "step": 5194 + }, + { + "epoch": 2.8203040173724214, + "grad_norm": 11.086905618662822, + "learning_rate": 8.418810958080832e-06, + "loss": 0.7101, + "step": 5195 + }, + { + "epoch": 2.820846905537459, + "grad_norm": 8.065915758013317, + "learning_rate": 8.415338458077293e-06, + "loss": 0.5744, + "step": 5196 + }, + { + "epoch": 2.8213897937024974, + "grad_norm": 9.063098353131359, + "learning_rate": 8.411866154067224e-06, + "loss": 0.5008, + "step": 5197 + }, + { + "epoch": 2.821932681867535, + "grad_norm": 12.457981789036095, + "learning_rate": 8.408394046480077e-06, + "loss": 0.6377, + "step": 5198 + }, + { + "epoch": 2.8224755700325734, + "grad_norm": 10.806638216082748, + "learning_rate": 8.404922135745295e-06, + "loss": 0.6137, + "step": 5199 + }, + { + "epoch": 2.823018458197611, + "grad_norm": 10.78339089292669, + "learning_rate": 8.401450422292281e-06, + "loss": 0.7984, + "step": 5200 + }, + { + "epoch": 2.8235613463626494, + "grad_norm": 9.574355357184858, + "learning_rate": 8.397978906550429e-06, + "loss": 0.6307, + "step": 5201 + }, + { + "epoch": 2.824104234527687, + "grad_norm": 9.457238410125305, + "learning_rate": 8.3945075889491e-06, + "loss": 0.6782, + "step": 5202 + }, + { + "epoch": 2.8246471226927254, + "grad_norm": 15.132207636686802, + "learning_rate": 8.391036469917626e-06, + "loss": 0.7843, + "step": 5203 + }, + { + "epoch": 2.8251900108577632, + "grad_norm": 12.425847301998207, + "learning_rate": 8.387565549885331e-06, + "loss": 0.9953, + "step": 5204 + }, + { + "epoch": 2.8257328990228014, + "grad_norm": 9.948899666849417, + "learning_rate": 8.384094829281495e-06, + "loss": 0.6601, + "step": 5205 + }, + { + "epoch": 2.8262757871878392, + "grad_norm": 9.412819573011207, + "learning_rate": 8.38062430853539e-06, + "loss": 0.3747, + "step": 5206 + }, + { + "epoch": 2.8268186753528775, + "grad_norm": 8.137925073552552, + "learning_rate": 8.377153988076247e-06, + "loss": 0.3818, + "step": 5207 + }, + { + "epoch": 2.8273615635179152, + "grad_norm": 8.994244424095461, + "learning_rate": 8.373683868333287e-06, + "loss": 0.796, + "step": 5208 + }, + { + "epoch": 2.8279044516829535, + "grad_norm": 7.570203929319876, + "learning_rate": 8.370213949735696e-06, + "loss": 0.5587, + "step": 5209 + }, + { + "epoch": 2.8284473398479912, + "grad_norm": 13.134200586671845, + "learning_rate": 8.366744232712645e-06, + "loss": 0.8274, + "step": 5210 + }, + { + "epoch": 2.8289902280130295, + "grad_norm": 9.833443120274195, + "learning_rate": 8.363274717693272e-06, + "loss": 0.7984, + "step": 5211 + }, + { + "epoch": 2.8295331161780672, + "grad_norm": 11.623259569780537, + "learning_rate": 8.359805405106685e-06, + "loss": 0.7305, + "step": 5212 + }, + { + "epoch": 2.8300760043431055, + "grad_norm": 10.494299983717905, + "learning_rate": 8.356336295381981e-06, + "loss": 0.6299, + "step": 5213 + }, + { + "epoch": 2.8306188925081432, + "grad_norm": 7.662506144803812, + "learning_rate": 8.352867388948223e-06, + "loss": 0.4432, + "step": 5214 + }, + { + "epoch": 2.8311617806731815, + "grad_norm": 11.250218590721435, + "learning_rate": 8.349398686234455e-06, + "loss": 0.5805, + "step": 5215 + }, + { + "epoch": 2.8317046688382193, + "grad_norm": 12.63080713053447, + "learning_rate": 8.345930187669685e-06, + "loss": 0.7236, + "step": 5216 + }, + { + "epoch": 2.8322475570032575, + "grad_norm": 12.902831533172366, + "learning_rate": 8.342461893682908e-06, + "loss": 0.6462, + "step": 5217 + }, + { + "epoch": 2.8327904451682953, + "grad_norm": 8.440334577273788, + "learning_rate": 8.338993804703082e-06, + "loss": 0.6557, + "step": 5218 + }, + { + "epoch": 2.8333333333333335, + "grad_norm": 8.946600771683327, + "learning_rate": 8.335525921159155e-06, + "loss": 0.4964, + "step": 5219 + }, + { + "epoch": 2.8338762214983713, + "grad_norm": 11.427720200073038, + "learning_rate": 8.332058243480032e-06, + "loss": 0.5832, + "step": 5220 + }, + { + "epoch": 2.8344191096634095, + "grad_norm": 10.663320470599537, + "learning_rate": 8.328590772094602e-06, + "loss": 0.385, + "step": 5221 + }, + { + "epoch": 2.8349619978284473, + "grad_norm": 10.091874752226794, + "learning_rate": 8.325123507431732e-06, + "loss": 0.8347, + "step": 5222 + }, + { + "epoch": 2.8355048859934855, + "grad_norm": 11.844959588854952, + "learning_rate": 8.321656449920252e-06, + "loss": 0.5577, + "step": 5223 + }, + { + "epoch": 2.8360477741585233, + "grad_norm": 10.895554793526724, + "learning_rate": 8.318189599988982e-06, + "loss": 0.7961, + "step": 5224 + }, + { + "epoch": 2.8365906623235615, + "grad_norm": 15.457894886801157, + "learning_rate": 8.314722958066699e-06, + "loss": 0.9784, + "step": 5225 + }, + { + "epoch": 2.8371335504885993, + "grad_norm": 9.287014113188532, + "learning_rate": 8.311256524582167e-06, + "loss": 0.5947, + "step": 5226 + }, + { + "epoch": 2.8376764386536375, + "grad_norm": 12.953808801302275, + "learning_rate": 8.307790299964119e-06, + "loss": 0.6073, + "step": 5227 + }, + { + "epoch": 2.8382193268186753, + "grad_norm": 12.53202947557571, + "learning_rate": 8.304324284641268e-06, + "loss": 0.8199, + "step": 5228 + }, + { + "epoch": 2.8387622149837135, + "grad_norm": 10.62038819495681, + "learning_rate": 8.300858479042291e-06, + "loss": 0.488, + "step": 5229 + }, + { + "epoch": 2.8393051031487513, + "grad_norm": 10.914681492949684, + "learning_rate": 8.29739288359584e-06, + "loss": 0.6159, + "step": 5230 + }, + { + "epoch": 2.8398479913137895, + "grad_norm": 9.514892864934003, + "learning_rate": 8.293927498730556e-06, + "loss": 0.8757, + "step": 5231 + }, + { + "epoch": 2.8403908794788273, + "grad_norm": 10.004218198768447, + "learning_rate": 8.290462324875036e-06, + "loss": 0.4543, + "step": 5232 + }, + { + "epoch": 2.8409337676438655, + "grad_norm": 7.955044675492623, + "learning_rate": 8.286997362457859e-06, + "loss": 0.3227, + "step": 5233 + }, + { + "epoch": 2.8414766558089033, + "grad_norm": 5.674531662978047, + "learning_rate": 8.283532611907577e-06, + "loss": 0.2674, + "step": 5234 + }, + { + "epoch": 2.8420195439739415, + "grad_norm": 10.825015665199876, + "learning_rate": 8.280068073652723e-06, + "loss": 0.7373, + "step": 5235 + }, + { + "epoch": 2.8425624321389793, + "grad_norm": 11.937523350465815, + "learning_rate": 8.276603748121786e-06, + "loss": 0.6019, + "step": 5236 + }, + { + "epoch": 2.8431053203040175, + "grad_norm": 11.660316243586271, + "learning_rate": 8.273139635743249e-06, + "loss": 0.8063, + "step": 5237 + }, + { + "epoch": 2.8436482084690553, + "grad_norm": 10.978384303044228, + "learning_rate": 8.269675736945547e-06, + "loss": 0.6178, + "step": 5238 + }, + { + "epoch": 2.8441910966340935, + "grad_norm": 10.489917172678325, + "learning_rate": 8.266212052157113e-06, + "loss": 0.4382, + "step": 5239 + }, + { + "epoch": 2.8447339847991313, + "grad_norm": 9.430697797795169, + "learning_rate": 8.262748581806336e-06, + "loss": 0.4811, + "step": 5240 + }, + { + "epoch": 2.8452768729641695, + "grad_norm": 13.552717272826865, + "learning_rate": 8.259285326321579e-06, + "loss": 0.6866, + "step": 5241 + }, + { + "epoch": 2.8458197611292073, + "grad_norm": 12.883094545932988, + "learning_rate": 8.25582228613119e-06, + "loss": 0.9637, + "step": 5242 + }, + { + "epoch": 2.8463626492942455, + "grad_norm": 18.04770906924292, + "learning_rate": 8.25235946166348e-06, + "loss": 0.801, + "step": 5243 + }, + { + "epoch": 2.8469055374592833, + "grad_norm": 12.482559439953985, + "learning_rate": 8.248896853346738e-06, + "loss": 0.7968, + "step": 5244 + }, + { + "epoch": 2.8474484256243215, + "grad_norm": 10.728211626455765, + "learning_rate": 8.245434461609221e-06, + "loss": 0.5192, + "step": 5245 + }, + { + "epoch": 2.8479913137893593, + "grad_norm": 10.793312471383818, + "learning_rate": 8.241972286879168e-06, + "loss": 0.6121, + "step": 5246 + }, + { + "epoch": 2.8485342019543975, + "grad_norm": 12.822804552881774, + "learning_rate": 8.238510329584782e-06, + "loss": 0.7807, + "step": 5247 + }, + { + "epoch": 2.8490770901194353, + "grad_norm": 14.208687960885756, + "learning_rate": 8.23504859015425e-06, + "loss": 0.73, + "step": 5248 + }, + { + "epoch": 2.8496199782844736, + "grad_norm": 13.561299495458854, + "learning_rate": 8.231587069015723e-06, + "loss": 1.269, + "step": 5249 + }, + { + "epoch": 2.8501628664495113, + "grad_norm": 12.542912517352129, + "learning_rate": 8.22812576659732e-06, + "loss": 0.5758, + "step": 5250 + }, + { + "epoch": 2.8507057546145496, + "grad_norm": 9.184884000564415, + "learning_rate": 8.22466468332715e-06, + "loss": 0.5323, + "step": 5251 + }, + { + "epoch": 2.8512486427795873, + "grad_norm": 10.234457404263107, + "learning_rate": 8.22120381963328e-06, + "loss": 0.5455, + "step": 5252 + }, + { + "epoch": 2.8517915309446256, + "grad_norm": 13.327787107696029, + "learning_rate": 8.217743175943756e-06, + "loss": 0.6097, + "step": 5253 + }, + { + "epoch": 2.8523344191096633, + "grad_norm": 10.375306748712186, + "learning_rate": 8.214282752686595e-06, + "loss": 0.9622, + "step": 5254 + }, + { + "epoch": 2.8528773072747016, + "grad_norm": 10.445896680603816, + "learning_rate": 8.210822550289794e-06, + "loss": 0.6347, + "step": 5255 + }, + { + "epoch": 2.8534201954397393, + "grad_norm": 8.375912125960554, + "learning_rate": 8.207362569181305e-06, + "loss": 0.419, + "step": 5256 + }, + { + "epoch": 2.8539630836047776, + "grad_norm": 12.786819256474223, + "learning_rate": 8.203902809789078e-06, + "loss": 1.0155, + "step": 5257 + }, + { + "epoch": 2.8545059717698154, + "grad_norm": 9.916362057542294, + "learning_rate": 8.200443272541007e-06, + "loss": 0.5501, + "step": 5258 + }, + { + "epoch": 2.8550488599348536, + "grad_norm": 12.303081851739867, + "learning_rate": 8.196983957864984e-06, + "loss": 0.5965, + "step": 5259 + }, + { + "epoch": 2.8555917480998914, + "grad_norm": 11.992761200744544, + "learning_rate": 8.19352486618886e-06, + "loss": 0.4804, + "step": 5260 + }, + { + "epoch": 2.8561346362649296, + "grad_norm": 11.342631954988665, + "learning_rate": 8.190065997940455e-06, + "loss": 0.6241, + "step": 5261 + }, + { + "epoch": 2.8566775244299674, + "grad_norm": 14.06120317718869, + "learning_rate": 8.186607353547578e-06, + "loss": 1.3883, + "step": 5262 + }, + { + "epoch": 2.8572204125950056, + "grad_norm": 9.985876763230308, + "learning_rate": 8.183148933437988e-06, + "loss": 0.6843, + "step": 5263 + }, + { + "epoch": 2.8577633007600434, + "grad_norm": 13.658693433537799, + "learning_rate": 8.179690738039436e-06, + "loss": 1.027, + "step": 5264 + }, + { + "epoch": 2.8583061889250816, + "grad_norm": 13.070549937452688, + "learning_rate": 8.176232767779634e-06, + "loss": 0.9171, + "step": 5265 + }, + { + "epoch": 2.8588490770901194, + "grad_norm": 10.202999590557772, + "learning_rate": 8.17277502308627e-06, + "loss": 0.6412, + "step": 5266 + }, + { + "epoch": 2.8593919652551576, + "grad_norm": 11.358682840396053, + "learning_rate": 8.169317504387002e-06, + "loss": 0.8102, + "step": 5267 + }, + { + "epoch": 2.8599348534201954, + "grad_norm": 12.795930497437475, + "learning_rate": 8.165860212109467e-06, + "loss": 0.7885, + "step": 5268 + }, + { + "epoch": 2.8604777415852336, + "grad_norm": 12.540977748153804, + "learning_rate": 8.162403146681263e-06, + "loss": 0.453, + "step": 5269 + }, + { + "epoch": 2.8610206297502714, + "grad_norm": 9.835742357069565, + "learning_rate": 8.158946308529964e-06, + "loss": 0.553, + "step": 5270 + }, + { + "epoch": 2.8615635179153096, + "grad_norm": 8.768245847600197, + "learning_rate": 8.15548969808312e-06, + "loss": 0.6084, + "step": 5271 + }, + { + "epoch": 2.8621064060803474, + "grad_norm": 11.27517155767251, + "learning_rate": 8.152033315768248e-06, + "loss": 0.708, + "step": 5272 + }, + { + "epoch": 2.8626492942453856, + "grad_norm": 10.175542619256326, + "learning_rate": 8.148577162012848e-06, + "loss": 0.567, + "step": 5273 + }, + { + "epoch": 2.8631921824104234, + "grad_norm": 13.49575371946775, + "learning_rate": 8.145121237244367e-06, + "loss": 0.7178, + "step": 5274 + }, + { + "epoch": 2.8637350705754616, + "grad_norm": 12.848120779354371, + "learning_rate": 8.141665541890254e-06, + "loss": 0.7601, + "step": 5275 + }, + { + "epoch": 2.8642779587404994, + "grad_norm": 9.660459046348308, + "learning_rate": 8.138210076377903e-06, + "loss": 0.5867, + "step": 5276 + }, + { + "epoch": 2.8648208469055376, + "grad_norm": 12.014506663974018, + "learning_rate": 8.134754841134704e-06, + "loss": 0.6157, + "step": 5277 + }, + { + "epoch": 2.8653637350705754, + "grad_norm": 9.006817793845983, + "learning_rate": 8.131299836587994e-06, + "loss": 0.417, + "step": 5278 + }, + { + "epoch": 2.8659066232356136, + "grad_norm": 14.075391102924492, + "learning_rate": 8.127845063165097e-06, + "loss": 0.9453, + "step": 5279 + }, + { + "epoch": 2.8664495114006514, + "grad_norm": 11.384071845529691, + "learning_rate": 8.124390521293311e-06, + "loss": 0.757, + "step": 5280 + }, + { + "epoch": 2.8669923995656896, + "grad_norm": 11.412348591705651, + "learning_rate": 8.12093621139989e-06, + "loss": 0.7149, + "step": 5281 + }, + { + "epoch": 2.8675352877307274, + "grad_norm": 10.533279510688539, + "learning_rate": 8.117482133912078e-06, + "loss": 0.4981, + "step": 5282 + }, + { + "epoch": 2.8680781758957656, + "grad_norm": 10.800216526898991, + "learning_rate": 8.114028289257071e-06, + "loss": 0.6033, + "step": 5283 + }, + { + "epoch": 2.8686210640608034, + "grad_norm": 10.834462944873792, + "learning_rate": 8.110574677862054e-06, + "loss": 0.5554, + "step": 5284 + }, + { + "epoch": 2.8691639522258416, + "grad_norm": 13.02642177423764, + "learning_rate": 8.10712130015417e-06, + "loss": 0.7303, + "step": 5285 + }, + { + "epoch": 2.8697068403908794, + "grad_norm": 13.210484866681897, + "learning_rate": 8.103668156560543e-06, + "loss": 0.9579, + "step": 5286 + }, + { + "epoch": 2.8702497285559176, + "grad_norm": 7.866855822388311, + "learning_rate": 8.100215247508259e-06, + "loss": 0.5878, + "step": 5287 + }, + { + "epoch": 2.8707926167209554, + "grad_norm": 10.857793278838274, + "learning_rate": 8.096762573424384e-06, + "loss": 0.6734, + "step": 5288 + }, + { + "epoch": 2.8713355048859937, + "grad_norm": 11.519175454268337, + "learning_rate": 8.093310134735945e-06, + "loss": 0.7203, + "step": 5289 + }, + { + "epoch": 2.8718783930510314, + "grad_norm": 12.16192132511212, + "learning_rate": 8.089857931869947e-06, + "loss": 0.5887, + "step": 5290 + }, + { + "epoch": 2.8724212812160697, + "grad_norm": 8.545914428475232, + "learning_rate": 8.086405965253364e-06, + "loss": 0.677, + "step": 5291 + }, + { + "epoch": 2.8729641693811074, + "grad_norm": 12.463412130388328, + "learning_rate": 8.08295423531314e-06, + "loss": 0.8954, + "step": 5292 + }, + { + "epoch": 2.8735070575461457, + "grad_norm": 10.296132022290973, + "learning_rate": 8.079502742476195e-06, + "loss": 0.6582, + "step": 5293 + }, + { + "epoch": 2.8740499457111834, + "grad_norm": 9.170071716678386, + "learning_rate": 8.076051487169407e-06, + "loss": 0.5406, + "step": 5294 + }, + { + "epoch": 2.8745928338762217, + "grad_norm": 14.43641149914591, + "learning_rate": 8.072600469819643e-06, + "loss": 0.7459, + "step": 5295 + }, + { + "epoch": 2.8751357220412594, + "grad_norm": 9.705510058137687, + "learning_rate": 8.069149690853719e-06, + "loss": 0.6341, + "step": 5296 + }, + { + "epoch": 2.8756786102062977, + "grad_norm": 11.720382904656155, + "learning_rate": 8.065699150698442e-06, + "loss": 0.6016, + "step": 5297 + }, + { + "epoch": 2.8762214983713354, + "grad_norm": 14.838955909579008, + "learning_rate": 8.062248849780578e-06, + "loss": 0.8757, + "step": 5298 + }, + { + "epoch": 2.8767643865363732, + "grad_norm": 10.594386765695665, + "learning_rate": 8.05879878852686e-06, + "loss": 0.4864, + "step": 5299 + }, + { + "epoch": 2.8773072747014115, + "grad_norm": 9.83553313698661, + "learning_rate": 8.055348967364007e-06, + "loss": 0.5116, + "step": 5300 + }, + { + "epoch": 2.8778501628664497, + "grad_norm": 12.500251504139728, + "learning_rate": 8.05189938671869e-06, + "loss": 0.891, + "step": 5301 + }, + { + "epoch": 2.8783930510314875, + "grad_norm": 8.988571871064325, + "learning_rate": 8.048450047017563e-06, + "loss": 0.3699, + "step": 5302 + }, + { + "epoch": 2.8789359391965252, + "grad_norm": 10.893701604379793, + "learning_rate": 8.045000948687242e-06, + "loss": 0.5933, + "step": 5303 + }, + { + "epoch": 2.8794788273615635, + "grad_norm": 10.474072412748654, + "learning_rate": 8.041552092154321e-06, + "loss": 0.8787, + "step": 5304 + }, + { + "epoch": 2.8800217155266017, + "grad_norm": 8.555744151752272, + "learning_rate": 8.038103477845357e-06, + "loss": 0.3793, + "step": 5305 + }, + { + "epoch": 2.8805646036916395, + "grad_norm": 9.792363571958322, + "learning_rate": 8.034655106186884e-06, + "loss": 0.5347, + "step": 5306 + }, + { + "epoch": 2.8811074918566772, + "grad_norm": 8.810561156197549, + "learning_rate": 8.031206977605399e-06, + "loss": 0.5128, + "step": 5307 + }, + { + "epoch": 2.8816503800217155, + "grad_norm": 12.383361133227218, + "learning_rate": 8.027759092527374e-06, + "loss": 0.7834, + "step": 5308 + }, + { + "epoch": 2.8821932681867537, + "grad_norm": 9.51185389588206, + "learning_rate": 8.024311451379247e-06, + "loss": 0.4661, + "step": 5309 + }, + { + "epoch": 2.8827361563517915, + "grad_norm": 10.564107506597352, + "learning_rate": 8.020864054587426e-06, + "loss": 0.8199, + "step": 5310 + }, + { + "epoch": 2.8832790445168293, + "grad_norm": 14.579672874726727, + "learning_rate": 8.017416902578296e-06, + "loss": 1.1114, + "step": 5311 + }, + { + "epoch": 2.8838219326818675, + "grad_norm": 10.739540875321001, + "learning_rate": 8.013969995778201e-06, + "loss": 0.5384, + "step": 5312 + }, + { + "epoch": 2.8843648208469057, + "grad_norm": 11.955074827477672, + "learning_rate": 8.010523334613466e-06, + "loss": 0.8475, + "step": 5313 + }, + { + "epoch": 2.8849077090119435, + "grad_norm": 12.842414101367579, + "learning_rate": 8.007076919510371e-06, + "loss": 0.7096, + "step": 5314 + }, + { + "epoch": 2.8854505971769813, + "grad_norm": 9.17766217916239, + "learning_rate": 8.003630750895183e-06, + "loss": 0.4673, + "step": 5315 + }, + { + "epoch": 2.8859934853420195, + "grad_norm": 13.581449358552742, + "learning_rate": 8.000184829194121e-06, + "loss": 0.7501, + "step": 5316 + }, + { + "epoch": 2.8865363735070577, + "grad_norm": 6.157129192897698, + "learning_rate": 7.99673915483339e-06, + "loss": 0.2521, + "step": 5317 + }, + { + "epoch": 2.8870792616720955, + "grad_norm": 9.179432551374715, + "learning_rate": 7.993293728239154e-06, + "loss": 0.5682, + "step": 5318 + }, + { + "epoch": 2.8876221498371333, + "grad_norm": 8.458089104581813, + "learning_rate": 7.989848549837544e-06, + "loss": 0.7131, + "step": 5319 + }, + { + "epoch": 2.8881650380021715, + "grad_norm": 11.039294062728864, + "learning_rate": 7.986403620054674e-06, + "loss": 0.5421, + "step": 5320 + }, + { + "epoch": 2.8887079261672097, + "grad_norm": 8.296966992960693, + "learning_rate": 7.982958939316607e-06, + "loss": 0.5809, + "step": 5321 + }, + { + "epoch": 2.8892508143322475, + "grad_norm": 9.862357699114755, + "learning_rate": 7.979514508049398e-06, + "loss": 0.7253, + "step": 5322 + }, + { + "epoch": 2.8897937024972853, + "grad_norm": 10.60481802559805, + "learning_rate": 7.976070326679053e-06, + "loss": 0.5593, + "step": 5323 + }, + { + "epoch": 2.8903365906623235, + "grad_norm": 12.452210665582081, + "learning_rate": 7.972626395631556e-06, + "loss": 0.9987, + "step": 5324 + }, + { + "epoch": 2.8908794788273617, + "grad_norm": 8.161513982505882, + "learning_rate": 7.969182715332855e-06, + "loss": 0.5242, + "step": 5325 + }, + { + "epoch": 2.8914223669923995, + "grad_norm": 9.073053317988423, + "learning_rate": 7.965739286208878e-06, + "loss": 0.5015, + "step": 5326 + }, + { + "epoch": 2.8919652551574373, + "grad_norm": 8.632672314975157, + "learning_rate": 7.962296108685507e-06, + "loss": 0.4398, + "step": 5327 + }, + { + "epoch": 2.8925081433224755, + "grad_norm": 12.024378540292089, + "learning_rate": 7.958853183188597e-06, + "loss": 0.7876, + "step": 5328 + }, + { + "epoch": 2.8930510314875137, + "grad_norm": 9.943181911450774, + "learning_rate": 7.955410510143982e-06, + "loss": 0.6408, + "step": 5329 + }, + { + "epoch": 2.8935939196525515, + "grad_norm": 11.580662527673274, + "learning_rate": 7.951968089977449e-06, + "loss": 0.6484, + "step": 5330 + }, + { + "epoch": 2.8941368078175893, + "grad_norm": 10.477418271397871, + "learning_rate": 7.948525923114773e-06, + "loss": 0.7197, + "step": 5331 + }, + { + "epoch": 2.8946796959826275, + "grad_norm": 8.959473056544534, + "learning_rate": 7.945084009981674e-06, + "loss": 0.5344, + "step": 5332 + }, + { + "epoch": 2.8952225841476658, + "grad_norm": 7.339268198721023, + "learning_rate": 7.941642351003867e-06, + "loss": 0.4373, + "step": 5333 + }, + { + "epoch": 2.8957654723127035, + "grad_norm": 11.710220973199451, + "learning_rate": 7.938200946607008e-06, + "loss": 0.7354, + "step": 5334 + }, + { + "epoch": 2.8963083604777413, + "grad_norm": 10.548034804633037, + "learning_rate": 7.934759797216744e-06, + "loss": 1.0912, + "step": 5335 + }, + { + "epoch": 2.8968512486427795, + "grad_norm": 7.427458411514844, + "learning_rate": 7.93131890325868e-06, + "loss": 0.4515, + "step": 5336 + }, + { + "epoch": 2.8973941368078178, + "grad_norm": 7.5336162811756475, + "learning_rate": 7.927878265158391e-06, + "loss": 0.4106, + "step": 5337 + }, + { + "epoch": 2.8979370249728555, + "grad_norm": 9.845844831585431, + "learning_rate": 7.924437883341424e-06, + "loss": 0.394, + "step": 5338 + }, + { + "epoch": 2.8984799131378933, + "grad_norm": 11.91533381613983, + "learning_rate": 7.920997758233282e-06, + "loss": 0.5863, + "step": 5339 + }, + { + "epoch": 2.8990228013029316, + "grad_norm": 7.229828747210386, + "learning_rate": 7.917557890259457e-06, + "loss": 0.5663, + "step": 5340 + }, + { + "epoch": 2.8995656894679698, + "grad_norm": 11.378354320604084, + "learning_rate": 7.914118279845385e-06, + "loss": 0.7622, + "step": 5341 + }, + { + "epoch": 2.9001085776330076, + "grad_norm": 7.591103634044025, + "learning_rate": 7.91067892741649e-06, + "loss": 0.3933, + "step": 5342 + }, + { + "epoch": 2.9006514657980453, + "grad_norm": 12.040725731955204, + "learning_rate": 7.907239833398154e-06, + "loss": 0.7143, + "step": 5343 + }, + { + "epoch": 2.9011943539630836, + "grad_norm": 14.022224272834306, + "learning_rate": 7.903800998215735e-06, + "loss": 0.7389, + "step": 5344 + }, + { + "epoch": 2.901737242128122, + "grad_norm": 9.316170523577583, + "learning_rate": 7.900362422294545e-06, + "loss": 0.6061, + "step": 5345 + }, + { + "epoch": 2.9022801302931596, + "grad_norm": 8.20692587330132, + "learning_rate": 7.896924106059882e-06, + "loss": 0.4652, + "step": 5346 + }, + { + "epoch": 2.9028230184581973, + "grad_norm": 12.086544028526333, + "learning_rate": 7.893486049936993e-06, + "loss": 0.6579, + "step": 5347 + }, + { + "epoch": 2.9033659066232356, + "grad_norm": 11.562435941032952, + "learning_rate": 7.89004825435111e-06, + "loss": 0.6619, + "step": 5348 + }, + { + "epoch": 2.903908794788274, + "grad_norm": 9.535770499152292, + "learning_rate": 7.88661071972742e-06, + "loss": 0.4646, + "step": 5349 + }, + { + "epoch": 2.9044516829533116, + "grad_norm": 9.216711023311518, + "learning_rate": 7.88317344649108e-06, + "loss": 0.8006, + "step": 5350 + }, + { + "epoch": 2.9049945711183494, + "grad_norm": 9.577156917577227, + "learning_rate": 7.879736435067228e-06, + "loss": 0.7352, + "step": 5351 + }, + { + "epoch": 2.9055374592833876, + "grad_norm": 7.559386359727195, + "learning_rate": 7.87629968588095e-06, + "loss": 0.3841, + "step": 5352 + }, + { + "epoch": 2.906080347448426, + "grad_norm": 13.704096139083338, + "learning_rate": 7.872863199357315e-06, + "loss": 0.992, + "step": 5353 + }, + { + "epoch": 2.9066232356134636, + "grad_norm": 9.988539472899605, + "learning_rate": 7.869426975921344e-06, + "loss": 0.455, + "step": 5354 + }, + { + "epoch": 2.9071661237785014, + "grad_norm": 11.253082081483269, + "learning_rate": 7.865991015998043e-06, + "loss": 0.8226, + "step": 5355 + }, + { + "epoch": 2.9077090119435396, + "grad_norm": 10.526429101674665, + "learning_rate": 7.862555320012373e-06, + "loss": 0.5357, + "step": 5356 + }, + { + "epoch": 2.908251900108578, + "grad_norm": 9.646031093569851, + "learning_rate": 7.859119888389271e-06, + "loss": 0.4366, + "step": 5357 + }, + { + "epoch": 2.9087947882736156, + "grad_norm": 11.695549963620051, + "learning_rate": 7.855684721553633e-06, + "loss": 0.5834, + "step": 5358 + }, + { + "epoch": 2.9093376764386534, + "grad_norm": 13.315809341009361, + "learning_rate": 7.852249819930324e-06, + "loss": 1.0645, + "step": 5359 + }, + { + "epoch": 2.9098805646036916, + "grad_norm": 12.349221876871807, + "learning_rate": 7.848815183944183e-06, + "loss": 0.577, + "step": 5360 + }, + { + "epoch": 2.91042345276873, + "grad_norm": 16.63305773978975, + "learning_rate": 7.845380814020004e-06, + "loss": 0.9412, + "step": 5361 + }, + { + "epoch": 2.9109663409337676, + "grad_norm": 8.450548593743887, + "learning_rate": 7.841946710582563e-06, + "loss": 0.4955, + "step": 5362 + }, + { + "epoch": 2.9115092290988054, + "grad_norm": 11.465063371310775, + "learning_rate": 7.838512874056589e-06, + "loss": 0.6399, + "step": 5363 + }, + { + "epoch": 2.9120521172638436, + "grad_norm": 8.712777448735437, + "learning_rate": 7.835079304866793e-06, + "loss": 0.309, + "step": 5364 + }, + { + "epoch": 2.912595005428882, + "grad_norm": 9.44188587266759, + "learning_rate": 7.831646003437835e-06, + "loss": 0.435, + "step": 5365 + }, + { + "epoch": 2.9131378935939196, + "grad_norm": 12.094488103222616, + "learning_rate": 7.828212970194357e-06, + "loss": 0.7245, + "step": 5366 + }, + { + "epoch": 2.9136807817589574, + "grad_norm": 11.219679589625631, + "learning_rate": 7.82478020556096e-06, + "loss": 0.8716, + "step": 5367 + }, + { + "epoch": 2.9142236699239956, + "grad_norm": 10.091805149105353, + "learning_rate": 7.821347709962211e-06, + "loss": 0.553, + "step": 5368 + }, + { + "epoch": 2.914766558089034, + "grad_norm": 12.047544009676127, + "learning_rate": 7.817915483822654e-06, + "loss": 0.7283, + "step": 5369 + }, + { + "epoch": 2.9153094462540716, + "grad_norm": 10.891198338724942, + "learning_rate": 7.814483527566783e-06, + "loss": 0.6233, + "step": 5370 + }, + { + "epoch": 2.9158523344191094, + "grad_norm": 8.372829160348498, + "learning_rate": 7.811051841619077e-06, + "loss": 0.3781, + "step": 5371 + }, + { + "epoch": 2.9163952225841476, + "grad_norm": 7.8341866326850464, + "learning_rate": 7.807620426403964e-06, + "loss": 0.3455, + "step": 5372 + }, + { + "epoch": 2.916938110749186, + "grad_norm": 11.836979275353194, + "learning_rate": 7.804189282345855e-06, + "loss": 0.5573, + "step": 5373 + }, + { + "epoch": 2.9174809989142236, + "grad_norm": 8.195935708953526, + "learning_rate": 7.800758409869111e-06, + "loss": 0.4219, + "step": 5374 + }, + { + "epoch": 2.9180238870792614, + "grad_norm": 12.93514826751349, + "learning_rate": 7.797327809398074e-06, + "loss": 1.0471, + "step": 5375 + }, + { + "epoch": 2.9185667752442996, + "grad_norm": 9.384641490219252, + "learning_rate": 7.793897481357047e-06, + "loss": 0.5004, + "step": 5376 + }, + { + "epoch": 2.919109663409338, + "grad_norm": 10.40559875162228, + "learning_rate": 7.79046742617029e-06, + "loss": 0.5434, + "step": 5377 + }, + { + "epoch": 2.9196525515743756, + "grad_norm": 10.11339680443781, + "learning_rate": 7.787037644262048e-06, + "loss": 0.6003, + "step": 5378 + }, + { + "epoch": 2.9201954397394134, + "grad_norm": 12.37057409853331, + "learning_rate": 7.783608136056512e-06, + "loss": 0.685, + "step": 5379 + }, + { + "epoch": 2.9207383279044516, + "grad_norm": 11.589913705537441, + "learning_rate": 7.780178901977857e-06, + "loss": 0.7646, + "step": 5380 + }, + { + "epoch": 2.92128121606949, + "grad_norm": 12.462452174896484, + "learning_rate": 7.77674994245021e-06, + "loss": 0.6442, + "step": 5381 + }, + { + "epoch": 2.9218241042345277, + "grad_norm": 9.567483949166586, + "learning_rate": 7.77332125789768e-06, + "loss": 0.4645, + "step": 5382 + }, + { + "epoch": 2.9223669923995654, + "grad_norm": 11.079212789248007, + "learning_rate": 7.76989284874432e-06, + "loss": 0.7285, + "step": 5383 + }, + { + "epoch": 2.9229098805646037, + "grad_norm": 14.661322701022721, + "learning_rate": 7.76646471541417e-06, + "loss": 0.7481, + "step": 5384 + }, + { + "epoch": 2.923452768729642, + "grad_norm": 13.968723101600672, + "learning_rate": 7.763036858331222e-06, + "loss": 0.6632, + "step": 5385 + }, + { + "epoch": 2.9239956568946797, + "grad_norm": 12.145893524297147, + "learning_rate": 7.759609277919442e-06, + "loss": 0.8389, + "step": 5386 + }, + { + "epoch": 2.9245385450597174, + "grad_norm": 10.052846347427483, + "learning_rate": 7.756181974602757e-06, + "loss": 0.5351, + "step": 5387 + }, + { + "epoch": 2.9250814332247557, + "grad_norm": 11.030654536959574, + "learning_rate": 7.75275494880506e-06, + "loss": 0.6469, + "step": 5388 + }, + { + "epoch": 2.925624321389794, + "grad_norm": 13.670311382356902, + "learning_rate": 7.749328200950215e-06, + "loss": 0.984, + "step": 5389 + }, + { + "epoch": 2.9261672095548317, + "grad_norm": 10.129392847361466, + "learning_rate": 7.745901731462041e-06, + "loss": 0.8149, + "step": 5390 + }, + { + "epoch": 2.9267100977198695, + "grad_norm": 7.867905327161546, + "learning_rate": 7.742475540764339e-06, + "loss": 0.4957, + "step": 5391 + }, + { + "epoch": 2.9272529858849077, + "grad_norm": 8.467285225937513, + "learning_rate": 7.739049629280854e-06, + "loss": 0.579, + "step": 5392 + }, + { + "epoch": 2.927795874049946, + "grad_norm": 8.215885752516279, + "learning_rate": 7.735623997435318e-06, + "loss": 0.4073, + "step": 5393 + }, + { + "epoch": 2.9283387622149837, + "grad_norm": 8.51797242171681, + "learning_rate": 7.732198645651413e-06, + "loss": 0.63, + "step": 5394 + }, + { + "epoch": 2.9288816503800215, + "grad_norm": 15.898718458126094, + "learning_rate": 7.728773574352795e-06, + "loss": 0.7468, + "step": 5395 + }, + { + "epoch": 2.9294245385450597, + "grad_norm": 11.412421000549452, + "learning_rate": 7.725348783963084e-06, + "loss": 0.9522, + "step": 5396 + }, + { + "epoch": 2.929967426710098, + "grad_norm": 9.428473061604137, + "learning_rate": 7.721924274905855e-06, + "loss": 0.5388, + "step": 5397 + }, + { + "epoch": 2.9305103148751357, + "grad_norm": 10.083367258939811, + "learning_rate": 7.718500047604667e-06, + "loss": 0.6605, + "step": 5398 + }, + { + "epoch": 2.9310532030401735, + "grad_norm": 12.85805147473796, + "learning_rate": 7.715076102483026e-06, + "loss": 0.7407, + "step": 5399 + }, + { + "epoch": 2.9315960912052117, + "grad_norm": 9.866769374539361, + "learning_rate": 7.711652439964415e-06, + "loss": 0.6007, + "step": 5400 + }, + { + "epoch": 2.93213897937025, + "grad_norm": 12.367813318049626, + "learning_rate": 7.708229060472277e-06, + "loss": 0.5462, + "step": 5401 + }, + { + "epoch": 2.9326818675352877, + "grad_norm": 13.515093387837554, + "learning_rate": 7.704805964430023e-06, + "loss": 1.0205, + "step": 5402 + }, + { + "epoch": 2.9332247557003255, + "grad_norm": 10.667525494454107, + "learning_rate": 7.701383152261022e-06, + "loss": 0.7873, + "step": 5403 + }, + { + "epoch": 2.9337676438653637, + "grad_norm": 13.024983806475014, + "learning_rate": 7.697960624388621e-06, + "loss": 0.6388, + "step": 5404 + }, + { + "epoch": 2.934310532030402, + "grad_norm": 7.885758920671933, + "learning_rate": 7.694538381236115e-06, + "loss": 0.44, + "step": 5405 + }, + { + "epoch": 2.9348534201954397, + "grad_norm": 9.862140672970398, + "learning_rate": 7.69111642322678e-06, + "loss": 0.4065, + "step": 5406 + }, + { + "epoch": 2.9353963083604775, + "grad_norm": 11.605538896695489, + "learning_rate": 7.687694750783844e-06, + "loss": 0.6574, + "step": 5407 + }, + { + "epoch": 2.9359391965255157, + "grad_norm": 9.384935873492674, + "learning_rate": 7.684273364330505e-06, + "loss": 0.573, + "step": 5408 + }, + { + "epoch": 2.936482084690554, + "grad_norm": 7.506351724565119, + "learning_rate": 7.68085226428993e-06, + "loss": 0.4421, + "step": 5409 + }, + { + "epoch": 2.9370249728555917, + "grad_norm": 9.783370092521682, + "learning_rate": 7.677431451085238e-06, + "loss": 0.49, + "step": 5410 + }, + { + "epoch": 2.9375678610206295, + "grad_norm": 10.612737595311591, + "learning_rate": 7.674010925139533e-06, + "loss": 0.6855, + "step": 5411 + }, + { + "epoch": 2.9381107491856677, + "grad_norm": 10.77127188283207, + "learning_rate": 7.670590686875856e-06, + "loss": 0.4701, + "step": 5412 + }, + { + "epoch": 2.938653637350706, + "grad_norm": 9.009892989770995, + "learning_rate": 7.66717073671724e-06, + "loss": 0.3802, + "step": 5413 + }, + { + "epoch": 2.9391965255157437, + "grad_norm": 9.718396629861862, + "learning_rate": 7.663751075086662e-06, + "loss": 0.6433, + "step": 5414 + }, + { + "epoch": 2.9397394136807815, + "grad_norm": 8.544955706387817, + "learning_rate": 7.66033170240708e-06, + "loss": 0.3449, + "step": 5415 + }, + { + "epoch": 2.9402823018458197, + "grad_norm": 12.174166457717448, + "learning_rate": 7.656912619101401e-06, + "loss": 0.4685, + "step": 5416 + }, + { + "epoch": 2.940825190010858, + "grad_norm": 8.829471274310071, + "learning_rate": 7.6534938255925e-06, + "loss": 0.6213, + "step": 5417 + }, + { + "epoch": 2.9413680781758957, + "grad_norm": 11.089620150396728, + "learning_rate": 7.650075322303222e-06, + "loss": 0.5682, + "step": 5418 + }, + { + "epoch": 2.9419109663409335, + "grad_norm": 7.099888540087664, + "learning_rate": 7.646657109656375e-06, + "loss": 0.2937, + "step": 5419 + }, + { + "epoch": 2.9424538545059717, + "grad_norm": 9.8313977098716, + "learning_rate": 7.64323918807473e-06, + "loss": 0.6224, + "step": 5420 + }, + { + "epoch": 2.94299674267101, + "grad_norm": 14.892485492602605, + "learning_rate": 7.639821557981013e-06, + "loss": 1.0678, + "step": 5421 + }, + { + "epoch": 2.9435396308360477, + "grad_norm": 8.648910922847081, + "learning_rate": 7.636404219797931e-06, + "loss": 0.3605, + "step": 5422 + }, + { + "epoch": 2.9440825190010855, + "grad_norm": 12.093378946161085, + "learning_rate": 7.63298717394814e-06, + "loss": 0.9133, + "step": 5423 + }, + { + "epoch": 2.9446254071661238, + "grad_norm": 9.302461570715176, + "learning_rate": 7.629570420854271e-06, + "loss": 0.5305, + "step": 5424 + }, + { + "epoch": 2.945168295331162, + "grad_norm": 10.492242594764365, + "learning_rate": 7.626153960938909e-06, + "loss": 0.6007, + "step": 5425 + }, + { + "epoch": 2.9457111834961998, + "grad_norm": 11.212789857802488, + "learning_rate": 7.622737794624605e-06, + "loss": 0.4854, + "step": 5426 + }, + { + "epoch": 2.9462540716612375, + "grad_norm": 11.607518266622822, + "learning_rate": 7.619321922333884e-06, + "loss": 0.5543, + "step": 5427 + }, + { + "epoch": 2.9467969598262758, + "grad_norm": 11.268089998182159, + "learning_rate": 7.615906344489219e-06, + "loss": 0.6998, + "step": 5428 + }, + { + "epoch": 2.947339847991314, + "grad_norm": 12.383131636281371, + "learning_rate": 7.61249106151306e-06, + "loss": 0.8179, + "step": 5429 + }, + { + "epoch": 2.9478827361563518, + "grad_norm": 11.603914807740104, + "learning_rate": 7.609076073827808e-06, + "loss": 0.9006, + "step": 5430 + }, + { + "epoch": 2.9484256243213895, + "grad_norm": 9.674307468707637, + "learning_rate": 7.6056613818558415e-06, + "loss": 0.6638, + "step": 5431 + }, + { + "epoch": 2.9489685124864278, + "grad_norm": 13.055177059547185, + "learning_rate": 7.602246986019487e-06, + "loss": 0.9642, + "step": 5432 + }, + { + "epoch": 2.949511400651466, + "grad_norm": 11.00057075160912, + "learning_rate": 7.598832886741049e-06, + "loss": 0.4843, + "step": 5433 + }, + { + "epoch": 2.950054288816504, + "grad_norm": 11.15214171419394, + "learning_rate": 7.5954190844427834e-06, + "loss": 0.6722, + "step": 5434 + }, + { + "epoch": 2.9505971769815416, + "grad_norm": 8.91145682769725, + "learning_rate": 7.592005579546924e-06, + "loss": 0.6258, + "step": 5435 + }, + { + "epoch": 2.95114006514658, + "grad_norm": 12.685207726915342, + "learning_rate": 7.588592372475651e-06, + "loss": 0.8481, + "step": 5436 + }, + { + "epoch": 2.951682953311618, + "grad_norm": 13.10934163141668, + "learning_rate": 7.585179463651112e-06, + "loss": 0.6655, + "step": 5437 + }, + { + "epoch": 2.952225841476656, + "grad_norm": 10.993826657178786, + "learning_rate": 7.5817668534954295e-06, + "loss": 0.5687, + "step": 5438 + }, + { + "epoch": 2.9527687296416936, + "grad_norm": 8.424743777393193, + "learning_rate": 7.578354542430673e-06, + "loss": 0.5485, + "step": 5439 + }, + { + "epoch": 2.953311617806732, + "grad_norm": 14.730970588587084, + "learning_rate": 7.574942530878891e-06, + "loss": 0.8759, + "step": 5440 + }, + { + "epoch": 2.95385450597177, + "grad_norm": 8.667781572678388, + "learning_rate": 7.571530819262077e-06, + "loss": 0.3704, + "step": 5441 + }, + { + "epoch": 2.954397394136808, + "grad_norm": 6.739834901590041, + "learning_rate": 7.568119408002206e-06, + "loss": 0.3619, + "step": 5442 + }, + { + "epoch": 2.9549402823018456, + "grad_norm": 5.9221212094491795, + "learning_rate": 7.564708297521197e-06, + "loss": 0.3177, + "step": 5443 + }, + { + "epoch": 2.955483170466884, + "grad_norm": 9.008687743468037, + "learning_rate": 7.561297488240953e-06, + "loss": 0.5756, + "step": 5444 + }, + { + "epoch": 2.956026058631922, + "grad_norm": 10.49570531884228, + "learning_rate": 7.55788698058332e-06, + "loss": 0.609, + "step": 5445 + }, + { + "epoch": 2.95656894679696, + "grad_norm": 11.053313913817204, + "learning_rate": 7.554476774970115e-06, + "loss": 0.6893, + "step": 5446 + }, + { + "epoch": 2.9571118349619976, + "grad_norm": 10.747800237741414, + "learning_rate": 7.551066871823125e-06, + "loss": 0.6361, + "step": 5447 + }, + { + "epoch": 2.957654723127036, + "grad_norm": 12.158436457559148, + "learning_rate": 7.547657271564083e-06, + "loss": 0.6427, + "step": 5448 + }, + { + "epoch": 2.958197611292074, + "grad_norm": 11.021084054397756, + "learning_rate": 7.544247974614701e-06, + "loss": 0.4399, + "step": 5449 + }, + { + "epoch": 2.958740499457112, + "grad_norm": 8.337942068125022, + "learning_rate": 7.540838981396641e-06, + "loss": 0.3781, + "step": 5450 + }, + { + "epoch": 2.9592833876221496, + "grad_norm": 9.86638573791189, + "learning_rate": 7.537430292331536e-06, + "loss": 0.4257, + "step": 5451 + }, + { + "epoch": 2.959826275787188, + "grad_norm": 11.512553822962575, + "learning_rate": 7.534021907840975e-06, + "loss": 0.64, + "step": 5452 + }, + { + "epoch": 2.960369163952226, + "grad_norm": 10.540338627981921, + "learning_rate": 7.530613828346519e-06, + "loss": 0.6669, + "step": 5453 + }, + { + "epoch": 2.960912052117264, + "grad_norm": 12.092382555989715, + "learning_rate": 7.5272060542696765e-06, + "loss": 0.6346, + "step": 5454 + }, + { + "epoch": 2.9614549402823016, + "grad_norm": 11.429168527550765, + "learning_rate": 7.5237985860319344e-06, + "loss": 0.7454, + "step": 5455 + }, + { + "epoch": 2.96199782844734, + "grad_norm": 10.331080624952307, + "learning_rate": 7.52039142405473e-06, + "loss": 0.5472, + "step": 5456 + }, + { + "epoch": 2.962540716612378, + "grad_norm": 16.484391160691512, + "learning_rate": 7.516984568759461e-06, + "loss": 1.3039, + "step": 5457 + }, + { + "epoch": 2.963083604777416, + "grad_norm": 12.853254391289898, + "learning_rate": 7.5135780205675e-06, + "loss": 0.6846, + "step": 5458 + }, + { + "epoch": 2.9636264929424536, + "grad_norm": 12.062646480747443, + "learning_rate": 7.510171779900171e-06, + "loss": 0.5656, + "step": 5459 + }, + { + "epoch": 2.964169381107492, + "grad_norm": 8.21924446991787, + "learning_rate": 7.506765847178768e-06, + "loss": 0.3401, + "step": 5460 + }, + { + "epoch": 2.96471226927253, + "grad_norm": 11.116098821110214, + "learning_rate": 7.503360222824535e-06, + "loss": 0.4931, + "step": 5461 + }, + { + "epoch": 2.965255157437568, + "grad_norm": 9.759573410672228, + "learning_rate": 7.499954907258693e-06, + "loss": 0.5363, + "step": 5462 + }, + { + "epoch": 2.9657980456026056, + "grad_norm": 15.195018271421278, + "learning_rate": 7.496549900902408e-06, + "loss": 0.5349, + "step": 5463 + }, + { + "epoch": 2.966340933767644, + "grad_norm": 11.555768004386108, + "learning_rate": 7.493145204176823e-06, + "loss": 0.8273, + "step": 5464 + }, + { + "epoch": 2.966883821932682, + "grad_norm": 12.516961882115927, + "learning_rate": 7.4897408175030366e-06, + "loss": 0.7826, + "step": 5465 + }, + { + "epoch": 2.96742671009772, + "grad_norm": 11.67450620684471, + "learning_rate": 7.486336741302103e-06, + "loss": 0.5689, + "step": 5466 + }, + { + "epoch": 2.9679695982627576, + "grad_norm": 13.075852566862391, + "learning_rate": 7.48293297599505e-06, + "loss": 0.8355, + "step": 5467 + }, + { + "epoch": 2.968512486427796, + "grad_norm": 8.943584587537828, + "learning_rate": 7.479529522002855e-06, + "loss": 0.3921, + "step": 5468 + }, + { + "epoch": 2.969055374592834, + "grad_norm": 10.91179454178022, + "learning_rate": 7.476126379746471e-06, + "loss": 0.7276, + "step": 5469 + }, + { + "epoch": 2.969598262757872, + "grad_norm": 11.79182407811329, + "learning_rate": 7.472723549646793e-06, + "loss": 1.012, + "step": 5470 + }, + { + "epoch": 2.9701411509229096, + "grad_norm": 12.476871332438998, + "learning_rate": 7.469321032124697e-06, + "loss": 0.7994, + "step": 5471 + }, + { + "epoch": 2.970684039087948, + "grad_norm": 14.350758882559228, + "learning_rate": 7.465918827601008e-06, + "loss": 0.8477, + "step": 5472 + }, + { + "epoch": 2.971226927252986, + "grad_norm": 13.79963097595528, + "learning_rate": 7.46251693649652e-06, + "loss": 1.0897, + "step": 5473 + }, + { + "epoch": 2.971769815418024, + "grad_norm": 9.061867840853571, + "learning_rate": 7.459115359231977e-06, + "loss": 0.2972, + "step": 5474 + }, + { + "epoch": 2.9723127035830617, + "grad_norm": 11.193898477701971, + "learning_rate": 7.455714096228102e-06, + "loss": 0.9238, + "step": 5475 + }, + { + "epoch": 2.9728555917481, + "grad_norm": 13.156649778942683, + "learning_rate": 7.452313147905559e-06, + "loss": 0.9945, + "step": 5476 + }, + { + "epoch": 2.973398479913138, + "grad_norm": 10.673119691620487, + "learning_rate": 7.448912514684985e-06, + "loss": 0.6296, + "step": 5477 + }, + { + "epoch": 2.973941368078176, + "grad_norm": 13.208883224921669, + "learning_rate": 7.44551219698698e-06, + "loss": 0.5018, + "step": 5478 + }, + { + "epoch": 2.9744842562432137, + "grad_norm": 10.477928943580013, + "learning_rate": 7.442112195232093e-06, + "loss": 0.4782, + "step": 5479 + }, + { + "epoch": 2.975027144408252, + "grad_norm": 10.052094162546567, + "learning_rate": 7.438712509840852e-06, + "loss": 0.5661, + "step": 5480 + }, + { + "epoch": 2.97557003257329, + "grad_norm": 14.921523553916886, + "learning_rate": 7.435313141233724e-06, + "loss": 1.3182, + "step": 5481 + }, + { + "epoch": 2.976112920738328, + "grad_norm": 9.3079131640488, + "learning_rate": 7.43191408983116e-06, + "loss": 0.5672, + "step": 5482 + }, + { + "epoch": 2.9766558089033657, + "grad_norm": 8.76037752811841, + "learning_rate": 7.428515356053551e-06, + "loss": 0.6026, + "step": 5483 + }, + { + "epoch": 2.977198697068404, + "grad_norm": 11.037798492427186, + "learning_rate": 7.425116940321262e-06, + "loss": 0.7053, + "step": 5484 + }, + { + "epoch": 2.977741585233442, + "grad_norm": 12.329608171901635, + "learning_rate": 7.421718843054615e-06, + "loss": 0.6576, + "step": 5485 + }, + { + "epoch": 2.97828447339848, + "grad_norm": 14.664205153544405, + "learning_rate": 7.4183210646738875e-06, + "loss": 0.7546, + "step": 5486 + }, + { + "epoch": 2.9788273615635177, + "grad_norm": 16.28337626285995, + "learning_rate": 7.414923605599329e-06, + "loss": 1.2774, + "step": 5487 + }, + { + "epoch": 2.979370249728556, + "grad_norm": 11.936294560502736, + "learning_rate": 7.411526466251135e-06, + "loss": 0.9397, + "step": 5488 + }, + { + "epoch": 2.979913137893594, + "grad_norm": 15.443918342772207, + "learning_rate": 7.408129647049474e-06, + "loss": 0.7591, + "step": 5489 + }, + { + "epoch": 2.980456026058632, + "grad_norm": 9.303363471188012, + "learning_rate": 7.404733148414471e-06, + "loss": 0.4917, + "step": 5490 + }, + { + "epoch": 2.9809989142236697, + "grad_norm": 16.111589693524394, + "learning_rate": 7.401336970766207e-06, + "loss": 0.8724, + "step": 5491 + }, + { + "epoch": 2.981541802388708, + "grad_norm": 15.913107093535366, + "learning_rate": 7.397941114524727e-06, + "loss": 1.0915, + "step": 5492 + }, + { + "epoch": 2.982084690553746, + "grad_norm": 12.432387434616404, + "learning_rate": 7.3945455801100404e-06, + "loss": 1.1143, + "step": 5493 + }, + { + "epoch": 2.982627578718784, + "grad_norm": 11.65910861056599, + "learning_rate": 7.39115036794211e-06, + "loss": 0.7012, + "step": 5494 + }, + { + "epoch": 2.9831704668838217, + "grad_norm": 10.645886415744744, + "learning_rate": 7.387755478440855e-06, + "loss": 0.5327, + "step": 5495 + }, + { + "epoch": 2.98371335504886, + "grad_norm": 11.340492191845808, + "learning_rate": 7.384360912026167e-06, + "loss": 0.8393, + "step": 5496 + }, + { + "epoch": 2.984256243213898, + "grad_norm": 11.222712079595647, + "learning_rate": 7.38096666911789e-06, + "loss": 0.4731, + "step": 5497 + }, + { + "epoch": 2.984799131378936, + "grad_norm": 14.06199638971423, + "learning_rate": 7.377572750135833e-06, + "loss": 1.0083, + "step": 5498 + }, + { + "epoch": 2.9853420195439737, + "grad_norm": 11.609516661749161, + "learning_rate": 7.374179155499752e-06, + "loss": 0.5397, + "step": 5499 + }, + { + "epoch": 2.985884907709012, + "grad_norm": 11.809076211376697, + "learning_rate": 7.370785885629384e-06, + "loss": 0.5726, + "step": 5500 + }, + { + "epoch": 2.98642779587405, + "grad_norm": 12.767099228571903, + "learning_rate": 7.367392940944403e-06, + "loss": 0.8919, + "step": 5501 + }, + { + "epoch": 2.986970684039088, + "grad_norm": 8.511059646511987, + "learning_rate": 7.364000321864464e-06, + "loss": 0.4431, + "step": 5502 + }, + { + "epoch": 2.9875135722041257, + "grad_norm": 14.166198872023001, + "learning_rate": 7.360608028809161e-06, + "loss": 0.8709, + "step": 5503 + }, + { + "epoch": 2.988056460369164, + "grad_norm": 11.774867588892974, + "learning_rate": 7.357216062198066e-06, + "loss": 0.4857, + "step": 5504 + }, + { + "epoch": 2.988599348534202, + "grad_norm": 9.622067683785492, + "learning_rate": 7.353824422450702e-06, + "loss": 0.4385, + "step": 5505 + }, + { + "epoch": 2.98914223669924, + "grad_norm": 9.959533869172905, + "learning_rate": 7.3504331099865474e-06, + "loss": 0.6041, + "step": 5506 + }, + { + "epoch": 2.9896851248642777, + "grad_norm": 12.611515654376284, + "learning_rate": 7.347042125225052e-06, + "loss": 0.5667, + "step": 5507 + }, + { + "epoch": 2.990228013029316, + "grad_norm": 16.346151732068346, + "learning_rate": 7.343651468585611e-06, + "loss": 0.8017, + "step": 5508 + }, + { + "epoch": 2.990770901194354, + "grad_norm": 14.039933672423732, + "learning_rate": 7.340261140487593e-06, + "loss": 0.6905, + "step": 5509 + }, + { + "epoch": 2.991313789359392, + "grad_norm": 9.68037751098424, + "learning_rate": 7.336871141350313e-06, + "loss": 0.7553, + "step": 5510 + }, + { + "epoch": 2.9918566775244297, + "grad_norm": 11.6044953546391, + "learning_rate": 7.333481471593058e-06, + "loss": 0.7525, + "step": 5511 + }, + { + "epoch": 2.992399565689468, + "grad_norm": 10.257246759089158, + "learning_rate": 7.330092131635061e-06, + "loss": 0.5479, + "step": 5512 + }, + { + "epoch": 2.992942453854506, + "grad_norm": 9.37947259585779, + "learning_rate": 7.326703121895528e-06, + "loss": 0.6117, + "step": 5513 + }, + { + "epoch": 2.993485342019544, + "grad_norm": 10.344183596900065, + "learning_rate": 7.323314442793614e-06, + "loss": 0.5074, + "step": 5514 + }, + { + "epoch": 2.9940282301845818, + "grad_norm": 8.940259405187144, + "learning_rate": 7.319926094748433e-06, + "loss": 0.2863, + "step": 5515 + }, + { + "epoch": 2.99457111834962, + "grad_norm": 10.920924886109798, + "learning_rate": 7.316538078179065e-06, + "loss": 0.7706, + "step": 5516 + }, + { + "epoch": 2.995114006514658, + "grad_norm": 4.8493484123114445, + "learning_rate": 7.3131503935045424e-06, + "loss": 0.2066, + "step": 5517 + }, + { + "epoch": 2.995656894679696, + "grad_norm": 11.385458894978902, + "learning_rate": 7.309763041143865e-06, + "loss": 0.5706, + "step": 5518 + }, + { + "epoch": 2.9961997828447338, + "grad_norm": 8.263697834815542, + "learning_rate": 7.306376021515977e-06, + "loss": 0.667, + "step": 5519 + }, + { + "epoch": 2.996742671009772, + "grad_norm": 11.30709409496411, + "learning_rate": 7.302989335039801e-06, + "loss": 0.5522, + "step": 5520 + }, + { + "epoch": 2.99728555917481, + "grad_norm": 9.914253293704078, + "learning_rate": 7.299602982134199e-06, + "loss": 0.6109, + "step": 5521 + }, + { + "epoch": 2.997828447339848, + "grad_norm": 15.471798238376472, + "learning_rate": 7.2962169632180055e-06, + "loss": 0.8321, + "step": 5522 + }, + { + "epoch": 2.9983713355048858, + "grad_norm": 10.956384342902831, + "learning_rate": 7.292831278710007e-06, + "loss": 0.5791, + "step": 5523 + }, + { + "epoch": 2.998914223669924, + "grad_norm": 8.8527583386149, + "learning_rate": 7.289445929028951e-06, + "loss": 0.5807, + "step": 5524 + }, + { + "epoch": 2.999457111834962, + "grad_norm": 10.520949488289421, + "learning_rate": 7.286060914593544e-06, + "loss": 0.6861, + "step": 5525 + }, + { + "epoch": 3.0, + "grad_norm": 11.327416720286067, + "learning_rate": 7.282676235822444e-06, + "loss": 0.4657, + "step": 5526 + }, + { + "epoch": 3.000542888165038, + "grad_norm": 8.97355367199935, + "learning_rate": 7.279291893134283e-06, + "loss": 0.2931, + "step": 5527 + }, + { + "epoch": 3.001085776330076, + "grad_norm": 8.169786294777076, + "learning_rate": 7.275907886947632e-06, + "loss": 0.3065, + "step": 5528 + }, + { + "epoch": 3.001628664495114, + "grad_norm": 9.498228913363004, + "learning_rate": 7.272524217681036e-06, + "loss": 0.3704, + "step": 5529 + }, + { + "epoch": 3.002171552660152, + "grad_norm": 11.241888839923089, + "learning_rate": 7.269140885752992e-06, + "loss": 0.5917, + "step": 5530 + }, + { + "epoch": 3.00271444082519, + "grad_norm": 8.725213535150235, + "learning_rate": 7.26575789158196e-06, + "loss": 0.4699, + "step": 5531 + }, + { + "epoch": 3.003257328990228, + "grad_norm": 7.545852057856303, + "learning_rate": 7.262375235586343e-06, + "loss": 0.5126, + "step": 5532 + }, + { + "epoch": 3.003800217155266, + "grad_norm": 11.581979137257708, + "learning_rate": 7.2589929181845255e-06, + "loss": 0.4424, + "step": 5533 + }, + { + "epoch": 3.004343105320304, + "grad_norm": 9.70311467595886, + "learning_rate": 7.255610939794831e-06, + "loss": 0.4074, + "step": 5534 + }, + { + "epoch": 3.004885993485342, + "grad_norm": 7.016594112536396, + "learning_rate": 7.252229300835545e-06, + "loss": 0.2748, + "step": 5535 + }, + { + "epoch": 3.00542888165038, + "grad_norm": 11.131101403236986, + "learning_rate": 7.248848001724924e-06, + "loss": 0.4696, + "step": 5536 + }, + { + "epoch": 3.005971769815418, + "grad_norm": 9.402705826901306, + "learning_rate": 7.2454670428811625e-06, + "loss": 0.4554, + "step": 5537 + }, + { + "epoch": 3.006514657980456, + "grad_norm": 10.04584435406275, + "learning_rate": 7.242086424722432e-06, + "loss": 0.5017, + "step": 5538 + }, + { + "epoch": 3.007057546145494, + "grad_norm": 9.814079675077302, + "learning_rate": 7.238706147666843e-06, + "loss": 0.3366, + "step": 5539 + }, + { + "epoch": 3.007600434310532, + "grad_norm": 8.970687114442066, + "learning_rate": 7.235326212132483e-06, + "loss": 0.5007, + "step": 5540 + }, + { + "epoch": 3.00814332247557, + "grad_norm": 10.57886281786491, + "learning_rate": 7.23194661853738e-06, + "loss": 0.3796, + "step": 5541 + }, + { + "epoch": 3.008686210640608, + "grad_norm": 10.781975432954557, + "learning_rate": 7.228567367299532e-06, + "loss": 0.5567, + "step": 5542 + }, + { + "epoch": 3.009229098805646, + "grad_norm": 11.394439293433617, + "learning_rate": 7.2251884588368916e-06, + "loss": 0.521, + "step": 5543 + }, + { + "epoch": 3.009771986970684, + "grad_norm": 10.408091638796, + "learning_rate": 7.22180989356736e-06, + "loss": 0.4805, + "step": 5544 + }, + { + "epoch": 3.010314875135722, + "grad_norm": 11.389146042667898, + "learning_rate": 7.218431671908812e-06, + "loss": 0.49, + "step": 5545 + }, + { + "epoch": 3.01085776330076, + "grad_norm": 12.57922467984807, + "learning_rate": 7.215053794279066e-06, + "loss": 0.5101, + "step": 5546 + }, + { + "epoch": 3.011400651465798, + "grad_norm": 14.591941806963382, + "learning_rate": 7.211676261095905e-06, + "loss": 0.6942, + "step": 5547 + }, + { + "epoch": 3.011943539630836, + "grad_norm": 12.696865069952242, + "learning_rate": 7.20829907277707e-06, + "loss": 0.6091, + "step": 5548 + }, + { + "epoch": 3.012486427795874, + "grad_norm": 8.118652375371086, + "learning_rate": 7.204922229740255e-06, + "loss": 0.346, + "step": 5549 + }, + { + "epoch": 3.013029315960912, + "grad_norm": 10.942362893265422, + "learning_rate": 7.2015457324031105e-06, + "loss": 0.6571, + "step": 5550 + }, + { + "epoch": 3.01357220412595, + "grad_norm": 10.204677993952515, + "learning_rate": 7.198169581183254e-06, + "loss": 0.5563, + "step": 5551 + }, + { + "epoch": 3.014115092290988, + "grad_norm": 9.634479609863043, + "learning_rate": 7.194793776498245e-06, + "loss": 0.5768, + "step": 5552 + }, + { + "epoch": 3.014657980456026, + "grad_norm": 9.591830793887732, + "learning_rate": 7.191418318765616e-06, + "loss": 0.3283, + "step": 5553 + }, + { + "epoch": 3.015200868621064, + "grad_norm": 13.201493691800454, + "learning_rate": 7.188043208402844e-06, + "loss": 0.6985, + "step": 5554 + }, + { + "epoch": 3.015743756786102, + "grad_norm": 8.863552231556248, + "learning_rate": 7.184668445827367e-06, + "loss": 0.569, + "step": 5555 + }, + { + "epoch": 3.01628664495114, + "grad_norm": 16.6519621896966, + "learning_rate": 7.181294031456589e-06, + "loss": 0.5339, + "step": 5556 + }, + { + "epoch": 3.016829533116178, + "grad_norm": 10.16813004001019, + "learning_rate": 7.177919965707853e-06, + "loss": 0.4042, + "step": 5557 + }, + { + "epoch": 3.017372421281216, + "grad_norm": 9.578956705487652, + "learning_rate": 7.174546248998477e-06, + "loss": 0.454, + "step": 5558 + }, + { + "epoch": 3.017915309446254, + "grad_norm": 12.884511869394336, + "learning_rate": 7.1711728817457204e-06, + "loss": 0.769, + "step": 5559 + }, + { + "epoch": 3.018458197611292, + "grad_norm": 9.438716836740053, + "learning_rate": 7.1677998643668124e-06, + "loss": 0.3148, + "step": 5560 + }, + { + "epoch": 3.01900108577633, + "grad_norm": 10.176350005774413, + "learning_rate": 7.164427197278931e-06, + "loss": 0.4634, + "step": 5561 + }, + { + "epoch": 3.019543973941368, + "grad_norm": 13.390457057860667, + "learning_rate": 7.161054880899212e-06, + "loss": 0.8095, + "step": 5562 + }, + { + "epoch": 3.020086862106406, + "grad_norm": 15.075886108978443, + "learning_rate": 7.157682915644754e-06, + "loss": 0.4937, + "step": 5563 + }, + { + "epoch": 3.020629750271444, + "grad_norm": 10.999178702680004, + "learning_rate": 7.154311301932597e-06, + "loss": 0.6033, + "step": 5564 + }, + { + "epoch": 3.021172638436482, + "grad_norm": 11.230531078526122, + "learning_rate": 7.150940040179759e-06, + "loss": 0.4776, + "step": 5565 + }, + { + "epoch": 3.02171552660152, + "grad_norm": 10.476946027170913, + "learning_rate": 7.147569130803193e-06, + "loss": 0.3336, + "step": 5566 + }, + { + "epoch": 3.022258414766558, + "grad_norm": 11.753790547662213, + "learning_rate": 7.144198574219824e-06, + "loss": 0.5004, + "step": 5567 + }, + { + "epoch": 3.022801302931596, + "grad_norm": 8.64883980987418, + "learning_rate": 7.140828370846525e-06, + "loss": 0.3042, + "step": 5568 + }, + { + "epoch": 3.023344191096634, + "grad_norm": 13.493035832393069, + "learning_rate": 7.137458521100136e-06, + "loss": 0.5849, + "step": 5569 + }, + { + "epoch": 3.023887079261672, + "grad_norm": 13.029908881843049, + "learning_rate": 7.1340890253974324e-06, + "loss": 0.583, + "step": 5570 + }, + { + "epoch": 3.02442996742671, + "grad_norm": 11.852047436985602, + "learning_rate": 7.130719884155173e-06, + "loss": 0.4575, + "step": 5571 + }, + { + "epoch": 3.024972855591748, + "grad_norm": 13.476829762238244, + "learning_rate": 7.127351097790045e-06, + "loss": 0.5722, + "step": 5572 + }, + { + "epoch": 3.025515743756786, + "grad_norm": 11.710899279093313, + "learning_rate": 7.123982666718716e-06, + "loss": 0.5369, + "step": 5573 + }, + { + "epoch": 3.026058631921824, + "grad_norm": 12.982691203975197, + "learning_rate": 7.120614591357793e-06, + "loss": 0.4587, + "step": 5574 + }, + { + "epoch": 3.026601520086862, + "grad_norm": 13.494356693487626, + "learning_rate": 7.117246872123843e-06, + "loss": 0.5962, + "step": 5575 + }, + { + "epoch": 3.0271444082519, + "grad_norm": 9.838797633608028, + "learning_rate": 7.113879509433399e-06, + "loss": 0.3549, + "step": 5576 + }, + { + "epoch": 3.027687296416938, + "grad_norm": 15.816616895711462, + "learning_rate": 7.110512503702933e-06, + "loss": 0.6669, + "step": 5577 + }, + { + "epoch": 3.028230184581976, + "grad_norm": 11.513602729284349, + "learning_rate": 7.1071458553488904e-06, + "loss": 0.3573, + "step": 5578 + }, + { + "epoch": 3.028773072747014, + "grad_norm": 14.040376133978413, + "learning_rate": 7.103779564787654e-06, + "loss": 0.5898, + "step": 5579 + }, + { + "epoch": 3.029315960912052, + "grad_norm": 12.839234013609035, + "learning_rate": 7.10041363243558e-06, + "loss": 0.525, + "step": 5580 + }, + { + "epoch": 3.02985884907709, + "grad_norm": 12.773102345996788, + "learning_rate": 7.097048058708966e-06, + "loss": 0.7338, + "step": 5581 + }, + { + "epoch": 3.030401737242128, + "grad_norm": 9.757116728754855, + "learning_rate": 7.093682844024079e-06, + "loss": 0.4107, + "step": 5582 + }, + { + "epoch": 3.030944625407166, + "grad_norm": 13.284417042689952, + "learning_rate": 7.090317988797131e-06, + "loss": 0.4909, + "step": 5583 + }, + { + "epoch": 3.031487513572204, + "grad_norm": 12.858923373950708, + "learning_rate": 7.086953493444286e-06, + "loss": 0.4164, + "step": 5584 + }, + { + "epoch": 3.032030401737242, + "grad_norm": 11.460500988764892, + "learning_rate": 7.083589358381681e-06, + "loss": 0.6339, + "step": 5585 + }, + { + "epoch": 3.03257328990228, + "grad_norm": 11.76507179913234, + "learning_rate": 7.0802255840253865e-06, + "loss": 0.33, + "step": 5586 + }, + { + "epoch": 3.033116178067318, + "grad_norm": 14.990076792803963, + "learning_rate": 7.076862170791449e-06, + "loss": 0.5647, + "step": 5587 + }, + { + "epoch": 3.033659066232356, + "grad_norm": 12.953159557836509, + "learning_rate": 7.0734991190958545e-06, + "loss": 0.5641, + "step": 5588 + }, + { + "epoch": 3.034201954397394, + "grad_norm": 9.746562255040114, + "learning_rate": 7.070136429354557e-06, + "loss": 0.5755, + "step": 5589 + }, + { + "epoch": 3.034744842562432, + "grad_norm": 14.867296485205143, + "learning_rate": 7.066774101983452e-06, + "loss": 0.4915, + "step": 5590 + }, + { + "epoch": 3.03528773072747, + "grad_norm": 12.725313171891754, + "learning_rate": 7.063412137398402e-06, + "loss": 0.541, + "step": 5591 + }, + { + "epoch": 3.035830618892508, + "grad_norm": 11.023036821532774, + "learning_rate": 7.060050536015219e-06, + "loss": 0.3524, + "step": 5592 + }, + { + "epoch": 3.036373507057546, + "grad_norm": 9.747447295285967, + "learning_rate": 7.05668929824967e-06, + "loss": 0.4588, + "step": 5593 + }, + { + "epoch": 3.036916395222584, + "grad_norm": 12.477427859321583, + "learning_rate": 7.0533284245174826e-06, + "loss": 0.382, + "step": 5594 + }, + { + "epoch": 3.037459283387622, + "grad_norm": 9.62366763242101, + "learning_rate": 7.049967915234329e-06, + "loss": 0.2984, + "step": 5595 + }, + { + "epoch": 3.03800217155266, + "grad_norm": 13.368913470296512, + "learning_rate": 7.046607770815849e-06, + "loss": 0.6245, + "step": 5596 + }, + { + "epoch": 3.038545059717698, + "grad_norm": 10.380221393303437, + "learning_rate": 7.043247991677622e-06, + "loss": 0.4542, + "step": 5597 + }, + { + "epoch": 3.039087947882736, + "grad_norm": 9.778914279269939, + "learning_rate": 7.039888578235201e-06, + "loss": 0.3281, + "step": 5598 + }, + { + "epoch": 3.039630836047774, + "grad_norm": 7.360801727275878, + "learning_rate": 7.036529530904073e-06, + "loss": 0.2668, + "step": 5599 + }, + { + "epoch": 3.040173724212812, + "grad_norm": 7.962043569210736, + "learning_rate": 7.033170850099699e-06, + "loss": 0.2437, + "step": 5600 + }, + { + "epoch": 3.04071661237785, + "grad_norm": 13.434858118835157, + "learning_rate": 7.029812536237479e-06, + "loss": 0.5657, + "step": 5601 + }, + { + "epoch": 3.041259500542888, + "grad_norm": 12.720111358562143, + "learning_rate": 7.026454589732784e-06, + "loss": 0.5799, + "step": 5602 + }, + { + "epoch": 3.041802388707926, + "grad_norm": 13.271222462410066, + "learning_rate": 7.023097011000923e-06, + "loss": 0.4124, + "step": 5603 + }, + { + "epoch": 3.042345276872964, + "grad_norm": 12.713200398376557, + "learning_rate": 7.019739800457166e-06, + "loss": 0.5288, + "step": 5604 + }, + { + "epoch": 3.042888165038002, + "grad_norm": 8.938306532181711, + "learning_rate": 7.016382958516741e-06, + "loss": 0.3436, + "step": 5605 + }, + { + "epoch": 3.04343105320304, + "grad_norm": 11.284866978643858, + "learning_rate": 7.0130264855948245e-06, + "loss": 0.3379, + "step": 5606 + }, + { + "epoch": 3.043973941368078, + "grad_norm": 13.192559496466789, + "learning_rate": 7.009670382106558e-06, + "loss": 0.4862, + "step": 5607 + }, + { + "epoch": 3.044516829533116, + "grad_norm": 11.479680729723038, + "learning_rate": 7.006314648467021e-06, + "loss": 0.3483, + "step": 5608 + }, + { + "epoch": 3.045059717698154, + "grad_norm": 12.417526118207578, + "learning_rate": 7.002959285091262e-06, + "loss": 0.4803, + "step": 5609 + }, + { + "epoch": 3.045602605863192, + "grad_norm": 10.73078215472077, + "learning_rate": 6.999604292394271e-06, + "loss": 0.391, + "step": 5610 + }, + { + "epoch": 3.04614549402823, + "grad_norm": 11.713115986580751, + "learning_rate": 6.9962496707910085e-06, + "loss": 0.4674, + "step": 5611 + }, + { + "epoch": 3.046688382193268, + "grad_norm": 11.477819508258428, + "learning_rate": 6.992895420696371e-06, + "loss": 0.4974, + "step": 5612 + }, + { + "epoch": 3.047231270358306, + "grad_norm": 13.625151931379403, + "learning_rate": 6.989541542525218e-06, + "loss": 0.7687, + "step": 5613 + }, + { + "epoch": 3.047774158523344, + "grad_norm": 11.52719950465663, + "learning_rate": 6.986188036692369e-06, + "loss": 0.5544, + "step": 5614 + }, + { + "epoch": 3.048317046688382, + "grad_norm": 13.283507119767767, + "learning_rate": 6.982834903612584e-06, + "loss": 0.6122, + "step": 5615 + }, + { + "epoch": 3.04885993485342, + "grad_norm": 13.11709341600981, + "learning_rate": 6.979482143700591e-06, + "loss": 0.5184, + "step": 5616 + }, + { + "epoch": 3.049402823018458, + "grad_norm": 14.076144429789945, + "learning_rate": 6.976129757371055e-06, + "loss": 0.5594, + "step": 5617 + }, + { + "epoch": 3.049945711183496, + "grad_norm": 13.73222380193425, + "learning_rate": 6.9727777450386124e-06, + "loss": 0.4793, + "step": 5618 + }, + { + "epoch": 3.050488599348534, + "grad_norm": 10.519849404041729, + "learning_rate": 6.969426107117845e-06, + "loss": 0.3503, + "step": 5619 + }, + { + "epoch": 3.0510314875135722, + "grad_norm": 11.598423103761737, + "learning_rate": 6.966074844023284e-06, + "loss": 0.8233, + "step": 5620 + }, + { + "epoch": 3.05157437567861, + "grad_norm": 10.260836656524097, + "learning_rate": 6.9627239561694205e-06, + "loss": 0.4298, + "step": 5621 + }, + { + "epoch": 3.0521172638436482, + "grad_norm": 12.202759177421122, + "learning_rate": 6.9593734439707024e-06, + "loss": 0.5644, + "step": 5622 + }, + { + "epoch": 3.052660152008686, + "grad_norm": 21.73953611396027, + "learning_rate": 6.956023307841524e-06, + "loss": 0.9458, + "step": 5623 + }, + { + "epoch": 3.0532030401737242, + "grad_norm": 12.609994641445128, + "learning_rate": 6.952673548196229e-06, + "loss": 0.5591, + "step": 5624 + }, + { + "epoch": 3.053745928338762, + "grad_norm": 9.019788767418095, + "learning_rate": 6.949324165449129e-06, + "loss": 0.2823, + "step": 5625 + }, + { + "epoch": 3.0542888165038002, + "grad_norm": 9.597041690728307, + "learning_rate": 6.945975160014476e-06, + "loss": 0.339, + "step": 5626 + }, + { + "epoch": 3.054831704668838, + "grad_norm": 9.413655638291356, + "learning_rate": 6.9426265323064865e-06, + "loss": 0.3789, + "step": 5627 + }, + { + "epoch": 3.0553745928338762, + "grad_norm": 11.15027104482265, + "learning_rate": 6.939278282739317e-06, + "loss": 0.5436, + "step": 5628 + }, + { + "epoch": 3.055917480998914, + "grad_norm": 8.752949901728641, + "learning_rate": 6.935930411727091e-06, + "loss": 0.3561, + "step": 5629 + }, + { + "epoch": 3.0564603691639523, + "grad_norm": 13.044879197154156, + "learning_rate": 6.93258291968387e-06, + "loss": 0.5744, + "step": 5630 + }, + { + "epoch": 3.05700325732899, + "grad_norm": 8.940911398612299, + "learning_rate": 6.929235807023689e-06, + "loss": 0.3882, + "step": 5631 + }, + { + "epoch": 3.0575461454940283, + "grad_norm": 11.140820597123383, + "learning_rate": 6.925889074160512e-06, + "loss": 0.4042, + "step": 5632 + }, + { + "epoch": 3.058089033659066, + "grad_norm": 11.477795419029349, + "learning_rate": 6.9225427215082745e-06, + "loss": 0.531, + "step": 5633 + }, + { + "epoch": 3.0586319218241043, + "grad_norm": 11.606939647718864, + "learning_rate": 6.919196749480861e-06, + "loss": 0.3965, + "step": 5634 + }, + { + "epoch": 3.059174809989142, + "grad_norm": 12.112684503158684, + "learning_rate": 6.915851158492099e-06, + "loss": 0.3518, + "step": 5635 + }, + { + "epoch": 3.0597176981541803, + "grad_norm": 11.996768432451738, + "learning_rate": 6.9125059489557855e-06, + "loss": 0.4507, + "step": 5636 + }, + { + "epoch": 3.060260586319218, + "grad_norm": 11.43143134688868, + "learning_rate": 6.909161121285653e-06, + "loss": 0.3375, + "step": 5637 + }, + { + "epoch": 3.0608034744842563, + "grad_norm": 11.944059447932277, + "learning_rate": 6.9058166758954015e-06, + "loss": 0.5155, + "step": 5638 + }, + { + "epoch": 3.061346362649294, + "grad_norm": 13.048822571970177, + "learning_rate": 6.902472613198672e-06, + "loss": 0.4834, + "step": 5639 + }, + { + "epoch": 3.0618892508143323, + "grad_norm": 12.031888695640616, + "learning_rate": 6.899128933609071e-06, + "loss": 0.5558, + "step": 5640 + }, + { + "epoch": 3.06243213897937, + "grad_norm": 12.352798864366527, + "learning_rate": 6.8957856375401466e-06, + "loss": 0.6299, + "step": 5641 + }, + { + "epoch": 3.0629750271444083, + "grad_norm": 8.205311159067556, + "learning_rate": 6.892442725405396e-06, + "loss": 0.2726, + "step": 5642 + }, + { + "epoch": 3.063517915309446, + "grad_norm": 10.36891977803926, + "learning_rate": 6.8891001976182845e-06, + "loss": 0.3787, + "step": 5643 + }, + { + "epoch": 3.0640608034744843, + "grad_norm": 13.053698798672075, + "learning_rate": 6.88575805459222e-06, + "loss": 0.8244, + "step": 5644 + }, + { + "epoch": 3.064603691639522, + "grad_norm": 9.553838049974857, + "learning_rate": 6.88241629674056e-06, + "loss": 0.3611, + "step": 5645 + }, + { + "epoch": 3.0651465798045603, + "grad_norm": 13.221077748060404, + "learning_rate": 6.879074924476621e-06, + "loss": 0.3718, + "step": 5646 + }, + { + "epoch": 3.065689467969598, + "grad_norm": 13.072535493727939, + "learning_rate": 6.8757339382136735e-06, + "loss": 0.4338, + "step": 5647 + }, + { + "epoch": 3.0662323561346363, + "grad_norm": 13.8353344886429, + "learning_rate": 6.872393338364927e-06, + "loss": 0.4158, + "step": 5648 + }, + { + "epoch": 3.066775244299674, + "grad_norm": 12.713981935614036, + "learning_rate": 6.869053125343561e-06, + "loss": 0.4537, + "step": 5649 + }, + { + "epoch": 3.0673181324647123, + "grad_norm": 9.412074844790709, + "learning_rate": 6.865713299562691e-06, + "loss": 0.3842, + "step": 5650 + }, + { + "epoch": 3.06786102062975, + "grad_norm": 12.671904220941121, + "learning_rate": 6.862373861435397e-06, + "loss": 0.5343, + "step": 5651 + }, + { + "epoch": 3.0684039087947883, + "grad_norm": 8.312202499711958, + "learning_rate": 6.859034811374707e-06, + "loss": 0.4265, + "step": 5652 + }, + { + "epoch": 3.068946796959826, + "grad_norm": 8.89496055231579, + "learning_rate": 6.8556961497935925e-06, + "loss": 0.2747, + "step": 5653 + }, + { + "epoch": 3.0694896851248643, + "grad_norm": 12.131106832543901, + "learning_rate": 6.852357877104995e-06, + "loss": 0.4964, + "step": 5654 + }, + { + "epoch": 3.070032573289902, + "grad_norm": 11.860190856186797, + "learning_rate": 6.849019993721786e-06, + "loss": 0.6045, + "step": 5655 + }, + { + "epoch": 3.0705754614549403, + "grad_norm": 9.678033292318382, + "learning_rate": 6.845682500056811e-06, + "loss": 0.331, + "step": 5656 + }, + { + "epoch": 3.071118349619978, + "grad_norm": 15.602537064720222, + "learning_rate": 6.8423453965228485e-06, + "loss": 0.586, + "step": 5657 + }, + { + "epoch": 3.0716612377850163, + "grad_norm": 14.701485694410227, + "learning_rate": 6.839008683532641e-06, + "loss": 0.4876, + "step": 5658 + }, + { + "epoch": 3.072204125950054, + "grad_norm": 10.535230213477728, + "learning_rate": 6.835672361498875e-06, + "loss": 0.6123, + "step": 5659 + }, + { + "epoch": 3.0727470141150923, + "grad_norm": 11.131327768224557, + "learning_rate": 6.832336430834199e-06, + "loss": 0.6257, + "step": 5660 + }, + { + "epoch": 3.07328990228013, + "grad_norm": 8.624075967779987, + "learning_rate": 6.829000891951202e-06, + "loss": 0.3361, + "step": 5661 + }, + { + "epoch": 3.0738327904451683, + "grad_norm": 13.128895205370725, + "learning_rate": 6.825665745262424e-06, + "loss": 0.5965, + "step": 5662 + }, + { + "epoch": 3.074375678610206, + "grad_norm": 12.64008926373084, + "learning_rate": 6.822330991180368e-06, + "loss": 0.62, + "step": 5663 + }, + { + "epoch": 3.0749185667752443, + "grad_norm": 13.416120809361393, + "learning_rate": 6.8189966301174785e-06, + "loss": 0.6674, + "step": 5664 + }, + { + "epoch": 3.075461454940282, + "grad_norm": 9.203438718678973, + "learning_rate": 6.815662662486158e-06, + "loss": 0.3816, + "step": 5665 + }, + { + "epoch": 3.0760043431053203, + "grad_norm": 13.276570487788803, + "learning_rate": 6.81232908869875e-06, + "loss": 0.8179, + "step": 5666 + }, + { + "epoch": 3.076547231270358, + "grad_norm": 12.280157213200626, + "learning_rate": 6.808995909167566e-06, + "loss": 0.5559, + "step": 5667 + }, + { + "epoch": 3.0770901194353963, + "grad_norm": 10.175753421889684, + "learning_rate": 6.805663124304848e-06, + "loss": 0.6191, + "step": 5668 + }, + { + "epoch": 3.077633007600434, + "grad_norm": 12.523871003768551, + "learning_rate": 6.802330734522813e-06, + "loss": 0.4992, + "step": 5669 + }, + { + "epoch": 3.0781758957654723, + "grad_norm": 9.976975930711898, + "learning_rate": 6.798998740233602e-06, + "loss": 0.4104, + "step": 5670 + }, + { + "epoch": 3.07871878393051, + "grad_norm": 11.57486842983445, + "learning_rate": 6.795667141849333e-06, + "loss": 0.5803, + "step": 5671 + }, + { + "epoch": 3.0792616720955484, + "grad_norm": 13.330648489276534, + "learning_rate": 6.79233593978206e-06, + "loss": 0.5116, + "step": 5672 + }, + { + "epoch": 3.079804560260586, + "grad_norm": 10.454358932554449, + "learning_rate": 6.789005134443785e-06, + "loss": 0.4015, + "step": 5673 + }, + { + "epoch": 3.0803474484256244, + "grad_norm": 10.39956810217567, + "learning_rate": 6.785674726246477e-06, + "loss": 0.5832, + "step": 5674 + }, + { + "epoch": 3.080890336590662, + "grad_norm": 10.388603711886958, + "learning_rate": 6.782344715602038e-06, + "loss": 0.2995, + "step": 5675 + }, + { + "epoch": 3.0814332247557004, + "grad_norm": 14.02200163557337, + "learning_rate": 6.779015102922335e-06, + "loss": 0.5132, + "step": 5676 + }, + { + "epoch": 3.081976112920738, + "grad_norm": 9.590771798908678, + "learning_rate": 6.775685888619174e-06, + "loss": 0.2876, + "step": 5677 + }, + { + "epoch": 3.0825190010857764, + "grad_norm": 15.505233514983502, + "learning_rate": 6.7723570731043275e-06, + "loss": 0.8031, + "step": 5678 + }, + { + "epoch": 3.083061889250814, + "grad_norm": 11.180083592225891, + "learning_rate": 6.769028656789497e-06, + "loss": 0.5965, + "step": 5679 + }, + { + "epoch": 3.0836047774158524, + "grad_norm": 12.733678586969875, + "learning_rate": 6.765700640086356e-06, + "loss": 0.4956, + "step": 5680 + }, + { + "epoch": 3.08414766558089, + "grad_norm": 6.888856085950901, + "learning_rate": 6.762373023406515e-06, + "loss": 0.2367, + "step": 5681 + }, + { + "epoch": 3.0846905537459284, + "grad_norm": 12.334377562502302, + "learning_rate": 6.759045807161532e-06, + "loss": 0.5297, + "step": 5682 + }, + { + "epoch": 3.085233441910966, + "grad_norm": 10.637487438782616, + "learning_rate": 6.755718991762934e-06, + "loss": 0.4714, + "step": 5683 + }, + { + "epoch": 3.0857763300760044, + "grad_norm": 10.636835957652801, + "learning_rate": 6.7523925776221775e-06, + "loss": 0.5023, + "step": 5684 + }, + { + "epoch": 3.086319218241042, + "grad_norm": 9.416617539617917, + "learning_rate": 6.749066565150688e-06, + "loss": 0.4012, + "step": 5685 + }, + { + "epoch": 3.0868621064060804, + "grad_norm": 12.384873763197831, + "learning_rate": 6.745740954759823e-06, + "loss": 0.4588, + "step": 5686 + }, + { + "epoch": 3.087404994571118, + "grad_norm": 16.770871731703775, + "learning_rate": 6.7424157468609075e-06, + "loss": 0.8317, + "step": 5687 + }, + { + "epoch": 3.0879478827361564, + "grad_norm": 8.972498742892359, + "learning_rate": 6.7390909418652e-06, + "loss": 0.376, + "step": 5688 + }, + { + "epoch": 3.088490770901194, + "grad_norm": 12.96316859851228, + "learning_rate": 6.735766540183923e-06, + "loss": 0.5285, + "step": 5689 + }, + { + "epoch": 3.0890336590662324, + "grad_norm": 15.891256870281978, + "learning_rate": 6.732442542228245e-06, + "loss": 0.6191, + "step": 5690 + }, + { + "epoch": 3.08957654723127, + "grad_norm": 17.35289550101855, + "learning_rate": 6.729118948409278e-06, + "loss": 0.6878, + "step": 5691 + }, + { + "epoch": 3.0901194353963084, + "grad_norm": 9.022024416599917, + "learning_rate": 6.7257957591380965e-06, + "loss": 0.417, + "step": 5692 + }, + { + "epoch": 3.090662323561346, + "grad_norm": 14.27557655416772, + "learning_rate": 6.722472974825709e-06, + "loss": 0.5535, + "step": 5693 + }, + { + "epoch": 3.0912052117263844, + "grad_norm": 13.610102273496864, + "learning_rate": 6.7191505958830916e-06, + "loss": 0.542, + "step": 5694 + }, + { + "epoch": 3.091748099891422, + "grad_norm": 9.424378370747705, + "learning_rate": 6.715828622721154e-06, + "loss": 0.3447, + "step": 5695 + }, + { + "epoch": 3.0922909880564604, + "grad_norm": 10.815334224243674, + "learning_rate": 6.712507055750768e-06, + "loss": 0.5058, + "step": 5696 + }, + { + "epoch": 3.092833876221498, + "grad_norm": 9.390605474272599, + "learning_rate": 6.709185895382746e-06, + "loss": 0.3778, + "step": 5697 + }, + { + "epoch": 3.0933767643865364, + "grad_norm": 9.621100751873954, + "learning_rate": 6.705865142027863e-06, + "loss": 0.5128, + "step": 5698 + }, + { + "epoch": 3.093919652551574, + "grad_norm": 10.148509981020817, + "learning_rate": 6.7025447960968236e-06, + "loss": 0.3857, + "step": 5699 + }, + { + "epoch": 3.0944625407166124, + "grad_norm": 12.950847437294314, + "learning_rate": 6.699224858000305e-06, + "loss": 0.5364, + "step": 5700 + }, + { + "epoch": 3.09500542888165, + "grad_norm": 11.169096513096612, + "learning_rate": 6.695905328148914e-06, + "loss": 0.7072, + "step": 5701 + }, + { + "epoch": 3.0955483170466884, + "grad_norm": 9.231874354349005, + "learning_rate": 6.692586206953218e-06, + "loss": 0.5096, + "step": 5702 + }, + { + "epoch": 3.096091205211726, + "grad_norm": 13.629340417763906, + "learning_rate": 6.68926749482373e-06, + "loss": 0.5454, + "step": 5703 + }, + { + "epoch": 3.0966340933767644, + "grad_norm": 12.18004127970413, + "learning_rate": 6.685949192170913e-06, + "loss": 0.4463, + "step": 5704 + }, + { + "epoch": 3.097176981541802, + "grad_norm": 10.488473413353695, + "learning_rate": 6.682631299405187e-06, + "loss": 0.4041, + "step": 5705 + }, + { + "epoch": 3.0977198697068404, + "grad_norm": 9.624411116526867, + "learning_rate": 6.679313816936904e-06, + "loss": 0.5316, + "step": 5706 + }, + { + "epoch": 3.098262757871878, + "grad_norm": 13.907233176818153, + "learning_rate": 6.675996745176385e-06, + "loss": 0.6742, + "step": 5707 + }, + { + "epoch": 3.0988056460369164, + "grad_norm": 9.086606489202428, + "learning_rate": 6.672680084533882e-06, + "loss": 0.4101, + "step": 5708 + }, + { + "epoch": 3.099348534201954, + "grad_norm": 9.283349632521654, + "learning_rate": 6.669363835419611e-06, + "loss": 0.318, + "step": 5709 + }, + { + "epoch": 3.0998914223669924, + "grad_norm": 9.668872732852824, + "learning_rate": 6.666047998243732e-06, + "loss": 0.3352, + "step": 5710 + }, + { + "epoch": 3.1004343105320302, + "grad_norm": 9.343099296154486, + "learning_rate": 6.662732573416345e-06, + "loss": 0.4697, + "step": 5711 + }, + { + "epoch": 3.1009771986970684, + "grad_norm": 11.924198282652524, + "learning_rate": 6.659417561347516e-06, + "loss": 0.5743, + "step": 5712 + }, + { + "epoch": 3.1015200868621062, + "grad_norm": 8.560855249443614, + "learning_rate": 6.656102962447244e-06, + "loss": 0.3406, + "step": 5713 + }, + { + "epoch": 3.1020629750271445, + "grad_norm": 14.378777363618713, + "learning_rate": 6.652788777125488e-06, + "loss": 0.6911, + "step": 5714 + }, + { + "epoch": 3.1026058631921822, + "grad_norm": 10.29068899101719, + "learning_rate": 6.649475005792152e-06, + "loss": 0.4449, + "step": 5715 + }, + { + "epoch": 3.1031487513572205, + "grad_norm": 14.02995967568585, + "learning_rate": 6.646161648857085e-06, + "loss": 0.6167, + "step": 5716 + }, + { + "epoch": 3.1036916395222582, + "grad_norm": 9.518238552951749, + "learning_rate": 6.642848706730092e-06, + "loss": 0.3222, + "step": 5717 + }, + { + "epoch": 3.1042345276872965, + "grad_norm": 11.82162879354163, + "learning_rate": 6.639536179820922e-06, + "loss": 0.541, + "step": 5718 + }, + { + "epoch": 3.1047774158523342, + "grad_norm": 10.063084650592682, + "learning_rate": 6.636224068539271e-06, + "loss": 0.3842, + "step": 5719 + }, + { + "epoch": 3.1053203040173725, + "grad_norm": 10.94706260488087, + "learning_rate": 6.632912373294792e-06, + "loss": 0.4513, + "step": 5720 + }, + { + "epoch": 3.1058631921824102, + "grad_norm": 18.39843990173123, + "learning_rate": 6.6296010944970745e-06, + "loss": 0.6059, + "step": 5721 + }, + { + "epoch": 3.1064060803474485, + "grad_norm": 12.609614695754882, + "learning_rate": 6.6262902325556635e-06, + "loss": 0.5504, + "step": 5722 + }, + { + "epoch": 3.1069489685124863, + "grad_norm": 10.109861615464935, + "learning_rate": 6.6229797878800575e-06, + "loss": 0.4609, + "step": 5723 + }, + { + "epoch": 3.1074918566775245, + "grad_norm": 10.309200757998537, + "learning_rate": 6.61966976087969e-06, + "loss": 0.3823, + "step": 5724 + }, + { + "epoch": 3.1080347448425623, + "grad_norm": 13.21038451142413, + "learning_rate": 6.616360151963956e-06, + "loss": 1.0012, + "step": 5725 + }, + { + "epoch": 3.1085776330076005, + "grad_norm": 10.685970514335464, + "learning_rate": 6.613050961542189e-06, + "loss": 0.4587, + "step": 5726 + }, + { + "epoch": 3.1091205211726383, + "grad_norm": 10.001481339897822, + "learning_rate": 6.60974219002368e-06, + "loss": 0.4954, + "step": 5727 + }, + { + "epoch": 3.1096634093376765, + "grad_norm": 10.414474916866299, + "learning_rate": 6.6064338378176585e-06, + "loss": 0.5087, + "step": 5728 + }, + { + "epoch": 3.1102062975027143, + "grad_norm": 12.456658351321629, + "learning_rate": 6.603125905333307e-06, + "loss": 0.5004, + "step": 5729 + }, + { + "epoch": 3.1107491856677525, + "grad_norm": 11.228566931054628, + "learning_rate": 6.599818392979762e-06, + "loss": 0.6843, + "step": 5730 + }, + { + "epoch": 3.1112920738327903, + "grad_norm": 11.262471951903663, + "learning_rate": 6.596511301166093e-06, + "loss": 0.4458, + "step": 5731 + }, + { + "epoch": 3.1118349619978285, + "grad_norm": 14.452306212538934, + "learning_rate": 6.593204630301333e-06, + "loss": 0.4319, + "step": 5732 + }, + { + "epoch": 3.1123778501628663, + "grad_norm": 11.527781393494147, + "learning_rate": 6.589898380794451e-06, + "loss": 0.6469, + "step": 5733 + }, + { + "epoch": 3.1129207383279045, + "grad_norm": 13.390783514445054, + "learning_rate": 6.586592553054374e-06, + "loss": 0.5311, + "step": 5734 + }, + { + "epoch": 3.1134636264929423, + "grad_norm": 9.033138513383328, + "learning_rate": 6.583287147489968e-06, + "loss": 0.4243, + "step": 5735 + }, + { + "epoch": 3.1140065146579805, + "grad_norm": 8.69133142787827, + "learning_rate": 6.579982164510057e-06, + "loss": 0.3874, + "step": 5736 + }, + { + "epoch": 3.1145494028230183, + "grad_norm": 8.998980230350567, + "learning_rate": 6.576677604523399e-06, + "loss": 0.3375, + "step": 5737 + }, + { + "epoch": 3.1150922909880565, + "grad_norm": 13.718371009849239, + "learning_rate": 6.573373467938715e-06, + "loss": 0.8846, + "step": 5738 + }, + { + "epoch": 3.1156351791530943, + "grad_norm": 9.65489128269965, + "learning_rate": 6.5700697551646595e-06, + "loss": 0.5469, + "step": 5739 + }, + { + "epoch": 3.1161780673181325, + "grad_norm": 12.925637209034505, + "learning_rate": 6.566766466609844e-06, + "loss": 0.4983, + "step": 5740 + }, + { + "epoch": 3.1167209554831703, + "grad_norm": 8.77821937172237, + "learning_rate": 6.563463602682823e-06, + "loss": 0.3698, + "step": 5741 + }, + { + "epoch": 3.1172638436482085, + "grad_norm": 9.642465181954359, + "learning_rate": 6.560161163792098e-06, + "loss": 0.3764, + "step": 5742 + }, + { + "epoch": 3.1178067318132463, + "grad_norm": 13.798683831974408, + "learning_rate": 6.5568591503461266e-06, + "loss": 0.6993, + "step": 5743 + }, + { + "epoch": 3.1183496199782845, + "grad_norm": 9.872788283771353, + "learning_rate": 6.553557562753299e-06, + "loss": 0.3386, + "step": 5744 + }, + { + "epoch": 3.1188925081433223, + "grad_norm": 8.862249503192198, + "learning_rate": 6.550256401421969e-06, + "loss": 0.2179, + "step": 5745 + }, + { + "epoch": 3.1194353963083605, + "grad_norm": 12.08451508737181, + "learning_rate": 6.546955666760421e-06, + "loss": 0.6611, + "step": 5746 + }, + { + "epoch": 3.1199782844733983, + "grad_norm": 12.735515016235034, + "learning_rate": 6.543655359176904e-06, + "loss": 0.497, + "step": 5747 + }, + { + "epoch": 3.1205211726384365, + "grad_norm": 11.84865199124047, + "learning_rate": 6.5403554790795974e-06, + "loss": 0.7862, + "step": 5748 + }, + { + "epoch": 3.1210640608034743, + "grad_norm": 10.714296855189131, + "learning_rate": 6.53705602687664e-06, + "loss": 0.4053, + "step": 5749 + }, + { + "epoch": 3.1216069489685125, + "grad_norm": 13.603289352492634, + "learning_rate": 6.533757002976115e-06, + "loss": 0.8514, + "step": 5750 + }, + { + "epoch": 3.1221498371335503, + "grad_norm": 13.12224567132668, + "learning_rate": 6.530458407786045e-06, + "loss": 0.5521, + "step": 5751 + }, + { + "epoch": 3.1226927252985885, + "grad_norm": 12.175088945770037, + "learning_rate": 6.527160241714412e-06, + "loss": 0.5578, + "step": 5752 + }, + { + "epoch": 3.1232356134636263, + "grad_norm": 8.84737172176379, + "learning_rate": 6.52386250516913e-06, + "loss": 0.628, + "step": 5753 + }, + { + "epoch": 3.1237785016286646, + "grad_norm": 9.436585430499584, + "learning_rate": 6.520565198558076e-06, + "loss": 0.3497, + "step": 5754 + }, + { + "epoch": 3.1243213897937023, + "grad_norm": 19.037012868331093, + "learning_rate": 6.517268322289063e-06, + "loss": 0.7797, + "step": 5755 + }, + { + "epoch": 3.1248642779587406, + "grad_norm": 10.13012645392841, + "learning_rate": 6.513971876769857e-06, + "loss": 0.4892, + "step": 5756 + }, + { + "epoch": 3.1254071661237783, + "grad_norm": 12.402167588058873, + "learning_rate": 6.510675862408162e-06, + "loss": 0.4358, + "step": 5757 + }, + { + "epoch": 3.1259500542888166, + "grad_norm": 13.090215605994585, + "learning_rate": 6.5073802796116415e-06, + "loss": 0.4757, + "step": 5758 + }, + { + "epoch": 3.1264929424538543, + "grad_norm": 12.046620866150537, + "learning_rate": 6.504085128787892e-06, + "loss": 0.6104, + "step": 5759 + }, + { + "epoch": 3.1270358306188926, + "grad_norm": 13.860233354070779, + "learning_rate": 6.500790410344463e-06, + "loss": 0.5933, + "step": 5760 + }, + { + "epoch": 3.1275787187839303, + "grad_norm": 10.049188091595793, + "learning_rate": 6.497496124688858e-06, + "loss": 0.294, + "step": 5761 + }, + { + "epoch": 3.1281216069489686, + "grad_norm": 14.96534379979502, + "learning_rate": 6.494202272228509e-06, + "loss": 0.5983, + "step": 5762 + }, + { + "epoch": 3.1286644951140063, + "grad_norm": 12.09845857427764, + "learning_rate": 6.490908853370816e-06, + "loss": 0.6995, + "step": 5763 + }, + { + "epoch": 3.1292073832790446, + "grad_norm": 11.420498115973434, + "learning_rate": 6.487615868523105e-06, + "loss": 0.5189, + "step": 5764 + }, + { + "epoch": 3.1297502714440824, + "grad_norm": 9.433130521020503, + "learning_rate": 6.4843233180926646e-06, + "loss": 0.5145, + "step": 5765 + }, + { + "epoch": 3.1302931596091206, + "grad_norm": 13.10906497158192, + "learning_rate": 6.481031202486714e-06, + "loss": 0.4662, + "step": 5766 + }, + { + "epoch": 3.1308360477741584, + "grad_norm": 11.61717934785526, + "learning_rate": 6.477739522112436e-06, + "loss": 0.4496, + "step": 5767 + }, + { + "epoch": 3.1313789359391966, + "grad_norm": 15.125919718339924, + "learning_rate": 6.474448277376946e-06, + "loss": 0.6205, + "step": 5768 + }, + { + "epoch": 3.1319218241042344, + "grad_norm": 11.779008588612678, + "learning_rate": 6.471157468687315e-06, + "loss": 0.337, + "step": 5769 + }, + { + "epoch": 3.1324647122692726, + "grad_norm": 11.960250814404343, + "learning_rate": 6.467867096450553e-06, + "loss": 0.6284, + "step": 5770 + }, + { + "epoch": 3.1330076004343104, + "grad_norm": 11.356052401682772, + "learning_rate": 6.464577161073613e-06, + "loss": 0.5146, + "step": 5771 + }, + { + "epoch": 3.1335504885993486, + "grad_norm": 12.23671034812476, + "learning_rate": 6.461287662963407e-06, + "loss": 0.4118, + "step": 5772 + }, + { + "epoch": 3.1340933767643864, + "grad_norm": 9.3897356425462, + "learning_rate": 6.457998602526783e-06, + "loss": 0.3312, + "step": 5773 + }, + { + "epoch": 3.1346362649294246, + "grad_norm": 8.285183279593793, + "learning_rate": 6.454709980170537e-06, + "loss": 0.3857, + "step": 5774 + }, + { + "epoch": 3.1351791530944624, + "grad_norm": 10.70178891805582, + "learning_rate": 6.451421796301407e-06, + "loss": 0.4596, + "step": 5775 + }, + { + "epoch": 3.1357220412595006, + "grad_norm": 9.966009510827126, + "learning_rate": 6.448134051326092e-06, + "loss": 0.5223, + "step": 5776 + }, + { + "epoch": 3.1362649294245384, + "grad_norm": 11.165316341245944, + "learning_rate": 6.444846745651212e-06, + "loss": 0.482, + "step": 5777 + }, + { + "epoch": 3.1368078175895766, + "grad_norm": 10.091328730281877, + "learning_rate": 6.441559879683357e-06, + "loss": 0.4232, + "step": 5778 + }, + { + "epoch": 3.1373507057546144, + "grad_norm": 9.885329689068836, + "learning_rate": 6.438273453829047e-06, + "loss": 0.3008, + "step": 5779 + }, + { + "epoch": 3.1378935939196526, + "grad_norm": 11.500516853477965, + "learning_rate": 6.434987468494749e-06, + "loss": 0.5076, + "step": 5780 + }, + { + "epoch": 3.1384364820846904, + "grad_norm": 11.00413836223355, + "learning_rate": 6.431701924086887e-06, + "loss": 0.3688, + "step": 5781 + }, + { + "epoch": 3.1389793702497286, + "grad_norm": 11.998875560827074, + "learning_rate": 6.428416821011814e-06, + "loss": 0.6377, + "step": 5782 + }, + { + "epoch": 3.1395222584147664, + "grad_norm": 13.007472970217544, + "learning_rate": 6.425132159675845e-06, + "loss": 0.5721, + "step": 5783 + }, + { + "epoch": 3.1400651465798046, + "grad_norm": 11.356611803430905, + "learning_rate": 6.4218479404852244e-06, + "loss": 0.3753, + "step": 5784 + }, + { + "epoch": 3.1406080347448424, + "grad_norm": 12.12920282380304, + "learning_rate": 6.418564163846155e-06, + "loss": 0.4842, + "step": 5785 + }, + { + "epoch": 3.1411509229098806, + "grad_norm": 10.041071912027489, + "learning_rate": 6.415280830164779e-06, + "loss": 0.4939, + "step": 5786 + }, + { + "epoch": 3.1416938110749184, + "grad_norm": 12.998407630945819, + "learning_rate": 6.411997939847181e-06, + "loss": 0.636, + "step": 5787 + }, + { + "epoch": 3.1422366992399566, + "grad_norm": 9.655708839558747, + "learning_rate": 6.408715493299398e-06, + "loss": 0.4262, + "step": 5788 + }, + { + "epoch": 3.1427795874049944, + "grad_norm": 18.92187273336628, + "learning_rate": 6.405433490927404e-06, + "loss": 0.6645, + "step": 5789 + }, + { + "epoch": 3.1433224755700326, + "grad_norm": 7.288861789816135, + "learning_rate": 6.402151933137128e-06, + "loss": 0.3247, + "step": 5790 + }, + { + "epoch": 3.1438653637350704, + "grad_norm": 15.171324384191708, + "learning_rate": 6.398870820334431e-06, + "loss": 0.5835, + "step": 5791 + }, + { + "epoch": 3.1444082519001086, + "grad_norm": 19.537893777194515, + "learning_rate": 6.395590152925133e-06, + "loss": 0.4794, + "step": 5792 + }, + { + "epoch": 3.1449511400651464, + "grad_norm": 15.604738367299253, + "learning_rate": 6.3923099313149865e-06, + "loss": 0.8757, + "step": 5793 + }, + { + "epoch": 3.1454940282301846, + "grad_norm": 11.401596673156138, + "learning_rate": 6.389030155909702e-06, + "loss": 0.5278, + "step": 5794 + }, + { + "epoch": 3.1460369163952224, + "grad_norm": 15.026441679185012, + "learning_rate": 6.385750827114919e-06, + "loss": 0.7473, + "step": 5795 + }, + { + "epoch": 3.1465798045602607, + "grad_norm": 10.75641281038809, + "learning_rate": 6.382471945336238e-06, + "loss": 0.7022, + "step": 5796 + }, + { + "epoch": 3.1471226927252984, + "grad_norm": 13.14436770595673, + "learning_rate": 6.379193510979188e-06, + "loss": 0.5036, + "step": 5797 + }, + { + "epoch": 3.1476655808903367, + "grad_norm": 12.687239439350824, + "learning_rate": 6.375915524449259e-06, + "loss": 0.7799, + "step": 5798 + }, + { + "epoch": 3.1482084690553744, + "grad_norm": 12.06538329633808, + "learning_rate": 6.372637986151873e-06, + "loss": 0.6135, + "step": 5799 + }, + { + "epoch": 3.1487513572204127, + "grad_norm": 14.280970019829384, + "learning_rate": 6.369360896492398e-06, + "loss": 0.5631, + "step": 5800 + }, + { + "epoch": 3.1492942453854504, + "grad_norm": 13.49612526729136, + "learning_rate": 6.366084255876159e-06, + "loss": 0.8147, + "step": 5801 + }, + { + "epoch": 3.1498371335504887, + "grad_norm": 16.44633506143303, + "learning_rate": 6.3628080647084074e-06, + "loss": 0.705, + "step": 5802 + }, + { + "epoch": 3.1503800217155264, + "grad_norm": 11.659496492366037, + "learning_rate": 6.359532323394355e-06, + "loss": 0.4872, + "step": 5803 + }, + { + "epoch": 3.1509229098805647, + "grad_norm": 9.20526792646835, + "learning_rate": 6.356257032339141e-06, + "loss": 0.3279, + "step": 5804 + }, + { + "epoch": 3.1514657980456025, + "grad_norm": 15.869195168975136, + "learning_rate": 6.35298219194787e-06, + "loss": 0.6893, + "step": 5805 + }, + { + "epoch": 3.1520086862106407, + "grad_norm": 12.918024481861224, + "learning_rate": 6.349707802625569e-06, + "loss": 0.3561, + "step": 5806 + }, + { + "epoch": 3.1525515743756785, + "grad_norm": 11.948249403281372, + "learning_rate": 6.34643386477723e-06, + "loss": 0.6678, + "step": 5807 + }, + { + "epoch": 3.1530944625407167, + "grad_norm": 10.53882181570608, + "learning_rate": 6.3431603788077736e-06, + "loss": 0.2843, + "step": 5808 + }, + { + "epoch": 3.1536373507057545, + "grad_norm": 14.890522782673512, + "learning_rate": 6.3398873451220644e-06, + "loss": 0.7516, + "step": 5809 + }, + { + "epoch": 3.1541802388707927, + "grad_norm": 13.860961190955267, + "learning_rate": 6.3366147641249266e-06, + "loss": 0.697, + "step": 5810 + }, + { + "epoch": 3.1547231270358305, + "grad_norm": 13.048567845980694, + "learning_rate": 6.3333426362211096e-06, + "loss": 0.5362, + "step": 5811 + }, + { + "epoch": 3.1552660152008687, + "grad_norm": 11.430913085361082, + "learning_rate": 6.33007096181532e-06, + "loss": 0.3698, + "step": 5812 + }, + { + "epoch": 3.1558089033659065, + "grad_norm": 13.799412170578423, + "learning_rate": 6.326799741312202e-06, + "loss": 0.579, + "step": 5813 + }, + { + "epoch": 3.1563517915309447, + "grad_norm": 12.205698077175509, + "learning_rate": 6.323528975116349e-06, + "loss": 0.6956, + "step": 5814 + }, + { + "epoch": 3.1568946796959825, + "grad_norm": 14.923158817152583, + "learning_rate": 6.320258663632288e-06, + "loss": 0.7094, + "step": 5815 + }, + { + "epoch": 3.1574375678610207, + "grad_norm": 9.221494784749213, + "learning_rate": 6.316988807264506e-06, + "loss": 0.3346, + "step": 5816 + }, + { + "epoch": 3.1579804560260585, + "grad_norm": 12.263733681763519, + "learning_rate": 6.3137194064174124e-06, + "loss": 0.6395, + "step": 5817 + }, + { + "epoch": 3.1585233441910967, + "grad_norm": 12.198844326189183, + "learning_rate": 6.31045046149538e-06, + "loss": 0.4894, + "step": 5818 + }, + { + "epoch": 3.1590662323561345, + "grad_norm": 11.16927244530179, + "learning_rate": 6.3071819729027175e-06, + "loss": 0.4653, + "step": 5819 + }, + { + "epoch": 3.1596091205211727, + "grad_norm": 11.182448096771813, + "learning_rate": 6.303913941043669e-06, + "loss": 0.3997, + "step": 5820 + }, + { + "epoch": 3.1601520086862105, + "grad_norm": 9.56488246617197, + "learning_rate": 6.300646366322439e-06, + "loss": 0.6558, + "step": 5821 + }, + { + "epoch": 3.1606948968512487, + "grad_norm": 12.485736110882538, + "learning_rate": 6.297379249143158e-06, + "loss": 0.2747, + "step": 5822 + }, + { + "epoch": 3.1612377850162865, + "grad_norm": 10.29008763858223, + "learning_rate": 6.294112589909919e-06, + "loss": 0.3593, + "step": 5823 + }, + { + "epoch": 3.1617806731813247, + "grad_norm": 10.281965116706075, + "learning_rate": 6.2908463890267345e-06, + "loss": 0.4736, + "step": 5824 + }, + { + "epoch": 3.1623235613463625, + "grad_norm": 9.464092471296539, + "learning_rate": 6.287580646897584e-06, + "loss": 0.4445, + "step": 5825 + }, + { + "epoch": 3.1628664495114007, + "grad_norm": 17.834508294283445, + "learning_rate": 6.284315363926372e-06, + "loss": 0.7063, + "step": 5826 + }, + { + "epoch": 3.1634093376764385, + "grad_norm": 13.155791246426181, + "learning_rate": 6.281050540516963e-06, + "loss": 0.5295, + "step": 5827 + }, + { + "epoch": 3.1639522258414767, + "grad_norm": 11.87732424107669, + "learning_rate": 6.277786177073151e-06, + "loss": 0.6095, + "step": 5828 + }, + { + "epoch": 3.1644951140065145, + "grad_norm": 9.937871872571362, + "learning_rate": 6.274522273998671e-06, + "loss": 0.3788, + "step": 5829 + }, + { + "epoch": 3.1650380021715527, + "grad_norm": 10.26395895690379, + "learning_rate": 6.2712588316972155e-06, + "loss": 0.4352, + "step": 5830 + }, + { + "epoch": 3.1655808903365905, + "grad_norm": 8.627642026627477, + "learning_rate": 6.26799585057241e-06, + "loss": 0.3298, + "step": 5831 + }, + { + "epoch": 3.1661237785016287, + "grad_norm": 12.940542818393178, + "learning_rate": 6.2647333310278295e-06, + "loss": 0.5464, + "step": 5832 + }, + { + "epoch": 3.1666666666666665, + "grad_norm": 11.72079573230472, + "learning_rate": 6.261471273466979e-06, + "loss": 0.3785, + "step": 5833 + }, + { + "epoch": 3.1672095548317047, + "grad_norm": 13.338454285585954, + "learning_rate": 6.258209678293324e-06, + "loss": 0.4455, + "step": 5834 + }, + { + "epoch": 3.1677524429967425, + "grad_norm": 10.755191139797633, + "learning_rate": 6.254948545910256e-06, + "loss": 0.492, + "step": 5835 + }, + { + "epoch": 3.1682953311617807, + "grad_norm": 19.472452158150972, + "learning_rate": 6.251687876721124e-06, + "loss": 0.5031, + "step": 5836 + }, + { + "epoch": 3.1688382193268185, + "grad_norm": 17.265109190709392, + "learning_rate": 6.248427671129207e-06, + "loss": 0.7912, + "step": 5837 + }, + { + "epoch": 3.1693811074918568, + "grad_norm": 15.65566631740727, + "learning_rate": 6.245167929537733e-06, + "loss": 0.5431, + "step": 5838 + }, + { + "epoch": 3.1699239956568945, + "grad_norm": 11.65166282725162, + "learning_rate": 6.241908652349879e-06, + "loss": 0.3528, + "step": 5839 + }, + { + "epoch": 3.1704668838219328, + "grad_norm": 10.29071410809139, + "learning_rate": 6.238649839968746e-06, + "loss": 0.314, + "step": 5840 + }, + { + "epoch": 3.1710097719869705, + "grad_norm": 8.17516522370587, + "learning_rate": 6.2353914927974024e-06, + "loss": 0.3906, + "step": 5841 + }, + { + "epoch": 3.1715526601520088, + "grad_norm": 10.555776499314874, + "learning_rate": 6.232133611238833e-06, + "loss": 0.4222, + "step": 5842 + }, + { + "epoch": 3.1720955483170465, + "grad_norm": 15.841225391604809, + "learning_rate": 6.228876195695986e-06, + "loss": 0.8112, + "step": 5843 + }, + { + "epoch": 3.1726384364820848, + "grad_norm": 12.430872813466289, + "learning_rate": 6.225619246571741e-06, + "loss": 0.5456, + "step": 5844 + }, + { + "epoch": 3.1731813246471225, + "grad_norm": 13.424033486912041, + "learning_rate": 6.222362764268924e-06, + "loss": 0.5942, + "step": 5845 + }, + { + "epoch": 3.1737242128121608, + "grad_norm": 13.261228349891576, + "learning_rate": 6.219106749190298e-06, + "loss": 0.6697, + "step": 5846 + }, + { + "epoch": 3.1742671009771986, + "grad_norm": 12.710362718713364, + "learning_rate": 6.215851201738578e-06, + "loss": 0.5753, + "step": 5847 + }, + { + "epoch": 3.1748099891422368, + "grad_norm": 14.841367294849315, + "learning_rate": 6.212596122316415e-06, + "loss": 0.965, + "step": 5848 + }, + { + "epoch": 3.1753528773072746, + "grad_norm": 13.885344637636523, + "learning_rate": 6.209341511326394e-06, + "loss": 0.5931, + "step": 5849 + }, + { + "epoch": 3.175895765472313, + "grad_norm": 11.460331492201762, + "learning_rate": 6.20608736917106e-06, + "loss": 0.5989, + "step": 5850 + }, + { + "epoch": 3.1764386536373506, + "grad_norm": 11.70057563506978, + "learning_rate": 6.202833696252884e-06, + "loss": 0.4908, + "step": 5851 + }, + { + "epoch": 3.176981541802389, + "grad_norm": 10.645588299206919, + "learning_rate": 6.199580492974291e-06, + "loss": 0.4098, + "step": 5852 + }, + { + "epoch": 3.1775244299674266, + "grad_norm": 12.755952884154276, + "learning_rate": 6.196327759737637e-06, + "loss": 0.6992, + "step": 5853 + }, + { + "epoch": 3.178067318132465, + "grad_norm": 14.709812641950414, + "learning_rate": 6.193075496945231e-06, + "loss": 0.6841, + "step": 5854 + }, + { + "epoch": 3.1786102062975026, + "grad_norm": 13.875280510305055, + "learning_rate": 6.189823704999312e-06, + "loss": 0.7106, + "step": 5855 + }, + { + "epoch": 3.179153094462541, + "grad_norm": 9.281312738229682, + "learning_rate": 6.186572384302072e-06, + "loss": 0.3546, + "step": 5856 + }, + { + "epoch": 3.1796959826275786, + "grad_norm": 12.958826598950415, + "learning_rate": 6.1833215352556375e-06, + "loss": 0.6019, + "step": 5857 + }, + { + "epoch": 3.180238870792617, + "grad_norm": 11.464928690074025, + "learning_rate": 6.180071158262075e-06, + "loss": 0.408, + "step": 5858 + }, + { + "epoch": 3.1807817589576546, + "grad_norm": 9.768529641559107, + "learning_rate": 6.176821253723404e-06, + "loss": 0.4072, + "step": 5859 + }, + { + "epoch": 3.181324647122693, + "grad_norm": 9.678665348787685, + "learning_rate": 6.173571822041568e-06, + "loss": 0.4732, + "step": 5860 + }, + { + "epoch": 3.1818675352877306, + "grad_norm": 14.30027690416035, + "learning_rate": 6.170322863618474e-06, + "loss": 0.7642, + "step": 5861 + }, + { + "epoch": 3.182410423452769, + "grad_norm": 14.961370236846983, + "learning_rate": 6.167074378855946e-06, + "loss": 0.6362, + "step": 5862 + }, + { + "epoch": 3.1829533116178066, + "grad_norm": 9.357640920276147, + "learning_rate": 6.1638263681557695e-06, + "loss": 0.4049, + "step": 5863 + }, + { + "epoch": 3.183496199782845, + "grad_norm": 10.782208269160673, + "learning_rate": 6.160578831919662e-06, + "loss": 0.336, + "step": 5864 + }, + { + "epoch": 3.1840390879478826, + "grad_norm": 14.384452607422604, + "learning_rate": 6.157331770549285e-06, + "loss": 0.5396, + "step": 5865 + }, + { + "epoch": 3.184581976112921, + "grad_norm": 9.250100344453772, + "learning_rate": 6.154085184446237e-06, + "loss": 0.3785, + "step": 5866 + }, + { + "epoch": 3.1851248642779586, + "grad_norm": 15.376158526777193, + "learning_rate": 6.150839074012065e-06, + "loss": 0.5189, + "step": 5867 + }, + { + "epoch": 3.185667752442997, + "grad_norm": 10.177494395248477, + "learning_rate": 6.14759343964825e-06, + "loss": 0.3756, + "step": 5868 + }, + { + "epoch": 3.1862106406080346, + "grad_norm": 17.67659076253852, + "learning_rate": 6.144348281756218e-06, + "loss": 0.7381, + "step": 5869 + }, + { + "epoch": 3.186753528773073, + "grad_norm": 11.56583944116506, + "learning_rate": 6.141103600737337e-06, + "loss": 0.7712, + "step": 5870 + }, + { + "epoch": 3.1872964169381106, + "grad_norm": 15.096190908138054, + "learning_rate": 6.137859396992909e-06, + "loss": 0.6608, + "step": 5871 + }, + { + "epoch": 3.187839305103149, + "grad_norm": 14.055042562247136, + "learning_rate": 6.134615670924191e-06, + "loss": 0.6762, + "step": 5872 + }, + { + "epoch": 3.1883821932681866, + "grad_norm": 12.50506943616142, + "learning_rate": 6.131372422932363e-06, + "loss": 0.4721, + "step": 5873 + }, + { + "epoch": 3.188925081433225, + "grad_norm": 12.679042521649755, + "learning_rate": 6.128129653418562e-06, + "loss": 0.4277, + "step": 5874 + }, + { + "epoch": 3.1894679695982626, + "grad_norm": 13.177050387778413, + "learning_rate": 6.124887362783854e-06, + "loss": 0.7357, + "step": 5875 + }, + { + "epoch": 3.190010857763301, + "grad_norm": 9.589104761627354, + "learning_rate": 6.121645551429255e-06, + "loss": 0.3541, + "step": 5876 + }, + { + "epoch": 3.1905537459283386, + "grad_norm": 15.986950102257723, + "learning_rate": 6.118404219755717e-06, + "loss": 0.5148, + "step": 5877 + }, + { + "epoch": 3.191096634093377, + "grad_norm": 14.855627037048395, + "learning_rate": 6.1151633681641275e-06, + "loss": 0.6874, + "step": 5878 + }, + { + "epoch": 3.1916395222584146, + "grad_norm": 9.68244350868187, + "learning_rate": 6.111922997055328e-06, + "loss": 0.2941, + "step": 5879 + }, + { + "epoch": 3.192182410423453, + "grad_norm": 10.654011696718703, + "learning_rate": 6.108683106830085e-06, + "loss": 0.5009, + "step": 5880 + }, + { + "epoch": 3.1927252985884906, + "grad_norm": 12.855166875044116, + "learning_rate": 6.105443697889121e-06, + "loss": 0.4806, + "step": 5881 + }, + { + "epoch": 3.193268186753529, + "grad_norm": 10.436692542811286, + "learning_rate": 6.102204770633083e-06, + "loss": 0.5723, + "step": 5882 + }, + { + "epoch": 3.1938110749185666, + "grad_norm": 9.990036928653613, + "learning_rate": 6.098966325462574e-06, + "loss": 0.5778, + "step": 5883 + }, + { + "epoch": 3.194353963083605, + "grad_norm": 12.046567955894364, + "learning_rate": 6.095728362778125e-06, + "loss": 0.5015, + "step": 5884 + }, + { + "epoch": 3.1948968512486426, + "grad_norm": 8.725155173492308, + "learning_rate": 6.092490882980219e-06, + "loss": 0.3857, + "step": 5885 + }, + { + "epoch": 3.195439739413681, + "grad_norm": 11.013208621693263, + "learning_rate": 6.089253886469267e-06, + "loss": 0.6389, + "step": 5886 + }, + { + "epoch": 3.1959826275787186, + "grad_norm": 12.794704970363892, + "learning_rate": 6.086017373645624e-06, + "loss": 0.6362, + "step": 5887 + }, + { + "epoch": 3.196525515743757, + "grad_norm": 10.413523492938069, + "learning_rate": 6.082781344909592e-06, + "loss": 0.477, + "step": 5888 + }, + { + "epoch": 3.1970684039087947, + "grad_norm": 19.28502803064406, + "learning_rate": 6.079545800661404e-06, + "loss": 0.6435, + "step": 5889 + }, + { + "epoch": 3.197611292073833, + "grad_norm": 12.8419010375632, + "learning_rate": 6.076310741301244e-06, + "loss": 0.816, + "step": 5890 + }, + { + "epoch": 3.1981541802388707, + "grad_norm": 13.169440134578803, + "learning_rate": 6.0730761672292215e-06, + "loss": 0.4307, + "step": 5891 + }, + { + "epoch": 3.198697068403909, + "grad_norm": 12.487488215101045, + "learning_rate": 6.0698420788454e-06, + "loss": 0.6785, + "step": 5892 + }, + { + "epoch": 3.1992399565689467, + "grad_norm": 13.786560936789252, + "learning_rate": 6.066608476549771e-06, + "loss": 0.866, + "step": 5893 + }, + { + "epoch": 3.199782844733985, + "grad_norm": 14.699571921737691, + "learning_rate": 6.0633753607422785e-06, + "loss": 0.682, + "step": 5894 + }, + { + "epoch": 3.2003257328990227, + "grad_norm": 10.298137592669294, + "learning_rate": 6.060142731822792e-06, + "loss": 0.4765, + "step": 5895 + }, + { + "epoch": 3.200868621064061, + "grad_norm": 11.57501814659113, + "learning_rate": 6.056910590191132e-06, + "loss": 0.4907, + "step": 5896 + }, + { + "epoch": 3.2014115092290987, + "grad_norm": 10.861108561617078, + "learning_rate": 6.0536789362470575e-06, + "loss": 0.4201, + "step": 5897 + }, + { + "epoch": 3.201954397394137, + "grad_norm": 14.726234720260377, + "learning_rate": 6.050447770390258e-06, + "loss": 0.5714, + "step": 5898 + }, + { + "epoch": 3.2024972855591747, + "grad_norm": 10.353294495922004, + "learning_rate": 6.047217093020374e-06, + "loss": 0.3656, + "step": 5899 + }, + { + "epoch": 3.203040173724213, + "grad_norm": 7.811262042752878, + "learning_rate": 6.043986904536979e-06, + "loss": 0.3013, + "step": 5900 + }, + { + "epoch": 3.2035830618892507, + "grad_norm": 9.961324210163603, + "learning_rate": 6.040757205339589e-06, + "loss": 0.3498, + "step": 5901 + }, + { + "epoch": 3.204125950054289, + "grad_norm": 11.897050086158284, + "learning_rate": 6.037527995827659e-06, + "loss": 0.4876, + "step": 5902 + }, + { + "epoch": 3.2046688382193267, + "grad_norm": 10.075300527690798, + "learning_rate": 6.034299276400582e-06, + "loss": 0.3781, + "step": 5903 + }, + { + "epoch": 3.205211726384365, + "grad_norm": 20.471624027064, + "learning_rate": 6.031071047457689e-06, + "loss": 0.7781, + "step": 5904 + }, + { + "epoch": 3.2057546145494027, + "grad_norm": 13.759518261461583, + "learning_rate": 6.027843309398256e-06, + "loss": 0.6616, + "step": 5905 + }, + { + "epoch": 3.206297502714441, + "grad_norm": 11.734095364989608, + "learning_rate": 6.024616062621496e-06, + "loss": 0.4283, + "step": 5906 + }, + { + "epoch": 3.2068403908794787, + "grad_norm": 11.862203608845647, + "learning_rate": 6.02138930752655e-06, + "loss": 0.5662, + "step": 5907 + }, + { + "epoch": 3.207383279044517, + "grad_norm": 15.72669110441235, + "learning_rate": 6.018163044512521e-06, + "loss": 0.5579, + "step": 5908 + }, + { + "epoch": 3.2079261672095547, + "grad_norm": 19.31549988156055, + "learning_rate": 6.014937273978429e-06, + "loss": 0.7692, + "step": 5909 + }, + { + "epoch": 3.208469055374593, + "grad_norm": 10.302000372364747, + "learning_rate": 6.011711996323251e-06, + "loss": 0.5133, + "step": 5910 + }, + { + "epoch": 3.2090119435396307, + "grad_norm": 12.224649649591383, + "learning_rate": 6.008487211945884e-06, + "loss": 0.7677, + "step": 5911 + }, + { + "epoch": 3.209554831704669, + "grad_norm": 17.460170685666615, + "learning_rate": 6.005262921245185e-06, + "loss": 0.6836, + "step": 5912 + }, + { + "epoch": 3.2100977198697067, + "grad_norm": 19.301567461364495, + "learning_rate": 6.00203912461993e-06, + "loss": 0.9095, + "step": 5913 + }, + { + "epoch": 3.210640608034745, + "grad_norm": 10.852675563314923, + "learning_rate": 5.998815822468851e-06, + "loss": 0.5342, + "step": 5914 + }, + { + "epoch": 3.2111834961997827, + "grad_norm": 10.649138913539469, + "learning_rate": 5.995593015190608e-06, + "loss": 0.5665, + "step": 5915 + }, + { + "epoch": 3.211726384364821, + "grad_norm": 10.875890162606636, + "learning_rate": 5.992370703183803e-06, + "loss": 0.3634, + "step": 5916 + }, + { + "epoch": 3.2122692725298587, + "grad_norm": 16.640296114311138, + "learning_rate": 5.9891488868469775e-06, + "loss": 0.7804, + "step": 5917 + }, + { + "epoch": 3.212812160694897, + "grad_norm": 9.73069162630754, + "learning_rate": 5.9859275665786065e-06, + "loss": 0.35, + "step": 5918 + }, + { + "epoch": 3.2133550488599347, + "grad_norm": 10.657768777546748, + "learning_rate": 5.982706742777116e-06, + "loss": 0.529, + "step": 5919 + }, + { + "epoch": 3.213897937024973, + "grad_norm": 12.700699759951497, + "learning_rate": 5.9794864158408535e-06, + "loss": 0.5159, + "step": 5920 + }, + { + "epoch": 3.2144408251900107, + "grad_norm": 8.375283482104031, + "learning_rate": 5.976266586168121e-06, + "loss": 0.2818, + "step": 5921 + }, + { + "epoch": 3.214983713355049, + "grad_norm": 18.061811128959963, + "learning_rate": 5.973047254157148e-06, + "loss": 1.2271, + "step": 5922 + }, + { + "epoch": 3.2155266015200867, + "grad_norm": 10.630424335111323, + "learning_rate": 5.969828420206114e-06, + "loss": 0.3874, + "step": 5923 + }, + { + "epoch": 3.216069489685125, + "grad_norm": 11.324230234064501, + "learning_rate": 5.966610084713118e-06, + "loss": 0.6559, + "step": 5924 + }, + { + "epoch": 3.2166123778501627, + "grad_norm": 12.436990138030758, + "learning_rate": 5.9633922480762205e-06, + "loss": 0.6468, + "step": 5925 + }, + { + "epoch": 3.217155266015201, + "grad_norm": 16.354922341711003, + "learning_rate": 5.960174910693401e-06, + "loss": 0.8733, + "step": 5926 + }, + { + "epoch": 3.2176981541802387, + "grad_norm": 12.756426829873451, + "learning_rate": 5.956958072962586e-06, + "loss": 0.5399, + "step": 5927 + }, + { + "epoch": 3.218241042345277, + "grad_norm": 11.493661345744254, + "learning_rate": 5.953741735281641e-06, + "loss": 0.5487, + "step": 5928 + }, + { + "epoch": 3.2187839305103148, + "grad_norm": 11.431750755108203, + "learning_rate": 5.9505258980483645e-06, + "loss": 0.3997, + "step": 5929 + }, + { + "epoch": 3.219326818675353, + "grad_norm": 12.940367227326554, + "learning_rate": 5.947310561660503e-06, + "loss": 0.8178, + "step": 5930 + }, + { + "epoch": 3.2198697068403908, + "grad_norm": 11.421594805016625, + "learning_rate": 5.9440957265157265e-06, + "loss": 0.5015, + "step": 5931 + }, + { + "epoch": 3.220412595005429, + "grad_norm": 12.215300050799616, + "learning_rate": 5.940881393011659e-06, + "loss": 0.4044, + "step": 5932 + }, + { + "epoch": 3.2209554831704668, + "grad_norm": 11.265138511012573, + "learning_rate": 5.937667561545845e-06, + "loss": 0.6294, + "step": 5933 + }, + { + "epoch": 3.221498371335505, + "grad_norm": 11.883510126446058, + "learning_rate": 5.934454232515785e-06, + "loss": 0.5257, + "step": 5934 + }, + { + "epoch": 3.2220412595005428, + "grad_norm": 14.862260710742142, + "learning_rate": 5.931241406318906e-06, + "loss": 0.7435, + "step": 5935 + }, + { + "epoch": 3.222584147665581, + "grad_norm": 11.010381613248398, + "learning_rate": 5.92802908335257e-06, + "loss": 0.4852, + "step": 5936 + }, + { + "epoch": 3.2231270358306188, + "grad_norm": 10.894562800140209, + "learning_rate": 5.924817264014091e-06, + "loss": 0.4694, + "step": 5937 + }, + { + "epoch": 3.223669923995657, + "grad_norm": 9.91297174934936, + "learning_rate": 5.921605948700704e-06, + "loss": 0.4031, + "step": 5938 + }, + { + "epoch": 3.2242128121606948, + "grad_norm": 10.829816377096837, + "learning_rate": 5.918395137809596e-06, + "loss": 0.5139, + "step": 5939 + }, + { + "epoch": 3.224755700325733, + "grad_norm": 9.940894286293783, + "learning_rate": 5.915184831737881e-06, + "loss": 0.3475, + "step": 5940 + }, + { + "epoch": 3.225298588490771, + "grad_norm": 12.818677443045534, + "learning_rate": 5.911975030882617e-06, + "loss": 0.4293, + "step": 5941 + }, + { + "epoch": 3.225841476655809, + "grad_norm": 8.859395484195392, + "learning_rate": 5.908765735640795e-06, + "loss": 0.4065, + "step": 5942 + }, + { + "epoch": 3.226384364820847, + "grad_norm": 10.145624567540983, + "learning_rate": 5.90555694640935e-06, + "loss": 0.3713, + "step": 5943 + }, + { + "epoch": 3.226927252985885, + "grad_norm": 12.596453073542072, + "learning_rate": 5.902348663585144e-06, + "loss": 0.5128, + "step": 5944 + }, + { + "epoch": 3.227470141150923, + "grad_norm": 13.220304215230376, + "learning_rate": 5.899140887564991e-06, + "loss": 0.7293, + "step": 5945 + }, + { + "epoch": 3.228013029315961, + "grad_norm": 15.436135404867363, + "learning_rate": 5.8959336187456265e-06, + "loss": 0.8522, + "step": 5946 + }, + { + "epoch": 3.228555917480999, + "grad_norm": 11.078819738735637, + "learning_rate": 5.89272685752373e-06, + "loss": 0.5559, + "step": 5947 + }, + { + "epoch": 3.229098805646037, + "grad_norm": 11.732786443102684, + "learning_rate": 5.889520604295926e-06, + "loss": 0.472, + "step": 5948 + }, + { + "epoch": 3.229641693811075, + "grad_norm": 12.744163420827304, + "learning_rate": 5.88631485945876e-06, + "loss": 0.5843, + "step": 5949 + }, + { + "epoch": 3.230184581976113, + "grad_norm": 12.717140075489619, + "learning_rate": 5.8831096234087335e-06, + "loss": 0.4826, + "step": 5950 + }, + { + "epoch": 3.230727470141151, + "grad_norm": 14.460440748409685, + "learning_rate": 5.8799048965422665e-06, + "loss": 0.5057, + "step": 5951 + }, + { + "epoch": 3.231270358306189, + "grad_norm": 13.034475639862867, + "learning_rate": 5.8767006792557316e-06, + "loss": 0.6811, + "step": 5952 + }, + { + "epoch": 3.231813246471227, + "grad_norm": 12.638392990019069, + "learning_rate": 5.873496971945425e-06, + "loss": 0.4532, + "step": 5953 + }, + { + "epoch": 3.232356134636265, + "grad_norm": 11.04321550599593, + "learning_rate": 5.8702937750075914e-06, + "loss": 0.5711, + "step": 5954 + }, + { + "epoch": 3.232899022801303, + "grad_norm": 14.964397050426088, + "learning_rate": 5.867091088838406e-06, + "loss": 0.6479, + "step": 5955 + }, + { + "epoch": 3.233441910966341, + "grad_norm": 10.647120251880176, + "learning_rate": 5.863888913833979e-06, + "loss": 0.5262, + "step": 5956 + }, + { + "epoch": 3.233984799131379, + "grad_norm": 11.288671625393498, + "learning_rate": 5.860687250390367e-06, + "loss": 0.479, + "step": 5957 + }, + { + "epoch": 3.234527687296417, + "grad_norm": 11.49866073465499, + "learning_rate": 5.857486098903549e-06, + "loss": 0.2939, + "step": 5958 + }, + { + "epoch": 3.235070575461455, + "grad_norm": 11.340345898343635, + "learning_rate": 5.854285459769454e-06, + "loss": 0.3908, + "step": 5959 + }, + { + "epoch": 3.235613463626493, + "grad_norm": 10.20477531793921, + "learning_rate": 5.851085333383939e-06, + "loss": 0.4133, + "step": 5960 + }, + { + "epoch": 3.236156351791531, + "grad_norm": 9.818700824916304, + "learning_rate": 5.8478857201428075e-06, + "loss": 0.3324, + "step": 5961 + }, + { + "epoch": 3.236699239956569, + "grad_norm": 9.73255562956133, + "learning_rate": 5.844686620441782e-06, + "loss": 0.7815, + "step": 5962 + }, + { + "epoch": 3.237242128121607, + "grad_norm": 17.504088445923752, + "learning_rate": 5.841488034676542e-06, + "loss": 0.693, + "step": 5963 + }, + { + "epoch": 3.237785016286645, + "grad_norm": 15.010551187805474, + "learning_rate": 5.838289963242684e-06, + "loss": 0.453, + "step": 5964 + }, + { + "epoch": 3.238327904451683, + "grad_norm": 11.512605593104578, + "learning_rate": 5.835092406535763e-06, + "loss": 0.792, + "step": 5965 + }, + { + "epoch": 3.238870792616721, + "grad_norm": 15.034973727458302, + "learning_rate": 5.831895364951247e-06, + "loss": 0.6216, + "step": 5966 + }, + { + "epoch": 3.239413680781759, + "grad_norm": 15.522277540385527, + "learning_rate": 5.828698838884558e-06, + "loss": 0.4771, + "step": 5967 + }, + { + "epoch": 3.239956568946797, + "grad_norm": 11.391715611260866, + "learning_rate": 5.825502828731042e-06, + "loss": 0.4213, + "step": 5968 + }, + { + "epoch": 3.240499457111835, + "grad_norm": 15.534988403627082, + "learning_rate": 5.82230733488599e-06, + "loss": 0.6494, + "step": 5969 + }, + { + "epoch": 3.241042345276873, + "grad_norm": 18.181760899119496, + "learning_rate": 5.819112357744626e-06, + "loss": 0.5825, + "step": 5970 + }, + { + "epoch": 3.241585233441911, + "grad_norm": 10.971869300023029, + "learning_rate": 5.815917897702104e-06, + "loss": 0.7647, + "step": 5971 + }, + { + "epoch": 3.242128121606949, + "grad_norm": 11.760933490969961, + "learning_rate": 5.812723955153533e-06, + "loss": 0.6203, + "step": 5972 + }, + { + "epoch": 3.242671009771987, + "grad_norm": 17.51328901218973, + "learning_rate": 5.809530530493929e-06, + "loss": 0.6598, + "step": 5973 + }, + { + "epoch": 3.243213897937025, + "grad_norm": 12.500120263524371, + "learning_rate": 5.8063376241182745e-06, + "loss": 0.4958, + "step": 5974 + }, + { + "epoch": 3.243756786102063, + "grad_norm": 11.050563890247632, + "learning_rate": 5.803145236421464e-06, + "loss": 0.4204, + "step": 5975 + }, + { + "epoch": 3.244299674267101, + "grad_norm": 6.058184111347752, + "learning_rate": 5.799953367798335e-06, + "loss": 0.2181, + "step": 5976 + }, + { + "epoch": 3.244842562432139, + "grad_norm": 9.648373148728487, + "learning_rate": 5.796762018643675e-06, + "loss": 0.4304, + "step": 5977 + }, + { + "epoch": 3.245385450597177, + "grad_norm": 11.906735912997219, + "learning_rate": 5.793571189352179e-06, + "loss": 0.5365, + "step": 5978 + }, + { + "epoch": 3.245928338762215, + "grad_norm": 12.12325593772168, + "learning_rate": 5.790380880318511e-06, + "loss": 0.3811, + "step": 5979 + }, + { + "epoch": 3.246471226927253, + "grad_norm": 18.625337422182657, + "learning_rate": 5.787191091937236e-06, + "loss": 0.6784, + "step": 5980 + }, + { + "epoch": 3.247014115092291, + "grad_norm": 9.093728620631289, + "learning_rate": 5.784001824602891e-06, + "loss": 0.4545, + "step": 5981 + }, + { + "epoch": 3.247557003257329, + "grad_norm": 12.360404516544222, + "learning_rate": 5.78081307870991e-06, + "loss": 0.5982, + "step": 5982 + }, + { + "epoch": 3.248099891422367, + "grad_norm": 12.023338602328888, + "learning_rate": 5.777624854652696e-06, + "loss": 0.4595, + "step": 5983 + }, + { + "epoch": 3.248642779587405, + "grad_norm": 11.60238178463911, + "learning_rate": 5.774437152825574e-06, + "loss": 0.4674, + "step": 5984 + }, + { + "epoch": 3.249185667752443, + "grad_norm": 13.826357134916073, + "learning_rate": 5.77124997362279e-06, + "loss": 0.4552, + "step": 5985 + }, + { + "epoch": 3.249728555917481, + "grad_norm": 13.531388043207002, + "learning_rate": 5.7680633174385595e-06, + "loss": 0.4926, + "step": 5986 + }, + { + "epoch": 3.250271444082519, + "grad_norm": 11.384053416563145, + "learning_rate": 5.764877184666993e-06, + "loss": 0.5127, + "step": 5987 + }, + { + "epoch": 3.250814332247557, + "grad_norm": 12.390918037366044, + "learning_rate": 5.761691575702168e-06, + "loss": 0.6428, + "step": 5988 + }, + { + "epoch": 3.251357220412595, + "grad_norm": 12.158948584345922, + "learning_rate": 5.758506490938087e-06, + "loss": 0.5374, + "step": 5989 + }, + { + "epoch": 3.251900108577633, + "grad_norm": 12.352114570156113, + "learning_rate": 5.755321930768683e-06, + "loss": 0.4149, + "step": 5990 + }, + { + "epoch": 3.252442996742671, + "grad_norm": 11.451850663530495, + "learning_rate": 5.752137895587826e-06, + "loss": 0.4864, + "step": 5991 + }, + { + "epoch": 3.252985884907709, + "grad_norm": 14.59595398014034, + "learning_rate": 5.748954385789325e-06, + "loss": 0.6791, + "step": 5992 + }, + { + "epoch": 3.253528773072747, + "grad_norm": 10.572257438915425, + "learning_rate": 5.74577140176692e-06, + "loss": 0.3851, + "step": 5993 + }, + { + "epoch": 3.254071661237785, + "grad_norm": 13.496500778871066, + "learning_rate": 5.742588943914289e-06, + "loss": 0.8371, + "step": 5994 + }, + { + "epoch": 3.254614549402823, + "grad_norm": 10.366843508091446, + "learning_rate": 5.739407012625043e-06, + "loss": 0.5289, + "step": 5995 + }, + { + "epoch": 3.255157437567861, + "grad_norm": 13.59778716822848, + "learning_rate": 5.736225608292727e-06, + "loss": 0.6353, + "step": 5996 + }, + { + "epoch": 3.255700325732899, + "grad_norm": 8.13083770907386, + "learning_rate": 5.7330447313108236e-06, + "loss": 0.2591, + "step": 5997 + }, + { + "epoch": 3.256243213897937, + "grad_norm": 11.705959913597166, + "learning_rate": 5.72986438207275e-06, + "loss": 0.4058, + "step": 5998 + }, + { + "epoch": 3.256786102062975, + "grad_norm": 9.25230495085822, + "learning_rate": 5.726684560971852e-06, + "loss": 0.3647, + "step": 5999 + }, + { + "epoch": 3.257328990228013, + "grad_norm": 10.705513260623167, + "learning_rate": 5.72350526840142e-06, + "loss": 0.6986, + "step": 6000 + }, + { + "epoch": 3.257871878393051, + "grad_norm": 11.426736023101553, + "learning_rate": 5.720326504754672e-06, + "loss": 0.548, + "step": 6001 + }, + { + "epoch": 3.258414766558089, + "grad_norm": 11.105330737807535, + "learning_rate": 5.71714827042476e-06, + "loss": 0.5345, + "step": 6002 + }, + { + "epoch": 3.258957654723127, + "grad_norm": 11.734545793783724, + "learning_rate": 5.713970565804782e-06, + "loss": 0.4435, + "step": 6003 + }, + { + "epoch": 3.259500542888165, + "grad_norm": 13.234029513207872, + "learning_rate": 5.710793391287751e-06, + "loss": 0.538, + "step": 6004 + }, + { + "epoch": 3.260043431053203, + "grad_norm": 9.429540505637686, + "learning_rate": 5.707616747266631e-06, + "loss": 0.4042, + "step": 6005 + }, + { + "epoch": 3.260586319218241, + "grad_norm": 10.73830038342973, + "learning_rate": 5.704440634134312e-06, + "loss": 0.4394, + "step": 6006 + }, + { + "epoch": 3.261129207383279, + "grad_norm": 15.154563307319142, + "learning_rate": 5.70126505228362e-06, + "loss": 0.5733, + "step": 6007 + }, + { + "epoch": 3.261672095548317, + "grad_norm": 11.231865527791925, + "learning_rate": 5.6980900021073196e-06, + "loss": 0.3472, + "step": 6008 + }, + { + "epoch": 3.262214983713355, + "grad_norm": 10.315608836480246, + "learning_rate": 5.694915483998099e-06, + "loss": 0.3231, + "step": 6009 + }, + { + "epoch": 3.262757871878393, + "grad_norm": 13.017299228674494, + "learning_rate": 5.691741498348601e-06, + "loss": 0.4652, + "step": 6010 + }, + { + "epoch": 3.263300760043431, + "grad_norm": 13.428357165583693, + "learning_rate": 5.688568045551373e-06, + "loss": 0.51, + "step": 6011 + }, + { + "epoch": 3.263843648208469, + "grad_norm": 12.887391562992057, + "learning_rate": 5.685395125998927e-06, + "loss": 0.7156, + "step": 6012 + }, + { + "epoch": 3.264386536373507, + "grad_norm": 11.112402459616494, + "learning_rate": 5.682222740083683e-06, + "loss": 0.3077, + "step": 6013 + }, + { + "epoch": 3.264929424538545, + "grad_norm": 17.334186329467677, + "learning_rate": 5.6790508881980136e-06, + "loss": 0.487, + "step": 6014 + }, + { + "epoch": 3.265472312703583, + "grad_norm": 11.272673037009941, + "learning_rate": 5.675879570734222e-06, + "loss": 0.5596, + "step": 6015 + }, + { + "epoch": 3.266015200868621, + "grad_norm": 10.57062964165952, + "learning_rate": 5.672708788084528e-06, + "loss": 0.4693, + "step": 6016 + }, + { + "epoch": 3.266558089033659, + "grad_norm": 12.532994743872049, + "learning_rate": 5.669538540641116e-06, + "loss": 0.5709, + "step": 6017 + }, + { + "epoch": 3.267100977198697, + "grad_norm": 12.467434828440602, + "learning_rate": 5.666368828796072e-06, + "loss": 0.5328, + "step": 6018 + }, + { + "epoch": 3.267643865363735, + "grad_norm": 10.349926895360463, + "learning_rate": 5.663199652941444e-06, + "loss": 0.3587, + "step": 6019 + }, + { + "epoch": 3.268186753528773, + "grad_norm": 10.889914966114198, + "learning_rate": 5.660031013469189e-06, + "loss": 0.469, + "step": 6020 + }, + { + "epoch": 3.268729641693811, + "grad_norm": 16.63961667582981, + "learning_rate": 5.656862910771218e-06, + "loss": 0.7146, + "step": 6021 + }, + { + "epoch": 3.269272529858849, + "grad_norm": 10.447294419775371, + "learning_rate": 5.653695345239365e-06, + "loss": 0.3975, + "step": 6022 + }, + { + "epoch": 3.269815418023887, + "grad_norm": 12.85070217834024, + "learning_rate": 5.650528317265398e-06, + "loss": 0.7395, + "step": 6023 + }, + { + "epoch": 3.270358306188925, + "grad_norm": 8.341742485534219, + "learning_rate": 5.647361827241025e-06, + "loss": 0.2861, + "step": 6024 + }, + { + "epoch": 3.270901194353963, + "grad_norm": 9.898309581037385, + "learning_rate": 5.64419587555787e-06, + "loss": 0.6192, + "step": 6025 + }, + { + "epoch": 3.271444082519001, + "grad_norm": 12.335345691658691, + "learning_rate": 5.641030462607514e-06, + "loss": 0.5161, + "step": 6026 + }, + { + "epoch": 3.271986970684039, + "grad_norm": 10.304455881328025, + "learning_rate": 5.637865588781459e-06, + "loss": 0.5773, + "step": 6027 + }, + { + "epoch": 3.272529858849077, + "grad_norm": 11.527834619584397, + "learning_rate": 5.63470125447114e-06, + "loss": 0.4415, + "step": 6028 + }, + { + "epoch": 3.273072747014115, + "grad_norm": 12.835285732552766, + "learning_rate": 5.631537460067926e-06, + "loss": 0.4513, + "step": 6029 + }, + { + "epoch": 3.273615635179153, + "grad_norm": 15.34754753916867, + "learning_rate": 5.628374205963123e-06, + "loss": 0.9626, + "step": 6030 + }, + { + "epoch": 3.274158523344191, + "grad_norm": 13.537553719954104, + "learning_rate": 5.625211492547965e-06, + "loss": 0.6022, + "step": 6031 + }, + { + "epoch": 3.274701411509229, + "grad_norm": 13.03189959063881, + "learning_rate": 5.622049320213622e-06, + "loss": 0.3918, + "step": 6032 + }, + { + "epoch": 3.275244299674267, + "grad_norm": 11.883924111949161, + "learning_rate": 5.618887689351195e-06, + "loss": 0.6166, + "step": 6033 + }, + { + "epoch": 3.2757871878393052, + "grad_norm": 12.574278462374288, + "learning_rate": 5.615726600351723e-06, + "loss": 0.364, + "step": 6034 + }, + { + "epoch": 3.276330076004343, + "grad_norm": 11.646167790458746, + "learning_rate": 5.612566053606172e-06, + "loss": 0.4893, + "step": 6035 + }, + { + "epoch": 3.2768729641693812, + "grad_norm": 15.464095407664189, + "learning_rate": 5.6094060495054435e-06, + "loss": 0.5033, + "step": 6036 + }, + { + "epoch": 3.277415852334419, + "grad_norm": 9.470859172197285, + "learning_rate": 5.606246588440374e-06, + "loss": 0.3201, + "step": 6037 + }, + { + "epoch": 3.2779587404994572, + "grad_norm": 10.186080712958546, + "learning_rate": 5.603087670801728e-06, + "loss": 0.375, + "step": 6038 + }, + { + "epoch": 3.278501628664495, + "grad_norm": 12.069192340552016, + "learning_rate": 5.599929296980205e-06, + "loss": 0.521, + "step": 6039 + }, + { + "epoch": 3.2790445168295332, + "grad_norm": 13.470203218211036, + "learning_rate": 5.596771467366442e-06, + "loss": 0.6999, + "step": 6040 + }, + { + "epoch": 3.279587404994571, + "grad_norm": 15.234328984284964, + "learning_rate": 5.5936141823509995e-06, + "loss": 0.658, + "step": 6041 + }, + { + "epoch": 3.2801302931596092, + "grad_norm": 10.28714486065803, + "learning_rate": 5.590457442324374e-06, + "loss": 0.49, + "step": 6042 + }, + { + "epoch": 3.280673181324647, + "grad_norm": 13.229082345644139, + "learning_rate": 5.587301247677009e-06, + "loss": 0.8617, + "step": 6043 + }, + { + "epoch": 3.2812160694896852, + "grad_norm": 19.161743435780973, + "learning_rate": 5.5841455987992536e-06, + "loss": 0.673, + "step": 6044 + }, + { + "epoch": 3.281758957654723, + "grad_norm": 14.576260981710798, + "learning_rate": 5.580990496081407e-06, + "loss": 0.4843, + "step": 6045 + }, + { + "epoch": 3.2823018458197613, + "grad_norm": 10.509886153036398, + "learning_rate": 5.577835939913701e-06, + "loss": 0.3372, + "step": 6046 + }, + { + "epoch": 3.282844733984799, + "grad_norm": 14.23830213747384, + "learning_rate": 5.574681930686289e-06, + "loss": 0.5041, + "step": 6047 + }, + { + "epoch": 3.2833876221498373, + "grad_norm": 16.736802729512746, + "learning_rate": 5.571528468789276e-06, + "loss": 0.6261, + "step": 6048 + }, + { + "epoch": 3.283930510314875, + "grad_norm": 16.53917203611799, + "learning_rate": 5.5683755546126724e-06, + "loss": 0.6671, + "step": 6049 + }, + { + "epoch": 3.2844733984799133, + "grad_norm": 9.26057962235423, + "learning_rate": 5.565223188546452e-06, + "loss": 0.4135, + "step": 6050 + }, + { + "epoch": 3.285016286644951, + "grad_norm": 13.238225399242737, + "learning_rate": 5.562071370980486e-06, + "loss": 0.558, + "step": 6051 + }, + { + "epoch": 3.2855591748099893, + "grad_norm": 9.005338285525495, + "learning_rate": 5.558920102304615e-06, + "loss": 0.3677, + "step": 6052 + }, + { + "epoch": 3.286102062975027, + "grad_norm": 14.118931865854162, + "learning_rate": 5.55576938290858e-06, + "loss": 0.8303, + "step": 6053 + }, + { + "epoch": 3.2866449511400653, + "grad_norm": 9.596979737331942, + "learning_rate": 5.5526192131820665e-06, + "loss": 0.3322, + "step": 6054 + }, + { + "epoch": 3.287187839305103, + "grad_norm": 11.143732344567033, + "learning_rate": 5.549469593514705e-06, + "loss": 0.4143, + "step": 6055 + }, + { + "epoch": 3.2877307274701413, + "grad_norm": 12.135282131920155, + "learning_rate": 5.546320524296028e-06, + "loss": 0.5302, + "step": 6056 + }, + { + "epoch": 3.288273615635179, + "grad_norm": 12.209025529059957, + "learning_rate": 5.543172005915536e-06, + "loss": 0.5643, + "step": 6057 + }, + { + "epoch": 3.2888165038002173, + "grad_norm": 16.880330158776534, + "learning_rate": 5.540024038762623e-06, + "loss": 0.6743, + "step": 6058 + }, + { + "epoch": 3.289359391965255, + "grad_norm": 13.623361718010658, + "learning_rate": 5.536876623226652e-06, + "loss": 0.6925, + "step": 6059 + }, + { + "epoch": 3.2899022801302933, + "grad_norm": 11.161796182879957, + "learning_rate": 5.533729759696893e-06, + "loss": 0.5875, + "step": 6060 + }, + { + "epoch": 3.290445168295331, + "grad_norm": 14.324802623539963, + "learning_rate": 5.5305834485625545e-06, + "loss": 0.9071, + "step": 6061 + }, + { + "epoch": 3.2909880564603693, + "grad_norm": 12.43112855099258, + "learning_rate": 5.527437690212778e-06, + "loss": 0.3761, + "step": 6062 + }, + { + "epoch": 3.291530944625407, + "grad_norm": 14.231447955496243, + "learning_rate": 5.524292485036638e-06, + "loss": 0.7001, + "step": 6063 + }, + { + "epoch": 3.2920738327904453, + "grad_norm": 11.474598912194304, + "learning_rate": 5.521147833423135e-06, + "loss": 0.419, + "step": 6064 + }, + { + "epoch": 3.292616720955483, + "grad_norm": 13.615753477381617, + "learning_rate": 5.518003735761206e-06, + "loss": 0.7142, + "step": 6065 + }, + { + "epoch": 3.2931596091205213, + "grad_norm": 10.341494655988399, + "learning_rate": 5.514860192439719e-06, + "loss": 0.4286, + "step": 6066 + }, + { + "epoch": 3.293702497285559, + "grad_norm": 11.37742694154389, + "learning_rate": 5.511717203847472e-06, + "loss": 0.4302, + "step": 6067 + }, + { + "epoch": 3.2942453854505973, + "grad_norm": 14.04977404747662, + "learning_rate": 5.508574770373194e-06, + "loss": 0.4538, + "step": 6068 + }, + { + "epoch": 3.294788273615635, + "grad_norm": 13.008904129849688, + "learning_rate": 5.505432892405547e-06, + "loss": 0.419, + "step": 6069 + }, + { + "epoch": 3.2953311617806733, + "grad_norm": 8.086590828855625, + "learning_rate": 5.502291570333122e-06, + "loss": 0.276, + "step": 6070 + }, + { + "epoch": 3.295874049945711, + "grad_norm": 8.20494775069595, + "learning_rate": 5.499150804544445e-06, + "loss": 0.349, + "step": 6071 + }, + { + "epoch": 3.2964169381107493, + "grad_norm": 13.336711424526726, + "learning_rate": 5.496010595427969e-06, + "loss": 0.3958, + "step": 6072 + }, + { + "epoch": 3.296959826275787, + "grad_norm": 21.927197817240263, + "learning_rate": 5.492870943372082e-06, + "loss": 0.5561, + "step": 6073 + }, + { + "epoch": 3.2975027144408253, + "grad_norm": 14.168102563986727, + "learning_rate": 5.489731848765098e-06, + "loss": 0.6231, + "step": 6074 + }, + { + "epoch": 3.298045602605863, + "grad_norm": 9.428786171987674, + "learning_rate": 5.48659331199527e-06, + "loss": 0.3004, + "step": 6075 + }, + { + "epoch": 3.2985884907709013, + "grad_norm": 13.58619098409203, + "learning_rate": 5.483455333450774e-06, + "loss": 0.7655, + "step": 6076 + }, + { + "epoch": 3.299131378935939, + "grad_norm": 10.799157987601124, + "learning_rate": 5.480317913519718e-06, + "loss": 0.4688, + "step": 6077 + }, + { + "epoch": 3.2996742671009773, + "grad_norm": 11.88277850667338, + "learning_rate": 5.477181052590148e-06, + "loss": 0.576, + "step": 6078 + }, + { + "epoch": 3.300217155266015, + "grad_norm": 9.399361473432771, + "learning_rate": 5.474044751050032e-06, + "loss": 0.3612, + "step": 6079 + }, + { + "epoch": 3.3007600434310533, + "grad_norm": 12.83439735807638, + "learning_rate": 5.470909009287273e-06, + "loss": 0.4007, + "step": 6080 + }, + { + "epoch": 3.301302931596091, + "grad_norm": 10.935369972874188, + "learning_rate": 5.467773827689712e-06, + "loss": 0.5349, + "step": 6081 + }, + { + "epoch": 3.3018458197611293, + "grad_norm": 13.981231839826476, + "learning_rate": 5.464639206645104e-06, + "loss": 0.4123, + "step": 6082 + }, + { + "epoch": 3.302388707926167, + "grad_norm": 10.388999075429536, + "learning_rate": 5.461505146541147e-06, + "loss": 0.666, + "step": 6083 + }, + { + "epoch": 3.3029315960912053, + "grad_norm": 10.483727188037403, + "learning_rate": 5.4583716477654665e-06, + "loss": 0.4385, + "step": 6084 + }, + { + "epoch": 3.303474484256243, + "grad_norm": 15.004888765224539, + "learning_rate": 5.455238710705616e-06, + "loss": 0.6089, + "step": 6085 + }, + { + "epoch": 3.3040173724212814, + "grad_norm": 12.510560050118539, + "learning_rate": 5.452106335749092e-06, + "loss": 0.5481, + "step": 6086 + }, + { + "epoch": 3.304560260586319, + "grad_norm": 12.402037551065197, + "learning_rate": 5.448974523283297e-06, + "loss": 0.5909, + "step": 6087 + }, + { + "epoch": 3.3051031487513574, + "grad_norm": 9.456679882341124, + "learning_rate": 5.445843273695595e-06, + "loss": 0.4422, + "step": 6088 + }, + { + "epoch": 3.305646036916395, + "grad_norm": 12.165182627083702, + "learning_rate": 5.4427125873732455e-06, + "loss": 0.8632, + "step": 6089 + }, + { + "epoch": 3.3061889250814334, + "grad_norm": 10.815051939615287, + "learning_rate": 5.439582464703476e-06, + "loss": 0.5143, + "step": 6090 + }, + { + "epoch": 3.306731813246471, + "grad_norm": 8.659965055909717, + "learning_rate": 5.436452906073406e-06, + "loss": 0.3732, + "step": 6091 + }, + { + "epoch": 3.3072747014115094, + "grad_norm": 12.566885836457084, + "learning_rate": 5.433323911870119e-06, + "loss": 0.7708, + "step": 6092 + }, + { + "epoch": 3.307817589576547, + "grad_norm": 12.162201810127142, + "learning_rate": 5.430195482480611e-06, + "loss": 0.5414, + "step": 6093 + }, + { + "epoch": 3.3083604777415854, + "grad_norm": 10.50777462312628, + "learning_rate": 5.427067618291804e-06, + "loss": 0.4301, + "step": 6094 + }, + { + "epoch": 3.308903365906623, + "grad_norm": 10.640857438281406, + "learning_rate": 5.423940319690568e-06, + "loss": 0.4682, + "step": 6095 + }, + { + "epoch": 3.3094462540716614, + "grad_norm": 10.370610442117792, + "learning_rate": 5.420813587063678e-06, + "loss": 0.3056, + "step": 6096 + }, + { + "epoch": 3.309989142236699, + "grad_norm": 10.60821300635486, + "learning_rate": 5.417687420797867e-06, + "loss": 0.3344, + "step": 6097 + }, + { + "epoch": 3.3105320304017374, + "grad_norm": 10.353507117674113, + "learning_rate": 5.414561821279778e-06, + "loss": 0.656, + "step": 6098 + }, + { + "epoch": 3.311074918566775, + "grad_norm": 8.344962502930883, + "learning_rate": 5.411436788895992e-06, + "loss": 0.3413, + "step": 6099 + }, + { + "epoch": 3.3116178067318134, + "grad_norm": 9.483199213032712, + "learning_rate": 5.408312324033016e-06, + "loss": 0.4156, + "step": 6100 + }, + { + "epoch": 3.312160694896851, + "grad_norm": 10.849967748759967, + "learning_rate": 5.40518842707729e-06, + "loss": 0.3849, + "step": 6101 + }, + { + "epoch": 3.3127035830618894, + "grad_norm": 11.385869965644408, + "learning_rate": 5.402065098415188e-06, + "loss": 0.5767, + "step": 6102 + }, + { + "epoch": 3.313246471226927, + "grad_norm": 9.846646631682042, + "learning_rate": 5.398942338432993e-06, + "loss": 0.3517, + "step": 6103 + }, + { + "epoch": 3.3137893593919654, + "grad_norm": 12.208760557718977, + "learning_rate": 5.3958201475169455e-06, + "loss": 0.3989, + "step": 6104 + }, + { + "epoch": 3.314332247557003, + "grad_norm": 14.044852806162847, + "learning_rate": 5.392698526053203e-06, + "loss": 0.6161, + "step": 6105 + }, + { + "epoch": 3.3148751357220414, + "grad_norm": 16.302573867426716, + "learning_rate": 5.389577474427848e-06, + "loss": 0.5668, + "step": 6106 + }, + { + "epoch": 3.315418023887079, + "grad_norm": 15.241837519747031, + "learning_rate": 5.3864569930268986e-06, + "loss": 0.531, + "step": 6107 + }, + { + "epoch": 3.3159609120521174, + "grad_norm": 16.054606664926766, + "learning_rate": 5.383337082236302e-06, + "loss": 0.5549, + "step": 6108 + }, + { + "epoch": 3.316503800217155, + "grad_norm": 7.873315866697446, + "learning_rate": 5.3802177424419335e-06, + "loss": 0.3505, + "step": 6109 + }, + { + "epoch": 3.3170466883821934, + "grad_norm": 7.233865234327245, + "learning_rate": 5.377098974029595e-06, + "loss": 0.3402, + "step": 6110 + }, + { + "epoch": 3.317589576547231, + "grad_norm": 13.489019875241555, + "learning_rate": 5.373980777385026e-06, + "loss": 0.5829, + "step": 6111 + }, + { + "epoch": 3.3181324647122694, + "grad_norm": 8.944352234530756, + "learning_rate": 5.370863152893886e-06, + "loss": 0.2591, + "step": 6112 + }, + { + "epoch": 3.318675352877307, + "grad_norm": 11.421952434092379, + "learning_rate": 5.367746100941769e-06, + "loss": 0.4549, + "step": 6113 + }, + { + "epoch": 3.3192182410423454, + "grad_norm": 9.61822328391865, + "learning_rate": 5.364629621914197e-06, + "loss": 0.3805, + "step": 6114 + }, + { + "epoch": 3.319761129207383, + "grad_norm": 9.466765582290753, + "learning_rate": 5.3615137161966205e-06, + "loss": 0.2804, + "step": 6115 + }, + { + "epoch": 3.3203040173724214, + "grad_norm": 10.228565001712072, + "learning_rate": 5.358398384174421e-06, + "loss": 0.3802, + "step": 6116 + }, + { + "epoch": 3.320846905537459, + "grad_norm": 12.471667235628182, + "learning_rate": 5.355283626232907e-06, + "loss": 0.7176, + "step": 6117 + }, + { + "epoch": 3.3213897937024974, + "grad_norm": 8.698625687388546, + "learning_rate": 5.352169442757312e-06, + "loss": 0.2962, + "step": 6118 + }, + { + "epoch": 3.321932681867535, + "grad_norm": 14.02974515499815, + "learning_rate": 5.349055834132817e-06, + "loss": 0.4804, + "step": 6119 + }, + { + "epoch": 3.3224755700325734, + "grad_norm": 14.91877137041168, + "learning_rate": 5.345942800744499e-06, + "loss": 0.4992, + "step": 6120 + }, + { + "epoch": 3.323018458197611, + "grad_norm": 13.54006240651423, + "learning_rate": 5.342830342977403e-06, + "loss": 0.822, + "step": 6121 + }, + { + "epoch": 3.3235613463626494, + "grad_norm": 11.467004290914351, + "learning_rate": 5.3397184612164676e-06, + "loss": 0.4878, + "step": 6122 + }, + { + "epoch": 3.324104234527687, + "grad_norm": 11.864551909227606, + "learning_rate": 5.33660715584658e-06, + "loss": 0.5312, + "step": 6123 + }, + { + "epoch": 3.3246471226927254, + "grad_norm": 10.288196378126806, + "learning_rate": 5.333496427252551e-06, + "loss": 0.4778, + "step": 6124 + }, + { + "epoch": 3.3251900108577632, + "grad_norm": 11.010968624281599, + "learning_rate": 5.330386275819119e-06, + "loss": 0.4471, + "step": 6125 + }, + { + "epoch": 3.3257328990228014, + "grad_norm": 9.902175725807343, + "learning_rate": 5.327276701930961e-06, + "loss": 0.3581, + "step": 6126 + }, + { + "epoch": 3.3262757871878392, + "grad_norm": 14.462369688525524, + "learning_rate": 5.324167705972661e-06, + "loss": 0.4262, + "step": 6127 + }, + { + "epoch": 3.3268186753528775, + "grad_norm": 12.522646032715151, + "learning_rate": 5.32105928832876e-06, + "loss": 0.3605, + "step": 6128 + }, + { + "epoch": 3.3273615635179152, + "grad_norm": 8.804303263881113, + "learning_rate": 5.317951449383693e-06, + "loss": 0.2589, + "step": 6129 + }, + { + "epoch": 3.3279044516829535, + "grad_norm": 16.231015853377418, + "learning_rate": 5.314844189521859e-06, + "loss": 0.6517, + "step": 6130 + }, + { + "epoch": 3.3284473398479912, + "grad_norm": 15.588313322776349, + "learning_rate": 5.311737509127561e-06, + "loss": 0.6392, + "step": 6131 + }, + { + "epoch": 3.3289902280130295, + "grad_norm": 8.118329570429164, + "learning_rate": 5.308631408585041e-06, + "loss": 0.2433, + "step": 6132 + }, + { + "epoch": 3.3295331161780672, + "grad_norm": 17.927534810088048, + "learning_rate": 5.305525888278469e-06, + "loss": 0.8314, + "step": 6133 + }, + { + "epoch": 3.3300760043431055, + "grad_norm": 12.7719407908283, + "learning_rate": 5.302420948591929e-06, + "loss": 0.5422, + "step": 6134 + }, + { + "epoch": 3.3306188925081432, + "grad_norm": 13.715317083116945, + "learning_rate": 5.29931658990946e-06, + "loss": 0.4756, + "step": 6135 + }, + { + "epoch": 3.3311617806731815, + "grad_norm": 14.232906623385299, + "learning_rate": 5.296212812615001e-06, + "loss": 0.5507, + "step": 6136 + }, + { + "epoch": 3.3317046688382193, + "grad_norm": 11.07024486678308, + "learning_rate": 5.29310961709244e-06, + "loss": 0.4916, + "step": 6137 + }, + { + "epoch": 3.3322475570032575, + "grad_norm": 12.322287003987269, + "learning_rate": 5.290007003725585e-06, + "loss": 0.426, + "step": 6138 + }, + { + "epoch": 3.3327904451682953, + "grad_norm": 11.084747737931762, + "learning_rate": 5.286904972898168e-06, + "loss": 0.2768, + "step": 6139 + }, + { + "epoch": 3.3333333333333335, + "grad_norm": 12.758536777666697, + "learning_rate": 5.283803524993858e-06, + "loss": 0.9266, + "step": 6140 + }, + { + "epoch": 3.3338762214983713, + "grad_norm": 13.098970900751402, + "learning_rate": 5.280702660396243e-06, + "loss": 0.512, + "step": 6141 + }, + { + "epoch": 3.3344191096634095, + "grad_norm": 12.317106256281363, + "learning_rate": 5.277602379488844e-06, + "loss": 0.4155, + "step": 6142 + }, + { + "epoch": 3.3349619978284473, + "grad_norm": 11.38584068826573, + "learning_rate": 5.274502682655112e-06, + "loss": 0.4051, + "step": 6143 + }, + { + "epoch": 3.3355048859934855, + "grad_norm": 14.501004041816849, + "learning_rate": 5.271403570278417e-06, + "loss": 0.5548, + "step": 6144 + }, + { + "epoch": 3.3360477741585233, + "grad_norm": 12.998451360538299, + "learning_rate": 5.268305042742065e-06, + "loss": 0.4021, + "step": 6145 + }, + { + "epoch": 3.3365906623235615, + "grad_norm": 9.038839543671822, + "learning_rate": 5.2652071004292855e-06, + "loss": 0.2625, + "step": 6146 + }, + { + "epoch": 3.3371335504885993, + "grad_norm": 14.728843022444776, + "learning_rate": 5.26210974372324e-06, + "loss": 0.7124, + "step": 6147 + }, + { + "epoch": 3.3376764386536375, + "grad_norm": 12.24691457854214, + "learning_rate": 5.259012973007011e-06, + "loss": 0.3889, + "step": 6148 + }, + { + "epoch": 3.3382193268186753, + "grad_norm": 12.889746719808702, + "learning_rate": 5.255916788663614e-06, + "loss": 0.6392, + "step": 6149 + }, + { + "epoch": 3.3387622149837135, + "grad_norm": 15.16855584494826, + "learning_rate": 5.252821191075989e-06, + "loss": 0.6308, + "step": 6150 + }, + { + "epoch": 3.3393051031487513, + "grad_norm": 15.137292971607204, + "learning_rate": 5.249726180627006e-06, + "loss": 0.4732, + "step": 6151 + }, + { + "epoch": 3.3398479913137895, + "grad_norm": 12.827169669915257, + "learning_rate": 5.246631757699461e-06, + "loss": 0.4235, + "step": 6152 + }, + { + "epoch": 3.3403908794788273, + "grad_norm": 8.000573554523031, + "learning_rate": 5.243537922676074e-06, + "loss": 0.3355, + "step": 6153 + }, + { + "epoch": 3.3409337676438655, + "grad_norm": 11.071696970347325, + "learning_rate": 5.240444675939498e-06, + "loss": 0.3129, + "step": 6154 + }, + { + "epoch": 3.3414766558089033, + "grad_norm": 13.751963016664462, + "learning_rate": 5.237352017872308e-06, + "loss": 0.7001, + "step": 6155 + }, + { + "epoch": 3.3420195439739415, + "grad_norm": 11.736421365008567, + "learning_rate": 5.23425994885701e-06, + "loss": 0.5775, + "step": 6156 + }, + { + "epoch": 3.3425624321389793, + "grad_norm": 13.034138205582762, + "learning_rate": 5.231168469276044e-06, + "loss": 0.4829, + "step": 6157 + }, + { + "epoch": 3.3431053203040175, + "grad_norm": 13.897122730322542, + "learning_rate": 5.228077579511754e-06, + "loss": 0.6586, + "step": 6158 + }, + { + "epoch": 3.3436482084690553, + "grad_norm": 13.765426696215664, + "learning_rate": 5.224987279946441e-06, + "loss": 0.4817, + "step": 6159 + }, + { + "epoch": 3.3441910966340935, + "grad_norm": 13.446494398430909, + "learning_rate": 5.221897570962304e-06, + "loss": 0.5121, + "step": 6160 + }, + { + "epoch": 3.3447339847991313, + "grad_norm": 9.817713520136255, + "learning_rate": 5.2188084529415e-06, + "loss": 0.2662, + "step": 6161 + }, + { + "epoch": 3.3452768729641695, + "grad_norm": 10.025373895120913, + "learning_rate": 5.215719926266082e-06, + "loss": 0.4928, + "step": 6162 + }, + { + "epoch": 3.3458197611292073, + "grad_norm": 13.267267741014804, + "learning_rate": 5.212631991318044e-06, + "loss": 0.6133, + "step": 6163 + }, + { + "epoch": 3.3463626492942455, + "grad_norm": 13.654831265704226, + "learning_rate": 5.209544648479319e-06, + "loss": 0.4202, + "step": 6164 + }, + { + "epoch": 3.3469055374592833, + "grad_norm": 14.003058740046605, + "learning_rate": 5.20645789813174e-06, + "loss": 0.6629, + "step": 6165 + }, + { + "epoch": 3.3474484256243215, + "grad_norm": 8.761766797305249, + "learning_rate": 5.203371740657095e-06, + "loss": 0.3107, + "step": 6166 + }, + { + "epoch": 3.3479913137893593, + "grad_norm": 13.961944426146035, + "learning_rate": 5.2002861764370705e-06, + "loss": 0.4029, + "step": 6167 + }, + { + "epoch": 3.3485342019543975, + "grad_norm": 11.942844125447042, + "learning_rate": 5.1972012058533035e-06, + "loss": 0.2327, + "step": 6168 + }, + { + "epoch": 3.3490770901194353, + "grad_norm": 15.94835762092326, + "learning_rate": 5.194116829287348e-06, + "loss": 0.4844, + "step": 6169 + }, + { + "epoch": 3.3496199782844736, + "grad_norm": 14.492034902597197, + "learning_rate": 5.191033047120682e-06, + "loss": 0.7219, + "step": 6170 + }, + { + "epoch": 3.3501628664495113, + "grad_norm": 16.04200227643707, + "learning_rate": 5.187949859734715e-06, + "loss": 0.4706, + "step": 6171 + }, + { + "epoch": 3.3507057546145496, + "grad_norm": 16.558616346746106, + "learning_rate": 5.184867267510774e-06, + "loss": 0.6355, + "step": 6172 + }, + { + "epoch": 3.3512486427795873, + "grad_norm": 14.202794625614665, + "learning_rate": 5.18178527083013e-06, + "loss": 0.6307, + "step": 6173 + }, + { + "epoch": 3.3517915309446256, + "grad_norm": 13.034510344424431, + "learning_rate": 5.178703870073954e-06, + "loss": 0.614, + "step": 6174 + }, + { + "epoch": 3.3523344191096633, + "grad_norm": 8.768533900570704, + "learning_rate": 5.1756230656233715e-06, + "loss": 0.4036, + "step": 6175 + }, + { + "epoch": 3.3528773072747016, + "grad_norm": 16.580405700504947, + "learning_rate": 5.172542857859418e-06, + "loss": 0.5804, + "step": 6176 + }, + { + "epoch": 3.3534201954397393, + "grad_norm": 11.843457868265668, + "learning_rate": 5.169463247163058e-06, + "loss": 0.5303, + "step": 6177 + }, + { + "epoch": 3.3539630836047776, + "grad_norm": 12.879389378162015, + "learning_rate": 5.166384233915182e-06, + "loss": 0.5253, + "step": 6178 + }, + { + "epoch": 3.3545059717698154, + "grad_norm": 9.5479358372862, + "learning_rate": 5.163305818496607e-06, + "loss": 0.2689, + "step": 6179 + }, + { + "epoch": 3.3550488599348536, + "grad_norm": 11.148812089885375, + "learning_rate": 5.160228001288077e-06, + "loss": 0.5212, + "step": 6180 + }, + { + "epoch": 3.3555917480998914, + "grad_norm": 10.77047326599204, + "learning_rate": 5.157150782670261e-06, + "loss": 0.3674, + "step": 6181 + }, + { + "epoch": 3.3561346362649296, + "grad_norm": 11.804189327484117, + "learning_rate": 5.154074163023756e-06, + "loss": 0.5972, + "step": 6182 + }, + { + "epoch": 3.3566775244299674, + "grad_norm": 11.540083386779209, + "learning_rate": 5.15099814272908e-06, + "loss": 0.703, + "step": 6183 + }, + { + "epoch": 3.3572204125950056, + "grad_norm": 12.780871236150638, + "learning_rate": 5.147922722166683e-06, + "loss": 0.4786, + "step": 6184 + }, + { + "epoch": 3.3577633007600434, + "grad_norm": 11.170640555611453, + "learning_rate": 5.144847901716936e-06, + "loss": 0.6836, + "step": 6185 + }, + { + "epoch": 3.3583061889250816, + "grad_norm": 11.202376321881136, + "learning_rate": 5.1417736817601386e-06, + "loss": 0.327, + "step": 6186 + }, + { + "epoch": 3.3588490770901194, + "grad_norm": 10.62227316496751, + "learning_rate": 5.138700062676516e-06, + "loss": 0.4885, + "step": 6187 + }, + { + "epoch": 3.3593919652551576, + "grad_norm": 11.784247976226597, + "learning_rate": 5.135627044846216e-06, + "loss": 0.4551, + "step": 6188 + }, + { + "epoch": 3.3599348534201954, + "grad_norm": 9.425213682768682, + "learning_rate": 5.132554628649313e-06, + "loss": 0.3285, + "step": 6189 + }, + { + "epoch": 3.3604777415852336, + "grad_norm": 10.622928852300923, + "learning_rate": 5.1294828144658185e-06, + "loss": 0.4265, + "step": 6190 + }, + { + "epoch": 3.3610206297502714, + "grad_norm": 11.666591539214851, + "learning_rate": 5.126411602675649e-06, + "loss": 0.2681, + "step": 6191 + }, + { + "epoch": 3.3615635179153096, + "grad_norm": 19.598513215176627, + "learning_rate": 5.123340993658658e-06, + "loss": 0.7577, + "step": 6192 + }, + { + "epoch": 3.3621064060803474, + "grad_norm": 13.077915822970434, + "learning_rate": 5.120270987794627e-06, + "loss": 0.653, + "step": 6193 + }, + { + "epoch": 3.3626492942453856, + "grad_norm": 15.592283919989452, + "learning_rate": 5.117201585463256e-06, + "loss": 0.4488, + "step": 6194 + }, + { + "epoch": 3.3631921824104234, + "grad_norm": 9.288688824641968, + "learning_rate": 5.114132787044175e-06, + "loss": 0.4185, + "step": 6195 + }, + { + "epoch": 3.3637350705754616, + "grad_norm": 15.306973087542367, + "learning_rate": 5.111064592916935e-06, + "loss": 0.8354, + "step": 6196 + }, + { + "epoch": 3.3642779587404994, + "grad_norm": 11.499674684661942, + "learning_rate": 5.107997003461023e-06, + "loss": 0.5953, + "step": 6197 + }, + { + "epoch": 3.3648208469055376, + "grad_norm": 12.276252733631965, + "learning_rate": 5.104930019055834e-06, + "loss": 0.5574, + "step": 6198 + }, + { + "epoch": 3.3653637350705754, + "grad_norm": 15.0044996135686, + "learning_rate": 5.1018636400807075e-06, + "loss": 0.403, + "step": 6199 + }, + { + "epoch": 3.3659066232356136, + "grad_norm": 8.91478305391761, + "learning_rate": 5.098797866914889e-06, + "loss": 0.3447, + "step": 6200 + }, + { + "epoch": 3.3664495114006514, + "grad_norm": 12.707988441543339, + "learning_rate": 5.095732699937559e-06, + "loss": 0.6121, + "step": 6201 + }, + { + "epoch": 3.3669923995656896, + "grad_norm": 11.893526329599132, + "learning_rate": 5.092668139527831e-06, + "loss": 0.4349, + "step": 6202 + }, + { + "epoch": 3.3675352877307274, + "grad_norm": 9.573100623698757, + "learning_rate": 5.08960418606472e-06, + "loss": 0.3979, + "step": 6203 + }, + { + "epoch": 3.3680781758957656, + "grad_norm": 12.055211170267912, + "learning_rate": 5.0865408399271995e-06, + "loss": 0.7061, + "step": 6204 + }, + { + "epoch": 3.3686210640608034, + "grad_norm": 10.746677950946136, + "learning_rate": 5.08347810149413e-06, + "loss": 0.4687, + "step": 6205 + }, + { + "epoch": 3.3691639522258416, + "grad_norm": 11.592472853876425, + "learning_rate": 5.080415971144332e-06, + "loss": 0.4447, + "step": 6206 + }, + { + "epoch": 3.3697068403908794, + "grad_norm": 14.931359329572437, + "learning_rate": 5.077354449256521e-06, + "loss": 0.7571, + "step": 6207 + }, + { + "epoch": 3.3702497285559176, + "grad_norm": 11.016079444425253, + "learning_rate": 5.07429353620936e-06, + "loss": 0.3964, + "step": 6208 + }, + { + "epoch": 3.3707926167209554, + "grad_norm": 9.567910765955927, + "learning_rate": 5.071233232381425e-06, + "loss": 0.395, + "step": 6209 + }, + { + "epoch": 3.3713355048859937, + "grad_norm": 14.85470239623418, + "learning_rate": 5.0681735381512195e-06, + "loss": 0.557, + "step": 6210 + }, + { + "epoch": 3.3718783930510314, + "grad_norm": 11.383297704214762, + "learning_rate": 5.0651144538971746e-06, + "loss": 0.4522, + "step": 6211 + }, + { + "epoch": 3.3724212812160697, + "grad_norm": 13.445463967563244, + "learning_rate": 5.062055979997631e-06, + "loss": 0.5483, + "step": 6212 + }, + { + "epoch": 3.3729641693811074, + "grad_norm": 12.602441542764687, + "learning_rate": 5.058998116830878e-06, + "loss": 0.6498, + "step": 6213 + }, + { + "epoch": 3.3735070575461457, + "grad_norm": 11.110392326995717, + "learning_rate": 5.055940864775113e-06, + "loss": 0.4893, + "step": 6214 + }, + { + "epoch": 3.3740499457111834, + "grad_norm": 12.47623770525584, + "learning_rate": 5.052884224208461e-06, + "loss": 0.4237, + "step": 6215 + }, + { + "epoch": 3.3745928338762217, + "grad_norm": 9.744420395590724, + "learning_rate": 5.049828195508972e-06, + "loss": 0.2669, + "step": 6216 + }, + { + "epoch": 3.3751357220412594, + "grad_norm": 15.871527506273058, + "learning_rate": 5.046772779054622e-06, + "loss": 0.4374, + "step": 6217 + }, + { + "epoch": 3.3756786102062977, + "grad_norm": 9.442731773451024, + "learning_rate": 5.043717975223308e-06, + "loss": 0.3363, + "step": 6218 + }, + { + "epoch": 3.3762214983713354, + "grad_norm": 9.122598498293966, + "learning_rate": 5.040663784392855e-06, + "loss": 0.366, + "step": 6219 + }, + { + "epoch": 3.3767643865363737, + "grad_norm": 11.713343466114539, + "learning_rate": 5.037610206941009e-06, + "loss": 0.4296, + "step": 6220 + }, + { + "epoch": 3.3773072747014115, + "grad_norm": 9.906918527878904, + "learning_rate": 5.034557243245441e-06, + "loss": 0.3189, + "step": 6221 + }, + { + "epoch": 3.3778501628664497, + "grad_norm": 12.91483922171168, + "learning_rate": 5.031504893683748e-06, + "loss": 0.4095, + "step": 6222 + }, + { + "epoch": 3.3783930510314875, + "grad_norm": 12.288487294242882, + "learning_rate": 5.028453158633448e-06, + "loss": 0.5457, + "step": 6223 + }, + { + "epoch": 3.3789359391965257, + "grad_norm": 8.504130528670414, + "learning_rate": 5.025402038471984e-06, + "loss": 0.3373, + "step": 6224 + }, + { + "epoch": 3.3794788273615635, + "grad_norm": 13.190667855419926, + "learning_rate": 5.022351533576725e-06, + "loss": 0.4901, + "step": 6225 + }, + { + "epoch": 3.3800217155266017, + "grad_norm": 15.552140760061034, + "learning_rate": 5.019301644324961e-06, + "loss": 1.0988, + "step": 6226 + }, + { + "epoch": 3.3805646036916395, + "grad_norm": 9.04502851979091, + "learning_rate": 5.016252371093904e-06, + "loss": 0.3088, + "step": 6227 + }, + { + "epoch": 3.3811074918566777, + "grad_norm": 10.595816125042607, + "learning_rate": 5.0132037142607035e-06, + "loss": 0.4645, + "step": 6228 + }, + { + "epoch": 3.3816503800217155, + "grad_norm": 12.60889622464924, + "learning_rate": 5.010155674202409e-06, + "loss": 0.4484, + "step": 6229 + }, + { + "epoch": 3.3821932681867537, + "grad_norm": 11.424705925472448, + "learning_rate": 5.00710825129602e-06, + "loss": 0.3788, + "step": 6230 + }, + { + "epoch": 3.3827361563517915, + "grad_norm": 12.859629128090067, + "learning_rate": 5.004061445918438e-06, + "loss": 0.6222, + "step": 6231 + }, + { + "epoch": 3.3832790445168297, + "grad_norm": 9.436094386037109, + "learning_rate": 5.001015258446497e-06, + "loss": 0.4005, + "step": 6232 + }, + { + "epoch": 3.3838219326818675, + "grad_norm": 13.270237716590554, + "learning_rate": 4.997969689256957e-06, + "loss": 0.5421, + "step": 6233 + }, + { + "epoch": 3.3843648208469057, + "grad_norm": 8.823605358833275, + "learning_rate": 4.994924738726493e-06, + "loss": 0.3196, + "step": 6234 + }, + { + "epoch": 3.3849077090119435, + "grad_norm": 15.35267771565606, + "learning_rate": 4.991880407231722e-06, + "loss": 0.7645, + "step": 6235 + }, + { + "epoch": 3.3854505971769817, + "grad_norm": 12.948816197321962, + "learning_rate": 4.988836695149156e-06, + "loss": 0.4939, + "step": 6236 + }, + { + "epoch": 3.3859934853420195, + "grad_norm": 11.08333997503354, + "learning_rate": 4.985793602855264e-06, + "loss": 0.386, + "step": 6237 + }, + { + "epoch": 3.3865363735070577, + "grad_norm": 11.954902260666392, + "learning_rate": 4.9827511307264006e-06, + "loss": 0.3809, + "step": 6238 + }, + { + "epoch": 3.3870792616720955, + "grad_norm": 13.826055371275897, + "learning_rate": 4.979709279138879e-06, + "loss": 0.488, + "step": 6239 + }, + { + "epoch": 3.3876221498371337, + "grad_norm": 12.084219110446965, + "learning_rate": 4.976668048468918e-06, + "loss": 0.5239, + "step": 6240 + }, + { + "epoch": 3.3881650380021715, + "grad_norm": 10.947811118815906, + "learning_rate": 4.973627439092651e-06, + "loss": 0.3714, + "step": 6241 + }, + { + "epoch": 3.3887079261672097, + "grad_norm": 12.522059184173699, + "learning_rate": 4.97058745138616e-06, + "loss": 0.689, + "step": 6242 + }, + { + "epoch": 3.3892508143322475, + "grad_norm": 10.527706094398768, + "learning_rate": 4.967548085725423e-06, + "loss": 0.6032, + "step": 6243 + }, + { + "epoch": 3.3897937024972857, + "grad_norm": 13.88984081976927, + "learning_rate": 4.964509342486365e-06, + "loss": 0.7107, + "step": 6244 + }, + { + "epoch": 3.3903365906623235, + "grad_norm": 12.254968564513517, + "learning_rate": 4.961471222044811e-06, + "loss": 0.5961, + "step": 6245 + }, + { + "epoch": 3.3908794788273617, + "grad_norm": 14.460789356302293, + "learning_rate": 4.95843372477653e-06, + "loss": 0.4468, + "step": 6246 + }, + { + "epoch": 3.3914223669923995, + "grad_norm": 11.641301023056258, + "learning_rate": 4.955396851057201e-06, + "loss": 0.4814, + "step": 6247 + }, + { + "epoch": 3.3919652551574377, + "grad_norm": 16.08871509775364, + "learning_rate": 4.9523606012624285e-06, + "loss": 0.6589, + "step": 6248 + }, + { + "epoch": 3.3925081433224755, + "grad_norm": 10.999987937636945, + "learning_rate": 4.9493249757677454e-06, + "loss": 0.4282, + "step": 6249 + }, + { + "epoch": 3.3930510314875137, + "grad_norm": 13.140914544731421, + "learning_rate": 4.946289974948591e-06, + "loss": 0.5447, + "step": 6250 + }, + { + "epoch": 3.3935939196525515, + "grad_norm": 16.948288206914583, + "learning_rate": 4.943255599180352e-06, + "loss": 0.3733, + "step": 6251 + }, + { + "epoch": 3.3941368078175898, + "grad_norm": 13.872557767694476, + "learning_rate": 4.940221848838319e-06, + "loss": 0.407, + "step": 6252 + }, + { + "epoch": 3.3946796959826275, + "grad_norm": 12.01068343710924, + "learning_rate": 4.937188724297713e-06, + "loss": 0.3866, + "step": 6253 + }, + { + "epoch": 3.3952225841476658, + "grad_norm": 12.980406743985375, + "learning_rate": 4.934156225933673e-06, + "loss": 0.6519, + "step": 6254 + }, + { + "epoch": 3.3957654723127035, + "grad_norm": 11.103766528107464, + "learning_rate": 4.931124354121265e-06, + "loss": 0.2982, + "step": 6255 + }, + { + "epoch": 3.3963083604777418, + "grad_norm": 7.2969268723114675, + "learning_rate": 4.928093109235476e-06, + "loss": 0.3553, + "step": 6256 + }, + { + "epoch": 3.3968512486427795, + "grad_norm": 8.742701237748966, + "learning_rate": 4.925062491651213e-06, + "loss": 0.3673, + "step": 6257 + }, + { + "epoch": 3.3973941368078178, + "grad_norm": 11.343078244137898, + "learning_rate": 4.922032501743311e-06, + "loss": 0.5849, + "step": 6258 + }, + { + "epoch": 3.3979370249728555, + "grad_norm": 10.498271159334314, + "learning_rate": 4.919003139886522e-06, + "loss": 0.4092, + "step": 6259 + }, + { + "epoch": 3.3984799131378938, + "grad_norm": 10.97821864845177, + "learning_rate": 4.915974406455522e-06, + "loss": 0.3967, + "step": 6260 + }, + { + "epoch": 3.3990228013029316, + "grad_norm": 13.345358070056497, + "learning_rate": 4.912946301824911e-06, + "loss": 0.61, + "step": 6261 + }, + { + "epoch": 3.3995656894679698, + "grad_norm": 11.235526091701708, + "learning_rate": 4.909918826369209e-06, + "loss": 0.577, + "step": 6262 + }, + { + "epoch": 3.4001085776330076, + "grad_norm": 11.63018111198774, + "learning_rate": 4.9068919804628575e-06, + "loss": 0.5611, + "step": 6263 + }, + { + "epoch": 3.400651465798046, + "grad_norm": 14.484517727976659, + "learning_rate": 4.903865764480224e-06, + "loss": 0.8804, + "step": 6264 + }, + { + "epoch": 3.4011943539630836, + "grad_norm": 13.063076605432173, + "learning_rate": 4.9008401787955964e-06, + "loss": 0.6872, + "step": 6265 + }, + { + "epoch": 3.401737242128122, + "grad_norm": 10.26193448056392, + "learning_rate": 4.89781522378318e-06, + "loss": 0.4086, + "step": 6266 + }, + { + "epoch": 3.4022801302931596, + "grad_norm": 14.7618293483203, + "learning_rate": 4.894790899817106e-06, + "loss": 0.5349, + "step": 6267 + }, + { + "epoch": 3.402823018458198, + "grad_norm": 15.246141002277422, + "learning_rate": 4.8917672072714364e-06, + "loss": 0.4647, + "step": 6268 + }, + { + "epoch": 3.4033659066232356, + "grad_norm": 12.999965847014597, + "learning_rate": 4.888744146520137e-06, + "loss": 0.6288, + "step": 6269 + }, + { + "epoch": 3.403908794788274, + "grad_norm": 12.403108581616499, + "learning_rate": 4.885721717937106e-06, + "loss": 0.4168, + "step": 6270 + }, + { + "epoch": 3.4044516829533116, + "grad_norm": 9.010405275529319, + "learning_rate": 4.882699921896166e-06, + "loss": 0.2848, + "step": 6271 + }, + { + "epoch": 3.40499457111835, + "grad_norm": 9.731197165536528, + "learning_rate": 4.87967875877105e-06, + "loss": 0.4276, + "step": 6272 + }, + { + "epoch": 3.4055374592833876, + "grad_norm": 16.141362422365116, + "learning_rate": 4.876658228935434e-06, + "loss": 0.425, + "step": 6273 + }, + { + "epoch": 3.406080347448426, + "grad_norm": 10.09711287157442, + "learning_rate": 4.873638332762887e-06, + "loss": 0.6085, + "step": 6274 + }, + { + "epoch": 3.4066232356134636, + "grad_norm": 15.877829530437113, + "learning_rate": 4.8706190706269276e-06, + "loss": 0.7737, + "step": 6275 + }, + { + "epoch": 3.407166123778502, + "grad_norm": 16.848740635164727, + "learning_rate": 4.867600442900969e-06, + "loss": 0.9117, + "step": 6276 + }, + { + "epoch": 3.4077090119435396, + "grad_norm": 12.439996859861845, + "learning_rate": 4.8645824499583764e-06, + "loss": 0.6655, + "step": 6277 + }, + { + "epoch": 3.408251900108578, + "grad_norm": 15.742400137527083, + "learning_rate": 4.861565092172402e-06, + "loss": 0.4276, + "step": 6278 + }, + { + "epoch": 3.4087947882736156, + "grad_norm": 10.574321270699965, + "learning_rate": 4.8585483699162505e-06, + "loss": 0.3849, + "step": 6279 + }, + { + "epoch": 3.409337676438654, + "grad_norm": 14.691129894472994, + "learning_rate": 4.8555322835630345e-06, + "loss": 0.8061, + "step": 6280 + }, + { + "epoch": 3.4098805646036916, + "grad_norm": 15.252415938946724, + "learning_rate": 4.852516833485778e-06, + "loss": 0.8019, + "step": 6281 + }, + { + "epoch": 3.41042345276873, + "grad_norm": 15.141237929554537, + "learning_rate": 4.849502020057449e-06, + "loss": 0.8201, + "step": 6282 + }, + { + "epoch": 3.4109663409337676, + "grad_norm": 12.751756702153507, + "learning_rate": 4.846487843650914e-06, + "loss": 0.5871, + "step": 6283 + }, + { + "epoch": 3.411509229098806, + "grad_norm": 10.510847411766605, + "learning_rate": 4.843474304638977e-06, + "loss": 0.5543, + "step": 6284 + }, + { + "epoch": 3.4120521172638436, + "grad_norm": 11.747946589267375, + "learning_rate": 4.8404614033943586e-06, + "loss": 0.5981, + "step": 6285 + }, + { + "epoch": 3.412595005428882, + "grad_norm": 10.03814207660896, + "learning_rate": 4.837449140289696e-06, + "loss": 0.3838, + "step": 6286 + }, + { + "epoch": 3.4131378935939196, + "grad_norm": 9.955014298679115, + "learning_rate": 4.8344375156975525e-06, + "loss": 0.4807, + "step": 6287 + }, + { + "epoch": 3.413680781758958, + "grad_norm": 16.70257123694898, + "learning_rate": 4.8314265299904085e-06, + "loss": 0.5431, + "step": 6288 + }, + { + "epoch": 3.4142236699239956, + "grad_norm": 12.494957483834643, + "learning_rate": 4.828416183540668e-06, + "loss": 0.428, + "step": 6289 + }, + { + "epoch": 3.414766558089034, + "grad_norm": 13.556838718061059, + "learning_rate": 4.825406476720658e-06, + "loss": 0.5487, + "step": 6290 + }, + { + "epoch": 3.4153094462540716, + "grad_norm": 11.716720789445835, + "learning_rate": 4.822397409902622e-06, + "loss": 0.8306, + "step": 6291 + }, + { + "epoch": 3.41585233441911, + "grad_norm": 18.495547243648275, + "learning_rate": 4.819388983458725e-06, + "loss": 1.0386, + "step": 6292 + }, + { + "epoch": 3.4163952225841476, + "grad_norm": 11.600851672889053, + "learning_rate": 4.816381197761055e-06, + "loss": 0.3111, + "step": 6293 + }, + { + "epoch": 3.416938110749186, + "grad_norm": 10.513433670937403, + "learning_rate": 4.813374053181621e-06, + "loss": 0.3881, + "step": 6294 + }, + { + "epoch": 3.4174809989142236, + "grad_norm": 12.120445230198527, + "learning_rate": 4.810367550092349e-06, + "loss": 0.6089, + "step": 6295 + }, + { + "epoch": 3.418023887079262, + "grad_norm": 14.617438317762286, + "learning_rate": 4.807361688865091e-06, + "loss": 0.725, + "step": 6296 + }, + { + "epoch": 3.4185667752442996, + "grad_norm": 13.459119769065767, + "learning_rate": 4.804356469871615e-06, + "loss": 0.678, + "step": 6297 + }, + { + "epoch": 3.419109663409338, + "grad_norm": 11.259674657240234, + "learning_rate": 4.801351893483611e-06, + "loss": 0.4629, + "step": 6298 + }, + { + "epoch": 3.4196525515743756, + "grad_norm": 12.169301874189793, + "learning_rate": 4.7983479600726904e-06, + "loss": 0.6371, + "step": 6299 + }, + { + "epoch": 3.420195439739414, + "grad_norm": 10.211316992248761, + "learning_rate": 4.795344670010385e-06, + "loss": 0.591, + "step": 6300 + }, + { + "epoch": 3.4207383279044516, + "grad_norm": 10.221898080949522, + "learning_rate": 4.792342023668144e-06, + "loss": 0.4058, + "step": 6301 + }, + { + "epoch": 3.42128121606949, + "grad_norm": 11.161796287973816, + "learning_rate": 4.789340021417343e-06, + "loss": 0.6044, + "step": 6302 + }, + { + "epoch": 3.4218241042345277, + "grad_norm": 13.57180649881648, + "learning_rate": 4.7863386636292705e-06, + "loss": 0.5895, + "step": 6303 + }, + { + "epoch": 3.422366992399566, + "grad_norm": 14.249903450132871, + "learning_rate": 4.783337950675143e-06, + "loss": 0.9115, + "step": 6304 + }, + { + "epoch": 3.4229098805646037, + "grad_norm": 8.778670420410771, + "learning_rate": 4.780337882926088e-06, + "loss": 0.368, + "step": 6305 + }, + { + "epoch": 3.423452768729642, + "grad_norm": 14.418437807091719, + "learning_rate": 4.77733846075317e-06, + "loss": 0.7248, + "step": 6306 + }, + { + "epoch": 3.4239956568946797, + "grad_norm": 12.207383551156253, + "learning_rate": 4.774339684527348e-06, + "loss": 0.715, + "step": 6307 + }, + { + "epoch": 3.424538545059718, + "grad_norm": 12.625928846965705, + "learning_rate": 4.7713415546195285e-06, + "loss": 0.4691, + "step": 6308 + }, + { + "epoch": 3.4250814332247557, + "grad_norm": 12.54338617690595, + "learning_rate": 4.768344071400516e-06, + "loss": 0.7746, + "step": 6309 + }, + { + "epoch": 3.425624321389794, + "grad_norm": 12.757091279673624, + "learning_rate": 4.765347235241042e-06, + "loss": 0.5129, + "step": 6310 + }, + { + "epoch": 3.4261672095548317, + "grad_norm": 14.719338319419373, + "learning_rate": 4.762351046511774e-06, + "loss": 0.6281, + "step": 6311 + }, + { + "epoch": 3.42671009771987, + "grad_norm": 14.600430485134494, + "learning_rate": 4.759355505583267e-06, + "loss": 0.6332, + "step": 6312 + }, + { + "epoch": 3.4272529858849077, + "grad_norm": 12.499065575741712, + "learning_rate": 4.756360612826032e-06, + "loss": 0.5188, + "step": 6313 + }, + { + "epoch": 3.427795874049946, + "grad_norm": 13.595282233685417, + "learning_rate": 4.753366368610466e-06, + "loss": 0.6934, + "step": 6314 + }, + { + "epoch": 3.4283387622149837, + "grad_norm": 18.513402859066886, + "learning_rate": 4.750372773306916e-06, + "loss": 0.8336, + "step": 6315 + }, + { + "epoch": 3.428881650380022, + "grad_norm": 12.541674490834746, + "learning_rate": 4.747379827285621e-06, + "loss": 0.4741, + "step": 6316 + }, + { + "epoch": 3.4294245385450597, + "grad_norm": 13.06139830486867, + "learning_rate": 4.744387530916764e-06, + "loss": 0.6489, + "step": 6317 + }, + { + "epoch": 3.429967426710098, + "grad_norm": 9.57665402322532, + "learning_rate": 4.741395884570437e-06, + "loss": 0.4976, + "step": 6318 + }, + { + "epoch": 3.4305103148751357, + "grad_norm": 9.338701533011545, + "learning_rate": 4.738404888616641e-06, + "loss": 0.4947, + "step": 6319 + }, + { + "epoch": 3.431053203040174, + "grad_norm": 13.17347762890872, + "learning_rate": 4.735414543425321e-06, + "loss": 0.5455, + "step": 6320 + }, + { + "epoch": 3.4315960912052117, + "grad_norm": 14.408997044186131, + "learning_rate": 4.732424849366314e-06, + "loss": 0.5761, + "step": 6321 + }, + { + "epoch": 3.43213897937025, + "grad_norm": 10.201121637580389, + "learning_rate": 4.729435806809401e-06, + "loss": 0.4271, + "step": 6322 + }, + { + "epoch": 3.4326818675352877, + "grad_norm": 11.714865692839483, + "learning_rate": 4.726447416124266e-06, + "loss": 0.4182, + "step": 6323 + }, + { + "epoch": 3.433224755700326, + "grad_norm": 21.671744658819392, + "learning_rate": 4.72345967768052e-06, + "loss": 1.1518, + "step": 6324 + }, + { + "epoch": 3.4337676438653637, + "grad_norm": 10.904130560504068, + "learning_rate": 4.72047259184769e-06, + "loss": 0.533, + "step": 6325 + }, + { + "epoch": 3.434310532030402, + "grad_norm": 17.26082351569084, + "learning_rate": 4.717486158995225e-06, + "loss": 0.5579, + "step": 6326 + }, + { + "epoch": 3.4348534201954397, + "grad_norm": 13.562029451071414, + "learning_rate": 4.7145003794924905e-06, + "loss": 0.6275, + "step": 6327 + }, + { + "epoch": 3.435396308360478, + "grad_norm": 10.679128408418185, + "learning_rate": 4.711515253708774e-06, + "loss": 0.3116, + "step": 6328 + }, + { + "epoch": 3.4359391965255157, + "grad_norm": 6.813829973820327, + "learning_rate": 4.708530782013277e-06, + "loss": 0.2313, + "step": 6329 + }, + { + "epoch": 3.436482084690554, + "grad_norm": 10.696300631140947, + "learning_rate": 4.705546964775128e-06, + "loss": 0.4345, + "step": 6330 + }, + { + "epoch": 3.4370249728555917, + "grad_norm": 15.665510743577896, + "learning_rate": 4.702563802363369e-06, + "loss": 0.6155, + "step": 6331 + }, + { + "epoch": 3.4375678610206295, + "grad_norm": 9.437457272265185, + "learning_rate": 4.699581295146961e-06, + "loss": 0.3972, + "step": 6332 + }, + { + "epoch": 3.4381107491856677, + "grad_norm": 11.253233608357364, + "learning_rate": 4.696599443494787e-06, + "loss": 0.4871, + "step": 6333 + }, + { + "epoch": 3.438653637350706, + "grad_norm": 13.323571403470238, + "learning_rate": 4.693618247775645e-06, + "loss": 0.4443, + "step": 6334 + }, + { + "epoch": 3.4391965255157437, + "grad_norm": 10.853976251462788, + "learning_rate": 4.6906377083582556e-06, + "loss": 0.4012, + "step": 6335 + }, + { + "epoch": 3.4397394136807815, + "grad_norm": 11.483324693854579, + "learning_rate": 4.687657825611256e-06, + "loss": 0.4913, + "step": 6336 + }, + { + "epoch": 3.4402823018458197, + "grad_norm": 16.90515830826938, + "learning_rate": 4.684678599903204e-06, + "loss": 0.5872, + "step": 6337 + }, + { + "epoch": 3.440825190010858, + "grad_norm": 11.339161987387357, + "learning_rate": 4.681700031602573e-06, + "loss": 0.484, + "step": 6338 + }, + { + "epoch": 3.4413680781758957, + "grad_norm": 13.116517374892055, + "learning_rate": 4.678722121077759e-06, + "loss": 0.8588, + "step": 6339 + }, + { + "epoch": 3.4419109663409335, + "grad_norm": 10.313990325741717, + "learning_rate": 4.675744868697073e-06, + "loss": 0.4864, + "step": 6340 + }, + { + "epoch": 3.4424538545059717, + "grad_norm": 12.464768763015064, + "learning_rate": 4.672768274828748e-06, + "loss": 0.7793, + "step": 6341 + }, + { + "epoch": 3.44299674267101, + "grad_norm": 7.491565779729114, + "learning_rate": 4.669792339840933e-06, + "loss": 0.2551, + "step": 6342 + }, + { + "epoch": 3.4435396308360477, + "grad_norm": 11.582532266159841, + "learning_rate": 4.666817064101693e-06, + "loss": 0.4687, + "step": 6343 + }, + { + "epoch": 3.4440825190010855, + "grad_norm": 13.223648734250315, + "learning_rate": 4.663842447979026e-06, + "loss": 0.5676, + "step": 6344 + }, + { + "epoch": 3.4446254071661238, + "grad_norm": 9.420907388669441, + "learning_rate": 4.660868491840821e-06, + "loss": 0.4091, + "step": 6345 + }, + { + "epoch": 3.445168295331162, + "grad_norm": 10.937341000869324, + "learning_rate": 4.657895196054919e-06, + "loss": 0.4299, + "step": 6346 + }, + { + "epoch": 3.4457111834961998, + "grad_norm": 10.997320970308403, + "learning_rate": 4.654922560989049e-06, + "loss": 0.4176, + "step": 6347 + }, + { + "epoch": 3.4462540716612375, + "grad_norm": 11.008951815202312, + "learning_rate": 4.651950587010875e-06, + "loss": 0.3343, + "step": 6348 + }, + { + "epoch": 3.4467969598262758, + "grad_norm": 13.714805717613569, + "learning_rate": 4.6489792744879755e-06, + "loss": 0.8672, + "step": 6349 + }, + { + "epoch": 3.447339847991314, + "grad_norm": 12.722366469633661, + "learning_rate": 4.646008623787845e-06, + "loss": 0.601, + "step": 6350 + }, + { + "epoch": 3.4478827361563518, + "grad_norm": 10.3662276435613, + "learning_rate": 4.643038635277908e-06, + "loss": 0.4597, + "step": 6351 + }, + { + "epoch": 3.4484256243213895, + "grad_norm": 11.28259690900768, + "learning_rate": 4.640069309325484e-06, + "loss": 0.4909, + "step": 6352 + }, + { + "epoch": 3.4489685124864278, + "grad_norm": 15.163346859296523, + "learning_rate": 4.6371006462978355e-06, + "loss": 0.8422, + "step": 6353 + }, + { + "epoch": 3.449511400651466, + "grad_norm": 9.625566221898923, + "learning_rate": 4.634132646562119e-06, + "loss": 0.2858, + "step": 6354 + }, + { + "epoch": 3.450054288816504, + "grad_norm": 14.927715657666415, + "learning_rate": 4.631165310485434e-06, + "loss": 0.5556, + "step": 6355 + }, + { + "epoch": 3.4505971769815416, + "grad_norm": 14.575103603282974, + "learning_rate": 4.62819863843478e-06, + "loss": 0.5422, + "step": 6356 + }, + { + "epoch": 3.45114006514658, + "grad_norm": 11.84763157429604, + "learning_rate": 4.625232630777079e-06, + "loss": 0.5382, + "step": 6357 + }, + { + "epoch": 3.451682953311618, + "grad_norm": 14.334352386311854, + "learning_rate": 4.622267287879176e-06, + "loss": 0.6771, + "step": 6358 + }, + { + "epoch": 3.452225841476656, + "grad_norm": 13.03121716154961, + "learning_rate": 4.619302610107819e-06, + "loss": 0.702, + "step": 6359 + }, + { + "epoch": 3.4527687296416936, + "grad_norm": 9.331217131822793, + "learning_rate": 4.616338597829697e-06, + "loss": 0.4664, + "step": 6360 + }, + { + "epoch": 3.453311617806732, + "grad_norm": 16.384420267273573, + "learning_rate": 4.61337525141139e-06, + "loss": 1.1202, + "step": 6361 + }, + { + "epoch": 3.45385450597177, + "grad_norm": 12.656834373054709, + "learning_rate": 4.610412571219421e-06, + "loss": 0.8405, + "step": 6362 + }, + { + "epoch": 3.454397394136808, + "grad_norm": 10.6285388663689, + "learning_rate": 4.607450557620216e-06, + "loss": 0.4027, + "step": 6363 + }, + { + "epoch": 3.4549402823018456, + "grad_norm": 14.149860401799353, + "learning_rate": 4.60448921098012e-06, + "loss": 0.5762, + "step": 6364 + }, + { + "epoch": 3.455483170466884, + "grad_norm": 15.249693808816065, + "learning_rate": 4.601528531665397e-06, + "loss": 1.0724, + "step": 6365 + }, + { + "epoch": 3.456026058631922, + "grad_norm": 11.967042611547605, + "learning_rate": 4.598568520042229e-06, + "loss": 0.572, + "step": 6366 + }, + { + "epoch": 3.45656894679696, + "grad_norm": 13.806580531191415, + "learning_rate": 4.595609176476715e-06, + "loss": 0.7437, + "step": 6367 + }, + { + "epoch": 3.4571118349619976, + "grad_norm": 15.062073427987208, + "learning_rate": 4.592650501334872e-06, + "loss": 0.6357, + "step": 6368 + }, + { + "epoch": 3.457654723127036, + "grad_norm": 8.804207344489894, + "learning_rate": 4.589692494982632e-06, + "loss": 0.3561, + "step": 6369 + }, + { + "epoch": 3.458197611292074, + "grad_norm": 13.940651501182272, + "learning_rate": 4.5867351577858475e-06, + "loss": 0.6055, + "step": 6370 + }, + { + "epoch": 3.458740499457112, + "grad_norm": 12.874778392486414, + "learning_rate": 4.583778490110287e-06, + "loss": 0.466, + "step": 6371 + }, + { + "epoch": 3.4592833876221496, + "grad_norm": 13.362591165530915, + "learning_rate": 4.580822492321634e-06, + "loss": 0.837, + "step": 6372 + }, + { + "epoch": 3.459826275787188, + "grad_norm": 13.501338998690153, + "learning_rate": 4.577867164785492e-06, + "loss": 0.4919, + "step": 6373 + }, + { + "epoch": 3.460369163952226, + "grad_norm": 8.394152440486542, + "learning_rate": 4.574912507867382e-06, + "loss": 0.3555, + "step": 6374 + }, + { + "epoch": 3.460912052117264, + "grad_norm": 11.611965148146934, + "learning_rate": 4.571958521932738e-06, + "loss": 0.4808, + "step": 6375 + }, + { + "epoch": 3.4614549402823016, + "grad_norm": 11.297446435669148, + "learning_rate": 4.569005207346911e-06, + "loss": 0.4093, + "step": 6376 + }, + { + "epoch": 3.46199782844734, + "grad_norm": 13.040685712021983, + "learning_rate": 4.566052564475184e-06, + "loss": 0.5487, + "step": 6377 + }, + { + "epoch": 3.462540716612378, + "grad_norm": 10.160191141450042, + "learning_rate": 4.563100593682732e-06, + "loss": 0.5555, + "step": 6378 + }, + { + "epoch": 3.463083604777416, + "grad_norm": 10.641601715805423, + "learning_rate": 4.560149295334664e-06, + "loss": 0.4319, + "step": 6379 + }, + { + "epoch": 3.4636264929424536, + "grad_norm": 11.742982035764232, + "learning_rate": 4.557198669796001e-06, + "loss": 0.77, + "step": 6380 + }, + { + "epoch": 3.464169381107492, + "grad_norm": 15.297545906946976, + "learning_rate": 4.554248717431678e-06, + "loss": 0.753, + "step": 6381 + }, + { + "epoch": 3.46471226927253, + "grad_norm": 11.061047467202666, + "learning_rate": 4.551299438606559e-06, + "loss": 0.428, + "step": 6382 + }, + { + "epoch": 3.465255157437568, + "grad_norm": 12.238086621939612, + "learning_rate": 4.548350833685402e-06, + "loss": 0.4874, + "step": 6383 + }, + { + "epoch": 3.4657980456026056, + "grad_norm": 13.316349313453275, + "learning_rate": 4.54540290303291e-06, + "loss": 0.7391, + "step": 6384 + }, + { + "epoch": 3.466340933767644, + "grad_norm": 14.524029697923662, + "learning_rate": 4.5424556470136735e-06, + "loss": 0.6782, + "step": 6385 + }, + { + "epoch": 3.466883821932682, + "grad_norm": 11.699705038001243, + "learning_rate": 4.539509065992229e-06, + "loss": 0.6561, + "step": 6386 + }, + { + "epoch": 3.46742671009772, + "grad_norm": 16.5251401693806, + "learning_rate": 4.536563160333001e-06, + "loss": 0.6788, + "step": 6387 + }, + { + "epoch": 3.4679695982627576, + "grad_norm": 8.228129211788293, + "learning_rate": 4.533617930400345e-06, + "loss": 0.4383, + "step": 6388 + }, + { + "epoch": 3.468512486427796, + "grad_norm": 10.016870671639179, + "learning_rate": 4.530673376558543e-06, + "loss": 0.5374, + "step": 6389 + }, + { + "epoch": 3.469055374592834, + "grad_norm": 12.529286864818076, + "learning_rate": 4.527729499171767e-06, + "loss": 0.3919, + "step": 6390 + }, + { + "epoch": 3.469598262757872, + "grad_norm": 13.31372906596201, + "learning_rate": 4.524786298604136e-06, + "loss": 0.7436, + "step": 6391 + }, + { + "epoch": 3.4701411509229096, + "grad_norm": 11.420174160084546, + "learning_rate": 4.521843775219654e-06, + "loss": 0.4285, + "step": 6392 + }, + { + "epoch": 3.470684039087948, + "grad_norm": 11.461560335128281, + "learning_rate": 4.518901929382267e-06, + "loss": 0.5651, + "step": 6393 + }, + { + "epoch": 3.471226927252986, + "grad_norm": 14.298269585226816, + "learning_rate": 4.515960761455827e-06, + "loss": 0.3989, + "step": 6394 + }, + { + "epoch": 3.471769815418024, + "grad_norm": 12.421864505486035, + "learning_rate": 4.5130202718041004e-06, + "loss": 0.8773, + "step": 6395 + }, + { + "epoch": 3.4723127035830617, + "grad_norm": 12.947288721903634, + "learning_rate": 4.510080460790775e-06, + "loss": 0.5396, + "step": 6396 + }, + { + "epoch": 3.4728555917481, + "grad_norm": 10.339717202308544, + "learning_rate": 4.507141328779439e-06, + "loss": 0.5154, + "step": 6397 + }, + { + "epoch": 3.473398479913138, + "grad_norm": 8.810723375318315, + "learning_rate": 4.504202876133627e-06, + "loss": 0.4513, + "step": 6398 + }, + { + "epoch": 3.473941368078176, + "grad_norm": 13.655631619886284, + "learning_rate": 4.501265103216755e-06, + "loss": 0.5043, + "step": 6399 + }, + { + "epoch": 3.4744842562432137, + "grad_norm": 9.947920274498516, + "learning_rate": 4.49832801039218e-06, + "loss": 0.6907, + "step": 6400 + }, + { + "epoch": 3.475027144408252, + "grad_norm": 9.800586651795994, + "learning_rate": 4.495391598023167e-06, + "loss": 0.3988, + "step": 6401 + }, + { + "epoch": 3.47557003257329, + "grad_norm": 14.696382692377027, + "learning_rate": 4.492455866472894e-06, + "loss": 0.5805, + "step": 6402 + }, + { + "epoch": 3.476112920738328, + "grad_norm": 13.06051257277803, + "learning_rate": 4.489520816104455e-06, + "loss": 0.7204, + "step": 6403 + }, + { + "epoch": 3.4766558089033657, + "grad_norm": 12.99399781439784, + "learning_rate": 4.486586447280866e-06, + "loss": 0.7112, + "step": 6404 + }, + { + "epoch": 3.477198697068404, + "grad_norm": 12.422813336616981, + "learning_rate": 4.483652760365052e-06, + "loss": 0.3927, + "step": 6405 + }, + { + "epoch": 3.477741585233442, + "grad_norm": 10.948217532391471, + "learning_rate": 4.480719755719857e-06, + "loss": 0.5587, + "step": 6406 + }, + { + "epoch": 3.47828447339848, + "grad_norm": 11.657205565439646, + "learning_rate": 4.477787433708038e-06, + "loss": 0.5584, + "step": 6407 + }, + { + "epoch": 3.4788273615635177, + "grad_norm": 7.795661051863107, + "learning_rate": 4.474855794692271e-06, + "loss": 0.3232, + "step": 6408 + }, + { + "epoch": 3.479370249728556, + "grad_norm": 7.775932230943303, + "learning_rate": 4.4719248390351446e-06, + "loss": 0.3613, + "step": 6409 + }, + { + "epoch": 3.479913137893594, + "grad_norm": 9.097718161588768, + "learning_rate": 4.468994567099165e-06, + "loss": 0.3896, + "step": 6410 + }, + { + "epoch": 3.480456026058632, + "grad_norm": 7.087231016744389, + "learning_rate": 4.466064979246751e-06, + "loss": 0.2486, + "step": 6411 + }, + { + "epoch": 3.4809989142236697, + "grad_norm": 17.009119363822553, + "learning_rate": 4.463136075840242e-06, + "loss": 0.7777, + "step": 6412 + }, + { + "epoch": 3.481541802388708, + "grad_norm": 11.54915190749092, + "learning_rate": 4.460207857241887e-06, + "loss": 0.4528, + "step": 6413 + }, + { + "epoch": 3.482084690553746, + "grad_norm": 14.237806827589361, + "learning_rate": 4.45728032381385e-06, + "loss": 0.8941, + "step": 6414 + }, + { + "epoch": 3.482627578718784, + "grad_norm": 11.85837340138276, + "learning_rate": 4.454353475918223e-06, + "loss": 0.5289, + "step": 6415 + }, + { + "epoch": 3.4831704668838217, + "grad_norm": 13.987937632959078, + "learning_rate": 4.4514273139169925e-06, + "loss": 0.5152, + "step": 6416 + }, + { + "epoch": 3.48371335504886, + "grad_norm": 10.643372071284285, + "learning_rate": 4.4485018381720755e-06, + "loss": 0.4758, + "step": 6417 + }, + { + "epoch": 3.484256243213898, + "grad_norm": 8.692254926051978, + "learning_rate": 4.445577049045299e-06, + "loss": 0.382, + "step": 6418 + }, + { + "epoch": 3.484799131378936, + "grad_norm": 9.767782914786244, + "learning_rate": 4.4426529468984055e-06, + "loss": 0.4674, + "step": 6419 + }, + { + "epoch": 3.4853420195439737, + "grad_norm": 14.118719943852668, + "learning_rate": 4.4397295320930525e-06, + "loss": 0.5859, + "step": 6420 + }, + { + "epoch": 3.485884907709012, + "grad_norm": 9.120918716734044, + "learning_rate": 4.4368068049908085e-06, + "loss": 0.37, + "step": 6421 + }, + { + "epoch": 3.48642779587405, + "grad_norm": 12.706215416375834, + "learning_rate": 4.4338847659531735e-06, + "loss": 0.8322, + "step": 6422 + }, + { + "epoch": 3.486970684039088, + "grad_norm": 10.926968726715666, + "learning_rate": 4.430963415341533e-06, + "loss": 0.4146, + "step": 6423 + }, + { + "epoch": 3.4875135722041257, + "grad_norm": 14.53230992435977, + "learning_rate": 4.428042753517222e-06, + "loss": 0.5612, + "step": 6424 + }, + { + "epoch": 3.488056460369164, + "grad_norm": 12.614428085329433, + "learning_rate": 4.425122780841456e-06, + "loss": 0.4886, + "step": 6425 + }, + { + "epoch": 3.488599348534202, + "grad_norm": 9.07248478126251, + "learning_rate": 4.422203497675394e-06, + "loss": 0.3342, + "step": 6426 + }, + { + "epoch": 3.48914223669924, + "grad_norm": 7.63850442052429, + "learning_rate": 4.419284904380095e-06, + "loss": 0.2709, + "step": 6427 + }, + { + "epoch": 3.4896851248642777, + "grad_norm": 8.139657158702438, + "learning_rate": 4.416367001316526e-06, + "loss": 0.3276, + "step": 6428 + }, + { + "epoch": 3.490228013029316, + "grad_norm": 12.800427760538255, + "learning_rate": 4.413449788845594e-06, + "loss": 0.8097, + "step": 6429 + }, + { + "epoch": 3.490770901194354, + "grad_norm": 9.88357123496715, + "learning_rate": 4.410533267328087e-06, + "loss": 0.3224, + "step": 6430 + }, + { + "epoch": 3.491313789359392, + "grad_norm": 10.15573231324521, + "learning_rate": 4.407617437124741e-06, + "loss": 0.5535, + "step": 6431 + }, + { + "epoch": 3.4918566775244297, + "grad_norm": 17.972024507407102, + "learning_rate": 4.404702298596177e-06, + "loss": 0.9413, + "step": 6432 + }, + { + "epoch": 3.492399565689468, + "grad_norm": 11.366140193603957, + "learning_rate": 4.401787852102955e-06, + "loss": 0.7224, + "step": 6433 + }, + { + "epoch": 3.492942453854506, + "grad_norm": 11.183293759896978, + "learning_rate": 4.398874098005532e-06, + "loss": 0.6597, + "step": 6434 + }, + { + "epoch": 3.493485342019544, + "grad_norm": 11.409722406543372, + "learning_rate": 4.395961036664288e-06, + "loss": 0.6165, + "step": 6435 + }, + { + "epoch": 3.4940282301845818, + "grad_norm": 8.532518886519954, + "learning_rate": 4.393048668439518e-06, + "loss": 0.3045, + "step": 6436 + }, + { + "epoch": 3.49457111834962, + "grad_norm": 15.499278790978442, + "learning_rate": 4.390136993691417e-06, + "loss": 0.6505, + "step": 6437 + }, + { + "epoch": 3.495114006514658, + "grad_norm": 9.855565799480534, + "learning_rate": 4.387226012780117e-06, + "loss": 0.4787, + "step": 6438 + }, + { + "epoch": 3.495656894679696, + "grad_norm": 13.510584259969823, + "learning_rate": 4.38431572606565e-06, + "loss": 0.5388, + "step": 6439 + }, + { + "epoch": 3.4961997828447338, + "grad_norm": 8.040242399067179, + "learning_rate": 4.381406133907964e-06, + "loss": 0.3596, + "step": 6440 + }, + { + "epoch": 3.496742671009772, + "grad_norm": 9.402081395938351, + "learning_rate": 4.378497236666922e-06, + "loss": 0.3326, + "step": 6441 + }, + { + "epoch": 3.49728555917481, + "grad_norm": 15.188738381211488, + "learning_rate": 4.3755890347023e-06, + "loss": 0.7356, + "step": 6442 + }, + { + "epoch": 3.497828447339848, + "grad_norm": 13.644691532416777, + "learning_rate": 4.372681528373791e-06, + "loss": 0.5727, + "step": 6443 + }, + { + "epoch": 3.4983713355048858, + "grad_norm": 12.036428423892223, + "learning_rate": 4.369774718041e-06, + "loss": 0.4333, + "step": 6444 + }, + { + "epoch": 3.498914223669924, + "grad_norm": 9.814264562422439, + "learning_rate": 4.366868604063444e-06, + "loss": 0.4396, + "step": 6445 + }, + { + "epoch": 3.499457111834962, + "grad_norm": 10.623481464500886, + "learning_rate": 4.363963186800557e-06, + "loss": 0.509, + "step": 6446 + }, + { + "epoch": 3.5, + "grad_norm": 9.931711278540687, + "learning_rate": 4.361058466611686e-06, + "loss": 0.352, + "step": 6447 + }, + { + "epoch": 3.500542888165038, + "grad_norm": 10.709530760490054, + "learning_rate": 4.358154443856091e-06, + "loss": 0.3233, + "step": 6448 + }, + { + "epoch": 3.501085776330076, + "grad_norm": 12.489815680915486, + "learning_rate": 4.355251118892946e-06, + "loss": 0.8317, + "step": 6449 + }, + { + "epoch": 3.5016286644951142, + "grad_norm": 13.400967427781875, + "learning_rate": 4.35234849208134e-06, + "loss": 0.601, + "step": 6450 + }, + { + "epoch": 3.502171552660152, + "grad_norm": 11.710683499628859, + "learning_rate": 4.349446563780272e-06, + "loss": 0.7362, + "step": 6451 + }, + { + "epoch": 3.50271444082519, + "grad_norm": 14.51448957958314, + "learning_rate": 4.346545334348658e-06, + "loss": 0.8007, + "step": 6452 + }, + { + "epoch": 3.503257328990228, + "grad_norm": 10.464776296844045, + "learning_rate": 4.343644804145329e-06, + "loss": 0.3993, + "step": 6453 + }, + { + "epoch": 3.5038002171552662, + "grad_norm": 9.718005412495163, + "learning_rate": 4.340744973529022e-06, + "loss": 0.269, + "step": 6454 + }, + { + "epoch": 3.504343105320304, + "grad_norm": 10.85577962939331, + "learning_rate": 4.337845842858402e-06, + "loss": 0.5342, + "step": 6455 + }, + { + "epoch": 3.504885993485342, + "grad_norm": 12.181308768808934, + "learning_rate": 4.33494741249203e-06, + "loss": 0.4005, + "step": 6456 + }, + { + "epoch": 3.50542888165038, + "grad_norm": 11.779746087085687, + "learning_rate": 4.332049682788391e-06, + "loss": 0.4414, + "step": 6457 + }, + { + "epoch": 3.5059717698154182, + "grad_norm": 11.664912784701475, + "learning_rate": 4.3291526541058795e-06, + "loss": 0.548, + "step": 6458 + }, + { + "epoch": 3.506514657980456, + "grad_norm": 13.373343929291233, + "learning_rate": 4.326256326802802e-06, + "loss": 0.564, + "step": 6459 + }, + { + "epoch": 3.507057546145494, + "grad_norm": 11.223466720082536, + "learning_rate": 4.3233607012373925e-06, + "loss": 0.3611, + "step": 6460 + }, + { + "epoch": 3.507600434310532, + "grad_norm": 13.530570397967848, + "learning_rate": 4.32046577776777e-06, + "loss": 0.6113, + "step": 6461 + }, + { + "epoch": 3.5081433224755703, + "grad_norm": 12.62657182669542, + "learning_rate": 4.3175715567520015e-06, + "loss": 0.4457, + "step": 6462 + }, + { + "epoch": 3.508686210640608, + "grad_norm": 14.69985392835765, + "learning_rate": 4.31467803854803e-06, + "loss": 0.8015, + "step": 6463 + }, + { + "epoch": 3.509229098805646, + "grad_norm": 12.910242277352095, + "learning_rate": 4.311785223513744e-06, + "loss": 0.5957, + "step": 6464 + }, + { + "epoch": 3.509771986970684, + "grad_norm": 14.38667526612287, + "learning_rate": 4.308893112006932e-06, + "loss": 0.6333, + "step": 6465 + }, + { + "epoch": 3.5103148751357223, + "grad_norm": 10.811689404016661, + "learning_rate": 4.306001704385282e-06, + "loss": 0.379, + "step": 6466 + }, + { + "epoch": 3.51085776330076, + "grad_norm": 12.950655194202454, + "learning_rate": 4.303111001006424e-06, + "loss": 0.3791, + "step": 6467 + }, + { + "epoch": 3.511400651465798, + "grad_norm": 10.403413919758718, + "learning_rate": 4.300221002227869e-06, + "loss": 0.5023, + "step": 6468 + }, + { + "epoch": 3.511943539630836, + "grad_norm": 14.86724875684143, + "learning_rate": 4.297331708407072e-06, + "loss": 0.5985, + "step": 6469 + }, + { + "epoch": 3.5124864277958743, + "grad_norm": 9.211711120681478, + "learning_rate": 4.29444311990137e-06, + "loss": 0.2738, + "step": 6470 + }, + { + "epoch": 3.513029315960912, + "grad_norm": 11.398896094597806, + "learning_rate": 4.29155523706804e-06, + "loss": 0.6349, + "step": 6471 + }, + { + "epoch": 3.51357220412595, + "grad_norm": 9.97395797119566, + "learning_rate": 4.288668060264257e-06, + "loss": 0.3945, + "step": 6472 + }, + { + "epoch": 3.514115092290988, + "grad_norm": 10.965919373159068, + "learning_rate": 4.2857815898471114e-06, + "loss": 0.6313, + "step": 6473 + }, + { + "epoch": 3.5146579804560263, + "grad_norm": 14.272003668320222, + "learning_rate": 4.2828958261736045e-06, + "loss": 0.8566, + "step": 6474 + }, + { + "epoch": 3.515200868621064, + "grad_norm": 9.155542594309077, + "learning_rate": 4.280010769600653e-06, + "loss": 0.4155, + "step": 6475 + }, + { + "epoch": 3.515743756786102, + "grad_norm": 12.664316762774577, + "learning_rate": 4.277126420485087e-06, + "loss": 0.4668, + "step": 6476 + }, + { + "epoch": 3.51628664495114, + "grad_norm": 13.218980598800185, + "learning_rate": 4.274242779183646e-06, + "loss": 0.5638, + "step": 6477 + }, + { + "epoch": 3.5168295331161783, + "grad_norm": 10.791302144957823, + "learning_rate": 4.271359846052983e-06, + "loss": 0.5458, + "step": 6478 + }, + { + "epoch": 3.517372421281216, + "grad_norm": 11.294781461737724, + "learning_rate": 4.268477621449666e-06, + "loss": 0.5746, + "step": 6479 + }, + { + "epoch": 3.517915309446254, + "grad_norm": 8.318333652865173, + "learning_rate": 4.2655961057301695e-06, + "loss": 0.3447, + "step": 6480 + }, + { + "epoch": 3.518458197611292, + "grad_norm": 14.570486409127577, + "learning_rate": 4.2627152992508865e-06, + "loss": 0.6757, + "step": 6481 + }, + { + "epoch": 3.5190010857763303, + "grad_norm": 12.036103572116089, + "learning_rate": 4.25983520236812e-06, + "loss": 0.4259, + "step": 6482 + }, + { + "epoch": 3.519543973941368, + "grad_norm": 9.254932202972295, + "learning_rate": 4.256955815438084e-06, + "loss": 0.4689, + "step": 6483 + }, + { + "epoch": 3.520086862106406, + "grad_norm": 10.562719431036207, + "learning_rate": 4.254077138816905e-06, + "loss": 0.3951, + "step": 6484 + }, + { + "epoch": 3.520629750271444, + "grad_norm": 10.475306089406416, + "learning_rate": 4.251199172860624e-06, + "loss": 0.6583, + "step": 6485 + }, + { + "epoch": 3.5211726384364823, + "grad_norm": 11.146639535946015, + "learning_rate": 4.2483219179251915e-06, + "loss": 0.5873, + "step": 6486 + }, + { + "epoch": 3.52171552660152, + "grad_norm": 13.607190728220559, + "learning_rate": 4.245445374366469e-06, + "loss": 0.5111, + "step": 6487 + }, + { + "epoch": 3.522258414766558, + "grad_norm": 13.73052324797901, + "learning_rate": 4.242569542540236e-06, + "loss": 0.6546, + "step": 6488 + }, + { + "epoch": 3.522801302931596, + "grad_norm": 11.846325790998266, + "learning_rate": 4.239694422802177e-06, + "loss": 0.3475, + "step": 6489 + }, + { + "epoch": 3.5233441910966343, + "grad_norm": 9.944462539493102, + "learning_rate": 4.236820015507893e-06, + "loss": 0.3798, + "step": 6490 + }, + { + "epoch": 3.523887079261672, + "grad_norm": 11.901268081862414, + "learning_rate": 4.233946321012895e-06, + "loss": 0.503, + "step": 6491 + }, + { + "epoch": 3.52442996742671, + "grad_norm": 11.373264829742677, + "learning_rate": 4.231073339672601e-06, + "loss": 0.3985, + "step": 6492 + }, + { + "epoch": 3.524972855591748, + "grad_norm": 8.462656388647403, + "learning_rate": 4.228201071842358e-06, + "loss": 0.3526, + "step": 6493 + }, + { + "epoch": 3.5255157437567863, + "grad_norm": 13.319319520255643, + "learning_rate": 4.2253295178774036e-06, + "loss": 0.3119, + "step": 6494 + }, + { + "epoch": 3.526058631921824, + "grad_norm": 11.873988390722962, + "learning_rate": 4.222458678132896e-06, + "loss": 0.3681, + "step": 6495 + }, + { + "epoch": 3.526601520086862, + "grad_norm": 14.615618896842006, + "learning_rate": 4.219588552963907e-06, + "loss": 0.5678, + "step": 6496 + }, + { + "epoch": 3.5271444082519, + "grad_norm": 12.902856396087687, + "learning_rate": 4.216719142725415e-06, + "loss": 0.4386, + "step": 6497 + }, + { + "epoch": 3.5276872964169383, + "grad_norm": 14.124175164997677, + "learning_rate": 4.213850447772324e-06, + "loss": 0.752, + "step": 6498 + }, + { + "epoch": 3.528230184581976, + "grad_norm": 17.881101453026922, + "learning_rate": 4.210982468459425e-06, + "loss": 1.0417, + "step": 6499 + }, + { + "epoch": 3.528773072747014, + "grad_norm": 12.44176561816019, + "learning_rate": 4.2081152051414464e-06, + "loss": 0.5422, + "step": 6500 + }, + { + "epoch": 3.529315960912052, + "grad_norm": 11.595014008440414, + "learning_rate": 4.205248658173005e-06, + "loss": 0.5242, + "step": 6501 + }, + { + "epoch": 3.5298588490770904, + "grad_norm": 13.42551753781462, + "learning_rate": 4.20238282790865e-06, + "loss": 0.4613, + "step": 6502 + }, + { + "epoch": 3.530401737242128, + "grad_norm": 14.699795423214242, + "learning_rate": 4.1995177147028195e-06, + "loss": 0.7931, + "step": 6503 + }, + { + "epoch": 3.530944625407166, + "grad_norm": 13.875599451841557, + "learning_rate": 4.196653318909888e-06, + "loss": 0.7584, + "step": 6504 + }, + { + "epoch": 3.531487513572204, + "grad_norm": 10.802073588313325, + "learning_rate": 4.193789640884126e-06, + "loss": 0.4051, + "step": 6505 + }, + { + "epoch": 3.5320304017372424, + "grad_norm": 12.895721698436635, + "learning_rate": 4.190926680979708e-06, + "loss": 0.5394, + "step": 6506 + }, + { + "epoch": 3.53257328990228, + "grad_norm": 9.06469669507275, + "learning_rate": 4.188064439550743e-06, + "loss": 0.3421, + "step": 6507 + }, + { + "epoch": 3.533116178067318, + "grad_norm": 9.872074130422732, + "learning_rate": 4.185202916951224e-06, + "loss": 0.4934, + "step": 6508 + }, + { + "epoch": 3.533659066232356, + "grad_norm": 14.460060519698754, + "learning_rate": 4.1823421135350796e-06, + "loss": 0.9002, + "step": 6509 + }, + { + "epoch": 3.5342019543973944, + "grad_norm": 10.80679699799082, + "learning_rate": 4.179482029656134e-06, + "loss": 0.5466, + "step": 6510 + }, + { + "epoch": 3.534744842562432, + "grad_norm": 17.39434448772562, + "learning_rate": 4.176622665668127e-06, + "loss": 0.7686, + "step": 6511 + }, + { + "epoch": 3.53528773072747, + "grad_norm": 18.02312862709536, + "learning_rate": 4.173764021924711e-06, + "loss": 0.6285, + "step": 6512 + }, + { + "epoch": 3.535830618892508, + "grad_norm": 16.367851061651017, + "learning_rate": 4.170906098779446e-06, + "loss": 0.7154, + "step": 6513 + }, + { + "epoch": 3.5363735070575464, + "grad_norm": 14.114924740024234, + "learning_rate": 4.168048896585809e-06, + "loss": 0.6001, + "step": 6514 + }, + { + "epoch": 3.536916395222584, + "grad_norm": 11.148204893242914, + "learning_rate": 4.165192415697171e-06, + "loss": 0.4029, + "step": 6515 + }, + { + "epoch": 3.537459283387622, + "grad_norm": 12.586445991725615, + "learning_rate": 4.162336656466839e-06, + "loss": 0.53, + "step": 6516 + }, + { + "epoch": 3.53800217155266, + "grad_norm": 10.601436138316005, + "learning_rate": 4.159481619248012e-06, + "loss": 0.5671, + "step": 6517 + }, + { + "epoch": 3.5385450597176984, + "grad_norm": 21.51381975898863, + "learning_rate": 4.156627304393808e-06, + "loss": 0.8123, + "step": 6518 + }, + { + "epoch": 3.539087947882736, + "grad_norm": 11.786864237322536, + "learning_rate": 4.153773712257251e-06, + "loss": 0.5383, + "step": 6519 + }, + { + "epoch": 3.539630836047774, + "grad_norm": 13.251189336378648, + "learning_rate": 4.1509208431912785e-06, + "loss": 0.4292, + "step": 6520 + }, + { + "epoch": 3.540173724212812, + "grad_norm": 11.978453022863066, + "learning_rate": 4.14806869754874e-06, + "loss": 0.5242, + "step": 6521 + }, + { + "epoch": 3.5407166123778504, + "grad_norm": 14.020243792854204, + "learning_rate": 4.145217275682389e-06, + "loss": 0.4147, + "step": 6522 + }, + { + "epoch": 3.541259500542888, + "grad_norm": 12.871747424704319, + "learning_rate": 4.142366577944897e-06, + "loss": 0.479, + "step": 6523 + }, + { + "epoch": 3.541802388707926, + "grad_norm": 16.745498758328235, + "learning_rate": 4.139516604688843e-06, + "loss": 0.732, + "step": 6524 + }, + { + "epoch": 3.542345276872964, + "grad_norm": 18.803948548990295, + "learning_rate": 4.136667356266716e-06, + "loss": 0.6007, + "step": 6525 + }, + { + "epoch": 3.5428881650380024, + "grad_norm": 12.83744283380288, + "learning_rate": 4.133818833030915e-06, + "loss": 0.556, + "step": 6526 + }, + { + "epoch": 3.54343105320304, + "grad_norm": 10.490416435869522, + "learning_rate": 4.13097103533375e-06, + "loss": 0.3933, + "step": 6527 + }, + { + "epoch": 3.543973941368078, + "grad_norm": 13.443902999757336, + "learning_rate": 4.128123963527441e-06, + "loss": 0.7184, + "step": 6528 + }, + { + "epoch": 3.544516829533116, + "grad_norm": 9.68823576515441, + "learning_rate": 4.125277617964119e-06, + "loss": 0.4799, + "step": 6529 + }, + { + "epoch": 3.5450597176981544, + "grad_norm": 8.977409878928121, + "learning_rate": 4.122431998995819e-06, + "loss": 0.2998, + "step": 6530 + }, + { + "epoch": 3.545602605863192, + "grad_norm": 10.738202771196406, + "learning_rate": 4.119587106974505e-06, + "loss": 0.4038, + "step": 6531 + }, + { + "epoch": 3.54614549402823, + "grad_norm": 13.534073718206239, + "learning_rate": 4.116742942252024e-06, + "loss": 0.6041, + "step": 6532 + }, + { + "epoch": 3.546688382193268, + "grad_norm": 9.554994157525224, + "learning_rate": 4.113899505180157e-06, + "loss": 0.3955, + "step": 6533 + }, + { + "epoch": 3.5472312703583064, + "grad_norm": 14.083745919964446, + "learning_rate": 4.1110567961105775e-06, + "loss": 0.5232, + "step": 6534 + }, + { + "epoch": 3.547774158523344, + "grad_norm": 9.270850995609294, + "learning_rate": 4.108214815394876e-06, + "loss": 0.3932, + "step": 6535 + }, + { + "epoch": 3.548317046688382, + "grad_norm": 8.658156641649512, + "learning_rate": 4.105373563384563e-06, + "loss": 0.3587, + "step": 6536 + }, + { + "epoch": 3.54885993485342, + "grad_norm": 22.179187215860352, + "learning_rate": 4.102533040431034e-06, + "loss": 0.6502, + "step": 6537 + }, + { + "epoch": 3.5494028230184584, + "grad_norm": 12.115851203141874, + "learning_rate": 4.0996932468856265e-06, + "loss": 0.6969, + "step": 6538 + }, + { + "epoch": 3.549945711183496, + "grad_norm": 9.686560788090254, + "learning_rate": 4.096854183099554e-06, + "loss": 0.4376, + "step": 6539 + }, + { + "epoch": 3.550488599348534, + "grad_norm": 14.280429673703154, + "learning_rate": 4.0940158494239725e-06, + "loss": 0.7158, + "step": 6540 + }, + { + "epoch": 3.5510314875135722, + "grad_norm": 12.581936500402087, + "learning_rate": 4.091178246209916e-06, + "loss": 0.5323, + "step": 6541 + }, + { + "epoch": 3.5515743756786105, + "grad_norm": 14.994050222936808, + "learning_rate": 4.088341373808354e-06, + "loss": 0.6231, + "step": 6542 + }, + { + "epoch": 3.5521172638436482, + "grad_norm": 11.419896198480089, + "learning_rate": 4.085505232570157e-06, + "loss": 0.4173, + "step": 6543 + }, + { + "epoch": 3.552660152008686, + "grad_norm": 9.002355774397408, + "learning_rate": 4.082669822846092e-06, + "loss": 0.4046, + "step": 6544 + }, + { + "epoch": 3.5532030401737242, + "grad_norm": 8.59747484019196, + "learning_rate": 4.079835144986861e-06, + "loss": 0.3399, + "step": 6545 + }, + { + "epoch": 3.5537459283387625, + "grad_norm": 15.204451286324845, + "learning_rate": 4.0770011993430465e-06, + "loss": 0.6488, + "step": 6546 + }, + { + "epoch": 3.5542888165038002, + "grad_norm": 12.454918944013878, + "learning_rate": 4.074167986265169e-06, + "loss": 0.7895, + "step": 6547 + }, + { + "epoch": 3.554831704668838, + "grad_norm": 12.640544797370467, + "learning_rate": 4.071335506103639e-06, + "loss": 0.3486, + "step": 6548 + }, + { + "epoch": 3.5553745928338762, + "grad_norm": 14.54014915273436, + "learning_rate": 4.068503759208782e-06, + "loss": 0.8116, + "step": 6549 + }, + { + "epoch": 3.5559174809989145, + "grad_norm": 13.706653138925564, + "learning_rate": 4.065672745930833e-06, + "loss": 0.6351, + "step": 6550 + }, + { + "epoch": 3.5564603691639523, + "grad_norm": 7.990687725437862, + "learning_rate": 4.062842466619937e-06, + "loss": 0.356, + "step": 6551 + }, + { + "epoch": 3.55700325732899, + "grad_norm": 12.892934153832375, + "learning_rate": 4.060012921626146e-06, + "loss": 0.5433, + "step": 6552 + }, + { + "epoch": 3.5575461454940283, + "grad_norm": 11.135918453001386, + "learning_rate": 4.057184111299426e-06, + "loss": 0.5487, + "step": 6553 + }, + { + "epoch": 3.5580890336590665, + "grad_norm": 8.023158669990128, + "learning_rate": 4.054356035989645e-06, + "loss": 0.2137, + "step": 6554 + }, + { + "epoch": 3.5586319218241043, + "grad_norm": 15.187504399840787, + "learning_rate": 4.051528696046586e-06, + "loss": 0.513, + "step": 6555 + }, + { + "epoch": 3.559174809989142, + "grad_norm": 11.432478939914837, + "learning_rate": 4.048702091819938e-06, + "loss": 0.4349, + "step": 6556 + }, + { + "epoch": 3.5597176981541803, + "grad_norm": 8.365367809450516, + "learning_rate": 4.045876223659301e-06, + "loss": 0.3435, + "step": 6557 + }, + { + "epoch": 3.5602605863192185, + "grad_norm": 12.186833110184857, + "learning_rate": 4.0430510919141815e-06, + "loss": 0.6071, + "step": 6558 + }, + { + "epoch": 3.5608034744842563, + "grad_norm": 12.640045326528098, + "learning_rate": 4.040226696933998e-06, + "loss": 0.4771, + "step": 6559 + }, + { + "epoch": 3.561346362649294, + "grad_norm": 12.859046455514006, + "learning_rate": 4.037403039068073e-06, + "loss": 0.6651, + "step": 6560 + }, + { + "epoch": 3.5618892508143323, + "grad_norm": 10.219552402377104, + "learning_rate": 4.034580118665644e-06, + "loss": 0.3973, + "step": 6561 + }, + { + "epoch": 3.5624321389793705, + "grad_norm": 11.468743456972462, + "learning_rate": 4.031757936075854e-06, + "loss": 0.5699, + "step": 6562 + }, + { + "epoch": 3.5629750271444083, + "grad_norm": 14.399787814630601, + "learning_rate": 4.028936491647753e-06, + "loss": 0.7208, + "step": 6563 + }, + { + "epoch": 3.563517915309446, + "grad_norm": 12.303594318406011, + "learning_rate": 4.026115785730305e-06, + "loss": 0.4314, + "step": 6564 + }, + { + "epoch": 3.5640608034744843, + "grad_norm": 9.21496866787658, + "learning_rate": 4.023295818672377e-06, + "loss": 0.4055, + "step": 6565 + }, + { + "epoch": 3.5646036916395225, + "grad_norm": 13.534731784395452, + "learning_rate": 4.0204765908227475e-06, + "loss": 0.4986, + "step": 6566 + }, + { + "epoch": 3.5651465798045603, + "grad_norm": 10.67198457926589, + "learning_rate": 4.017658102530103e-06, + "loss": 0.424, + "step": 6567 + }, + { + "epoch": 3.565689467969598, + "grad_norm": 10.45173885950783, + "learning_rate": 4.014840354143035e-06, + "loss": 0.3562, + "step": 6568 + }, + { + "epoch": 3.5662323561346363, + "grad_norm": 17.950537818667204, + "learning_rate": 4.012023346010059e-06, + "loss": 0.6163, + "step": 6569 + }, + { + "epoch": 3.5667752442996745, + "grad_norm": 12.324798504977831, + "learning_rate": 4.009207078479571e-06, + "loss": 0.4225, + "step": 6570 + }, + { + "epoch": 3.5673181324647123, + "grad_norm": 14.602931671691664, + "learning_rate": 4.006391551899906e-06, + "loss": 0.5519, + "step": 6571 + }, + { + "epoch": 3.56786102062975, + "grad_norm": 11.70892852397016, + "learning_rate": 4.00357676661928e-06, + "loss": 0.483, + "step": 6572 + }, + { + "epoch": 3.5684039087947883, + "grad_norm": 9.372765530009781, + "learning_rate": 4.000762722985844e-06, + "loss": 0.2418, + "step": 6573 + }, + { + "epoch": 3.5689467969598265, + "grad_norm": 9.365246601466223, + "learning_rate": 3.997949421347631e-06, + "loss": 0.3748, + "step": 6574 + }, + { + "epoch": 3.5694896851248643, + "grad_norm": 21.027585535755822, + "learning_rate": 3.995136862052597e-06, + "loss": 0.6116, + "step": 6575 + }, + { + "epoch": 3.570032573289902, + "grad_norm": 9.558242987641426, + "learning_rate": 3.992325045448613e-06, + "loss": 0.3608, + "step": 6576 + }, + { + "epoch": 3.5705754614549403, + "grad_norm": 15.196445736424556, + "learning_rate": 3.989513971883434e-06, + "loss": 0.5847, + "step": 6577 + }, + { + "epoch": 3.5711183496199785, + "grad_norm": 13.049400371079434, + "learning_rate": 3.9867036417047546e-06, + "loss": 0.9054, + "step": 6578 + }, + { + "epoch": 3.5716612377850163, + "grad_norm": 10.359620340512786, + "learning_rate": 3.983894055260146e-06, + "loss": 0.3928, + "step": 6579 + }, + { + "epoch": 3.572204125950054, + "grad_norm": 11.96747351920452, + "learning_rate": 3.981085212897111e-06, + "loss": 0.4581, + "step": 6580 + }, + { + "epoch": 3.5727470141150923, + "grad_norm": 13.611366572987158, + "learning_rate": 3.97827711496305e-06, + "loss": 0.8939, + "step": 6581 + }, + { + "epoch": 3.5732899022801305, + "grad_norm": 12.527292279303031, + "learning_rate": 3.975469761805273e-06, + "loss": 0.3713, + "step": 6582 + }, + { + "epoch": 3.5738327904451683, + "grad_norm": 11.7570760452819, + "learning_rate": 3.9726631537710005e-06, + "loss": 0.442, + "step": 6583 + }, + { + "epoch": 3.574375678610206, + "grad_norm": 10.8606079711249, + "learning_rate": 3.969857291207349e-06, + "loss": 0.3717, + "step": 6584 + }, + { + "epoch": 3.5749185667752443, + "grad_norm": 10.89094058468825, + "learning_rate": 3.9670521744613645e-06, + "loss": 0.5867, + "step": 6585 + }, + { + "epoch": 3.5754614549402826, + "grad_norm": 11.08527189234627, + "learning_rate": 3.964247803879976e-06, + "loss": 0.4183, + "step": 6586 + }, + { + "epoch": 3.5760043431053203, + "grad_norm": 14.584182784882016, + "learning_rate": 3.9614441798100415e-06, + "loss": 0.6388, + "step": 6587 + }, + { + "epoch": 3.576547231270358, + "grad_norm": 12.70805104327354, + "learning_rate": 3.958641302598315e-06, + "loss": 0.4211, + "step": 6588 + }, + { + "epoch": 3.5770901194353963, + "grad_norm": 13.935385901164445, + "learning_rate": 3.95583917259146e-06, + "loss": 0.4955, + "step": 6589 + }, + { + "epoch": 3.5776330076004346, + "grad_norm": 8.25831878298027, + "learning_rate": 3.953037790136051e-06, + "loss": 0.3929, + "step": 6590 + }, + { + "epoch": 3.5781758957654723, + "grad_norm": 13.224695924777794, + "learning_rate": 3.950237155578563e-06, + "loss": 0.9157, + "step": 6591 + }, + { + "epoch": 3.57871878393051, + "grad_norm": 14.476470240562103, + "learning_rate": 3.947437269265387e-06, + "loss": 0.6185, + "step": 6592 + }, + { + "epoch": 3.5792616720955484, + "grad_norm": 14.623277057541618, + "learning_rate": 3.944638131542816e-06, + "loss": 0.415, + "step": 6593 + }, + { + "epoch": 3.5798045602605866, + "grad_norm": 10.928897970897149, + "learning_rate": 3.941839742757052e-06, + "loss": 0.4612, + "step": 6594 + }, + { + "epoch": 3.5803474484256244, + "grad_norm": 11.795057907246525, + "learning_rate": 3.939042103254204e-06, + "loss": 0.449, + "step": 6595 + }, + { + "epoch": 3.580890336590662, + "grad_norm": 13.011176416375783, + "learning_rate": 3.9362452133802866e-06, + "loss": 0.4674, + "step": 6596 + }, + { + "epoch": 3.5814332247557004, + "grad_norm": 12.733097120202023, + "learning_rate": 3.933449073481227e-06, + "loss": 0.5925, + "step": 6597 + }, + { + "epoch": 3.5819761129207386, + "grad_norm": 11.53786371590528, + "learning_rate": 3.930653683902854e-06, + "loss": 0.6089, + "step": 6598 + }, + { + "epoch": 3.5825190010857764, + "grad_norm": 10.840934046340113, + "learning_rate": 3.927859044990907e-06, + "loss": 0.3937, + "step": 6599 + }, + { + "epoch": 3.583061889250814, + "grad_norm": 13.435349756607788, + "learning_rate": 3.92506515709103e-06, + "loss": 0.5642, + "step": 6600 + }, + { + "epoch": 3.5836047774158524, + "grad_norm": 15.725827415410208, + "learning_rate": 3.922272020548773e-06, + "loss": 0.7689, + "step": 6601 + }, + { + "epoch": 3.5841476655808906, + "grad_norm": 10.5862295430844, + "learning_rate": 3.919479635709606e-06, + "loss": 0.4151, + "step": 6602 + }, + { + "epoch": 3.5846905537459284, + "grad_norm": 13.70191819426941, + "learning_rate": 3.9166880029188855e-06, + "loss": 0.5176, + "step": 6603 + }, + { + "epoch": 3.585233441910966, + "grad_norm": 20.653564386817997, + "learning_rate": 3.9138971225218895e-06, + "loss": 0.5782, + "step": 6604 + }, + { + "epoch": 3.5857763300760044, + "grad_norm": 13.964645747661887, + "learning_rate": 3.911106994863794e-06, + "loss": 0.629, + "step": 6605 + }, + { + "epoch": 3.5863192182410426, + "grad_norm": 15.152047460754902, + "learning_rate": 3.908317620289689e-06, + "loss": 0.7885, + "step": 6606 + }, + { + "epoch": 3.5868621064060804, + "grad_norm": 10.596374095925908, + "learning_rate": 3.905528999144576e-06, + "loss": 0.3883, + "step": 6607 + }, + { + "epoch": 3.587404994571118, + "grad_norm": 11.600211796857451, + "learning_rate": 3.902741131773341e-06, + "loss": 0.4589, + "step": 6608 + }, + { + "epoch": 3.5879478827361564, + "grad_norm": 9.294471096012794, + "learning_rate": 3.899954018520808e-06, + "loss": 0.334, + "step": 6609 + }, + { + "epoch": 3.5884907709011946, + "grad_norm": 12.555701694026313, + "learning_rate": 3.8971676597316775e-06, + "loss": 0.3707, + "step": 6610 + }, + { + "epoch": 3.5890336590662324, + "grad_norm": 10.572122929609495, + "learning_rate": 3.894382055750584e-06, + "loss": 0.5635, + "step": 6611 + }, + { + "epoch": 3.58957654723127, + "grad_norm": 9.702545363892131, + "learning_rate": 3.891597206922046e-06, + "loss": 0.3634, + "step": 6612 + }, + { + "epoch": 3.5901194353963084, + "grad_norm": 10.430633527197228, + "learning_rate": 3.888813113590496e-06, + "loss": 0.3777, + "step": 6613 + }, + { + "epoch": 3.5906623235613466, + "grad_norm": 16.05235504376427, + "learning_rate": 3.886029776100287e-06, + "loss": 0.5759, + "step": 6614 + }, + { + "epoch": 3.5912052117263844, + "grad_norm": 9.348911210053458, + "learning_rate": 3.883247194795653e-06, + "loss": 0.2619, + "step": 6615 + }, + { + "epoch": 3.591748099891422, + "grad_norm": 8.462834598999683, + "learning_rate": 3.8804653700207615e-06, + "loss": 0.3012, + "step": 6616 + }, + { + "epoch": 3.5922909880564604, + "grad_norm": 16.64696159175473, + "learning_rate": 3.87768430211966e-06, + "loss": 0.9137, + "step": 6617 + }, + { + "epoch": 3.5928338762214986, + "grad_norm": 12.223148301717917, + "learning_rate": 3.874903991436324e-06, + "loss": 0.4287, + "step": 6618 + }, + { + "epoch": 3.5933767643865364, + "grad_norm": 8.953587319006854, + "learning_rate": 3.872124438314624e-06, + "loss": 0.3932, + "step": 6619 + }, + { + "epoch": 3.593919652551574, + "grad_norm": 13.721613828536574, + "learning_rate": 3.869345643098341e-06, + "loss": 0.5449, + "step": 6620 + }, + { + "epoch": 3.5944625407166124, + "grad_norm": 10.931345128421508, + "learning_rate": 3.866567606131159e-06, + "loss": 0.3144, + "step": 6621 + }, + { + "epoch": 3.5950054288816506, + "grad_norm": 16.911421303952586, + "learning_rate": 3.86379032775667e-06, + "loss": 0.8115, + "step": 6622 + }, + { + "epoch": 3.5955483170466884, + "grad_norm": 13.179299764136056, + "learning_rate": 3.861013808318378e-06, + "loss": 0.557, + "step": 6623 + }, + { + "epoch": 3.596091205211726, + "grad_norm": 14.513913457624946, + "learning_rate": 3.858238048159674e-06, + "loss": 0.7132, + "step": 6624 + }, + { + "epoch": 3.5966340933767644, + "grad_norm": 18.20977238222908, + "learning_rate": 3.8554630476238805e-06, + "loss": 0.9001, + "step": 6625 + }, + { + "epoch": 3.5971769815418027, + "grad_norm": 13.411920870051427, + "learning_rate": 3.852688807054211e-06, + "loss": 0.6691, + "step": 6626 + }, + { + "epoch": 3.5977198697068404, + "grad_norm": 12.94431912351694, + "learning_rate": 3.849915326793786e-06, + "loss": 0.8245, + "step": 6627 + }, + { + "epoch": 3.598262757871878, + "grad_norm": 11.378136214690231, + "learning_rate": 3.847142607185636e-06, + "loss": 0.5369, + "step": 6628 + }, + { + "epoch": 3.5988056460369164, + "grad_norm": 18.107740640174118, + "learning_rate": 3.844370648572694e-06, + "loss": 0.5532, + "step": 6629 + }, + { + "epoch": 3.5993485342019547, + "grad_norm": 13.376518628320737, + "learning_rate": 3.841599451297802e-06, + "loss": 0.6283, + "step": 6630 + }, + { + "epoch": 3.5998914223669924, + "grad_norm": 17.68256418630599, + "learning_rate": 3.8388290157037034e-06, + "loss": 0.5916, + "step": 6631 + }, + { + "epoch": 3.6004343105320302, + "grad_norm": 9.436411945807468, + "learning_rate": 3.836059342133052e-06, + "loss": 0.488, + "step": 6632 + }, + { + "epoch": 3.6009771986970684, + "grad_norm": 12.744056075258795, + "learning_rate": 3.8332904309284045e-06, + "loss": 0.568, + "step": 6633 + }, + { + "epoch": 3.6015200868621067, + "grad_norm": 15.484356986063371, + "learning_rate": 3.830522282432227e-06, + "loss": 0.5619, + "step": 6634 + }, + { + "epoch": 3.6020629750271445, + "grad_norm": 22.466452052149346, + "learning_rate": 3.827754896986884e-06, + "loss": 0.754, + "step": 6635 + }, + { + "epoch": 3.6026058631921822, + "grad_norm": 10.907712543607193, + "learning_rate": 3.824988274934654e-06, + "loss": 0.3684, + "step": 6636 + }, + { + "epoch": 3.6031487513572205, + "grad_norm": 11.436863024139019, + "learning_rate": 3.822222416617714e-06, + "loss": 0.3772, + "step": 6637 + }, + { + "epoch": 3.6036916395222587, + "grad_norm": 8.249398777832742, + "learning_rate": 3.8194573223781515e-06, + "loss": 0.3908, + "step": 6638 + }, + { + "epoch": 3.6042345276872965, + "grad_norm": 10.129204140499496, + "learning_rate": 3.816692992557955e-06, + "loss": 0.5793, + "step": 6639 + }, + { + "epoch": 3.6047774158523342, + "grad_norm": 16.22320006976241, + "learning_rate": 3.8139294274990313e-06, + "loss": 0.7805, + "step": 6640 + }, + { + "epoch": 3.6053203040173725, + "grad_norm": 16.13099309974979, + "learning_rate": 3.8111666275431715e-06, + "loss": 0.5848, + "step": 6641 + }, + { + "epoch": 3.6058631921824107, + "grad_norm": 16.513602515759278, + "learning_rate": 3.8084045930320855e-06, + "loss": 0.5467, + "step": 6642 + }, + { + "epoch": 3.6064060803474485, + "grad_norm": 17.235827177127607, + "learning_rate": 3.8056433243073888e-06, + "loss": 0.7889, + "step": 6643 + }, + { + "epoch": 3.6069489685124863, + "grad_norm": 8.437058780004433, + "learning_rate": 3.8028828217105974e-06, + "loss": 0.4132, + "step": 6644 + }, + { + "epoch": 3.6074918566775245, + "grad_norm": 10.802576839366907, + "learning_rate": 3.800123085583135e-06, + "loss": 0.4123, + "step": 6645 + }, + { + "epoch": 3.6080347448425627, + "grad_norm": 11.08981584907767, + "learning_rate": 3.7973641162663276e-06, + "loss": 0.3947, + "step": 6646 + }, + { + "epoch": 3.6085776330076005, + "grad_norm": 9.95657545789331, + "learning_rate": 3.794605914101419e-06, + "loss": 0.6197, + "step": 6647 + }, + { + "epoch": 3.6091205211726383, + "grad_norm": 8.688663473460801, + "learning_rate": 3.791848479429534e-06, + "loss": 0.2914, + "step": 6648 + }, + { + "epoch": 3.6096634093376765, + "grad_norm": 13.569135821875664, + "learning_rate": 3.78909181259173e-06, + "loss": 0.4108, + "step": 6649 + }, + { + "epoch": 3.6102062975027147, + "grad_norm": 12.845103922950058, + "learning_rate": 3.7863359139289425e-06, + "loss": 0.5261, + "step": 6650 + }, + { + "epoch": 3.6107491856677525, + "grad_norm": 19.751616468555607, + "learning_rate": 3.783580783782035e-06, + "loss": 0.7023, + "step": 6651 + }, + { + "epoch": 3.6112920738327903, + "grad_norm": 13.094345121958364, + "learning_rate": 3.780826422491768e-06, + "loss": 0.5129, + "step": 6652 + }, + { + "epoch": 3.6118349619978285, + "grad_norm": 9.608815012306174, + "learning_rate": 3.7780728303987935e-06, + "loss": 0.3692, + "step": 6653 + }, + { + "epoch": 3.6123778501628667, + "grad_norm": 12.846531429400725, + "learning_rate": 3.7753200078436947e-06, + "loss": 0.5278, + "step": 6654 + }, + { + "epoch": 3.6129207383279045, + "grad_norm": 10.62078519937276, + "learning_rate": 3.7725679551669316e-06, + "loss": 0.5868, + "step": 6655 + }, + { + "epoch": 3.6134636264929423, + "grad_norm": 12.42176232319324, + "learning_rate": 3.769816672708895e-06, + "loss": 0.5746, + "step": 6656 + }, + { + "epoch": 3.6140065146579805, + "grad_norm": 15.37568549128414, + "learning_rate": 3.7670661608098545e-06, + "loss": 0.5671, + "step": 6657 + }, + { + "epoch": 3.6145494028230187, + "grad_norm": 14.651318253713534, + "learning_rate": 3.764316419810009e-06, + "loss": 0.4817, + "step": 6658 + }, + { + "epoch": 3.6150922909880565, + "grad_norm": 17.772625754725265, + "learning_rate": 3.761567450049448e-06, + "loss": 0.6106, + "step": 6659 + }, + { + "epoch": 3.6156351791530943, + "grad_norm": 12.785953562325862, + "learning_rate": 3.7588192518681664e-06, + "loss": 0.458, + "step": 6660 + }, + { + "epoch": 3.6161780673181325, + "grad_norm": 14.872223886784575, + "learning_rate": 3.7560718256060704e-06, + "loss": 0.5833, + "step": 6661 + }, + { + "epoch": 3.6167209554831707, + "grad_norm": 10.482699567034725, + "learning_rate": 3.753325171602955e-06, + "loss": 0.4262, + "step": 6662 + }, + { + "epoch": 3.6172638436482085, + "grad_norm": 12.097306912879557, + "learning_rate": 3.7505792901985417e-06, + "loss": 0.4847, + "step": 6663 + }, + { + "epoch": 3.6178067318132463, + "grad_norm": 11.56080631055532, + "learning_rate": 3.7478341817324414e-06, + "loss": 0.4421, + "step": 6664 + }, + { + "epoch": 3.6183496199782845, + "grad_norm": 17.08246749281531, + "learning_rate": 3.7450898465441744e-06, + "loss": 0.6724, + "step": 6665 + }, + { + "epoch": 3.6188925081433228, + "grad_norm": 14.083035375418559, + "learning_rate": 3.742346284973164e-06, + "loss": 0.7099, + "step": 6666 + }, + { + "epoch": 3.6194353963083605, + "grad_norm": 12.257812595787682, + "learning_rate": 3.7396034973587392e-06, + "loss": 0.5108, + "step": 6667 + }, + { + "epoch": 3.6199782844733983, + "grad_norm": 16.583982679880926, + "learning_rate": 3.7368614840401297e-06, + "loss": 0.5608, + "step": 6668 + }, + { + "epoch": 3.6205211726384365, + "grad_norm": 9.31804613123421, + "learning_rate": 3.734120245356475e-06, + "loss": 0.2673, + "step": 6669 + }, + { + "epoch": 3.6210640608034748, + "grad_norm": 8.73729193728082, + "learning_rate": 3.7313797816468145e-06, + "loss": 0.2867, + "step": 6670 + }, + { + "epoch": 3.6216069489685125, + "grad_norm": 10.574625431794137, + "learning_rate": 3.728640093250093e-06, + "loss": 0.6508, + "step": 6671 + }, + { + "epoch": 3.6221498371335503, + "grad_norm": 13.116215370859619, + "learning_rate": 3.7259011805051605e-06, + "loss": 0.8055, + "step": 6672 + }, + { + "epoch": 3.6226927252985885, + "grad_norm": 12.674214863430533, + "learning_rate": 3.723163043750768e-06, + "loss": 0.4149, + "step": 6673 + }, + { + "epoch": 3.6232356134636268, + "grad_norm": 15.457787982715566, + "learning_rate": 3.720425683325576e-06, + "loss": 0.5407, + "step": 6674 + }, + { + "epoch": 3.6237785016286646, + "grad_norm": 13.509659223593033, + "learning_rate": 3.7176890995681426e-06, + "loss": 0.3499, + "step": 6675 + }, + { + "epoch": 3.6243213897937023, + "grad_norm": 16.313287760559906, + "learning_rate": 3.714953292816934e-06, + "loss": 0.697, + "step": 6676 + }, + { + "epoch": 3.6248642779587406, + "grad_norm": 14.440828134958558, + "learning_rate": 3.712218263410319e-06, + "loss": 0.5698, + "step": 6677 + }, + { + "epoch": 3.6254071661237783, + "grad_norm": 9.354441770636624, + "learning_rate": 3.7094840116865693e-06, + "loss": 0.3452, + "step": 6678 + }, + { + "epoch": 3.6259500542888166, + "grad_norm": 9.584690153683102, + "learning_rate": 3.706750537983861e-06, + "loss": 0.4305, + "step": 6679 + }, + { + "epoch": 3.6264929424538543, + "grad_norm": 18.051709305408423, + "learning_rate": 3.7040178426402817e-06, + "loss": 0.7387, + "step": 6680 + }, + { + "epoch": 3.6270358306188926, + "grad_norm": 10.05950366415881, + "learning_rate": 3.7012859259938074e-06, + "loss": 0.3443, + "step": 6681 + }, + { + "epoch": 3.6275787187839303, + "grad_norm": 12.034931076516903, + "learning_rate": 3.698554788382328e-06, + "loss": 0.542, + "step": 6682 + }, + { + "epoch": 3.6281216069489686, + "grad_norm": 12.814483154092642, + "learning_rate": 3.6958244301436353e-06, + "loss": 0.6022, + "step": 6683 + }, + { + "epoch": 3.6286644951140063, + "grad_norm": 12.602310642914164, + "learning_rate": 3.693094851615421e-06, + "loss": 0.5764, + "step": 6684 + }, + { + "epoch": 3.6292073832790446, + "grad_norm": 13.840451235628782, + "learning_rate": 3.6903660531352947e-06, + "loss": 0.673, + "step": 6685 + }, + { + "epoch": 3.6297502714440824, + "grad_norm": 11.319388737736466, + "learning_rate": 3.6876380350407428e-06, + "loss": 0.5201, + "step": 6686 + }, + { + "epoch": 3.6302931596091206, + "grad_norm": 8.195460173712483, + "learning_rate": 3.6849107976691867e-06, + "loss": 0.2834, + "step": 6687 + }, + { + "epoch": 3.6308360477741584, + "grad_norm": 11.24454405051031, + "learning_rate": 3.6821843413579217e-06, + "loss": 0.4174, + "step": 6688 + }, + { + "epoch": 3.6313789359391966, + "grad_norm": 13.272520776849182, + "learning_rate": 3.679458666444169e-06, + "loss": 0.5333, + "step": 6689 + }, + { + "epoch": 3.6319218241042344, + "grad_norm": 11.044629755858516, + "learning_rate": 3.6767337732650454e-06, + "loss": 0.4734, + "step": 6690 + }, + { + "epoch": 3.6324647122692726, + "grad_norm": 9.202516399757954, + "learning_rate": 3.6740096621575604e-06, + "loss": 0.28, + "step": 6691 + }, + { + "epoch": 3.6330076004343104, + "grad_norm": 11.820877288548346, + "learning_rate": 3.6712863334586478e-06, + "loss": 0.2672, + "step": 6692 + }, + { + "epoch": 3.6335504885993486, + "grad_norm": 11.093852797013197, + "learning_rate": 3.668563787505123e-06, + "loss": 0.3923, + "step": 6693 + }, + { + "epoch": 3.6340933767643864, + "grad_norm": 10.10902067133091, + "learning_rate": 3.6658420246337256e-06, + "loss": 0.3633, + "step": 6694 + }, + { + "epoch": 3.6346362649294246, + "grad_norm": 11.211935387408433, + "learning_rate": 3.663121045181075e-06, + "loss": 0.4325, + "step": 6695 + }, + { + "epoch": 3.6351791530944624, + "grad_norm": 7.899984820961417, + "learning_rate": 3.660400849483716e-06, + "loss": 0.2753, + "step": 6696 + }, + { + "epoch": 3.6357220412595006, + "grad_norm": 15.593004160513892, + "learning_rate": 3.6576814378780844e-06, + "loss": 0.5612, + "step": 6697 + }, + { + "epoch": 3.6362649294245384, + "grad_norm": 14.607948256302272, + "learning_rate": 3.654962810700521e-06, + "loss": 0.5311, + "step": 6698 + }, + { + "epoch": 3.6368078175895766, + "grad_norm": 14.159583660700658, + "learning_rate": 3.6522449682872687e-06, + "loss": 0.4704, + "step": 6699 + }, + { + "epoch": 3.6373507057546144, + "grad_norm": 12.5139018167431, + "learning_rate": 3.6495279109744764e-06, + "loss": 0.5437, + "step": 6700 + }, + { + "epoch": 3.6378935939196526, + "grad_norm": 13.188903276553058, + "learning_rate": 3.6468116390981922e-06, + "loss": 0.5713, + "step": 6701 + }, + { + "epoch": 3.6384364820846904, + "grad_norm": 9.269396910628474, + "learning_rate": 3.644096152994371e-06, + "loss": 0.3362, + "step": 6702 + }, + { + "epoch": 3.6389793702497286, + "grad_norm": 15.76289925225522, + "learning_rate": 3.6413814529988666e-06, + "loss": 0.6549, + "step": 6703 + }, + { + "epoch": 3.6395222584147664, + "grad_norm": 14.218821330155238, + "learning_rate": 3.6386675394474393e-06, + "loss": 0.5017, + "step": 6704 + }, + { + "epoch": 3.6400651465798046, + "grad_norm": 13.035114169534788, + "learning_rate": 3.6359544126757483e-06, + "loss": 0.4331, + "step": 6705 + }, + { + "epoch": 3.6406080347448424, + "grad_norm": 20.82770787825795, + "learning_rate": 3.6332420730193584e-06, + "loss": 0.7242, + "step": 6706 + }, + { + "epoch": 3.6411509229098806, + "grad_norm": 14.084479195896975, + "learning_rate": 3.630530520813735e-06, + "loss": 0.5915, + "step": 6707 + }, + { + "epoch": 3.6416938110749184, + "grad_norm": 12.24391802241245, + "learning_rate": 3.6278197563942497e-06, + "loss": 0.3637, + "step": 6708 + }, + { + "epoch": 3.6422366992399566, + "grad_norm": 19.068246989978615, + "learning_rate": 3.625109780096173e-06, + "loss": 0.5401, + "step": 6709 + }, + { + "epoch": 3.6427795874049944, + "grad_norm": 12.923571390979863, + "learning_rate": 3.6224005922546777e-06, + "loss": 0.4372, + "step": 6710 + }, + { + "epoch": 3.6433224755700326, + "grad_norm": 18.72956431478281, + "learning_rate": 3.6196921932048424e-06, + "loss": 0.5212, + "step": 6711 + }, + { + "epoch": 3.6438653637350704, + "grad_norm": 9.246217525160361, + "learning_rate": 3.6169845832816443e-06, + "loss": 0.2875, + "step": 6712 + }, + { + "epoch": 3.6444082519001086, + "grad_norm": 12.538873429253352, + "learning_rate": 3.6142777628199667e-06, + "loss": 0.7951, + "step": 6713 + }, + { + "epoch": 3.6449511400651464, + "grad_norm": 9.970157389027023, + "learning_rate": 3.611571732154592e-06, + "loss": 0.3658, + "step": 6714 + }, + { + "epoch": 3.6454940282301846, + "grad_norm": 12.203072582682385, + "learning_rate": 3.6088664916202077e-06, + "loss": 0.5721, + "step": 6715 + }, + { + "epoch": 3.6460369163952224, + "grad_norm": 12.279248739912763, + "learning_rate": 3.606162041551402e-06, + "loss": 0.4408, + "step": 6716 + }, + { + "epoch": 3.6465798045602607, + "grad_norm": 12.802459902824975, + "learning_rate": 3.6034583822826607e-06, + "loss": 0.6177, + "step": 6717 + }, + { + "epoch": 3.6471226927252984, + "grad_norm": 14.71635978144862, + "learning_rate": 3.6007555141483897e-06, + "loss": 0.6333, + "step": 6718 + }, + { + "epoch": 3.6476655808903367, + "grad_norm": 15.254287066882977, + "learning_rate": 3.5980534374828667e-06, + "loss": 0.6541, + "step": 6719 + }, + { + "epoch": 3.6482084690553744, + "grad_norm": 17.306678251097242, + "learning_rate": 3.595352152620306e-06, + "loss": 0.5605, + "step": 6720 + }, + { + "epoch": 3.6487513572204127, + "grad_norm": 13.28374885202886, + "learning_rate": 3.5926516598947946e-06, + "loss": 0.553, + "step": 6721 + }, + { + "epoch": 3.6492942453854504, + "grad_norm": 15.674888174232077, + "learning_rate": 3.5899519596403356e-06, + "loss": 0.5473, + "step": 6722 + }, + { + "epoch": 3.6498371335504887, + "grad_norm": 9.387854198306167, + "learning_rate": 3.58725305219084e-06, + "loss": 0.4405, + "step": 6723 + }, + { + "epoch": 3.6503800217155264, + "grad_norm": 9.686200542254724, + "learning_rate": 3.5845549378801005e-06, + "loss": 0.4019, + "step": 6724 + }, + { + "epoch": 3.6509229098805647, + "grad_norm": 15.112849538872657, + "learning_rate": 3.5818576170418373e-06, + "loss": 0.7064, + "step": 6725 + }, + { + "epoch": 3.6514657980456025, + "grad_norm": 11.208536381794502, + "learning_rate": 3.5791610900096465e-06, + "loss": 0.3352, + "step": 6726 + }, + { + "epoch": 3.6520086862106407, + "grad_norm": 7.334988680617685, + "learning_rate": 3.5764653571170527e-06, + "loss": 0.288, + "step": 6727 + }, + { + "epoch": 3.6525515743756785, + "grad_norm": 11.838685637818697, + "learning_rate": 3.573770418697454e-06, + "loss": 0.6447, + "step": 6728 + }, + { + "epoch": 3.6530944625407167, + "grad_norm": 15.976115688824532, + "learning_rate": 3.5710762750841756e-06, + "loss": 0.4407, + "step": 6729 + }, + { + "epoch": 3.6536373507057545, + "grad_norm": 12.25069694802214, + "learning_rate": 3.5683829266104318e-06, + "loss": 0.4395, + "step": 6730 + }, + { + "epoch": 3.6541802388707927, + "grad_norm": 9.192123102498591, + "learning_rate": 3.5656903736093328e-06, + "loss": 0.3363, + "step": 6731 + }, + { + "epoch": 3.6547231270358305, + "grad_norm": 14.524314713398583, + "learning_rate": 3.5629986164139095e-06, + "loss": 0.3778, + "step": 6732 + }, + { + "epoch": 3.6552660152008687, + "grad_norm": 16.927339643078007, + "learning_rate": 3.5603076553570692e-06, + "loss": 0.6248, + "step": 6733 + }, + { + "epoch": 3.6558089033659065, + "grad_norm": 12.599717116787803, + "learning_rate": 3.5576174907716455e-06, + "loss": 0.6944, + "step": 6734 + }, + { + "epoch": 3.6563517915309447, + "grad_norm": 10.646444115088963, + "learning_rate": 3.5549281229903575e-06, + "loss": 0.3772, + "step": 6735 + }, + { + "epoch": 3.6568946796959825, + "grad_norm": 15.498764867333419, + "learning_rate": 3.552239552345831e-06, + "loss": 0.5415, + "step": 6736 + }, + { + "epoch": 3.6574375678610207, + "grad_norm": 15.747483640049493, + "learning_rate": 3.5495517791705937e-06, + "loss": 0.4526, + "step": 6737 + }, + { + "epoch": 3.6579804560260585, + "grad_norm": 8.811218308037992, + "learning_rate": 3.546864803797072e-06, + "loss": 0.2853, + "step": 6738 + }, + { + "epoch": 3.6585233441910967, + "grad_norm": 13.307618617651949, + "learning_rate": 3.5441786265576006e-06, + "loss": 0.5929, + "step": 6739 + }, + { + "epoch": 3.6590662323561345, + "grad_norm": 16.35655181341609, + "learning_rate": 3.541493247784398e-06, + "loss": 0.5969, + "step": 6740 + }, + { + "epoch": 3.6596091205211727, + "grad_norm": 13.468394744914283, + "learning_rate": 3.5388086678096077e-06, + "loss": 0.7131, + "step": 6741 + }, + { + "epoch": 3.6601520086862105, + "grad_norm": 10.610479158018709, + "learning_rate": 3.5361248869652586e-06, + "loss": 0.3491, + "step": 6742 + }, + { + "epoch": 3.6606948968512487, + "grad_norm": 10.63033092512971, + "learning_rate": 3.5334419055832857e-06, + "loss": 0.4148, + "step": 6743 + }, + { + "epoch": 3.6612377850162865, + "grad_norm": 12.066673558862, + "learning_rate": 3.530759723995524e-06, + "loss": 0.5922, + "step": 6744 + }, + { + "epoch": 3.6617806731813247, + "grad_norm": 13.669464818674667, + "learning_rate": 3.5280783425337097e-06, + "loss": 0.4984, + "step": 6745 + }, + { + "epoch": 3.6623235613463625, + "grad_norm": 11.385901052590954, + "learning_rate": 3.525397761529481e-06, + "loss": 0.4451, + "step": 6746 + }, + { + "epoch": 3.6628664495114007, + "grad_norm": 13.127975944644733, + "learning_rate": 3.5227179813143763e-06, + "loss": 0.4786, + "step": 6747 + }, + { + "epoch": 3.6634093376764385, + "grad_norm": 8.286555642541957, + "learning_rate": 3.520039002219835e-06, + "loss": 0.3308, + "step": 6748 + }, + { + "epoch": 3.6639522258414767, + "grad_norm": 13.989369716944896, + "learning_rate": 3.5173608245771964e-06, + "loss": 0.6301, + "step": 6749 + }, + { + "epoch": 3.6644951140065145, + "grad_norm": 11.280241380801893, + "learning_rate": 3.514683448717704e-06, + "loss": 0.4333, + "step": 6750 + }, + { + "epoch": 3.6650380021715527, + "grad_norm": 14.109069012865124, + "learning_rate": 3.5120068749724977e-06, + "loss": 0.4928, + "step": 6751 + }, + { + "epoch": 3.6655808903365905, + "grad_norm": 9.420479780236201, + "learning_rate": 3.509331103672623e-06, + "loss": 0.3409, + "step": 6752 + }, + { + "epoch": 3.6661237785016287, + "grad_norm": 13.634334295467264, + "learning_rate": 3.5066561351490202e-06, + "loss": 0.4104, + "step": 6753 + }, + { + "epoch": 3.6666666666666665, + "grad_norm": 12.151334809320293, + "learning_rate": 3.503981969732536e-06, + "loss": 0.6371, + "step": 6754 + }, + { + "epoch": 3.6672095548317047, + "grad_norm": 15.410529391564157, + "learning_rate": 3.5013086077539125e-06, + "loss": 0.6359, + "step": 6755 + }, + { + "epoch": 3.6677524429967425, + "grad_norm": 14.234740703322796, + "learning_rate": 3.4986360495438033e-06, + "loss": 0.5397, + "step": 6756 + }, + { + "epoch": 3.6682953311617807, + "grad_norm": 12.465036402535198, + "learning_rate": 3.4959642954327435e-06, + "loss": 0.5274, + "step": 6757 + }, + { + "epoch": 3.6688382193268185, + "grad_norm": 12.485874013130712, + "learning_rate": 3.4932933457511907e-06, + "loss": 0.4669, + "step": 6758 + }, + { + "epoch": 3.6693811074918568, + "grad_norm": 13.300754823528772, + "learning_rate": 3.490623200829485e-06, + "loss": 0.5991, + "step": 6759 + }, + { + "epoch": 3.6699239956568945, + "grad_norm": 13.6162114238038, + "learning_rate": 3.4879538609978724e-06, + "loss": 0.5753, + "step": 6760 + }, + { + "epoch": 3.6704668838219328, + "grad_norm": 15.055930735004507, + "learning_rate": 3.4852853265865117e-06, + "loss": 0.5937, + "step": 6761 + }, + { + "epoch": 3.6710097719869705, + "grad_norm": 19.021493329558147, + "learning_rate": 3.4826175979254382e-06, + "loss": 0.7309, + "step": 6762 + }, + { + "epoch": 3.6715526601520088, + "grad_norm": 7.879252217720616, + "learning_rate": 3.4799506753446133e-06, + "loss": 0.2324, + "step": 6763 + }, + { + "epoch": 3.6720955483170465, + "grad_norm": 15.566429874979375, + "learning_rate": 3.477284559173875e-06, + "loss": 0.8685, + "step": 6764 + }, + { + "epoch": 3.6726384364820848, + "grad_norm": 14.932288315207366, + "learning_rate": 3.474619249742983e-06, + "loss": 0.4121, + "step": 6765 + }, + { + "epoch": 3.6731813246471225, + "grad_norm": 12.826517078026129, + "learning_rate": 3.471954747381577e-06, + "loss": 0.4172, + "step": 6766 + }, + { + "epoch": 3.6737242128121608, + "grad_norm": 11.27094341605798, + "learning_rate": 3.4692910524192145e-06, + "loss": 0.6047, + "step": 6767 + }, + { + "epoch": 3.6742671009771986, + "grad_norm": 10.227384928715546, + "learning_rate": 3.4666281651853426e-06, + "loss": 0.4978, + "step": 6768 + }, + { + "epoch": 3.6748099891422368, + "grad_norm": 9.140068130601126, + "learning_rate": 3.463966086009313e-06, + "loss": 0.3314, + "step": 6769 + }, + { + "epoch": 3.6753528773072746, + "grad_norm": 8.57213348310477, + "learning_rate": 3.4613048152203776e-06, + "loss": 0.3135, + "step": 6770 + }, + { + "epoch": 3.675895765472313, + "grad_norm": 18.06864303821088, + "learning_rate": 3.4586443531476777e-06, + "loss": 0.7083, + "step": 6771 + }, + { + "epoch": 3.6764386536373506, + "grad_norm": 12.06612974538307, + "learning_rate": 3.455984700120272e-06, + "loss": 0.4263, + "step": 6772 + }, + { + "epoch": 3.676981541802389, + "grad_norm": 15.51068609579363, + "learning_rate": 3.453325856467108e-06, + "loss": 0.5577, + "step": 6773 + }, + { + "epoch": 3.6775244299674266, + "grad_norm": 10.276565099526616, + "learning_rate": 3.4506678225170363e-06, + "loss": 0.3197, + "step": 6774 + }, + { + "epoch": 3.678067318132465, + "grad_norm": 9.001248775476501, + "learning_rate": 3.4480105985988065e-06, + "loss": 0.376, + "step": 6775 + }, + { + "epoch": 3.6786102062975026, + "grad_norm": 10.679849220418427, + "learning_rate": 3.4453541850410686e-06, + "loss": 0.3383, + "step": 6776 + }, + { + "epoch": 3.679153094462541, + "grad_norm": 10.331306790662499, + "learning_rate": 3.4426985821723714e-06, + "loss": 0.5359, + "step": 6777 + }, + { + "epoch": 3.6796959826275786, + "grad_norm": 12.802407333669652, + "learning_rate": 3.4400437903211635e-06, + "loss": 0.4397, + "step": 6778 + }, + { + "epoch": 3.680238870792617, + "grad_norm": 14.73981252395514, + "learning_rate": 3.4373898098157953e-06, + "loss": 0.5717, + "step": 6779 + }, + { + "epoch": 3.6807817589576546, + "grad_norm": 16.362575863101522, + "learning_rate": 3.434736640984515e-06, + "loss": 0.5125, + "step": 6780 + }, + { + "epoch": 3.681324647122693, + "grad_norm": 11.576699027697012, + "learning_rate": 3.4320842841554702e-06, + "loss": 0.3924, + "step": 6781 + }, + { + "epoch": 3.6818675352877306, + "grad_norm": 12.430829059754513, + "learning_rate": 3.4294327396567085e-06, + "loss": 0.5816, + "step": 6782 + }, + { + "epoch": 3.682410423452769, + "grad_norm": 12.320695413208654, + "learning_rate": 3.4267820078161785e-06, + "loss": 0.5416, + "step": 6783 + }, + { + "epoch": 3.6829533116178066, + "grad_norm": 13.000253896056515, + "learning_rate": 3.4241320889617245e-06, + "loss": 0.536, + "step": 6784 + }, + { + "epoch": 3.683496199782845, + "grad_norm": 10.91584072091396, + "learning_rate": 3.4214829834210948e-06, + "loss": 0.4491, + "step": 6785 + }, + { + "epoch": 3.6840390879478826, + "grad_norm": 13.085054837243629, + "learning_rate": 3.418834691521934e-06, + "loss": 0.378, + "step": 6786 + }, + { + "epoch": 3.684581976112921, + "grad_norm": 11.116151442052336, + "learning_rate": 3.4161872135917873e-06, + "loss": 0.4053, + "step": 6787 + }, + { + "epoch": 3.6851248642779586, + "grad_norm": 14.044458874643864, + "learning_rate": 3.4135405499580998e-06, + "loss": 0.3851, + "step": 6788 + }, + { + "epoch": 3.685667752442997, + "grad_norm": 15.351406534798858, + "learning_rate": 3.410894700948214e-06, + "loss": 0.7129, + "step": 6789 + }, + { + "epoch": 3.6862106406080346, + "grad_norm": 18.553137629768052, + "learning_rate": 3.408249666889373e-06, + "loss": 0.9605, + "step": 6790 + }, + { + "epoch": 3.686753528773073, + "grad_norm": 12.059229166245107, + "learning_rate": 3.4056054481087187e-06, + "loss": 0.6246, + "step": 6791 + }, + { + "epoch": 3.6872964169381106, + "grad_norm": 8.164629976709701, + "learning_rate": 3.4029620449332925e-06, + "loss": 0.2484, + "step": 6792 + }, + { + "epoch": 3.687839305103149, + "grad_norm": 12.956574155251525, + "learning_rate": 3.400319457690031e-06, + "loss": 0.4846, + "step": 6793 + }, + { + "epoch": 3.6883821932681866, + "grad_norm": 13.218562644434163, + "learning_rate": 3.397677686705785e-06, + "loss": 0.4956, + "step": 6794 + }, + { + "epoch": 3.688925081433225, + "grad_norm": 12.52521210536022, + "learning_rate": 3.3950367323072777e-06, + "loss": 0.5609, + "step": 6795 + }, + { + "epoch": 3.6894679695982626, + "grad_norm": 14.858442377207329, + "learning_rate": 3.3923965948211612e-06, + "loss": 0.5361, + "step": 6796 + }, + { + "epoch": 3.690010857763301, + "grad_norm": 10.357147809279667, + "learning_rate": 3.3897572745739584e-06, + "loss": 0.4834, + "step": 6797 + }, + { + "epoch": 3.6905537459283386, + "grad_norm": 10.579748653808505, + "learning_rate": 3.3871187718921173e-06, + "loss": 0.4685, + "step": 6798 + }, + { + "epoch": 3.691096634093377, + "grad_norm": 16.620322523965804, + "learning_rate": 3.3844810871019617e-06, + "loss": 0.6373, + "step": 6799 + }, + { + "epoch": 3.6916395222584146, + "grad_norm": 15.999425270589017, + "learning_rate": 3.3818442205297262e-06, + "loss": 0.895, + "step": 6800 + }, + { + "epoch": 3.692182410423453, + "grad_norm": 10.78206305263471, + "learning_rate": 3.3792081725015525e-06, + "loss": 0.5564, + "step": 6801 + }, + { + "epoch": 3.6927252985884906, + "grad_norm": 11.641189355388065, + "learning_rate": 3.376572943343457e-06, + "loss": 0.5319, + "step": 6802 + }, + { + "epoch": 3.693268186753529, + "grad_norm": 14.695430550731649, + "learning_rate": 3.3739385333813824e-06, + "loss": 0.5868, + "step": 6803 + }, + { + "epoch": 3.6938110749185666, + "grad_norm": 10.078064673313023, + "learning_rate": 3.3713049429411435e-06, + "loss": 0.5002, + "step": 6804 + }, + { + "epoch": 3.694353963083605, + "grad_norm": 13.958804490496405, + "learning_rate": 3.368672172348478e-06, + "loss": 0.5521, + "step": 6805 + }, + { + "epoch": 3.6948968512486426, + "grad_norm": 14.349798638841735, + "learning_rate": 3.366040221929007e-06, + "loss": 0.8658, + "step": 6806 + }, + { + "epoch": 3.695439739413681, + "grad_norm": 19.918583954387678, + "learning_rate": 3.363409092008253e-06, + "loss": 0.7743, + "step": 6807 + }, + { + "epoch": 3.6959826275787186, + "grad_norm": 17.742647882095138, + "learning_rate": 3.3607787829116446e-06, + "loss": 0.4515, + "step": 6808 + }, + { + "epoch": 3.696525515743757, + "grad_norm": 10.287874260684818, + "learning_rate": 3.35814929496449e-06, + "loss": 0.4691, + "step": 6809 + }, + { + "epoch": 3.6970684039087947, + "grad_norm": 13.988040829949796, + "learning_rate": 3.3555206284920226e-06, + "loss": 0.5869, + "step": 6810 + }, + { + "epoch": 3.697611292073833, + "grad_norm": 12.599509307267757, + "learning_rate": 3.352892783819347e-06, + "loss": 0.5899, + "step": 6811 + }, + { + "epoch": 3.6981541802388707, + "grad_norm": 9.433309182729886, + "learning_rate": 3.3502657612714894e-06, + "loss": 0.3442, + "step": 6812 + }, + { + "epoch": 3.698697068403909, + "grad_norm": 16.010015776538683, + "learning_rate": 3.3476395611733616e-06, + "loss": 0.5644, + "step": 6813 + }, + { + "epoch": 3.6992399565689467, + "grad_norm": 12.170086596108076, + "learning_rate": 3.345014183849774e-06, + "loss": 0.5416, + "step": 6814 + }, + { + "epoch": 3.699782844733985, + "grad_norm": 9.466907002380811, + "learning_rate": 3.342389629625439e-06, + "loss": 0.5223, + "step": 6815 + }, + { + "epoch": 3.7003257328990227, + "grad_norm": 10.374676588070637, + "learning_rate": 3.3397658988249647e-06, + "loss": 0.327, + "step": 6816 + }, + { + "epoch": 3.700868621064061, + "grad_norm": 16.837753149163387, + "learning_rate": 3.337142991772859e-06, + "loss": 0.3853, + "step": 6817 + }, + { + "epoch": 3.7014115092290987, + "grad_norm": 19.187866841083977, + "learning_rate": 3.334520908793527e-06, + "loss": 0.7495, + "step": 6818 + }, + { + "epoch": 3.701954397394137, + "grad_norm": 12.434378548251997, + "learning_rate": 3.3318996502112733e-06, + "loss": 0.4615, + "step": 6819 + }, + { + "epoch": 3.7024972855591747, + "grad_norm": 14.472283943179146, + "learning_rate": 3.3292792163502973e-06, + "loss": 0.6001, + "step": 6820 + }, + { + "epoch": 3.703040173724213, + "grad_norm": 10.597662524177974, + "learning_rate": 3.3266596075346993e-06, + "loss": 0.4247, + "step": 6821 + }, + { + "epoch": 3.7035830618892507, + "grad_norm": 10.287091606357066, + "learning_rate": 3.324040824088477e-06, + "loss": 0.3372, + "step": 6822 + }, + { + "epoch": 3.704125950054289, + "grad_norm": 16.36152453578531, + "learning_rate": 3.3214228663355252e-06, + "loss": 0.4507, + "step": 6823 + }, + { + "epoch": 3.7046688382193267, + "grad_norm": 11.166011131196704, + "learning_rate": 3.3188057345996372e-06, + "loss": 0.4297, + "step": 6824 + }, + { + "epoch": 3.705211726384365, + "grad_norm": 13.97058303490548, + "learning_rate": 3.316189429204505e-06, + "loss": 0.5737, + "step": 6825 + }, + { + "epoch": 3.7057546145494027, + "grad_norm": 13.078068910284376, + "learning_rate": 3.313573950473713e-06, + "loss": 0.3998, + "step": 6826 + }, + { + "epoch": 3.706297502714441, + "grad_norm": 17.144621177815008, + "learning_rate": 3.3109592987307583e-06, + "loss": 0.6534, + "step": 6827 + }, + { + "epoch": 3.7068403908794787, + "grad_norm": 12.887234362542056, + "learning_rate": 3.308345474299014e-06, + "loss": 0.6388, + "step": 6828 + }, + { + "epoch": 3.707383279044517, + "grad_norm": 14.117206302860215, + "learning_rate": 3.305732477501766e-06, + "loss": 0.8498, + "step": 6829 + }, + { + "epoch": 3.7079261672095547, + "grad_norm": 16.668445763886474, + "learning_rate": 3.3031203086621953e-06, + "loss": 0.5491, + "step": 6830 + }, + { + "epoch": 3.708469055374593, + "grad_norm": 9.687167039444047, + "learning_rate": 3.300508968103375e-06, + "loss": 0.3493, + "step": 6831 + }, + { + "epoch": 3.7090119435396307, + "grad_norm": 8.884631092991913, + "learning_rate": 3.2978984561482898e-06, + "loss": 0.3227, + "step": 6832 + }, + { + "epoch": 3.709554831704669, + "grad_norm": 10.708989687832828, + "learning_rate": 3.295288773119797e-06, + "loss": 0.5266, + "step": 6833 + }, + { + "epoch": 3.7100977198697067, + "grad_norm": 11.602581058072978, + "learning_rate": 3.292679919340682e-06, + "loss": 0.4903, + "step": 6834 + }, + { + "epoch": 3.710640608034745, + "grad_norm": 11.22107577843576, + "learning_rate": 3.2900718951335973e-06, + "loss": 0.2758, + "step": 6835 + }, + { + "epoch": 3.7111834961997827, + "grad_norm": 11.654780601665442, + "learning_rate": 3.2874647008211223e-06, + "loss": 0.4478, + "step": 6836 + }, + { + "epoch": 3.711726384364821, + "grad_norm": 13.952341454584447, + "learning_rate": 3.2848583367257037e-06, + "loss": 0.6784, + "step": 6837 + }, + { + "epoch": 3.7122692725298587, + "grad_norm": 9.937530151430312, + "learning_rate": 3.2822528031697123e-06, + "loss": 0.3417, + "step": 6838 + }, + { + "epoch": 3.712812160694897, + "grad_norm": 12.66955971380617, + "learning_rate": 3.279648100475403e-06, + "loss": 0.5978, + "step": 6839 + }, + { + "epoch": 3.7133550488599347, + "grad_norm": 11.746942398326137, + "learning_rate": 3.2770442289649218e-06, + "loss": 0.422, + "step": 6840 + }, + { + "epoch": 3.713897937024973, + "grad_norm": 9.742871784049646, + "learning_rate": 3.274441188960331e-06, + "loss": 0.2878, + "step": 6841 + }, + { + "epoch": 3.7144408251900107, + "grad_norm": 16.102195766631716, + "learning_rate": 3.271838980783567e-06, + "loss": 0.6011, + "step": 6842 + }, + { + "epoch": 3.714983713355049, + "grad_norm": 12.78519494193334, + "learning_rate": 3.2692376047564843e-06, + "loss": 0.5416, + "step": 6843 + }, + { + "epoch": 3.7155266015200867, + "grad_norm": 7.51189016165619, + "learning_rate": 3.266637061200821e-06, + "loss": 0.2993, + "step": 6844 + }, + { + "epoch": 3.716069489685125, + "grad_norm": 11.448398733111881, + "learning_rate": 3.264037350438218e-06, + "loss": 0.469, + "step": 6845 + }, + { + "epoch": 3.7166123778501627, + "grad_norm": 11.057137423687962, + "learning_rate": 3.2614384727902106e-06, + "loss": 0.5315, + "step": 6846 + }, + { + "epoch": 3.717155266015201, + "grad_norm": 10.948114702379822, + "learning_rate": 3.2588404285782327e-06, + "loss": 0.5224, + "step": 6847 + }, + { + "epoch": 3.7176981541802387, + "grad_norm": 14.169026887876377, + "learning_rate": 3.2562432181236182e-06, + "loss": 0.4101, + "step": 6848 + }, + { + "epoch": 3.718241042345277, + "grad_norm": 10.211418365179567, + "learning_rate": 3.253646841747583e-06, + "loss": 0.356, + "step": 6849 + }, + { + "epoch": 3.7187839305103148, + "grad_norm": 10.750722744873066, + "learning_rate": 3.251051299771263e-06, + "loss": 0.5381, + "step": 6850 + }, + { + "epoch": 3.719326818675353, + "grad_norm": 10.07832966516926, + "learning_rate": 3.2484565925156732e-06, + "loss": 0.3729, + "step": 6851 + }, + { + "epoch": 3.7198697068403908, + "grad_norm": 12.95382559219359, + "learning_rate": 3.2458627203017335e-06, + "loss": 0.5247, + "step": 6852 + }, + { + "epoch": 3.720412595005429, + "grad_norm": 11.982999422321642, + "learning_rate": 3.243269683450256e-06, + "loss": 0.4333, + "step": 6853 + }, + { + "epoch": 3.7209554831704668, + "grad_norm": 14.445578614379583, + "learning_rate": 3.240677482281953e-06, + "loss": 0.6296, + "step": 6854 + }, + { + "epoch": 3.721498371335505, + "grad_norm": 14.769553336755148, + "learning_rate": 3.2380861171174317e-06, + "loss": 0.48, + "step": 6855 + }, + { + "epoch": 3.7220412595005428, + "grad_norm": 13.97227360884286, + "learning_rate": 3.2354955882771965e-06, + "loss": 0.8019, + "step": 6856 + }, + { + "epoch": 3.722584147665581, + "grad_norm": 10.17994965485879, + "learning_rate": 3.2329058960816483e-06, + "loss": 0.4003, + "step": 6857 + }, + { + "epoch": 3.7231270358306188, + "grad_norm": 11.647057134479718, + "learning_rate": 3.2303170408510832e-06, + "loss": 0.367, + "step": 6858 + }, + { + "epoch": 3.723669923995657, + "grad_norm": 13.75957119148339, + "learning_rate": 3.227729022905697e-06, + "loss": 0.5791, + "step": 6859 + }, + { + "epoch": 3.7242128121606948, + "grad_norm": 9.205512068687561, + "learning_rate": 3.225141842565577e-06, + "loss": 0.3107, + "step": 6860 + }, + { + "epoch": 3.724755700325733, + "grad_norm": 11.81216687228549, + "learning_rate": 3.2225555001507124e-06, + "loss": 0.555, + "step": 6861 + }, + { + "epoch": 3.725298588490771, + "grad_norm": 9.593384384681405, + "learning_rate": 3.219969995980985e-06, + "loss": 0.4445, + "step": 6862 + }, + { + "epoch": 3.725841476655809, + "grad_norm": 6.990210033223102, + "learning_rate": 3.217385330376174e-06, + "loss": 0.1827, + "step": 6863 + }, + { + "epoch": 3.726384364820847, + "grad_norm": 16.17725365334651, + "learning_rate": 3.2148015036559522e-06, + "loss": 0.6442, + "step": 6864 + }, + { + "epoch": 3.726927252985885, + "grad_norm": 9.936484066410749, + "learning_rate": 3.2122185161399012e-06, + "loss": 0.3459, + "step": 6865 + }, + { + "epoch": 3.727470141150923, + "grad_norm": 16.59297040971604, + "learning_rate": 3.209636368147475e-06, + "loss": 0.5169, + "step": 6866 + }, + { + "epoch": 3.728013029315961, + "grad_norm": 14.156312439459766, + "learning_rate": 3.207055059998052e-06, + "loss": 0.6236, + "step": 6867 + }, + { + "epoch": 3.728555917480999, + "grad_norm": 10.18075916418277, + "learning_rate": 3.2044745920108832e-06, + "loss": 0.3108, + "step": 6868 + }, + { + "epoch": 3.729098805646037, + "grad_norm": 11.677366741727328, + "learning_rate": 3.2018949645051257e-06, + "loss": 0.5047, + "step": 6869 + }, + { + "epoch": 3.729641693811075, + "grad_norm": 12.56566346772868, + "learning_rate": 3.199316177799835e-06, + "loss": 0.4698, + "step": 6870 + }, + { + "epoch": 3.730184581976113, + "grad_norm": 14.518568356077754, + "learning_rate": 3.1967382322139538e-06, + "loss": 0.6106, + "step": 6871 + }, + { + "epoch": 3.730727470141151, + "grad_norm": 12.595862204241472, + "learning_rate": 3.194161128066338e-06, + "loss": 0.5617, + "step": 6872 + }, + { + "epoch": 3.731270358306189, + "grad_norm": 14.836027445862813, + "learning_rate": 3.191584865675713e-06, + "loss": 0.5989, + "step": 6873 + }, + { + "epoch": 3.731813246471227, + "grad_norm": 15.308949249609173, + "learning_rate": 3.189009445360731e-06, + "loss": 0.5385, + "step": 6874 + }, + { + "epoch": 3.732356134636265, + "grad_norm": 12.000029094911469, + "learning_rate": 3.1864348674399083e-06, + "loss": 0.4519, + "step": 6875 + }, + { + "epoch": 3.732899022801303, + "grad_norm": 15.618812442704373, + "learning_rate": 3.1838611322316836e-06, + "loss": 0.7854, + "step": 6876 + }, + { + "epoch": 3.733441910966341, + "grad_norm": 11.71029558715352, + "learning_rate": 3.181288240054381e-06, + "loss": 0.4303, + "step": 6877 + }, + { + "epoch": 3.733984799131379, + "grad_norm": 10.712473236039532, + "learning_rate": 3.17871619122621e-06, + "loss": 0.4896, + "step": 6878 + }, + { + "epoch": 3.734527687296417, + "grad_norm": 10.854545722563955, + "learning_rate": 3.176144986065298e-06, + "loss": 0.4795, + "step": 6879 + }, + { + "epoch": 3.735070575461455, + "grad_norm": 10.845848101056692, + "learning_rate": 3.1735746248896436e-06, + "loss": 0.3512, + "step": 6880 + }, + { + "epoch": 3.735613463626493, + "grad_norm": 11.675294486688255, + "learning_rate": 3.1710051080171657e-06, + "loss": 0.3521, + "step": 6881 + }, + { + "epoch": 3.736156351791531, + "grad_norm": 13.325204505254437, + "learning_rate": 3.168436435765654e-06, + "loss": 0.9413, + "step": 6882 + }, + { + "epoch": 3.736699239956569, + "grad_norm": 11.623023017043877, + "learning_rate": 3.1658686084528146e-06, + "loss": 0.491, + "step": 6883 + }, + { + "epoch": 3.737242128121607, + "grad_norm": 7.846579009143442, + "learning_rate": 3.163301626396237e-06, + "loss": 0.2837, + "step": 6884 + }, + { + "epoch": 3.737785016286645, + "grad_norm": 13.308504927624488, + "learning_rate": 3.1607354899134102e-06, + "loss": 0.5548, + "step": 6885 + }, + { + "epoch": 3.738327904451683, + "grad_norm": 8.7685040860308, + "learning_rate": 3.1581701993217197e-06, + "loss": 0.2844, + "step": 6886 + }, + { + "epoch": 3.738870792616721, + "grad_norm": 12.979195325374219, + "learning_rate": 3.1556057549384424e-06, + "loss": 0.6242, + "step": 6887 + }, + { + "epoch": 3.739413680781759, + "grad_norm": 15.955341391994917, + "learning_rate": 3.1530421570807536e-06, + "loss": 0.5053, + "step": 6888 + }, + { + "epoch": 3.739956568946797, + "grad_norm": 15.39892028189493, + "learning_rate": 3.1504794060657228e-06, + "loss": 0.8675, + "step": 6889 + }, + { + "epoch": 3.740499457111835, + "grad_norm": 15.189388217562758, + "learning_rate": 3.147917502210316e-06, + "loss": 0.6521, + "step": 6890 + }, + { + "epoch": 3.741042345276873, + "grad_norm": 15.85941037907273, + "learning_rate": 3.1453564458313923e-06, + "loss": 0.6711, + "step": 6891 + }, + { + "epoch": 3.741585233441911, + "grad_norm": 10.599235827332738, + "learning_rate": 3.142796237245709e-06, + "loss": 0.3489, + "step": 6892 + }, + { + "epoch": 3.742128121606949, + "grad_norm": 11.991913744676399, + "learning_rate": 3.1402368767699152e-06, + "loss": 0.4865, + "step": 6893 + }, + { + "epoch": 3.742671009771987, + "grad_norm": 15.429458030904042, + "learning_rate": 3.137678364720559e-06, + "loss": 0.8425, + "step": 6894 + }, + { + "epoch": 3.743213897937025, + "grad_norm": 13.89952284029879, + "learning_rate": 3.135120701414077e-06, + "loss": 0.5919, + "step": 6895 + }, + { + "epoch": 3.743756786102063, + "grad_norm": 14.355639876320694, + "learning_rate": 3.1325638871668094e-06, + "loss": 0.517, + "step": 6896 + }, + { + "epoch": 3.744299674267101, + "grad_norm": 9.998711427681842, + "learning_rate": 3.1300079222949854e-06, + "loss": 0.4837, + "step": 6897 + }, + { + "epoch": 3.744842562432139, + "grad_norm": 12.948353433621449, + "learning_rate": 3.1274528071147303e-06, + "loss": 0.4702, + "step": 6898 + }, + { + "epoch": 3.745385450597177, + "grad_norm": 11.300075344262957, + "learning_rate": 3.124898541942065e-06, + "loss": 0.4163, + "step": 6899 + }, + { + "epoch": 3.745928338762215, + "grad_norm": 10.014842311292028, + "learning_rate": 3.122345127092905e-06, + "loss": 0.4597, + "step": 6900 + }, + { + "epoch": 3.746471226927253, + "grad_norm": 14.937187078292135, + "learning_rate": 3.1197925628830616e-06, + "loss": 1.0359, + "step": 6901 + }, + { + "epoch": 3.747014115092291, + "grad_norm": 11.821812911772865, + "learning_rate": 3.117240849628239e-06, + "loss": 0.369, + "step": 6902 + }, + { + "epoch": 3.747557003257329, + "grad_norm": 18.601455125565817, + "learning_rate": 3.1146899876440383e-06, + "loss": 0.3967, + "step": 6903 + }, + { + "epoch": 3.748099891422367, + "grad_norm": 26.38167902221625, + "learning_rate": 3.1121399772459495e-06, + "loss": 0.7955, + "step": 6904 + }, + { + "epoch": 3.748642779587405, + "grad_norm": 9.036657138578375, + "learning_rate": 3.109590818749373e-06, + "loss": 0.3234, + "step": 6905 + }, + { + "epoch": 3.749185667752443, + "grad_norm": 12.029611743069063, + "learning_rate": 3.1070425124695825e-06, + "loss": 0.4689, + "step": 6906 + }, + { + "epoch": 3.749728555917481, + "grad_norm": 11.16075799831913, + "learning_rate": 3.1044950587217595e-06, + "loss": 0.455, + "step": 6907 + }, + { + "epoch": 3.750271444082519, + "grad_norm": 12.64123781208839, + "learning_rate": 3.1019484578209768e-06, + "loss": 0.4525, + "step": 6908 + }, + { + "epoch": 3.750814332247557, + "grad_norm": 10.873865130895668, + "learning_rate": 3.0994027100822e-06, + "loss": 0.7957, + "step": 6909 + }, + { + "epoch": 3.751357220412595, + "grad_norm": 15.826997708730714, + "learning_rate": 3.0968578158203e-06, + "loss": 0.4531, + "step": 6910 + }, + { + "epoch": 3.751900108577633, + "grad_norm": 16.864174578861103, + "learning_rate": 3.094313775350022e-06, + "loss": 0.6061, + "step": 6911 + }, + { + "epoch": 3.752442996742671, + "grad_norm": 16.360017183257842, + "learning_rate": 3.0917705889860283e-06, + "loss": 0.849, + "step": 6912 + }, + { + "epoch": 3.752985884907709, + "grad_norm": 11.473341112084933, + "learning_rate": 3.0892282570428513e-06, + "loss": 0.4494, + "step": 6913 + }, + { + "epoch": 3.753528773072747, + "grad_norm": 12.456658860480365, + "learning_rate": 3.0866867798349407e-06, + "loss": 0.4774, + "step": 6914 + }, + { + "epoch": 3.754071661237785, + "grad_norm": 16.476960663849457, + "learning_rate": 3.0841461576766284e-06, + "loss": 0.8323, + "step": 6915 + }, + { + "epoch": 3.754614549402823, + "grad_norm": 13.888913904073622, + "learning_rate": 3.0816063908821403e-06, + "loss": 0.5861, + "step": 6916 + }, + { + "epoch": 3.755157437567861, + "grad_norm": 10.764080500005052, + "learning_rate": 3.079067479765604e-06, + "loss": 0.3865, + "step": 6917 + }, + { + "epoch": 3.755700325732899, + "grad_norm": 13.47504295740381, + "learning_rate": 3.076529424641026e-06, + "loss": 0.628, + "step": 6918 + }, + { + "epoch": 3.756243213897937, + "grad_norm": 14.742498671541298, + "learning_rate": 3.0739922258223286e-06, + "loss": 0.6586, + "step": 6919 + }, + { + "epoch": 3.756786102062975, + "grad_norm": 11.002483024658138, + "learning_rate": 3.0714558836233056e-06, + "loss": 0.539, + "step": 6920 + }, + { + "epoch": 3.757328990228013, + "grad_norm": 13.952905049231472, + "learning_rate": 3.068920398357663e-06, + "loss": 0.8338, + "step": 6921 + }, + { + "epoch": 3.757871878393051, + "grad_norm": 10.43091762446858, + "learning_rate": 3.0663857703389943e-06, + "loss": 0.4779, + "step": 6922 + }, + { + "epoch": 3.758414766558089, + "grad_norm": 16.415629535468835, + "learning_rate": 3.0638519998807826e-06, + "loss": 0.4277, + "step": 6923 + }, + { + "epoch": 3.758957654723127, + "grad_norm": 8.867208839527924, + "learning_rate": 3.0613190872964104e-06, + "loss": 0.3777, + "step": 6924 + }, + { + "epoch": 3.759500542888165, + "grad_norm": 11.642294707216093, + "learning_rate": 3.058787032899152e-06, + "loss": 0.5056, + "step": 6925 + }, + { + "epoch": 3.760043431053203, + "grad_norm": 11.206786718860933, + "learning_rate": 3.0562558370021765e-06, + "loss": 0.446, + "step": 6926 + }, + { + "epoch": 3.760586319218241, + "grad_norm": 10.068491984488928, + "learning_rate": 3.053725499918545e-06, + "loss": 0.5076, + "step": 6927 + }, + { + "epoch": 3.761129207383279, + "grad_norm": 16.07206135268971, + "learning_rate": 3.0511960219612147e-06, + "loss": 0.5143, + "step": 6928 + }, + { + "epoch": 3.761672095548317, + "grad_norm": 11.602660115945492, + "learning_rate": 3.0486674034430352e-06, + "loss": 0.508, + "step": 6929 + }, + { + "epoch": 3.762214983713355, + "grad_norm": 13.3862945751605, + "learning_rate": 3.046139644676751e-06, + "loss": 0.3828, + "step": 6930 + }, + { + "epoch": 3.762757871878393, + "grad_norm": 10.733311225895106, + "learning_rate": 3.0436127459749975e-06, + "loss": 0.37, + "step": 6931 + }, + { + "epoch": 3.763300760043431, + "grad_norm": 11.261346397862807, + "learning_rate": 3.041086707650306e-06, + "loss": 0.4808, + "step": 6932 + }, + { + "epoch": 3.763843648208469, + "grad_norm": 16.501582349438152, + "learning_rate": 3.0385615300151018e-06, + "loss": 0.6239, + "step": 6933 + }, + { + "epoch": 3.764386536373507, + "grad_norm": 11.553010310131304, + "learning_rate": 3.036037213381702e-06, + "loss": 0.4541, + "step": 6934 + }, + { + "epoch": 3.764929424538545, + "grad_norm": 11.723556454514561, + "learning_rate": 3.0335137580623154e-06, + "loss": 0.4217, + "step": 6935 + }, + { + "epoch": 3.765472312703583, + "grad_norm": 15.843208183133134, + "learning_rate": 3.030991164369057e-06, + "loss": 0.628, + "step": 6936 + }, + { + "epoch": 3.766015200868621, + "grad_norm": 8.046500189967162, + "learning_rate": 3.028469432613915e-06, + "loss": 0.288, + "step": 6937 + }, + { + "epoch": 3.766558089033659, + "grad_norm": 10.115796532018264, + "learning_rate": 3.0259485631087846e-06, + "loss": 0.3899, + "step": 6938 + }, + { + "epoch": 3.767100977198697, + "grad_norm": 13.289116360712832, + "learning_rate": 3.023428556165451e-06, + "loss": 0.6068, + "step": 6939 + }, + { + "epoch": 3.767643865363735, + "grad_norm": 11.871048514073903, + "learning_rate": 3.020909412095592e-06, + "loss": 0.6089, + "step": 6940 + }, + { + "epoch": 3.768186753528773, + "grad_norm": 8.084122023965138, + "learning_rate": 3.0183911312107806e-06, + "loss": 0.3247, + "step": 6941 + }, + { + "epoch": 3.768729641693811, + "grad_norm": 10.871234822080087, + "learning_rate": 3.015873713822478e-06, + "loss": 0.4928, + "step": 6942 + }, + { + "epoch": 3.769272529858849, + "grad_norm": 13.956354691016482, + "learning_rate": 3.013357160242051e-06, + "loss": 0.7454, + "step": 6943 + }, + { + "epoch": 3.769815418023887, + "grad_norm": 17.342203574294, + "learning_rate": 3.01084147078074e-06, + "loss": 0.6629, + "step": 6944 + }, + { + "epoch": 3.770358306188925, + "grad_norm": 9.435776085708705, + "learning_rate": 3.008326645749701e-06, + "loss": 0.2969, + "step": 6945 + }, + { + "epoch": 3.770901194353963, + "grad_norm": 14.35523924414514, + "learning_rate": 3.0058126854599624e-06, + "loss": 0.6859, + "step": 6946 + }, + { + "epoch": 3.771444082519001, + "grad_norm": 10.162640497221313, + "learning_rate": 3.003299590222454e-06, + "loss": 0.4144, + "step": 6947 + }, + { + "epoch": 3.771986970684039, + "grad_norm": 10.611516903885335, + "learning_rate": 3.0007873603480107e-06, + "loss": 0.5975, + "step": 6948 + }, + { + "epoch": 3.772529858849077, + "grad_norm": 13.809727664609994, + "learning_rate": 2.998275996147335e-06, + "loss": 0.4714, + "step": 6949 + }, + { + "epoch": 3.773072747014115, + "grad_norm": 11.317211520834723, + "learning_rate": 2.995765497931049e-06, + "loss": 0.4646, + "step": 6950 + }, + { + "epoch": 3.773615635179153, + "grad_norm": 10.86923796914853, + "learning_rate": 2.9932558660096443e-06, + "loss": 0.3228, + "step": 6951 + }, + { + "epoch": 3.774158523344191, + "grad_norm": 7.470997903259584, + "learning_rate": 2.990747100693526e-06, + "loss": 0.2531, + "step": 6952 + }, + { + "epoch": 3.774701411509229, + "grad_norm": 8.627824874780737, + "learning_rate": 2.988239202292972e-06, + "loss": 0.361, + "step": 6953 + }, + { + "epoch": 3.775244299674267, + "grad_norm": 9.14607939221467, + "learning_rate": 2.985732171118172e-06, + "loss": 0.3116, + "step": 6954 + }, + { + "epoch": 3.7757871878393052, + "grad_norm": 10.155112379070209, + "learning_rate": 2.9832260074791984e-06, + "loss": 0.4723, + "step": 6955 + }, + { + "epoch": 3.776330076004343, + "grad_norm": 10.506852840497306, + "learning_rate": 2.9807207116860094e-06, + "loss": 0.4232, + "step": 6956 + }, + { + "epoch": 3.7768729641693812, + "grad_norm": 11.193678853889507, + "learning_rate": 2.9782162840484765e-06, + "loss": 0.4915, + "step": 6957 + }, + { + "epoch": 3.777415852334419, + "grad_norm": 14.323380391376173, + "learning_rate": 2.9757127248763375e-06, + "loss": 0.6207, + "step": 6958 + }, + { + "epoch": 3.7779587404994572, + "grad_norm": 11.773695664037353, + "learning_rate": 2.973210034479247e-06, + "loss": 0.3957, + "step": 6959 + }, + { + "epoch": 3.778501628664495, + "grad_norm": 8.978033406900165, + "learning_rate": 2.97070821316674e-06, + "loss": 0.2935, + "step": 6960 + }, + { + "epoch": 3.7790445168295332, + "grad_norm": 12.4187581608995, + "learning_rate": 2.9682072612482427e-06, + "loss": 0.515, + "step": 6961 + }, + { + "epoch": 3.779587404994571, + "grad_norm": 12.551466191074608, + "learning_rate": 2.965707179033078e-06, + "loss": 0.4014, + "step": 6962 + }, + { + "epoch": 3.7801302931596092, + "grad_norm": 8.193678184944888, + "learning_rate": 2.9632079668304613e-06, + "loss": 0.2947, + "step": 6963 + }, + { + "epoch": 3.780673181324647, + "grad_norm": 12.277305597806611, + "learning_rate": 2.960709624949498e-06, + "loss": 0.4283, + "step": 6964 + }, + { + "epoch": 3.7812160694896852, + "grad_norm": 12.873825056251757, + "learning_rate": 2.958212153699187e-06, + "loss": 0.3719, + "step": 6965 + }, + { + "epoch": 3.781758957654723, + "grad_norm": 11.733322711394022, + "learning_rate": 2.955715553388421e-06, + "loss": 0.4547, + "step": 6966 + }, + { + "epoch": 3.7823018458197613, + "grad_norm": 13.421562379885513, + "learning_rate": 2.95321982432598e-06, + "loss": 0.8095, + "step": 6967 + }, + { + "epoch": 3.782844733984799, + "grad_norm": 10.751529697853956, + "learning_rate": 2.950724966820544e-06, + "loss": 0.3795, + "step": 6968 + }, + { + "epoch": 3.7833876221498373, + "grad_norm": 11.23769589292945, + "learning_rate": 2.9482309811806785e-06, + "loss": 0.4324, + "step": 6969 + }, + { + "epoch": 3.783930510314875, + "grad_norm": 10.760821500115565, + "learning_rate": 2.945737867714844e-06, + "loss": 0.4112, + "step": 6970 + }, + { + "epoch": 3.7844733984799133, + "grad_norm": 10.577088653058517, + "learning_rate": 2.9432456267313924e-06, + "loss": 0.3819, + "step": 6971 + }, + { + "epoch": 3.785016286644951, + "grad_norm": 9.877099189181129, + "learning_rate": 2.9407542585385683e-06, + "loss": 0.2344, + "step": 6972 + }, + { + "epoch": 3.7855591748099893, + "grad_norm": 18.577198876664223, + "learning_rate": 2.9382637634445087e-06, + "loss": 0.3254, + "step": 6973 + }, + { + "epoch": 3.786102062975027, + "grad_norm": 14.663768331220036, + "learning_rate": 2.9357741417572415e-06, + "loss": 0.5764, + "step": 6974 + }, + { + "epoch": 3.7866449511400653, + "grad_norm": 12.695557652262636, + "learning_rate": 2.9332853937846873e-06, + "loss": 0.4016, + "step": 6975 + }, + { + "epoch": 3.787187839305103, + "grad_norm": 13.456679312848303, + "learning_rate": 2.930797519834658e-06, + "loss": 0.438, + "step": 6976 + }, + { + "epoch": 3.7877307274701413, + "grad_norm": 10.441195425871536, + "learning_rate": 2.928310520214859e-06, + "loss": 0.4062, + "step": 6977 + }, + { + "epoch": 3.788273615635179, + "grad_norm": 10.300692057422426, + "learning_rate": 2.9258243952328846e-06, + "loss": 0.368, + "step": 6978 + }, + { + "epoch": 3.7888165038002173, + "grad_norm": 11.532450200794159, + "learning_rate": 2.9233391451962233e-06, + "loss": 0.5835, + "step": 6979 + }, + { + "epoch": 3.789359391965255, + "grad_norm": 13.465329246550354, + "learning_rate": 2.9208547704122525e-06, + "loss": 0.5226, + "step": 6980 + }, + { + "epoch": 3.7899022801302933, + "grad_norm": 14.107590055474905, + "learning_rate": 2.9183712711882526e-06, + "loss": 0.6195, + "step": 6981 + }, + { + "epoch": 3.790445168295331, + "grad_norm": 10.577784567014287, + "learning_rate": 2.915888647831374e-06, + "loss": 0.3655, + "step": 6982 + }, + { + "epoch": 3.7909880564603693, + "grad_norm": 13.136687160957166, + "learning_rate": 2.913406900648683e-06, + "loss": 0.6708, + "step": 6983 + }, + { + "epoch": 3.791530944625407, + "grad_norm": 15.57844160050923, + "learning_rate": 2.9109260299471165e-06, + "loss": 0.7229, + "step": 6984 + }, + { + "epoch": 3.7920738327904453, + "grad_norm": 14.528051815895886, + "learning_rate": 2.908446036033519e-06, + "loss": 0.4592, + "step": 6985 + }, + { + "epoch": 3.792616720955483, + "grad_norm": 20.212297672580796, + "learning_rate": 2.90596691921462e-06, + "loss": 0.759, + "step": 6986 + }, + { + "epoch": 3.7931596091205213, + "grad_norm": 9.84665764740946, + "learning_rate": 2.9034886797970342e-06, + "loss": 0.3681, + "step": 6987 + }, + { + "epoch": 3.793702497285559, + "grad_norm": 8.475948428830446, + "learning_rate": 2.9010113180872847e-06, + "loss": 0.2304, + "step": 6988 + }, + { + "epoch": 3.7942453854505973, + "grad_norm": 11.955148009858453, + "learning_rate": 2.898534834391762e-06, + "loss": 0.5278, + "step": 6989 + }, + { + "epoch": 3.794788273615635, + "grad_norm": 18.8535666021868, + "learning_rate": 2.8960592290167754e-06, + "loss": 1.1773, + "step": 6990 + }, + { + "epoch": 3.7953311617806733, + "grad_norm": 14.603134582974139, + "learning_rate": 2.8935845022685006e-06, + "loss": 0.7039, + "step": 6991 + }, + { + "epoch": 3.795874049945711, + "grad_norm": 17.09419124573872, + "learning_rate": 2.891110654453022e-06, + "loss": 0.6918, + "step": 6992 + }, + { + "epoch": 3.7964169381107493, + "grad_norm": 14.480524368793315, + "learning_rate": 2.888637685876309e-06, + "loss": 0.5987, + "step": 6993 + }, + { + "epoch": 3.796959826275787, + "grad_norm": 13.958888615564357, + "learning_rate": 2.8861655968442192e-06, + "loss": 0.7819, + "step": 6994 + }, + { + "epoch": 3.7975027144408253, + "grad_norm": 13.785514235662655, + "learning_rate": 2.883694387662511e-06, + "loss": 0.6469, + "step": 6995 + }, + { + "epoch": 3.798045602605863, + "grad_norm": 15.113325361621122, + "learning_rate": 2.8812240586368157e-06, + "loss": 0.6732, + "step": 6996 + }, + { + "epoch": 3.7985884907709013, + "grad_norm": 13.447315748094697, + "learning_rate": 2.8787546100726773e-06, + "loss": 0.8885, + "step": 6997 + }, + { + "epoch": 3.799131378935939, + "grad_norm": 15.942374082087648, + "learning_rate": 2.876286042275518e-06, + "loss": 0.5496, + "step": 6998 + }, + { + "epoch": 3.7996742671009773, + "grad_norm": 13.841633437355465, + "learning_rate": 2.873818355550655e-06, + "loss": 0.2924, + "step": 6999 + }, + { + "epoch": 3.800217155266015, + "grad_norm": 12.814203732560337, + "learning_rate": 2.871351550203295e-06, + "loss": 0.5538, + "step": 7000 + }, + { + "epoch": 3.8007600434310533, + "grad_norm": 12.191028501042764, + "learning_rate": 2.8688856265385367e-06, + "loss": 0.4576, + "step": 7001 + }, + { + "epoch": 3.801302931596091, + "grad_norm": 10.517009605166134, + "learning_rate": 2.86642058486137e-06, + "loss": 0.4275, + "step": 7002 + }, + { + "epoch": 3.8018458197611293, + "grad_norm": 13.976454434703868, + "learning_rate": 2.863956425476674e-06, + "loss": 0.5466, + "step": 7003 + }, + { + "epoch": 3.802388707926167, + "grad_norm": 11.5016640245614, + "learning_rate": 2.8614931486892207e-06, + "loss": 0.366, + "step": 7004 + }, + { + "epoch": 3.8029315960912053, + "grad_norm": 11.939582442434286, + "learning_rate": 2.8590307548036712e-06, + "loss": 0.4141, + "step": 7005 + }, + { + "epoch": 3.803474484256243, + "grad_norm": 12.987510367990025, + "learning_rate": 2.8565692441245796e-06, + "loss": 0.5152, + "step": 7006 + }, + { + "epoch": 3.8040173724212814, + "grad_norm": 12.832896642663844, + "learning_rate": 2.8541086169563894e-06, + "loss": 0.4585, + "step": 7007 + }, + { + "epoch": 3.804560260586319, + "grad_norm": 11.863347803098113, + "learning_rate": 2.851648873603433e-06, + "loss": 0.4196, + "step": 7008 + }, + { + "epoch": 3.8051031487513574, + "grad_norm": 13.239707667599935, + "learning_rate": 2.8491900143699367e-06, + "loss": 0.5748, + "step": 7009 + }, + { + "epoch": 3.805646036916395, + "grad_norm": 12.665696921336465, + "learning_rate": 2.846732039560017e-06, + "loss": 0.4688, + "step": 7010 + }, + { + "epoch": 3.8061889250814334, + "grad_norm": 10.268698833296764, + "learning_rate": 2.844274949477679e-06, + "loss": 0.3858, + "step": 7011 + }, + { + "epoch": 3.806731813246471, + "grad_norm": 12.660908637644715, + "learning_rate": 2.8418187444268198e-06, + "loss": 0.445, + "step": 7012 + }, + { + "epoch": 3.8072747014115094, + "grad_norm": 10.571667634245497, + "learning_rate": 2.8393634247112232e-06, + "loss": 0.4066, + "step": 7013 + }, + { + "epoch": 3.807817589576547, + "grad_norm": 19.94546394586301, + "learning_rate": 2.836908990634578e-06, + "loss": 0.5101, + "step": 7014 + }, + { + "epoch": 3.8083604777415854, + "grad_norm": 13.85083509805195, + "learning_rate": 2.834455442500441e-06, + "loss": 0.4073, + "step": 7015 + }, + { + "epoch": 3.808903365906623, + "grad_norm": 10.847008608128625, + "learning_rate": 2.8320027806122753e-06, + "loss": 0.3582, + "step": 7016 + }, + { + "epoch": 3.8094462540716614, + "grad_norm": 13.798257428735926, + "learning_rate": 2.82955100527343e-06, + "loss": 0.7057, + "step": 7017 + }, + { + "epoch": 3.809989142236699, + "grad_norm": 15.062376260700791, + "learning_rate": 2.8271001167871405e-06, + "loss": 0.6019, + "step": 7018 + }, + { + "epoch": 3.8105320304017374, + "grad_norm": 11.109879992701345, + "learning_rate": 2.8246501154565476e-06, + "loss": 0.3265, + "step": 7019 + }, + { + "epoch": 3.811074918566775, + "grad_norm": 13.70633009915418, + "learning_rate": 2.822201001584657e-06, + "loss": 0.4019, + "step": 7020 + }, + { + "epoch": 3.8116178067318134, + "grad_norm": 12.339132223947464, + "learning_rate": 2.8197527754743926e-06, + "loss": 0.3272, + "step": 7021 + }, + { + "epoch": 3.812160694896851, + "grad_norm": 9.817254835111205, + "learning_rate": 2.8173054374285434e-06, + "loss": 0.4118, + "step": 7022 + }, + { + "epoch": 3.8127035830618894, + "grad_norm": 15.18363193003276, + "learning_rate": 2.81485898774981e-06, + "loss": 0.7261, + "step": 7023 + }, + { + "epoch": 3.813246471226927, + "grad_norm": 11.984870936412664, + "learning_rate": 2.8124134267407665e-06, + "loss": 0.3734, + "step": 7024 + }, + { + "epoch": 3.8137893593919654, + "grad_norm": 10.678523359563961, + "learning_rate": 2.8099687547038824e-06, + "loss": 0.5318, + "step": 7025 + }, + { + "epoch": 3.814332247557003, + "grad_norm": 18.489591388355425, + "learning_rate": 2.807524971941528e-06, + "loss": 0.8301, + "step": 7026 + }, + { + "epoch": 3.8148751357220414, + "grad_norm": 10.459115256144122, + "learning_rate": 2.805082078755942e-06, + "loss": 0.3638, + "step": 7027 + }, + { + "epoch": 3.815418023887079, + "grad_norm": 12.778148708574054, + "learning_rate": 2.8026400754492767e-06, + "loss": 0.6202, + "step": 7028 + }, + { + "epoch": 3.8159609120521174, + "grad_norm": 14.76075800024929, + "learning_rate": 2.8001989623235528e-06, + "loss": 0.8516, + "step": 7029 + }, + { + "epoch": 3.816503800217155, + "grad_norm": 15.243047141569159, + "learning_rate": 2.797758739680698e-06, + "loss": 0.7363, + "step": 7030 + }, + { + "epoch": 3.8170466883821934, + "grad_norm": 12.332962959117395, + "learning_rate": 2.7953194078225223e-06, + "loss": 0.5528, + "step": 7031 + }, + { + "epoch": 3.817589576547231, + "grad_norm": 14.15721396911131, + "learning_rate": 2.792880967050724e-06, + "loss": 0.6093, + "step": 7032 + }, + { + "epoch": 3.8181324647122694, + "grad_norm": 16.149812128817175, + "learning_rate": 2.790443417666894e-06, + "loss": 0.7232, + "step": 7033 + }, + { + "epoch": 3.818675352877307, + "grad_norm": 6.884614594609857, + "learning_rate": 2.7880067599725123e-06, + "loss": 0.2341, + "step": 7034 + }, + { + "epoch": 3.8192182410423454, + "grad_norm": 13.604715542482158, + "learning_rate": 2.7855709942689525e-06, + "loss": 0.5522, + "step": 7035 + }, + { + "epoch": 3.819761129207383, + "grad_norm": 12.822857567671976, + "learning_rate": 2.7831361208574636e-06, + "loss": 0.459, + "step": 7036 + }, + { + "epoch": 3.8203040173724214, + "grad_norm": 16.80060104026762, + "learning_rate": 2.780702140039204e-06, + "loss": 0.6408, + "step": 7037 + }, + { + "epoch": 3.820846905537459, + "grad_norm": 9.25304440984111, + "learning_rate": 2.77826905211521e-06, + "loss": 0.3628, + "step": 7038 + }, + { + "epoch": 3.8213897937024974, + "grad_norm": 9.5263200358754, + "learning_rate": 2.775836857386409e-06, + "loss": 0.4398, + "step": 7039 + }, + { + "epoch": 3.821932681867535, + "grad_norm": 9.394535823999632, + "learning_rate": 2.77340555615362e-06, + "loss": 0.4318, + "step": 7040 + }, + { + "epoch": 3.8224755700325734, + "grad_norm": 16.75570479373163, + "learning_rate": 2.7709751487175486e-06, + "loss": 0.621, + "step": 7041 + }, + { + "epoch": 3.823018458197611, + "grad_norm": 11.68838601944592, + "learning_rate": 2.768545635378792e-06, + "loss": 0.4332, + "step": 7042 + }, + { + "epoch": 3.8235613463626494, + "grad_norm": 13.068740742878463, + "learning_rate": 2.7661170164378372e-06, + "loss": 0.6055, + "step": 7043 + }, + { + "epoch": 3.824104234527687, + "grad_norm": 12.808750119841285, + "learning_rate": 2.7636892921950586e-06, + "loss": 0.3966, + "step": 7044 + }, + { + "epoch": 3.8246471226927254, + "grad_norm": 13.368527360017245, + "learning_rate": 2.7612624629507213e-06, + "loss": 0.4222, + "step": 7045 + }, + { + "epoch": 3.8251900108577632, + "grad_norm": 14.624704158928514, + "learning_rate": 2.758836529004979e-06, + "loss": 0.5196, + "step": 7046 + }, + { + "epoch": 3.8257328990228014, + "grad_norm": 11.425494087567786, + "learning_rate": 2.756411490657875e-06, + "loss": 0.5476, + "step": 7047 + }, + { + "epoch": 3.8262757871878392, + "grad_norm": 11.438998129235056, + "learning_rate": 2.753987348209344e-06, + "loss": 0.3902, + "step": 7048 + }, + { + "epoch": 3.8268186753528775, + "grad_norm": 13.662522916551874, + "learning_rate": 2.7515641019592053e-06, + "loss": 0.494, + "step": 7049 + }, + { + "epoch": 3.8273615635179152, + "grad_norm": 10.296671849651366, + "learning_rate": 2.7491417522071706e-06, + "loss": 0.3226, + "step": 7050 + }, + { + "epoch": 3.8279044516829535, + "grad_norm": 10.454941310693076, + "learning_rate": 2.7467202992528376e-06, + "loss": 0.2803, + "step": 7051 + }, + { + "epoch": 3.8284473398479912, + "grad_norm": 11.648384847247753, + "learning_rate": 2.744299743395703e-06, + "loss": 0.5601, + "step": 7052 + }, + { + "epoch": 3.8289902280130295, + "grad_norm": 8.944598429537628, + "learning_rate": 2.741880084935138e-06, + "loss": 0.3296, + "step": 7053 + }, + { + "epoch": 3.8295331161780672, + "grad_norm": 13.218169787059113, + "learning_rate": 2.7394613241704117e-06, + "loss": 0.5443, + "step": 7054 + }, + { + "epoch": 3.8300760043431055, + "grad_norm": 9.804324070437223, + "learning_rate": 2.7370434614006803e-06, + "loss": 0.4269, + "step": 7055 + }, + { + "epoch": 3.8306188925081432, + "grad_norm": 16.684309046147042, + "learning_rate": 2.734626496924986e-06, + "loss": 0.7608, + "step": 7056 + }, + { + "epoch": 3.8311617806731815, + "grad_norm": 11.594370437365152, + "learning_rate": 2.7322104310422713e-06, + "loss": 0.4286, + "step": 7057 + }, + { + "epoch": 3.8317046688382193, + "grad_norm": 9.387708800824917, + "learning_rate": 2.7297952640513483e-06, + "loss": 0.4659, + "step": 7058 + }, + { + "epoch": 3.8322475570032575, + "grad_norm": 15.36709432735338, + "learning_rate": 2.727380996250939e-06, + "loss": 0.4572, + "step": 7059 + }, + { + "epoch": 3.8327904451682953, + "grad_norm": 10.866529562034719, + "learning_rate": 2.724967627939634e-06, + "loss": 0.5208, + "step": 7060 + }, + { + "epoch": 3.8333333333333335, + "grad_norm": 11.215185901261178, + "learning_rate": 2.722555159415934e-06, + "loss": 0.4077, + "step": 7061 + }, + { + "epoch": 3.8338762214983713, + "grad_norm": 13.560818928747905, + "learning_rate": 2.7201435909782027e-06, + "loss": 0.4849, + "step": 7062 + }, + { + "epoch": 3.8344191096634095, + "grad_norm": 12.76281749363346, + "learning_rate": 2.7177329229247186e-06, + "loss": 0.6375, + "step": 7063 + }, + { + "epoch": 3.8349619978284473, + "grad_norm": 11.109626055006531, + "learning_rate": 2.715323155553635e-06, + "loss": 0.6538, + "step": 7064 + }, + { + "epoch": 3.8355048859934855, + "grad_norm": 16.73906444542459, + "learning_rate": 2.712914289162989e-06, + "loss": 0.4548, + "step": 7065 + }, + { + "epoch": 3.8360477741585233, + "grad_norm": 15.832313422557572, + "learning_rate": 2.7105063240507222e-06, + "loss": 0.5417, + "step": 7066 + }, + { + "epoch": 3.8365906623235615, + "grad_norm": 10.615862418603832, + "learning_rate": 2.708099260514645e-06, + "loss": 0.6034, + "step": 7067 + }, + { + "epoch": 3.8371335504885993, + "grad_norm": 16.513578882781225, + "learning_rate": 2.7056930988524763e-06, + "loss": 0.9556, + "step": 7068 + }, + { + "epoch": 3.8376764386536375, + "grad_norm": 17.31989234293219, + "learning_rate": 2.703287839361811e-06, + "loss": 0.744, + "step": 7069 + }, + { + "epoch": 3.8382193268186753, + "grad_norm": 10.336380611592002, + "learning_rate": 2.7008834823401344e-06, + "loss": 0.5354, + "step": 7070 + }, + { + "epoch": 3.8387622149837135, + "grad_norm": 13.577288380592488, + "learning_rate": 2.698480028084821e-06, + "loss": 0.5647, + "step": 7071 + }, + { + "epoch": 3.8393051031487513, + "grad_norm": 15.062012119646445, + "learning_rate": 2.6960774768931365e-06, + "loss": 0.6897, + "step": 7072 + }, + { + "epoch": 3.8398479913137895, + "grad_norm": 9.679061664826015, + "learning_rate": 2.6936758290622324e-06, + "loss": 0.3364, + "step": 7073 + }, + { + "epoch": 3.8403908794788273, + "grad_norm": 12.06717080820273, + "learning_rate": 2.69127508488914e-06, + "loss": 0.473, + "step": 7074 + }, + { + "epoch": 3.8409337676438655, + "grad_norm": 12.515990246581914, + "learning_rate": 2.6888752446707965e-06, + "loss": 0.7188, + "step": 7075 + }, + { + "epoch": 3.8414766558089033, + "grad_norm": 12.151182717520161, + "learning_rate": 2.6864763087040145e-06, + "loss": 0.4494, + "step": 7076 + }, + { + "epoch": 3.8420195439739415, + "grad_norm": 13.03785855389188, + "learning_rate": 2.6840782772855002e-06, + "loss": 0.6012, + "step": 7077 + }, + { + "epoch": 3.8425624321389793, + "grad_norm": 13.049506054659528, + "learning_rate": 2.6816811507118436e-06, + "loss": 0.6763, + "step": 7078 + }, + { + "epoch": 3.8431053203040175, + "grad_norm": 10.797345758374234, + "learning_rate": 2.6792849292795253e-06, + "loss": 0.3712, + "step": 7079 + }, + { + "epoch": 3.8436482084690553, + "grad_norm": 7.303579416108625, + "learning_rate": 2.6768896132849144e-06, + "loss": 0.2622, + "step": 7080 + }, + { + "epoch": 3.8441910966340935, + "grad_norm": 12.992302429769211, + "learning_rate": 2.674495203024269e-06, + "loss": 0.5022, + "step": 7081 + }, + { + "epoch": 3.8447339847991313, + "grad_norm": 11.355323594435934, + "learning_rate": 2.67210169879373e-06, + "loss": 0.3898, + "step": 7082 + }, + { + "epoch": 3.8452768729641695, + "grad_norm": 10.043141357807462, + "learning_rate": 2.669709100889333e-06, + "loss": 0.4849, + "step": 7083 + }, + { + "epoch": 3.8458197611292073, + "grad_norm": 10.154968040172625, + "learning_rate": 2.6673174096069976e-06, + "loss": 0.6587, + "step": 7084 + }, + { + "epoch": 3.8463626492942455, + "grad_norm": 15.332496592868017, + "learning_rate": 2.6649266252425297e-06, + "loss": 0.5297, + "step": 7085 + }, + { + "epoch": 3.8469055374592833, + "grad_norm": 12.176208814956482, + "learning_rate": 2.6625367480916285e-06, + "loss": 0.603, + "step": 7086 + }, + { + "epoch": 3.8474484256243215, + "grad_norm": 12.08600431730549, + "learning_rate": 2.660147778449876e-06, + "loss": 0.6609, + "step": 7087 + }, + { + "epoch": 3.8479913137893593, + "grad_norm": 12.157193141505328, + "learning_rate": 2.6577597166127455e-06, + "loss": 0.4869, + "step": 7088 + }, + { + "epoch": 3.8485342019543975, + "grad_norm": 13.235050132598122, + "learning_rate": 2.655372562875591e-06, + "loss": 0.4279, + "step": 7089 + }, + { + "epoch": 3.8490770901194353, + "grad_norm": 12.171527155216868, + "learning_rate": 2.652986317533669e-06, + "loss": 0.4388, + "step": 7090 + }, + { + "epoch": 3.8496199782844736, + "grad_norm": 10.123066437797494, + "learning_rate": 2.650600980882104e-06, + "loss": 0.3086, + "step": 7091 + }, + { + "epoch": 3.8501628664495113, + "grad_norm": 12.006465921202063, + "learning_rate": 2.6482165532159275e-06, + "loss": 0.4117, + "step": 7092 + }, + { + "epoch": 3.8507057546145496, + "grad_norm": 15.608180907150922, + "learning_rate": 2.645833034830043e-06, + "loss": 0.5819, + "step": 7093 + }, + { + "epoch": 3.8512486427795873, + "grad_norm": 18.214304962348578, + "learning_rate": 2.6434504260192485e-06, + "loss": 0.6414, + "step": 7094 + }, + { + "epoch": 3.8517915309446256, + "grad_norm": 9.263854181848101, + "learning_rate": 2.64106872707823e-06, + "loss": 0.3855, + "step": 7095 + }, + { + "epoch": 3.8523344191096633, + "grad_norm": 11.56441830671296, + "learning_rate": 2.638687938301557e-06, + "loss": 0.3565, + "step": 7096 + }, + { + "epoch": 3.8528773072747016, + "grad_norm": 13.250463401216958, + "learning_rate": 2.636308059983699e-06, + "loss": 0.6425, + "step": 7097 + }, + { + "epoch": 3.8534201954397393, + "grad_norm": 14.438578870413643, + "learning_rate": 2.6339290924189886e-06, + "loss": 0.5401, + "step": 7098 + }, + { + "epoch": 3.8539630836047776, + "grad_norm": 19.454185693944922, + "learning_rate": 2.6315510359016736e-06, + "loss": 0.5113, + "step": 7099 + }, + { + "epoch": 3.8545059717698154, + "grad_norm": 14.281154826114456, + "learning_rate": 2.629173890725866e-06, + "loss": 0.6589, + "step": 7100 + }, + { + "epoch": 3.8550488599348536, + "grad_norm": 10.71424151041571, + "learning_rate": 2.62679765718558e-06, + "loss": 0.5477, + "step": 7101 + }, + { + "epoch": 3.8555917480998914, + "grad_norm": 15.907412596919608, + "learning_rate": 2.6244223355747144e-06, + "loss": 0.5586, + "step": 7102 + }, + { + "epoch": 3.8561346362649296, + "grad_norm": 14.500284217616562, + "learning_rate": 2.6220479261870436e-06, + "loss": 0.4131, + "step": 7103 + }, + { + "epoch": 3.8566775244299674, + "grad_norm": 16.114243210903357, + "learning_rate": 2.6196744293162503e-06, + "loss": 0.8063, + "step": 7104 + }, + { + "epoch": 3.8572204125950056, + "grad_norm": 10.762654220587946, + "learning_rate": 2.6173018452558786e-06, + "loss": 0.3736, + "step": 7105 + }, + { + "epoch": 3.8577633007600434, + "grad_norm": 8.619518240537495, + "learning_rate": 2.614930174299388e-06, + "loss": 0.4294, + "step": 7106 + }, + { + "epoch": 3.8583061889250816, + "grad_norm": 13.731123733938277, + "learning_rate": 2.612559416740098e-06, + "loss": 0.5039, + "step": 7107 + }, + { + "epoch": 3.8588490770901194, + "grad_norm": 9.925845082716284, + "learning_rate": 2.6101895728712355e-06, + "loss": 0.4914, + "step": 7108 + }, + { + "epoch": 3.8593919652551576, + "grad_norm": 9.305166559901256, + "learning_rate": 2.6078206429859044e-06, + "loss": 0.2875, + "step": 7109 + }, + { + "epoch": 3.8599348534201954, + "grad_norm": 11.732054868905763, + "learning_rate": 2.6054526273770975e-06, + "loss": 0.3843, + "step": 7110 + }, + { + "epoch": 3.8604777415852336, + "grad_norm": 11.857351610067585, + "learning_rate": 2.603085526337694e-06, + "loss": 0.431, + "step": 7111 + }, + { + "epoch": 3.8610206297502714, + "grad_norm": 8.758502750249699, + "learning_rate": 2.6007193401604626e-06, + "loss": 0.2799, + "step": 7112 + }, + { + "epoch": 3.8615635179153096, + "grad_norm": 11.732493468231032, + "learning_rate": 2.598354069138056e-06, + "loss": 0.4618, + "step": 7113 + }, + { + "epoch": 3.8621064060803474, + "grad_norm": 12.849925851696895, + "learning_rate": 2.595989713563014e-06, + "loss": 0.6998, + "step": 7114 + }, + { + "epoch": 3.8626492942453856, + "grad_norm": 8.79473787534307, + "learning_rate": 2.593626273727765e-06, + "loss": 0.2626, + "step": 7115 + }, + { + "epoch": 3.8631921824104234, + "grad_norm": 10.941535841138563, + "learning_rate": 2.5912637499246218e-06, + "loss": 0.4595, + "step": 7116 + }, + { + "epoch": 3.8637350705754616, + "grad_norm": 12.201214800033737, + "learning_rate": 2.588902142445786e-06, + "loss": 0.3282, + "step": 7117 + }, + { + "epoch": 3.8642779587404994, + "grad_norm": 12.348184440426978, + "learning_rate": 2.5865414515833455e-06, + "loss": 0.5828, + "step": 7118 + }, + { + "epoch": 3.8648208469055376, + "grad_norm": 13.160180910276585, + "learning_rate": 2.5841816776292727e-06, + "loss": 0.3516, + "step": 7119 + }, + { + "epoch": 3.8653637350705754, + "grad_norm": 19.57808856867751, + "learning_rate": 2.581822820875429e-06, + "loss": 0.8939, + "step": 7120 + }, + { + "epoch": 3.8659066232356136, + "grad_norm": 10.92966233467966, + "learning_rate": 2.5794648816135627e-06, + "loss": 0.3509, + "step": 7121 + }, + { + "epoch": 3.8664495114006514, + "grad_norm": 14.189811177607414, + "learning_rate": 2.577107860135305e-06, + "loss": 0.5024, + "step": 7122 + }, + { + "epoch": 3.8669923995656896, + "grad_norm": 10.963014195472391, + "learning_rate": 2.5747517567321776e-06, + "loss": 0.3253, + "step": 7123 + }, + { + "epoch": 3.8675352877307274, + "grad_norm": 16.06753478220038, + "learning_rate": 2.5723965716955878e-06, + "loss": 0.5281, + "step": 7124 + }, + { + "epoch": 3.8680781758957656, + "grad_norm": 10.96537764415174, + "learning_rate": 2.5700423053168276e-06, + "loss": 0.5059, + "step": 7125 + }, + { + "epoch": 3.8686210640608034, + "grad_norm": 9.185810250309261, + "learning_rate": 2.567688957887077e-06, + "loss": 0.3768, + "step": 7126 + }, + { + "epoch": 3.8691639522258416, + "grad_norm": 13.984895212503027, + "learning_rate": 2.565336529697401e-06, + "loss": 0.5294, + "step": 7127 + }, + { + "epoch": 3.8697068403908794, + "grad_norm": 14.97489290708819, + "learning_rate": 2.562985021038752e-06, + "loss": 0.6177, + "step": 7128 + }, + { + "epoch": 3.8702497285559176, + "grad_norm": 15.68383969722191, + "learning_rate": 2.560634432201966e-06, + "loss": 0.4874, + "step": 7129 + }, + { + "epoch": 3.8707926167209554, + "grad_norm": 13.06746691168529, + "learning_rate": 2.5582847634777753e-06, + "loss": 0.771, + "step": 7130 + }, + { + "epoch": 3.8713355048859937, + "grad_norm": 11.661980178584857, + "learning_rate": 2.55593601515678e-06, + "loss": 0.4957, + "step": 7131 + }, + { + "epoch": 3.8718783930510314, + "grad_norm": 9.687387371930855, + "learning_rate": 2.5535881875294875e-06, + "loss": 0.238, + "step": 7132 + }, + { + "epoch": 3.8724212812160697, + "grad_norm": 12.53249739465081, + "learning_rate": 2.551241280886274e-06, + "loss": 0.3521, + "step": 7133 + }, + { + "epoch": 3.8729641693811074, + "grad_norm": 18.057246108983335, + "learning_rate": 2.548895295517406e-06, + "loss": 0.9307, + "step": 7134 + }, + { + "epoch": 3.8735070575461457, + "grad_norm": 15.513117876289204, + "learning_rate": 2.546550231713051e-06, + "loss": 0.6529, + "step": 7135 + }, + { + "epoch": 3.8740499457111834, + "grad_norm": 11.883743610451969, + "learning_rate": 2.544206089763235e-06, + "loss": 0.7861, + "step": 7136 + }, + { + "epoch": 3.8745928338762217, + "grad_norm": 14.684059381744154, + "learning_rate": 2.5418628699578986e-06, + "loss": 0.412, + "step": 7137 + }, + { + "epoch": 3.8751357220412594, + "grad_norm": 7.135402796193207, + "learning_rate": 2.539520572586843e-06, + "loss": 0.2079, + "step": 7138 + }, + { + "epoch": 3.8756786102062977, + "grad_norm": 10.289771258088079, + "learning_rate": 2.5371791979397766e-06, + "loss": 0.3274, + "step": 7139 + }, + { + "epoch": 3.8762214983713354, + "grad_norm": 9.949109620860526, + "learning_rate": 2.5348387463062808e-06, + "loss": 0.4355, + "step": 7140 + }, + { + "epoch": 3.8767643865363732, + "grad_norm": 10.991037946192776, + "learning_rate": 2.5324992179758268e-06, + "loss": 0.4432, + "step": 7141 + }, + { + "epoch": 3.8773072747014115, + "grad_norm": 15.18957067683597, + "learning_rate": 2.5301606132377733e-06, + "loss": 0.8666, + "step": 7142 + }, + { + "epoch": 3.8778501628664497, + "grad_norm": 14.818504355332566, + "learning_rate": 2.5278229323813553e-06, + "loss": 0.6253, + "step": 7143 + }, + { + "epoch": 3.8783930510314875, + "grad_norm": 13.754502247150116, + "learning_rate": 2.5254861756957115e-06, + "loss": 0.5842, + "step": 7144 + }, + { + "epoch": 3.8789359391965252, + "grad_norm": 19.3020908869945, + "learning_rate": 2.5231503434698435e-06, + "loss": 0.6372, + "step": 7145 + }, + { + "epoch": 3.8794788273615635, + "grad_norm": 8.868163440639846, + "learning_rate": 2.5208154359926606e-06, + "loss": 0.4229, + "step": 7146 + }, + { + "epoch": 3.8800217155266017, + "grad_norm": 10.130535143618188, + "learning_rate": 2.5184814535529457e-06, + "loss": 0.3024, + "step": 7147 + }, + { + "epoch": 3.8805646036916395, + "grad_norm": 10.20257149827293, + "learning_rate": 2.5161483964393676e-06, + "loss": 0.3697, + "step": 7148 + }, + { + "epoch": 3.8811074918566772, + "grad_norm": 17.03033594136655, + "learning_rate": 2.513816264940483e-06, + "loss": 0.6084, + "step": 7149 + }, + { + "epoch": 3.8816503800217155, + "grad_norm": 9.767302667294265, + "learning_rate": 2.5114850593447336e-06, + "loss": 0.2872, + "step": 7150 + }, + { + "epoch": 3.8821932681867537, + "grad_norm": 11.550149537445538, + "learning_rate": 2.509154779940447e-06, + "loss": 0.4796, + "step": 7151 + }, + { + "epoch": 3.8827361563517915, + "grad_norm": 13.607872338694929, + "learning_rate": 2.5068254270158364e-06, + "loss": 0.6471, + "step": 7152 + }, + { + "epoch": 3.8832790445168293, + "grad_norm": 13.786046110369192, + "learning_rate": 2.504497000859e-06, + "loss": 0.6666, + "step": 7153 + }, + { + "epoch": 3.8838219326818675, + "grad_norm": 11.281665232889665, + "learning_rate": 2.5021695017579193e-06, + "loss": 0.5587, + "step": 7154 + }, + { + "epoch": 3.8843648208469057, + "grad_norm": 11.481761454219619, + "learning_rate": 2.4998429300004657e-06, + "loss": 0.3071, + "step": 7155 + }, + { + "epoch": 3.8849077090119435, + "grad_norm": 13.28075154587505, + "learning_rate": 2.4975172858743914e-06, + "loss": 0.3665, + "step": 7156 + }, + { + "epoch": 3.8854505971769813, + "grad_norm": 9.802453085562453, + "learning_rate": 2.4951925696673352e-06, + "loss": 0.3148, + "step": 7157 + }, + { + "epoch": 3.8859934853420195, + "grad_norm": 14.520688569604944, + "learning_rate": 2.492868781666824e-06, + "loss": 0.4942, + "step": 7158 + }, + { + "epoch": 3.8865363735070577, + "grad_norm": 10.711275878066797, + "learning_rate": 2.4905459221602667e-06, + "loss": 0.3353, + "step": 7159 + }, + { + "epoch": 3.8870792616720955, + "grad_norm": 12.549002018795433, + "learning_rate": 2.488223991434955e-06, + "loss": 0.3241, + "step": 7160 + }, + { + "epoch": 3.8876221498371333, + "grad_norm": 13.961472675915795, + "learning_rate": 2.485902989778077e-06, + "loss": 0.5143, + "step": 7161 + }, + { + "epoch": 3.8881650380021715, + "grad_norm": 10.194253119170678, + "learning_rate": 2.483582917476691e-06, + "loss": 0.3732, + "step": 7162 + }, + { + "epoch": 3.8887079261672097, + "grad_norm": 11.408277950635817, + "learning_rate": 2.481263774817748e-06, + "loss": 0.5475, + "step": 7163 + }, + { + "epoch": 3.8892508143322475, + "grad_norm": 11.720624145965107, + "learning_rate": 2.4789455620880855e-06, + "loss": 0.4152, + "step": 7164 + }, + { + "epoch": 3.8897937024972853, + "grad_norm": 17.498990068476704, + "learning_rate": 2.4766282795744225e-06, + "loss": 0.4959, + "step": 7165 + }, + { + "epoch": 3.8903365906623235, + "grad_norm": 10.025081073758155, + "learning_rate": 2.474311927563364e-06, + "loss": 0.2563, + "step": 7166 + }, + { + "epoch": 3.8908794788273617, + "grad_norm": 11.064975553515668, + "learning_rate": 2.4719965063413975e-06, + "loss": 0.4892, + "step": 7167 + }, + { + "epoch": 3.8914223669923995, + "grad_norm": 11.4169177186445, + "learning_rate": 2.4696820161949076e-06, + "loss": 0.4661, + "step": 7168 + }, + { + "epoch": 3.8919652551574373, + "grad_norm": 11.127726171001362, + "learning_rate": 2.4673684574101407e-06, + "loss": 0.5133, + "step": 7169 + }, + { + "epoch": 3.8925081433224755, + "grad_norm": 14.4532269361977, + "learning_rate": 2.4650558302732554e-06, + "loss": 0.6211, + "step": 7170 + }, + { + "epoch": 3.8930510314875137, + "grad_norm": 11.501842216375938, + "learning_rate": 2.4627441350702697e-06, + "loss": 0.43, + "step": 7171 + }, + { + "epoch": 3.8935939196525515, + "grad_norm": 11.827951767183402, + "learning_rate": 2.460433372087099e-06, + "loss": 0.4745, + "step": 7172 + }, + { + "epoch": 3.8941368078175893, + "grad_norm": 16.63470564409548, + "learning_rate": 2.4581235416095516e-06, + "loss": 0.7871, + "step": 7173 + }, + { + "epoch": 3.8946796959826275, + "grad_norm": 13.957180122419292, + "learning_rate": 2.455814643923298e-06, + "loss": 0.435, + "step": 7174 + }, + { + "epoch": 3.8952225841476658, + "grad_norm": 14.844456130947902, + "learning_rate": 2.4535066793139194e-06, + "loss": 0.3498, + "step": 7175 + }, + { + "epoch": 3.8957654723127035, + "grad_norm": 13.063011297960943, + "learning_rate": 2.4511996480668554e-06, + "loss": 0.4807, + "step": 7176 + }, + { + "epoch": 3.8963083604777413, + "grad_norm": 11.343167538907203, + "learning_rate": 2.448893550467456e-06, + "loss": 0.6566, + "step": 7177 + }, + { + "epoch": 3.8968512486427795, + "grad_norm": 21.134566936836904, + "learning_rate": 2.4465883868009323e-06, + "loss": 0.7527, + "step": 7178 + }, + { + "epoch": 3.8973941368078178, + "grad_norm": 12.658548113922626, + "learning_rate": 2.4442841573523967e-06, + "loss": 0.388, + "step": 7179 + }, + { + "epoch": 3.8979370249728555, + "grad_norm": 10.941499945328948, + "learning_rate": 2.4419808624068396e-06, + "loss": 0.585, + "step": 7180 + }, + { + "epoch": 3.8984799131378933, + "grad_norm": 14.25189382311016, + "learning_rate": 2.4396785022491343e-06, + "loss": 0.6702, + "step": 7181 + }, + { + "epoch": 3.8990228013029316, + "grad_norm": 16.90343774367846, + "learning_rate": 2.4373770771640448e-06, + "loss": 0.5498, + "step": 7182 + }, + { + "epoch": 3.8995656894679698, + "grad_norm": 14.478488557861258, + "learning_rate": 2.4350765874362047e-06, + "loss": 0.6703, + "step": 7183 + }, + { + "epoch": 3.9001085776330076, + "grad_norm": 11.280629928495003, + "learning_rate": 2.4327770333501522e-06, + "loss": 0.387, + "step": 7184 + }, + { + "epoch": 3.9006514657980453, + "grad_norm": 10.389018994464337, + "learning_rate": 2.430478415190297e-06, + "loss": 0.4396, + "step": 7185 + }, + { + "epoch": 3.9011943539630836, + "grad_norm": 12.229139584290909, + "learning_rate": 2.4281807332409358e-06, + "loss": 0.639, + "step": 7186 + }, + { + "epoch": 3.901737242128122, + "grad_norm": 18.238591257772253, + "learning_rate": 2.425883987786248e-06, + "loss": 0.6162, + "step": 7187 + }, + { + "epoch": 3.9022801302931596, + "grad_norm": 14.326334376777995, + "learning_rate": 2.423588179110301e-06, + "loss": 0.8227, + "step": 7188 + }, + { + "epoch": 3.9028230184581973, + "grad_norm": 14.43500891639456, + "learning_rate": 2.4212933074970423e-06, + "loss": 0.5057, + "step": 7189 + }, + { + "epoch": 3.9033659066232356, + "grad_norm": 14.30031793020835, + "learning_rate": 2.4189993732303063e-06, + "loss": 0.4514, + "step": 7190 + }, + { + "epoch": 3.903908794788274, + "grad_norm": 9.814097341569619, + "learning_rate": 2.4167063765938103e-06, + "loss": 0.3212, + "step": 7191 + }, + { + "epoch": 3.9044516829533116, + "grad_norm": 13.100322061273497, + "learning_rate": 2.4144143178711555e-06, + "loss": 0.5785, + "step": 7192 + }, + { + "epoch": 3.9049945711183494, + "grad_norm": 15.547319070400764, + "learning_rate": 2.412123197345827e-06, + "loss": 0.6, + "step": 7193 + }, + { + "epoch": 3.9055374592833876, + "grad_norm": 10.568502118830102, + "learning_rate": 2.409833015301195e-06, + "loss": 0.4499, + "step": 7194 + }, + { + "epoch": 3.906080347448426, + "grad_norm": 12.216551577948717, + "learning_rate": 2.4075437720205132e-06, + "loss": 0.3987, + "step": 7195 + }, + { + "epoch": 3.9066232356134636, + "grad_norm": 14.203010493004756, + "learning_rate": 2.4052554677869165e-06, + "loss": 0.7274, + "step": 7196 + }, + { + "epoch": 3.9071661237785014, + "grad_norm": 16.842358223683046, + "learning_rate": 2.4029681028834293e-06, + "loss": 0.6862, + "step": 7197 + }, + { + "epoch": 3.9077090119435396, + "grad_norm": 17.087528448096148, + "learning_rate": 2.4006816775929553e-06, + "loss": 0.6367, + "step": 7198 + }, + { + "epoch": 3.908251900108578, + "grad_norm": 13.559074748741233, + "learning_rate": 2.3983961921982823e-06, + "loss": 0.497, + "step": 7199 + }, + { + "epoch": 3.9087947882736156, + "grad_norm": 11.82349002805914, + "learning_rate": 2.3961116469820834e-06, + "loss": 0.3808, + "step": 7200 + }, + { + "epoch": 3.9093376764386534, + "grad_norm": 9.536111713709516, + "learning_rate": 2.3938280422269143e-06, + "loss": 0.2577, + "step": 7201 + }, + { + "epoch": 3.9098805646036916, + "grad_norm": 13.40690087309448, + "learning_rate": 2.3915453782152166e-06, + "loss": 0.5696, + "step": 7202 + }, + { + "epoch": 3.91042345276873, + "grad_norm": 11.767180310470792, + "learning_rate": 2.3892636552293114e-06, + "loss": 0.447, + "step": 7203 + }, + { + "epoch": 3.9109663409337676, + "grad_norm": 15.579650605609013, + "learning_rate": 2.3869828735514076e-06, + "loss": 0.5241, + "step": 7204 + }, + { + "epoch": 3.9115092290988054, + "grad_norm": 10.113811350963385, + "learning_rate": 2.3847030334635923e-06, + "loss": 0.3542, + "step": 7205 + }, + { + "epoch": 3.9120521172638436, + "grad_norm": 10.589793992165902, + "learning_rate": 2.3824241352478484e-06, + "loss": 0.2795, + "step": 7206 + }, + { + "epoch": 3.912595005428882, + "grad_norm": 7.9421075555522505, + "learning_rate": 2.3801461791860226e-06, + "loss": 0.2936, + "step": 7207 + }, + { + "epoch": 3.9131378935939196, + "grad_norm": 9.535162659306835, + "learning_rate": 2.377869165559867e-06, + "loss": 0.5889, + "step": 7208 + }, + { + "epoch": 3.9136807817589574, + "grad_norm": 11.900400034485365, + "learning_rate": 2.375593094650995e-06, + "loss": 0.5422, + "step": 7209 + }, + { + "epoch": 3.9142236699239956, + "grad_norm": 11.395565814320083, + "learning_rate": 2.3733179667409247e-06, + "loss": 0.4194, + "step": 7210 + }, + { + "epoch": 3.914766558089034, + "grad_norm": 13.14302661684852, + "learning_rate": 2.3710437821110456e-06, + "loss": 0.5611, + "step": 7211 + }, + { + "epoch": 3.9153094462540716, + "grad_norm": 14.854169528681664, + "learning_rate": 2.3687705410426242e-06, + "loss": 0.676, + "step": 7212 + }, + { + "epoch": 3.9158523344191094, + "grad_norm": 14.377106944895534, + "learning_rate": 2.3664982438168305e-06, + "loss": 0.5029, + "step": 7213 + }, + { + "epoch": 3.9163952225841476, + "grad_norm": 16.2949751889776, + "learning_rate": 2.364226890714694e-06, + "loss": 0.4424, + "step": 7214 + }, + { + "epoch": 3.916938110749186, + "grad_norm": 12.453892587793705, + "learning_rate": 2.3619564820171515e-06, + "loss": 0.3498, + "step": 7215 + }, + { + "epoch": 3.9174809989142236, + "grad_norm": 10.510207014160567, + "learning_rate": 2.359687018004998e-06, + "loss": 0.4104, + "step": 7216 + }, + { + "epoch": 3.9180238870792614, + "grad_norm": 10.817797408933902, + "learning_rate": 2.3574184989589345e-06, + "loss": 0.466, + "step": 7217 + }, + { + "epoch": 3.9185667752442996, + "grad_norm": 9.615419411562353, + "learning_rate": 2.3551509251595315e-06, + "loss": 0.3975, + "step": 7218 + }, + { + "epoch": 3.919109663409338, + "grad_norm": 10.629184513933357, + "learning_rate": 2.3528842968872456e-06, + "loss": 0.3305, + "step": 7219 + }, + { + "epoch": 3.9196525515743756, + "grad_norm": 14.741934283339353, + "learning_rate": 2.350618614422421e-06, + "loss": 0.5822, + "step": 7220 + }, + { + "epoch": 3.9201954397394134, + "grad_norm": 10.145231872720867, + "learning_rate": 2.348353878045272e-06, + "loss": 0.3223, + "step": 7221 + }, + { + "epoch": 3.9207383279044516, + "grad_norm": 9.719525157659627, + "learning_rate": 2.346090088035913e-06, + "loss": 0.323, + "step": 7222 + }, + { + "epoch": 3.92128121606949, + "grad_norm": 11.153861466428655, + "learning_rate": 2.3438272446743293e-06, + "loss": 0.4453, + "step": 7223 + }, + { + "epoch": 3.9218241042345277, + "grad_norm": 16.881204557198902, + "learning_rate": 2.3415653482403954e-06, + "loss": 0.8422, + "step": 7224 + }, + { + "epoch": 3.9223669923995654, + "grad_norm": 13.409647968671297, + "learning_rate": 2.339304399013864e-06, + "loss": 0.5316, + "step": 7225 + }, + { + "epoch": 3.9229098805646037, + "grad_norm": 17.78528943034469, + "learning_rate": 2.337044397274375e-06, + "loss": 1.0815, + "step": 7226 + }, + { + "epoch": 3.923452768729642, + "grad_norm": 13.521330366302392, + "learning_rate": 2.3347853433014467e-06, + "loss": 0.6285, + "step": 7227 + }, + { + "epoch": 3.9239956568946797, + "grad_norm": 15.033911656714881, + "learning_rate": 2.3325272373744844e-06, + "loss": 0.4578, + "step": 7228 + }, + { + "epoch": 3.9245385450597174, + "grad_norm": 13.063924995112636, + "learning_rate": 2.330270079772774e-06, + "loss": 0.4246, + "step": 7229 + }, + { + "epoch": 3.9250814332247557, + "grad_norm": 11.872225399295557, + "learning_rate": 2.328013870775483e-06, + "loss": 0.354, + "step": 7230 + }, + { + "epoch": 3.925624321389794, + "grad_norm": 10.802228004866322, + "learning_rate": 2.325758610661664e-06, + "loss": 0.2926, + "step": 7231 + }, + { + "epoch": 3.9261672095548317, + "grad_norm": 10.24423290229698, + "learning_rate": 2.323504299710251e-06, + "loss": 0.3887, + "step": 7232 + }, + { + "epoch": 3.9267100977198695, + "grad_norm": 10.40790875242795, + "learning_rate": 2.32125093820006e-06, + "loss": 0.3774, + "step": 7233 + }, + { + "epoch": 3.9272529858849077, + "grad_norm": 10.819416071470908, + "learning_rate": 2.3189985264097925e-06, + "loss": 0.4433, + "step": 7234 + }, + { + "epoch": 3.927795874049946, + "grad_norm": 14.400478824494995, + "learning_rate": 2.316747064618028e-06, + "loss": 0.6139, + "step": 7235 + }, + { + "epoch": 3.9283387622149837, + "grad_norm": 13.400576026026663, + "learning_rate": 2.3144965531032314e-06, + "loss": 0.6807, + "step": 7236 + }, + { + "epoch": 3.9288816503800215, + "grad_norm": 12.144971157259214, + "learning_rate": 2.3122469921437507e-06, + "loss": 0.406, + "step": 7237 + }, + { + "epoch": 3.9294245385450597, + "grad_norm": 13.566955931406746, + "learning_rate": 2.3099983820178116e-06, + "loss": 0.6583, + "step": 7238 + }, + { + "epoch": 3.929967426710098, + "grad_norm": 13.453764479761338, + "learning_rate": 2.3077507230035345e-06, + "loss": 0.8706, + "step": 7239 + }, + { + "epoch": 3.9305103148751357, + "grad_norm": 15.149717368244767, + "learning_rate": 2.305504015378904e-06, + "loss": 0.6808, + "step": 7240 + }, + { + "epoch": 3.9310532030401735, + "grad_norm": 15.68496135669598, + "learning_rate": 2.303258259421801e-06, + "loss": 0.4163, + "step": 7241 + }, + { + "epoch": 3.9315960912052117, + "grad_norm": 18.73387442818072, + "learning_rate": 2.301013455409983e-06, + "loss": 0.972, + "step": 7242 + }, + { + "epoch": 3.93213897937025, + "grad_norm": 11.659704005587447, + "learning_rate": 2.298769603621088e-06, + "loss": 0.4766, + "step": 7243 + }, + { + "epoch": 3.9326818675352877, + "grad_norm": 12.907778650132892, + "learning_rate": 2.296526704332648e-06, + "loss": 0.5669, + "step": 7244 + }, + { + "epoch": 3.9332247557003255, + "grad_norm": 14.095811490299557, + "learning_rate": 2.294284757822057e-06, + "loss": 0.6699, + "step": 7245 + }, + { + "epoch": 3.9337676438653637, + "grad_norm": 14.347126906125892, + "learning_rate": 2.292043764366615e-06, + "loss": 0.5478, + "step": 7246 + }, + { + "epoch": 3.934310532030402, + "grad_norm": 14.15574104757881, + "learning_rate": 2.289803724243478e-06, + "loss": 0.3451, + "step": 7247 + }, + { + "epoch": 3.9348534201954397, + "grad_norm": 16.273077860218656, + "learning_rate": 2.28756463772971e-06, + "loss": 0.433, + "step": 7248 + }, + { + "epoch": 3.9353963083604775, + "grad_norm": 12.905791138340465, + "learning_rate": 2.2853265051022376e-06, + "loss": 0.3549, + "step": 7249 + }, + { + "epoch": 3.9359391965255157, + "grad_norm": 10.740402790195867, + "learning_rate": 2.283089326637875e-06, + "loss": 0.294, + "step": 7250 + }, + { + "epoch": 3.936482084690554, + "grad_norm": 16.206654981483148, + "learning_rate": 2.2808531026133297e-06, + "loss": 0.4773, + "step": 7251 + }, + { + "epoch": 3.9370249728555917, + "grad_norm": 13.723535114050927, + "learning_rate": 2.278617833305169e-06, + "loss": 0.5248, + "step": 7252 + }, + { + "epoch": 3.9375678610206295, + "grad_norm": 12.638788461264767, + "learning_rate": 2.2763835189898665e-06, + "loss": 0.6087, + "step": 7253 + }, + { + "epoch": 3.9381107491856677, + "grad_norm": 12.333038982803954, + "learning_rate": 2.2741501599437543e-06, + "loss": 0.3829, + "step": 7254 + }, + { + "epoch": 3.938653637350706, + "grad_norm": 11.07699141293442, + "learning_rate": 2.2719177564430662e-06, + "loss": 0.6625, + "step": 7255 + }, + { + "epoch": 3.9391965255157437, + "grad_norm": 14.354975596263804, + "learning_rate": 2.2696863087639063e-06, + "loss": 0.6906, + "step": 7256 + }, + { + "epoch": 3.9397394136807815, + "grad_norm": 10.159046635813173, + "learning_rate": 2.2674558171822646e-06, + "loss": 0.4371, + "step": 7257 + }, + { + "epoch": 3.9402823018458197, + "grad_norm": 18.017466534966523, + "learning_rate": 2.265226281974011e-06, + "loss": 0.5347, + "step": 7258 + }, + { + "epoch": 3.940825190010858, + "grad_norm": 13.962772198264076, + "learning_rate": 2.2629977034148988e-06, + "loss": 0.4712, + "step": 7259 + }, + { + "epoch": 3.9413680781758957, + "grad_norm": 10.058544580261096, + "learning_rate": 2.2607700817805635e-06, + "loss": 0.3168, + "step": 7260 + }, + { + "epoch": 3.9419109663409335, + "grad_norm": 18.61661644453424, + "learning_rate": 2.258543417346514e-06, + "loss": 0.5639, + "step": 7261 + }, + { + "epoch": 3.9424538545059717, + "grad_norm": 12.38740485358688, + "learning_rate": 2.256317710388155e-06, + "loss": 0.4182, + "step": 7262 + }, + { + "epoch": 3.94299674267101, + "grad_norm": 12.633562464203537, + "learning_rate": 2.254092961180764e-06, + "loss": 0.5299, + "step": 7263 + }, + { + "epoch": 3.9435396308360477, + "grad_norm": 15.51128932246205, + "learning_rate": 2.2518691699995e-06, + "loss": 0.6563, + "step": 7264 + }, + { + "epoch": 3.9440825190010855, + "grad_norm": 14.124949663680717, + "learning_rate": 2.2496463371194065e-06, + "loss": 0.7836, + "step": 7265 + }, + { + "epoch": 3.9446254071661238, + "grad_norm": 14.035611568552724, + "learning_rate": 2.247424462815405e-06, + "loss": 0.4113, + "step": 7266 + }, + { + "epoch": 3.945168295331162, + "grad_norm": 13.30563635940546, + "learning_rate": 2.2452035473623022e-06, + "loss": 0.449, + "step": 7267 + }, + { + "epoch": 3.9457111834961998, + "grad_norm": 12.093768184800798, + "learning_rate": 2.242983591034784e-06, + "loss": 0.5745, + "step": 7268 + }, + { + "epoch": 3.9462540716612375, + "grad_norm": 13.19161002685443, + "learning_rate": 2.2407645941074185e-06, + "loss": 0.6073, + "step": 7269 + }, + { + "epoch": 3.9467969598262758, + "grad_norm": 10.454523102469183, + "learning_rate": 2.238546556854655e-06, + "loss": 0.265, + "step": 7270 + }, + { + "epoch": 3.947339847991314, + "grad_norm": 12.97022293790845, + "learning_rate": 2.2363294795508217e-06, + "loss": 0.5445, + "step": 7271 + }, + { + "epoch": 3.9478827361563518, + "grad_norm": 12.83418150037974, + "learning_rate": 2.234113362470133e-06, + "loss": 0.4941, + "step": 7272 + }, + { + "epoch": 3.9484256243213895, + "grad_norm": 12.615689050586889, + "learning_rate": 2.23189820588668e-06, + "loss": 0.8588, + "step": 7273 + }, + { + "epoch": 3.9489685124864278, + "grad_norm": 14.878456274053418, + "learning_rate": 2.2296840100744375e-06, + "loss": 0.4019, + "step": 7274 + }, + { + "epoch": 3.949511400651466, + "grad_norm": 13.017249525734119, + "learning_rate": 2.227470775307261e-06, + "loss": 0.5343, + "step": 7275 + }, + { + "epoch": 3.950054288816504, + "grad_norm": 11.575062166378828, + "learning_rate": 2.2252585018588836e-06, + "loss": 0.3775, + "step": 7276 + }, + { + "epoch": 3.9505971769815416, + "grad_norm": 15.290489125989911, + "learning_rate": 2.2230471900029303e-06, + "loss": 0.4912, + "step": 7277 + }, + { + "epoch": 3.95114006514658, + "grad_norm": 17.764334443185927, + "learning_rate": 2.220836840012891e-06, + "loss": 0.7215, + "step": 7278 + }, + { + "epoch": 3.951682953311618, + "grad_norm": 14.154407180706324, + "learning_rate": 2.218627452162154e-06, + "loss": 0.6472, + "step": 7279 + }, + { + "epoch": 3.952225841476656, + "grad_norm": 11.225658955685093, + "learning_rate": 2.2164190267239737e-06, + "loss": 0.2478, + "step": 7280 + }, + { + "epoch": 3.9527687296416936, + "grad_norm": 12.814332016509574, + "learning_rate": 2.2142115639714935e-06, + "loss": 0.5622, + "step": 7281 + }, + { + "epoch": 3.953311617806732, + "grad_norm": 19.419294319476467, + "learning_rate": 2.2120050641777345e-06, + "loss": 0.7996, + "step": 7282 + }, + { + "epoch": 3.95385450597177, + "grad_norm": 19.022041488664335, + "learning_rate": 2.209799527615599e-06, + "loss": 0.7117, + "step": 7283 + }, + { + "epoch": 3.954397394136808, + "grad_norm": 12.40634141862146, + "learning_rate": 2.20759495455788e-06, + "loss": 0.3793, + "step": 7284 + }, + { + "epoch": 3.9549402823018456, + "grad_norm": 8.873847512885336, + "learning_rate": 2.20539134527723e-06, + "loss": 0.3723, + "step": 7285 + }, + { + "epoch": 3.955483170466884, + "grad_norm": 15.833917455665446, + "learning_rate": 2.203188700046207e-06, + "loss": 0.6743, + "step": 7286 + }, + { + "epoch": 3.956026058631922, + "grad_norm": 12.353488572936646, + "learning_rate": 2.2009870191372263e-06, + "loss": 0.4195, + "step": 7287 + }, + { + "epoch": 3.95656894679696, + "grad_norm": 19.03028178846917, + "learning_rate": 2.198786302822603e-06, + "loss": 0.9055, + "step": 7288 + }, + { + "epoch": 3.9571118349619976, + "grad_norm": 15.81794212526427, + "learning_rate": 2.1965865513745265e-06, + "loss": 0.4088, + "step": 7289 + }, + { + "epoch": 3.957654723127036, + "grad_norm": 14.909566500608973, + "learning_rate": 2.1943877650650556e-06, + "loss": 0.9969, + "step": 7290 + }, + { + "epoch": 3.958197611292074, + "grad_norm": 12.332907917932245, + "learning_rate": 2.192189944166153e-06, + "loss": 0.3762, + "step": 7291 + }, + { + "epoch": 3.958740499457112, + "grad_norm": 13.036980814291068, + "learning_rate": 2.189993088949636e-06, + "loss": 0.4931, + "step": 7292 + }, + { + "epoch": 3.9592833876221496, + "grad_norm": 15.24047006419779, + "learning_rate": 2.187797199687224e-06, + "loss": 0.5143, + "step": 7293 + }, + { + "epoch": 3.959826275787188, + "grad_norm": 12.214872928235861, + "learning_rate": 2.185602276650505e-06, + "loss": 0.4249, + "step": 7294 + }, + { + "epoch": 3.960369163952226, + "grad_norm": 11.734417487796039, + "learning_rate": 2.18340832011095e-06, + "loss": 0.5159, + "step": 7295 + }, + { + "epoch": 3.960912052117264, + "grad_norm": 16.316727205367513, + "learning_rate": 2.181215330339912e-06, + "loss": 0.5141, + "step": 7296 + }, + { + "epoch": 3.9614549402823016, + "grad_norm": 13.896917482997276, + "learning_rate": 2.1790233076086243e-06, + "loss": 1.0396, + "step": 7297 + }, + { + "epoch": 3.96199782844734, + "grad_norm": 17.3480021128049, + "learning_rate": 2.1768322521882003e-06, + "loss": 0.43, + "step": 7298 + }, + { + "epoch": 3.962540716612378, + "grad_norm": 11.942778635499716, + "learning_rate": 2.1746421643496264e-06, + "loss": 0.383, + "step": 7299 + }, + { + "epoch": 3.963083604777416, + "grad_norm": 10.46804392176075, + "learning_rate": 2.1724530443637836e-06, + "loss": 0.6297, + "step": 7300 + }, + { + "epoch": 3.9636264929424536, + "grad_norm": 12.678863612803186, + "learning_rate": 2.1702648925014248e-06, + "loss": 0.4674, + "step": 7301 + }, + { + "epoch": 3.964169381107492, + "grad_norm": 7.353045500973433, + "learning_rate": 2.1680777090331816e-06, + "loss": 0.3103, + "step": 7302 + }, + { + "epoch": 3.96471226927253, + "grad_norm": 12.29878912174572, + "learning_rate": 2.1658914942295706e-06, + "loss": 0.5663, + "step": 7303 + }, + { + "epoch": 3.965255157437568, + "grad_norm": 8.497343343626396, + "learning_rate": 2.163706248360985e-06, + "loss": 0.3524, + "step": 7304 + }, + { + "epoch": 3.9657980456026056, + "grad_norm": 11.764041467674721, + "learning_rate": 2.1615219716977e-06, + "loss": 0.4548, + "step": 7305 + }, + { + "epoch": 3.966340933767644, + "grad_norm": 17.313346132201072, + "learning_rate": 2.1593386645098692e-06, + "loss": 1.0529, + "step": 7306 + }, + { + "epoch": 3.966883821932682, + "grad_norm": 13.90349930971142, + "learning_rate": 2.15715632706753e-06, + "loss": 0.5808, + "step": 7307 + }, + { + "epoch": 3.96742671009772, + "grad_norm": 13.776936265783004, + "learning_rate": 2.1549749596405945e-06, + "loss": 0.6697, + "step": 7308 + }, + { + "epoch": 3.9679695982627576, + "grad_norm": 16.504532446022996, + "learning_rate": 2.15279456249886e-06, + "loss": 0.5508, + "step": 7309 + }, + { + "epoch": 3.968512486427796, + "grad_norm": 9.948445138257954, + "learning_rate": 2.1506151359119997e-06, + "loss": 0.4461, + "step": 7310 + }, + { + "epoch": 3.969055374592834, + "grad_norm": 14.495691133016251, + "learning_rate": 2.1484366801495705e-06, + "loss": 0.7312, + "step": 7311 + }, + { + "epoch": 3.969598262757872, + "grad_norm": 16.82715961877643, + "learning_rate": 2.1462591954810054e-06, + "loss": 0.7524, + "step": 7312 + }, + { + "epoch": 3.9701411509229096, + "grad_norm": 11.225827985724983, + "learning_rate": 2.1440826821756213e-06, + "loss": 0.4354, + "step": 7313 + }, + { + "epoch": 3.970684039087948, + "grad_norm": 9.865289892982581, + "learning_rate": 2.141907140502607e-06, + "loss": 0.2906, + "step": 7314 + }, + { + "epoch": 3.971226927252986, + "grad_norm": 8.988273380587568, + "learning_rate": 2.139732570731048e-06, + "loss": 0.3062, + "step": 7315 + }, + { + "epoch": 3.971769815418024, + "grad_norm": 13.659816709804225, + "learning_rate": 2.1375589731298864e-06, + "loss": 0.382, + "step": 7316 + }, + { + "epoch": 3.9723127035830617, + "grad_norm": 15.628687568254344, + "learning_rate": 2.135386347967967e-06, + "loss": 0.6687, + "step": 7317 + }, + { + "epoch": 3.9728555917481, + "grad_norm": 10.910710203208446, + "learning_rate": 2.1332146955139963e-06, + "loss": 0.3909, + "step": 7318 + }, + { + "epoch": 3.973398479913138, + "grad_norm": 13.270954739017608, + "learning_rate": 2.13104401603657e-06, + "loss": 0.4808, + "step": 7319 + }, + { + "epoch": 3.973941368078176, + "grad_norm": 13.799228933349236, + "learning_rate": 2.12887430980416e-06, + "loss": 0.4717, + "step": 7320 + }, + { + "epoch": 3.9744842562432137, + "grad_norm": 12.928285689167119, + "learning_rate": 2.1267055770851185e-06, + "loss": 0.5013, + "step": 7321 + }, + { + "epoch": 3.975027144408252, + "grad_norm": 11.400185046713785, + "learning_rate": 2.124537818147684e-06, + "loss": 0.304, + "step": 7322 + }, + { + "epoch": 3.97557003257329, + "grad_norm": 10.537157565034327, + "learning_rate": 2.122371033259959e-06, + "loss": 0.3975, + "step": 7323 + }, + { + "epoch": 3.976112920738328, + "grad_norm": 14.40105822826117, + "learning_rate": 2.120205222689944e-06, + "loss": 0.6778, + "step": 7324 + }, + { + "epoch": 3.9766558089033657, + "grad_norm": 13.229824974968945, + "learning_rate": 2.118040386705501e-06, + "loss": 0.6712, + "step": 7325 + }, + { + "epoch": 3.977198697068404, + "grad_norm": 13.793638194835786, + "learning_rate": 2.1158765255743872e-06, + "loss": 0.4601, + "step": 7326 + }, + { + "epoch": 3.977741585233442, + "grad_norm": 15.505430296521286, + "learning_rate": 2.113713639564231e-06, + "loss": 0.8212, + "step": 7327 + }, + { + "epoch": 3.97828447339848, + "grad_norm": 17.09119757360545, + "learning_rate": 2.11155172894254e-06, + "loss": 0.7094, + "step": 7328 + }, + { + "epoch": 3.9788273615635177, + "grad_norm": 13.556531251232467, + "learning_rate": 2.1093907939767065e-06, + "loss": 0.4832, + "step": 7329 + }, + { + "epoch": 3.979370249728556, + "grad_norm": 14.237329918169115, + "learning_rate": 2.10723083493399e-06, + "loss": 0.3663, + "step": 7330 + }, + { + "epoch": 3.979913137893594, + "grad_norm": 9.012447122900646, + "learning_rate": 2.1050718520815485e-06, + "loss": 0.5217, + "step": 7331 + }, + { + "epoch": 3.980456026058632, + "grad_norm": 12.91832304111153, + "learning_rate": 2.1029138456863973e-06, + "loss": 0.7424, + "step": 7332 + }, + { + "epoch": 3.9809989142236697, + "grad_norm": 17.140779656879335, + "learning_rate": 2.1007568160154502e-06, + "loss": 0.5607, + "step": 7333 + }, + { + "epoch": 3.981541802388708, + "grad_norm": 15.535441443258492, + "learning_rate": 2.09860076333549e-06, + "loss": 0.6682, + "step": 7334 + }, + { + "epoch": 3.982084690553746, + "grad_norm": 9.108166871624157, + "learning_rate": 2.09644568791318e-06, + "loss": 0.3471, + "step": 7335 + }, + { + "epoch": 3.982627578718784, + "grad_norm": 13.8585496563886, + "learning_rate": 2.094291590015064e-06, + "loss": 0.4907, + "step": 7336 + }, + { + "epoch": 3.9831704668838217, + "grad_norm": 17.309058810158508, + "learning_rate": 2.0921384699075644e-06, + "loss": 0.7292, + "step": 7337 + }, + { + "epoch": 3.98371335504886, + "grad_norm": 11.888938470455745, + "learning_rate": 2.089986327856981e-06, + "loss": 0.5343, + "step": 7338 + }, + { + "epoch": 3.984256243213898, + "grad_norm": 14.96573909686316, + "learning_rate": 2.087835164129496e-06, + "loss": 0.436, + "step": 7339 + }, + { + "epoch": 3.984799131378936, + "grad_norm": 11.965090702213786, + "learning_rate": 2.085684978991168e-06, + "loss": 0.4307, + "step": 7340 + }, + { + "epoch": 3.9853420195439737, + "grad_norm": 9.089611131661139, + "learning_rate": 2.083535772707935e-06, + "loss": 0.3474, + "step": 7341 + }, + { + "epoch": 3.985884907709012, + "grad_norm": 15.457161419441805, + "learning_rate": 2.0813875455456156e-06, + "loss": 0.6491, + "step": 7342 + }, + { + "epoch": 3.98642779587405, + "grad_norm": 10.250596479107204, + "learning_rate": 2.0792402977699033e-06, + "loss": 0.3589, + "step": 7343 + }, + { + "epoch": 3.986970684039088, + "grad_norm": 14.949877353645148, + "learning_rate": 2.077094029646376e-06, + "loss": 0.8685, + "step": 7344 + }, + { + "epoch": 3.9875135722041257, + "grad_norm": 15.23923375936191, + "learning_rate": 2.074948741440486e-06, + "loss": 0.744, + "step": 7345 + }, + { + "epoch": 3.988056460369164, + "grad_norm": 11.133935275967207, + "learning_rate": 2.0728044334175667e-06, + "loss": 0.7823, + "step": 7346 + }, + { + "epoch": 3.988599348534202, + "grad_norm": 10.744562653679704, + "learning_rate": 2.0706611058428285e-06, + "loss": 0.3669, + "step": 7347 + }, + { + "epoch": 3.98914223669924, + "grad_norm": 12.814655704488615, + "learning_rate": 2.0685187589813625e-06, + "loss": 0.3812, + "step": 7348 + }, + { + "epoch": 3.9896851248642777, + "grad_norm": 10.282758906730834, + "learning_rate": 2.0663773930981367e-06, + "loss": 0.3229, + "step": 7349 + }, + { + "epoch": 3.990228013029316, + "grad_norm": 14.644174681230187, + "learning_rate": 2.064237008458e-06, + "loss": 0.7615, + "step": 7350 + }, + { + "epoch": 3.990770901194354, + "grad_norm": 16.98502298622554, + "learning_rate": 2.062097605325678e-06, + "loss": 0.5847, + "step": 7351 + }, + { + "epoch": 3.991313789359392, + "grad_norm": 12.815769655827422, + "learning_rate": 2.059959183965775e-06, + "loss": 0.3939, + "step": 7352 + }, + { + "epoch": 3.9918566775244297, + "grad_norm": 13.188644656309128, + "learning_rate": 2.057821744642774e-06, + "loss": 0.6187, + "step": 7353 + }, + { + "epoch": 3.992399565689468, + "grad_norm": 12.875041049936685, + "learning_rate": 2.0556852876210354e-06, + "loss": 0.3272, + "step": 7354 + }, + { + "epoch": 3.992942453854506, + "grad_norm": 13.754004827000545, + "learning_rate": 2.053549813164808e-06, + "loss": 0.4637, + "step": 7355 + }, + { + "epoch": 3.993485342019544, + "grad_norm": 10.58203405942224, + "learning_rate": 2.0514153215381983e-06, + "loss": 0.5784, + "step": 7356 + }, + { + "epoch": 3.9940282301845818, + "grad_norm": 10.96146456022653, + "learning_rate": 2.049281813005215e-06, + "loss": 0.4389, + "step": 7357 + }, + { + "epoch": 3.99457111834962, + "grad_norm": 12.017836415062291, + "learning_rate": 2.047149287829726e-06, + "loss": 0.5839, + "step": 7358 + }, + { + "epoch": 3.995114006514658, + "grad_norm": 10.822043166102832, + "learning_rate": 2.045017746275485e-06, + "loss": 0.4467, + "step": 7359 + }, + { + "epoch": 3.995656894679696, + "grad_norm": 13.151255799546034, + "learning_rate": 2.0428871886061343e-06, + "loss": 0.6173, + "step": 7360 + }, + { + "epoch": 3.9961997828447338, + "grad_norm": 11.118960416841484, + "learning_rate": 2.0407576150851705e-06, + "loss": 0.3935, + "step": 7361 + }, + { + "epoch": 3.996742671009772, + "grad_norm": 12.928271003217045, + "learning_rate": 2.0386290259759967e-06, + "loss": 0.5222, + "step": 7362 + }, + { + "epoch": 3.99728555917481, + "grad_norm": 14.237952289938509, + "learning_rate": 2.0365014215418676e-06, + "loss": 0.495, + "step": 7363 + }, + { + "epoch": 3.997828447339848, + "grad_norm": 15.550337043796231, + "learning_rate": 2.0343748020459374e-06, + "loss": 0.4867, + "step": 7364 + }, + { + "epoch": 3.9983713355048858, + "grad_norm": 12.364600226217249, + "learning_rate": 2.032249167751228e-06, + "loss": 0.6296, + "step": 7365 + }, + { + "epoch": 3.998914223669924, + "grad_norm": 13.569416929145026, + "learning_rate": 2.0301245189206385e-06, + "loss": 0.5857, + "step": 7366 + }, + { + "epoch": 3.999457111834962, + "grad_norm": 14.915873617390455, + "learning_rate": 2.028000855816954e-06, + "loss": 0.7152, + "step": 7367 + }, + { + "epoch": 4.0, + "grad_norm": 13.686794191538802, + "learning_rate": 2.025878178702825e-06, + "loss": 0.3451, + "step": 7368 + }, + { + "epoch": 4.000542888165038, + "grad_norm": 14.607866561626649, + "learning_rate": 2.0237564878407956e-06, + "loss": 0.5273, + "step": 7369 + }, + { + "epoch": 4.001085776330076, + "grad_norm": 10.130331812248723, + "learning_rate": 2.021635783493271e-06, + "loss": 0.2723, + "step": 7370 + }, + { + "epoch": 4.001628664495114, + "grad_norm": 10.032934671154406, + "learning_rate": 2.019516065922551e-06, + "loss": 0.3576, + "step": 7371 + }, + { + "epoch": 4.002171552660152, + "grad_norm": 11.657828066715073, + "learning_rate": 2.017397335390803e-06, + "loss": 0.3434, + "step": 7372 + }, + { + "epoch": 4.00271444082519, + "grad_norm": 13.248765523472253, + "learning_rate": 2.0152795921600745e-06, + "loss": 0.5575, + "step": 7373 + }, + { + "epoch": 4.003257328990228, + "grad_norm": 13.375690344309364, + "learning_rate": 2.013162836492293e-06, + "loss": 0.4909, + "step": 7374 + }, + { + "epoch": 4.003800217155266, + "grad_norm": 13.29236800364953, + "learning_rate": 2.01104706864926e-06, + "loss": 0.4056, + "step": 7375 + }, + { + "epoch": 4.004343105320304, + "grad_norm": 11.112608581516877, + "learning_rate": 2.0089322888926577e-06, + "loss": 0.3764, + "step": 7376 + }, + { + "epoch": 4.004885993485342, + "grad_norm": 13.400379725287804, + "learning_rate": 2.0068184974840464e-06, + "loss": 0.5315, + "step": 7377 + }, + { + "epoch": 4.00542888165038, + "grad_norm": 12.399245820402392, + "learning_rate": 2.004705694684863e-06, + "loss": 0.4755, + "step": 7378 + }, + { + "epoch": 4.005971769815418, + "grad_norm": 7.235424189620777, + "learning_rate": 2.0025938807564207e-06, + "loss": 0.2616, + "step": 7379 + }, + { + "epoch": 4.006514657980456, + "grad_norm": 14.18643915790045, + "learning_rate": 2.0004830559599143e-06, + "loss": 0.3706, + "step": 7380 + }, + { + "epoch": 4.007057546145494, + "grad_norm": 10.127056422809925, + "learning_rate": 1.998373220556412e-06, + "loss": 0.3258, + "step": 7381 + }, + { + "epoch": 4.007600434310532, + "grad_norm": 17.273359103095373, + "learning_rate": 1.9962643748068633e-06, + "loss": 0.6007, + "step": 7382 + }, + { + "epoch": 4.00814332247557, + "grad_norm": 10.953175473484771, + "learning_rate": 1.994156518972092e-06, + "loss": 0.4094, + "step": 7383 + }, + { + "epoch": 4.008686210640608, + "grad_norm": 10.927954583193763, + "learning_rate": 1.992049653312802e-06, + "loss": 0.4199, + "step": 7384 + }, + { + "epoch": 4.009229098805646, + "grad_norm": 13.711331521508846, + "learning_rate": 1.9899437780895716e-06, + "loss": 0.6639, + "step": 7385 + }, + { + "epoch": 4.009771986970684, + "grad_norm": 12.537819678821979, + "learning_rate": 1.9878388935628655e-06, + "loss": 0.3551, + "step": 7386 + }, + { + "epoch": 4.010314875135722, + "grad_norm": 11.448288732996893, + "learning_rate": 1.985734999993013e-06, + "loss": 0.4247, + "step": 7387 + }, + { + "epoch": 4.01085776330076, + "grad_norm": 17.00322545288514, + "learning_rate": 1.9836320976402267e-06, + "loss": 0.5907, + "step": 7388 + }, + { + "epoch": 4.011400651465798, + "grad_norm": 10.797125749762163, + "learning_rate": 1.981530186764601e-06, + "loss": 0.4138, + "step": 7389 + }, + { + "epoch": 4.011943539630836, + "grad_norm": 17.95998607345293, + "learning_rate": 1.9794292676260997e-06, + "loss": 0.6222, + "step": 7390 + }, + { + "epoch": 4.012486427795874, + "grad_norm": 14.411778567127596, + "learning_rate": 1.9773293404845696e-06, + "loss": 0.421, + "step": 7391 + }, + { + "epoch": 4.013029315960912, + "grad_norm": 10.2409856593537, + "learning_rate": 1.9752304055997305e-06, + "loss": 0.3125, + "step": 7392 + }, + { + "epoch": 4.01357220412595, + "grad_norm": 11.292522294180968, + "learning_rate": 1.97313246323119e-06, + "loss": 0.2754, + "step": 7393 + }, + { + "epoch": 4.014115092290988, + "grad_norm": 10.928696930486545, + "learning_rate": 1.971035513638414e-06, + "loss": 0.4488, + "step": 7394 + }, + { + "epoch": 4.014657980456026, + "grad_norm": 12.415811489917902, + "learning_rate": 1.968939557080767e-06, + "loss": 0.436, + "step": 7395 + }, + { + "epoch": 4.015200868621064, + "grad_norm": 11.13510857929724, + "learning_rate": 1.9668445938174717e-06, + "loss": 0.3495, + "step": 7396 + }, + { + "epoch": 4.015743756786102, + "grad_norm": 9.229951548837793, + "learning_rate": 1.9647506241076387e-06, + "loss": 0.2861, + "step": 7397 + }, + { + "epoch": 4.01628664495114, + "grad_norm": 16.847132987227848, + "learning_rate": 1.96265764821026e-06, + "loss": 0.5219, + "step": 7398 + }, + { + "epoch": 4.016829533116178, + "grad_norm": 11.84958702376863, + "learning_rate": 1.9605656663841867e-06, + "loss": 0.4128, + "step": 7399 + }, + { + "epoch": 4.017372421281216, + "grad_norm": 14.742001988102418, + "learning_rate": 1.95847467888817e-06, + "loss": 0.4796, + "step": 7400 + }, + { + "epoch": 4.017915309446254, + "grad_norm": 11.28536118564977, + "learning_rate": 1.956384685980818e-06, + "loss": 0.3599, + "step": 7401 + }, + { + "epoch": 4.018458197611292, + "grad_norm": 15.721826059355514, + "learning_rate": 1.954295687920631e-06, + "loss": 0.5619, + "step": 7402 + }, + { + "epoch": 4.01900108577633, + "grad_norm": 15.983663047461183, + "learning_rate": 1.952207684965971e-06, + "loss": 0.6122, + "step": 7403 + }, + { + "epoch": 4.019543973941368, + "grad_norm": 12.128334107658644, + "learning_rate": 1.9501206773750947e-06, + "loss": 0.4605, + "step": 7404 + }, + { + "epoch": 4.020086862106406, + "grad_norm": 11.36806204402688, + "learning_rate": 1.948034665406121e-06, + "loss": 0.3116, + "step": 7405 + }, + { + "epoch": 4.020629750271444, + "grad_norm": 10.68555301238784, + "learning_rate": 1.9459496493170536e-06, + "loss": 0.321, + "step": 7406 + }, + { + "epoch": 4.021172638436482, + "grad_norm": 9.879738893549789, + "learning_rate": 1.9438656293657733e-06, + "loss": 0.2991, + "step": 7407 + }, + { + "epoch": 4.02171552660152, + "grad_norm": 13.176679205393846, + "learning_rate": 1.9417826058100253e-06, + "loss": 0.6426, + "step": 7408 + }, + { + "epoch": 4.022258414766558, + "grad_norm": 9.5138082645493, + "learning_rate": 1.9397005789074497e-06, + "loss": 0.3611, + "step": 7409 + }, + { + "epoch": 4.022801302931596, + "grad_norm": 9.82645446405319, + "learning_rate": 1.9376195489155537e-06, + "loss": 0.2537, + "step": 7410 + }, + { + "epoch": 4.023344191096634, + "grad_norm": 12.161440619011307, + "learning_rate": 1.935539516091721e-06, + "loss": 0.3498, + "step": 7411 + }, + { + "epoch": 4.023887079261672, + "grad_norm": 14.534994929925375, + "learning_rate": 1.9334604806932143e-06, + "loss": 0.5933, + "step": 7412 + }, + { + "epoch": 4.02442996742671, + "grad_norm": 13.507360209479582, + "learning_rate": 1.931382442977171e-06, + "loss": 0.3563, + "step": 7413 + }, + { + "epoch": 4.024972855591748, + "grad_norm": 15.046753479367776, + "learning_rate": 1.9293054032006063e-06, + "loss": 0.4387, + "step": 7414 + }, + { + "epoch": 4.025515743756786, + "grad_norm": 12.258959062712908, + "learning_rate": 1.927229361620412e-06, + "loss": 0.4434, + "step": 7415 + }, + { + "epoch": 4.026058631921824, + "grad_norm": 12.015424398969936, + "learning_rate": 1.9251543184933576e-06, + "loss": 0.4404, + "step": 7416 + }, + { + "epoch": 4.026601520086862, + "grad_norm": 9.477634559865093, + "learning_rate": 1.9230802740760857e-06, + "loss": 0.3345, + "step": 7417 + }, + { + "epoch": 4.0271444082519, + "grad_norm": 14.279619141455536, + "learning_rate": 1.921007228625118e-06, + "loss": 0.3984, + "step": 7418 + }, + { + "epoch": 4.027687296416938, + "grad_norm": 21.075479101285925, + "learning_rate": 1.9189351823968518e-06, + "loss": 0.4166, + "step": 7419 + }, + { + "epoch": 4.028230184581976, + "grad_norm": 9.220949774925229, + "learning_rate": 1.9168641356475625e-06, + "loss": 0.2762, + "step": 7420 + }, + { + "epoch": 4.028773072747014, + "grad_norm": 14.44225601755817, + "learning_rate": 1.9147940886333994e-06, + "loss": 0.564, + "step": 7421 + }, + { + "epoch": 4.029315960912052, + "grad_norm": 9.06166197984856, + "learning_rate": 1.912725041610388e-06, + "loss": 0.3853, + "step": 7422 + }, + { + "epoch": 4.02985884907709, + "grad_norm": 13.498032050668492, + "learning_rate": 1.9106569948344344e-06, + "loss": 0.4768, + "step": 7423 + }, + { + "epoch": 4.030401737242128, + "grad_norm": 13.191503934950113, + "learning_rate": 1.908589948561316e-06, + "loss": 0.3802, + "step": 7424 + }, + { + "epoch": 4.030944625407166, + "grad_norm": 12.094983329355872, + "learning_rate": 1.9065239030466854e-06, + "loss": 0.3355, + "step": 7425 + }, + { + "epoch": 4.031487513572204, + "grad_norm": 10.332292929620861, + "learning_rate": 1.9044588585460821e-06, + "loss": 0.3172, + "step": 7426 + }, + { + "epoch": 4.032030401737242, + "grad_norm": 13.8560757790004, + "learning_rate": 1.9023948153149075e-06, + "loss": 0.3627, + "step": 7427 + }, + { + "epoch": 4.03257328990228, + "grad_norm": 10.601587706717948, + "learning_rate": 1.9003317736084481e-06, + "loss": 0.3288, + "step": 7428 + }, + { + "epoch": 4.033116178067318, + "grad_norm": 11.935383011132533, + "learning_rate": 1.8982697336818633e-06, + "loss": 0.5207, + "step": 7429 + }, + { + "epoch": 4.033659066232356, + "grad_norm": 16.729428248551148, + "learning_rate": 1.8962086957901882e-06, + "loss": 0.495, + "step": 7430 + }, + { + "epoch": 4.034201954397394, + "grad_norm": 9.607193659544246, + "learning_rate": 1.894148660188342e-06, + "loss": 0.326, + "step": 7431 + }, + { + "epoch": 4.034744842562432, + "grad_norm": 12.36629523451521, + "learning_rate": 1.892089627131103e-06, + "loss": 0.5407, + "step": 7432 + }, + { + "epoch": 4.03528773072747, + "grad_norm": 12.984997271109028, + "learning_rate": 1.8900315968731465e-06, + "loss": 0.401, + "step": 7433 + }, + { + "epoch": 4.035830618892508, + "grad_norm": 10.733134132257513, + "learning_rate": 1.8879745696690022e-06, + "loss": 0.2782, + "step": 7434 + }, + { + "epoch": 4.036373507057546, + "grad_norm": 18.013649168300645, + "learning_rate": 1.8859185457730944e-06, + "loss": 0.8034, + "step": 7435 + }, + { + "epoch": 4.036916395222584, + "grad_norm": 13.544413626676313, + "learning_rate": 1.8838635254397154e-06, + "loss": 0.3762, + "step": 7436 + }, + { + "epoch": 4.037459283387622, + "grad_norm": 11.586743250886352, + "learning_rate": 1.8818095089230248e-06, + "loss": 0.5606, + "step": 7437 + }, + { + "epoch": 4.03800217155266, + "grad_norm": 10.323656236449189, + "learning_rate": 1.8797564964770787e-06, + "loss": 0.3772, + "step": 7438 + }, + { + "epoch": 4.038545059717698, + "grad_norm": 11.956263255604744, + "learning_rate": 1.877704488355785e-06, + "loss": 0.3292, + "step": 7439 + }, + { + "epoch": 4.039087947882736, + "grad_norm": 11.706565985563376, + "learning_rate": 1.8756534848129504e-06, + "loss": 0.3045, + "step": 7440 + }, + { + "epoch": 4.039630836047774, + "grad_norm": 17.55780625111388, + "learning_rate": 1.8736034861022368e-06, + "loss": 0.6189, + "step": 7441 + }, + { + "epoch": 4.040173724212812, + "grad_norm": 12.815942770561511, + "learning_rate": 1.8715544924771977e-06, + "loss": 0.3457, + "step": 7442 + }, + { + "epoch": 4.04071661237785, + "grad_norm": 15.230354926789687, + "learning_rate": 1.869506504191253e-06, + "loss": 0.3725, + "step": 7443 + }, + { + "epoch": 4.041259500542888, + "grad_norm": 12.610726529266895, + "learning_rate": 1.8674595214977031e-06, + "loss": 0.4368, + "step": 7444 + }, + { + "epoch": 4.041802388707926, + "grad_norm": 16.959731764835915, + "learning_rate": 1.8654135446497234e-06, + "loss": 0.7666, + "step": 7445 + }, + { + "epoch": 4.042345276872964, + "grad_norm": 13.908609561441018, + "learning_rate": 1.8633685739003548e-06, + "loss": 0.2762, + "step": 7446 + }, + { + "epoch": 4.042888165038002, + "grad_norm": 12.835858815757566, + "learning_rate": 1.8613246095025329e-06, + "loss": 0.4223, + "step": 7447 + }, + { + "epoch": 4.04343105320304, + "grad_norm": 15.055795233826663, + "learning_rate": 1.859281651709053e-06, + "loss": 0.4317, + "step": 7448 + }, + { + "epoch": 4.043973941368078, + "grad_norm": 13.581558077614412, + "learning_rate": 1.857239700772594e-06, + "loss": 0.4312, + "step": 7449 + }, + { + "epoch": 4.044516829533116, + "grad_norm": 13.841266542296328, + "learning_rate": 1.8551987569457053e-06, + "loss": 0.3242, + "step": 7450 + }, + { + "epoch": 4.045059717698154, + "grad_norm": 10.057661085861465, + "learning_rate": 1.853158820480816e-06, + "loss": 0.4803, + "step": 7451 + }, + { + "epoch": 4.045602605863192, + "grad_norm": 11.127274589006348, + "learning_rate": 1.8511198916302274e-06, + "loss": 0.3005, + "step": 7452 + }, + { + "epoch": 4.04614549402823, + "grad_norm": 10.414927056352687, + "learning_rate": 1.8490819706461193e-06, + "loss": 0.4651, + "step": 7453 + }, + { + "epoch": 4.046688382193268, + "grad_norm": 11.548866791691525, + "learning_rate": 1.847045057780542e-06, + "loss": 0.416, + "step": 7454 + }, + { + "epoch": 4.047231270358306, + "grad_norm": 8.995318003205785, + "learning_rate": 1.8450091532854264e-06, + "loss": 0.1938, + "step": 7455 + }, + { + "epoch": 4.047774158523344, + "grad_norm": 12.54663924123166, + "learning_rate": 1.8429742574125765e-06, + "loss": 0.4278, + "step": 7456 + }, + { + "epoch": 4.048317046688382, + "grad_norm": 9.475847861618263, + "learning_rate": 1.8409403704136707e-06, + "loss": 0.2686, + "step": 7457 + }, + { + "epoch": 4.04885993485342, + "grad_norm": 14.596699092222988, + "learning_rate": 1.8389074925402628e-06, + "loss": 0.6494, + "step": 7458 + }, + { + "epoch": 4.049402823018458, + "grad_norm": 8.923910821563455, + "learning_rate": 1.8368756240437836e-06, + "loss": 0.2667, + "step": 7459 + }, + { + "epoch": 4.049945711183496, + "grad_norm": 12.475431384264864, + "learning_rate": 1.8348447651755364e-06, + "loss": 0.2887, + "step": 7460 + }, + { + "epoch": 4.050488599348534, + "grad_norm": 10.730423522471147, + "learning_rate": 1.8328149161867025e-06, + "loss": 0.3498, + "step": 7461 + }, + { + "epoch": 4.051031487513572, + "grad_norm": 16.724674239679032, + "learning_rate": 1.830786077328337e-06, + "loss": 0.7322, + "step": 7462 + }, + { + "epoch": 4.0515743756786105, + "grad_norm": 12.182586543041394, + "learning_rate": 1.8287582488513656e-06, + "loss": 0.4257, + "step": 7463 + }, + { + "epoch": 4.052117263843648, + "grad_norm": 6.932472045654237, + "learning_rate": 1.826731431006603e-06, + "loss": 0.2157, + "step": 7464 + }, + { + "epoch": 4.052660152008686, + "grad_norm": 8.646135191685339, + "learning_rate": 1.8247056240447203e-06, + "loss": 0.2646, + "step": 7465 + }, + { + "epoch": 4.053203040173724, + "grad_norm": 16.802820609812404, + "learning_rate": 1.8226808282162755e-06, + "loss": 0.5392, + "step": 7466 + }, + { + "epoch": 4.0537459283387625, + "grad_norm": 13.75433513899018, + "learning_rate": 1.8206570437716986e-06, + "loss": 0.367, + "step": 7467 + }, + { + "epoch": 4.0542888165038, + "grad_norm": 12.642507240641134, + "learning_rate": 1.8186342709612925e-06, + "loss": 0.4884, + "step": 7468 + }, + { + "epoch": 4.054831704668838, + "grad_norm": 12.446058308802613, + "learning_rate": 1.8166125100352427e-06, + "loss": 0.3996, + "step": 7469 + }, + { + "epoch": 4.055374592833876, + "grad_norm": 9.780058065307577, + "learning_rate": 1.814591761243596e-06, + "loss": 0.2016, + "step": 7470 + }, + { + "epoch": 4.0559174809989145, + "grad_norm": 13.137545091734312, + "learning_rate": 1.812572024836291e-06, + "loss": 0.5646, + "step": 7471 + }, + { + "epoch": 4.056460369163952, + "grad_norm": 13.982451209627083, + "learning_rate": 1.8105533010631215e-06, + "loss": 0.3899, + "step": 7472 + }, + { + "epoch": 4.05700325732899, + "grad_norm": 14.830274354521338, + "learning_rate": 1.8085355901737767e-06, + "loss": 0.6667, + "step": 7473 + }, + { + "epoch": 4.057546145494028, + "grad_norm": 15.746342293850457, + "learning_rate": 1.8065188924178002e-06, + "loss": 0.6769, + "step": 7474 + }, + { + "epoch": 4.0580890336590665, + "grad_norm": 19.880439031191987, + "learning_rate": 1.8045032080446279e-06, + "loss": 0.6128, + "step": 7475 + }, + { + "epoch": 4.058631921824104, + "grad_norm": 12.174293630332373, + "learning_rate": 1.8024885373035628e-06, + "loss": 0.351, + "step": 7476 + }, + { + "epoch": 4.059174809989142, + "grad_norm": 13.340029936420176, + "learning_rate": 1.8004748804437755e-06, + "loss": 0.4044, + "step": 7477 + }, + { + "epoch": 4.05971769815418, + "grad_norm": 10.710849412967622, + "learning_rate": 1.798462237714329e-06, + "loss": 0.2545, + "step": 7478 + }, + { + "epoch": 4.0602605863192185, + "grad_norm": 14.145700312877072, + "learning_rate": 1.7964506093641388e-06, + "loss": 0.442, + "step": 7479 + }, + { + "epoch": 4.060803474484256, + "grad_norm": 11.782368982891633, + "learning_rate": 1.7944399956420133e-06, + "loss": 0.3699, + "step": 7480 + }, + { + "epoch": 4.061346362649294, + "grad_norm": 16.400383314329044, + "learning_rate": 1.7924303967966283e-06, + "loss": 0.4159, + "step": 7481 + }, + { + "epoch": 4.061889250814332, + "grad_norm": 14.158517117689607, + "learning_rate": 1.7904218130765316e-06, + "loss": 0.3477, + "step": 7482 + }, + { + "epoch": 4.0624321389793705, + "grad_norm": 10.538895009930492, + "learning_rate": 1.78841424473015e-06, + "loss": 0.4439, + "step": 7483 + }, + { + "epoch": 4.062975027144408, + "grad_norm": 15.749637206952377, + "learning_rate": 1.7864076920057815e-06, + "loss": 0.4063, + "step": 7484 + }, + { + "epoch": 4.063517915309446, + "grad_norm": 14.701502876295633, + "learning_rate": 1.7844021551516023e-06, + "loss": 0.293, + "step": 7485 + }, + { + "epoch": 4.064060803474484, + "grad_norm": 13.441957282892526, + "learning_rate": 1.7823976344156534e-06, + "loss": 0.3922, + "step": 7486 + }, + { + "epoch": 4.0646036916395225, + "grad_norm": 12.860794488151207, + "learning_rate": 1.780394130045865e-06, + "loss": 0.3637, + "step": 7487 + }, + { + "epoch": 4.06514657980456, + "grad_norm": 10.713125214412898, + "learning_rate": 1.7783916422900304e-06, + "loss": 0.3995, + "step": 7488 + }, + { + "epoch": 4.065689467969598, + "grad_norm": 12.736414280259371, + "learning_rate": 1.776390171395821e-06, + "loss": 0.4328, + "step": 7489 + }, + { + "epoch": 4.066232356134636, + "grad_norm": 13.781401823444826, + "learning_rate": 1.7743897176107817e-06, + "loss": 0.5039, + "step": 7490 + }, + { + "epoch": 4.0667752442996745, + "grad_norm": 12.032700692680026, + "learning_rate": 1.7723902811823312e-06, + "loss": 0.2701, + "step": 7491 + }, + { + "epoch": 4.067318132464712, + "grad_norm": 13.357720909095406, + "learning_rate": 1.7703918623577642e-06, + "loss": 0.472, + "step": 7492 + }, + { + "epoch": 4.06786102062975, + "grad_norm": 13.666181670657824, + "learning_rate": 1.7683944613842474e-06, + "loss": 0.4525, + "step": 7493 + }, + { + "epoch": 4.068403908794788, + "grad_norm": 15.204188285424376, + "learning_rate": 1.766398078508823e-06, + "loss": 0.5151, + "step": 7494 + }, + { + "epoch": 4.0689467969598265, + "grad_norm": 14.986125649693381, + "learning_rate": 1.7644027139784058e-06, + "loss": 0.5068, + "step": 7495 + }, + { + "epoch": 4.069489685124864, + "grad_norm": 17.414959041982844, + "learning_rate": 1.7624083680397874e-06, + "loss": 0.5285, + "step": 7496 + }, + { + "epoch": 4.070032573289902, + "grad_norm": 15.018593246149907, + "learning_rate": 1.7604150409396315e-06, + "loss": 0.4034, + "step": 7497 + }, + { + "epoch": 4.07057546145494, + "grad_norm": 7.740045825599994, + "learning_rate": 1.758422732924474e-06, + "loss": 0.2231, + "step": 7498 + }, + { + "epoch": 4.0711183496199785, + "grad_norm": 15.973649660315148, + "learning_rate": 1.7564314442407282e-06, + "loss": 0.4157, + "step": 7499 + }, + { + "epoch": 4.071661237785016, + "grad_norm": 15.522023726804239, + "learning_rate": 1.7544411751346802e-06, + "loss": 0.5029, + "step": 7500 + }, + { + "epoch": 4.072204125950054, + "grad_norm": 19.621921290245965, + "learning_rate": 1.7524519258524863e-06, + "loss": 0.4998, + "step": 7501 + }, + { + "epoch": 4.072747014115092, + "grad_norm": 13.460020491045686, + "learning_rate": 1.7504636966401877e-06, + "loss": 0.4035, + "step": 7502 + }, + { + "epoch": 4.0732899022801305, + "grad_norm": 12.082530806707663, + "learning_rate": 1.7484764877436822e-06, + "loss": 0.4754, + "step": 7503 + }, + { + "epoch": 4.073832790445168, + "grad_norm": 10.41118329664297, + "learning_rate": 1.7464902994087607e-06, + "loss": 0.331, + "step": 7504 + }, + { + "epoch": 4.074375678610206, + "grad_norm": 13.189232599075739, + "learning_rate": 1.7445051318810702e-06, + "loss": 0.4648, + "step": 7505 + }, + { + "epoch": 4.074918566775244, + "grad_norm": 9.349455348221264, + "learning_rate": 1.742520985406143e-06, + "loss": 0.2363, + "step": 7506 + }, + { + "epoch": 4.075461454940283, + "grad_norm": 11.755562149227849, + "learning_rate": 1.74053786022938e-06, + "loss": 0.2546, + "step": 7507 + }, + { + "epoch": 4.07600434310532, + "grad_norm": 13.819551647694206, + "learning_rate": 1.7385557565960564e-06, + "loss": 0.508, + "step": 7508 + }, + { + "epoch": 4.076547231270358, + "grad_norm": 9.181009774294514, + "learning_rate": 1.7365746747513278e-06, + "loss": 0.2214, + "step": 7509 + }, + { + "epoch": 4.077090119435396, + "grad_norm": 8.97042904079199, + "learning_rate": 1.7345946149402094e-06, + "loss": 0.2924, + "step": 7510 + }, + { + "epoch": 4.077633007600435, + "grad_norm": 12.257238316765228, + "learning_rate": 1.7326155774076058e-06, + "loss": 0.446, + "step": 7511 + }, + { + "epoch": 4.078175895765472, + "grad_norm": 18.213109072198556, + "learning_rate": 1.7306375623982796e-06, + "loss": 0.5576, + "step": 7512 + }, + { + "epoch": 4.07871878393051, + "grad_norm": 15.252433630945886, + "learning_rate": 1.7286605701568815e-06, + "loss": 0.5242, + "step": 7513 + }, + { + "epoch": 4.079261672095548, + "grad_norm": 12.724698295224274, + "learning_rate": 1.7266846009279292e-06, + "loss": 0.4852, + "step": 7514 + }, + { + "epoch": 4.079804560260587, + "grad_norm": 14.292147421433285, + "learning_rate": 1.7247096549558062e-06, + "loss": 0.3125, + "step": 7515 + }, + { + "epoch": 4.080347448425624, + "grad_norm": 11.689927807096288, + "learning_rate": 1.722735732484786e-06, + "loss": 0.3077, + "step": 7516 + }, + { + "epoch": 4.080890336590662, + "grad_norm": 17.211462481334554, + "learning_rate": 1.7207628337589988e-06, + "loss": 0.6293, + "step": 7517 + }, + { + "epoch": 4.0814332247557, + "grad_norm": 11.662256116349747, + "learning_rate": 1.7187909590224604e-06, + "loss": 0.3055, + "step": 7518 + }, + { + "epoch": 4.081976112920739, + "grad_norm": 10.371623662202015, + "learning_rate": 1.7168201085190562e-06, + "loss": 0.3695, + "step": 7519 + }, + { + "epoch": 4.082519001085776, + "grad_norm": 14.563522671402897, + "learning_rate": 1.7148502824925418e-06, + "loss": 0.3588, + "step": 7520 + }, + { + "epoch": 4.083061889250814, + "grad_norm": 12.36778161398001, + "learning_rate": 1.712881481186548e-06, + "loss": 0.3054, + "step": 7521 + }, + { + "epoch": 4.083604777415852, + "grad_norm": 13.99470872596806, + "learning_rate": 1.7109137048445812e-06, + "loss": 0.4663, + "step": 7522 + }, + { + "epoch": 4.084147665580891, + "grad_norm": 11.566073320833054, + "learning_rate": 1.7089469537100178e-06, + "loss": 0.4081, + "step": 7523 + }, + { + "epoch": 4.084690553745928, + "grad_norm": 10.823284667656079, + "learning_rate": 1.70698122802611e-06, + "loss": 0.3027, + "step": 7524 + }, + { + "epoch": 4.085233441910966, + "grad_norm": 14.505207806467645, + "learning_rate": 1.705016528035981e-06, + "loss": 0.7583, + "step": 7525 + }, + { + "epoch": 4.085776330076004, + "grad_norm": 13.977656869068836, + "learning_rate": 1.7030528539826275e-06, + "loss": 0.4479, + "step": 7526 + }, + { + "epoch": 4.086319218241043, + "grad_norm": 14.88627029394164, + "learning_rate": 1.7010902061089197e-06, + "loss": 0.5773, + "step": 7527 + }, + { + "epoch": 4.08686210640608, + "grad_norm": 9.998415444955523, + "learning_rate": 1.6991285846576022e-06, + "loss": 0.2414, + "step": 7528 + }, + { + "epoch": 4.087404994571118, + "grad_norm": 11.634301676106471, + "learning_rate": 1.6971679898712912e-06, + "loss": 0.3949, + "step": 7529 + }, + { + "epoch": 4.087947882736156, + "grad_norm": 12.469794055243895, + "learning_rate": 1.6952084219924757e-06, + "loss": 0.4851, + "step": 7530 + }, + { + "epoch": 4.088490770901195, + "grad_norm": 11.162959342238105, + "learning_rate": 1.6932498812635189e-06, + "loss": 0.4459, + "step": 7531 + }, + { + "epoch": 4.089033659066232, + "grad_norm": 17.59176356543345, + "learning_rate": 1.6912923679266557e-06, + "loss": 0.6036, + "step": 7532 + }, + { + "epoch": 4.08957654723127, + "grad_norm": 10.674032941701148, + "learning_rate": 1.6893358822239947e-06, + "loss": 0.3718, + "step": 7533 + }, + { + "epoch": 4.090119435396308, + "grad_norm": 9.119363490702213, + "learning_rate": 1.6873804243975167e-06, + "loss": 0.2871, + "step": 7534 + }, + { + "epoch": 4.090662323561347, + "grad_norm": 13.955880230291728, + "learning_rate": 1.6854259946890762e-06, + "loss": 0.384, + "step": 7535 + }, + { + "epoch": 4.091205211726384, + "grad_norm": 15.830781520521217, + "learning_rate": 1.6834725933403995e-06, + "loss": 0.59, + "step": 7536 + }, + { + "epoch": 4.091748099891422, + "grad_norm": 11.306287185705354, + "learning_rate": 1.681520220593088e-06, + "loss": 0.3796, + "step": 7537 + }, + { + "epoch": 4.09229098805646, + "grad_norm": 14.796920699824259, + "learning_rate": 1.6795688766886132e-06, + "loss": 0.4181, + "step": 7538 + }, + { + "epoch": 4.092833876221499, + "grad_norm": 15.602066419097026, + "learning_rate": 1.6776185618683171e-06, + "loss": 0.5769, + "step": 7539 + }, + { + "epoch": 4.093376764386536, + "grad_norm": 10.75921703150943, + "learning_rate": 1.6756692763734272e-06, + "loss": 0.3376, + "step": 7540 + }, + { + "epoch": 4.093919652551574, + "grad_norm": 12.134257694162375, + "learning_rate": 1.6737210204450226e-06, + "loss": 0.3719, + "step": 7541 + }, + { + "epoch": 4.094462540716612, + "grad_norm": 14.106939136679763, + "learning_rate": 1.6717737943240774e-06, + "loss": 0.6107, + "step": 7542 + }, + { + "epoch": 4.095005428881651, + "grad_norm": 12.295261891318004, + "learning_rate": 1.6698275982514178e-06, + "loss": 0.35, + "step": 7543 + }, + { + "epoch": 4.095548317046688, + "grad_norm": 11.519476162199268, + "learning_rate": 1.6678824324677623e-06, + "loss": 0.3783, + "step": 7544 + }, + { + "epoch": 4.096091205211726, + "grad_norm": 8.77456108175982, + "learning_rate": 1.6659382972136839e-06, + "loss": 0.2232, + "step": 7545 + }, + { + "epoch": 4.096634093376764, + "grad_norm": 11.479634703602272, + "learning_rate": 1.6639951927296371e-06, + "loss": 0.3152, + "step": 7546 + }, + { + "epoch": 4.097176981541803, + "grad_norm": 11.756787071755726, + "learning_rate": 1.6620531192559552e-06, + "loss": 0.4063, + "step": 7547 + }, + { + "epoch": 4.09771986970684, + "grad_norm": 16.583849337588653, + "learning_rate": 1.6601120770328283e-06, + "loss": 0.5936, + "step": 7548 + }, + { + "epoch": 4.098262757871878, + "grad_norm": 12.762971955755509, + "learning_rate": 1.6581720663003354e-06, + "loss": 0.4476, + "step": 7549 + }, + { + "epoch": 4.098805646036916, + "grad_norm": 18.75880484333442, + "learning_rate": 1.6562330872984122e-06, + "loss": 0.6011, + "step": 7550 + }, + { + "epoch": 4.099348534201955, + "grad_norm": 18.29186988565711, + "learning_rate": 1.6542951402668805e-06, + "loss": 0.5015, + "step": 7551 + }, + { + "epoch": 4.099891422366992, + "grad_norm": 12.057222479862308, + "learning_rate": 1.6523582254454273e-06, + "loss": 0.6185, + "step": 7552 + }, + { + "epoch": 4.10043431053203, + "grad_norm": 15.90807833403242, + "learning_rate": 1.6504223430736122e-06, + "loss": 0.5644, + "step": 7553 + }, + { + "epoch": 4.100977198697068, + "grad_norm": 15.27338292218376, + "learning_rate": 1.6484874933908723e-06, + "loss": 0.4299, + "step": 7554 + }, + { + "epoch": 4.101520086862107, + "grad_norm": 12.827779978027912, + "learning_rate": 1.6465536766365043e-06, + "loss": 0.4768, + "step": 7555 + }, + { + "epoch": 4.1020629750271445, + "grad_norm": 19.22780764109623, + "learning_rate": 1.6446208930496954e-06, + "loss": 0.3838, + "step": 7556 + }, + { + "epoch": 4.102605863192182, + "grad_norm": 14.630988970729627, + "learning_rate": 1.642689142869487e-06, + "loss": 0.3607, + "step": 7557 + }, + { + "epoch": 4.10314875135722, + "grad_norm": 11.07230018602903, + "learning_rate": 1.6407584263348065e-06, + "loss": 0.2996, + "step": 7558 + }, + { + "epoch": 4.103691639522259, + "grad_norm": 12.881805511729306, + "learning_rate": 1.6388287436844474e-06, + "loss": 0.4539, + "step": 7559 + }, + { + "epoch": 4.1042345276872965, + "grad_norm": 12.06776906866114, + "learning_rate": 1.6369000951570746e-06, + "loss": 0.534, + "step": 7560 + }, + { + "epoch": 4.104777415852334, + "grad_norm": 16.25382663633683, + "learning_rate": 1.634972480991226e-06, + "loss": 0.5308, + "step": 7561 + }, + { + "epoch": 4.105320304017372, + "grad_norm": 10.586138014834138, + "learning_rate": 1.6330459014253141e-06, + "loss": 0.3352, + "step": 7562 + }, + { + "epoch": 4.105863192182411, + "grad_norm": 15.890912108947397, + "learning_rate": 1.6311203566976196e-06, + "loss": 0.5464, + "step": 7563 + }, + { + "epoch": 4.1064060803474485, + "grad_norm": 12.756769761348481, + "learning_rate": 1.6291958470462966e-06, + "loss": 0.3651, + "step": 7564 + }, + { + "epoch": 4.106948968512486, + "grad_norm": 10.76450128852776, + "learning_rate": 1.6272723727093732e-06, + "loss": 0.3137, + "step": 7565 + }, + { + "epoch": 4.107491856677524, + "grad_norm": 13.308264213836948, + "learning_rate": 1.6253499339247468e-06, + "loss": 0.5733, + "step": 7566 + }, + { + "epoch": 4.108034744842563, + "grad_norm": 13.76589828286844, + "learning_rate": 1.6234285309301878e-06, + "loss": 0.473, + "step": 7567 + }, + { + "epoch": 4.1085776330076005, + "grad_norm": 14.628341432791526, + "learning_rate": 1.621508163963338e-06, + "loss": 0.2876, + "step": 7568 + }, + { + "epoch": 4.109120521172638, + "grad_norm": 11.72578228984837, + "learning_rate": 1.6195888332617126e-06, + "loss": 0.4284, + "step": 7569 + }, + { + "epoch": 4.109663409337676, + "grad_norm": 20.77607641436691, + "learning_rate": 1.6176705390626956e-06, + "loss": 0.7359, + "step": 7570 + }, + { + "epoch": 4.110206297502715, + "grad_norm": 15.592080343031311, + "learning_rate": 1.615753281603546e-06, + "loss": 0.5176, + "step": 7571 + }, + { + "epoch": 4.1107491856677525, + "grad_norm": 14.765460153242566, + "learning_rate": 1.6138370611213917e-06, + "loss": 0.3779, + "step": 7572 + }, + { + "epoch": 4.11129207383279, + "grad_norm": 11.227048300024913, + "learning_rate": 1.6119218778532386e-06, + "loss": 0.2835, + "step": 7573 + }, + { + "epoch": 4.111834961997828, + "grad_norm": 10.73327552489785, + "learning_rate": 1.6100077320359552e-06, + "loss": 0.3355, + "step": 7574 + }, + { + "epoch": 4.112377850162867, + "grad_norm": 12.220262589971611, + "learning_rate": 1.6080946239062856e-06, + "loss": 0.2647, + "step": 7575 + }, + { + "epoch": 4.1129207383279045, + "grad_norm": 13.578576106400867, + "learning_rate": 1.6061825537008491e-06, + "loss": 0.4504, + "step": 7576 + }, + { + "epoch": 4.113463626492942, + "grad_norm": 12.666206872990253, + "learning_rate": 1.6042715216561312e-06, + "loss": 0.4362, + "step": 7577 + }, + { + "epoch": 4.11400651465798, + "grad_norm": 14.81928570898845, + "learning_rate": 1.6023615280084925e-06, + "loss": 0.5685, + "step": 7578 + }, + { + "epoch": 4.114549402823019, + "grad_norm": 13.058193252197759, + "learning_rate": 1.6004525729941622e-06, + "loss": 0.5339, + "step": 7579 + }, + { + "epoch": 4.1150922909880565, + "grad_norm": 16.238471411046245, + "learning_rate": 1.598544656849248e-06, + "loss": 0.6546, + "step": 7580 + }, + { + "epoch": 4.115635179153094, + "grad_norm": 11.090517158246884, + "learning_rate": 1.5966377798097165e-06, + "loss": 0.4045, + "step": 7581 + }, + { + "epoch": 4.116178067318132, + "grad_norm": 15.25648505046589, + "learning_rate": 1.5947319421114227e-06, + "loss": 0.5787, + "step": 7582 + }, + { + "epoch": 4.116720955483171, + "grad_norm": 11.253131068704839, + "learning_rate": 1.5928271439900755e-06, + "loss": 0.3468, + "step": 7583 + }, + { + "epoch": 4.1172638436482085, + "grad_norm": 22.87822267893679, + "learning_rate": 1.5909233856812634e-06, + "loss": 0.5457, + "step": 7584 + }, + { + "epoch": 4.117806731813246, + "grad_norm": 11.373191676935885, + "learning_rate": 1.5890206674204535e-06, + "loss": 0.4627, + "step": 7585 + }, + { + "epoch": 4.118349619978284, + "grad_norm": 14.408977573167448, + "learning_rate": 1.5871189894429683e-06, + "loss": 0.3613, + "step": 7586 + }, + { + "epoch": 4.118892508143323, + "grad_norm": 13.14025988474886, + "learning_rate": 1.5852183519840203e-06, + "loss": 0.5449, + "step": 7587 + }, + { + "epoch": 4.1194353963083605, + "grad_norm": 13.859357618288538, + "learning_rate": 1.583318755278671e-06, + "loss": 0.5978, + "step": 7588 + }, + { + "epoch": 4.119978284473398, + "grad_norm": 12.173600067172606, + "learning_rate": 1.581420199561876e-06, + "loss": 0.4165, + "step": 7589 + }, + { + "epoch": 4.120521172638436, + "grad_norm": 11.35642341805059, + "learning_rate": 1.5795226850684476e-06, + "loss": 0.4934, + "step": 7590 + }, + { + "epoch": 4.121064060803475, + "grad_norm": 15.267752687607832, + "learning_rate": 1.5776262120330743e-06, + "loss": 0.4089, + "step": 7591 + }, + { + "epoch": 4.1216069489685125, + "grad_norm": 14.238479740595348, + "learning_rate": 1.5757307806903133e-06, + "loss": 0.4299, + "step": 7592 + }, + { + "epoch": 4.12214983713355, + "grad_norm": 11.160357781980832, + "learning_rate": 1.5738363912745959e-06, + "loss": 0.4948, + "step": 7593 + }, + { + "epoch": 4.122692725298588, + "grad_norm": 16.47054898460848, + "learning_rate": 1.5719430440202255e-06, + "loss": 0.5639, + "step": 7594 + }, + { + "epoch": 4.123235613463627, + "grad_norm": 15.099982308232326, + "learning_rate": 1.5700507391613673e-06, + "loss": 0.3724, + "step": 7595 + }, + { + "epoch": 4.1237785016286646, + "grad_norm": 13.904934094160584, + "learning_rate": 1.5681594769320706e-06, + "loss": 0.5423, + "step": 7596 + }, + { + "epoch": 4.124321389793702, + "grad_norm": 10.17978139897218, + "learning_rate": 1.5662692575662486e-06, + "loss": 0.3509, + "step": 7597 + }, + { + "epoch": 4.12486427795874, + "grad_norm": 10.662732400077363, + "learning_rate": 1.5643800812976873e-06, + "loss": 0.306, + "step": 7598 + }, + { + "epoch": 4.125407166123779, + "grad_norm": 10.833314489423646, + "learning_rate": 1.56249194836004e-06, + "loss": 0.3767, + "step": 7599 + }, + { + "epoch": 4.125950054288817, + "grad_norm": 14.593277788934097, + "learning_rate": 1.5606048589868372e-06, + "loss": 0.3886, + "step": 7600 + }, + { + "epoch": 4.126492942453854, + "grad_norm": 19.995368160150967, + "learning_rate": 1.558718813411475e-06, + "loss": 0.3951, + "step": 7601 + }, + { + "epoch": 4.127035830618892, + "grad_norm": 8.750505077849805, + "learning_rate": 1.5568338118672243e-06, + "loss": 0.2172, + "step": 7602 + }, + { + "epoch": 4.127578718783931, + "grad_norm": 13.314482415981553, + "learning_rate": 1.5549498545872233e-06, + "loss": 0.2903, + "step": 7603 + }, + { + "epoch": 4.128121606948969, + "grad_norm": 10.865437650317455, + "learning_rate": 1.5530669418044842e-06, + "loss": 0.3271, + "step": 7604 + }, + { + "epoch": 4.128664495114006, + "grad_norm": 13.939733125814856, + "learning_rate": 1.5511850737518875e-06, + "loss": 0.3687, + "step": 7605 + }, + { + "epoch": 4.129207383279044, + "grad_norm": 13.380151193528324, + "learning_rate": 1.549304250662187e-06, + "loss": 0.3158, + "step": 7606 + }, + { + "epoch": 4.129750271444083, + "grad_norm": 9.920776483442411, + "learning_rate": 1.5474244727680055e-06, + "loss": 0.3943, + "step": 7607 + }, + { + "epoch": 4.130293159609121, + "grad_norm": 11.896762787826454, + "learning_rate": 1.545545740301836e-06, + "loss": 0.5542, + "step": 7608 + }, + { + "epoch": 4.130836047774158, + "grad_norm": 12.91558903974088, + "learning_rate": 1.5436680534960434e-06, + "loss": 0.4152, + "step": 7609 + }, + { + "epoch": 4.131378935939196, + "grad_norm": 10.961854763205224, + "learning_rate": 1.5417914125828616e-06, + "loss": 0.4836, + "step": 7610 + }, + { + "epoch": 4.131921824104235, + "grad_norm": 10.548604561082739, + "learning_rate": 1.5399158177944019e-06, + "loss": 0.4001, + "step": 7611 + }, + { + "epoch": 4.132464712269273, + "grad_norm": 15.534076149917789, + "learning_rate": 1.5380412693626356e-06, + "loss": 0.5564, + "step": 7612 + }, + { + "epoch": 4.13300760043431, + "grad_norm": 12.354229348605834, + "learning_rate": 1.536167767519412e-06, + "loss": 0.5188, + "step": 7613 + }, + { + "epoch": 4.133550488599348, + "grad_norm": 15.550407127657197, + "learning_rate": 1.5342953124964477e-06, + "loss": 0.7004, + "step": 7614 + }, + { + "epoch": 4.134093376764387, + "grad_norm": 13.960316578712156, + "learning_rate": 1.532423904525332e-06, + "loss": 0.3999, + "step": 7615 + }, + { + "epoch": 4.134636264929425, + "grad_norm": 12.1140917102908, + "learning_rate": 1.5305535438375218e-06, + "loss": 0.3274, + "step": 7616 + }, + { + "epoch": 4.135179153094462, + "grad_norm": 12.38487299229727, + "learning_rate": 1.5286842306643468e-06, + "loss": 0.4094, + "step": 7617 + }, + { + "epoch": 4.1357220412595, + "grad_norm": 10.14740725625142, + "learning_rate": 1.5268159652370108e-06, + "loss": 0.4163, + "step": 7618 + }, + { + "epoch": 4.136264929424539, + "grad_norm": 14.207848794460347, + "learning_rate": 1.5249487477865765e-06, + "loss": 0.3613, + "step": 7619 + }, + { + "epoch": 4.136807817589577, + "grad_norm": 17.170203385892172, + "learning_rate": 1.523082578543993e-06, + "loss": 0.4012, + "step": 7620 + }, + { + "epoch": 4.137350705754614, + "grad_norm": 11.31582486367915, + "learning_rate": 1.5212174577400618e-06, + "loss": 0.3245, + "step": 7621 + }, + { + "epoch": 4.137893593919652, + "grad_norm": 15.179364893086714, + "learning_rate": 1.519353385605471e-06, + "loss": 0.418, + "step": 7622 + }, + { + "epoch": 4.138436482084691, + "grad_norm": 16.498670863026067, + "learning_rate": 1.5174903623707716e-06, + "loss": 0.4128, + "step": 7623 + }, + { + "epoch": 4.138979370249729, + "grad_norm": 19.148612028425138, + "learning_rate": 1.515628388266378e-06, + "loss": 0.6773, + "step": 7624 + }, + { + "epoch": 4.139522258414766, + "grad_norm": 21.1656943566629, + "learning_rate": 1.513767463522593e-06, + "loss": 0.6674, + "step": 7625 + }, + { + "epoch": 4.140065146579804, + "grad_norm": 13.460184953663427, + "learning_rate": 1.5119075883695678e-06, + "loss": 0.4615, + "step": 7626 + }, + { + "epoch": 4.140608034744843, + "grad_norm": 13.260423844727587, + "learning_rate": 1.510048763037345e-06, + "loss": 0.3272, + "step": 7627 + }, + { + "epoch": 4.141150922909881, + "grad_norm": 11.331130692905933, + "learning_rate": 1.5081909877558177e-06, + "loss": 0.2903, + "step": 7628 + }, + { + "epoch": 4.141693811074918, + "grad_norm": 15.865149055955152, + "learning_rate": 1.506334262754765e-06, + "loss": 0.4817, + "step": 7629 + }, + { + "epoch": 4.142236699239956, + "grad_norm": 11.349271790017943, + "learning_rate": 1.5044785882638279e-06, + "loss": 0.3778, + "step": 7630 + }, + { + "epoch": 4.142779587404995, + "grad_norm": 12.519516513475779, + "learning_rate": 1.5026239645125196e-06, + "loss": 0.2995, + "step": 7631 + }, + { + "epoch": 4.143322475570033, + "grad_norm": 13.209913219504104, + "learning_rate": 1.500770391730223e-06, + "loss": 0.4709, + "step": 7632 + }, + { + "epoch": 4.14386536373507, + "grad_norm": 9.929305640423868, + "learning_rate": 1.498917870146186e-06, + "loss": 0.269, + "step": 7633 + }, + { + "epoch": 4.144408251900108, + "grad_norm": 17.43652576717896, + "learning_rate": 1.4970663999895384e-06, + "loss": 0.7616, + "step": 7634 + }, + { + "epoch": 4.144951140065147, + "grad_norm": 10.810139421001562, + "learning_rate": 1.4952159814892697e-06, + "loss": 0.3408, + "step": 7635 + }, + { + "epoch": 4.145494028230185, + "grad_norm": 17.00710504331134, + "learning_rate": 1.4933666148742432e-06, + "loss": 0.6018, + "step": 7636 + }, + { + "epoch": 4.146036916395222, + "grad_norm": 8.623042522508, + "learning_rate": 1.4915183003731904e-06, + "loss": 0.2758, + "step": 7637 + }, + { + "epoch": 4.14657980456026, + "grad_norm": 16.04766738068284, + "learning_rate": 1.489671038214715e-06, + "loss": 0.6531, + "step": 7638 + }, + { + "epoch": 4.147122692725299, + "grad_norm": 11.752307004240341, + "learning_rate": 1.48782482862729e-06, + "loss": 0.3995, + "step": 7639 + }, + { + "epoch": 4.147665580890337, + "grad_norm": 12.146925988953807, + "learning_rate": 1.485979671839256e-06, + "loss": 0.4834, + "step": 7640 + }, + { + "epoch": 4.148208469055374, + "grad_norm": 17.18792212776091, + "learning_rate": 1.4841355680788261e-06, + "loss": 0.5534, + "step": 7641 + }, + { + "epoch": 4.148751357220412, + "grad_norm": 12.648802834009173, + "learning_rate": 1.4822925175740798e-06, + "loss": 0.5194, + "step": 7642 + }, + { + "epoch": 4.149294245385451, + "grad_norm": 12.067595847261051, + "learning_rate": 1.4804505205529718e-06, + "loss": 0.3579, + "step": 7643 + }, + { + "epoch": 4.149837133550489, + "grad_norm": 10.279884523990727, + "learning_rate": 1.4786095772433206e-06, + "loss": 0.3087, + "step": 7644 + }, + { + "epoch": 4.1503800217155264, + "grad_norm": 12.341347959843054, + "learning_rate": 1.4767696878728188e-06, + "loss": 0.314, + "step": 7645 + }, + { + "epoch": 4.150922909880564, + "grad_norm": 9.540652635109405, + "learning_rate": 1.4749308526690253e-06, + "loss": 0.3572, + "step": 7646 + }, + { + "epoch": 4.151465798045603, + "grad_norm": 12.941719739828546, + "learning_rate": 1.4730930718593705e-06, + "loss": 0.3057, + "step": 7647 + }, + { + "epoch": 4.152008686210641, + "grad_norm": 10.404139051977397, + "learning_rate": 1.471256345671156e-06, + "loss": 0.3796, + "step": 7648 + }, + { + "epoch": 4.1525515743756785, + "grad_norm": 11.552988822806185, + "learning_rate": 1.4694206743315476e-06, + "loss": 0.3922, + "step": 7649 + }, + { + "epoch": 4.153094462540716, + "grad_norm": 10.698397299218179, + "learning_rate": 1.467586058067585e-06, + "loss": 0.3779, + "step": 7650 + }, + { + "epoch": 4.153637350705755, + "grad_norm": 14.80756632794264, + "learning_rate": 1.465752497106181e-06, + "loss": 0.2783, + "step": 7651 + }, + { + "epoch": 4.154180238870793, + "grad_norm": 14.09754720490473, + "learning_rate": 1.4639199916741087e-06, + "loss": 0.315, + "step": 7652 + }, + { + "epoch": 4.1547231270358305, + "grad_norm": 8.268573012134576, + "learning_rate": 1.4620885419980147e-06, + "loss": 0.321, + "step": 7653 + }, + { + "epoch": 4.155266015200868, + "grad_norm": 18.11398789359469, + "learning_rate": 1.4602581483044188e-06, + "loss": 0.5637, + "step": 7654 + }, + { + "epoch": 4.155808903365907, + "grad_norm": 13.40679373737216, + "learning_rate": 1.458428810819702e-06, + "loss": 0.3844, + "step": 7655 + }, + { + "epoch": 4.156351791530945, + "grad_norm": 19.530436228333638, + "learning_rate": 1.456600529770128e-06, + "loss": 0.7487, + "step": 7656 + }, + { + "epoch": 4.1568946796959825, + "grad_norm": 15.655440408772145, + "learning_rate": 1.4547733053818125e-06, + "loss": 0.5907, + "step": 7657 + }, + { + "epoch": 4.15743756786102, + "grad_norm": 11.05518654892871, + "learning_rate": 1.452947137880757e-06, + "loss": 0.3539, + "step": 7658 + }, + { + "epoch": 4.157980456026059, + "grad_norm": 13.535214104936339, + "learning_rate": 1.4511220274928183e-06, + "loss": 0.4481, + "step": 7659 + }, + { + "epoch": 4.158523344191097, + "grad_norm": 15.227632590619127, + "learning_rate": 1.4492979744437342e-06, + "loss": 0.5095, + "step": 7660 + }, + { + "epoch": 4.1590662323561345, + "grad_norm": 12.117485924978645, + "learning_rate": 1.4474749789591058e-06, + "loss": 0.3035, + "step": 7661 + }, + { + "epoch": 4.159609120521172, + "grad_norm": 15.47621703654388, + "learning_rate": 1.4456530412643988e-06, + "loss": 0.7481, + "step": 7662 + }, + { + "epoch": 4.160152008686211, + "grad_norm": 16.71458160401543, + "learning_rate": 1.4438321615849615e-06, + "loss": 0.7957, + "step": 7663 + }, + { + "epoch": 4.160694896851249, + "grad_norm": 16.66028628014589, + "learning_rate": 1.442012340145995e-06, + "loss": 0.4658, + "step": 7664 + }, + { + "epoch": 4.1612377850162865, + "grad_norm": 14.178694096135883, + "learning_rate": 1.440193577172585e-06, + "loss": 0.5093, + "step": 7665 + }, + { + "epoch": 4.161780673181324, + "grad_norm": 13.385243486849467, + "learning_rate": 1.4383758728896724e-06, + "loss": 0.448, + "step": 7666 + }, + { + "epoch": 4.162323561346363, + "grad_norm": 12.920254235657909, + "learning_rate": 1.4365592275220787e-06, + "loss": 0.3069, + "step": 7667 + }, + { + "epoch": 4.162866449511401, + "grad_norm": 17.879050333979382, + "learning_rate": 1.4347436412944882e-06, + "loss": 0.6627, + "step": 7668 + }, + { + "epoch": 4.1634093376764385, + "grad_norm": 13.04178990878481, + "learning_rate": 1.4329291144314549e-06, + "loss": 0.3993, + "step": 7669 + }, + { + "epoch": 4.163952225841476, + "grad_norm": 13.341317379498156, + "learning_rate": 1.4311156471574027e-06, + "loss": 0.2989, + "step": 7670 + }, + { + "epoch": 4.164495114006515, + "grad_norm": 15.262211702534875, + "learning_rate": 1.4293032396966234e-06, + "loss": 0.6915, + "step": 7671 + }, + { + "epoch": 4.165038002171553, + "grad_norm": 10.866372137845064, + "learning_rate": 1.4274918922732805e-06, + "loss": 0.3446, + "step": 7672 + }, + { + "epoch": 4.1655808903365905, + "grad_norm": 9.213769718587553, + "learning_rate": 1.425681605111402e-06, + "loss": 0.2418, + "step": 7673 + }, + { + "epoch": 4.166123778501628, + "grad_norm": 14.937646538347469, + "learning_rate": 1.4238723784348873e-06, + "loss": 0.4886, + "step": 7674 + }, + { + "epoch": 4.166666666666667, + "grad_norm": 14.913733879373519, + "learning_rate": 1.4220642124675066e-06, + "loss": 0.549, + "step": 7675 + }, + { + "epoch": 4.167209554831705, + "grad_norm": 13.717360959964097, + "learning_rate": 1.4202571074328941e-06, + "loss": 0.5719, + "step": 7676 + }, + { + "epoch": 4.1677524429967425, + "grad_norm": 16.071560326957982, + "learning_rate": 1.4184510635545568e-06, + "loss": 0.6427, + "step": 7677 + }, + { + "epoch": 4.16829533116178, + "grad_norm": 12.694831131159624, + "learning_rate": 1.4166460810558691e-06, + "loss": 0.502, + "step": 7678 + }, + { + "epoch": 4.168838219326819, + "grad_norm": 9.505941995564902, + "learning_rate": 1.4148421601600749e-06, + "loss": 0.357, + "step": 7679 + }, + { + "epoch": 4.169381107491857, + "grad_norm": 10.441603292556694, + "learning_rate": 1.4130393010902831e-06, + "loss": 0.3607, + "step": 7680 + }, + { + "epoch": 4.1699239956568945, + "grad_norm": 11.694087750701021, + "learning_rate": 1.4112375040694781e-06, + "loss": 0.4678, + "step": 7681 + }, + { + "epoch": 4.170466883821932, + "grad_norm": 14.852563594982858, + "learning_rate": 1.4094367693205057e-06, + "loss": 0.5183, + "step": 7682 + }, + { + "epoch": 4.171009771986971, + "grad_norm": 13.451301475700863, + "learning_rate": 1.4076370970660858e-06, + "loss": 0.402, + "step": 7683 + }, + { + "epoch": 4.171552660152009, + "grad_norm": 12.466975166766947, + "learning_rate": 1.4058384875288034e-06, + "loss": 0.5729, + "step": 7684 + }, + { + "epoch": 4.1720955483170465, + "grad_norm": 14.018971166921462, + "learning_rate": 1.404040940931114e-06, + "loss": 0.3046, + "step": 7685 + }, + { + "epoch": 4.172638436482084, + "grad_norm": 9.441888668504816, + "learning_rate": 1.402244457495341e-06, + "loss": 0.2794, + "step": 7686 + }, + { + "epoch": 4.173181324647123, + "grad_norm": 14.181424515746851, + "learning_rate": 1.4004490374436764e-06, + "loss": 0.584, + "step": 7687 + }, + { + "epoch": 4.173724212812161, + "grad_norm": 10.806677133942568, + "learning_rate": 1.3986546809981783e-06, + "loss": 0.4671, + "step": 7688 + }, + { + "epoch": 4.1742671009771986, + "grad_norm": 13.238385393014816, + "learning_rate": 1.3968613883807814e-06, + "loss": 0.8039, + "step": 7689 + }, + { + "epoch": 4.174809989142236, + "grad_norm": 10.501288736320749, + "learning_rate": 1.395069159813276e-06, + "loss": 0.2945, + "step": 7690 + }, + { + "epoch": 4.175352877307275, + "grad_norm": 15.172510573487276, + "learning_rate": 1.3932779955173349e-06, + "loss": 0.4151, + "step": 7691 + }, + { + "epoch": 4.175895765472313, + "grad_norm": 15.13893230910218, + "learning_rate": 1.3914878957144862e-06, + "loss": 0.6144, + "step": 7692 + }, + { + "epoch": 4.176438653637351, + "grad_norm": 9.703023506384875, + "learning_rate": 1.389698860626133e-06, + "loss": 0.2173, + "step": 7693 + }, + { + "epoch": 4.176981541802388, + "grad_norm": 9.6655026395734, + "learning_rate": 1.3879108904735505e-06, + "loss": 0.3047, + "step": 7694 + }, + { + "epoch": 4.177524429967427, + "grad_norm": 13.323140178847439, + "learning_rate": 1.3861239854778719e-06, + "loss": 0.4717, + "step": 7695 + }, + { + "epoch": 4.178067318132465, + "grad_norm": 9.405943562091032, + "learning_rate": 1.38433814586011e-06, + "loss": 0.2174, + "step": 7696 + }, + { + "epoch": 4.178610206297503, + "grad_norm": 13.771494832808791, + "learning_rate": 1.3825533718411343e-06, + "loss": 0.3924, + "step": 7697 + }, + { + "epoch": 4.17915309446254, + "grad_norm": 15.26443665837414, + "learning_rate": 1.3807696636416956e-06, + "loss": 0.4529, + "step": 7698 + }, + { + "epoch": 4.179695982627579, + "grad_norm": 19.71010753607603, + "learning_rate": 1.3789870214823987e-06, + "loss": 1.155, + "step": 7699 + }, + { + "epoch": 4.180238870792617, + "grad_norm": 10.482042942898147, + "learning_rate": 1.377205445583728e-06, + "loss": 0.2945, + "step": 7700 + }, + { + "epoch": 4.180781758957655, + "grad_norm": 23.55094181120616, + "learning_rate": 1.375424936166032e-06, + "loss": 0.5429, + "step": 7701 + }, + { + "epoch": 4.181324647122692, + "grad_norm": 12.840338456448858, + "learning_rate": 1.3736454934495225e-06, + "loss": 0.5054, + "step": 7702 + }, + { + "epoch": 4.181867535287731, + "grad_norm": 15.318914270528728, + "learning_rate": 1.3718671176542897e-06, + "loss": 0.5164, + "step": 7703 + }, + { + "epoch": 4.182410423452769, + "grad_norm": 12.76929564122516, + "learning_rate": 1.3700898090002802e-06, + "loss": 0.5994, + "step": 7704 + }, + { + "epoch": 4.182953311617807, + "grad_norm": 10.099666037602478, + "learning_rate": 1.368313567707318e-06, + "loss": 0.2488, + "step": 7705 + }, + { + "epoch": 4.183496199782844, + "grad_norm": 14.142637199722767, + "learning_rate": 1.3665383939950916e-06, + "loss": 0.5153, + "step": 7706 + }, + { + "epoch": 4.184039087947883, + "grad_norm": 17.926261733912252, + "learning_rate": 1.3647642880831556e-06, + "loss": 0.647, + "step": 7707 + }, + { + "epoch": 4.184581976112921, + "grad_norm": 17.298123683774026, + "learning_rate": 1.3629912501909347e-06, + "loss": 0.8169, + "step": 7708 + }, + { + "epoch": 4.185124864277959, + "grad_norm": 12.451648500428053, + "learning_rate": 1.3612192805377222e-06, + "loss": 0.3808, + "step": 7709 + }, + { + "epoch": 4.185667752442996, + "grad_norm": 9.322957640674206, + "learning_rate": 1.3594483793426794e-06, + "loss": 0.2851, + "step": 7710 + }, + { + "epoch": 4.186210640608035, + "grad_norm": 13.229946902809969, + "learning_rate": 1.3576785468248265e-06, + "loss": 0.4352, + "step": 7711 + }, + { + "epoch": 4.186753528773073, + "grad_norm": 11.991137980793829, + "learning_rate": 1.3559097832030677e-06, + "loss": 0.3591, + "step": 7712 + }, + { + "epoch": 4.187296416938111, + "grad_norm": 13.197661861619876, + "learning_rate": 1.3541420886961631e-06, + "loss": 0.3863, + "step": 7713 + }, + { + "epoch": 4.187839305103148, + "grad_norm": 15.68139113867389, + "learning_rate": 1.352375463522746e-06, + "loss": 0.6056, + "step": 7714 + }, + { + "epoch": 4.188382193268187, + "grad_norm": 12.308133143281433, + "learning_rate": 1.3506099079013124e-06, + "loss": 0.3132, + "step": 7715 + }, + { + "epoch": 4.188925081433225, + "grad_norm": 11.907660119383362, + "learning_rate": 1.34884542205023e-06, + "loss": 0.3939, + "step": 7716 + }, + { + "epoch": 4.189467969598263, + "grad_norm": 19.521959289662206, + "learning_rate": 1.3470820061877344e-06, + "loss": 0.5875, + "step": 7717 + }, + { + "epoch": 4.1900108577633, + "grad_norm": 12.438243625833383, + "learning_rate": 1.3453196605319274e-06, + "loss": 0.307, + "step": 7718 + }, + { + "epoch": 4.190553745928339, + "grad_norm": 12.048829129615415, + "learning_rate": 1.3435583853007783e-06, + "loss": 0.3221, + "step": 7719 + }, + { + "epoch": 4.191096634093377, + "grad_norm": 10.135648901522302, + "learning_rate": 1.3417981807121239e-06, + "loss": 0.2947, + "step": 7720 + }, + { + "epoch": 4.191639522258415, + "grad_norm": 12.238680241464417, + "learning_rate": 1.34003904698367e-06, + "loss": 0.4503, + "step": 7721 + }, + { + "epoch": 4.192182410423452, + "grad_norm": 15.537341199882544, + "learning_rate": 1.338280984332988e-06, + "loss": 0.6186, + "step": 7722 + }, + { + "epoch": 4.192725298588491, + "grad_norm": 14.855361885159425, + "learning_rate": 1.3365239929775186e-06, + "loss": 0.5375, + "step": 7723 + }, + { + "epoch": 4.193268186753529, + "grad_norm": 12.128751807129769, + "learning_rate": 1.3347680731345703e-06, + "loss": 0.3844, + "step": 7724 + }, + { + "epoch": 4.193811074918567, + "grad_norm": 14.352544667740425, + "learning_rate": 1.3330132250213158e-06, + "loss": 0.3476, + "step": 7725 + }, + { + "epoch": 4.194353963083604, + "grad_norm": 11.233393992477202, + "learning_rate": 1.331259448854798e-06, + "loss": 0.3616, + "step": 7726 + }, + { + "epoch": 4.194896851248643, + "grad_norm": 16.82629514482139, + "learning_rate": 1.3295067448519305e-06, + "loss": 0.5184, + "step": 7727 + }, + { + "epoch": 4.195439739413681, + "grad_norm": 16.36127203758974, + "learning_rate": 1.3277551132294842e-06, + "loss": 0.3893, + "step": 7728 + }, + { + "epoch": 4.195982627578719, + "grad_norm": 12.108570761810808, + "learning_rate": 1.32600455420411e-06, + "loss": 0.4403, + "step": 7729 + }, + { + "epoch": 4.196525515743756, + "grad_norm": 9.120521528279125, + "learning_rate": 1.3242550679923162e-06, + "loss": 0.2306, + "step": 7730 + }, + { + "epoch": 4.197068403908795, + "grad_norm": 10.947328867337708, + "learning_rate": 1.3225066548104814e-06, + "loss": 0.4367, + "step": 7731 + }, + { + "epoch": 4.197611292073833, + "grad_norm": 13.012382022226626, + "learning_rate": 1.3207593148748543e-06, + "loss": 0.3955, + "step": 7732 + }, + { + "epoch": 4.198154180238871, + "grad_norm": 12.631500933000643, + "learning_rate": 1.3190130484015452e-06, + "loss": 0.3416, + "step": 7733 + }, + { + "epoch": 4.198697068403908, + "grad_norm": 10.924514493419068, + "learning_rate": 1.3172678556065421e-06, + "loss": 0.2377, + "step": 7734 + }, + { + "epoch": 4.199239956568947, + "grad_norm": 17.37092245490247, + "learning_rate": 1.315523736705684e-06, + "loss": 0.3534, + "step": 7735 + }, + { + "epoch": 4.199782844733985, + "grad_norm": 10.606509484945631, + "learning_rate": 1.3137806919146967e-06, + "loss": 0.367, + "step": 7736 + }, + { + "epoch": 4.200325732899023, + "grad_norm": 12.557115475350406, + "learning_rate": 1.3120387214491514e-06, + "loss": 0.3562, + "step": 7737 + }, + { + "epoch": 4.2008686210640604, + "grad_norm": 12.911761209858065, + "learning_rate": 1.310297825524507e-06, + "loss": 0.4865, + "step": 7738 + }, + { + "epoch": 4.201411509229099, + "grad_norm": 16.757054584318308, + "learning_rate": 1.3085580043560763e-06, + "loss": 0.4604, + "step": 7739 + }, + { + "epoch": 4.201954397394137, + "grad_norm": 11.675532522546053, + "learning_rate": 1.3068192581590433e-06, + "loss": 0.403, + "step": 7740 + }, + { + "epoch": 4.202497285559175, + "grad_norm": 16.159527030741106, + "learning_rate": 1.3050815871484622e-06, + "loss": 0.4909, + "step": 7741 + }, + { + "epoch": 4.2030401737242125, + "grad_norm": 16.38078706465185, + "learning_rate": 1.3033449915392437e-06, + "loss": 0.6048, + "step": 7742 + }, + { + "epoch": 4.203583061889251, + "grad_norm": 12.471147411789133, + "learning_rate": 1.301609471546179e-06, + "loss": 0.5796, + "step": 7743 + }, + { + "epoch": 4.204125950054289, + "grad_norm": 12.706419284404552, + "learning_rate": 1.2998750273839179e-06, + "loss": 0.2934, + "step": 7744 + }, + { + "epoch": 4.204668838219327, + "grad_norm": 11.938920053403065, + "learning_rate": 1.2981416592669794e-06, + "loss": 0.4079, + "step": 7745 + }, + { + "epoch": 4.2052117263843645, + "grad_norm": 14.656307060991027, + "learning_rate": 1.2964093674097488e-06, + "loss": 0.4356, + "step": 7746 + }, + { + "epoch": 4.205754614549403, + "grad_norm": 12.993216203856585, + "learning_rate": 1.2946781520264795e-06, + "loss": 0.3562, + "step": 7747 + }, + { + "epoch": 4.206297502714441, + "grad_norm": 14.182239777867444, + "learning_rate": 1.292948013331291e-06, + "loss": 0.4632, + "step": 7748 + }, + { + "epoch": 4.206840390879479, + "grad_norm": 17.007767789893737, + "learning_rate": 1.291218951538169e-06, + "loss": 0.5598, + "step": 7749 + }, + { + "epoch": 4.2073832790445165, + "grad_norm": 10.545945938754782, + "learning_rate": 1.2894909668609657e-06, + "loss": 0.3098, + "step": 7750 + }, + { + "epoch": 4.207926167209555, + "grad_norm": 14.053283729714321, + "learning_rate": 1.2877640595134033e-06, + "loss": 0.5349, + "step": 7751 + }, + { + "epoch": 4.208469055374593, + "grad_norm": 11.754708732243547, + "learning_rate": 1.2860382297090669e-06, + "loss": 0.2941, + "step": 7752 + }, + { + "epoch": 4.209011943539631, + "grad_norm": 13.75780850163015, + "learning_rate": 1.2843134776614108e-06, + "loss": 0.4709, + "step": 7753 + }, + { + "epoch": 4.2095548317046685, + "grad_norm": 14.679629107183144, + "learning_rate": 1.2825898035837537e-06, + "loss": 0.4888, + "step": 7754 + }, + { + "epoch": 4.210097719869707, + "grad_norm": 11.061342428653278, + "learning_rate": 1.2808672076892837e-06, + "loss": 0.2758, + "step": 7755 + }, + { + "epoch": 4.210640608034745, + "grad_norm": 11.723468756168664, + "learning_rate": 1.2791456901910526e-06, + "loss": 0.4185, + "step": 7756 + }, + { + "epoch": 4.211183496199783, + "grad_norm": 10.0084471239469, + "learning_rate": 1.2774252513019825e-06, + "loss": 0.2693, + "step": 7757 + }, + { + "epoch": 4.2117263843648205, + "grad_norm": 14.164307521757856, + "learning_rate": 1.2757058912348596e-06, + "loss": 0.5625, + "step": 7758 + }, + { + "epoch": 4.212269272529859, + "grad_norm": 15.322157151421756, + "learning_rate": 1.2739876102023352e-06, + "loss": 0.4828, + "step": 7759 + }, + { + "epoch": 4.212812160694897, + "grad_norm": 17.051847694097003, + "learning_rate": 1.2722704084169313e-06, + "loss": 0.5323, + "step": 7760 + }, + { + "epoch": 4.213355048859935, + "grad_norm": 11.059167577716101, + "learning_rate": 1.2705542860910325e-06, + "loss": 0.2758, + "step": 7761 + }, + { + "epoch": 4.2138979370249725, + "grad_norm": 11.89278028580855, + "learning_rate": 1.2688392434368924e-06, + "loss": 0.315, + "step": 7762 + }, + { + "epoch": 4.214440825190011, + "grad_norm": 10.03832961212838, + "learning_rate": 1.2671252806666312e-06, + "loss": 0.316, + "step": 7763 + }, + { + "epoch": 4.214983713355049, + "grad_norm": 14.358273720821714, + "learning_rate": 1.2654123979922294e-06, + "loss": 0.447, + "step": 7764 + }, + { + "epoch": 4.215526601520087, + "grad_norm": 13.267216152626093, + "learning_rate": 1.2637005956255489e-06, + "loss": 0.3846, + "step": 7765 + }, + { + "epoch": 4.2160694896851245, + "grad_norm": 10.109215242139403, + "learning_rate": 1.2619898737782977e-06, + "loss": 0.2809, + "step": 7766 + }, + { + "epoch": 4.216612377850163, + "grad_norm": 16.400315992543526, + "learning_rate": 1.2602802326620688e-06, + "loss": 0.632, + "step": 7767 + }, + { + "epoch": 4.217155266015201, + "grad_norm": 11.109933889769897, + "learning_rate": 1.2585716724883057e-06, + "loss": 0.484, + "step": 7768 + }, + { + "epoch": 4.217698154180239, + "grad_norm": 14.07156919017978, + "learning_rate": 1.2568641934683345e-06, + "loss": 0.7668, + "step": 7769 + }, + { + "epoch": 4.2182410423452765, + "grad_norm": 11.176348337381958, + "learning_rate": 1.255157795813332e-06, + "loss": 0.4213, + "step": 7770 + }, + { + "epoch": 4.218783930510315, + "grad_norm": 16.063587698095667, + "learning_rate": 1.2534524797343494e-06, + "loss": 0.6027, + "step": 7771 + }, + { + "epoch": 4.219326818675353, + "grad_norm": 15.762904336294534, + "learning_rate": 1.2517482454423068e-06, + "loss": 0.6919, + "step": 7772 + }, + { + "epoch": 4.219869706840391, + "grad_norm": 11.914370317943318, + "learning_rate": 1.2500450931479801e-06, + "loss": 0.4781, + "step": 7773 + }, + { + "epoch": 4.2204125950054285, + "grad_norm": 17.015123181348883, + "learning_rate": 1.2483430230620264e-06, + "loss": 0.31, + "step": 7774 + }, + { + "epoch": 4.220955483170467, + "grad_norm": 12.337173743701737, + "learning_rate": 1.2466420353949515e-06, + "loss": 0.3487, + "step": 7775 + }, + { + "epoch": 4.221498371335505, + "grad_norm": 9.196709204828075, + "learning_rate": 1.2449421303571418e-06, + "loss": 0.2396, + "step": 7776 + }, + { + "epoch": 4.222041259500543, + "grad_norm": 14.267880372674293, + "learning_rate": 1.2432433081588424e-06, + "loss": 0.7252, + "step": 7777 + }, + { + "epoch": 4.2225841476655805, + "grad_norm": 10.078600378968735, + "learning_rate": 1.2415455690101685e-06, + "loss": 0.3152, + "step": 7778 + }, + { + "epoch": 4.223127035830619, + "grad_norm": 15.372509732415313, + "learning_rate": 1.2398489131210989e-06, + "loss": 0.5178, + "step": 7779 + }, + { + "epoch": 4.223669923995657, + "grad_norm": 11.547384831594652, + "learning_rate": 1.2381533407014722e-06, + "loss": 0.3999, + "step": 7780 + }, + { + "epoch": 4.224212812160695, + "grad_norm": 11.21628120223001, + "learning_rate": 1.23645885196101e-06, + "loss": 0.2856, + "step": 7781 + }, + { + "epoch": 4.2247557003257326, + "grad_norm": 12.171889903747351, + "learning_rate": 1.2347654471092797e-06, + "loss": 0.3303, + "step": 7782 + }, + { + "epoch": 4.225298588490771, + "grad_norm": 11.210591190805236, + "learning_rate": 1.2330731263557293e-06, + "loss": 0.321, + "step": 7783 + }, + { + "epoch": 4.225841476655809, + "grad_norm": 11.56900032090966, + "learning_rate": 1.2313818899096685e-06, + "loss": 0.3416, + "step": 7784 + }, + { + "epoch": 4.226384364820847, + "grad_norm": 10.580053290763194, + "learning_rate": 1.2296917379802709e-06, + "loss": 0.3201, + "step": 7785 + }, + { + "epoch": 4.226927252985885, + "grad_norm": 18.553406932221385, + "learning_rate": 1.2280026707765758e-06, + "loss": 0.4094, + "step": 7786 + }, + { + "epoch": 4.227470141150923, + "grad_norm": 11.769072597193505, + "learning_rate": 1.2263146885074917e-06, + "loss": 0.3517, + "step": 7787 + }, + { + "epoch": 4.228013029315961, + "grad_norm": 14.383741980090374, + "learning_rate": 1.2246277913817895e-06, + "loss": 0.4665, + "step": 7788 + }, + { + "epoch": 4.228555917480999, + "grad_norm": 13.049438484607592, + "learning_rate": 1.2229419796081077e-06, + "loss": 0.4222, + "step": 7789 + }, + { + "epoch": 4.229098805646037, + "grad_norm": 9.676387763683955, + "learning_rate": 1.2212572533949506e-06, + "loss": 0.3977, + "step": 7790 + }, + { + "epoch": 4.229641693811075, + "grad_norm": 20.33530097040713, + "learning_rate": 1.2195736129506885e-06, + "loss": 0.446, + "step": 7791 + }, + { + "epoch": 4.230184581976113, + "grad_norm": 14.282100526459736, + "learning_rate": 1.2178910584835556e-06, + "loss": 0.3286, + "step": 7792 + }, + { + "epoch": 4.230727470141151, + "grad_norm": 19.14182283529482, + "learning_rate": 1.2162095902016525e-06, + "loss": 0.6531, + "step": 7793 + }, + { + "epoch": 4.231270358306189, + "grad_norm": 9.949859543104305, + "learning_rate": 1.2145292083129467e-06, + "loss": 0.3108, + "step": 7794 + }, + { + "epoch": 4.231813246471227, + "grad_norm": 12.280893090679601, + "learning_rate": 1.2128499130252714e-06, + "loss": 0.3321, + "step": 7795 + }, + { + "epoch": 4.232356134636265, + "grad_norm": 13.441725051837208, + "learning_rate": 1.2111717045463223e-06, + "loss": 0.4196, + "step": 7796 + }, + { + "epoch": 4.232899022801303, + "grad_norm": 15.03461419306544, + "learning_rate": 1.2094945830836625e-06, + "loss": 0.3983, + "step": 7797 + }, + { + "epoch": 4.233441910966341, + "grad_norm": 16.773668359582217, + "learning_rate": 1.207818548844727e-06, + "loss": 0.386, + "step": 7798 + }, + { + "epoch": 4.233984799131379, + "grad_norm": 11.119171257862224, + "learning_rate": 1.2061436020368033e-06, + "loss": 0.4611, + "step": 7799 + }, + { + "epoch": 4.234527687296417, + "grad_norm": 8.229976665110055, + "learning_rate": 1.204469742867055e-06, + "loss": 0.2615, + "step": 7800 + }, + { + "epoch": 4.235070575461455, + "grad_norm": 13.435979628871456, + "learning_rate": 1.2027969715425058e-06, + "loss": 0.4595, + "step": 7801 + }, + { + "epoch": 4.235613463626493, + "grad_norm": 17.70018909852596, + "learning_rate": 1.2011252882700486e-06, + "loss": 0.56, + "step": 7802 + }, + { + "epoch": 4.236156351791531, + "grad_norm": 13.260555683605508, + "learning_rate": 1.1994546932564388e-06, + "loss": 0.384, + "step": 7803 + }, + { + "epoch": 4.236699239956569, + "grad_norm": 10.203419005145452, + "learning_rate": 1.1977851867082968e-06, + "loss": 0.3906, + "step": 7804 + }, + { + "epoch": 4.237242128121607, + "grad_norm": 19.37091442524948, + "learning_rate": 1.1961167688321151e-06, + "loss": 0.5527, + "step": 7805 + }, + { + "epoch": 4.237785016286645, + "grad_norm": 17.208395947554312, + "learning_rate": 1.1944494398342388e-06, + "loss": 0.597, + "step": 7806 + }, + { + "epoch": 4.238327904451683, + "grad_norm": 12.251023694428005, + "learning_rate": 1.1927831999208938e-06, + "loss": 0.4034, + "step": 7807 + }, + { + "epoch": 4.238870792616721, + "grad_norm": 16.49490715109582, + "learning_rate": 1.1911180492981578e-06, + "loss": 0.6092, + "step": 7808 + }, + { + "epoch": 4.239413680781759, + "grad_norm": 14.390923091386458, + "learning_rate": 1.189453988171978e-06, + "loss": 0.5611, + "step": 7809 + }, + { + "epoch": 4.239956568946797, + "grad_norm": 16.171314155557898, + "learning_rate": 1.1877910167481754e-06, + "loss": 0.3232, + "step": 7810 + }, + { + "epoch": 4.240499457111835, + "grad_norm": 15.308750990763075, + "learning_rate": 1.18612913523242e-06, + "loss": 0.3301, + "step": 7811 + }, + { + "epoch": 4.241042345276873, + "grad_norm": 11.75895557202086, + "learning_rate": 1.1844683438302652e-06, + "loss": 0.4324, + "step": 7812 + }, + { + "epoch": 4.241585233441911, + "grad_norm": 11.620344969621105, + "learning_rate": 1.1828086427471107e-06, + "loss": 0.5246, + "step": 7813 + }, + { + "epoch": 4.242128121606949, + "grad_norm": 11.17860914452438, + "learning_rate": 1.1811500321882374e-06, + "loss": 0.4626, + "step": 7814 + }, + { + "epoch": 4.242671009771987, + "grad_norm": 12.479770513980384, + "learning_rate": 1.1794925123587841e-06, + "loss": 0.375, + "step": 7815 + }, + { + "epoch": 4.243213897937025, + "grad_norm": 13.677272569244249, + "learning_rate": 1.1778360834637549e-06, + "loss": 0.3313, + "step": 7816 + }, + { + "epoch": 4.243756786102063, + "grad_norm": 11.135594932532166, + "learning_rate": 1.1761807457080187e-06, + "loss": 0.4987, + "step": 7817 + }, + { + "epoch": 4.244299674267101, + "grad_norm": 16.738993364102154, + "learning_rate": 1.1745264992963112e-06, + "loss": 0.367, + "step": 7818 + }, + { + "epoch": 4.244842562432139, + "grad_norm": 12.305866749951548, + "learning_rate": 1.1728733444332342e-06, + "loss": 0.3499, + "step": 7819 + }, + { + "epoch": 4.245385450597177, + "grad_norm": 13.571484126012804, + "learning_rate": 1.1712212813232448e-06, + "loss": 0.4086, + "step": 7820 + }, + { + "epoch": 4.245928338762215, + "grad_norm": 18.75127331038876, + "learning_rate": 1.1695703101706823e-06, + "loss": 0.6246, + "step": 7821 + }, + { + "epoch": 4.246471226927253, + "grad_norm": 17.48505499787076, + "learning_rate": 1.1679204311797365e-06, + "loss": 0.5487, + "step": 7822 + }, + { + "epoch": 4.247014115092291, + "grad_norm": 9.684180280120152, + "learning_rate": 1.1662716445544676e-06, + "loss": 0.2527, + "step": 7823 + }, + { + "epoch": 4.247557003257329, + "grad_norm": 13.169958031763949, + "learning_rate": 1.164623950498801e-06, + "loss": 0.3898, + "step": 7824 + }, + { + "epoch": 4.248099891422367, + "grad_norm": 11.019724381747643, + "learning_rate": 1.1629773492165264e-06, + "loss": 0.3471, + "step": 7825 + }, + { + "epoch": 4.248642779587405, + "grad_norm": 9.968852095122674, + "learning_rate": 1.1613318409112962e-06, + "loss": 0.3911, + "step": 7826 + }, + { + "epoch": 4.249185667752443, + "grad_norm": 15.553115617722527, + "learning_rate": 1.159687425786632e-06, + "loss": 0.425, + "step": 7827 + }, + { + "epoch": 4.249728555917481, + "grad_norm": 12.82652568103273, + "learning_rate": 1.1580441040459146e-06, + "loss": 0.436, + "step": 7828 + }, + { + "epoch": 4.250271444082519, + "grad_norm": 10.72479783442799, + "learning_rate": 1.1564018758923956e-06, + "loss": 0.3967, + "step": 7829 + }, + { + "epoch": 4.250814332247557, + "grad_norm": 17.35073007540839, + "learning_rate": 1.1547607415291874e-06, + "loss": 0.473, + "step": 7830 + }, + { + "epoch": 4.251357220412595, + "grad_norm": 11.61546854660444, + "learning_rate": 1.153120701159267e-06, + "loss": 0.3997, + "step": 7831 + }, + { + "epoch": 4.251900108577633, + "grad_norm": 14.988521210943322, + "learning_rate": 1.1514817549854784e-06, + "loss": 0.5253, + "step": 7832 + }, + { + "epoch": 4.252442996742671, + "grad_norm": 12.668110351197043, + "learning_rate": 1.149843903210529e-06, + "loss": 0.4776, + "step": 7833 + }, + { + "epoch": 4.252985884907709, + "grad_norm": 15.448405948912091, + "learning_rate": 1.1482071460369903e-06, + "loss": 0.4487, + "step": 7834 + }, + { + "epoch": 4.253528773072747, + "grad_norm": 13.582390046184774, + "learning_rate": 1.146571483667298e-06, + "loss": 0.3464, + "step": 7835 + }, + { + "epoch": 4.254071661237785, + "grad_norm": 16.689486575643958, + "learning_rate": 1.1449369163037582e-06, + "loss": 0.4516, + "step": 7836 + }, + { + "epoch": 4.254614549402823, + "grad_norm": 12.868709622460063, + "learning_rate": 1.1433034441485303e-06, + "loss": 0.2851, + "step": 7837 + }, + { + "epoch": 4.255157437567861, + "grad_norm": 8.159166476090338, + "learning_rate": 1.1416710674036512e-06, + "loss": 0.1686, + "step": 7838 + }, + { + "epoch": 4.255700325732899, + "grad_norm": 10.431332020179449, + "learning_rate": 1.1400397862710112e-06, + "loss": 0.3455, + "step": 7839 + }, + { + "epoch": 4.256243213897937, + "grad_norm": 12.657036094463699, + "learning_rate": 1.1384096009523705e-06, + "loss": 0.3586, + "step": 7840 + }, + { + "epoch": 4.256786102062975, + "grad_norm": 17.536752353552888, + "learning_rate": 1.1367805116493534e-06, + "loss": 0.5641, + "step": 7841 + }, + { + "epoch": 4.257328990228013, + "grad_norm": 12.422330381694959, + "learning_rate": 1.135152518563447e-06, + "loss": 0.5808, + "step": 7842 + }, + { + "epoch": 4.257871878393051, + "grad_norm": 14.406429659207616, + "learning_rate": 1.13352562189601e-06, + "loss": 0.4333, + "step": 7843 + }, + { + "epoch": 4.258414766558089, + "grad_norm": 13.093016383546724, + "learning_rate": 1.1318998218482501e-06, + "loss": 0.4737, + "step": 7844 + }, + { + "epoch": 4.258957654723127, + "grad_norm": 12.327912324997381, + "learning_rate": 1.130275118621259e-06, + "loss": 0.3371, + "step": 7845 + }, + { + "epoch": 4.259500542888165, + "grad_norm": 12.784200365328555, + "learning_rate": 1.1286515124159726e-06, + "loss": 0.4967, + "step": 7846 + }, + { + "epoch": 4.260043431053203, + "grad_norm": 14.609662199743113, + "learning_rate": 1.1270290034332076e-06, + "loss": 0.5266, + "step": 7847 + }, + { + "epoch": 4.260586319218241, + "grad_norm": 16.132913961447098, + "learning_rate": 1.1254075918736395e-06, + "loss": 0.7818, + "step": 7848 + }, + { + "epoch": 4.261129207383279, + "grad_norm": 12.362451065508255, + "learning_rate": 1.1237872779378013e-06, + "loss": 0.3534, + "step": 7849 + }, + { + "epoch": 4.261672095548317, + "grad_norm": 11.787990488474275, + "learning_rate": 1.1221680618261022e-06, + "loss": 0.3316, + "step": 7850 + }, + { + "epoch": 4.262214983713355, + "grad_norm": 16.520849175369758, + "learning_rate": 1.1205499437388024e-06, + "loss": 0.6686, + "step": 7851 + }, + { + "epoch": 4.262757871878393, + "grad_norm": 15.087215559431693, + "learning_rate": 1.1189329238760416e-06, + "loss": 0.4386, + "step": 7852 + }, + { + "epoch": 4.263300760043431, + "grad_norm": 14.92562475156193, + "learning_rate": 1.1173170024378067e-06, + "loss": 0.6797, + "step": 7853 + }, + { + "epoch": 4.263843648208469, + "grad_norm": 10.695763423794212, + "learning_rate": 1.1157021796239631e-06, + "loss": 0.2609, + "step": 7854 + }, + { + "epoch": 4.264386536373507, + "grad_norm": 13.337827871285958, + "learning_rate": 1.1140884556342346e-06, + "loss": 0.6186, + "step": 7855 + }, + { + "epoch": 4.264929424538545, + "grad_norm": 13.122182163353113, + "learning_rate": 1.1124758306682081e-06, + "loss": 0.518, + "step": 7856 + }, + { + "epoch": 4.265472312703583, + "grad_norm": 8.714454332149657, + "learning_rate": 1.1108643049253364e-06, + "loss": 0.3084, + "step": 7857 + }, + { + "epoch": 4.266015200868621, + "grad_norm": 12.760955138975337, + "learning_rate": 1.1092538786049311e-06, + "loss": 0.5716, + "step": 7858 + }, + { + "epoch": 4.266558089033659, + "grad_norm": 17.235261430649693, + "learning_rate": 1.1076445519061784e-06, + "loss": 0.4657, + "step": 7859 + }, + { + "epoch": 4.267100977198697, + "grad_norm": 14.044151036119151, + "learning_rate": 1.1060363250281192e-06, + "loss": 0.6461, + "step": 7860 + }, + { + "epoch": 4.267643865363735, + "grad_norm": 12.741420750608738, + "learning_rate": 1.104429198169662e-06, + "loss": 0.4156, + "step": 7861 + }, + { + "epoch": 4.268186753528773, + "grad_norm": 16.38667997198235, + "learning_rate": 1.10282317152958e-06, + "loss": 0.4167, + "step": 7862 + }, + { + "epoch": 4.268729641693811, + "grad_norm": 12.512768079843084, + "learning_rate": 1.1012182453065069e-06, + "loss": 0.4918, + "step": 7863 + }, + { + "epoch": 4.269272529858849, + "grad_norm": 13.720175584312965, + "learning_rate": 1.0996144196989444e-06, + "loss": 0.3484, + "step": 7864 + }, + { + "epoch": 4.269815418023887, + "grad_norm": 12.963348508780628, + "learning_rate": 1.0980116949052554e-06, + "loss": 0.4639, + "step": 7865 + }, + { + "epoch": 4.270358306188925, + "grad_norm": 16.8876044604694, + "learning_rate": 1.096410071123668e-06, + "loss": 0.4278, + "step": 7866 + }, + { + "epoch": 4.270901194353963, + "grad_norm": 14.644807785059461, + "learning_rate": 1.0948095485522725e-06, + "loss": 0.3409, + "step": 7867 + }, + { + "epoch": 4.271444082519001, + "grad_norm": 13.719922365450167, + "learning_rate": 1.093210127389025e-06, + "loss": 0.5096, + "step": 7868 + }, + { + "epoch": 4.271986970684039, + "grad_norm": 15.219211444329652, + "learning_rate": 1.091611807831745e-06, + "loss": 0.4472, + "step": 7869 + }, + { + "epoch": 4.272529858849077, + "grad_norm": 12.392349389977586, + "learning_rate": 1.0900145900781133e-06, + "loss": 0.2404, + "step": 7870 + }, + { + "epoch": 4.273072747014115, + "grad_norm": 15.937024579734995, + "learning_rate": 1.0884184743256788e-06, + "loss": 0.5215, + "step": 7871 + }, + { + "epoch": 4.273615635179153, + "grad_norm": 11.853999121791691, + "learning_rate": 1.0868234607718498e-06, + "loss": 0.4883, + "step": 7872 + }, + { + "epoch": 4.274158523344191, + "grad_norm": 13.690567615009499, + "learning_rate": 1.0852295496139008e-06, + "loss": 0.5029, + "step": 7873 + }, + { + "epoch": 4.274701411509229, + "grad_norm": 9.345078157701355, + "learning_rate": 1.083636741048969e-06, + "loss": 0.34, + "step": 7874 + }, + { + "epoch": 4.2752442996742674, + "grad_norm": 10.355626815319667, + "learning_rate": 1.082045035274053e-06, + "loss": 0.2761, + "step": 7875 + }, + { + "epoch": 4.275787187839305, + "grad_norm": 9.570407579073933, + "learning_rate": 1.0804544324860245e-06, + "loss": 0.2851, + "step": 7876 + }, + { + "epoch": 4.276330076004343, + "grad_norm": 12.52101649250042, + "learning_rate": 1.0788649328816059e-06, + "loss": 0.4681, + "step": 7877 + }, + { + "epoch": 4.276872964169381, + "grad_norm": 10.848327417784757, + "learning_rate": 1.0772765366573902e-06, + "loss": 0.4557, + "step": 7878 + }, + { + "epoch": 4.2774158523344195, + "grad_norm": 14.851238098236974, + "learning_rate": 1.0756892440098333e-06, + "loss": 0.3483, + "step": 7879 + }, + { + "epoch": 4.277958740499457, + "grad_norm": 10.516600101157945, + "learning_rate": 1.0741030551352517e-06, + "loss": 0.2466, + "step": 7880 + }, + { + "epoch": 4.278501628664495, + "grad_norm": 13.065799904684205, + "learning_rate": 1.0725179702298338e-06, + "loss": 0.3965, + "step": 7881 + }, + { + "epoch": 4.279044516829533, + "grad_norm": 11.826945310503726, + "learning_rate": 1.0709339894896186e-06, + "loss": 0.3297, + "step": 7882 + }, + { + "epoch": 4.2795874049945715, + "grad_norm": 12.014578152352707, + "learning_rate": 1.0693511131105216e-06, + "loss": 0.6114, + "step": 7883 + }, + { + "epoch": 4.280130293159609, + "grad_norm": 13.074176249077697, + "learning_rate": 1.0677693412883082e-06, + "loss": 0.4467, + "step": 7884 + }, + { + "epoch": 4.280673181324647, + "grad_norm": 16.7050290896531, + "learning_rate": 1.0661886742186212e-06, + "loss": 0.7551, + "step": 7885 + }, + { + "epoch": 4.281216069489685, + "grad_norm": 9.762649259962595, + "learning_rate": 1.064609112096956e-06, + "loss": 0.3769, + "step": 7886 + }, + { + "epoch": 4.2817589576547235, + "grad_norm": 13.322662274779637, + "learning_rate": 1.0630306551186775e-06, + "loss": 0.6132, + "step": 7887 + }, + { + "epoch": 4.282301845819761, + "grad_norm": 14.512051944801728, + "learning_rate": 1.061453303479013e-06, + "loss": 0.2763, + "step": 7888 + }, + { + "epoch": 4.282844733984799, + "grad_norm": 11.38093344044884, + "learning_rate": 1.0598770573730465e-06, + "loss": 0.4843, + "step": 7889 + }, + { + "epoch": 4.283387622149837, + "grad_norm": 12.749261916645077, + "learning_rate": 1.0583019169957377e-06, + "loss": 0.3451, + "step": 7890 + }, + { + "epoch": 4.2839305103148755, + "grad_norm": 11.929936221189454, + "learning_rate": 1.0567278825418937e-06, + "loss": 0.7333, + "step": 7891 + }, + { + "epoch": 4.284473398479913, + "grad_norm": 7.644818532912615, + "learning_rate": 1.0551549542062023e-06, + "loss": 0.2397, + "step": 7892 + }, + { + "epoch": 4.285016286644951, + "grad_norm": 13.22419891221089, + "learning_rate": 1.0535831321832003e-06, + "loss": 0.348, + "step": 7893 + }, + { + "epoch": 4.285559174809989, + "grad_norm": 16.960704066599227, + "learning_rate": 1.0520124166672964e-06, + "loss": 0.5281, + "step": 7894 + }, + { + "epoch": 4.2861020629750275, + "grad_norm": 13.957960998474602, + "learning_rate": 1.0504428078527573e-06, + "loss": 0.4479, + "step": 7895 + }, + { + "epoch": 4.286644951140065, + "grad_norm": 10.380635061146368, + "learning_rate": 1.0488743059337148e-06, + "loss": 0.2593, + "step": 7896 + }, + { + "epoch": 4.287187839305103, + "grad_norm": 14.044042926670437, + "learning_rate": 1.0473069111041644e-06, + "loss": 0.3819, + "step": 7897 + }, + { + "epoch": 4.287730727470141, + "grad_norm": 11.351044141908234, + "learning_rate": 1.0457406235579638e-06, + "loss": 0.3519, + "step": 7898 + }, + { + "epoch": 4.2882736156351795, + "grad_norm": 15.23295931267919, + "learning_rate": 1.044175443488834e-06, + "loss": 0.4457, + "step": 7899 + }, + { + "epoch": 4.288816503800217, + "grad_norm": 15.871156671188771, + "learning_rate": 1.0426113710903586e-06, + "loss": 0.4451, + "step": 7900 + }, + { + "epoch": 4.289359391965255, + "grad_norm": 15.223931858136798, + "learning_rate": 1.0410484065559845e-06, + "loss": 0.7813, + "step": 7901 + }, + { + "epoch": 4.289902280130293, + "grad_norm": 17.47270174804404, + "learning_rate": 1.039486550079023e-06, + "loss": 0.6555, + "step": 7902 + }, + { + "epoch": 4.2904451682953315, + "grad_norm": 14.66163351293163, + "learning_rate": 1.0379258018526451e-06, + "loss": 0.4451, + "step": 7903 + }, + { + "epoch": 4.290988056460369, + "grad_norm": 14.775375275089438, + "learning_rate": 1.0363661620698872e-06, + "loss": 0.5579, + "step": 7904 + }, + { + "epoch": 4.291530944625407, + "grad_norm": 17.26244179697533, + "learning_rate": 1.0348076309236487e-06, + "loss": 0.436, + "step": 7905 + }, + { + "epoch": 4.292073832790445, + "grad_norm": 11.967809939669856, + "learning_rate": 1.0332502086066909e-06, + "loss": 0.251, + "step": 7906 + }, + { + "epoch": 4.2926167209554835, + "grad_norm": 18.886544353437575, + "learning_rate": 1.0316938953116374e-06, + "loss": 0.7729, + "step": 7907 + }, + { + "epoch": 4.293159609120521, + "grad_norm": 16.082653194833412, + "learning_rate": 1.0301386912309763e-06, + "loss": 0.6706, + "step": 7908 + }, + { + "epoch": 4.293702497285559, + "grad_norm": 10.115336308458334, + "learning_rate": 1.0285845965570584e-06, + "loss": 0.238, + "step": 7909 + }, + { + "epoch": 4.294245385450597, + "grad_norm": 10.909597072320835, + "learning_rate": 1.0270316114820954e-06, + "loss": 0.2976, + "step": 7910 + }, + { + "epoch": 4.2947882736156355, + "grad_norm": 13.390757276403917, + "learning_rate": 1.025479736198164e-06, + "loss": 0.8118, + "step": 7911 + }, + { + "epoch": 4.295331161780673, + "grad_norm": 10.149705690991324, + "learning_rate": 1.0239289708972022e-06, + "loss": 0.3365, + "step": 7912 + }, + { + "epoch": 4.295874049945711, + "grad_norm": 14.439616764678872, + "learning_rate": 1.0223793157710083e-06, + "loss": 0.4143, + "step": 7913 + }, + { + "epoch": 4.296416938110749, + "grad_norm": 12.606369995049853, + "learning_rate": 1.020830771011253e-06, + "loss": 0.4796, + "step": 7914 + }, + { + "epoch": 4.2969598262757875, + "grad_norm": 12.593118847476834, + "learning_rate": 1.0192833368094556e-06, + "loss": 0.4738, + "step": 7915 + }, + { + "epoch": 4.297502714440825, + "grad_norm": 16.435528985746867, + "learning_rate": 1.0177370133570108e-06, + "loss": 0.4431, + "step": 7916 + }, + { + "epoch": 4.298045602605863, + "grad_norm": 10.261492535138068, + "learning_rate": 1.0161918008451665e-06, + "loss": 0.3084, + "step": 7917 + }, + { + "epoch": 4.298588490770901, + "grad_norm": 19.67761715432872, + "learning_rate": 1.0146476994650367e-06, + "loss": 0.4107, + "step": 7918 + }, + { + "epoch": 4.2991313789359396, + "grad_norm": 12.628487019222309, + "learning_rate": 1.013104709407603e-06, + "loss": 0.3636, + "step": 7919 + }, + { + "epoch": 4.299674267100977, + "grad_norm": 8.48683461551969, + "learning_rate": 1.0115628308636981e-06, + "loss": 0.2398, + "step": 7920 + }, + { + "epoch": 4.300217155266015, + "grad_norm": 10.946283608256392, + "learning_rate": 1.010022064024032e-06, + "loss": 0.3367, + "step": 7921 + }, + { + "epoch": 4.300760043431053, + "grad_norm": 11.690515621064748, + "learning_rate": 1.0084824090791612e-06, + "loss": 0.3374, + "step": 7922 + }, + { + "epoch": 4.301302931596092, + "grad_norm": 13.693996659420456, + "learning_rate": 1.0069438662195197e-06, + "loss": 0.5506, + "step": 7923 + }, + { + "epoch": 4.301845819761129, + "grad_norm": 10.325683049665914, + "learning_rate": 1.0054064356353898e-06, + "loss": 0.3755, + "step": 7924 + }, + { + "epoch": 4.302388707926167, + "grad_norm": 10.906414174966567, + "learning_rate": 1.0038701175169295e-06, + "loss": 0.3714, + "step": 7925 + }, + { + "epoch": 4.302931596091205, + "grad_norm": 11.218531762718778, + "learning_rate": 1.0023349120541526e-06, + "loss": 0.256, + "step": 7926 + }, + { + "epoch": 4.303474484256244, + "grad_norm": 11.942829819879181, + "learning_rate": 1.00080081943693e-06, + "loss": 0.4458, + "step": 7927 + }, + { + "epoch": 4.304017372421281, + "grad_norm": 10.47979684406596, + "learning_rate": 9.992678398550082e-07, + "loss": 0.2921, + "step": 7928 + }, + { + "epoch": 4.304560260586319, + "grad_norm": 13.649113498748063, + "learning_rate": 9.977359734979818e-07, + "loss": 0.3592, + "step": 7929 + }, + { + "epoch": 4.305103148751357, + "grad_norm": 15.946994134404349, + "learning_rate": 9.962052205553196e-07, + "loss": 0.5153, + "step": 7930 + }, + { + "epoch": 4.305646036916396, + "grad_norm": 11.373437657544107, + "learning_rate": 9.946755812163445e-07, + "loss": 0.34, + "step": 7931 + }, + { + "epoch": 4.306188925081433, + "grad_norm": 13.701873624030634, + "learning_rate": 9.931470556702472e-07, + "loss": 0.4921, + "step": 7932 + }, + { + "epoch": 4.306731813246471, + "grad_norm": 15.895989124158662, + "learning_rate": 9.916196441060765e-07, + "loss": 0.5287, + "step": 7933 + }, + { + "epoch": 4.307274701411509, + "grad_norm": 12.624151740529374, + "learning_rate": 9.900933467127448e-07, + "loss": 0.4054, + "step": 7934 + }, + { + "epoch": 4.307817589576548, + "grad_norm": 10.664486127172733, + "learning_rate": 9.88568163679028e-07, + "loss": 0.2721, + "step": 7935 + }, + { + "epoch": 4.308360477741585, + "grad_norm": 11.459203733504268, + "learning_rate": 9.870440951935633e-07, + "loss": 0.4345, + "step": 7936 + }, + { + "epoch": 4.308903365906623, + "grad_norm": 9.387573986721975, + "learning_rate": 9.855211414448484e-07, + "loss": 0.2966, + "step": 7937 + }, + { + "epoch": 4.309446254071661, + "grad_norm": 11.914282073062253, + "learning_rate": 9.839993026212458e-07, + "loss": 0.4385, + "step": 7938 + }, + { + "epoch": 4.3099891422367, + "grad_norm": 13.814254907233607, + "learning_rate": 9.824785789109792e-07, + "loss": 0.4864, + "step": 7939 + }, + { + "epoch": 4.310532030401737, + "grad_norm": 9.848682371430808, + "learning_rate": 9.809589705021339e-07, + "loss": 0.3087, + "step": 7940 + }, + { + "epoch": 4.311074918566775, + "grad_norm": 12.717733643585042, + "learning_rate": 9.79440477582657e-07, + "loss": 0.3316, + "step": 7941 + }, + { + "epoch": 4.311617806731813, + "grad_norm": 8.913158382091432, + "learning_rate": 9.77923100340359e-07, + "loss": 0.2116, + "step": 7942 + }, + { + "epoch": 4.312160694896852, + "grad_norm": 15.897903840435498, + "learning_rate": 9.764068389629112e-07, + "loss": 0.4687, + "step": 7943 + }, + { + "epoch": 4.312703583061889, + "grad_norm": 16.56909882802291, + "learning_rate": 9.748916936378461e-07, + "loss": 0.4037, + "step": 7944 + }, + { + "epoch": 4.313246471226927, + "grad_norm": 11.499341380465028, + "learning_rate": 9.733776645525606e-07, + "loss": 0.4861, + "step": 7945 + }, + { + "epoch": 4.313789359391965, + "grad_norm": 14.229871658466864, + "learning_rate": 9.718647518943115e-07, + "loss": 0.4858, + "step": 7946 + }, + { + "epoch": 4.314332247557004, + "grad_norm": 11.12286071870711, + "learning_rate": 9.70352955850219e-07, + "loss": 0.3882, + "step": 7947 + }, + { + "epoch": 4.314875135722041, + "grad_norm": 19.942030070834193, + "learning_rate": 9.688422766072648e-07, + "loss": 0.7211, + "step": 7948 + }, + { + "epoch": 4.315418023887079, + "grad_norm": 12.959497734741237, + "learning_rate": 9.673327143522904e-07, + "loss": 0.5216, + "step": 7949 + }, + { + "epoch": 4.315960912052117, + "grad_norm": 13.592821886751958, + "learning_rate": 9.65824269272001e-07, + "loss": 0.586, + "step": 7950 + }, + { + "epoch": 4.316503800217156, + "grad_norm": 16.14045501283915, + "learning_rate": 9.64316941552964e-07, + "loss": 0.448, + "step": 7951 + }, + { + "epoch": 4.317046688382193, + "grad_norm": 14.78354837719301, + "learning_rate": 9.628107313816116e-07, + "loss": 0.5419, + "step": 7952 + }, + { + "epoch": 4.317589576547231, + "grad_norm": 11.53798450893877, + "learning_rate": 9.613056389442265e-07, + "loss": 0.3079, + "step": 7953 + }, + { + "epoch": 4.318132464712269, + "grad_norm": 19.061498107117767, + "learning_rate": 9.598016644269703e-07, + "loss": 0.6412, + "step": 7954 + }, + { + "epoch": 4.318675352877308, + "grad_norm": 14.807149115023895, + "learning_rate": 9.582988080158506e-07, + "loss": 0.4823, + "step": 7955 + }, + { + "epoch": 4.319218241042345, + "grad_norm": 15.900590171517315, + "learning_rate": 9.567970698967434e-07, + "loss": 0.4381, + "step": 7956 + }, + { + "epoch": 4.319761129207383, + "grad_norm": 14.580684073605896, + "learning_rate": 9.552964502553885e-07, + "loss": 0.7024, + "step": 7957 + }, + { + "epoch": 4.320304017372421, + "grad_norm": 13.182118947543977, + "learning_rate": 9.537969492773814e-07, + "loss": 0.4845, + "step": 7958 + }, + { + "epoch": 4.32084690553746, + "grad_norm": 16.398747723242373, + "learning_rate": 9.522985671481888e-07, + "loss": 0.4973, + "step": 7959 + }, + { + "epoch": 4.321389793702497, + "grad_norm": 12.433647949014988, + "learning_rate": 9.50801304053125e-07, + "loss": 0.2912, + "step": 7960 + }, + { + "epoch": 4.321932681867535, + "grad_norm": 18.17349554177303, + "learning_rate": 9.493051601773828e-07, + "loss": 0.3915, + "step": 7961 + }, + { + "epoch": 4.322475570032573, + "grad_norm": 8.718427912132642, + "learning_rate": 9.478101357059988e-07, + "loss": 0.2589, + "step": 7962 + }, + { + "epoch": 4.323018458197612, + "grad_norm": 15.540646935471532, + "learning_rate": 9.46316230823886e-07, + "loss": 0.5203, + "step": 7963 + }, + { + "epoch": 4.323561346362649, + "grad_norm": 10.368742305387377, + "learning_rate": 9.448234457158112e-07, + "loss": 0.337, + "step": 7964 + }, + { + "epoch": 4.324104234527687, + "grad_norm": 15.68035062917153, + "learning_rate": 9.433317805664055e-07, + "loss": 0.5318, + "step": 7965 + }, + { + "epoch": 4.324647122692725, + "grad_norm": 11.340254631487928, + "learning_rate": 9.418412355601603e-07, + "loss": 0.4241, + "step": 7966 + }, + { + "epoch": 4.325190010857764, + "grad_norm": 13.451210184334698, + "learning_rate": 9.403518108814236e-07, + "loss": 0.5082, + "step": 7967 + }, + { + "epoch": 4.3257328990228014, + "grad_norm": 11.17927635762652, + "learning_rate": 9.388635067144169e-07, + "loss": 0.3793, + "step": 7968 + }, + { + "epoch": 4.326275787187839, + "grad_norm": 13.902709479850872, + "learning_rate": 9.373763232432121e-07, + "loss": 0.5547, + "step": 7969 + }, + { + "epoch": 4.326818675352877, + "grad_norm": 18.295405200485035, + "learning_rate": 9.358902606517472e-07, + "loss": 0.3459, + "step": 7970 + }, + { + "epoch": 4.327361563517916, + "grad_norm": 11.600290394966255, + "learning_rate": 9.344053191238211e-07, + "loss": 0.3111, + "step": 7971 + }, + { + "epoch": 4.3279044516829535, + "grad_norm": 14.034291419295219, + "learning_rate": 9.329214988430934e-07, + "loss": 0.4003, + "step": 7972 + }, + { + "epoch": 4.328447339847991, + "grad_norm": 11.650796163329682, + "learning_rate": 9.314387999930863e-07, + "loss": 0.5023, + "step": 7973 + }, + { + "epoch": 4.328990228013029, + "grad_norm": 12.709065524457731, + "learning_rate": 9.299572227571807e-07, + "loss": 0.5242, + "step": 7974 + }, + { + "epoch": 4.329533116178068, + "grad_norm": 14.580030445612671, + "learning_rate": 9.284767673186213e-07, + "loss": 0.4893, + "step": 7975 + }, + { + "epoch": 4.3300760043431055, + "grad_norm": 17.03339708205366, + "learning_rate": 9.269974338605137e-07, + "loss": 0.3985, + "step": 7976 + }, + { + "epoch": 4.330618892508143, + "grad_norm": 15.730665291068982, + "learning_rate": 9.255192225658239e-07, + "loss": 0.335, + "step": 7977 + }, + { + "epoch": 4.331161780673181, + "grad_norm": 10.567417866924043, + "learning_rate": 9.240421336173788e-07, + "loss": 0.2627, + "step": 7978 + }, + { + "epoch": 4.33170466883822, + "grad_norm": 10.543118083895102, + "learning_rate": 9.22566167197867e-07, + "loss": 0.5185, + "step": 7979 + }, + { + "epoch": 4.3322475570032575, + "grad_norm": 12.182864373567915, + "learning_rate": 9.210913234898389e-07, + "loss": 0.3888, + "step": 7980 + }, + { + "epoch": 4.332790445168295, + "grad_norm": 11.069697536316085, + "learning_rate": 9.196176026757064e-07, + "loss": 0.4811, + "step": 7981 + }, + { + "epoch": 4.333333333333333, + "grad_norm": 16.843361756349292, + "learning_rate": 9.181450049377405e-07, + "loss": 0.5628, + "step": 7982 + }, + { + "epoch": 4.333876221498372, + "grad_norm": 16.78954259852446, + "learning_rate": 9.166735304580743e-07, + "loss": 0.5249, + "step": 7983 + }, + { + "epoch": 4.3344191096634095, + "grad_norm": 10.463116306289768, + "learning_rate": 9.15203179418701e-07, + "loss": 0.3087, + "step": 7984 + }, + { + "epoch": 4.334961997828447, + "grad_norm": 10.619811686151522, + "learning_rate": 9.13733952001481e-07, + "loss": 0.4424, + "step": 7985 + }, + { + "epoch": 4.335504885993485, + "grad_norm": 9.230465655882801, + "learning_rate": 9.122658483881264e-07, + "loss": 0.307, + "step": 7986 + }, + { + "epoch": 4.336047774158524, + "grad_norm": 9.602884171912605, + "learning_rate": 9.107988687602142e-07, + "loss": 0.4543, + "step": 7987 + }, + { + "epoch": 4.3365906623235615, + "grad_norm": 11.802560272642094, + "learning_rate": 9.09333013299184e-07, + "loss": 0.293, + "step": 7988 + }, + { + "epoch": 4.337133550488599, + "grad_norm": 19.98627599580886, + "learning_rate": 9.078682821863339e-07, + "loss": 0.5988, + "step": 7989 + }, + { + "epoch": 4.337676438653637, + "grad_norm": 14.01016302793737, + "learning_rate": 9.064046756028278e-07, + "loss": 0.3797, + "step": 7990 + }, + { + "epoch": 4.338219326818676, + "grad_norm": 10.54886354090915, + "learning_rate": 9.049421937296821e-07, + "loss": 0.2179, + "step": 7991 + }, + { + "epoch": 4.3387622149837135, + "grad_norm": 14.088344911170685, + "learning_rate": 9.034808367477844e-07, + "loss": 0.5632, + "step": 7992 + }, + { + "epoch": 4.339305103148751, + "grad_norm": 18.19756991978728, + "learning_rate": 9.0202060483787e-07, + "loss": 0.6184, + "step": 7993 + }, + { + "epoch": 4.339847991313789, + "grad_norm": 13.111955449734706, + "learning_rate": 9.005614981805522e-07, + "loss": 0.2811, + "step": 7994 + }, + { + "epoch": 4.340390879478828, + "grad_norm": 16.45443112800284, + "learning_rate": 8.991035169562879e-07, + "loss": 0.7305, + "step": 7995 + }, + { + "epoch": 4.3409337676438655, + "grad_norm": 15.613823352223783, + "learning_rate": 8.976466613454038e-07, + "loss": 0.6787, + "step": 7996 + }, + { + "epoch": 4.341476655808903, + "grad_norm": 12.421367112935704, + "learning_rate": 8.961909315280914e-07, + "loss": 0.3399, + "step": 7997 + }, + { + "epoch": 4.342019543973941, + "grad_norm": 12.609970065901182, + "learning_rate": 8.947363276843901e-07, + "loss": 0.3253, + "step": 7998 + }, + { + "epoch": 4.34256243213898, + "grad_norm": 10.696313488480198, + "learning_rate": 8.932828499942159e-07, + "loss": 0.3197, + "step": 7999 + }, + { + "epoch": 4.3431053203040175, + "grad_norm": 13.140351352844807, + "learning_rate": 8.918304986373283e-07, + "loss": 0.3766, + "step": 8000 + }, + { + "epoch": 4.343648208469055, + "grad_norm": 9.439332843485358, + "learning_rate": 8.903792737933626e-07, + "loss": 0.3762, + "step": 8001 + }, + { + "epoch": 4.344191096634093, + "grad_norm": 17.7660438004401, + "learning_rate": 8.889291756418073e-07, + "loss": 0.4339, + "step": 8002 + }, + { + "epoch": 4.344733984799132, + "grad_norm": 12.278395472072871, + "learning_rate": 8.874802043620112e-07, + "loss": 0.4535, + "step": 8003 + }, + { + "epoch": 4.3452768729641695, + "grad_norm": 16.55900970801957, + "learning_rate": 8.860323601331899e-07, + "loss": 0.511, + "step": 8004 + }, + { + "epoch": 4.345819761129207, + "grad_norm": 11.14928157377559, + "learning_rate": 8.84585643134408e-07, + "loss": 0.417, + "step": 8005 + }, + { + "epoch": 4.346362649294245, + "grad_norm": 21.55707350645272, + "learning_rate": 8.831400535446044e-07, + "loss": 0.5588, + "step": 8006 + }, + { + "epoch": 4.346905537459284, + "grad_norm": 14.787630844212787, + "learning_rate": 8.816955915425663e-07, + "loss": 0.4701, + "step": 8007 + }, + { + "epoch": 4.3474484256243215, + "grad_norm": 12.957973486859906, + "learning_rate": 8.802522573069505e-07, + "loss": 0.3898, + "step": 8008 + }, + { + "epoch": 4.347991313789359, + "grad_norm": 13.858932325098396, + "learning_rate": 8.788100510162712e-07, + "loss": 0.4319, + "step": 8009 + }, + { + "epoch": 4.348534201954397, + "grad_norm": 11.44820285977616, + "learning_rate": 8.773689728489021e-07, + "loss": 0.346, + "step": 8010 + }, + { + "epoch": 4.349077090119436, + "grad_norm": 13.32788926130813, + "learning_rate": 8.759290229830774e-07, + "loss": 0.5118, + "step": 8011 + }, + { + "epoch": 4.3496199782844736, + "grad_norm": 13.65313572883658, + "learning_rate": 8.744902015968926e-07, + "loss": 0.6354, + "step": 8012 + }, + { + "epoch": 4.350162866449511, + "grad_norm": 15.440111761472027, + "learning_rate": 8.730525088683028e-07, + "loss": 0.3655, + "step": 8013 + }, + { + "epoch": 4.350705754614549, + "grad_norm": 13.392699739564616, + "learning_rate": 8.71615944975126e-07, + "loss": 0.5795, + "step": 8014 + }, + { + "epoch": 4.351248642779588, + "grad_norm": 12.572118935109678, + "learning_rate": 8.701805100950378e-07, + "loss": 0.4298, + "step": 8015 + }, + { + "epoch": 4.351791530944626, + "grad_norm": 13.635858966392217, + "learning_rate": 8.687462044055739e-07, + "loss": 0.3448, + "step": 8016 + }, + { + "epoch": 4.352334419109663, + "grad_norm": 13.828043330695678, + "learning_rate": 8.673130280841335e-07, + "loss": 0.645, + "step": 8017 + }, + { + "epoch": 4.352877307274701, + "grad_norm": 15.111038785972921, + "learning_rate": 8.658809813079716e-07, + "loss": 0.451, + "step": 8018 + }, + { + "epoch": 4.35342019543974, + "grad_norm": 11.79984417087872, + "learning_rate": 8.644500642542086e-07, + "loss": 0.3503, + "step": 8019 + }, + { + "epoch": 4.353963083604778, + "grad_norm": 11.058205432853121, + "learning_rate": 8.63020277099822e-07, + "loss": 0.3616, + "step": 8020 + }, + { + "epoch": 4.354505971769815, + "grad_norm": 17.658740606783095, + "learning_rate": 8.61591620021649e-07, + "loss": 0.4659, + "step": 8021 + }, + { + "epoch": 4.355048859934853, + "grad_norm": 15.417869712043583, + "learning_rate": 8.601640931963873e-07, + "loss": 0.5888, + "step": 8022 + }, + { + "epoch": 4.355591748099892, + "grad_norm": 17.637210382444906, + "learning_rate": 8.587376968006011e-07, + "loss": 0.6608, + "step": 8023 + }, + { + "epoch": 4.35613463626493, + "grad_norm": 11.883245195971504, + "learning_rate": 8.573124310107028e-07, + "loss": 0.3882, + "step": 8024 + }, + { + "epoch": 4.356677524429967, + "grad_norm": 13.09142985069161, + "learning_rate": 8.558882960029758e-07, + "loss": 0.4295, + "step": 8025 + }, + { + "epoch": 4.357220412595005, + "grad_norm": 13.42588355451286, + "learning_rate": 8.544652919535568e-07, + "loss": 0.5258, + "step": 8026 + }, + { + "epoch": 4.357763300760044, + "grad_norm": 11.055862131666451, + "learning_rate": 8.530434190384463e-07, + "loss": 0.3189, + "step": 8027 + }, + { + "epoch": 4.358306188925082, + "grad_norm": 14.270581711188854, + "learning_rate": 8.516226774335046e-07, + "loss": 0.4594, + "step": 8028 + }, + { + "epoch": 4.358849077090119, + "grad_norm": 13.14559769122887, + "learning_rate": 8.502030673144479e-07, + "loss": 0.6237, + "step": 8029 + }, + { + "epoch": 4.359391965255157, + "grad_norm": 16.47710832181487, + "learning_rate": 8.487845888568625e-07, + "loss": 0.4926, + "step": 8030 + }, + { + "epoch": 4.359934853420196, + "grad_norm": 9.957560938239181, + "learning_rate": 8.473672422361801e-07, + "loss": 0.3599, + "step": 8031 + }, + { + "epoch": 4.360477741585234, + "grad_norm": 13.852815438372643, + "learning_rate": 8.459510276277083e-07, + "loss": 0.3684, + "step": 8032 + }, + { + "epoch": 4.361020629750271, + "grad_norm": 15.597466583644815, + "learning_rate": 8.445359452065982e-07, + "loss": 0.6425, + "step": 8033 + }, + { + "epoch": 4.361563517915309, + "grad_norm": 10.260784495869965, + "learning_rate": 8.431219951478775e-07, + "loss": 0.3258, + "step": 8034 + }, + { + "epoch": 4.362106406080348, + "grad_norm": 12.005771351192994, + "learning_rate": 8.41709177626423e-07, + "loss": 0.3723, + "step": 8035 + }, + { + "epoch": 4.362649294245386, + "grad_norm": 12.407275509701295, + "learning_rate": 8.402974928169705e-07, + "loss": 0.4967, + "step": 8036 + }, + { + "epoch": 4.363192182410423, + "grad_norm": 14.727533661174913, + "learning_rate": 8.388869408941247e-07, + "loss": 0.4133, + "step": 8037 + }, + { + "epoch": 4.363735070575461, + "grad_norm": 11.403317480094385, + "learning_rate": 8.374775220323406e-07, + "loss": 0.3032, + "step": 8038 + }, + { + "epoch": 4.3642779587405, + "grad_norm": 12.660090294532273, + "learning_rate": 8.360692364059409e-07, + "loss": 0.3985, + "step": 8039 + }, + { + "epoch": 4.364820846905538, + "grad_norm": 9.17954467960599, + "learning_rate": 8.346620841891029e-07, + "loss": 0.287, + "step": 8040 + }, + { + "epoch": 4.365363735070575, + "grad_norm": 16.649840062856516, + "learning_rate": 8.332560655558663e-07, + "loss": 0.5538, + "step": 8041 + }, + { + "epoch": 4.365906623235613, + "grad_norm": 12.11674130758851, + "learning_rate": 8.318511806801288e-07, + "loss": 0.6752, + "step": 8042 + }, + { + "epoch": 4.366449511400652, + "grad_norm": 15.055685550752939, + "learning_rate": 8.304474297356479e-07, + "loss": 0.4629, + "step": 8043 + }, + { + "epoch": 4.36699239956569, + "grad_norm": 13.344971903872295, + "learning_rate": 8.29044812896046e-07, + "loss": 0.3308, + "step": 8044 + }, + { + "epoch": 4.367535287730727, + "grad_norm": 10.304984524921624, + "learning_rate": 8.276433303347953e-07, + "loss": 0.7107, + "step": 8045 + }, + { + "epoch": 4.368078175895765, + "grad_norm": 11.722406609137764, + "learning_rate": 8.26242982225236e-07, + "loss": 0.5312, + "step": 8046 + }, + { + "epoch": 4.368621064060804, + "grad_norm": 9.83110831313041, + "learning_rate": 8.248437687405664e-07, + "loss": 0.2615, + "step": 8047 + }, + { + "epoch": 4.369163952225842, + "grad_norm": 12.152410682750114, + "learning_rate": 8.234456900538424e-07, + "loss": 0.3105, + "step": 8048 + }, + { + "epoch": 4.369706840390879, + "grad_norm": 13.31087133977307, + "learning_rate": 8.220487463379811e-07, + "loss": 0.4315, + "step": 8049 + }, + { + "epoch": 4.370249728555917, + "grad_norm": 11.94526177677569, + "learning_rate": 8.206529377657579e-07, + "loss": 0.3136, + "step": 8050 + }, + { + "epoch": 4.370792616720956, + "grad_norm": 16.337088129910562, + "learning_rate": 8.192582645098091e-07, + "loss": 0.6422, + "step": 8051 + }, + { + "epoch": 4.371335504885994, + "grad_norm": 15.612171822665877, + "learning_rate": 8.178647267426299e-07, + "loss": 0.4182, + "step": 8052 + }, + { + "epoch": 4.371878393051031, + "grad_norm": 15.057359339910816, + "learning_rate": 8.164723246365758e-07, + "loss": 0.5622, + "step": 8053 + }, + { + "epoch": 4.372421281216069, + "grad_norm": 14.162937228652032, + "learning_rate": 8.150810583638602e-07, + "loss": 0.4681, + "step": 8054 + }, + { + "epoch": 4.372964169381108, + "grad_norm": 15.27678092027012, + "learning_rate": 8.136909280965589e-07, + "loss": 0.5663, + "step": 8055 + }, + { + "epoch": 4.373507057546146, + "grad_norm": 9.416661735214964, + "learning_rate": 8.123019340066029e-07, + "loss": 0.4337, + "step": 8056 + }, + { + "epoch": 4.374049945711183, + "grad_norm": 8.853547844230004, + "learning_rate": 8.109140762657885e-07, + "loss": 0.2355, + "step": 8057 + }, + { + "epoch": 4.374592833876221, + "grad_norm": 13.636809573527607, + "learning_rate": 8.095273550457649e-07, + "loss": 0.3061, + "step": 8058 + }, + { + "epoch": 4.37513572204126, + "grad_norm": 10.074225429783628, + "learning_rate": 8.081417705180461e-07, + "loss": 0.5573, + "step": 8059 + }, + { + "epoch": 4.375678610206298, + "grad_norm": 16.626739722709576, + "learning_rate": 8.067573228540015e-07, + "loss": 0.8822, + "step": 8060 + }, + { + "epoch": 4.3762214983713354, + "grad_norm": 13.021514377648314, + "learning_rate": 8.053740122248665e-07, + "loss": 0.5488, + "step": 8061 + }, + { + "epoch": 4.376764386536373, + "grad_norm": 22.411620281238108, + "learning_rate": 8.039918388017231e-07, + "loss": 1.0347, + "step": 8062 + }, + { + "epoch": 4.377307274701412, + "grad_norm": 9.303857597914824, + "learning_rate": 8.026108027555302e-07, + "loss": 0.4273, + "step": 8063 + }, + { + "epoch": 4.37785016286645, + "grad_norm": 13.977505292581387, + "learning_rate": 8.012309042570887e-07, + "loss": 0.4826, + "step": 8064 + }, + { + "epoch": 4.3783930510314875, + "grad_norm": 16.681445518035993, + "learning_rate": 7.9985214347707e-07, + "loss": 0.6674, + "step": 8065 + }, + { + "epoch": 4.378935939196525, + "grad_norm": 15.765471997364271, + "learning_rate": 7.984745205860022e-07, + "loss": 0.4246, + "step": 8066 + }, + { + "epoch": 4.379478827361564, + "grad_norm": 12.65619134059525, + "learning_rate": 7.970980357542679e-07, + "loss": 0.4719, + "step": 8067 + }, + { + "epoch": 4.380021715526602, + "grad_norm": 14.558989497065397, + "learning_rate": 7.957226891521185e-07, + "loss": 0.3499, + "step": 8068 + }, + { + "epoch": 4.3805646036916395, + "grad_norm": 10.308079110981643, + "learning_rate": 7.943484809496549e-07, + "loss": 0.2406, + "step": 8069 + }, + { + "epoch": 4.381107491856677, + "grad_norm": 10.563122177925374, + "learning_rate": 7.929754113168442e-07, + "loss": 0.4711, + "step": 8070 + }, + { + "epoch": 4.381650380021716, + "grad_norm": 16.713706157595226, + "learning_rate": 7.916034804235062e-07, + "loss": 0.4097, + "step": 8071 + }, + { + "epoch": 4.382193268186754, + "grad_norm": 13.661917790603933, + "learning_rate": 7.902326884393275e-07, + "loss": 0.421, + "step": 8072 + }, + { + "epoch": 4.3827361563517915, + "grad_norm": 11.192296315079558, + "learning_rate": 7.888630355338489e-07, + "loss": 0.2375, + "step": 8073 + }, + { + "epoch": 4.383279044516829, + "grad_norm": 10.108216567015306, + "learning_rate": 7.874945218764685e-07, + "loss": 0.2637, + "step": 8074 + }, + { + "epoch": 4.383821932681868, + "grad_norm": 16.67098266690521, + "learning_rate": 7.861271476364496e-07, + "loss": 0.7568, + "step": 8075 + }, + { + "epoch": 4.384364820846906, + "grad_norm": 14.31792498162585, + "learning_rate": 7.847609129829081e-07, + "loss": 0.5722, + "step": 8076 + }, + { + "epoch": 4.3849077090119435, + "grad_norm": 11.783645413114316, + "learning_rate": 7.833958180848267e-07, + "loss": 0.4316, + "step": 8077 + }, + { + "epoch": 4.385450597176981, + "grad_norm": 12.328500975614492, + "learning_rate": 7.820318631110357e-07, + "loss": 0.3031, + "step": 8078 + }, + { + "epoch": 4.38599348534202, + "grad_norm": 15.932687060138331, + "learning_rate": 7.806690482302371e-07, + "loss": 0.5638, + "step": 8079 + }, + { + "epoch": 4.386536373507058, + "grad_norm": 11.335548440871465, + "learning_rate": 7.793073736109846e-07, + "loss": 0.2642, + "step": 8080 + }, + { + "epoch": 4.3870792616720955, + "grad_norm": 14.219082798674243, + "learning_rate": 7.779468394216905e-07, + "loss": 0.5914, + "step": 8081 + }, + { + "epoch": 4.387622149837133, + "grad_norm": 11.171865534559577, + "learning_rate": 7.765874458306299e-07, + "loss": 0.3573, + "step": 8082 + }, + { + "epoch": 4.388165038002172, + "grad_norm": 14.949429459159015, + "learning_rate": 7.752291930059341e-07, + "loss": 0.491, + "step": 8083 + }, + { + "epoch": 4.38870792616721, + "grad_norm": 11.682058767447176, + "learning_rate": 7.738720811155931e-07, + "loss": 0.4278, + "step": 8084 + }, + { + "epoch": 4.3892508143322475, + "grad_norm": 9.720345707653586, + "learning_rate": 7.725161103274581e-07, + "loss": 0.3931, + "step": 8085 + }, + { + "epoch": 4.389793702497285, + "grad_norm": 13.696977621677252, + "learning_rate": 7.711612808092372e-07, + "loss": 0.4812, + "step": 8086 + }, + { + "epoch": 4.390336590662324, + "grad_norm": 13.326736087170834, + "learning_rate": 7.698075927284964e-07, + "loss": 0.5377, + "step": 8087 + }, + { + "epoch": 4.390879478827362, + "grad_norm": 13.222977582929962, + "learning_rate": 7.684550462526641e-07, + "loss": 0.4528, + "step": 8088 + }, + { + "epoch": 4.3914223669923995, + "grad_norm": 15.11572868941985, + "learning_rate": 7.671036415490252e-07, + "loss": 0.6123, + "step": 8089 + }, + { + "epoch": 4.391965255157437, + "grad_norm": 13.903887430270185, + "learning_rate": 7.657533787847215e-07, + "loss": 0.493, + "step": 8090 + }, + { + "epoch": 4.392508143322476, + "grad_norm": 13.728638124426194, + "learning_rate": 7.644042581267585e-07, + "loss": 0.5071, + "step": 8091 + }, + { + "epoch": 4.393051031487514, + "grad_norm": 16.05244491657159, + "learning_rate": 7.630562797419949e-07, + "loss": 0.4212, + "step": 8092 + }, + { + "epoch": 4.3935939196525515, + "grad_norm": 11.275459672619295, + "learning_rate": 7.617094437971529e-07, + "loss": 0.2391, + "step": 8093 + }, + { + "epoch": 4.394136807817589, + "grad_norm": 15.521537268341413, + "learning_rate": 7.603637504588112e-07, + "loss": 0.4127, + "step": 8094 + }, + { + "epoch": 4.394679695982628, + "grad_norm": 11.264052644917689, + "learning_rate": 7.59019199893406e-07, + "loss": 0.3287, + "step": 8095 + }, + { + "epoch": 4.395222584147666, + "grad_norm": 11.628986278608336, + "learning_rate": 7.576757922672339e-07, + "loss": 0.3109, + "step": 8096 + }, + { + "epoch": 4.3957654723127035, + "grad_norm": 12.05990932087969, + "learning_rate": 7.56333527746449e-07, + "loss": 0.4798, + "step": 8097 + }, + { + "epoch": 4.396308360477741, + "grad_norm": 11.758858029678734, + "learning_rate": 7.54992406497067e-07, + "loss": 0.3736, + "step": 8098 + }, + { + "epoch": 4.39685124864278, + "grad_norm": 16.559994897045023, + "learning_rate": 7.536524286849578e-07, + "loss": 0.5436, + "step": 8099 + }, + { + "epoch": 4.397394136807818, + "grad_norm": 12.81236009916768, + "learning_rate": 7.523135944758519e-07, + "loss": 0.5033, + "step": 8100 + }, + { + "epoch": 4.3979370249728555, + "grad_norm": 11.076059660720759, + "learning_rate": 7.509759040353415e-07, + "loss": 0.337, + "step": 8101 + }, + { + "epoch": 4.398479913137893, + "grad_norm": 9.275703031226358, + "learning_rate": 7.496393575288708e-07, + "loss": 0.2984, + "step": 8102 + }, + { + "epoch": 4.399022801302932, + "grad_norm": 15.86969949487026, + "learning_rate": 7.483039551217475e-07, + "loss": 0.5653, + "step": 8103 + }, + { + "epoch": 4.39956568946797, + "grad_norm": 12.679331406935141, + "learning_rate": 7.469696969791351e-07, + "loss": 0.3381, + "step": 8104 + }, + { + "epoch": 4.400108577633008, + "grad_norm": 14.791912312085929, + "learning_rate": 7.456365832660572e-07, + "loss": 0.3494, + "step": 8105 + }, + { + "epoch": 4.400651465798045, + "grad_norm": 10.790537111566087, + "learning_rate": 7.443046141473986e-07, + "loss": 0.3024, + "step": 8106 + }, + { + "epoch": 4.401194353963084, + "grad_norm": 16.74012144354708, + "learning_rate": 7.429737897878942e-07, + "loss": 0.7146, + "step": 8107 + }, + { + "epoch": 4.401737242128122, + "grad_norm": 18.112779551139106, + "learning_rate": 7.416441103521476e-07, + "loss": 0.787, + "step": 8108 + }, + { + "epoch": 4.40228013029316, + "grad_norm": 18.192800558391657, + "learning_rate": 7.403155760046099e-07, + "loss": 0.6249, + "step": 8109 + }, + { + "epoch": 4.402823018458197, + "grad_norm": 12.394483609056916, + "learning_rate": 7.389881869096027e-07, + "loss": 0.3235, + "step": 8110 + }, + { + "epoch": 4.403365906623236, + "grad_norm": 12.486317574320127, + "learning_rate": 7.376619432312926e-07, + "loss": 0.4458, + "step": 8111 + }, + { + "epoch": 4.403908794788274, + "grad_norm": 11.546942923476625, + "learning_rate": 7.363368451337183e-07, + "loss": 0.2904, + "step": 8112 + }, + { + "epoch": 4.404451682953312, + "grad_norm": 11.661763155608599, + "learning_rate": 7.350128927807676e-07, + "loss": 0.6103, + "step": 8113 + }, + { + "epoch": 4.404994571118349, + "grad_norm": 12.401624870044014, + "learning_rate": 7.336900863361851e-07, + "loss": 0.5543, + "step": 8114 + }, + { + "epoch": 4.405537459283388, + "grad_norm": 14.577278093491197, + "learning_rate": 7.323684259635855e-07, + "loss": 0.5578, + "step": 8115 + }, + { + "epoch": 4.406080347448426, + "grad_norm": 12.374504432300087, + "learning_rate": 7.310479118264247e-07, + "loss": 0.4786, + "step": 8116 + }, + { + "epoch": 4.406623235613464, + "grad_norm": 13.53250221886951, + "learning_rate": 7.297285440880331e-07, + "loss": 0.5691, + "step": 8117 + }, + { + "epoch": 4.407166123778501, + "grad_norm": 15.930406260794717, + "learning_rate": 7.284103229115891e-07, + "loss": 0.4541, + "step": 8118 + }, + { + "epoch": 4.40770901194354, + "grad_norm": 12.90844460573453, + "learning_rate": 7.270932484601345e-07, + "loss": 0.43, + "step": 8119 + }, + { + "epoch": 4.408251900108578, + "grad_norm": 11.272038151188799, + "learning_rate": 7.257773208965646e-07, + "loss": 0.5185, + "step": 8120 + }, + { + "epoch": 4.408794788273616, + "grad_norm": 17.139510748072922, + "learning_rate": 7.24462540383637e-07, + "loss": 0.6829, + "step": 8121 + }, + { + "epoch": 4.409337676438653, + "grad_norm": 12.62321799336504, + "learning_rate": 7.231489070839648e-07, + "loss": 0.3177, + "step": 8122 + }, + { + "epoch": 4.409880564603692, + "grad_norm": 11.177325186499452, + "learning_rate": 7.218364211600215e-07, + "loss": 0.3812, + "step": 8123 + }, + { + "epoch": 4.41042345276873, + "grad_norm": 16.25406531253722, + "learning_rate": 7.205250827741361e-07, + "loss": 0.4482, + "step": 8124 + }, + { + "epoch": 4.410966340933768, + "grad_norm": 16.76577388765637, + "learning_rate": 7.192148920884989e-07, + "loss": 0.4272, + "step": 8125 + }, + { + "epoch": 4.411509229098805, + "grad_norm": 14.687321011203691, + "learning_rate": 7.179058492651547e-07, + "loss": 0.5207, + "step": 8126 + }, + { + "epoch": 4.412052117263844, + "grad_norm": 11.956370304080279, + "learning_rate": 7.165979544660085e-07, + "loss": 0.3902, + "step": 8127 + }, + { + "epoch": 4.412595005428882, + "grad_norm": 16.74860718331433, + "learning_rate": 7.152912078528229e-07, + "loss": 0.4425, + "step": 8128 + }, + { + "epoch": 4.41313789359392, + "grad_norm": 12.718796959860708, + "learning_rate": 7.139856095872177e-07, + "loss": 0.4485, + "step": 8129 + }, + { + "epoch": 4.413680781758957, + "grad_norm": 11.351426129816947, + "learning_rate": 7.126811598306726e-07, + "loss": 0.4284, + "step": 8130 + }, + { + "epoch": 4.414223669923996, + "grad_norm": 15.528891800944992, + "learning_rate": 7.11377858744522e-07, + "loss": 0.4391, + "step": 8131 + }, + { + "epoch": 4.414766558089034, + "grad_norm": 15.062549425859391, + "learning_rate": 7.100757064899644e-07, + "loss": 0.4085, + "step": 8132 + }, + { + "epoch": 4.415309446254072, + "grad_norm": 13.147141392832742, + "learning_rate": 7.087747032280479e-07, + "loss": 0.5011, + "step": 8133 + }, + { + "epoch": 4.415852334419109, + "grad_norm": 13.452451139567685, + "learning_rate": 7.074748491196837e-07, + "loss": 0.5419, + "step": 8134 + }, + { + "epoch": 4.416395222584148, + "grad_norm": 15.154494684843028, + "learning_rate": 7.061761443256409e-07, + "loss": 0.4214, + "step": 8135 + }, + { + "epoch": 4.416938110749186, + "grad_norm": 14.435292009300346, + "learning_rate": 7.048785890065446e-07, + "loss": 0.3093, + "step": 8136 + }, + { + "epoch": 4.417480998914224, + "grad_norm": 14.372369930518246, + "learning_rate": 7.035821833228785e-07, + "loss": 0.4875, + "step": 8137 + }, + { + "epoch": 4.418023887079261, + "grad_norm": 13.382996066133376, + "learning_rate": 7.02286927434983e-07, + "loss": 0.5495, + "step": 8138 + }, + { + "epoch": 4.4185667752443, + "grad_norm": 18.053750531135215, + "learning_rate": 7.009928215030615e-07, + "loss": 0.4188, + "step": 8139 + }, + { + "epoch": 4.419109663409338, + "grad_norm": 16.22341652670191, + "learning_rate": 6.996998656871646e-07, + "loss": 0.6233, + "step": 8140 + }, + { + "epoch": 4.419652551574376, + "grad_norm": 12.414389082202195, + "learning_rate": 6.984080601472143e-07, + "loss": 0.441, + "step": 8141 + }, + { + "epoch": 4.420195439739413, + "grad_norm": 15.971472074811146, + "learning_rate": 6.971174050429786e-07, + "loss": 0.4502, + "step": 8142 + }, + { + "epoch": 4.420738327904452, + "grad_norm": 8.57673948818228, + "learning_rate": 6.958279005340874e-07, + "loss": 0.3876, + "step": 8143 + }, + { + "epoch": 4.42128121606949, + "grad_norm": 14.907293994220137, + "learning_rate": 6.94539546780032e-07, + "loss": 0.4499, + "step": 8144 + }, + { + "epoch": 4.421824104234528, + "grad_norm": 11.784072399397415, + "learning_rate": 6.93252343940154e-07, + "loss": 0.4506, + "step": 8145 + }, + { + "epoch": 4.422366992399565, + "grad_norm": 13.773309280609004, + "learning_rate": 6.919662921736614e-07, + "loss": 0.6268, + "step": 8146 + }, + { + "epoch": 4.422909880564604, + "grad_norm": 14.034219630419418, + "learning_rate": 6.906813916396104e-07, + "loss": 0.557, + "step": 8147 + }, + { + "epoch": 4.423452768729642, + "grad_norm": 11.54689354028201, + "learning_rate": 6.89397642496924e-07, + "loss": 0.293, + "step": 8148 + }, + { + "epoch": 4.42399565689468, + "grad_norm": 12.196466085866508, + "learning_rate": 6.881150449043727e-07, + "loss": 0.4063, + "step": 8149 + }, + { + "epoch": 4.424538545059717, + "grad_norm": 13.943110002700577, + "learning_rate": 6.868335990205965e-07, + "loss": 0.4629, + "step": 8150 + }, + { + "epoch": 4.425081433224756, + "grad_norm": 11.136718980358994, + "learning_rate": 6.855533050040841e-07, + "loss": 0.353, + "step": 8151 + }, + { + "epoch": 4.425624321389794, + "grad_norm": 12.815867518018083, + "learning_rate": 6.84274163013181e-07, + "loss": 0.3692, + "step": 8152 + }, + { + "epoch": 4.426167209554832, + "grad_norm": 15.673409819920156, + "learning_rate": 6.829961732060997e-07, + "loss": 0.7404, + "step": 8153 + }, + { + "epoch": 4.4267100977198695, + "grad_norm": 15.185723434348903, + "learning_rate": 6.81719335740898e-07, + "loss": 0.6198, + "step": 8154 + }, + { + "epoch": 4.427252985884908, + "grad_norm": 13.985897143479551, + "learning_rate": 6.804436507755008e-07, + "loss": 0.4197, + "step": 8155 + }, + { + "epoch": 4.427795874049946, + "grad_norm": 15.277569295336207, + "learning_rate": 6.791691184676863e-07, + "loss": 0.5533, + "step": 8156 + }, + { + "epoch": 4.428338762214984, + "grad_norm": 16.725284925242832, + "learning_rate": 6.778957389750907e-07, + "loss": 0.6159, + "step": 8157 + }, + { + "epoch": 4.4288816503800215, + "grad_norm": 14.027576538249749, + "learning_rate": 6.766235124552079e-07, + "loss": 0.3277, + "step": 8158 + }, + { + "epoch": 4.42942453854506, + "grad_norm": 9.599302156075451, + "learning_rate": 6.753524390653876e-07, + "loss": 0.2496, + "step": 8159 + }, + { + "epoch": 4.429967426710098, + "grad_norm": 15.48284468466842, + "learning_rate": 6.740825189628386e-07, + "loss": 0.3937, + "step": 8160 + }, + { + "epoch": 4.430510314875136, + "grad_norm": 13.417678670406026, + "learning_rate": 6.728137523046274e-07, + "loss": 0.4335, + "step": 8161 + }, + { + "epoch": 4.4310532030401735, + "grad_norm": 16.002263273785378, + "learning_rate": 6.715461392476763e-07, + "loss": 0.7239, + "step": 8162 + }, + { + "epoch": 4.431596091205212, + "grad_norm": 13.513757533100362, + "learning_rate": 6.702796799487665e-07, + "loss": 0.5011, + "step": 8163 + }, + { + "epoch": 4.43213897937025, + "grad_norm": 17.068661392730355, + "learning_rate": 6.69014374564535e-07, + "loss": 0.5128, + "step": 8164 + }, + { + "epoch": 4.432681867535288, + "grad_norm": 16.016384575594365, + "learning_rate": 6.677502232514777e-07, + "loss": 0.2884, + "step": 8165 + }, + { + "epoch": 4.4332247557003255, + "grad_norm": 14.947992178628231, + "learning_rate": 6.66487226165945e-07, + "loss": 0.603, + "step": 8166 + }, + { + "epoch": 4.433767643865364, + "grad_norm": 14.940395600883463, + "learning_rate": 6.652253834641487e-07, + "loss": 0.4589, + "step": 8167 + }, + { + "epoch": 4.434310532030402, + "grad_norm": 10.553864429327028, + "learning_rate": 6.639646953021539e-07, + "loss": 0.2977, + "step": 8168 + }, + { + "epoch": 4.43485342019544, + "grad_norm": 13.081302650262185, + "learning_rate": 6.627051618358848e-07, + "loss": 0.5568, + "step": 8169 + }, + { + "epoch": 4.4353963083604775, + "grad_norm": 11.615908937272211, + "learning_rate": 6.614467832211224e-07, + "loss": 0.3839, + "step": 8170 + }, + { + "epoch": 4.435939196525516, + "grad_norm": 13.81874812862262, + "learning_rate": 6.601895596135055e-07, + "loss": 0.4983, + "step": 8171 + }, + { + "epoch": 4.436482084690554, + "grad_norm": 10.994119413078606, + "learning_rate": 6.589334911685297e-07, + "loss": 0.2608, + "step": 8172 + }, + { + "epoch": 4.437024972855592, + "grad_norm": 13.512853171310855, + "learning_rate": 6.576785780415474e-07, + "loss": 0.3852, + "step": 8173 + }, + { + "epoch": 4.4375678610206295, + "grad_norm": 13.099445891838716, + "learning_rate": 6.564248203877677e-07, + "loss": 0.3673, + "step": 8174 + }, + { + "epoch": 4.438110749185668, + "grad_norm": 13.344678881473703, + "learning_rate": 6.551722183622578e-07, + "loss": 0.3684, + "step": 8175 + }, + { + "epoch": 4.438653637350706, + "grad_norm": 11.89837504148065, + "learning_rate": 6.539207721199392e-07, + "loss": 0.3663, + "step": 8176 + }, + { + "epoch": 4.439196525515744, + "grad_norm": 16.57103010045736, + "learning_rate": 6.526704818155983e-07, + "loss": 0.3226, + "step": 8177 + }, + { + "epoch": 4.4397394136807815, + "grad_norm": 13.302285956970346, + "learning_rate": 6.514213476038667e-07, + "loss": 0.5893, + "step": 8178 + }, + { + "epoch": 4.440282301845819, + "grad_norm": 16.633046645169372, + "learning_rate": 6.501733696392454e-07, + "loss": 0.3461, + "step": 8179 + }, + { + "epoch": 4.440825190010858, + "grad_norm": 11.411100884112152, + "learning_rate": 6.489265480760798e-07, + "loss": 0.3867, + "step": 8180 + }, + { + "epoch": 4.441368078175896, + "grad_norm": 11.251827734607014, + "learning_rate": 6.476808830685855e-07, + "loss": 0.4055, + "step": 8181 + }, + { + "epoch": 4.4419109663409335, + "grad_norm": 12.272578715881838, + "learning_rate": 6.464363747708236e-07, + "loss": 0.489, + "step": 8182 + }, + { + "epoch": 4.442453854505972, + "grad_norm": 10.150536439617065, + "learning_rate": 6.451930233367154e-07, + "loss": 0.2645, + "step": 8183 + }, + { + "epoch": 4.44299674267101, + "grad_norm": 16.617783744787634, + "learning_rate": 6.43950828920048e-07, + "loss": 0.452, + "step": 8184 + }, + { + "epoch": 4.443539630836048, + "grad_norm": 13.318664745403998, + "learning_rate": 6.427097916744496e-07, + "loss": 0.2832, + "step": 8185 + }, + { + "epoch": 4.4440825190010855, + "grad_norm": 14.768021540885634, + "learning_rate": 6.414699117534207e-07, + "loss": 0.6652, + "step": 8186 + }, + { + "epoch": 4.444625407166123, + "grad_norm": 12.248982189139824, + "learning_rate": 6.402311893103052e-07, + "loss": 0.3722, + "step": 8187 + }, + { + "epoch": 4.445168295331162, + "grad_norm": 17.38625490353128, + "learning_rate": 6.389936244983153e-07, + "loss": 0.4537, + "step": 8188 + }, + { + "epoch": 4.4457111834962, + "grad_norm": 11.805512735858951, + "learning_rate": 6.377572174705127e-07, + "loss": 0.3552, + "step": 8189 + }, + { + "epoch": 4.4462540716612375, + "grad_norm": 11.017672732553946, + "learning_rate": 6.365219683798197e-07, + "loss": 0.3821, + "step": 8190 + }, + { + "epoch": 4.446796959826276, + "grad_norm": 13.850511847705805, + "learning_rate": 6.352878773790128e-07, + "loss": 0.3766, + "step": 8191 + }, + { + "epoch": 4.447339847991314, + "grad_norm": 11.980286969929338, + "learning_rate": 6.340549446207245e-07, + "loss": 0.3077, + "step": 8192 + }, + { + "epoch": 4.447882736156352, + "grad_norm": 12.593566275833842, + "learning_rate": 6.328231702574483e-07, + "loss": 0.3866, + "step": 8193 + }, + { + "epoch": 4.4484256243213895, + "grad_norm": 16.906397392736164, + "learning_rate": 6.315925544415324e-07, + "loss": 0.6914, + "step": 8194 + }, + { + "epoch": 4.448968512486427, + "grad_norm": 13.626782408196018, + "learning_rate": 6.303630973251795e-07, + "loss": 0.3687, + "step": 8195 + }, + { + "epoch": 4.449511400651466, + "grad_norm": 12.864518300164796, + "learning_rate": 6.291347990604524e-07, + "loss": 0.3668, + "step": 8196 + }, + { + "epoch": 4.450054288816504, + "grad_norm": 14.217617041842738, + "learning_rate": 6.279076597992683e-07, + "loss": 0.4862, + "step": 8197 + }, + { + "epoch": 4.450597176981542, + "grad_norm": 10.62667862211288, + "learning_rate": 6.266816796934016e-07, + "loss": 0.5427, + "step": 8198 + }, + { + "epoch": 4.45114006514658, + "grad_norm": 13.714935436924003, + "learning_rate": 6.254568588944832e-07, + "loss": 0.5349, + "step": 8199 + }, + { + "epoch": 4.451682953311618, + "grad_norm": 10.890652969571276, + "learning_rate": 6.242331975540017e-07, + "loss": 0.3663, + "step": 8200 + }, + { + "epoch": 4.452225841476656, + "grad_norm": 15.543319095934837, + "learning_rate": 6.230106958233006e-07, + "loss": 0.4374, + "step": 8201 + }, + { + "epoch": 4.452768729641694, + "grad_norm": 13.007089382527154, + "learning_rate": 6.217893538535813e-07, + "loss": 0.5068, + "step": 8202 + }, + { + "epoch": 4.453311617806731, + "grad_norm": 10.246126043767955, + "learning_rate": 6.205691717959017e-07, + "loss": 0.2651, + "step": 8203 + }, + { + "epoch": 4.45385450597177, + "grad_norm": 14.357913078114876, + "learning_rate": 6.193501498011756e-07, + "loss": 0.7913, + "step": 8204 + }, + { + "epoch": 4.454397394136808, + "grad_norm": 9.017306171160634, + "learning_rate": 6.181322880201734e-07, + "loss": 0.3367, + "step": 8205 + }, + { + "epoch": 4.454940282301846, + "grad_norm": 9.76917208852768, + "learning_rate": 6.169155866035226e-07, + "loss": 0.3176, + "step": 8206 + }, + { + "epoch": 4.455483170466884, + "grad_norm": 12.851456706342148, + "learning_rate": 6.15700045701706e-07, + "loss": 0.363, + "step": 8207 + }, + { + "epoch": 4.456026058631922, + "grad_norm": 12.21030677086537, + "learning_rate": 6.144856654650644e-07, + "loss": 0.3758, + "step": 8208 + }, + { + "epoch": 4.45656894679696, + "grad_norm": 12.105703346136071, + "learning_rate": 6.132724460437923e-07, + "loss": 0.4042, + "step": 8209 + }, + { + "epoch": 4.457111834961998, + "grad_norm": 9.134328872162778, + "learning_rate": 6.120603875879472e-07, + "loss": 0.2771, + "step": 8210 + }, + { + "epoch": 4.457654723127035, + "grad_norm": 11.516060563585224, + "learning_rate": 6.108494902474349e-07, + "loss": 0.2969, + "step": 8211 + }, + { + "epoch": 4.458197611292074, + "grad_norm": 12.99985841750232, + "learning_rate": 6.096397541720201e-07, + "loss": 0.4357, + "step": 8212 + }, + { + "epoch": 4.458740499457112, + "grad_norm": 17.46011030710575, + "learning_rate": 6.084311795113273e-07, + "loss": 0.3445, + "step": 8213 + }, + { + "epoch": 4.45928338762215, + "grad_norm": 13.1182184160893, + "learning_rate": 6.072237664148328e-07, + "loss": 0.4747, + "step": 8214 + }, + { + "epoch": 4.459826275787188, + "grad_norm": 12.980936251031281, + "learning_rate": 6.060175150318759e-07, + "loss": 0.3913, + "step": 8215 + }, + { + "epoch": 4.460369163952226, + "grad_norm": 17.440326310172946, + "learning_rate": 6.048124255116417e-07, + "loss": 0.6855, + "step": 8216 + }, + { + "epoch": 4.460912052117264, + "grad_norm": 9.688408464276451, + "learning_rate": 6.036084980031831e-07, + "loss": 0.258, + "step": 8217 + }, + { + "epoch": 4.461454940282302, + "grad_norm": 12.394258791386804, + "learning_rate": 6.024057326553978e-07, + "loss": 0.2525, + "step": 8218 + }, + { + "epoch": 4.461997828447339, + "grad_norm": 22.30304317136707, + "learning_rate": 6.012041296170523e-07, + "loss": 0.9278, + "step": 8219 + }, + { + "epoch": 4.462540716612378, + "grad_norm": 15.580394413629943, + "learning_rate": 6.000036890367577e-07, + "loss": 0.5253, + "step": 8220 + }, + { + "epoch": 4.463083604777416, + "grad_norm": 12.8104075419022, + "learning_rate": 5.988044110629864e-07, + "loss": 0.4224, + "step": 8221 + }, + { + "epoch": 4.463626492942454, + "grad_norm": 15.277018990574911, + "learning_rate": 5.97606295844072e-07, + "loss": 0.4786, + "step": 8222 + }, + { + "epoch": 4.464169381107492, + "grad_norm": 10.424466200619575, + "learning_rate": 5.964093435281937e-07, + "loss": 0.4286, + "step": 8223 + }, + { + "epoch": 4.46471226927253, + "grad_norm": 10.219964992820138, + "learning_rate": 5.952135542633975e-07, + "loss": 0.4709, + "step": 8224 + }, + { + "epoch": 4.465255157437568, + "grad_norm": 15.141978803020056, + "learning_rate": 5.94018928197575e-07, + "loss": 0.5745, + "step": 8225 + }, + { + "epoch": 4.465798045602606, + "grad_norm": 12.814624758233451, + "learning_rate": 5.928254654784837e-07, + "loss": 0.451, + "step": 8226 + }, + { + "epoch": 4.466340933767643, + "grad_norm": 11.052255193283386, + "learning_rate": 5.916331662537322e-07, + "loss": 0.353, + "step": 8227 + }, + { + "epoch": 4.466883821932682, + "grad_norm": 10.286265305406184, + "learning_rate": 5.904420306707848e-07, + "loss": 0.2147, + "step": 8228 + }, + { + "epoch": 4.46742671009772, + "grad_norm": 8.65857853802515, + "learning_rate": 5.892520588769646e-07, + "loss": 0.2347, + "step": 8229 + }, + { + "epoch": 4.467969598262758, + "grad_norm": 12.354192118862283, + "learning_rate": 5.880632510194473e-07, + "loss": 0.4059, + "step": 8230 + }, + { + "epoch": 4.468512486427796, + "grad_norm": 11.57115618341546, + "learning_rate": 5.868756072452686e-07, + "loss": 0.3671, + "step": 8231 + }, + { + "epoch": 4.469055374592834, + "grad_norm": 12.575647382464236, + "learning_rate": 5.856891277013154e-07, + "loss": 0.557, + "step": 8232 + }, + { + "epoch": 4.469598262757872, + "grad_norm": 15.538783933288979, + "learning_rate": 5.84503812534335e-07, + "loss": 0.3137, + "step": 8233 + }, + { + "epoch": 4.47014115092291, + "grad_norm": 10.745613891988105, + "learning_rate": 5.833196618909309e-07, + "loss": 0.243, + "step": 8234 + }, + { + "epoch": 4.470684039087947, + "grad_norm": 14.21692386565347, + "learning_rate": 5.821366759175573e-07, + "loss": 0.3973, + "step": 8235 + }, + { + "epoch": 4.471226927252986, + "grad_norm": 14.050911061509822, + "learning_rate": 5.809548547605304e-07, + "loss": 0.4404, + "step": 8236 + }, + { + "epoch": 4.471769815418024, + "grad_norm": 13.715645322282858, + "learning_rate": 5.797741985660188e-07, + "loss": 0.4384, + "step": 8237 + }, + { + "epoch": 4.472312703583062, + "grad_norm": 14.288594139401848, + "learning_rate": 5.78594707480048e-07, + "loss": 0.5333, + "step": 8238 + }, + { + "epoch": 4.4728555917481, + "grad_norm": 8.76616154408981, + "learning_rate": 5.77416381648499e-07, + "loss": 0.3037, + "step": 8239 + }, + { + "epoch": 4.473398479913138, + "grad_norm": 19.474660708412426, + "learning_rate": 5.762392212171086e-07, + "loss": 0.6935, + "step": 8240 + }, + { + "epoch": 4.473941368078176, + "grad_norm": 10.918792962612473, + "learning_rate": 5.750632263314715e-07, + "loss": 0.5132, + "step": 8241 + }, + { + "epoch": 4.474484256243214, + "grad_norm": 12.225987090529173, + "learning_rate": 5.738883971370357e-07, + "loss": 0.3604, + "step": 8242 + }, + { + "epoch": 4.4750271444082514, + "grad_norm": 15.19011773226198, + "learning_rate": 5.72714733779105e-07, + "loss": 0.5368, + "step": 8243 + }, + { + "epoch": 4.47557003257329, + "grad_norm": 15.315623852151898, + "learning_rate": 5.715422364028423e-07, + "loss": 0.6617, + "step": 8244 + }, + { + "epoch": 4.476112920738328, + "grad_norm": 14.565005688916782, + "learning_rate": 5.703709051532613e-07, + "loss": 0.3753, + "step": 8245 + }, + { + "epoch": 4.476655808903366, + "grad_norm": 13.306086406222777, + "learning_rate": 5.692007401752353e-07, + "loss": 0.297, + "step": 8246 + }, + { + "epoch": 4.477198697068404, + "grad_norm": 11.685314077983634, + "learning_rate": 5.680317416134917e-07, + "loss": 0.3943, + "step": 8247 + }, + { + "epoch": 4.477741585233442, + "grad_norm": 15.78610523944711, + "learning_rate": 5.668639096126172e-07, + "loss": 0.568, + "step": 8248 + }, + { + "epoch": 4.47828447339848, + "grad_norm": 17.678808024002272, + "learning_rate": 5.65697244317045e-07, + "loss": 0.5074, + "step": 8249 + }, + { + "epoch": 4.478827361563518, + "grad_norm": 13.613139148009028, + "learning_rate": 5.645317458710775e-07, + "loss": 0.4749, + "step": 8250 + }, + { + "epoch": 4.4793702497285555, + "grad_norm": 11.810323226840032, + "learning_rate": 5.633674144188594e-07, + "loss": 0.3615, + "step": 8251 + }, + { + "epoch": 4.479913137893594, + "grad_norm": 12.479823316808458, + "learning_rate": 5.62204250104399e-07, + "loss": 0.3417, + "step": 8252 + }, + { + "epoch": 4.480456026058632, + "grad_norm": 10.507991166659478, + "learning_rate": 5.610422530715597e-07, + "loss": 0.2787, + "step": 8253 + }, + { + "epoch": 4.48099891422367, + "grad_norm": 11.553633131438758, + "learning_rate": 5.59881423464056e-07, + "loss": 0.304, + "step": 8254 + }, + { + "epoch": 4.481541802388708, + "grad_norm": 12.545712767514072, + "learning_rate": 5.587217614254658e-07, + "loss": 0.5601, + "step": 8255 + }, + { + "epoch": 4.482084690553746, + "grad_norm": 14.917437145721351, + "learning_rate": 5.575632670992126e-07, + "loss": 0.6768, + "step": 8256 + }, + { + "epoch": 4.482627578718784, + "grad_norm": 20.337480802460437, + "learning_rate": 5.56405940628586e-07, + "loss": 0.602, + "step": 8257 + }, + { + "epoch": 4.483170466883822, + "grad_norm": 11.359050219678316, + "learning_rate": 5.552497821567216e-07, + "loss": 0.3441, + "step": 8258 + }, + { + "epoch": 4.4837133550488595, + "grad_norm": 14.281903579588091, + "learning_rate": 5.540947918266171e-07, + "loss": 0.3607, + "step": 8259 + }, + { + "epoch": 4.484256243213898, + "grad_norm": 10.394969225634863, + "learning_rate": 5.52940969781125e-07, + "loss": 0.2491, + "step": 8260 + }, + { + "epoch": 4.484799131378936, + "grad_norm": 15.233116056219048, + "learning_rate": 5.517883161629478e-07, + "loss": 0.4715, + "step": 8261 + }, + { + "epoch": 4.485342019543974, + "grad_norm": 13.370733413078632, + "learning_rate": 5.506368311146526e-07, + "loss": 0.5931, + "step": 8262 + }, + { + "epoch": 4.485884907709012, + "grad_norm": 11.613070671019594, + "learning_rate": 5.49486514778651e-07, + "loss": 0.3663, + "step": 8263 + }, + { + "epoch": 4.48642779587405, + "grad_norm": 12.305931901351924, + "learning_rate": 5.483373672972215e-07, + "loss": 0.3761, + "step": 8264 + }, + { + "epoch": 4.486970684039088, + "grad_norm": 11.00494181177833, + "learning_rate": 5.471893888124891e-07, + "loss": 0.3874, + "step": 8265 + }, + { + "epoch": 4.487513572204126, + "grad_norm": 12.920317758044833, + "learning_rate": 5.460425794664381e-07, + "loss": 0.5323, + "step": 8266 + }, + { + "epoch": 4.4880564603691635, + "grad_norm": 16.420345201613245, + "learning_rate": 5.448969394009096e-07, + "loss": 0.4206, + "step": 8267 + }, + { + "epoch": 4.488599348534202, + "grad_norm": 14.573525619555683, + "learning_rate": 5.437524687575957e-07, + "loss": 0.4514, + "step": 8268 + }, + { + "epoch": 4.48914223669924, + "grad_norm": 22.781040116815415, + "learning_rate": 5.426091676780499e-07, + "loss": 0.7505, + "step": 8269 + }, + { + "epoch": 4.489685124864278, + "grad_norm": 8.882692706900785, + "learning_rate": 5.414670363036722e-07, + "loss": 0.2729, + "step": 8270 + }, + { + "epoch": 4.490228013029316, + "grad_norm": 14.363794979804172, + "learning_rate": 5.403260747757266e-07, + "loss": 0.453, + "step": 8271 + }, + { + "epoch": 4.490770901194354, + "grad_norm": 10.245019537386296, + "learning_rate": 5.391862832353289e-07, + "loss": 0.3372, + "step": 8272 + }, + { + "epoch": 4.491313789359392, + "grad_norm": 12.50265023247342, + "learning_rate": 5.380476618234498e-07, + "loss": 0.4035, + "step": 8273 + }, + { + "epoch": 4.49185667752443, + "grad_norm": 11.646457698141697, + "learning_rate": 5.369102106809165e-07, + "loss": 0.3997, + "step": 8274 + }, + { + "epoch": 4.4923995656894675, + "grad_norm": 9.464982035604422, + "learning_rate": 5.3577392994841e-07, + "loss": 0.2514, + "step": 8275 + }, + { + "epoch": 4.492942453854506, + "grad_norm": 17.529725496816823, + "learning_rate": 5.346388197664675e-07, + "loss": 0.4352, + "step": 8276 + }, + { + "epoch": 4.493485342019544, + "grad_norm": 17.57064929965146, + "learning_rate": 5.335048802754817e-07, + "loss": 0.4559, + "step": 8277 + }, + { + "epoch": 4.494028230184582, + "grad_norm": 13.71860363844268, + "learning_rate": 5.323721116156999e-07, + "loss": 0.3552, + "step": 8278 + }, + { + "epoch": 4.49457111834962, + "grad_norm": 10.880133558033618, + "learning_rate": 5.312405139272237e-07, + "loss": 0.3963, + "step": 8279 + }, + { + "epoch": 4.495114006514658, + "grad_norm": 20.247587867745704, + "learning_rate": 5.30110087350012e-07, + "loss": 0.5438, + "step": 8280 + }, + { + "epoch": 4.495656894679696, + "grad_norm": 9.454005269809635, + "learning_rate": 5.289808320238776e-07, + "loss": 0.2168, + "step": 8281 + }, + { + "epoch": 4.496199782844734, + "grad_norm": 13.194855975031016, + "learning_rate": 5.278527480884888e-07, + "loss": 0.3221, + "step": 8282 + }, + { + "epoch": 4.4967426710097715, + "grad_norm": 16.741747646625797, + "learning_rate": 5.267258356833671e-07, + "loss": 0.5343, + "step": 8283 + }, + { + "epoch": 4.49728555917481, + "grad_norm": 17.72898677538945, + "learning_rate": 5.256000949478934e-07, + "loss": 0.5213, + "step": 8284 + }, + { + "epoch": 4.497828447339848, + "grad_norm": 23.94089901495696, + "learning_rate": 5.244755260212986e-07, + "loss": 1.1702, + "step": 8285 + }, + { + "epoch": 4.498371335504886, + "grad_norm": 15.217551359960183, + "learning_rate": 5.233521290426746e-07, + "loss": 0.667, + "step": 8286 + }, + { + "epoch": 4.498914223669924, + "grad_norm": 15.41827808885883, + "learning_rate": 5.222299041509604e-07, + "loss": 0.3961, + "step": 8287 + }, + { + "epoch": 4.499457111834962, + "grad_norm": 10.397823177247577, + "learning_rate": 5.211088514849594e-07, + "loss": 0.3334, + "step": 8288 + }, + { + "epoch": 4.5, + "grad_norm": 10.913567291837683, + "learning_rate": 5.199889711833217e-07, + "loss": 0.3493, + "step": 8289 + }, + { + "epoch": 4.500542888165038, + "grad_norm": 12.37886318983014, + "learning_rate": 5.188702633845566e-07, + "loss": 0.3399, + "step": 8290 + }, + { + "epoch": 4.501085776330076, + "grad_norm": 11.119221436517403, + "learning_rate": 5.177527282270278e-07, + "loss": 0.2982, + "step": 8291 + }, + { + "epoch": 4.501628664495114, + "grad_norm": 11.838482844225242, + "learning_rate": 5.166363658489537e-07, + "loss": 0.4248, + "step": 8292 + }, + { + "epoch": 4.502171552660152, + "grad_norm": 13.111861380868095, + "learning_rate": 5.155211763884094e-07, + "loss": 0.521, + "step": 8293 + }, + { + "epoch": 4.50271444082519, + "grad_norm": 12.88810562354075, + "learning_rate": 5.1440715998332e-07, + "loss": 0.3804, + "step": 8294 + }, + { + "epoch": 4.5032573289902285, + "grad_norm": 10.514953303860349, + "learning_rate": 5.132943167714744e-07, + "loss": 0.3674, + "step": 8295 + }, + { + "epoch": 4.503800217155266, + "grad_norm": 10.023365825501816, + "learning_rate": 5.121826468905033e-07, + "loss": 0.5211, + "step": 8296 + }, + { + "epoch": 4.504343105320304, + "grad_norm": 11.606294239792984, + "learning_rate": 5.110721504779059e-07, + "loss": 0.7045, + "step": 8297 + }, + { + "epoch": 4.504885993485342, + "grad_norm": 15.029276069995117, + "learning_rate": 5.099628276710278e-07, + "loss": 0.602, + "step": 8298 + }, + { + "epoch": 4.50542888165038, + "grad_norm": 14.866091084726918, + "learning_rate": 5.088546786070714e-07, + "loss": 0.4688, + "step": 8299 + }, + { + "epoch": 4.505971769815418, + "grad_norm": 12.486305532545822, + "learning_rate": 5.07747703423097e-07, + "loss": 0.6218, + "step": 8300 + }, + { + "epoch": 4.506514657980456, + "grad_norm": 13.755190146974774, + "learning_rate": 5.066419022560121e-07, + "loss": 0.5051, + "step": 8301 + }, + { + "epoch": 4.507057546145494, + "grad_norm": 17.222866237868953, + "learning_rate": 5.055372752425902e-07, + "loss": 0.5394, + "step": 8302 + }, + { + "epoch": 4.5076004343105325, + "grad_norm": 13.245212020095746, + "learning_rate": 5.044338225194467e-07, + "loss": 0.2677, + "step": 8303 + }, + { + "epoch": 4.50814332247557, + "grad_norm": 17.027782293780568, + "learning_rate": 5.033315442230636e-07, + "loss": 0.6007, + "step": 8304 + }, + { + "epoch": 4.508686210640608, + "grad_norm": 13.708233013149469, + "learning_rate": 5.022304404897693e-07, + "loss": 0.623, + "step": 8305 + }, + { + "epoch": 4.509229098805646, + "grad_norm": 12.927772993543751, + "learning_rate": 5.011305114557519e-07, + "loss": 0.3889, + "step": 8306 + }, + { + "epoch": 4.509771986970684, + "grad_norm": 14.732891345640356, + "learning_rate": 5.000317572570523e-07, + "loss": 0.3646, + "step": 8307 + }, + { + "epoch": 4.510314875135722, + "grad_norm": 12.052951834491875, + "learning_rate": 4.989341780295654e-07, + "loss": 0.3066, + "step": 8308 + }, + { + "epoch": 4.51085776330076, + "grad_norm": 12.16750268825224, + "learning_rate": 4.978377739090424e-07, + "loss": 0.317, + "step": 8309 + }, + { + "epoch": 4.511400651465798, + "grad_norm": 15.533245706529485, + "learning_rate": 4.96742545031087e-07, + "loss": 0.2653, + "step": 8310 + }, + { + "epoch": 4.5119435396308365, + "grad_norm": 15.668556903572544, + "learning_rate": 4.956484915311599e-07, + "loss": 0.3729, + "step": 8311 + }, + { + "epoch": 4.512486427795874, + "grad_norm": 17.819603638078444, + "learning_rate": 4.945556135445739e-07, + "loss": 0.8422, + "step": 8312 + }, + { + "epoch": 4.513029315960912, + "grad_norm": 11.03718224282426, + "learning_rate": 4.934639112064998e-07, + "loss": 0.3522, + "step": 8313 + }, + { + "epoch": 4.51357220412595, + "grad_norm": 17.833077129344446, + "learning_rate": 4.923733846519607e-07, + "loss": 0.3477, + "step": 8314 + }, + { + "epoch": 4.514115092290988, + "grad_norm": 13.75370085535341, + "learning_rate": 4.912840340158343e-07, + "loss": 0.4303, + "step": 8315 + }, + { + "epoch": 4.514657980456026, + "grad_norm": 13.416980517253936, + "learning_rate": 4.901958594328527e-07, + "loss": 0.5717, + "step": 8316 + }, + { + "epoch": 4.515200868621064, + "grad_norm": 13.697716950123397, + "learning_rate": 4.891088610376038e-07, + "loss": 0.4465, + "step": 8317 + }, + { + "epoch": 4.515743756786102, + "grad_norm": 11.647982736650992, + "learning_rate": 4.880230389645291e-07, + "loss": 0.3537, + "step": 8318 + }, + { + "epoch": 4.5162866449511405, + "grad_norm": 17.71700440336686, + "learning_rate": 4.869383933479254e-07, + "loss": 0.5677, + "step": 8319 + }, + { + "epoch": 4.516829533116178, + "grad_norm": 8.473597132737073, + "learning_rate": 4.858549243219423e-07, + "loss": 0.2425, + "step": 8320 + }, + { + "epoch": 4.517372421281216, + "grad_norm": 9.186745419175947, + "learning_rate": 4.847726320205847e-07, + "loss": 0.2741, + "step": 8321 + }, + { + "epoch": 4.517915309446254, + "grad_norm": 10.521404049591816, + "learning_rate": 4.836915165777134e-07, + "loss": 0.3232, + "step": 8322 + }, + { + "epoch": 4.518458197611292, + "grad_norm": 18.349893458827204, + "learning_rate": 4.826115781270424e-07, + "loss": 0.5938, + "step": 8323 + }, + { + "epoch": 4.51900108577633, + "grad_norm": 10.506129387254754, + "learning_rate": 4.815328168021405e-07, + "loss": 0.2614, + "step": 8324 + }, + { + "epoch": 4.519543973941368, + "grad_norm": 13.11135730425014, + "learning_rate": 4.804552327364276e-07, + "loss": 0.3003, + "step": 8325 + }, + { + "epoch": 4.520086862106406, + "grad_norm": 9.56019008291403, + "learning_rate": 4.79378826063186e-07, + "loss": 0.2494, + "step": 8326 + }, + { + "epoch": 4.5206297502714445, + "grad_norm": 15.77492779940971, + "learning_rate": 4.783035969155425e-07, + "loss": 0.6032, + "step": 8327 + }, + { + "epoch": 4.521172638436482, + "grad_norm": 12.024645355321516, + "learning_rate": 4.772295454264886e-07, + "loss": 0.3562, + "step": 8328 + }, + { + "epoch": 4.52171552660152, + "grad_norm": 17.154707742654768, + "learning_rate": 4.7615667172885903e-07, + "loss": 0.4941, + "step": 8329 + }, + { + "epoch": 4.522258414766558, + "grad_norm": 11.691277186483006, + "learning_rate": 4.750849759553511e-07, + "loss": 0.3484, + "step": 8330 + }, + { + "epoch": 4.522801302931596, + "grad_norm": 11.471731467185506, + "learning_rate": 4.7401445823851534e-07, + "loss": 0.4616, + "step": 8331 + }, + { + "epoch": 4.523344191096634, + "grad_norm": 14.556670225131105, + "learning_rate": 4.729451187107514e-07, + "loss": 0.3747, + "step": 8332 + }, + { + "epoch": 4.523887079261672, + "grad_norm": 10.060443485377643, + "learning_rate": 4.718769575043214e-07, + "loss": 0.2432, + "step": 8333 + }, + { + "epoch": 4.52442996742671, + "grad_norm": 17.265452310056222, + "learning_rate": 4.708099747513328e-07, + "loss": 0.5352, + "step": 8334 + }, + { + "epoch": 4.5249728555917486, + "grad_norm": 14.537157483573262, + "learning_rate": 4.6974417058375574e-07, + "loss": 0.3997, + "step": 8335 + }, + { + "epoch": 4.525515743756786, + "grad_norm": 12.731997994806079, + "learning_rate": 4.68679545133407e-07, + "loss": 0.4316, + "step": 8336 + }, + { + "epoch": 4.526058631921824, + "grad_norm": 16.64941354972263, + "learning_rate": 4.676160985319633e-07, + "loss": 0.5068, + "step": 8337 + }, + { + "epoch": 4.526601520086862, + "grad_norm": 14.967442903073584, + "learning_rate": 4.6655383091095405e-07, + "loss": 0.5022, + "step": 8338 + }, + { + "epoch": 4.5271444082519, + "grad_norm": 14.760527489157035, + "learning_rate": 4.654927424017586e-07, + "loss": 0.5753, + "step": 8339 + }, + { + "epoch": 4.527687296416938, + "grad_norm": 18.12528620784058, + "learning_rate": 4.644328331356196e-07, + "loss": 0.5, + "step": 8340 + }, + { + "epoch": 4.528230184581976, + "grad_norm": 11.591072114682094, + "learning_rate": 4.633741032436223e-07, + "loss": 0.458, + "step": 8341 + }, + { + "epoch": 4.528773072747014, + "grad_norm": 14.15798620351717, + "learning_rate": 4.623165528567164e-07, + "loss": 0.3498, + "step": 8342 + }, + { + "epoch": 4.529315960912053, + "grad_norm": 23.12722823259664, + "learning_rate": 4.6126018210569946e-07, + "loss": 0.5397, + "step": 8343 + }, + { + "epoch": 4.52985884907709, + "grad_norm": 14.338174694861847, + "learning_rate": 4.602049911212259e-07, + "loss": 0.5037, + "step": 8344 + }, + { + "epoch": 4.530401737242128, + "grad_norm": 13.947445388030966, + "learning_rate": 4.5915098003380343e-07, + "loss": 0.7601, + "step": 8345 + }, + { + "epoch": 4.530944625407166, + "grad_norm": 15.423631300358158, + "learning_rate": 4.5809814897379345e-07, + "loss": 0.3918, + "step": 8346 + }, + { + "epoch": 4.531487513572204, + "grad_norm": 11.526636704758273, + "learning_rate": 4.570464980714129e-07, + "loss": 0.3192, + "step": 8347 + }, + { + "epoch": 4.532030401737242, + "grad_norm": 10.597564388470815, + "learning_rate": 4.559960274567299e-07, + "loss": 0.2513, + "step": 8348 + }, + { + "epoch": 4.53257328990228, + "grad_norm": 17.637678470826174, + "learning_rate": 4.5494673725966833e-07, + "loss": 0.3689, + "step": 8349 + }, + { + "epoch": 4.533116178067318, + "grad_norm": 15.332131840534869, + "learning_rate": 4.5389862761000767e-07, + "loss": 0.5437, + "step": 8350 + }, + { + "epoch": 4.533659066232357, + "grad_norm": 18.544251971087917, + "learning_rate": 4.5285169863737874e-07, + "loss": 0.9338, + "step": 8351 + }, + { + "epoch": 4.534201954397394, + "grad_norm": 14.845994288337273, + "learning_rate": 4.5180595047126795e-07, + "loss": 0.4528, + "step": 8352 + }, + { + "epoch": 4.534744842562432, + "grad_norm": 10.966600100926046, + "learning_rate": 4.5076138324101516e-07, + "loss": 0.3213, + "step": 8353 + }, + { + "epoch": 4.53528773072747, + "grad_norm": 13.979375013420652, + "learning_rate": 4.4971799707581254e-07, + "loss": 0.4328, + "step": 8354 + }, + { + "epoch": 4.535830618892508, + "grad_norm": 13.199579774530013, + "learning_rate": 4.4867579210470915e-07, + "loss": 0.4421, + "step": 8355 + }, + { + "epoch": 4.536373507057546, + "grad_norm": 15.679271986245535, + "learning_rate": 4.4763476845660627e-07, + "loss": 0.686, + "step": 8356 + }, + { + "epoch": 4.536916395222584, + "grad_norm": 10.32839293787231, + "learning_rate": 4.465949262602609e-07, + "loss": 0.3593, + "step": 8357 + }, + { + "epoch": 4.537459283387622, + "grad_norm": 19.97455202826578, + "learning_rate": 4.4555626564428024e-07, + "loss": 0.6335, + "step": 8358 + }, + { + "epoch": 4.538002171552661, + "grad_norm": 17.035674620541695, + "learning_rate": 4.44518786737127e-07, + "loss": 0.5835, + "step": 8359 + }, + { + "epoch": 4.538545059717698, + "grad_norm": 15.009806208164479, + "learning_rate": 4.434824896671208e-07, + "loss": 0.4077, + "step": 8360 + }, + { + "epoch": 4.539087947882736, + "grad_norm": 13.778069593523517, + "learning_rate": 4.4244737456243025e-07, + "loss": 0.3833, + "step": 8361 + }, + { + "epoch": 4.539630836047774, + "grad_norm": 12.494743874325179, + "learning_rate": 4.4141344155108066e-07, + "loss": 0.2996, + "step": 8362 + }, + { + "epoch": 4.540173724212812, + "grad_norm": 13.814732420411174, + "learning_rate": 4.4038069076094983e-07, + "loss": 0.319, + "step": 8363 + }, + { + "epoch": 4.54071661237785, + "grad_norm": 18.350946091244264, + "learning_rate": 4.3934912231977433e-07, + "loss": 0.7825, + "step": 8364 + }, + { + "epoch": 4.541259500542888, + "grad_norm": 11.565882281033971, + "learning_rate": 4.383187363551333e-07, + "loss": 0.3073, + "step": 8365 + }, + { + "epoch": 4.541802388707926, + "grad_norm": 16.013246124631713, + "learning_rate": 4.3728953299447265e-07, + "loss": 0.6382, + "step": 8366 + }, + { + "epoch": 4.542345276872965, + "grad_norm": 13.181941989917167, + "learning_rate": 4.362615123650815e-07, + "loss": 0.3552, + "step": 8367 + }, + { + "epoch": 4.542888165038002, + "grad_norm": 11.189597225295508, + "learning_rate": 4.352346745941083e-07, + "loss": 0.3848, + "step": 8368 + }, + { + "epoch": 4.54343105320304, + "grad_norm": 14.67310600028813, + "learning_rate": 4.342090198085569e-07, + "loss": 0.3936, + "step": 8369 + }, + { + "epoch": 4.543973941368078, + "grad_norm": 13.25839722725421, + "learning_rate": 4.33184548135277e-07, + "loss": 0.499, + "step": 8370 + }, + { + "epoch": 4.544516829533116, + "grad_norm": 9.9899345545398, + "learning_rate": 4.3216125970098164e-07, + "loss": 0.366, + "step": 8371 + }, + { + "epoch": 4.545059717698154, + "grad_norm": 8.836049199804664, + "learning_rate": 4.3113915463222855e-07, + "loss": 0.2981, + "step": 8372 + }, + { + "epoch": 4.545602605863192, + "grad_norm": 10.295970227608981, + "learning_rate": 4.3011823305543766e-07, + "loss": 0.3629, + "step": 8373 + }, + { + "epoch": 4.54614549402823, + "grad_norm": 12.00515544737705, + "learning_rate": 4.290984950968724e-07, + "loss": 0.3902, + "step": 8374 + }, + { + "epoch": 4.546688382193269, + "grad_norm": 15.20456661725038, + "learning_rate": 4.280799408826619e-07, + "loss": 0.3767, + "step": 8375 + }, + { + "epoch": 4.547231270358306, + "grad_norm": 16.982779080955037, + "learning_rate": 4.270625705387788e-07, + "loss": 0.3967, + "step": 8376 + }, + { + "epoch": 4.547774158523344, + "grad_norm": 10.63460587566383, + "learning_rate": 4.2604638419105336e-07, + "loss": 0.5597, + "step": 8377 + }, + { + "epoch": 4.548317046688382, + "grad_norm": 14.423686882901746, + "learning_rate": 4.250313819651719e-07, + "loss": 0.4968, + "step": 8378 + }, + { + "epoch": 4.54885993485342, + "grad_norm": 14.591468749508062, + "learning_rate": 4.240175639866662e-07, + "loss": 0.4191, + "step": 8379 + }, + { + "epoch": 4.549402823018458, + "grad_norm": 13.309197670984693, + "learning_rate": 4.230049303809325e-07, + "loss": 0.5255, + "step": 8380 + }, + { + "epoch": 4.549945711183496, + "grad_norm": 17.125364753150805, + "learning_rate": 4.219934812732107e-07, + "loss": 0.8478, + "step": 8381 + }, + { + "epoch": 4.550488599348534, + "grad_norm": 14.128057894087553, + "learning_rate": 4.209832167886019e-07, + "loss": 0.3886, + "step": 8382 + }, + { + "epoch": 4.551031487513573, + "grad_norm": 15.17581204817665, + "learning_rate": 4.199741370520538e-07, + "loss": 0.5868, + "step": 8383 + }, + { + "epoch": 4.5515743756786105, + "grad_norm": 13.255434170354311, + "learning_rate": 4.1896624218837444e-07, + "loss": 0.2633, + "step": 8384 + }, + { + "epoch": 4.552117263843648, + "grad_norm": 10.395013531026416, + "learning_rate": 4.1795953232221966e-07, + "loss": 0.2439, + "step": 8385 + }, + { + "epoch": 4.552660152008686, + "grad_norm": 16.020764006465836, + "learning_rate": 4.169540075781009e-07, + "loss": 0.4021, + "step": 8386 + }, + { + "epoch": 4.553203040173724, + "grad_norm": 16.237245900097843, + "learning_rate": 4.159496680803832e-07, + "loss": 0.4336, + "step": 8387 + }, + { + "epoch": 4.5537459283387625, + "grad_norm": 13.197573051421786, + "learning_rate": 4.14946513953286e-07, + "loss": 0.361, + "step": 8388 + }, + { + "epoch": 4.5542888165038, + "grad_norm": 14.796764385747446, + "learning_rate": 4.1394454532088015e-07, + "loss": 0.3395, + "step": 8389 + }, + { + "epoch": 4.554831704668838, + "grad_norm": 14.559402156578185, + "learning_rate": 4.129437623070909e-07, + "loss": 0.3893, + "step": 8390 + }, + { + "epoch": 4.555374592833877, + "grad_norm": 17.202126229361323, + "learning_rate": 4.1194416503569703e-07, + "loss": 0.7156, + "step": 8391 + }, + { + "epoch": 4.5559174809989145, + "grad_norm": 9.358596584111847, + "learning_rate": 4.109457536303285e-07, + "loss": 0.4116, + "step": 8392 + }, + { + "epoch": 4.556460369163952, + "grad_norm": 12.636931486569619, + "learning_rate": 4.099485282144733e-07, + "loss": 0.4355, + "step": 8393 + }, + { + "epoch": 4.55700325732899, + "grad_norm": 13.073013470376315, + "learning_rate": 4.089524889114671e-07, + "loss": 0.2467, + "step": 8394 + }, + { + "epoch": 4.557546145494028, + "grad_norm": 13.801266347348164, + "learning_rate": 4.0795763584450366e-07, + "loss": 0.7289, + "step": 8395 + }, + { + "epoch": 4.5580890336590665, + "grad_norm": 9.171162460594205, + "learning_rate": 4.069639691366267e-07, + "loss": 0.2739, + "step": 8396 + }, + { + "epoch": 4.558631921824104, + "grad_norm": 12.729495446250427, + "learning_rate": 4.059714889107369e-07, + "loss": 0.4215, + "step": 8397 + }, + { + "epoch": 4.559174809989142, + "grad_norm": 18.76182078049615, + "learning_rate": 4.0498019528958265e-07, + "loss": 0.4525, + "step": 8398 + }, + { + "epoch": 4.559717698154181, + "grad_norm": 10.419041741316235, + "learning_rate": 4.0399008839576927e-07, + "loss": 0.2068, + "step": 8399 + }, + { + "epoch": 4.5602605863192185, + "grad_norm": 16.500852833343252, + "learning_rate": 4.0300116835175653e-07, + "loss": 0.5465, + "step": 8400 + }, + { + "epoch": 4.560803474484256, + "grad_norm": 13.29028900926885, + "learning_rate": 4.020134352798533e-07, + "loss": 0.4574, + "step": 8401 + }, + { + "epoch": 4.561346362649294, + "grad_norm": 14.113439209552388, + "learning_rate": 4.010268893022273e-07, + "loss": 0.3419, + "step": 8402 + }, + { + "epoch": 4.561889250814332, + "grad_norm": 11.962313774958856, + "learning_rate": 4.0004153054089223e-07, + "loss": 0.2653, + "step": 8403 + }, + { + "epoch": 4.5624321389793705, + "grad_norm": 15.681458008075632, + "learning_rate": 3.990573591177227e-07, + "loss": 0.9879, + "step": 8404 + }, + { + "epoch": 4.562975027144408, + "grad_norm": 7.896889702285753, + "learning_rate": 3.9807437515443915e-07, + "loss": 0.266, + "step": 8405 + }, + { + "epoch": 4.563517915309446, + "grad_norm": 14.317793413141796, + "learning_rate": 3.97092578772621e-07, + "loss": 0.5589, + "step": 8406 + }, + { + "epoch": 4.564060803474485, + "grad_norm": 14.701295872646838, + "learning_rate": 3.9611197009369774e-07, + "loss": 0.4754, + "step": 8407 + }, + { + "epoch": 4.5646036916395225, + "grad_norm": 9.751091449679476, + "learning_rate": 3.9513254923895017e-07, + "loss": 0.3138, + "step": 8408 + }, + { + "epoch": 4.56514657980456, + "grad_norm": 13.368140777634292, + "learning_rate": 3.9415431632951917e-07, + "loss": 0.4099, + "step": 8409 + }, + { + "epoch": 4.565689467969598, + "grad_norm": 15.380332314106305, + "learning_rate": 3.9317727148638906e-07, + "loss": 0.528, + "step": 8410 + }, + { + "epoch": 4.566232356134636, + "grad_norm": 11.966998113136334, + "learning_rate": 3.9220141483040873e-07, + "loss": 0.3129, + "step": 8411 + }, + { + "epoch": 4.5667752442996745, + "grad_norm": 10.655786798583739, + "learning_rate": 3.912267464822661e-07, + "loss": 0.3956, + "step": 8412 + }, + { + "epoch": 4.567318132464712, + "grad_norm": 11.348329273545943, + "learning_rate": 3.9025326656251583e-07, + "loss": 0.2485, + "step": 8413 + }, + { + "epoch": 4.56786102062975, + "grad_norm": 10.93357120051505, + "learning_rate": 3.892809751915572e-07, + "loss": 0.1911, + "step": 8414 + }, + { + "epoch": 4.568403908794789, + "grad_norm": 16.65942853342247, + "learning_rate": 3.8830987248964394e-07, + "loss": 0.5467, + "step": 8415 + }, + { + "epoch": 4.5689467969598265, + "grad_norm": 14.213503704674388, + "learning_rate": 3.8733995857688664e-07, + "loss": 0.5076, + "step": 8416 + }, + { + "epoch": 4.569489685124864, + "grad_norm": 9.07435692786808, + "learning_rate": 3.8637123357324057e-07, + "loss": 0.314, + "step": 8417 + }, + { + "epoch": 4.570032573289902, + "grad_norm": 13.204052175322088, + "learning_rate": 3.8540369759852313e-07, + "loss": 0.4221, + "step": 8418 + }, + { + "epoch": 4.57057546145494, + "grad_norm": 13.165097473842113, + "learning_rate": 3.8443735077239975e-07, + "loss": 0.5271, + "step": 8419 + }, + { + "epoch": 4.5711183496199785, + "grad_norm": 14.175386650762487, + "learning_rate": 3.8347219321439033e-07, + "loss": 0.8023, + "step": 8420 + }, + { + "epoch": 4.571661237785016, + "grad_norm": 14.868459939871421, + "learning_rate": 3.825082250438661e-07, + "loss": 0.4034, + "step": 8421 + }, + { + "epoch": 4.572204125950054, + "grad_norm": 10.271953811670828, + "learning_rate": 3.8154544638005275e-07, + "loss": 0.2737, + "step": 8422 + }, + { + "epoch": 4.572747014115093, + "grad_norm": 12.599854900654838, + "learning_rate": 3.805838573420273e-07, + "loss": 0.3811, + "step": 8423 + }, + { + "epoch": 4.5732899022801305, + "grad_norm": 12.34104633408041, + "learning_rate": 3.7962345804872235e-07, + "loss": 0.4701, + "step": 8424 + }, + { + "epoch": 4.573832790445168, + "grad_norm": 16.049446226050716, + "learning_rate": 3.786642486189207e-07, + "loss": 0.6162, + "step": 8425 + }, + { + "epoch": 4.574375678610206, + "grad_norm": 12.429280513300323, + "learning_rate": 3.7770622917125857e-07, + "loss": 0.3345, + "step": 8426 + }, + { + "epoch": 4.574918566775244, + "grad_norm": 13.767089346919395, + "learning_rate": 3.7674939982422555e-07, + "loss": 0.4666, + "step": 8427 + }, + { + "epoch": 4.575461454940283, + "grad_norm": 11.650625569284712, + "learning_rate": 3.757937606961648e-07, + "loss": 0.446, + "step": 8428 + }, + { + "epoch": 4.57600434310532, + "grad_norm": 14.190570945854875, + "learning_rate": 3.7483931190526956e-07, + "loss": 0.5539, + "step": 8429 + }, + { + "epoch": 4.576547231270358, + "grad_norm": 14.141845515233648, + "learning_rate": 3.7388605356958873e-07, + "loss": 0.5402, + "step": 8430 + }, + { + "epoch": 4.577090119435397, + "grad_norm": 9.796163273061495, + "learning_rate": 3.7293398580702244e-07, + "loss": 0.1934, + "step": 8431 + }, + { + "epoch": 4.577633007600435, + "grad_norm": 14.965355025993395, + "learning_rate": 3.719831087353243e-07, + "loss": 0.6887, + "step": 8432 + }, + { + "epoch": 4.578175895765472, + "grad_norm": 10.131802869294864, + "learning_rate": 3.7103342247210015e-07, + "loss": 0.305, + "step": 8433 + }, + { + "epoch": 4.57871878393051, + "grad_norm": 13.343860286662204, + "learning_rate": 3.700849271348073e-07, + "loss": 0.573, + "step": 8434 + }, + { + "epoch": 4.579261672095548, + "grad_norm": 15.966222976200056, + "learning_rate": 3.691376228407606e-07, + "loss": 0.57, + "step": 8435 + }, + { + "epoch": 4.579804560260587, + "grad_norm": 11.11642134835898, + "learning_rate": 3.6819150970712093e-07, + "loss": 0.4264, + "step": 8436 + }, + { + "epoch": 4.580347448425624, + "grad_norm": 12.39535666720846, + "learning_rate": 3.672465878509057e-07, + "loss": 0.2841, + "step": 8437 + }, + { + "epoch": 4.580890336590662, + "grad_norm": 19.74473442455948, + "learning_rate": 3.66302857388986e-07, + "loss": 0.6739, + "step": 8438 + }, + { + "epoch": 4.581433224755701, + "grad_norm": 19.226999685559647, + "learning_rate": 3.653603184380805e-07, + "loss": 0.5112, + "step": 8439 + }, + { + "epoch": 4.581976112920739, + "grad_norm": 14.270097098199749, + "learning_rate": 3.644189711147683e-07, + "loss": 0.3839, + "step": 8440 + }, + { + "epoch": 4.582519001085776, + "grad_norm": 22.049747355061914, + "learning_rate": 3.634788155354729e-07, + "loss": 1.0213, + "step": 8441 + }, + { + "epoch": 4.583061889250814, + "grad_norm": 13.290670656397504, + "learning_rate": 3.62539851816478e-07, + "loss": 0.3051, + "step": 8442 + }, + { + "epoch": 4.583604777415852, + "grad_norm": 12.069309467099986, + "learning_rate": 3.616020800739117e-07, + "loss": 0.3633, + "step": 8443 + }, + { + "epoch": 4.584147665580891, + "grad_norm": 12.59014501860219, + "learning_rate": 3.606655004237647e-07, + "loss": 0.4773, + "step": 8444 + }, + { + "epoch": 4.584690553745928, + "grad_norm": 19.27546265345427, + "learning_rate": 3.5973011298186756e-07, + "loss": 0.6292, + "step": 8445 + }, + { + "epoch": 4.585233441910966, + "grad_norm": 12.759660559474787, + "learning_rate": 3.5879591786391667e-07, + "loss": 0.4249, + "step": 8446 + }, + { + "epoch": 4.585776330076005, + "grad_norm": 9.774391903683185, + "learning_rate": 3.5786291518545293e-07, + "loss": 0.2522, + "step": 8447 + }, + { + "epoch": 4.586319218241043, + "grad_norm": 17.765613416632792, + "learning_rate": 3.5693110506186956e-07, + "loss": 0.3391, + "step": 8448 + }, + { + "epoch": 4.58686210640608, + "grad_norm": 8.844583546256278, + "learning_rate": 3.5600048760841776e-07, + "loss": 0.29, + "step": 8449 + }, + { + "epoch": 4.587404994571118, + "grad_norm": 15.035031432995032, + "learning_rate": 3.5507106294019323e-07, + "loss": 0.517, + "step": 8450 + }, + { + "epoch": 4.587947882736156, + "grad_norm": 11.610533233131106, + "learning_rate": 3.5414283117215285e-07, + "loss": 0.3211, + "step": 8451 + }, + { + "epoch": 4.588490770901195, + "grad_norm": 16.920965677227937, + "learning_rate": 3.5321579241910043e-07, + "loss": 0.5429, + "step": 8452 + }, + { + "epoch": 4.589033659066232, + "grad_norm": 16.08329007012344, + "learning_rate": 3.5228994679569307e-07, + "loss": 0.535, + "step": 8453 + }, + { + "epoch": 4.58957654723127, + "grad_norm": 12.092722534315392, + "learning_rate": 3.513652944164414e-07, + "loss": 0.3239, + "step": 8454 + }, + { + "epoch": 4.590119435396309, + "grad_norm": 15.030040393058352, + "learning_rate": 3.504418353957073e-07, + "loss": 0.7184, + "step": 8455 + }, + { + "epoch": 4.590662323561347, + "grad_norm": 13.30266812560096, + "learning_rate": 3.495195698477083e-07, + "loss": 0.5521, + "step": 8456 + }, + { + "epoch": 4.591205211726384, + "grad_norm": 8.94936555843634, + "learning_rate": 3.4859849788650647e-07, + "loss": 0.2933, + "step": 8457 + }, + { + "epoch": 4.591748099891422, + "grad_norm": 12.005638724180985, + "learning_rate": 3.4767861962602624e-07, + "loss": 0.4297, + "step": 8458 + }, + { + "epoch": 4.59229098805646, + "grad_norm": 16.7349439675274, + "learning_rate": 3.467599351800366e-07, + "loss": 0.4244, + "step": 8459 + }, + { + "epoch": 4.592833876221499, + "grad_norm": 11.893802858804712, + "learning_rate": 3.458424446621644e-07, + "loss": 0.398, + "step": 8460 + }, + { + "epoch": 4.593376764386536, + "grad_norm": 14.139039290089297, + "learning_rate": 3.4492614818588457e-07, + "loss": 0.4137, + "step": 8461 + }, + { + "epoch": 4.593919652551574, + "grad_norm": 11.191416980837655, + "learning_rate": 3.440110458645263e-07, + "loss": 0.4034, + "step": 8462 + }, + { + "epoch": 4.594462540716613, + "grad_norm": 11.78684784372542, + "learning_rate": 3.430971378112724e-07, + "loss": 0.4019, + "step": 8463 + }, + { + "epoch": 4.595005428881651, + "grad_norm": 18.939664291756173, + "learning_rate": 3.4218442413915477e-07, + "loss": 0.7571, + "step": 8464 + }, + { + "epoch": 4.595548317046688, + "grad_norm": 15.556000892744493, + "learning_rate": 3.412729049610586e-07, + "loss": 0.6587, + "step": 8465 + }, + { + "epoch": 4.596091205211726, + "grad_norm": 11.440255717066862, + "learning_rate": 3.403625803897248e-07, + "loss": 0.5029, + "step": 8466 + }, + { + "epoch": 4.596634093376764, + "grad_norm": 13.988507935932384, + "learning_rate": 3.3945345053774115e-07, + "loss": 0.5368, + "step": 8467 + }, + { + "epoch": 4.597176981541803, + "grad_norm": 12.212309724820258, + "learning_rate": 3.385455155175521e-07, + "loss": 0.5259, + "step": 8468 + }, + { + "epoch": 4.59771986970684, + "grad_norm": 15.416156056006928, + "learning_rate": 3.376387754414523e-07, + "loss": 0.5056, + "step": 8469 + }, + { + "epoch": 4.598262757871878, + "grad_norm": 13.475505300424985, + "learning_rate": 3.3673323042158645e-07, + "loss": 0.4761, + "step": 8470 + }, + { + "epoch": 4.598805646036917, + "grad_norm": 14.065679411115752, + "learning_rate": 3.358288805699572e-07, + "loss": 0.42, + "step": 8471 + }, + { + "epoch": 4.599348534201955, + "grad_norm": 13.604312597253596, + "learning_rate": 3.349257259984129e-07, + "loss": 0.3117, + "step": 8472 + }, + { + "epoch": 4.599891422366992, + "grad_norm": 12.84454204935273, + "learning_rate": 3.3402376681866076e-07, + "loss": 0.3665, + "step": 8473 + }, + { + "epoch": 4.60043431053203, + "grad_norm": 11.499709529011515, + "learning_rate": 3.3312300314225166e-07, + "loss": 0.4646, + "step": 8474 + }, + { + "epoch": 4.600977198697068, + "grad_norm": 16.60616452066744, + "learning_rate": 3.3222343508059975e-07, + "loss": 0.6893, + "step": 8475 + }, + { + "epoch": 4.601520086862107, + "grad_norm": 11.281168428156999, + "learning_rate": 3.3132506274495936e-07, + "loss": 0.2993, + "step": 8476 + }, + { + "epoch": 4.6020629750271445, + "grad_norm": 9.791961525364469, + "learning_rate": 3.3042788624644496e-07, + "loss": 0.366, + "step": 8477 + }, + { + "epoch": 4.602605863192182, + "grad_norm": 13.850560788317166, + "learning_rate": 3.2953190569602e-07, + "loss": 0.657, + "step": 8478 + }, + { + "epoch": 4.603148751357221, + "grad_norm": 15.50838377377847, + "learning_rate": 3.2863712120450144e-07, + "loss": 0.4996, + "step": 8479 + }, + { + "epoch": 4.603691639522259, + "grad_norm": 14.002029740570508, + "learning_rate": 3.2774353288255957e-07, + "loss": 0.2919, + "step": 8480 + }, + { + "epoch": 4.6042345276872965, + "grad_norm": 11.327925279919503, + "learning_rate": 3.268511408407105e-07, + "loss": 0.3677, + "step": 8481 + }, + { + "epoch": 4.604777415852334, + "grad_norm": 12.329600797590148, + "learning_rate": 3.259599451893303e-07, + "loss": 0.5164, + "step": 8482 + }, + { + "epoch": 4.605320304017372, + "grad_norm": 12.05441506895834, + "learning_rate": 3.2506994603864085e-07, + "loss": 0.373, + "step": 8483 + }, + { + "epoch": 4.605863192182411, + "grad_norm": 14.077644499086377, + "learning_rate": 3.241811434987219e-07, + "loss": 0.5312, + "step": 8484 + }, + { + "epoch": 4.6064060803474485, + "grad_norm": 13.177790677829083, + "learning_rate": 3.232935376794999e-07, + "loss": 0.4356, + "step": 8485 + }, + { + "epoch": 4.606948968512486, + "grad_norm": 13.304561923438573, + "learning_rate": 3.224071286907537e-07, + "loss": 0.5409, + "step": 8486 + }, + { + "epoch": 4.607491856677525, + "grad_norm": 12.635364751343143, + "learning_rate": 3.2152191664212016e-07, + "loss": 0.364, + "step": 8487 + }, + { + "epoch": 4.608034744842563, + "grad_norm": 13.255155227139575, + "learning_rate": 3.2063790164307827e-07, + "loss": 0.5079, + "step": 8488 + }, + { + "epoch": 4.6085776330076005, + "grad_norm": 10.295823117219982, + "learning_rate": 3.197550838029684e-07, + "loss": 0.4099, + "step": 8489 + }, + { + "epoch": 4.609120521172638, + "grad_norm": 11.793953833291377, + "learning_rate": 3.188734632309787e-07, + "loss": 0.3837, + "step": 8490 + }, + { + "epoch": 4.609663409337676, + "grad_norm": 9.860632156842737, + "learning_rate": 3.179930400361475e-07, + "loss": 0.3377, + "step": 8491 + }, + { + "epoch": 4.610206297502715, + "grad_norm": 14.1155595223432, + "learning_rate": 3.171138143273689e-07, + "loss": 0.555, + "step": 8492 + }, + { + "epoch": 4.6107491856677525, + "grad_norm": 12.713917889288181, + "learning_rate": 3.162357862133858e-07, + "loss": 0.7882, + "step": 8493 + }, + { + "epoch": 4.61129207383279, + "grad_norm": 10.439021701558202, + "learning_rate": 3.1535895580279364e-07, + "loss": 0.2997, + "step": 8494 + }, + { + "epoch": 4.611834961997829, + "grad_norm": 20.571810234682463, + "learning_rate": 3.1448332320404116e-07, + "loss": 0.8567, + "step": 8495 + }, + { + "epoch": 4.612377850162867, + "grad_norm": 15.758885478695694, + "learning_rate": 3.1360888852542735e-07, + "loss": 0.6149, + "step": 8496 + }, + { + "epoch": 4.6129207383279045, + "grad_norm": 11.258643577667833, + "learning_rate": 3.1273565187510455e-07, + "loss": 0.3024, + "step": 8497 + }, + { + "epoch": 4.613463626492942, + "grad_norm": 8.375690337720608, + "learning_rate": 3.118636133610753e-07, + "loss": 0.2935, + "step": 8498 + }, + { + "epoch": 4.61400651465798, + "grad_norm": 14.247335073986578, + "learning_rate": 3.1099277309119544e-07, + "loss": 0.478, + "step": 8499 + }, + { + "epoch": 4.614549402823019, + "grad_norm": 13.654635451453982, + "learning_rate": 3.1012313117317007e-07, + "loss": 0.4477, + "step": 8500 + }, + { + "epoch": 4.6150922909880565, + "grad_norm": 18.13015910990287, + "learning_rate": 3.092546877145608e-07, + "loss": 0.4855, + "step": 8501 + }, + { + "epoch": 4.615635179153094, + "grad_norm": 12.036553317442285, + "learning_rate": 3.083874428227751e-07, + "loss": 0.3039, + "step": 8502 + }, + { + "epoch": 4.616178067318133, + "grad_norm": 11.335798067486618, + "learning_rate": 3.0752139660507716e-07, + "loss": 0.4776, + "step": 8503 + }, + { + "epoch": 4.616720955483171, + "grad_norm": 8.618445905749738, + "learning_rate": 3.0665654916858136e-07, + "loss": 0.2891, + "step": 8504 + }, + { + "epoch": 4.6172638436482085, + "grad_norm": 15.36813056497114, + "learning_rate": 3.057929006202509e-07, + "loss": 0.7977, + "step": 8505 + }, + { + "epoch": 4.617806731813246, + "grad_norm": 14.372663174892352, + "learning_rate": 3.04930451066906e-07, + "loss": 0.4023, + "step": 8506 + }, + { + "epoch": 4.618349619978284, + "grad_norm": 11.295565033762697, + "learning_rate": 3.040692006152135e-07, + "loss": 0.3676, + "step": 8507 + }, + { + "epoch": 4.618892508143323, + "grad_norm": 16.469747315227195, + "learning_rate": 3.03209149371696e-07, + "loss": 0.513, + "step": 8508 + }, + { + "epoch": 4.6194353963083605, + "grad_norm": 12.642088556027979, + "learning_rate": 3.0235029744272503e-07, + "loss": 0.2876, + "step": 8509 + }, + { + "epoch": 4.619978284473398, + "grad_norm": 11.969568380769612, + "learning_rate": 3.0149264493452345e-07, + "loss": 0.3763, + "step": 8510 + }, + { + "epoch": 4.620521172638437, + "grad_norm": 10.582102519076841, + "learning_rate": 3.0063619195317197e-07, + "loss": 0.3355, + "step": 8511 + }, + { + "epoch": 4.621064060803475, + "grad_norm": 16.902071395253756, + "learning_rate": 2.9978093860459133e-07, + "loss": 0.4931, + "step": 8512 + }, + { + "epoch": 4.6216069489685125, + "grad_norm": 12.087668761128803, + "learning_rate": 2.98926884994567e-07, + "loss": 0.5744, + "step": 8513 + }, + { + "epoch": 4.62214983713355, + "grad_norm": 12.845222810452263, + "learning_rate": 2.9807403122872556e-07, + "loss": 0.5582, + "step": 8514 + }, + { + "epoch": 4.622692725298588, + "grad_norm": 14.946886726818494, + "learning_rate": 2.972223774125504e-07, + "loss": 0.255, + "step": 8515 + }, + { + "epoch": 4.623235613463627, + "grad_norm": 11.713196783156775, + "learning_rate": 2.963719236513751e-07, + "loss": 0.3617, + "step": 8516 + }, + { + "epoch": 4.6237785016286646, + "grad_norm": 13.896782178184436, + "learning_rate": 2.955226700503855e-07, + "loss": 0.4964, + "step": 8517 + }, + { + "epoch": 4.624321389793702, + "grad_norm": 17.65687683800585, + "learning_rate": 2.946746167146197e-07, + "loss": 0.5491, + "step": 8518 + }, + { + "epoch": 4.624864277958741, + "grad_norm": 17.240713061010005, + "learning_rate": 2.938277637489639e-07, + "loss": 0.6051, + "step": 8519 + }, + { + "epoch": 4.625407166123779, + "grad_norm": 12.679756933018908, + "learning_rate": 2.929821112581621e-07, + "loss": 0.3541, + "step": 8520 + }, + { + "epoch": 4.625950054288817, + "grad_norm": 13.390576698021126, + "learning_rate": 2.921376593468006e-07, + "loss": 0.4184, + "step": 8521 + }, + { + "epoch": 4.626492942453854, + "grad_norm": 13.360144383030772, + "learning_rate": 2.9129440811932694e-07, + "loss": 0.4945, + "step": 8522 + }, + { + "epoch": 4.627035830618892, + "grad_norm": 11.822031606359241, + "learning_rate": 2.9045235768003334e-07, + "loss": 0.407, + "step": 8523 + }, + { + "epoch": 4.627578718783931, + "grad_norm": 15.383117428348534, + "learning_rate": 2.896115081330675e-07, + "loss": 0.3646, + "step": 8524 + }, + { + "epoch": 4.628121606948969, + "grad_norm": 9.313479496169867, + "learning_rate": 2.8877185958242846e-07, + "loss": 0.4177, + "step": 8525 + }, + { + "epoch": 4.628664495114006, + "grad_norm": 10.269974892795535, + "learning_rate": 2.879334121319599e-07, + "loss": 0.2665, + "step": 8526 + }, + { + "epoch": 4.629207383279044, + "grad_norm": 17.13009555473781, + "learning_rate": 2.870961658853677e-07, + "loss": 0.7038, + "step": 8527 + }, + { + "epoch": 4.629750271444083, + "grad_norm": 13.86007402381537, + "learning_rate": 2.8626012094620016e-07, + "loss": 0.4005, + "step": 8528 + }, + { + "epoch": 4.630293159609121, + "grad_norm": 12.672456984260394, + "learning_rate": 2.8542527741786343e-07, + "loss": 0.3733, + "step": 8529 + }, + { + "epoch": 4.630836047774158, + "grad_norm": 14.341898459464153, + "learning_rate": 2.8459163540361044e-07, + "loss": 0.44, + "step": 8530 + }, + { + "epoch": 4.631378935939196, + "grad_norm": 13.870748930087846, + "learning_rate": 2.837591950065477e-07, + "loss": 0.384, + "step": 8531 + }, + { + "epoch": 4.631921824104235, + "grad_norm": 13.917731753801313, + "learning_rate": 2.829279563296339e-07, + "loss": 0.4294, + "step": 8532 + }, + { + "epoch": 4.632464712269273, + "grad_norm": 11.929659844878131, + "learning_rate": 2.8209791947567786e-07, + "loss": 0.41, + "step": 8533 + }, + { + "epoch": 4.63300760043431, + "grad_norm": 13.166803611206449, + "learning_rate": 2.812690845473376e-07, + "loss": 0.5053, + "step": 8534 + }, + { + "epoch": 4.633550488599348, + "grad_norm": 14.961346446295355, + "learning_rate": 2.804414516471277e-07, + "loss": 0.5023, + "step": 8535 + }, + { + "epoch": 4.634093376764387, + "grad_norm": 12.220085895146651, + "learning_rate": 2.7961502087740845e-07, + "loss": 0.2651, + "step": 8536 + }, + { + "epoch": 4.634636264929425, + "grad_norm": 12.031722727757703, + "learning_rate": 2.787897923403959e-07, + "loss": 0.3387, + "step": 8537 + }, + { + "epoch": 4.635179153094462, + "grad_norm": 8.385703592908602, + "learning_rate": 2.779657661381552e-07, + "loss": 0.1904, + "step": 8538 + }, + { + "epoch": 4.6357220412595, + "grad_norm": 10.906284934193355, + "learning_rate": 2.7714294237260353e-07, + "loss": 0.2436, + "step": 8539 + }, + { + "epoch": 4.636264929424539, + "grad_norm": 10.369806572631566, + "learning_rate": 2.7632132114550846e-07, + "loss": 0.3501, + "step": 8540 + }, + { + "epoch": 4.636807817589577, + "grad_norm": 9.022715700382092, + "learning_rate": 2.7550090255848984e-07, + "loss": 0.2257, + "step": 8541 + }, + { + "epoch": 4.637350705754614, + "grad_norm": 12.181969040940151, + "learning_rate": 2.7468168671301756e-07, + "loss": 0.2761, + "step": 8542 + }, + { + "epoch": 4.637893593919652, + "grad_norm": 14.91305612612671, + "learning_rate": 2.7386367371041387e-07, + "loss": 0.5462, + "step": 8543 + }, + { + "epoch": 4.638436482084691, + "grad_norm": 14.378563278437591, + "learning_rate": 2.730468636518535e-07, + "loss": 0.465, + "step": 8544 + }, + { + "epoch": 4.638979370249729, + "grad_norm": 16.775535206614833, + "learning_rate": 2.722312566383589e-07, + "loss": 0.5409, + "step": 8545 + }, + { + "epoch": 4.639522258414766, + "grad_norm": 12.973657108911796, + "learning_rate": 2.714168527708061e-07, + "loss": 0.588, + "step": 8546 + }, + { + "epoch": 4.640065146579804, + "grad_norm": 11.59497961581488, + "learning_rate": 2.7060365214992114e-07, + "loss": 0.2145, + "step": 8547 + }, + { + "epoch": 4.640608034744843, + "grad_norm": 11.458027506031687, + "learning_rate": 2.6979165487628354e-07, + "loss": 0.4019, + "step": 8548 + }, + { + "epoch": 4.641150922909881, + "grad_norm": 10.405492855343923, + "learning_rate": 2.6898086105032193e-07, + "loss": 0.3135, + "step": 8549 + }, + { + "epoch": 4.641693811074918, + "grad_norm": 18.94716197151622, + "learning_rate": 2.6817127077231495e-07, + "loss": 0.681, + "step": 8550 + }, + { + "epoch": 4.642236699239956, + "grad_norm": 10.670452690665341, + "learning_rate": 2.6736288414239806e-07, + "loss": 0.3912, + "step": 8551 + }, + { + "epoch": 4.642779587404995, + "grad_norm": 12.348673232962904, + "learning_rate": 2.665557012605491e-07, + "loss": 0.388, + "step": 8552 + }, + { + "epoch": 4.643322475570033, + "grad_norm": 21.51174818589849, + "learning_rate": 2.6574972222660477e-07, + "loss": 0.5009, + "step": 8553 + }, + { + "epoch": 4.64386536373507, + "grad_norm": 12.513732617765974, + "learning_rate": 2.649449471402488e-07, + "loss": 0.5977, + "step": 8554 + }, + { + "epoch": 4.644408251900108, + "grad_norm": 15.098381270766813, + "learning_rate": 2.641413761010159e-07, + "loss": 0.661, + "step": 8555 + }, + { + "epoch": 4.644951140065147, + "grad_norm": 12.750147316660394, + "learning_rate": 2.633390092082966e-07, + "loss": 0.4127, + "step": 8556 + }, + { + "epoch": 4.645494028230185, + "grad_norm": 14.376855822812393, + "learning_rate": 2.6253784656132375e-07, + "loss": 0.4477, + "step": 8557 + }, + { + "epoch": 4.646036916395222, + "grad_norm": 9.209626776279498, + "learning_rate": 2.6173788825919253e-07, + "loss": 0.2738, + "step": 8558 + }, + { + "epoch": 4.64657980456026, + "grad_norm": 14.646741912366064, + "learning_rate": 2.6093913440083715e-07, + "loss": 0.5399, + "step": 8559 + }, + { + "epoch": 4.647122692725299, + "grad_norm": 13.357674411747405, + "learning_rate": 2.6014158508505304e-07, + "loss": 0.4451, + "step": 8560 + }, + { + "epoch": 4.647665580890337, + "grad_norm": 12.343655571585558, + "learning_rate": 2.5934524041047904e-07, + "loss": 0.25, + "step": 8561 + }, + { + "epoch": 4.648208469055374, + "grad_norm": 13.60791433197796, + "learning_rate": 2.585501004756108e-07, + "loss": 0.3931, + "step": 8562 + }, + { + "epoch": 4.648751357220412, + "grad_norm": 13.34834410008458, + "learning_rate": 2.577561653787919e-07, + "loss": 0.4463, + "step": 8563 + }, + { + "epoch": 4.649294245385451, + "grad_norm": 14.512470015045476, + "learning_rate": 2.569634352182171e-07, + "loss": 0.4353, + "step": 8564 + }, + { + "epoch": 4.649837133550489, + "grad_norm": 14.941554464032452, + "learning_rate": 2.5617191009193356e-07, + "loss": 0.5627, + "step": 8565 + }, + { + "epoch": 4.6503800217155264, + "grad_norm": 19.483475442382513, + "learning_rate": 2.5538159009783516e-07, + "loss": 0.4743, + "step": 8566 + }, + { + "epoch": 4.650922909880564, + "grad_norm": 14.823894906192926, + "learning_rate": 2.545924753336737e-07, + "loss": 0.7184, + "step": 8567 + }, + { + "epoch": 4.651465798045603, + "grad_norm": 10.450395109397446, + "learning_rate": 2.538045658970478e-07, + "loss": 0.2966, + "step": 8568 + }, + { + "epoch": 4.652008686210641, + "grad_norm": 15.67717478106929, + "learning_rate": 2.530178618854051e-07, + "loss": 0.7021, + "step": 8569 + }, + { + "epoch": 4.6525515743756785, + "grad_norm": 11.80163747115902, + "learning_rate": 2.5223236339604775e-07, + "loss": 0.8058, + "step": 8570 + }, + { + "epoch": 4.653094462540716, + "grad_norm": 13.651780146568179, + "learning_rate": 2.51448070526128e-07, + "loss": 0.6028, + "step": 8571 + }, + { + "epoch": 4.653637350705755, + "grad_norm": 11.185785067606833, + "learning_rate": 2.506649833726471e-07, + "loss": 0.3177, + "step": 8572 + }, + { + "epoch": 4.654180238870793, + "grad_norm": 12.995594956299897, + "learning_rate": 2.498831020324588e-07, + "loss": 0.4614, + "step": 8573 + }, + { + "epoch": 4.6547231270358305, + "grad_norm": 9.97604838386349, + "learning_rate": 2.49102426602269e-07, + "loss": 0.3827, + "step": 8574 + }, + { + "epoch": 4.655266015200868, + "grad_norm": 12.661162432576356, + "learning_rate": 2.4832295717863054e-07, + "loss": 0.4313, + "step": 8575 + }, + { + "epoch": 4.655808903365907, + "grad_norm": 17.062394951641533, + "learning_rate": 2.4754469385795177e-07, + "loss": 0.5858, + "step": 8576 + }, + { + "epoch": 4.656351791530945, + "grad_norm": 12.7320209210787, + "learning_rate": 2.4676763673648687e-07, + "loss": 0.3646, + "step": 8577 + }, + { + "epoch": 4.6568946796959825, + "grad_norm": 12.069345026621932, + "learning_rate": 2.4599178591034666e-07, + "loss": 0.4396, + "step": 8578 + }, + { + "epoch": 4.65743756786102, + "grad_norm": 11.365880065156547, + "learning_rate": 2.452171414754867e-07, + "loss": 0.27, + "step": 8579 + }, + { + "epoch": 4.657980456026059, + "grad_norm": 16.52465140027476, + "learning_rate": 2.4444370352771807e-07, + "loss": 0.3874, + "step": 8580 + }, + { + "epoch": 4.658523344191097, + "grad_norm": 9.574363596986716, + "learning_rate": 2.4367147216269873e-07, + "loss": 0.2478, + "step": 8581 + }, + { + "epoch": 4.6590662323561345, + "grad_norm": 15.080081475667454, + "learning_rate": 2.4290044747594333e-07, + "loss": 0.4246, + "step": 8582 + }, + { + "epoch": 4.659609120521172, + "grad_norm": 17.58550881221689, + "learning_rate": 2.421306295628101e-07, + "loss": 0.3626, + "step": 8583 + }, + { + "epoch": 4.660152008686211, + "grad_norm": 11.302404165947946, + "learning_rate": 2.413620185185128e-07, + "loss": 0.3163, + "step": 8584 + }, + { + "epoch": 4.660694896851249, + "grad_norm": 12.604954301784339, + "learning_rate": 2.4059461443811325e-07, + "loss": 0.4633, + "step": 8585 + }, + { + "epoch": 4.6612377850162865, + "grad_norm": 16.576550436899478, + "learning_rate": 2.398284174165255e-07, + "loss": 0.5843, + "step": 8586 + }, + { + "epoch": 4.661780673181324, + "grad_norm": 18.037796417236624, + "learning_rate": 2.3906342754851595e-07, + "loss": 0.8494, + "step": 8587 + }, + { + "epoch": 4.662323561346363, + "grad_norm": 10.736278337401655, + "learning_rate": 2.3829964492869673e-07, + "loss": 0.4509, + "step": 8588 + }, + { + "epoch": 4.662866449511401, + "grad_norm": 13.335894542183691, + "learning_rate": 2.3753706965153666e-07, + "loss": 0.6105, + "step": 8589 + }, + { + "epoch": 4.6634093376764385, + "grad_norm": 11.486510470019578, + "learning_rate": 2.3677570181134923e-07, + "loss": 0.3567, + "step": 8590 + }, + { + "epoch": 4.663952225841476, + "grad_norm": 19.53299668429002, + "learning_rate": 2.3601554150230577e-07, + "loss": 0.6259, + "step": 8591 + }, + { + "epoch": 4.664495114006515, + "grad_norm": 12.957043525497816, + "learning_rate": 2.3525658881841884e-07, + "loss": 0.4851, + "step": 8592 + }, + { + "epoch": 4.665038002171553, + "grad_norm": 11.145868206087584, + "learning_rate": 2.3449884385356114e-07, + "loss": 0.2943, + "step": 8593 + }, + { + "epoch": 4.6655808903365905, + "grad_norm": 17.411940658097325, + "learning_rate": 2.3374230670145104e-07, + "loss": 0.6137, + "step": 8594 + }, + { + "epoch": 4.666123778501628, + "grad_norm": 13.68690986112585, + "learning_rate": 2.329869774556559e-07, + "loss": 0.374, + "step": 8595 + }, + { + "epoch": 4.666666666666667, + "grad_norm": 15.428074760731318, + "learning_rate": 2.322328562095988e-07, + "loss": 0.3807, + "step": 8596 + }, + { + "epoch": 4.667209554831705, + "grad_norm": 14.483267133862492, + "learning_rate": 2.314799430565484e-07, + "loss": 0.4133, + "step": 8597 + }, + { + "epoch": 4.6677524429967425, + "grad_norm": 12.727450078217835, + "learning_rate": 2.3072823808962918e-07, + "loss": 0.424, + "step": 8598 + }, + { + "epoch": 4.66829533116178, + "grad_norm": 10.546825509905323, + "learning_rate": 2.2997774140181007e-07, + "loss": 0.3447, + "step": 8599 + }, + { + "epoch": 4.668838219326819, + "grad_norm": 12.26173003486407, + "learning_rate": 2.2922845308591457e-07, + "loss": 0.3995, + "step": 8600 + }, + { + "epoch": 4.669381107491857, + "grad_norm": 11.862665121581074, + "learning_rate": 2.2848037323461748e-07, + "loss": 0.396, + "step": 8601 + }, + { + "epoch": 4.6699239956568945, + "grad_norm": 11.53214186874608, + "learning_rate": 2.277335019404403e-07, + "loss": 0.3117, + "step": 8602 + }, + { + "epoch": 4.670466883821932, + "grad_norm": 9.677126503825807, + "learning_rate": 2.2698783929576029e-07, + "loss": 0.3299, + "step": 8603 + }, + { + "epoch": 4.671009771986971, + "grad_norm": 12.489313816621161, + "learning_rate": 2.26243385392797e-07, + "loss": 0.5942, + "step": 8604 + }, + { + "epoch": 4.671552660152009, + "grad_norm": 13.63702388723821, + "learning_rate": 2.255001403236312e-07, + "loss": 0.5161, + "step": 8605 + }, + { + "epoch": 4.6720955483170465, + "grad_norm": 12.047997675365789, + "learning_rate": 2.2475810418018496e-07, + "loss": 0.3852, + "step": 8606 + }, + { + "epoch": 4.672638436482084, + "grad_norm": 14.787350942388676, + "learning_rate": 2.2401727705423704e-07, + "loss": 0.4452, + "step": 8607 + }, + { + "epoch": 4.673181324647123, + "grad_norm": 15.453200600201, + "learning_rate": 2.2327765903741194e-07, + "loss": 0.4722, + "step": 8608 + }, + { + "epoch": 4.673724212812161, + "grad_norm": 14.118602165974508, + "learning_rate": 2.2253925022118872e-07, + "loss": 0.464, + "step": 8609 + }, + { + "epoch": 4.6742671009771986, + "grad_norm": 14.255945897987628, + "learning_rate": 2.218020506968932e-07, + "loss": 0.6369, + "step": 8610 + }, + { + "epoch": 4.674809989142236, + "grad_norm": 15.496989818750507, + "learning_rate": 2.2106606055570468e-07, + "loss": 0.2463, + "step": 8611 + }, + { + "epoch": 4.675352877307275, + "grad_norm": 11.272611567068735, + "learning_rate": 2.2033127988865032e-07, + "loss": 0.3691, + "step": 8612 + }, + { + "epoch": 4.675895765472313, + "grad_norm": 11.496100548853523, + "learning_rate": 2.195977087866097e-07, + "loss": 0.2911, + "step": 8613 + }, + { + "epoch": 4.676438653637351, + "grad_norm": 8.624224394988135, + "learning_rate": 2.1886534734031238e-07, + "loss": 0.2357, + "step": 8614 + }, + { + "epoch": 4.676981541802388, + "grad_norm": 15.681027668557938, + "learning_rate": 2.1813419564033823e-07, + "loss": 0.5629, + "step": 8615 + }, + { + "epoch": 4.677524429967427, + "grad_norm": 15.960892979111017, + "learning_rate": 2.1740425377711704e-07, + "loss": 0.5147, + "step": 8616 + }, + { + "epoch": 4.678067318132465, + "grad_norm": 14.11243172747272, + "learning_rate": 2.1667552184092778e-07, + "loss": 0.4392, + "step": 8617 + }, + { + "epoch": 4.678610206297503, + "grad_norm": 10.877174468176161, + "learning_rate": 2.1594799992190274e-07, + "loss": 0.3015, + "step": 8618 + }, + { + "epoch": 4.67915309446254, + "grad_norm": 15.805676119173855, + "learning_rate": 2.152216881100222e-07, + "loss": 0.6413, + "step": 8619 + }, + { + "epoch": 4.679695982627579, + "grad_norm": 10.914077977600847, + "learning_rate": 2.1449658649511873e-07, + "loss": 0.3409, + "step": 8620 + }, + { + "epoch": 4.680238870792617, + "grad_norm": 18.998010438264245, + "learning_rate": 2.137726951668717e-07, + "loss": 0.7325, + "step": 8621 + }, + { + "epoch": 4.680781758957655, + "grad_norm": 11.034223861305692, + "learning_rate": 2.130500142148162e-07, + "loss": 0.3421, + "step": 8622 + }, + { + "epoch": 4.681324647122692, + "grad_norm": 13.223093235714936, + "learning_rate": 2.123285437283329e-07, + "loss": 0.3714, + "step": 8623 + }, + { + "epoch": 4.681867535287731, + "grad_norm": 15.922027216206333, + "learning_rate": 2.1160828379665378e-07, + "loss": 0.6786, + "step": 8624 + }, + { + "epoch": 4.682410423452769, + "grad_norm": 12.103763081737, + "learning_rate": 2.1088923450886312e-07, + "loss": 0.2379, + "step": 8625 + }, + { + "epoch": 4.682953311617807, + "grad_norm": 8.921346684769288, + "learning_rate": 2.1017139595389202e-07, + "loss": 0.2147, + "step": 8626 + }, + { + "epoch": 4.683496199782844, + "grad_norm": 17.606964595238484, + "learning_rate": 2.0945476822052725e-07, + "loss": 0.5509, + "step": 8627 + }, + { + "epoch": 4.684039087947883, + "grad_norm": 16.30507726038034, + "learning_rate": 2.0873935139739898e-07, + "loss": 0.9206, + "step": 8628 + }, + { + "epoch": 4.684581976112921, + "grad_norm": 10.642327565305825, + "learning_rate": 2.0802514557299424e-07, + "loss": 0.4332, + "step": 8629 + }, + { + "epoch": 4.685124864277959, + "grad_norm": 11.145256035049321, + "learning_rate": 2.073121508356446e-07, + "loss": 0.3007, + "step": 8630 + }, + { + "epoch": 4.685667752442996, + "grad_norm": 12.907940052414924, + "learning_rate": 2.0660036727353615e-07, + "loss": 0.4258, + "step": 8631 + }, + { + "epoch": 4.686210640608035, + "grad_norm": 12.333985136599837, + "learning_rate": 2.0588979497470185e-07, + "loss": 0.3495, + "step": 8632 + }, + { + "epoch": 4.686753528773073, + "grad_norm": 15.062071959789007, + "learning_rate": 2.0518043402702692e-07, + "loss": 0.3947, + "step": 8633 + }, + { + "epoch": 4.687296416938111, + "grad_norm": 17.461595809708303, + "learning_rate": 2.0447228451824785e-07, + "loss": 0.6401, + "step": 8634 + }, + { + "epoch": 4.687839305103148, + "grad_norm": 15.753893232071881, + "learning_rate": 2.037653465359457e-07, + "loss": 0.5768, + "step": 8635 + }, + { + "epoch": 4.688382193268187, + "grad_norm": 7.552176112941645, + "learning_rate": 2.0305962016756164e-07, + "loss": 0.2491, + "step": 8636 + }, + { + "epoch": 4.688925081433225, + "grad_norm": 14.000517906223289, + "learning_rate": 2.0235510550037584e-07, + "loss": 0.4483, + "step": 8637 + }, + { + "epoch": 4.689467969598263, + "grad_norm": 11.530560060275077, + "learning_rate": 2.0165180262152528e-07, + "loss": 0.3759, + "step": 8638 + }, + { + "epoch": 4.6900108577633, + "grad_norm": 12.821236187937325, + "learning_rate": 2.0094971161799703e-07, + "loss": 0.3813, + "step": 8639 + }, + { + "epoch": 4.690553745928339, + "grad_norm": 10.067727608425038, + "learning_rate": 2.0024883257662608e-07, + "loss": 0.3769, + "step": 8640 + }, + { + "epoch": 4.691096634093377, + "grad_norm": 12.525413966092579, + "learning_rate": 1.9954916558409753e-07, + "loss": 0.5518, + "step": 8641 + }, + { + "epoch": 4.691639522258415, + "grad_norm": 10.599027738171559, + "learning_rate": 1.9885071072694773e-07, + "loss": 0.2527, + "step": 8642 + }, + { + "epoch": 4.692182410423452, + "grad_norm": 18.15731060573008, + "learning_rate": 1.9815346809156311e-07, + "loss": 0.4417, + "step": 8643 + }, + { + "epoch": 4.692725298588491, + "grad_norm": 10.69232848876475, + "learning_rate": 1.9745743776417914e-07, + "loss": 0.3218, + "step": 8644 + }, + { + "epoch": 4.693268186753529, + "grad_norm": 12.33953438764356, + "learning_rate": 1.967626198308814e-07, + "loss": 0.3422, + "step": 8645 + }, + { + "epoch": 4.693811074918567, + "grad_norm": 14.649644582793519, + "learning_rate": 1.9606901437760785e-07, + "loss": 0.5192, + "step": 8646 + }, + { + "epoch": 4.694353963083604, + "grad_norm": 13.239419010285507, + "learning_rate": 1.9537662149014312e-07, + "loss": 0.3623, + "step": 8647 + }, + { + "epoch": 4.694896851248643, + "grad_norm": 13.905675199978676, + "learning_rate": 1.9468544125412549e-07, + "loss": 0.5053, + "step": 8648 + }, + { + "epoch": 4.695439739413681, + "grad_norm": 12.990136427182705, + "learning_rate": 1.939954737550387e-07, + "loss": 0.3557, + "step": 8649 + }, + { + "epoch": 4.695982627578719, + "grad_norm": 17.66732769429285, + "learning_rate": 1.9330671907822006e-07, + "loss": 0.5037, + "step": 8650 + }, + { + "epoch": 4.696525515743756, + "grad_norm": 10.181756534303782, + "learning_rate": 1.9261917730885592e-07, + "loss": 0.3003, + "step": 8651 + }, + { + "epoch": 4.697068403908795, + "grad_norm": 13.590242162946018, + "learning_rate": 1.9193284853198268e-07, + "loss": 0.5647, + "step": 8652 + }, + { + "epoch": 4.697611292073833, + "grad_norm": 12.17175015122372, + "learning_rate": 1.9124773283248688e-07, + "loss": 0.4198, + "step": 8653 + }, + { + "epoch": 4.698154180238871, + "grad_norm": 11.105538961089028, + "learning_rate": 1.9056383029510406e-07, + "loss": 0.3301, + "step": 8654 + }, + { + "epoch": 4.698697068403908, + "grad_norm": 16.019376029880355, + "learning_rate": 1.898811410044199e-07, + "loss": 0.5822, + "step": 8655 + }, + { + "epoch": 4.699239956568947, + "grad_norm": 15.67141763731205, + "learning_rate": 1.8919966504487242e-07, + "loss": 0.5289, + "step": 8656 + }, + { + "epoch": 4.699782844733985, + "grad_norm": 12.06247311043435, + "learning_rate": 1.885194025007464e-07, + "loss": 0.305, + "step": 8657 + }, + { + "epoch": 4.700325732899023, + "grad_norm": 9.89364921552198, + "learning_rate": 1.878403534561768e-07, + "loss": 0.3231, + "step": 8658 + }, + { + "epoch": 4.7008686210640604, + "grad_norm": 14.67715265064961, + "learning_rate": 1.8716251799515083e-07, + "loss": 0.4446, + "step": 8659 + }, + { + "epoch": 4.701411509229099, + "grad_norm": 11.673006836898997, + "learning_rate": 1.864858962015048e-07, + "loss": 0.556, + "step": 8660 + }, + { + "epoch": 4.701954397394137, + "grad_norm": 10.608708966780606, + "learning_rate": 1.8581048815892288e-07, + "loss": 0.2792, + "step": 8661 + }, + { + "epoch": 4.702497285559175, + "grad_norm": 13.134592713431639, + "learning_rate": 1.8513629395094158e-07, + "loss": 0.3235, + "step": 8662 + }, + { + "epoch": 4.7030401737242125, + "grad_norm": 10.263003236237122, + "learning_rate": 1.844633136609453e-07, + "loss": 0.2699, + "step": 8663 + }, + { + "epoch": 4.703583061889251, + "grad_norm": 15.219462479114338, + "learning_rate": 1.8379154737216963e-07, + "loss": 0.3003, + "step": 8664 + }, + { + "epoch": 4.704125950054289, + "grad_norm": 13.049475534310673, + "learning_rate": 1.8312099516770265e-07, + "loss": 0.4857, + "step": 8665 + }, + { + "epoch": 4.704668838219327, + "grad_norm": 10.788197456357453, + "learning_rate": 1.824516571304735e-07, + "loss": 0.3011, + "step": 8666 + }, + { + "epoch": 4.7052117263843645, + "grad_norm": 9.043271241783842, + "learning_rate": 1.817835333432727e-07, + "loss": 0.2357, + "step": 8667 + }, + { + "epoch": 4.705754614549403, + "grad_norm": 9.656532376361252, + "learning_rate": 1.8111662388873074e-07, + "loss": 0.248, + "step": 8668 + }, + { + "epoch": 4.706297502714441, + "grad_norm": 16.70271141414241, + "learning_rate": 1.80450928849335e-07, + "loss": 0.4306, + "step": 8669 + }, + { + "epoch": 4.706840390879479, + "grad_norm": 13.146831118988962, + "learning_rate": 1.7978644830741743e-07, + "loss": 0.3665, + "step": 8670 + }, + { + "epoch": 4.7073832790445165, + "grad_norm": 15.26133993523942, + "learning_rate": 1.7912318234516334e-07, + "loss": 0.6296, + "step": 8671 + }, + { + "epoch": 4.707926167209555, + "grad_norm": 12.19166523931649, + "learning_rate": 1.7846113104460716e-07, + "loss": 0.3262, + "step": 8672 + }, + { + "epoch": 4.708469055374593, + "grad_norm": 14.50300563374095, + "learning_rate": 1.7780029448763004e-07, + "loss": 0.6108, + "step": 8673 + }, + { + "epoch": 4.709011943539631, + "grad_norm": 12.157023060827033, + "learning_rate": 1.771406727559688e-07, + "loss": 0.3338, + "step": 8674 + }, + { + "epoch": 4.7095548317046685, + "grad_norm": 16.358634798576084, + "learning_rate": 1.7648226593120154e-07, + "loss": 0.5102, + "step": 8675 + }, + { + "epoch": 4.710097719869707, + "grad_norm": 15.026104365171472, + "learning_rate": 1.7582507409476534e-07, + "loss": 0.5946, + "step": 8676 + }, + { + "epoch": 4.710640608034745, + "grad_norm": 10.191961599550515, + "learning_rate": 1.7516909732794186e-07, + "loss": 0.2927, + "step": 8677 + }, + { + "epoch": 4.711183496199783, + "grad_norm": 13.319008024488364, + "learning_rate": 1.7451433571186282e-07, + "loss": 0.3696, + "step": 8678 + }, + { + "epoch": 4.7117263843648205, + "grad_norm": 9.989147946007154, + "learning_rate": 1.7386078932751016e-07, + "loss": 0.2435, + "step": 8679 + }, + { + "epoch": 4.712269272529859, + "grad_norm": 14.90238420603907, + "learning_rate": 1.7320845825571474e-07, + "loss": 0.4526, + "step": 8680 + }, + { + "epoch": 4.712812160694897, + "grad_norm": 19.24089773059805, + "learning_rate": 1.7255734257716094e-07, + "loss": 0.6766, + "step": 8681 + }, + { + "epoch": 4.713355048859935, + "grad_norm": 14.187464466284949, + "learning_rate": 1.7190744237237545e-07, + "loss": 0.7976, + "step": 8682 + }, + { + "epoch": 4.7138979370249725, + "grad_norm": 14.841750204586576, + "learning_rate": 1.7125875772174172e-07, + "loss": 0.6627, + "step": 8683 + }, + { + "epoch": 4.714440825190011, + "grad_norm": 13.266714679527853, + "learning_rate": 1.7061128870548893e-07, + "loss": 0.3231, + "step": 8684 + }, + { + "epoch": 4.714983713355049, + "grad_norm": 10.735930720776885, + "learning_rate": 1.6996503540369857e-07, + "loss": 0.3564, + "step": 8685 + }, + { + "epoch": 4.715526601520087, + "grad_norm": 13.418944607227882, + "learning_rate": 1.6931999789629895e-07, + "loss": 0.4069, + "step": 8686 + }, + { + "epoch": 4.7160694896851245, + "grad_norm": 7.840135585116192, + "learning_rate": 1.6867617626307064e-07, + "loss": 0.2062, + "step": 8687 + }, + { + "epoch": 4.716612377850163, + "grad_norm": 16.197182822841835, + "learning_rate": 1.6803357058364113e-07, + "loss": 0.6624, + "step": 8688 + }, + { + "epoch": 4.717155266015201, + "grad_norm": 13.971201461249894, + "learning_rate": 1.6739218093748898e-07, + "loss": 0.6094, + "step": 8689 + }, + { + "epoch": 4.717698154180239, + "grad_norm": 13.176526767197032, + "learning_rate": 1.6675200740394303e-07, + "loss": 0.5601, + "step": 8690 + }, + { + "epoch": 4.7182410423452765, + "grad_norm": 11.043215518250438, + "learning_rate": 1.6611305006218103e-07, + "loss": 0.3768, + "step": 8691 + }, + { + "epoch": 4.718783930510315, + "grad_norm": 11.048074306636915, + "learning_rate": 1.6547530899122866e-07, + "loss": 0.3214, + "step": 8692 + }, + { + "epoch": 4.719326818675353, + "grad_norm": 12.953674925372624, + "learning_rate": 1.64838784269965e-07, + "loss": 0.3367, + "step": 8693 + }, + { + "epoch": 4.719869706840391, + "grad_norm": 19.600125509779307, + "learning_rate": 1.6420347597711607e-07, + "loss": 0.5111, + "step": 8694 + }, + { + "epoch": 4.7204125950054285, + "grad_norm": 12.73836642608448, + "learning_rate": 1.6356938419125556e-07, + "loss": 0.4027, + "step": 8695 + }, + { + "epoch": 4.720955483170467, + "grad_norm": 13.775692913408399, + "learning_rate": 1.6293650899081193e-07, + "loss": 0.6892, + "step": 8696 + }, + { + "epoch": 4.721498371335505, + "grad_norm": 13.475886438070386, + "learning_rate": 1.6230485045405697e-07, + "loss": 0.4, + "step": 8697 + }, + { + "epoch": 4.722041259500543, + "grad_norm": 14.97051098396697, + "learning_rate": 1.6167440865911932e-07, + "loss": 0.3445, + "step": 8698 + }, + { + "epoch": 4.7225841476655805, + "grad_norm": 16.47361547249056, + "learning_rate": 1.610451836839688e-07, + "loss": 0.3719, + "step": 8699 + }, + { + "epoch": 4.723127035830619, + "grad_norm": 11.078494155572205, + "learning_rate": 1.604171756064321e-07, + "loss": 0.3196, + "step": 8700 + }, + { + "epoch": 4.723669923995657, + "grad_norm": 17.74776123788687, + "learning_rate": 1.597903845041815e-07, + "loss": 0.686, + "step": 8701 + }, + { + "epoch": 4.724212812160695, + "grad_norm": 10.382852901941714, + "learning_rate": 1.591648104547383e-07, + "loss": 0.3855, + "step": 8702 + }, + { + "epoch": 4.7247557003257326, + "grad_norm": 13.409173794328078, + "learning_rate": 1.5854045353547508e-07, + "loss": 0.4286, + "step": 8703 + }, + { + "epoch": 4.725298588490771, + "grad_norm": 13.843911786331836, + "learning_rate": 1.579173138236134e-07, + "loss": 0.2781, + "step": 8704 + }, + { + "epoch": 4.725841476655809, + "grad_norm": 15.962352266628493, + "learning_rate": 1.5729539139622496e-07, + "loss": 0.573, + "step": 8705 + }, + { + "epoch": 4.726384364820847, + "grad_norm": 13.180235377705445, + "learning_rate": 1.566746863302293e-07, + "loss": 0.6014, + "step": 8706 + }, + { + "epoch": 4.726927252985885, + "grad_norm": 13.52518276203702, + "learning_rate": 1.5605519870239615e-07, + "loss": 0.4553, + "step": 8707 + }, + { + "epoch": 4.727470141150923, + "grad_norm": 16.51381961667992, + "learning_rate": 1.5543692858934535e-07, + "loss": 0.435, + "step": 8708 + }, + { + "epoch": 4.728013029315961, + "grad_norm": 7.676924114723452, + "learning_rate": 1.5481987606754456e-07, + "loss": 0.2235, + "step": 8709 + }, + { + "epoch": 4.728555917480999, + "grad_norm": 10.953711460471121, + "learning_rate": 1.542040412133128e-07, + "loss": 0.4336, + "step": 8710 + }, + { + "epoch": 4.729098805646037, + "grad_norm": 14.934986133675514, + "learning_rate": 1.5358942410281685e-07, + "loss": 0.461, + "step": 8711 + }, + { + "epoch": 4.729641693811075, + "grad_norm": 15.50829782966211, + "learning_rate": 1.5297602481207485e-07, + "loss": 0.4241, + "step": 8712 + }, + { + "epoch": 4.730184581976113, + "grad_norm": 12.740867628061949, + "learning_rate": 1.5236384341694944e-07, + "loss": 0.3727, + "step": 8713 + }, + { + "epoch": 4.730727470141151, + "grad_norm": 17.63779367168277, + "learning_rate": 1.5175287999316e-07, + "loss": 0.522, + "step": 8714 + }, + { + "epoch": 4.731270358306189, + "grad_norm": 12.751132476411902, + "learning_rate": 1.511431346162706e-07, + "loss": 0.3198, + "step": 8715 + }, + { + "epoch": 4.731813246471227, + "grad_norm": 13.490836748308888, + "learning_rate": 1.5053460736169535e-07, + "loss": 0.6525, + "step": 8716 + }, + { + "epoch": 4.732356134636265, + "grad_norm": 14.36545355486523, + "learning_rate": 1.499272983046962e-07, + "loss": 0.3776, + "step": 8717 + }, + { + "epoch": 4.732899022801303, + "grad_norm": 12.319339869403628, + "learning_rate": 1.493212075203887e-07, + "loss": 0.3082, + "step": 8718 + }, + { + "epoch": 4.733441910966341, + "grad_norm": 14.700540476410424, + "learning_rate": 1.4871633508373285e-07, + "loss": 0.5825, + "step": 8719 + }, + { + "epoch": 4.733984799131379, + "grad_norm": 12.437335028743151, + "learning_rate": 1.4811268106954212e-07, + "loss": 0.5425, + "step": 8720 + }, + { + "epoch": 4.734527687296417, + "grad_norm": 13.731965452187058, + "learning_rate": 1.4751024555247683e-07, + "loss": 0.3165, + "step": 8721 + }, + { + "epoch": 4.735070575461455, + "grad_norm": 12.201944872037965, + "learning_rate": 1.4690902860704626e-07, + "loss": 0.3219, + "step": 8722 + }, + { + "epoch": 4.735613463626493, + "grad_norm": 10.718132799597125, + "learning_rate": 1.463090303076109e-07, + "loss": 0.3031, + "step": 8723 + }, + { + "epoch": 4.736156351791531, + "grad_norm": 10.91023746799804, + "learning_rate": 1.457102507283781e-07, + "loss": 0.2936, + "step": 8724 + }, + { + "epoch": 4.736699239956569, + "grad_norm": 18.09494215671278, + "learning_rate": 1.4511268994340742e-07, + "loss": 0.481, + "step": 8725 + }, + { + "epoch": 4.737242128121607, + "grad_norm": 11.40586943270243, + "learning_rate": 1.4451634802660654e-07, + "loss": 0.335, + "step": 8726 + }, + { + "epoch": 4.737785016286645, + "grad_norm": 11.869445402678647, + "learning_rate": 1.4392122505172968e-07, + "loss": 0.4306, + "step": 8727 + }, + { + "epoch": 4.738327904451683, + "grad_norm": 15.51945714681897, + "learning_rate": 1.4332732109238466e-07, + "loss": 0.4721, + "step": 8728 + }, + { + "epoch": 4.738870792616721, + "grad_norm": 16.334459977316087, + "learning_rate": 1.427346362220261e-07, + "loss": 0.3783, + "step": 8729 + }, + { + "epoch": 4.739413680781759, + "grad_norm": 16.14244485338596, + "learning_rate": 1.4214317051395753e-07, + "loss": 0.5549, + "step": 8730 + }, + { + "epoch": 4.739956568946797, + "grad_norm": 10.731389089987651, + "learning_rate": 1.415529240413327e-07, + "loss": 0.2656, + "step": 8731 + }, + { + "epoch": 4.740499457111835, + "grad_norm": 11.843454339533057, + "learning_rate": 1.4096389687715427e-07, + "loss": 0.4719, + "step": 8732 + }, + { + "epoch": 4.741042345276873, + "grad_norm": 18.272514152254576, + "learning_rate": 1.4037608909427402e-07, + "loss": 0.6369, + "step": 8733 + }, + { + "epoch": 4.741585233441911, + "grad_norm": 12.608713904711752, + "learning_rate": 1.3978950076539265e-07, + "loss": 0.3475, + "step": 8734 + }, + { + "epoch": 4.742128121606949, + "grad_norm": 13.64628443407127, + "learning_rate": 1.3920413196305992e-07, + "loss": 0.578, + "step": 8735 + }, + { + "epoch": 4.742671009771987, + "grad_norm": 16.32438818863956, + "learning_rate": 1.3861998275967793e-07, + "loss": 0.6912, + "step": 8736 + }, + { + "epoch": 4.743213897937025, + "grad_norm": 11.713933611270436, + "learning_rate": 1.3803705322749107e-07, + "loss": 0.2049, + "step": 8737 + }, + { + "epoch": 4.743756786102063, + "grad_norm": 12.260983477641858, + "learning_rate": 1.3745534343860166e-07, + "loss": 0.4776, + "step": 8738 + }, + { + "epoch": 4.744299674267101, + "grad_norm": 12.840258489630633, + "learning_rate": 1.368748534649511e-07, + "loss": 0.3826, + "step": 8739 + }, + { + "epoch": 4.744842562432139, + "grad_norm": 16.03466717753298, + "learning_rate": 1.362955833783408e-07, + "loss": 0.6061, + "step": 8740 + }, + { + "epoch": 4.745385450597177, + "grad_norm": 9.356241312014683, + "learning_rate": 1.3571753325041126e-07, + "loss": 0.2633, + "step": 8741 + }, + { + "epoch": 4.745928338762215, + "grad_norm": 16.828204569000025, + "learning_rate": 1.351407031526586e-07, + "loss": 0.4258, + "step": 8742 + }, + { + "epoch": 4.746471226927253, + "grad_norm": 15.686474603102551, + "learning_rate": 1.3456509315642685e-07, + "loss": 0.3753, + "step": 8743 + }, + { + "epoch": 4.747014115092291, + "grad_norm": 9.198083120731805, + "learning_rate": 1.3399070333290688e-07, + "loss": 0.2984, + "step": 8744 + }, + { + "epoch": 4.747557003257329, + "grad_norm": 10.265356555589412, + "learning_rate": 1.3341753375314182e-07, + "loss": 0.3292, + "step": 8745 + }, + { + "epoch": 4.748099891422367, + "grad_norm": 10.896246230002951, + "learning_rate": 1.3284558448802054e-07, + "loss": 0.3922, + "step": 8746 + }, + { + "epoch": 4.748642779587405, + "grad_norm": 14.67283008436615, + "learning_rate": 1.3227485560828312e-07, + "loss": 0.5232, + "step": 8747 + }, + { + "epoch": 4.749185667752443, + "grad_norm": 9.52295940041419, + "learning_rate": 1.3170534718451754e-07, + "loss": 0.3544, + "step": 8748 + }, + { + "epoch": 4.749728555917481, + "grad_norm": 18.509918462924368, + "learning_rate": 1.3113705928716415e-07, + "loss": 0.4525, + "step": 8749 + }, + { + "epoch": 4.750271444082519, + "grad_norm": 12.780045325033498, + "learning_rate": 1.3056999198650778e-07, + "loss": 0.4545, + "step": 8750 + }, + { + "epoch": 4.750814332247557, + "grad_norm": 14.885728814999055, + "learning_rate": 1.3000414535268236e-07, + "loss": 0.2981, + "step": 8751 + }, + { + "epoch": 4.751357220412595, + "grad_norm": 13.266942664774517, + "learning_rate": 1.2943951945567633e-07, + "loss": 0.492, + "step": 8752 + }, + { + "epoch": 4.751900108577633, + "grad_norm": 16.094462901734268, + "learning_rate": 1.2887611436532167e-07, + "loss": 0.444, + "step": 8753 + }, + { + "epoch": 4.752442996742671, + "grad_norm": 14.798576298672852, + "learning_rate": 1.2831393015130034e-07, + "loss": 0.6268, + "step": 8754 + }, + { + "epoch": 4.752985884907709, + "grad_norm": 14.276704947243562, + "learning_rate": 1.2775296688314675e-07, + "loss": 0.4696, + "step": 8755 + }, + { + "epoch": 4.753528773072747, + "grad_norm": 11.206999668115289, + "learning_rate": 1.2719322463023877e-07, + "loss": 0.3501, + "step": 8756 + }, + { + "epoch": 4.754071661237785, + "grad_norm": 9.692669879532469, + "learning_rate": 1.2663470346180872e-07, + "loss": 0.317, + "step": 8757 + }, + { + "epoch": 4.754614549402823, + "grad_norm": 11.794412367069619, + "learning_rate": 1.2607740344693474e-07, + "loss": 0.5, + "step": 8758 + }, + { + "epoch": 4.755157437567861, + "grad_norm": 10.997145578080557, + "learning_rate": 1.2552132465454282e-07, + "loss": 0.2869, + "step": 8759 + }, + { + "epoch": 4.755700325732899, + "grad_norm": 11.55907046766932, + "learning_rate": 1.2496646715341231e-07, + "loss": 0.2764, + "step": 8760 + }, + { + "epoch": 4.756243213897937, + "grad_norm": 24.08335246288593, + "learning_rate": 1.2441283101216727e-07, + "loss": 1.1823, + "step": 8761 + }, + { + "epoch": 4.756786102062975, + "grad_norm": 14.279299755644365, + "learning_rate": 1.238604162992818e-07, + "loss": 0.7198, + "step": 8762 + }, + { + "epoch": 4.757328990228013, + "grad_norm": 13.336435628960723, + "learning_rate": 1.2330922308308125e-07, + "loss": 0.341, + "step": 8763 + }, + { + "epoch": 4.757871878393051, + "grad_norm": 13.97355889589316, + "learning_rate": 1.2275925143173662e-07, + "loss": 0.4514, + "step": 8764 + }, + { + "epoch": 4.758414766558089, + "grad_norm": 10.201594507019205, + "learning_rate": 1.2221050141327018e-07, + "loss": 0.4758, + "step": 8765 + }, + { + "epoch": 4.758957654723127, + "grad_norm": 8.477171079397465, + "learning_rate": 1.2166297309555098e-07, + "loss": 0.2141, + "step": 8766 + }, + { + "epoch": 4.759500542888165, + "grad_norm": 11.363514355058907, + "learning_rate": 1.211166665462993e-07, + "loss": 0.2775, + "step": 8767 + }, + { + "epoch": 4.760043431053203, + "grad_norm": 12.151898918162795, + "learning_rate": 1.2057158183308215e-07, + "loss": 0.3867, + "step": 8768 + }, + { + "epoch": 4.760586319218241, + "grad_norm": 12.369021167087942, + "learning_rate": 1.20027719023319e-07, + "loss": 0.3602, + "step": 8769 + }, + { + "epoch": 4.761129207383279, + "grad_norm": 16.318528391362594, + "learning_rate": 1.1948507818427269e-07, + "loss": 0.5529, + "step": 8770 + }, + { + "epoch": 4.761672095548317, + "grad_norm": 17.758768019553752, + "learning_rate": 1.1894365938305952e-07, + "loss": 0.7753, + "step": 8771 + }, + { + "epoch": 4.762214983713355, + "grad_norm": 11.920196045244229, + "learning_rate": 1.184034626866426e-07, + "loss": 0.36, + "step": 8772 + }, + { + "epoch": 4.762757871878393, + "grad_norm": 13.291930947753695, + "learning_rate": 1.1786448816183516e-07, + "loss": 0.5906, + "step": 8773 + }, + { + "epoch": 4.763300760043431, + "grad_norm": 14.292097175382699, + "learning_rate": 1.1732673587529609e-07, + "loss": 0.3195, + "step": 8774 + }, + { + "epoch": 4.763843648208469, + "grad_norm": 9.609372443037184, + "learning_rate": 1.1679020589353662e-07, + "loss": 0.2597, + "step": 8775 + }, + { + "epoch": 4.764386536373507, + "grad_norm": 13.447235691488418, + "learning_rate": 1.1625489828291814e-07, + "loss": 0.2508, + "step": 8776 + }, + { + "epoch": 4.764929424538545, + "grad_norm": 16.230999706702107, + "learning_rate": 1.1572081310964434e-07, + "loss": 0.7529, + "step": 8777 + }, + { + "epoch": 4.765472312703583, + "grad_norm": 12.727526228421432, + "learning_rate": 1.151879504397746e-07, + "loss": 0.3311, + "step": 8778 + }, + { + "epoch": 4.766015200868621, + "grad_norm": 10.438531330742874, + "learning_rate": 1.1465631033921288e-07, + "loss": 0.2289, + "step": 8779 + }, + { + "epoch": 4.766558089033659, + "grad_norm": 14.295286154658784, + "learning_rate": 1.1412589287371323e-07, + "loss": 0.3953, + "step": 8780 + }, + { + "epoch": 4.767100977198697, + "grad_norm": 9.693437885130846, + "learning_rate": 1.1359669810887985e-07, + "loss": 0.2993, + "step": 8781 + }, + { + "epoch": 4.767643865363735, + "grad_norm": 15.32504375625243, + "learning_rate": 1.1306872611016151e-07, + "loss": 0.4768, + "step": 8782 + }, + { + "epoch": 4.768186753528773, + "grad_norm": 11.211203320225184, + "learning_rate": 1.1254197694286262e-07, + "loss": 0.2661, + "step": 8783 + }, + { + "epoch": 4.768729641693811, + "grad_norm": 13.67914440990922, + "learning_rate": 1.1201645067212886e-07, + "loss": 0.5022, + "step": 8784 + }, + { + "epoch": 4.769272529858849, + "grad_norm": 11.569900535556993, + "learning_rate": 1.1149214736296044e-07, + "loss": 0.3133, + "step": 8785 + }, + { + "epoch": 4.769815418023887, + "grad_norm": 17.63101629527424, + "learning_rate": 1.1096906708020216e-07, + "loss": 0.5855, + "step": 8786 + }, + { + "epoch": 4.770358306188925, + "grad_norm": 13.27760517619257, + "learning_rate": 1.1044720988855118e-07, + "loss": 0.3421, + "step": 8787 + }, + { + "epoch": 4.770901194353963, + "grad_norm": 15.992691867357234, + "learning_rate": 1.099265758525514e-07, + "loss": 0.4832, + "step": 8788 + }, + { + "epoch": 4.771444082519001, + "grad_norm": 13.15390855364886, + "learning_rate": 1.0940716503659465e-07, + "loss": 0.3082, + "step": 8789 + }, + { + "epoch": 4.771986970684039, + "grad_norm": 18.263335547323074, + "learning_rate": 1.0888897750492289e-07, + "loss": 0.556, + "step": 8790 + }, + { + "epoch": 4.772529858849077, + "grad_norm": 10.441020903025926, + "learning_rate": 1.0837201332162595e-07, + "loss": 0.2007, + "step": 8791 + }, + { + "epoch": 4.773072747014115, + "grad_norm": 15.037462451331834, + "learning_rate": 1.0785627255064379e-07, + "loss": 0.6108, + "step": 8792 + }, + { + "epoch": 4.773615635179153, + "grad_norm": 14.518203414614236, + "learning_rate": 1.0734175525576429e-07, + "loss": 0.444, + "step": 8793 + }, + { + "epoch": 4.774158523344191, + "grad_norm": 11.643159629017385, + "learning_rate": 1.068284615006221e-07, + "loss": 0.4596, + "step": 8794 + }, + { + "epoch": 4.774701411509229, + "grad_norm": 15.528158500984132, + "learning_rate": 1.0631639134870308e-07, + "loss": 0.6026, + "step": 8795 + }, + { + "epoch": 4.7752442996742674, + "grad_norm": 19.44862903645615, + "learning_rate": 1.0580554486334216e-07, + "loss": 0.5733, + "step": 8796 + }, + { + "epoch": 4.775787187839305, + "grad_norm": 13.48756952731863, + "learning_rate": 1.0529592210771988e-07, + "loss": 0.3872, + "step": 8797 + }, + { + "epoch": 4.776330076004343, + "grad_norm": 13.820366819889347, + "learning_rate": 1.0478752314486695e-07, + "loss": 0.3807, + "step": 8798 + }, + { + "epoch": 4.776872964169381, + "grad_norm": 14.931351167549256, + "learning_rate": 1.0428034803766418e-07, + "loss": 0.4707, + "step": 8799 + }, + { + "epoch": 4.7774158523344195, + "grad_norm": 9.35754799853792, + "learning_rate": 1.0377439684884027e-07, + "loss": 0.2955, + "step": 8800 + }, + { + "epoch": 4.777958740499457, + "grad_norm": 15.469307360429545, + "learning_rate": 1.0326966964096963e-07, + "loss": 0.5782, + "step": 8801 + }, + { + "epoch": 4.778501628664495, + "grad_norm": 13.056079685648413, + "learning_rate": 1.0276616647648008e-07, + "loss": 0.5096, + "step": 8802 + }, + { + "epoch": 4.779044516829533, + "grad_norm": 13.862999489833173, + "learning_rate": 1.0226388741764514e-07, + "loss": 0.2736, + "step": 8803 + }, + { + "epoch": 4.7795874049945715, + "grad_norm": 10.890688599470478, + "learning_rate": 1.0176283252658737e-07, + "loss": 0.3334, + "step": 8804 + }, + { + "epoch": 4.780130293159609, + "grad_norm": 14.415866724042699, + "learning_rate": 1.0126300186527715e-07, + "loss": 0.5933, + "step": 8805 + }, + { + "epoch": 4.780673181324647, + "grad_norm": 13.555552920820668, + "learning_rate": 1.0076439549553507e-07, + "loss": 0.3958, + "step": 8806 + }, + { + "epoch": 4.781216069489685, + "grad_norm": 11.893874625329541, + "learning_rate": 1.0026701347903067e-07, + "loss": 0.3873, + "step": 8807 + }, + { + "epoch": 4.7817589576547235, + "grad_norm": 15.261946481554697, + "learning_rate": 9.977085587727919e-08, + "loss": 0.404, + "step": 8808 + }, + { + "epoch": 4.782301845819761, + "grad_norm": 11.516422308885032, + "learning_rate": 9.92759227516471e-08, + "loss": 0.318, + "step": 8809 + }, + { + "epoch": 4.782844733984799, + "grad_norm": 13.915845024156635, + "learning_rate": 9.878221416334877e-08, + "loss": 0.4227, + "step": 8810 + }, + { + "epoch": 4.783387622149837, + "grad_norm": 12.05067037271945, + "learning_rate": 9.828973017344534e-08, + "loss": 0.3802, + "step": 8811 + }, + { + "epoch": 4.7839305103148755, + "grad_norm": 12.366225498487221, + "learning_rate": 9.779847084284921e-08, + "loss": 0.5068, + "step": 8812 + }, + { + "epoch": 4.784473398479913, + "grad_norm": 11.020281825268318, + "learning_rate": 9.730843623231956e-08, + "loss": 0.4372, + "step": 8813 + }, + { + "epoch": 4.785016286644951, + "grad_norm": 11.115605935865265, + "learning_rate": 9.681962640246679e-08, + "loss": 0.3556, + "step": 8814 + }, + { + "epoch": 4.785559174809989, + "grad_norm": 15.746022339127645, + "learning_rate": 9.633204141374475e-08, + "loss": 0.4654, + "step": 8815 + }, + { + "epoch": 4.7861020629750275, + "grad_norm": 11.675199453622405, + "learning_rate": 9.584568132645966e-08, + "loss": 0.3, + "step": 8816 + }, + { + "epoch": 4.786644951140065, + "grad_norm": 15.071011265966053, + "learning_rate": 9.536054620076563e-08, + "loss": 0.3563, + "step": 8817 + }, + { + "epoch": 4.787187839305103, + "grad_norm": 14.392285757682815, + "learning_rate": 9.487663609666576e-08, + "loss": 0.3403, + "step": 8818 + }, + { + "epoch": 4.787730727470141, + "grad_norm": 14.478548975904134, + "learning_rate": 9.439395107400995e-08, + "loss": 0.4303, + "step": 8819 + }, + { + "epoch": 4.7882736156351795, + "grad_norm": 16.873014104043584, + "learning_rate": 9.391249119249601e-08, + "loss": 0.4669, + "step": 8820 + }, + { + "epoch": 4.788816503800217, + "grad_norm": 12.850529236209315, + "learning_rate": 9.34322565116752e-08, + "loss": 0.36, + "step": 8821 + }, + { + "epoch": 4.789359391965255, + "grad_norm": 12.049173435805207, + "learning_rate": 9.29532470909389e-08, + "loss": 0.4809, + "step": 8822 + }, + { + "epoch": 4.789902280130293, + "grad_norm": 14.445567418948913, + "learning_rate": 9.247546298953747e-08, + "loss": 0.375, + "step": 8823 + }, + { + "epoch": 4.7904451682953315, + "grad_norm": 16.331300178692935, + "learning_rate": 9.19989042665581e-08, + "loss": 0.5994, + "step": 8824 + }, + { + "epoch": 4.790988056460369, + "grad_norm": 16.563817844672233, + "learning_rate": 9.1523570980947e-08, + "loss": 0.4818, + "step": 8825 + }, + { + "epoch": 4.791530944625407, + "grad_norm": 12.9314417840519, + "learning_rate": 9.104946319149266e-08, + "loss": 0.5215, + "step": 8826 + }, + { + "epoch": 4.792073832790445, + "grad_norm": 11.600758934710594, + "learning_rate": 9.057658095683264e-08, + "loss": 0.4369, + "step": 8827 + }, + { + "epoch": 4.7926167209554835, + "grad_norm": 11.593366700305912, + "learning_rate": 9.010492433545459e-08, + "loss": 0.3912, + "step": 8828 + }, + { + "epoch": 4.793159609120521, + "grad_norm": 14.357902988841598, + "learning_rate": 8.963449338569297e-08, + "loss": 0.5229, + "step": 8829 + }, + { + "epoch": 4.793702497285559, + "grad_norm": 10.079996265707912, + "learning_rate": 8.916528816573234e-08, + "loss": 0.4725, + "step": 8830 + }, + { + "epoch": 4.794245385450597, + "grad_norm": 13.722106385687223, + "learning_rate": 8.869730873360405e-08, + "loss": 0.4548, + "step": 8831 + }, + { + "epoch": 4.7947882736156355, + "grad_norm": 14.066258092224395, + "learning_rate": 8.82305551471896e-08, + "loss": 0.5001, + "step": 8832 + }, + { + "epoch": 4.795331161780673, + "grad_norm": 12.082732968978428, + "learning_rate": 8.776502746421611e-08, + "loss": 0.3134, + "step": 8833 + }, + { + "epoch": 4.795874049945711, + "grad_norm": 14.022649629603004, + "learning_rate": 8.730072574226311e-08, + "loss": 0.3467, + "step": 8834 + }, + { + "epoch": 4.796416938110749, + "grad_norm": 15.002589519279939, + "learning_rate": 8.683765003875356e-08, + "loss": 0.4259, + "step": 8835 + }, + { + "epoch": 4.7969598262757875, + "grad_norm": 14.91562383479778, + "learning_rate": 8.637580041096382e-08, + "loss": 0.4919, + "step": 8836 + }, + { + "epoch": 4.797502714440825, + "grad_norm": 12.718843473649267, + "learning_rate": 8.59151769160138e-08, + "loss": 0.6194, + "step": 8837 + }, + { + "epoch": 4.798045602605863, + "grad_norm": 12.934801394638406, + "learning_rate": 8.545577961087681e-08, + "loss": 0.4032, + "step": 8838 + }, + { + "epoch": 4.798588490770901, + "grad_norm": 15.44257297198036, + "learning_rate": 8.49976085523696e-08, + "loss": 0.4059, + "step": 8839 + }, + { + "epoch": 4.7991313789359396, + "grad_norm": 11.274515134729002, + "learning_rate": 8.454066379716019e-08, + "loss": 0.2459, + "step": 8840 + }, + { + "epoch": 4.799674267100977, + "grad_norm": 13.639296201505058, + "learning_rate": 8.408494540176448e-08, + "loss": 0.4588, + "step": 8841 + }, + { + "epoch": 4.800217155266015, + "grad_norm": 10.677817137466716, + "learning_rate": 8.363045342254628e-08, + "loss": 0.2931, + "step": 8842 + }, + { + "epoch": 4.800760043431053, + "grad_norm": 14.87764794308729, + "learning_rate": 8.317718791571838e-08, + "loss": 0.5248, + "step": 8843 + }, + { + "epoch": 4.801302931596092, + "grad_norm": 11.064095389068989, + "learning_rate": 8.27251489373404e-08, + "loss": 0.2687, + "step": 8844 + }, + { + "epoch": 4.801845819761129, + "grad_norm": 22.55978148923217, + "learning_rate": 8.227433654332206e-08, + "loss": 0.862, + "step": 8845 + }, + { + "epoch": 4.802388707926167, + "grad_norm": 12.685033016571245, + "learning_rate": 8.182475078941987e-08, + "loss": 0.3498, + "step": 8846 + }, + { + "epoch": 4.802931596091205, + "grad_norm": 14.666620778959023, + "learning_rate": 8.137639173124046e-08, + "loss": 0.4637, + "step": 8847 + }, + { + "epoch": 4.803474484256244, + "grad_norm": 12.025170948479563, + "learning_rate": 8.092925942423613e-08, + "loss": 0.4287, + "step": 8848 + }, + { + "epoch": 4.804017372421281, + "grad_norm": 15.47756406381011, + "learning_rate": 8.048335392371042e-08, + "loss": 0.4397, + "step": 8849 + }, + { + "epoch": 4.804560260586319, + "grad_norm": 15.134699918755489, + "learning_rate": 8.003867528481257e-08, + "loss": 0.7081, + "step": 8850 + }, + { + "epoch": 4.805103148751357, + "grad_norm": 12.320316293618925, + "learning_rate": 7.959522356254079e-08, + "loss": 0.3828, + "step": 8851 + }, + { + "epoch": 4.805646036916396, + "grad_norm": 11.222062063383326, + "learning_rate": 7.915299881174344e-08, + "loss": 0.4006, + "step": 8852 + }, + { + "epoch": 4.806188925081433, + "grad_norm": 12.422853158493984, + "learning_rate": 7.871200108711341e-08, + "loss": 0.422, + "step": 8853 + }, + { + "epoch": 4.806731813246471, + "grad_norm": 9.781525068337004, + "learning_rate": 7.82722304431971e-08, + "loss": 0.2884, + "step": 8854 + }, + { + "epoch": 4.807274701411509, + "grad_norm": 11.760204421271188, + "learning_rate": 7.783368693438209e-08, + "loss": 0.4436, + "step": 8855 + }, + { + "epoch": 4.807817589576548, + "grad_norm": 10.811544328919764, + "learning_rate": 7.739637061491279e-08, + "loss": 0.2036, + "step": 8856 + }, + { + "epoch": 4.808360477741585, + "grad_norm": 16.51916863908083, + "learning_rate": 7.696028153887259e-08, + "loss": 0.7597, + "step": 8857 + }, + { + "epoch": 4.808903365906623, + "grad_norm": 18.728919815805536, + "learning_rate": 7.652541976019945e-08, + "loss": 0.4963, + "step": 8858 + }, + { + "epoch": 4.809446254071661, + "grad_norm": 12.95657669087096, + "learning_rate": 7.609178533268036e-08, + "loss": 0.4693, + "step": 8859 + }, + { + "epoch": 4.8099891422367, + "grad_norm": 11.87170444365128, + "learning_rate": 7.565937830994353e-08, + "loss": 0.4667, + "step": 8860 + }, + { + "epoch": 4.810532030401737, + "grad_norm": 10.885904559128528, + "learning_rate": 7.522819874547283e-08, + "loss": 0.2819, + "step": 8861 + }, + { + "epoch": 4.811074918566775, + "grad_norm": 11.241238204039677, + "learning_rate": 7.479824669259561e-08, + "loss": 0.2828, + "step": 8862 + }, + { + "epoch": 4.811617806731813, + "grad_norm": 12.844422586502446, + "learning_rate": 7.436952220449045e-08, + "loss": 0.4597, + "step": 8863 + }, + { + "epoch": 4.812160694896852, + "grad_norm": 10.410880488204011, + "learning_rate": 7.394202533418049e-08, + "loss": 0.3583, + "step": 8864 + }, + { + "epoch": 4.812703583061889, + "grad_norm": 10.695248824341306, + "learning_rate": 7.351575613454231e-08, + "loss": 0.3529, + "step": 8865 + }, + { + "epoch": 4.813246471226927, + "grad_norm": 15.957870026482412, + "learning_rate": 7.309071465829487e-08, + "loss": 0.4339, + "step": 8866 + }, + { + "epoch": 4.813789359391965, + "grad_norm": 10.337571437282099, + "learning_rate": 7.266690095800944e-08, + "loss": 0.2697, + "step": 8867 + }, + { + "epoch": 4.814332247557004, + "grad_norm": 11.474450789929184, + "learning_rate": 7.224431508610296e-08, + "loss": 0.2208, + "step": 8868 + }, + { + "epoch": 4.814875135722041, + "grad_norm": 9.448799109853498, + "learning_rate": 7.182295709484255e-08, + "loss": 0.24, + "step": 8869 + }, + { + "epoch": 4.815418023887079, + "grad_norm": 10.70255325665073, + "learning_rate": 7.140282703634204e-08, + "loss": 0.4463, + "step": 8870 + }, + { + "epoch": 4.815960912052117, + "grad_norm": 15.464670090867934, + "learning_rate": 7.098392496256324e-08, + "loss": 0.505, + "step": 8871 + }, + { + "epoch": 4.816503800217156, + "grad_norm": 12.515924687979576, + "learning_rate": 7.056625092531688e-08, + "loss": 0.3723, + "step": 8872 + }, + { + "epoch": 4.817046688382193, + "grad_norm": 12.22224030809159, + "learning_rate": 7.014980497626279e-08, + "loss": 0.3378, + "step": 8873 + }, + { + "epoch": 4.817589576547231, + "grad_norm": 18.212411386580676, + "learning_rate": 6.973458716690639e-08, + "loss": 0.5945, + "step": 8874 + }, + { + "epoch": 4.818132464712269, + "grad_norm": 12.548255509888588, + "learning_rate": 6.93205975486022e-08, + "loss": 0.3672, + "step": 8875 + }, + { + "epoch": 4.818675352877308, + "grad_norm": 10.738878454120973, + "learning_rate": 6.89078361725537e-08, + "loss": 0.3582, + "step": 8876 + }, + { + "epoch": 4.819218241042345, + "grad_norm": 13.763430071394062, + "learning_rate": 6.849630308981226e-08, + "loss": 0.4466, + "step": 8877 + }, + { + "epoch": 4.819761129207383, + "grad_norm": 14.698245992825742, + "learning_rate": 6.808599835127605e-08, + "loss": 0.4725, + "step": 8878 + }, + { + "epoch": 4.820304017372421, + "grad_norm": 17.5106844905948, + "learning_rate": 6.767692200769226e-08, + "loss": 0.582, + "step": 8879 + }, + { + "epoch": 4.82084690553746, + "grad_norm": 11.772458890524595, + "learning_rate": 6.726907410965822e-08, + "loss": 0.3311, + "step": 8880 + }, + { + "epoch": 4.821389793702497, + "grad_norm": 10.79164174819329, + "learning_rate": 6.686245470761355e-08, + "loss": 0.299, + "step": 8881 + }, + { + "epoch": 4.821932681867535, + "grad_norm": 12.844928905031605, + "learning_rate": 6.645706385185246e-08, + "loss": 0.3807, + "step": 8882 + }, + { + "epoch": 4.822475570032573, + "grad_norm": 21.17889161763075, + "learning_rate": 6.605290159251376e-08, + "loss": 0.6281, + "step": 8883 + }, + { + "epoch": 4.823018458197612, + "grad_norm": 10.727221853449628, + "learning_rate": 6.564996797958412e-08, + "loss": 0.3694, + "step": 8884 + }, + { + "epoch": 4.823561346362649, + "grad_norm": 16.28670128227416, + "learning_rate": 6.524826306289921e-08, + "loss": 0.4659, + "step": 8885 + }, + { + "epoch": 4.824104234527687, + "grad_norm": 17.1662871993217, + "learning_rate": 6.484778689214266e-08, + "loss": 0.5159, + "step": 8886 + }, + { + "epoch": 4.824647122692725, + "grad_norm": 12.292228244187442, + "learning_rate": 6.444853951684704e-08, + "loss": 0.2194, + "step": 8887 + }, + { + "epoch": 4.825190010857764, + "grad_norm": 13.565563908072575, + "learning_rate": 6.405052098639065e-08, + "loss": 0.4117, + "step": 8888 + }, + { + "epoch": 4.8257328990228014, + "grad_norm": 13.37247191878539, + "learning_rate": 6.365373135000075e-08, + "loss": 0.3119, + "step": 8889 + }, + { + "epoch": 4.826275787187839, + "grad_norm": 14.204941355106659, + "learning_rate": 6.325817065675366e-08, + "loss": 0.3838, + "step": 8890 + }, + { + "epoch": 4.826818675352877, + "grad_norm": 11.742712728879413, + "learning_rate": 6.286383895557357e-08, + "loss": 0.5932, + "step": 8891 + }, + { + "epoch": 4.827361563517916, + "grad_norm": 10.887835991446542, + "learning_rate": 6.247073629523037e-08, + "loss": 0.3227, + "step": 8892 + }, + { + "epoch": 4.8279044516829535, + "grad_norm": 12.144000261688534, + "learning_rate": 6.207886272434516e-08, + "loss": 0.3384, + "step": 8893 + }, + { + "epoch": 4.828447339847991, + "grad_norm": 12.385730983063644, + "learning_rate": 6.168821829138472e-08, + "loss": 0.2895, + "step": 8894 + }, + { + "epoch": 4.828990228013029, + "grad_norm": 15.682138391036652, + "learning_rate": 6.129880304466484e-08, + "loss": 0.5902, + "step": 8895 + }, + { + "epoch": 4.829533116178068, + "grad_norm": 16.02759596723009, + "learning_rate": 6.091061703234924e-08, + "loss": 0.5285, + "step": 8896 + }, + { + "epoch": 4.8300760043431055, + "grad_norm": 14.27361095848103, + "learning_rate": 6.052366030244838e-08, + "loss": 0.4956, + "step": 8897 + }, + { + "epoch": 4.830618892508143, + "grad_norm": 11.386206832016665, + "learning_rate": 6.013793290282289e-08, + "loss": 0.3047, + "step": 8898 + }, + { + "epoch": 4.831161780673181, + "grad_norm": 10.195950730327096, + "learning_rate": 5.975343488118013e-08, + "loss": 0.3495, + "step": 8899 + }, + { + "epoch": 4.83170466883822, + "grad_norm": 12.21542669488425, + "learning_rate": 5.9370166285073195e-08, + "loss": 0.5064, + "step": 8900 + }, + { + "epoch": 4.8322475570032575, + "grad_norm": 14.215817918389403, + "learning_rate": 5.8988127161908603e-08, + "loss": 0.478, + "step": 8901 + }, + { + "epoch": 4.832790445168295, + "grad_norm": 12.971102099976296, + "learning_rate": 5.8607317558935226e-08, + "loss": 0.374, + "step": 8902 + }, + { + "epoch": 4.833333333333333, + "grad_norm": 7.241451153701665, + "learning_rate": 5.822773752325317e-08, + "loss": 0.3883, + "step": 8903 + }, + { + "epoch": 4.833876221498372, + "grad_norm": 12.147250550465268, + "learning_rate": 5.784938710181043e-08, + "loss": 0.3978, + "step": 8904 + }, + { + "epoch": 4.8344191096634095, + "grad_norm": 18.246476649890155, + "learning_rate": 5.747226634139957e-08, + "loss": 0.3516, + "step": 8905 + }, + { + "epoch": 4.834961997828447, + "grad_norm": 10.69413118790195, + "learning_rate": 5.709637528866552e-08, + "loss": 0.4941, + "step": 8906 + }, + { + "epoch": 4.835504885993485, + "grad_norm": 16.53209092388979, + "learning_rate": 5.672171399009774e-08, + "loss": 0.4765, + "step": 8907 + }, + { + "epoch": 4.836047774158524, + "grad_norm": 13.199658324083233, + "learning_rate": 5.6348282492036946e-08, + "loss": 0.3345, + "step": 8908 + }, + { + "epoch": 4.8365906623235615, + "grad_norm": 15.65690883916084, + "learning_rate": 5.59760808406673e-08, + "loss": 0.4327, + "step": 8909 + }, + { + "epoch": 4.837133550488599, + "grad_norm": 13.187099385246231, + "learning_rate": 5.5605109082024215e-08, + "loss": 0.6508, + "step": 8910 + }, + { + "epoch": 4.837676438653637, + "grad_norm": 15.637262642220758, + "learning_rate": 5.523536726199097e-08, + "loss": 0.3265, + "step": 8911 + }, + { + "epoch": 4.838219326818676, + "grad_norm": 16.12059897202224, + "learning_rate": 5.486685542629655e-08, + "loss": 0.5249, + "step": 8912 + }, + { + "epoch": 4.8387622149837135, + "grad_norm": 13.559084799401953, + "learning_rate": 5.449957362052005e-08, + "loss": 0.5622, + "step": 8913 + }, + { + "epoch": 4.839305103148751, + "grad_norm": 10.42459560738085, + "learning_rate": 5.413352189008736e-08, + "loss": 0.2996, + "step": 8914 + }, + { + "epoch": 4.839847991313789, + "grad_norm": 16.46805590997443, + "learning_rate": 5.3768700280271146e-08, + "loss": 0.5508, + "step": 8915 + }, + { + "epoch": 4.840390879478828, + "grad_norm": 15.003668947871548, + "learning_rate": 5.340510883619421e-08, + "loss": 0.4085, + "step": 8916 + }, + { + "epoch": 4.8409337676438655, + "grad_norm": 8.64870666065282, + "learning_rate": 5.3042747602826125e-08, + "loss": 0.2792, + "step": 8917 + }, + { + "epoch": 4.841476655808903, + "grad_norm": 13.21055869887479, + "learning_rate": 5.268161662498439e-08, + "loss": 0.3822, + "step": 8918 + }, + { + "epoch": 4.842019543973941, + "grad_norm": 15.48767742576848, + "learning_rate": 5.2321715947333264e-08, + "loss": 0.6979, + "step": 8919 + }, + { + "epoch": 4.84256243213898, + "grad_norm": 13.032634397464806, + "learning_rate": 5.196304561438714e-08, + "loss": 0.4676, + "step": 8920 + }, + { + "epoch": 4.8431053203040175, + "grad_norm": 12.108816113672754, + "learning_rate": 5.1605605670506095e-08, + "loss": 0.4098, + "step": 8921 + }, + { + "epoch": 4.843648208469055, + "grad_norm": 9.883044917135045, + "learning_rate": 5.124939615989921e-08, + "loss": 0.3021, + "step": 8922 + }, + { + "epoch": 4.844191096634093, + "grad_norm": 15.02353598376969, + "learning_rate": 5.089441712662346e-08, + "loss": 0.5224, + "step": 8923 + }, + { + "epoch": 4.844733984799132, + "grad_norm": 14.500832790037261, + "learning_rate": 5.054066861458151e-08, + "loss": 0.3961, + "step": 8924 + }, + { + "epoch": 4.8452768729641695, + "grad_norm": 14.045099169119101, + "learning_rate": 5.0188150667528356e-08, + "loss": 0.2919, + "step": 8925 + }, + { + "epoch": 4.845819761129207, + "grad_norm": 11.608813846717336, + "learning_rate": 4.9836863329061346e-08, + "loss": 0.303, + "step": 8926 + }, + { + "epoch": 4.846362649294245, + "grad_norm": 13.766577335723845, + "learning_rate": 4.948680664262906e-08, + "loss": 0.4839, + "step": 8927 + }, + { + "epoch": 4.846905537459284, + "grad_norm": 16.843435011316796, + "learning_rate": 4.913798065152797e-08, + "loss": 0.5213, + "step": 8928 + }, + { + "epoch": 4.8474484256243215, + "grad_norm": 10.74962929723452, + "learning_rate": 4.879038539890024e-08, + "loss": 0.3224, + "step": 8929 + }, + { + "epoch": 4.847991313789359, + "grad_norm": 14.297515517628703, + "learning_rate": 4.844402092773815e-08, + "loss": 0.6642, + "step": 8930 + }, + { + "epoch": 4.848534201954397, + "grad_norm": 10.40231274048184, + "learning_rate": 4.809888728087853e-08, + "loss": 0.2788, + "step": 8931 + }, + { + "epoch": 4.849077090119436, + "grad_norm": 14.099191061940111, + "learning_rate": 4.775498450101057e-08, + "loss": 0.455, + "step": 8932 + }, + { + "epoch": 4.8496199782844736, + "grad_norm": 14.245133009805686, + "learning_rate": 4.7412312630666925e-08, + "loss": 0.4375, + "step": 8933 + }, + { + "epoch": 4.850162866449511, + "grad_norm": 13.88233182279541, + "learning_rate": 4.707087171223146e-08, + "loss": 0.367, + "step": 8934 + }, + { + "epoch": 4.850705754614549, + "grad_norm": 16.896162529240044, + "learning_rate": 4.673066178793262e-08, + "loss": 0.7301, + "step": 8935 + }, + { + "epoch": 4.851248642779588, + "grad_norm": 15.42411861883886, + "learning_rate": 4.6391682899848966e-08, + "loss": 0.5421, + "step": 8936 + }, + { + "epoch": 4.851791530944626, + "grad_norm": 13.430931121094932, + "learning_rate": 4.6053935089905856e-08, + "loss": 0.5728, + "step": 8937 + }, + { + "epoch": 4.852334419109663, + "grad_norm": 11.78549017995068, + "learning_rate": 4.5717418399875426e-08, + "loss": 0.3377, + "step": 8938 + }, + { + "epoch": 4.852877307274701, + "grad_norm": 9.979103448244798, + "learning_rate": 4.5382132871381046e-08, + "loss": 0.2352, + "step": 8939 + }, + { + "epoch": 4.85342019543974, + "grad_norm": 14.36690978126345, + "learning_rate": 4.504807854588844e-08, + "loss": 0.4036, + "step": 8940 + }, + { + "epoch": 4.853963083604778, + "grad_norm": 15.465505131974233, + "learning_rate": 4.4715255464715665e-08, + "loss": 0.3305, + "step": 8941 + }, + { + "epoch": 4.854505971769815, + "grad_norm": 11.075613879055394, + "learning_rate": 4.438366366902647e-08, + "loss": 0.3939, + "step": 8942 + }, + { + "epoch": 4.855048859934853, + "grad_norm": 9.494149902633026, + "learning_rate": 4.4053303199832474e-08, + "loss": 0.455, + "step": 8943 + }, + { + "epoch": 4.855591748099892, + "grad_norm": 13.435606504793506, + "learning_rate": 4.372417409799323e-08, + "loss": 0.7939, + "step": 8944 + }, + { + "epoch": 4.85613463626493, + "grad_norm": 15.032529072722495, + "learning_rate": 4.3396276404216175e-08, + "loss": 0.4496, + "step": 8945 + }, + { + "epoch": 4.856677524429967, + "grad_norm": 15.656375813559288, + "learning_rate": 4.3069610159056644e-08, + "loss": 0.4964, + "step": 8946 + }, + { + "epoch": 4.857220412595005, + "grad_norm": 10.906491414551686, + "learning_rate": 4.274417540291564e-08, + "loss": 0.3617, + "step": 8947 + }, + { + "epoch": 4.857763300760044, + "grad_norm": 13.569426587375315, + "learning_rate": 4.241997217604543e-08, + "loss": 0.5283, + "step": 8948 + }, + { + "epoch": 4.858306188925082, + "grad_norm": 10.691629929006671, + "learning_rate": 4.209700051854282e-08, + "loss": 0.1982, + "step": 8949 + }, + { + "epoch": 4.858849077090119, + "grad_norm": 10.71016148764362, + "learning_rate": 4.177526047035363e-08, + "loss": 0.4083, + "step": 8950 + }, + { + "epoch": 4.859391965255157, + "grad_norm": 14.30919702980143, + "learning_rate": 4.1454752071271586e-08, + "loss": 0.4441, + "step": 8951 + }, + { + "epoch": 4.859934853420196, + "grad_norm": 15.73961142029379, + "learning_rate": 4.11354753609372e-08, + "loss": 0.3984, + "step": 8952 + }, + { + "epoch": 4.860477741585234, + "grad_norm": 17.417654306288096, + "learning_rate": 4.0817430378840004e-08, + "loss": 0.4194, + "step": 8953 + }, + { + "epoch": 4.861020629750271, + "grad_norm": 13.52226655944115, + "learning_rate": 4.050061716431408e-08, + "loss": 0.3684, + "step": 8954 + }, + { + "epoch": 4.861563517915309, + "grad_norm": 9.814382598914019, + "learning_rate": 4.018503575654587e-08, + "loss": 0.7053, + "step": 8955 + }, + { + "epoch": 4.862106406080348, + "grad_norm": 11.24338506908338, + "learning_rate": 3.987068619456635e-08, + "loss": 0.3247, + "step": 8956 + }, + { + "epoch": 4.862649294245386, + "grad_norm": 12.546220960893276, + "learning_rate": 3.955756851725334e-08, + "loss": 0.2883, + "step": 8957 + }, + { + "epoch": 4.863192182410423, + "grad_norm": 9.883047461274098, + "learning_rate": 3.9245682763335846e-08, + "loss": 0.2081, + "step": 8958 + }, + { + "epoch": 4.863735070575461, + "grad_norm": 11.723178509404748, + "learning_rate": 3.8935028971386344e-08, + "loss": 0.2868, + "step": 8959 + }, + { + "epoch": 4.8642779587405, + "grad_norm": 14.517146742383616, + "learning_rate": 3.862560717982855e-08, + "loss": 0.4062, + "step": 8960 + }, + { + "epoch": 4.864820846905538, + "grad_norm": 14.948031344029658, + "learning_rate": 3.8317417426931845e-08, + "loss": 0.5253, + "step": 8961 + }, + { + "epoch": 4.865363735070575, + "grad_norm": 12.386658970702634, + "learning_rate": 3.801045975081352e-08, + "loss": 0.4839, + "step": 8962 + }, + { + "epoch": 4.865906623235613, + "grad_norm": 15.354605902256184, + "learning_rate": 3.770473418943876e-08, + "loss": 0.4133, + "step": 8963 + }, + { + "epoch": 4.866449511400652, + "grad_norm": 10.875285258256444, + "learning_rate": 3.7400240780619543e-08, + "loss": 0.3914, + "step": 8964 + }, + { + "epoch": 4.86699239956569, + "grad_norm": 12.106013147842344, + "learning_rate": 3.709697956201686e-08, + "loss": 0.3433, + "step": 8965 + }, + { + "epoch": 4.867535287730727, + "grad_norm": 13.309078271025745, + "learning_rate": 3.6794950571138464e-08, + "loss": 0.331, + "step": 8966 + }, + { + "epoch": 4.868078175895765, + "grad_norm": 10.487793011045266, + "learning_rate": 3.649415384533894e-08, + "loss": 0.2756, + "step": 8967 + }, + { + "epoch": 4.868621064060804, + "grad_norm": 15.14197511158265, + "learning_rate": 3.619458942182297e-08, + "loss": 0.854, + "step": 8968 + }, + { + "epoch": 4.869163952225842, + "grad_norm": 11.373840117881688, + "learning_rate": 3.589625733763869e-08, + "loss": 0.3237, + "step": 8969 + }, + { + "epoch": 4.869706840390879, + "grad_norm": 14.405255774825118, + "learning_rate": 3.5599157629686576e-08, + "loss": 0.5447, + "step": 8970 + }, + { + "epoch": 4.870249728555917, + "grad_norm": 9.844386093357125, + "learning_rate": 3.5303290334711695e-08, + "loss": 0.4378, + "step": 8971 + }, + { + "epoch": 4.870792616720956, + "grad_norm": 13.588627269277811, + "learning_rate": 3.5008655489306984e-08, + "loss": 0.3705, + "step": 8972 + }, + { + "epoch": 4.871335504885994, + "grad_norm": 10.470214013749253, + "learning_rate": 3.471525312991331e-08, + "loss": 0.3385, + "step": 8973 + }, + { + "epoch": 4.871878393051031, + "grad_norm": 22.377934873090744, + "learning_rate": 3.44230832928194e-08, + "loss": 0.6696, + "step": 8974 + }, + { + "epoch": 4.872421281216069, + "grad_norm": 10.012570917492981, + "learning_rate": 3.4132146014161925e-08, + "loss": 0.3796, + "step": 8975 + }, + { + "epoch": 4.872964169381108, + "grad_norm": 10.734857093886871, + "learning_rate": 3.38424413299232e-08, + "loss": 0.3492, + "step": 8976 + }, + { + "epoch": 4.873507057546146, + "grad_norm": 18.0435053595691, + "learning_rate": 3.3553969275935684e-08, + "loss": 0.5868, + "step": 8977 + }, + { + "epoch": 4.874049945711183, + "grad_norm": 10.021666813244016, + "learning_rate": 3.326672988787638e-08, + "loss": 0.2792, + "step": 8978 + }, + { + "epoch": 4.874592833876221, + "grad_norm": 11.363776145547698, + "learning_rate": 3.298072320127355e-08, + "loss": 0.2911, + "step": 8979 + }, + { + "epoch": 4.875135722041259, + "grad_norm": 16.234875062594668, + "learning_rate": 3.2695949251499994e-08, + "loss": 0.4444, + "step": 8980 + }, + { + "epoch": 4.875678610206298, + "grad_norm": 18.354132401016077, + "learning_rate": 3.2412408073776434e-08, + "loss": 0.8921, + "step": 8981 + }, + { + "epoch": 4.8762214983713354, + "grad_norm": 13.549633551421508, + "learning_rate": 3.2130099703171494e-08, + "loss": 0.4195, + "step": 8982 + }, + { + "epoch": 4.876764386536373, + "grad_norm": 9.25197339319036, + "learning_rate": 3.1849024174603895e-08, + "loss": 0.363, + "step": 8983 + }, + { + "epoch": 4.877307274701412, + "grad_norm": 14.423397162730842, + "learning_rate": 3.1569181522834724e-08, + "loss": 0.5159, + "step": 8984 + }, + { + "epoch": 4.87785016286645, + "grad_norm": 15.097447164063551, + "learning_rate": 3.1290571782476296e-08, + "loss": 0.5308, + "step": 8985 + }, + { + "epoch": 4.8783930510314875, + "grad_norm": 14.75727860602389, + "learning_rate": 3.101319498798883e-08, + "loss": 0.5018, + "step": 8986 + }, + { + "epoch": 4.878935939196525, + "grad_norm": 11.612303819226995, + "learning_rate": 3.073705117367709e-08, + "loss": 0.3928, + "step": 8987 + }, + { + "epoch": 4.879478827361563, + "grad_norm": 14.550764874493284, + "learning_rate": 3.046214037369488e-08, + "loss": 0.4143, + "step": 8988 + }, + { + "epoch": 4.880021715526602, + "grad_norm": 14.242337431144017, + "learning_rate": 3.0188462622045e-08, + "loss": 0.4503, + "step": 8989 + }, + { + "epoch": 4.8805646036916395, + "grad_norm": 14.734943195801437, + "learning_rate": 2.991601795257482e-08, + "loss": 0.3427, + "step": 8990 + }, + { + "epoch": 4.881107491856677, + "grad_norm": 14.592718045998431, + "learning_rate": 2.9644806398982928e-08, + "loss": 0.4573, + "step": 8991 + }, + { + "epoch": 4.881650380021716, + "grad_norm": 12.610312778876645, + "learning_rate": 2.9374827994810283e-08, + "loss": 0.4903, + "step": 8992 + }, + { + "epoch": 4.882193268186754, + "grad_norm": 12.128783807352118, + "learning_rate": 2.9106082773450172e-08, + "loss": 0.4459, + "step": 8993 + }, + { + "epoch": 4.8827361563517915, + "grad_norm": 17.48739197019621, + "learning_rate": 2.8838570768141562e-08, + "loss": 0.7127, + "step": 8994 + }, + { + "epoch": 4.883279044516829, + "grad_norm": 14.579178040970136, + "learning_rate": 2.8572292011969095e-08, + "loss": 0.3715, + "step": 8995 + }, + { + "epoch": 4.883821932681867, + "grad_norm": 11.025778405045106, + "learning_rate": 2.830724653786865e-08, + "loss": 0.2711, + "step": 8996 + }, + { + "epoch": 4.884364820846906, + "grad_norm": 20.20923605960106, + "learning_rate": 2.804343437861956e-08, + "loss": 0.6597, + "step": 8997 + }, + { + "epoch": 4.8849077090119435, + "grad_norm": 9.4089323273495, + "learning_rate": 2.7780855566852394e-08, + "loss": 0.2086, + "step": 8998 + }, + { + "epoch": 4.885450597176981, + "grad_norm": 10.97470388984354, + "learning_rate": 2.7519510135041173e-08, + "loss": 0.297, + "step": 8999 + }, + { + "epoch": 4.88599348534202, + "grad_norm": 12.570657915339005, + "learning_rate": 2.7259398115511148e-08, + "loss": 0.5147, + "step": 9000 + }, + { + "epoch": 4.886536373507058, + "grad_norm": 15.54569627943996, + "learning_rate": 2.7000519540434367e-08, + "loss": 0.4506, + "step": 9001 + }, + { + "epoch": 4.8870792616720955, + "grad_norm": 10.42963878683934, + "learning_rate": 2.6742874441826328e-08, + "loss": 0.3591, + "step": 9002 + }, + { + "epoch": 4.887622149837133, + "grad_norm": 12.875386611254111, + "learning_rate": 2.6486462851554874e-08, + "loss": 0.3134, + "step": 9003 + }, + { + "epoch": 4.888165038002171, + "grad_norm": 9.34456052093577, + "learning_rate": 2.6231284801333524e-08, + "loss": 0.2513, + "step": 9004 + }, + { + "epoch": 4.88870792616721, + "grad_norm": 12.286889941053696, + "learning_rate": 2.597734032272148e-08, + "loss": 0.258, + "step": 9005 + }, + { + "epoch": 4.8892508143322475, + "grad_norm": 18.670865914831435, + "learning_rate": 2.5724629447130278e-08, + "loss": 0.7677, + "step": 9006 + }, + { + "epoch": 4.889793702497285, + "grad_norm": 13.505630369786697, + "learning_rate": 2.547315220581159e-08, + "loss": 0.4204, + "step": 9007 + }, + { + "epoch": 4.890336590662324, + "grad_norm": 13.5847940651091, + "learning_rate": 2.5222908629871647e-08, + "loss": 0.3804, + "step": 9008 + }, + { + "epoch": 4.890879478827362, + "grad_norm": 18.30352602432072, + "learning_rate": 2.497389875025902e-08, + "loss": 0.7192, + "step": 9009 + }, + { + "epoch": 4.8914223669923995, + "grad_norm": 17.37999494699269, + "learning_rate": 2.4726122597773516e-08, + "loss": 0.6299, + "step": 9010 + }, + { + "epoch": 4.891965255157437, + "grad_norm": 10.851195927316295, + "learning_rate": 2.4479580203058408e-08, + "loss": 0.5514, + "step": 9011 + }, + { + "epoch": 4.892508143322475, + "grad_norm": 10.563754398632097, + "learning_rate": 2.423427159660707e-08, + "loss": 0.2451, + "step": 9012 + }, + { + "epoch": 4.893051031487514, + "grad_norm": 11.26826600207582, + "learning_rate": 2.39901968087608e-08, + "loss": 0.3286, + "step": 9013 + }, + { + "epoch": 4.8935939196525515, + "grad_norm": 17.973747561161268, + "learning_rate": 2.3747355869706555e-08, + "loss": 0.5601, + "step": 9014 + }, + { + "epoch": 4.894136807817589, + "grad_norm": 9.432719575335916, + "learning_rate": 2.35057488094792e-08, + "loss": 0.1905, + "step": 9015 + }, + { + "epoch": 4.894679695982628, + "grad_norm": 10.718766904219184, + "learning_rate": 2.3265375657960388e-08, + "loss": 0.4343, + "step": 9016 + }, + { + "epoch": 4.895222584147666, + "grad_norm": 13.174705545759618, + "learning_rate": 2.3026236444880778e-08, + "loss": 0.7121, + "step": 9017 + }, + { + "epoch": 4.8957654723127035, + "grad_norm": 14.581493576657595, + "learning_rate": 2.2788331199816715e-08, + "loss": 0.6438, + "step": 9018 + }, + { + "epoch": 4.896308360477741, + "grad_norm": 13.239127120829922, + "learning_rate": 2.255165995219355e-08, + "loss": 0.3997, + "step": 9019 + }, + { + "epoch": 4.896851248642779, + "grad_norm": 19.46295020201305, + "learning_rate": 2.2316222731282312e-08, + "loss": 0.4649, + "step": 9020 + }, + { + "epoch": 4.897394136807818, + "grad_norm": 10.875751867136424, + "learning_rate": 2.208201956620304e-08, + "loss": 0.3195, + "step": 9021 + }, + { + "epoch": 4.8979370249728555, + "grad_norm": 20.350514144081636, + "learning_rate": 2.184905048592256e-08, + "loss": 0.399, + "step": 9022 + }, + { + "epoch": 4.898479913137893, + "grad_norm": 14.514112929300751, + "learning_rate": 2.161731551925339e-08, + "loss": 0.3543, + "step": 9023 + }, + { + "epoch": 4.899022801302932, + "grad_norm": 14.478901017090632, + "learning_rate": 2.1386814694859258e-08, + "loss": 0.341, + "step": 9024 + }, + { + "epoch": 4.89956568946797, + "grad_norm": 13.744752043693394, + "learning_rate": 2.1157548041246258e-08, + "loss": 0.4873, + "step": 9025 + }, + { + "epoch": 4.900108577633008, + "grad_norm": 9.057494433239784, + "learning_rate": 2.09295155867717e-08, + "loss": 0.217, + "step": 9026 + }, + { + "epoch": 4.900651465798045, + "grad_norm": 15.4247048251968, + "learning_rate": 2.070271735963858e-08, + "loss": 0.5049, + "step": 9027 + }, + { + "epoch": 4.901194353963083, + "grad_norm": 11.973584500311773, + "learning_rate": 2.0477153387898908e-08, + "loss": 0.4135, + "step": 9028 + }, + { + "epoch": 4.901737242128122, + "grad_norm": 15.813841494616677, + "learning_rate": 2.0252823699449254e-08, + "loss": 0.5418, + "step": 9029 + }, + { + "epoch": 4.90228013029316, + "grad_norm": 12.85782607647325, + "learning_rate": 2.0029728322036312e-08, + "loss": 0.5075, + "step": 9030 + }, + { + "epoch": 4.902823018458197, + "grad_norm": 14.609268597455578, + "learning_rate": 1.9807867283251346e-08, + "loss": 0.633, + "step": 9031 + }, + { + "epoch": 4.903365906623236, + "grad_norm": 18.076004594690488, + "learning_rate": 1.958724061053574e-08, + "loss": 0.8642, + "step": 9032 + }, + { + "epoch": 4.903908794788274, + "grad_norm": 14.58349613179072, + "learning_rate": 1.9367848331176554e-08, + "loss": 0.4699, + "step": 9033 + }, + { + "epoch": 4.904451682953312, + "grad_norm": 8.68624975589409, + "learning_rate": 1.9149690472309856e-08, + "loss": 0.2521, + "step": 9034 + }, + { + "epoch": 4.904994571118349, + "grad_norm": 13.439742340445157, + "learning_rate": 1.8932767060915182e-08, + "loss": 0.5178, + "step": 9035 + }, + { + "epoch": 4.905537459283387, + "grad_norm": 10.992659888320869, + "learning_rate": 1.8717078123823286e-08, + "loss": 0.344, + "step": 9036 + }, + { + "epoch": 4.906080347448426, + "grad_norm": 13.160181116268273, + "learning_rate": 1.8502623687711717e-08, + "loss": 0.641, + "step": 9037 + }, + { + "epoch": 4.906623235613464, + "grad_norm": 12.762552089795363, + "learning_rate": 1.828940377910371e-08, + "loss": 0.3898, + "step": 9038 + }, + { + "epoch": 4.907166123778501, + "grad_norm": 9.15219473750769, + "learning_rate": 1.8077418424370387e-08, + "loss": 0.3201, + "step": 9039 + }, + { + "epoch": 4.90770901194354, + "grad_norm": 16.992302430875814, + "learning_rate": 1.7866667649730774e-08, + "loss": 0.5003, + "step": 9040 + }, + { + "epoch": 4.908251900108578, + "grad_norm": 13.106449507444296, + "learning_rate": 1.765715148125069e-08, + "loss": 0.4396, + "step": 9041 + }, + { + "epoch": 4.908794788273616, + "grad_norm": 17.602962269162354, + "learning_rate": 1.744886994484385e-08, + "loss": 0.5396, + "step": 9042 + }, + { + "epoch": 4.909337676438653, + "grad_norm": 12.698432190182416, + "learning_rate": 1.724182306627076e-08, + "loss": 0.3839, + "step": 9043 + }, + { + "epoch": 4.909880564603691, + "grad_norm": 18.435733381982295, + "learning_rate": 1.703601087113871e-08, + "loss": 0.523, + "step": 9044 + }, + { + "epoch": 4.91042345276873, + "grad_norm": 15.182522992411029, + "learning_rate": 1.6831433384904006e-08, + "loss": 0.4973, + "step": 9045 + }, + { + "epoch": 4.910966340933768, + "grad_norm": 16.90243953066863, + "learning_rate": 1.662809063286863e-08, + "loss": 0.419, + "step": 9046 + }, + { + "epoch": 4.911509229098805, + "grad_norm": 15.579122908193515, + "learning_rate": 1.642598264018025e-08, + "loss": 0.4699, + "step": 9047 + }, + { + "epoch": 4.912052117263844, + "grad_norm": 11.883257189390664, + "learning_rate": 1.6225109431839968e-08, + "loss": 0.387, + "step": 9048 + }, + { + "epoch": 4.912595005428882, + "grad_norm": 12.57956767087257, + "learning_rate": 1.6025471032689034e-08, + "loss": 0.3389, + "step": 9049 + }, + { + "epoch": 4.91313789359392, + "grad_norm": 10.023724036259201, + "learning_rate": 1.582706746741991e-08, + "loss": 0.181, + "step": 9050 + }, + { + "epoch": 4.913680781758957, + "grad_norm": 10.904566382599436, + "learning_rate": 1.562989876057075e-08, + "loss": 0.2939, + "step": 9051 + }, + { + "epoch": 4.914223669923995, + "grad_norm": 15.131794978062635, + "learning_rate": 1.5433964936529823e-08, + "loss": 0.5318, + "step": 9052 + }, + { + "epoch": 4.914766558089034, + "grad_norm": 11.260253598807191, + "learning_rate": 1.5239266019527743e-08, + "loss": 0.2825, + "step": 9053 + }, + { + "epoch": 4.915309446254072, + "grad_norm": 8.783484863701455, + "learning_rate": 1.504580203364747e-08, + "loss": 0.2619, + "step": 9054 + }, + { + "epoch": 4.915852334419109, + "grad_norm": 17.289518463767752, + "learning_rate": 1.4853573002815425e-08, + "loss": 0.6613, + "step": 9055 + }, + { + "epoch": 4.916395222584148, + "grad_norm": 10.577132965088307, + "learning_rate": 1.4662578950808137e-08, + "loss": 0.34, + "step": 9056 + }, + { + "epoch": 4.916938110749186, + "grad_norm": 13.481640817817254, + "learning_rate": 1.4472819901246715e-08, + "loss": 0.3756, + "step": 9057 + }, + { + "epoch": 4.917480998914224, + "grad_norm": 14.814161562886865, + "learning_rate": 1.4284295877602383e-08, + "loss": 0.5217, + "step": 9058 + }, + { + "epoch": 4.918023887079261, + "grad_norm": 13.911904587945742, + "learning_rate": 1.4097006903190935e-08, + "loss": 0.4312, + "step": 9059 + }, + { + "epoch": 4.918566775244299, + "grad_norm": 13.559974799811506, + "learning_rate": 1.3910953001176063e-08, + "loss": 0.5598, + "step": 9060 + }, + { + "epoch": 4.919109663409338, + "grad_norm": 12.999506118739305, + "learning_rate": 1.3726134194570472e-08, + "loss": 0.4866, + "step": 9061 + }, + { + "epoch": 4.919652551574376, + "grad_norm": 12.16516581528143, + "learning_rate": 1.3542550506232543e-08, + "loss": 0.4469, + "step": 9062 + }, + { + "epoch": 4.920195439739413, + "grad_norm": 13.76187292005182, + "learning_rate": 1.3360201958867447e-08, + "loss": 0.5088, + "step": 9063 + }, + { + "epoch": 4.920738327904452, + "grad_norm": 11.597669183780656, + "learning_rate": 1.3179088575029365e-08, + "loss": 0.3892, + "step": 9064 + }, + { + "epoch": 4.92128121606949, + "grad_norm": 13.03473175291288, + "learning_rate": 1.2999210377118155e-08, + "loss": 0.3606, + "step": 9065 + }, + { + "epoch": 4.921824104234528, + "grad_norm": 11.618399948981745, + "learning_rate": 1.2820567387381577e-08, + "loss": 0.431, + "step": 9066 + }, + { + "epoch": 4.922366992399565, + "grad_norm": 14.444651618891138, + "learning_rate": 1.2643159627914182e-08, + "loss": 0.4671, + "step": 9067 + }, + { + "epoch": 4.922909880564603, + "grad_norm": 19.413144717406254, + "learning_rate": 1.2466987120658413e-08, + "loss": 0.4479, + "step": 9068 + }, + { + "epoch": 4.923452768729642, + "grad_norm": 15.188364084941032, + "learning_rate": 1.2292049887403513e-08, + "loss": 0.518, + "step": 9069 + }, + { + "epoch": 4.92399565689468, + "grad_norm": 17.595339652436596, + "learning_rate": 1.2118347949785503e-08, + "loss": 0.5491, + "step": 9070 + }, + { + "epoch": 4.924538545059717, + "grad_norm": 13.811004803619682, + "learning_rate": 1.1945881329289422e-08, + "loss": 0.521, + "step": 9071 + }, + { + "epoch": 4.925081433224756, + "grad_norm": 17.210661605650973, + "learning_rate": 1.1774650047244873e-08, + "loss": 0.4881, + "step": 9072 + }, + { + "epoch": 4.925624321389794, + "grad_norm": 11.74172359244011, + "learning_rate": 1.1604654124830472e-08, + "loss": 0.2734, + "step": 9073 + }, + { + "epoch": 4.926167209554832, + "grad_norm": 12.999583685539983, + "learning_rate": 1.1435893583071623e-08, + "loss": 0.4139, + "step": 9074 + }, + { + "epoch": 4.9267100977198695, + "grad_norm": 18.891052821295602, + "learning_rate": 1.1268368442840516e-08, + "loss": 0.6411, + "step": 9075 + }, + { + "epoch": 4.927252985884907, + "grad_norm": 9.732481550986785, + "learning_rate": 1.1102078724857247e-08, + "loss": 0.2768, + "step": 9076 + }, + { + "epoch": 4.927795874049946, + "grad_norm": 13.879976567786102, + "learning_rate": 1.0937024449688693e-08, + "loss": 0.3334, + "step": 9077 + }, + { + "epoch": 4.928338762214984, + "grad_norm": 14.2590437939657, + "learning_rate": 1.0773205637749639e-08, + "loss": 0.5803, + "step": 9078 + }, + { + "epoch": 4.9288816503800215, + "grad_norm": 12.223213148169753, + "learning_rate": 1.0610622309300544e-08, + "loss": 0.3341, + "step": 9079 + }, + { + "epoch": 4.92942453854506, + "grad_norm": 9.799661483908018, + "learning_rate": 1.0449274484450877e-08, + "loss": 0.2622, + "step": 9080 + }, + { + "epoch": 4.929967426710098, + "grad_norm": 12.782439661304597, + "learning_rate": 1.0289162183155788e-08, + "loss": 0.4998, + "step": 9081 + }, + { + "epoch": 4.930510314875136, + "grad_norm": 17.86302762326864, + "learning_rate": 1.0130285425218322e-08, + "loss": 0.6778, + "step": 9082 + }, + { + "epoch": 4.9310532030401735, + "grad_norm": 7.999286008192372, + "learning_rate": 9.97264423028832e-09, + "loss": 0.2896, + "step": 9083 + }, + { + "epoch": 4.931596091205211, + "grad_norm": 18.706243688121948, + "learning_rate": 9.816238617864626e-09, + "loss": 0.5758, + "step": 9084 + }, + { + "epoch": 4.93213897937025, + "grad_norm": 12.391173965997686, + "learning_rate": 9.661068607288437e-09, + "loss": 0.2557, + "step": 9085 + }, + { + "epoch": 4.932681867535288, + "grad_norm": 13.206506481368905, + "learning_rate": 9.50713421775551e-09, + "loss": 0.4463, + "step": 9086 + }, + { + "epoch": 4.9332247557003255, + "grad_norm": 12.900701770071388, + "learning_rate": 9.354435468301726e-09, + "loss": 0.459, + "step": 9087 + }, + { + "epoch": 4.933767643865364, + "grad_norm": 12.068288986888435, + "learning_rate": 9.202972377814201e-09, + "loss": 0.3485, + "step": 9088 + }, + { + "epoch": 4.934310532030402, + "grad_norm": 12.75825787671443, + "learning_rate": 9.052744965025728e-09, + "loss": 0.4755, + "step": 9089 + }, + { + "epoch": 4.93485342019544, + "grad_norm": 11.99221513373725, + "learning_rate": 8.903753248516999e-09, + "loss": 0.3729, + "step": 9090 + }, + { + "epoch": 4.9353963083604775, + "grad_norm": 12.257794064544841, + "learning_rate": 8.755997246715498e-09, + "loss": 0.3006, + "step": 9091 + }, + { + "epoch": 4.935939196525515, + "grad_norm": 11.554301500218191, + "learning_rate": 8.609476977896602e-09, + "loss": 0.4357, + "step": 9092 + }, + { + "epoch": 4.936482084690554, + "grad_norm": 12.78413585140587, + "learning_rate": 8.464192460181376e-09, + "loss": 0.3639, + "step": 9093 + }, + { + "epoch": 4.937024972855592, + "grad_norm": 12.250703901070336, + "learning_rate": 8.320143711538775e-09, + "loss": 0.3514, + "step": 9094 + }, + { + "epoch": 4.9375678610206295, + "grad_norm": 14.649810047012926, + "learning_rate": 8.17733074978455e-09, + "loss": 0.3402, + "step": 9095 + }, + { + "epoch": 4.938110749185668, + "grad_norm": 17.43504427698499, + "learning_rate": 8.035753592583461e-09, + "loss": 0.5377, + "step": 9096 + }, + { + "epoch": 4.938653637350706, + "grad_norm": 14.034004629213161, + "learning_rate": 7.895412257444834e-09, + "loss": 0.397, + "step": 9097 + }, + { + "epoch": 4.939196525515744, + "grad_norm": 11.963493138309827, + "learning_rate": 7.756306761725895e-09, + "loss": 0.2919, + "step": 9098 + }, + { + "epoch": 4.9397394136807815, + "grad_norm": 15.061307897752748, + "learning_rate": 7.618437122632882e-09, + "loss": 0.6577, + "step": 9099 + }, + { + "epoch": 4.940282301845819, + "grad_norm": 14.852941299011642, + "learning_rate": 7.48180335721771e-09, + "loss": 0.5519, + "step": 9100 + }, + { + "epoch": 4.940825190010858, + "grad_norm": 12.814384068391547, + "learning_rate": 7.346405482377972e-09, + "loss": 0.5925, + "step": 9101 + }, + { + "epoch": 4.941368078175896, + "grad_norm": 9.563813812350007, + "learning_rate": 7.2122435148613835e-09, + "loss": 0.2204, + "step": 9102 + }, + { + "epoch": 4.9419109663409335, + "grad_norm": 9.564014825276837, + "learning_rate": 7.079317471260228e-09, + "loss": 0.2511, + "step": 9103 + }, + { + "epoch": 4.942453854505972, + "grad_norm": 17.38602510454561, + "learning_rate": 6.947627368015797e-09, + "loss": 0.5656, + "step": 9104 + }, + { + "epoch": 4.94299674267101, + "grad_norm": 9.577450195791936, + "learning_rate": 6.817173221415063e-09, + "loss": 0.2404, + "step": 9105 + }, + { + "epoch": 4.943539630836048, + "grad_norm": 11.92192298760108, + "learning_rate": 6.687955047592898e-09, + "loss": 0.3116, + "step": 9106 + }, + { + "epoch": 4.9440825190010855, + "grad_norm": 14.111860080166249, + "learning_rate": 6.559972862530961e-09, + "loss": 0.4142, + "step": 9107 + }, + { + "epoch": 4.944625407166123, + "grad_norm": 16.85584479160711, + "learning_rate": 6.433226682059923e-09, + "loss": 0.4438, + "step": 9108 + }, + { + "epoch": 4.945168295331162, + "grad_norm": 12.407679682981549, + "learning_rate": 6.307716521853912e-09, + "loss": 0.7122, + "step": 9109 + }, + { + "epoch": 4.9457111834962, + "grad_norm": 18.57218994539643, + "learning_rate": 6.183442397437178e-09, + "loss": 0.4333, + "step": 9110 + }, + { + "epoch": 4.9462540716612375, + "grad_norm": 10.179606005226988, + "learning_rate": 6.0604043241807574e-09, + "loss": 0.2016, + "step": 9111 + }, + { + "epoch": 4.946796959826276, + "grad_norm": 11.47232407870111, + "learning_rate": 5.938602317301368e-09, + "loss": 0.4277, + "step": 9112 + }, + { + "epoch": 4.947339847991314, + "grad_norm": 11.739066706652215, + "learning_rate": 5.8180363918625135e-09, + "loss": 0.5738, + "step": 9113 + }, + { + "epoch": 4.947882736156352, + "grad_norm": 10.786798778444865, + "learning_rate": 5.698706562778933e-09, + "loss": 0.3646, + "step": 9114 + }, + { + "epoch": 4.9484256243213895, + "grad_norm": 14.9560917250246, + "learning_rate": 5.580612844806599e-09, + "loss": 0.4187, + "step": 9115 + }, + { + "epoch": 4.948968512486427, + "grad_norm": 8.906425035542007, + "learning_rate": 5.4637552525538266e-09, + "loss": 0.2081, + "step": 9116 + }, + { + "epoch": 4.949511400651466, + "grad_norm": 15.606788834694017, + "learning_rate": 5.3481338004723886e-09, + "loss": 0.533, + "step": 9117 + }, + { + "epoch": 4.950054288816504, + "grad_norm": 12.28419381741561, + "learning_rate": 5.233748502861957e-09, + "loss": 0.5334, + "step": 9118 + }, + { + "epoch": 4.950597176981542, + "grad_norm": 12.227392912720493, + "learning_rate": 5.120599373872326e-09, + "loss": 0.2228, + "step": 9119 + }, + { + "epoch": 4.95114006514658, + "grad_norm": 16.11553058072441, + "learning_rate": 5.008686427495635e-09, + "loss": 0.4372, + "step": 9120 + }, + { + "epoch": 4.951682953311618, + "grad_norm": 13.761040611076448, + "learning_rate": 4.898009677574145e-09, + "loss": 0.4973, + "step": 9121 + }, + { + "epoch": 4.952225841476656, + "grad_norm": 10.73617260071943, + "learning_rate": 4.788569137796905e-09, + "loss": 0.4157, + "step": 9122 + }, + { + "epoch": 4.952768729641694, + "grad_norm": 20.53036734807641, + "learning_rate": 4.6803648216986465e-09, + "loss": 0.3275, + "step": 9123 + }, + { + "epoch": 4.953311617806731, + "grad_norm": 10.707480347228744, + "learning_rate": 4.5733967426642155e-09, + "loss": 0.3827, + "step": 9124 + }, + { + "epoch": 4.95385450597177, + "grad_norm": 17.496584887499267, + "learning_rate": 4.467664913921921e-09, + "loss": 0.7894, + "step": 9125 + }, + { + "epoch": 4.954397394136808, + "grad_norm": 13.01334205119194, + "learning_rate": 4.363169348549079e-09, + "loss": 0.2762, + "step": 9126 + }, + { + "epoch": 4.954940282301846, + "grad_norm": 19.033412027447728, + "learning_rate": 4.259910059470906e-09, + "loss": 0.4458, + "step": 9127 + }, + { + "epoch": 4.955483170466884, + "grad_norm": 12.446627215988766, + "learning_rate": 4.1578870594571885e-09, + "loss": 0.3732, + "step": 9128 + }, + { + "epoch": 4.956026058631922, + "grad_norm": 11.854175020272, + "learning_rate": 4.05710036112672e-09, + "loss": 0.412, + "step": 9129 + }, + { + "epoch": 4.95656894679696, + "grad_norm": 13.145274907827265, + "learning_rate": 3.957549976946195e-09, + "loss": 0.376, + "step": 9130 + }, + { + "epoch": 4.957111834961998, + "grad_norm": 14.159905703597753, + "learning_rate": 3.859235919226878e-09, + "loss": 0.4783, + "step": 9131 + }, + { + "epoch": 4.957654723127035, + "grad_norm": 10.76209944745829, + "learning_rate": 3.76215820012793e-09, + "loss": 0.3364, + "step": 9132 + }, + { + "epoch": 4.958197611292074, + "grad_norm": 18.14425836712623, + "learning_rate": 3.666316831657524e-09, + "loss": 0.3815, + "step": 9133 + }, + { + "epoch": 4.958740499457112, + "grad_norm": 9.071986096860721, + "learning_rate": 3.5717118256684004e-09, + "loss": 0.3018, + "step": 9134 + }, + { + "epoch": 4.95928338762215, + "grad_norm": 11.250761445029084, + "learning_rate": 3.4783431938623103e-09, + "loss": 0.3972, + "step": 9135 + }, + { + "epoch": 4.959826275787188, + "grad_norm": 14.397121807180921, + "learning_rate": 3.3862109477877935e-09, + "loss": 0.6723, + "step": 9136 + }, + { + "epoch": 4.960369163952226, + "grad_norm": 13.645836729945032, + "learning_rate": 3.295315098837959e-09, + "loss": 0.3039, + "step": 9137 + }, + { + "epoch": 4.960912052117264, + "grad_norm": 13.683432725518449, + "learning_rate": 3.2056556582549247e-09, + "loss": 0.4265, + "step": 9138 + }, + { + "epoch": 4.961454940282302, + "grad_norm": 13.894633750793059, + "learning_rate": 3.117232637130929e-09, + "loss": 0.3269, + "step": 9139 + }, + { + "epoch": 4.961997828447339, + "grad_norm": 12.71603084534765, + "learning_rate": 3.0300460463983384e-09, + "loss": 0.4258, + "step": 9140 + }, + { + "epoch": 4.962540716612378, + "grad_norm": 12.764000039058175, + "learning_rate": 2.944095896844079e-09, + "loss": 0.2995, + "step": 9141 + }, + { + "epoch": 4.963083604777416, + "grad_norm": 9.50398515456868, + "learning_rate": 2.8593821990963167e-09, + "loss": 0.3252, + "step": 9142 + }, + { + "epoch": 4.963626492942454, + "grad_norm": 10.26064116609969, + "learning_rate": 2.7759049636322254e-09, + "loss": 0.4138, + "step": 9143 + }, + { + "epoch": 4.964169381107492, + "grad_norm": 14.706063028722339, + "learning_rate": 2.6936642007791004e-09, + "loss": 0.3732, + "step": 9144 + }, + { + "epoch": 4.96471226927253, + "grad_norm": 16.446836987033343, + "learning_rate": 2.612659920705474e-09, + "loss": 0.6265, + "step": 9145 + }, + { + "epoch": 4.965255157437568, + "grad_norm": 13.694193607083827, + "learning_rate": 2.5328921334322186e-09, + "loss": 0.3205, + "step": 9146 + }, + { + "epoch": 4.965798045602606, + "grad_norm": 21.42453240685977, + "learning_rate": 2.4543608488247774e-09, + "loss": 0.5323, + "step": 9147 + }, + { + "epoch": 4.966340933767643, + "grad_norm": 17.278512554654093, + "learning_rate": 2.3770660765953803e-09, + "loss": 0.4915, + "step": 9148 + }, + { + "epoch": 4.966883821932682, + "grad_norm": 11.488248878514298, + "learning_rate": 2.301007826304158e-09, + "loss": 0.4663, + "step": 9149 + }, + { + "epoch": 4.96742671009772, + "grad_norm": 12.059564726501527, + "learning_rate": 2.2261861073580306e-09, + "loss": 0.354, + "step": 9150 + }, + { + "epoch": 4.967969598262758, + "grad_norm": 12.062263452100884, + "learning_rate": 2.1526009290118165e-09, + "loss": 0.4141, + "step": 9151 + }, + { + "epoch": 4.968512486427796, + "grad_norm": 13.560790820046249, + "learning_rate": 2.0802523003660146e-09, + "loss": 0.4742, + "step": 9152 + }, + { + "epoch": 4.969055374592834, + "grad_norm": 12.550658712203877, + "learning_rate": 2.009140230369022e-09, + "loss": 0.5253, + "step": 9153 + }, + { + "epoch": 4.969598262757872, + "grad_norm": 12.71176738765954, + "learning_rate": 1.939264727816026e-09, + "loss": 0.4024, + "step": 9154 + }, + { + "epoch": 4.97014115092291, + "grad_norm": 11.734799781581923, + "learning_rate": 1.870625801349002e-09, + "loss": 0.2679, + "step": 9155 + }, + { + "epoch": 4.970684039087947, + "grad_norm": 13.891313537034907, + "learning_rate": 1.8032234594578257e-09, + "loss": 0.2977, + "step": 9156 + }, + { + "epoch": 4.971226927252986, + "grad_norm": 13.309663904879017, + "learning_rate": 1.7370577104780518e-09, + "loss": 0.6025, + "step": 9157 + }, + { + "epoch": 4.971769815418024, + "grad_norm": 20.714422472932196, + "learning_rate": 1.672128562594244e-09, + "loss": 0.7059, + "step": 9158 + }, + { + "epoch": 4.972312703583062, + "grad_norm": 16.889615399656492, + "learning_rate": 1.6084360238366458e-09, + "loss": 0.5564, + "step": 9159 + }, + { + "epoch": 4.9728555917481, + "grad_norm": 16.399562625707034, + "learning_rate": 1.5459801020833999e-09, + "loss": 0.5455, + "step": 9160 + }, + { + "epoch": 4.973398479913138, + "grad_norm": 13.826143902617954, + "learning_rate": 1.484760805057217e-09, + "loss": 0.4126, + "step": 9161 + }, + { + "epoch": 4.973941368078176, + "grad_norm": 11.172428599335714, + "learning_rate": 1.4247781403320393e-09, + "loss": 0.4096, + "step": 9162 + }, + { + "epoch": 4.974484256243214, + "grad_norm": 16.468489977898997, + "learning_rate": 1.3660321153252665e-09, + "loss": 0.4705, + "step": 9163 + }, + { + "epoch": 4.9750271444082514, + "grad_norm": 14.439166503302797, + "learning_rate": 1.3085227373033082e-09, + "loss": 0.742, + "step": 9164 + }, + { + "epoch": 4.97557003257329, + "grad_norm": 13.576277807739485, + "learning_rate": 1.2522500133782533e-09, + "loss": 0.3955, + "step": 9165 + }, + { + "epoch": 4.976112920738328, + "grad_norm": 10.042503188679653, + "learning_rate": 1.1972139505112002e-09, + "loss": 0.2315, + "step": 9166 + }, + { + "epoch": 4.976655808903366, + "grad_norm": 15.56377868085397, + "learning_rate": 1.1434145555078157e-09, + "loss": 0.6752, + "step": 9167 + }, + { + "epoch": 4.977198697068404, + "grad_norm": 14.579485607298796, + "learning_rate": 1.0908518350238873e-09, + "loss": 0.4866, + "step": 9168 + }, + { + "epoch": 4.977741585233442, + "grad_norm": 16.774796454586042, + "learning_rate": 1.0395257955586602e-09, + "loss": 0.5845, + "step": 9169 + }, + { + "epoch": 4.97828447339848, + "grad_norm": 11.815529585109633, + "learning_rate": 9.894364434603899e-10, + "loss": 0.4916, + "step": 9170 + }, + { + "epoch": 4.978827361563518, + "grad_norm": 14.632698898536018, + "learning_rate": 9.405837849252308e-10, + "loss": 0.4883, + "step": 9171 + }, + { + "epoch": 4.9793702497285555, + "grad_norm": 14.57532070039393, + "learning_rate": 8.929678259939067e-10, + "loss": 0.3583, + "step": 9172 + }, + { + "epoch": 4.979913137893594, + "grad_norm": 18.102220051162018, + "learning_rate": 8.465885725572609e-10, + "loss": 0.5808, + "step": 9173 + }, + { + "epoch": 4.980456026058632, + "grad_norm": 15.979995931593416, + "learning_rate": 8.014460303507054e-10, + "loss": 0.2909, + "step": 9174 + }, + { + "epoch": 4.98099891422367, + "grad_norm": 12.010762984710725, + "learning_rate": 7.57540204957552e-10, + "loss": 0.4328, + "step": 9175 + }, + { + "epoch": 4.981541802388708, + "grad_norm": 9.665802442722407, + "learning_rate": 7.148711018090115e-10, + "loss": 0.3006, + "step": 9176 + }, + { + "epoch": 4.982084690553746, + "grad_norm": 12.841279252555506, + "learning_rate": 6.734387261808639e-10, + "loss": 0.4085, + "step": 9177 + }, + { + "epoch": 4.982627578718784, + "grad_norm": 17.41329959247634, + "learning_rate": 6.332430831978986e-10, + "loss": 0.4756, + "step": 9178 + }, + { + "epoch": 4.983170466883822, + "grad_norm": 15.463364790372648, + "learning_rate": 5.942841778328045e-10, + "loss": 0.4864, + "step": 9179 + }, + { + "epoch": 4.9837133550488595, + "grad_norm": 18.266485807419585, + "learning_rate": 5.565620149039497e-10, + "loss": 0.5961, + "step": 9180 + }, + { + "epoch": 4.984256243213898, + "grad_norm": 16.902373526165416, + "learning_rate": 5.200765990753809e-10, + "loss": 0.5701, + "step": 9181 + }, + { + "epoch": 4.984799131378936, + "grad_norm": 19.631248888899865, + "learning_rate": 4.848279348601547e-10, + "loss": 0.6822, + "step": 9182 + }, + { + "epoch": 4.985342019543974, + "grad_norm": 12.12267583241551, + "learning_rate": 4.5081602661922743e-10, + "loss": 0.4016, + "step": 9183 + }, + { + "epoch": 4.985884907709012, + "grad_norm": 13.057551391655595, + "learning_rate": 4.180408785570134e-10, + "loss": 0.415, + "step": 9184 + }, + { + "epoch": 4.98642779587405, + "grad_norm": 17.345548037710916, + "learning_rate": 3.8650249472915735e-10, + "loss": 0.6871, + "step": 9185 + }, + { + "epoch": 4.986970684039088, + "grad_norm": 13.256176178212948, + "learning_rate": 3.562008790358729e-10, + "loss": 0.5344, + "step": 9186 + }, + { + "epoch": 4.987513572204126, + "grad_norm": 14.72682874967264, + "learning_rate": 3.271360352252728e-10, + "loss": 0.4598, + "step": 9187 + }, + { + "epoch": 4.9880564603691635, + "grad_norm": 11.832103539399576, + "learning_rate": 2.993079668911492e-10, + "loss": 0.314, + "step": 9188 + }, + { + "epoch": 4.988599348534202, + "grad_norm": 16.48759487025403, + "learning_rate": 2.7271667747519327e-10, + "loss": 0.4666, + "step": 9189 + }, + { + "epoch": 4.98914223669924, + "grad_norm": 11.438831531197014, + "learning_rate": 2.473621702669959e-10, + "loss": 0.3144, + "step": 9190 + }, + { + "epoch": 4.989685124864278, + "grad_norm": 16.2603783146197, + "learning_rate": 2.2324444840293723e-10, + "loss": 0.4698, + "step": 9191 + }, + { + "epoch": 4.990228013029316, + "grad_norm": 13.65337346906678, + "learning_rate": 2.003635148650762e-10, + "loss": 0.296, + "step": 9192 + }, + { + "epoch": 4.990770901194354, + "grad_norm": 13.609445308581654, + "learning_rate": 1.7871937248337134e-10, + "loss": 0.411, + "step": 9193 + }, + { + "epoch": 4.991313789359392, + "grad_norm": 16.227775980791364, + "learning_rate": 1.5831202393457035e-10, + "loss": 0.3289, + "step": 9194 + }, + { + "epoch": 4.99185667752443, + "grad_norm": 18.790945613812905, + "learning_rate": 1.3914147174332037e-10, + "loss": 0.6441, + "step": 9195 + }, + { + "epoch": 4.9923995656894675, + "grad_norm": 11.25550609886848, + "learning_rate": 1.2120771828105783e-10, + "loss": 0.3092, + "step": 9196 + }, + { + "epoch": 4.992942453854506, + "grad_norm": 11.398167541292397, + "learning_rate": 1.0451076576489805e-10, + "loss": 0.2813, + "step": 9197 + }, + { + "epoch": 4.993485342019544, + "grad_norm": 11.282564692027671, + "learning_rate": 8.905061625985589e-11, + "loss": 0.3401, + "step": 9198 + }, + { + "epoch": 4.994028230184582, + "grad_norm": 20.099380241860953, + "learning_rate": 7.482727167884563e-11, + "loss": 0.5747, + "step": 9199 + }, + { + "epoch": 4.99457111834962, + "grad_norm": 11.476742325586816, + "learning_rate": 6.184073378046051e-11, + "loss": 0.2744, + "step": 9200 + }, + { + "epoch": 4.995114006514658, + "grad_norm": 9.30054079035995, + "learning_rate": 5.009100417119328e-11, + "loss": 0.1959, + "step": 9201 + }, + { + "epoch": 4.995656894679696, + "grad_norm": 17.192495762416968, + "learning_rate": 3.957808430432586e-11, + "loss": 0.4294, + "step": 9202 + }, + { + "epoch": 4.996199782844734, + "grad_norm": 15.84133950781789, + "learning_rate": 3.030197548103964e-11, + "loss": 0.4281, + "step": 9203 + }, + { + "epoch": 4.9967426710097715, + "grad_norm": 11.498444122453641, + "learning_rate": 2.226267884597455e-11, + "loss": 0.2538, + "step": 9204 + }, + { + "epoch": 4.99728555917481, + "grad_norm": 10.709506205461455, + "learning_rate": 1.5460195396110878e-11, + "loss": 0.3601, + "step": 9205 + }, + { + "epoch": 4.997828447339848, + "grad_norm": 11.943394957532275, + "learning_rate": 9.894525970777224e-12, + "loss": 0.2582, + "step": 9206 + }, + { + "epoch": 4.998371335504886, + "grad_norm": 18.015533305967693, + "learning_rate": 5.5656712605323126e-12, + "loss": 0.5477, + "step": 9207 + }, + { + "epoch": 4.998914223669924, + "grad_norm": 13.226875446527242, + "learning_rate": 2.4736317982831936e-12, + "loss": 0.4569, + "step": 9208 + }, + { + "epoch": 4.999457111834962, + "grad_norm": 13.424018021065068, + "learning_rate": 6.184079692772571e-13, + "loss": 0.3316, + "step": 9209 + }, + { + "epoch": 5.0, + "grad_norm": 15.798499583731147, + "learning_rate": 0.0, + "loss": 0.5066, + "step": 9210 + }, + { + "epoch": 5.0, + "step": 9210, + "total_flos": 1.098902245736448e+16, + "train_loss": 0.8259544500585617, + "train_runtime": 18528.2554, + "train_samples_per_second": 1.988, + "train_steps_per_second": 0.497 + } + ], + "logging_steps": 1.0, + "max_steps": 9210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.098902245736448e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/llava-next-video-7b_lora-True_qlora-False/training_args.bin b/llava-next-video-7b_lora-True_qlora-False/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..b16de63f342ea53d59c4d97aa8a3bd463a79531f --- /dev/null +++ b/llava-next-video-7b_lora-True_qlora-False/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9801fdd09c6ba798e12b77f77221db3b5747b6e1296785789d44cf2bc492ff9e +size 7160